Merge commit '29ce831000081dd757d3116bf774aafffc4b6b20' into next
* commit '29ce831000081dd757d3116bf774aafffc4b6b20': (34 commits) rcu: provide rcu_virt_note_context_switch() function. rcu: get rid of signed overflow in check_cpu_stall() rcu: optimize rcutiny rcu: prevent call_rcu() from diving into rcu core if irqs disabled rcu: further lower priority in rcu_yield() rcu: introduce kfree_rcu() rcu: fix spelling rcu: call __rcu_read_unlock() in exit_rcu for tree RCU rcu: Converge TINY_RCU expedited and normal boosting rcu: remove useless ->boosted_this_gp field rcu: code cleanups in TINY_RCU priority boosting. rcu: Switch to this_cpu() primitives rcu: Use WARN_ON_ONCE for DEBUG_OBJECTS_RCU_HEAD warnings rcu: mark rcutorture boosting callback as being on-stack rcu: add DEBUG_OBJECTS_RCU_HEAD check for alignment rcu: Enable DEBUG_OBJECTS_RCU_HEAD from !PREEMPT rcu: Add forward-progress diagnostic for per-CPU kthreads rcu: add grace-period age and more kthread state to tracing rcu: fix tracing bug thinko on boost-balk attribution rcu: update tracing documentation for new rcutorture and rcuboost ... Pulling in rcu_virt_note_context_switch(). Signed-off-by: Avi Kivity <avi@redhat.com> * commit '29ce831000081dd757d3116bf774aafffc4b6b20': (34 commits) rcu: provide rcu_virt_note_context_switch() function. rcu: get rid of signed overflow in check_cpu_stall() rcu: optimize rcutiny rcu: prevent call_rcu() from diving into rcu core if irqs disabled rcu: further lower priority in rcu_yield() rcu: introduce kfree_rcu() rcu: fix spelling rcu: call __rcu_read_unlock() in exit_rcu for tree RCU rcu: Converge TINY_RCU expedited and normal boosting rcu: remove useless ->boosted_this_gp field rcu: code cleanups in TINY_RCU priority boosting. rcu: Switch to this_cpu() primitives rcu: Use WARN_ON_ONCE for DEBUG_OBJECTS_RCU_HEAD warnings rcu: mark rcutorture boosting callback as being on-stack rcu: add DEBUG_OBJECTS_RCU_HEAD check for alignment rcu: Enable DEBUG_OBJECTS_RCU_HEAD from !PREEMPT rcu: Add forward-progress diagnostic for per-CPU kthreads rcu: add grace-period age and more kthread state to tracing rcu: fix tracing bug thinko on boost-balk attribution rcu: update tracing documentation for new rcutorture and rcuboost ...
This commit is contained in:
@@ -21,7 +21,7 @@ rcu.txt
|
|||||||
RTFP.txt
|
RTFP.txt
|
||||||
- List of RCU papers (bibliography) going back to 1980.
|
- List of RCU papers (bibliography) going back to 1980.
|
||||||
stallwarn.txt
|
stallwarn.txt
|
||||||
- RCU CPU stall warnings (CONFIG_RCU_CPU_STALL_DETECTOR)
|
- RCU CPU stall warnings (module parameter rcu_cpu_stall_suppress)
|
||||||
torture.txt
|
torture.txt
|
||||||
- RCU Torture Test Operation (CONFIG_RCU_TORTURE_TEST)
|
- RCU Torture Test Operation (CONFIG_RCU_TORTURE_TEST)
|
||||||
trace.txt
|
trace.txt
|
||||||
|
@@ -1,22 +1,25 @@
|
|||||||
Using RCU's CPU Stall Detector
|
Using RCU's CPU Stall Detector
|
||||||
|
|
||||||
The CONFIG_RCU_CPU_STALL_DETECTOR kernel config parameter enables
|
The rcu_cpu_stall_suppress module parameter enables RCU's CPU stall
|
||||||
RCU's CPU stall detector, which detects conditions that unduly delay
|
detector, which detects conditions that unduly delay RCU grace periods.
|
||||||
RCU grace periods. The stall detector's idea of what constitutes
|
This module parameter enables CPU stall detection by default, but
|
||||||
"unduly delayed" is controlled by a set of C preprocessor macros:
|
may be overridden via boot-time parameter or at runtime via sysfs.
|
||||||
|
The stall detector's idea of what constitutes "unduly delayed" is
|
||||||
|
controlled by a set of kernel configuration variables and cpp macros:
|
||||||
|
|
||||||
RCU_SECONDS_TILL_STALL_CHECK
|
CONFIG_RCU_CPU_STALL_TIMEOUT
|
||||||
|
|
||||||
This macro defines the period of time that RCU will wait from
|
This kernel configuration parameter defines the period of time
|
||||||
the beginning of a grace period until it issues an RCU CPU
|
that RCU will wait from the beginning of a grace period until it
|
||||||
stall warning. This time period is normally ten seconds.
|
issues an RCU CPU stall warning. This time period is normally
|
||||||
|
ten seconds.
|
||||||
|
|
||||||
RCU_SECONDS_TILL_STALL_RECHECK
|
RCU_SECONDS_TILL_STALL_RECHECK
|
||||||
|
|
||||||
This macro defines the period of time that RCU will wait after
|
This macro defines the period of time that RCU will wait after
|
||||||
issuing a stall warning until it issues another stall warning
|
issuing a stall warning until it issues another stall warning
|
||||||
for the same stall. This time period is normally set to thirty
|
for the same stall. This time period is normally set to three
|
||||||
seconds.
|
times the check interval plus thirty seconds.
|
||||||
|
|
||||||
RCU_STALL_RAT_DELAY
|
RCU_STALL_RAT_DELAY
|
||||||
|
|
||||||
|
@@ -10,34 +10,46 @@ for rcutree and next for rcutiny.
|
|||||||
|
|
||||||
CONFIG_TREE_RCU and CONFIG_TREE_PREEMPT_RCU debugfs Files and Formats
|
CONFIG_TREE_RCU and CONFIG_TREE_PREEMPT_RCU debugfs Files and Formats
|
||||||
|
|
||||||
These implementations of RCU provides five debugfs files under the
|
These implementations of RCU provides several debugfs files under the
|
||||||
top-level directory RCU: rcu/rcudata (which displays fields in struct
|
top-level directory "rcu":
|
||||||
rcu_data), rcu/rcudata.csv (which is a .csv spreadsheet version of
|
|
||||||
rcu/rcudata), rcu/rcugp (which displays grace-period counters),
|
rcu/rcudata:
|
||||||
rcu/rcuhier (which displays the struct rcu_node hierarchy), and
|
Displays fields in struct rcu_data.
|
||||||
rcu/rcu_pending (which displays counts of the reasons that the
|
rcu/rcudata.csv:
|
||||||
rcu_pending() function decided that there was core RCU work to do).
|
Comma-separated values spreadsheet version of rcudata.
|
||||||
|
rcu/rcugp:
|
||||||
|
Displays grace-period counters.
|
||||||
|
rcu/rcuhier:
|
||||||
|
Displays the struct rcu_node hierarchy.
|
||||||
|
rcu/rcu_pending:
|
||||||
|
Displays counts of the reasons rcu_pending() decided that RCU had
|
||||||
|
work to do.
|
||||||
|
rcu/rcutorture:
|
||||||
|
Displays rcutorture test progress.
|
||||||
|
rcu/rcuboost:
|
||||||
|
Displays RCU boosting statistics. Only present if
|
||||||
|
CONFIG_RCU_BOOST=y.
|
||||||
|
|
||||||
The output of "cat rcu/rcudata" looks as follows:
|
The output of "cat rcu/rcudata" looks as follows:
|
||||||
|
|
||||||
rcu_sched:
|
rcu_sched:
|
||||||
0 c=17829 g=17829 pq=1 pqc=17829 qp=0 dt=10951/1 dn=0 df=1101 of=0 ri=36 ql=0 b=10
|
0 c=20972 g=20973 pq=1 pqc=20972 qp=0 dt=545/1/0 df=50 of=0 ri=0 ql=163 qs=NRW. kt=0/W/0 ktl=ebc3 b=10 ci=153737 co=0 ca=0
|
||||||
1 c=17829 g=17829 pq=1 pqc=17829 qp=0 dt=16117/1 dn=0 df=1015 of=0 ri=0 ql=0 b=10
|
1 c=20972 g=20973 pq=1 pqc=20972 qp=0 dt=967/1/0 df=58 of=0 ri=0 ql=634 qs=NRW. kt=0/W/1 ktl=58c b=10 ci=191037 co=0 ca=0
|
||||||
2 c=17829 g=17829 pq=1 pqc=17829 qp=0 dt=1445/1 dn=0 df=1839 of=0 ri=0 ql=0 b=10
|
2 c=20972 g=20973 pq=1 pqc=20972 qp=0 dt=1081/1/0 df=175 of=0 ri=0 ql=74 qs=N.W. kt=0/W/2 ktl=da94 b=10 ci=75991 co=0 ca=0
|
||||||
3 c=17829 g=17829 pq=1 pqc=17829 qp=0 dt=6681/1 dn=0 df=1545 of=0 ri=0 ql=0 b=10
|
3 c=20942 g=20943 pq=1 pqc=20942 qp=1 dt=1846/0/0 df=404 of=0 ri=0 ql=0 qs=.... kt=0/W/3 ktl=d1cd b=10 ci=72261 co=0 ca=0
|
||||||
4 c=17829 g=17829 pq=1 pqc=17829 qp=0 dt=1003/1 dn=0 df=1992 of=0 ri=0 ql=0 b=10
|
4 c=20972 g=20973 pq=1 pqc=20972 qp=0 dt=369/1/0 df=83 of=0 ri=0 ql=48 qs=N.W. kt=0/W/4 ktl=e0e7 b=10 ci=128365 co=0 ca=0
|
||||||
5 c=17829 g=17830 pq=1 pqc=17829 qp=1 dt=3887/1 dn=0 df=3331 of=0 ri=4 ql=2 b=10
|
5 c=20972 g=20973 pq=1 pqc=20972 qp=0 dt=381/1/0 df=64 of=0 ri=0 ql=169 qs=NRW. kt=0/W/5 ktl=fb2f b=10 ci=164360 co=0 ca=0
|
||||||
6 c=17829 g=17829 pq=1 pqc=17829 qp=0 dt=859/1 dn=0 df=3224 of=0 ri=0 ql=0 b=10
|
6 c=20972 g=20973 pq=1 pqc=20972 qp=0 dt=1037/1/0 df=183 of=0 ri=0 ql=62 qs=N.W. kt=0/W/6 ktl=d2ad b=10 ci=65663 co=0 ca=0
|
||||||
7 c=17829 g=17830 pq=0 pqc=17829 qp=1 dt=3761/1 dn=0 df=1818 of=0 ri=0 ql=2 b=10
|
7 c=20897 g=20897 pq=1 pqc=20896 qp=0 dt=1572/0/0 df=382 of=0 ri=0 ql=0 qs=.... kt=0/W/7 ktl=cf15 b=10 ci=75006 co=0 ca=0
|
||||||
rcu_bh:
|
rcu_bh:
|
||||||
0 c=-275 g=-275 pq=1 pqc=-275 qp=0 dt=10951/1 dn=0 df=0 of=0 ri=0 ql=0 b=10
|
0 c=1480 g=1480 pq=1 pqc=1479 qp=0 dt=545/1/0 df=6 of=0 ri=1 ql=0 qs=.... kt=0/W/0 ktl=ebc3 b=10 ci=0 co=0 ca=0
|
||||||
1 c=-275 g=-275 pq=1 pqc=-275 qp=0 dt=16117/1 dn=0 df=13 of=0 ri=0 ql=0 b=10
|
1 c=1480 g=1480 pq=1 pqc=1479 qp=0 dt=967/1/0 df=3 of=0 ri=1 ql=0 qs=.... kt=0/W/1 ktl=58c b=10 ci=151 co=0 ca=0
|
||||||
2 c=-275 g=-275 pq=1 pqc=-275 qp=0 dt=1445/1 dn=0 df=15 of=0 ri=0 ql=0 b=10
|
2 c=1480 g=1480 pq=1 pqc=1479 qp=0 dt=1081/1/0 df=6 of=0 ri=1 ql=0 qs=.... kt=0/W/2 ktl=da94 b=10 ci=0 co=0 ca=0
|
||||||
3 c=-275 g=-275 pq=1 pqc=-275 qp=0 dt=6681/1 dn=0 df=9 of=0 ri=0 ql=0 b=10
|
3 c=1480 g=1480 pq=1 pqc=1479 qp=0 dt=1846/0/0 df=8 of=0 ri=1 ql=0 qs=.... kt=0/W/3 ktl=d1cd b=10 ci=0 co=0 ca=0
|
||||||
4 c=-275 g=-275 pq=1 pqc=-275 qp=0 dt=1003/1 dn=0 df=15 of=0 ri=0 ql=0 b=10
|
4 c=1480 g=1480 pq=1 pqc=1479 qp=0 dt=369/1/0 df=6 of=0 ri=1 ql=0 qs=.... kt=0/W/4 ktl=e0e7 b=10 ci=0 co=0 ca=0
|
||||||
5 c=-275 g=-275 pq=1 pqc=-275 qp=0 dt=3887/1 dn=0 df=15 of=0 ri=0 ql=0 b=10
|
5 c=1480 g=1480 pq=1 pqc=1479 qp=0 dt=381/1/0 df=4 of=0 ri=1 ql=0 qs=.... kt=0/W/5 ktl=fb2f b=10 ci=0 co=0 ca=0
|
||||||
6 c=-275 g=-275 pq=1 pqc=-275 qp=0 dt=859/1 dn=0 df=15 of=0 ri=0 ql=0 b=10
|
6 c=1480 g=1480 pq=1 pqc=1479 qp=0 dt=1037/1/0 df=6 of=0 ri=1 ql=0 qs=.... kt=0/W/6 ktl=d2ad b=10 ci=0 co=0 ca=0
|
||||||
7 c=-275 g=-275 pq=1 pqc=-275 qp=0 dt=3761/1 dn=0 df=15 of=0 ri=0 ql=0 b=10
|
7 c=1474 g=1474 pq=1 pqc=1473 qp=0 dt=1572/0/0 df=8 of=0 ri=1 ql=0 qs=.... kt=0/W/7 ktl=cf15 b=10 ci=0 co=0 ca=0
|
||||||
|
|
||||||
The first section lists the rcu_data structures for rcu_sched, the second
|
The first section lists the rcu_data structures for rcu_sched, the second
|
||||||
for rcu_bh. Note that CONFIG_TREE_PREEMPT_RCU kernels will have an
|
for rcu_bh. Note that CONFIG_TREE_PREEMPT_RCU kernels will have an
|
||||||
@@ -52,17 +64,18 @@ o The number at the beginning of each line is the CPU number.
|
|||||||
substantially larger than the number of actual CPUs.
|
substantially larger than the number of actual CPUs.
|
||||||
|
|
||||||
o "c" is the count of grace periods that this CPU believes have
|
o "c" is the count of grace periods that this CPU believes have
|
||||||
completed. CPUs in dynticks idle mode may lag quite a ways
|
completed. Offlined CPUs and CPUs in dynticks idle mode may
|
||||||
behind, for example, CPU 4 under "rcu_sched" above, which has
|
lag quite a ways behind, for example, CPU 6 under "rcu_sched"
|
||||||
slept through the past 25 RCU grace periods. It is not unusual
|
above, which has been offline through not quite 40,000 RCU grace
|
||||||
to see CPUs lagging by thousands of grace periods.
|
periods. It is not unusual to see CPUs lagging by thousands of
|
||||||
|
grace periods.
|
||||||
|
|
||||||
o "g" is the count of grace periods that this CPU believes have
|
o "g" is the count of grace periods that this CPU believes have
|
||||||
started. Again, CPUs in dynticks idle mode may lag behind.
|
started. Again, offlined CPUs and CPUs in dynticks idle mode
|
||||||
If the "c" and "g" values are equal, this CPU has already
|
may lag behind. If the "c" and "g" values are equal, this CPU
|
||||||
reported a quiescent state for the last RCU grace period that
|
has already reported a quiescent state for the last RCU grace
|
||||||
it is aware of, otherwise, the CPU believes that it owes RCU a
|
period that it is aware of, otherwise, the CPU believes that it
|
||||||
quiescent state.
|
owes RCU a quiescent state.
|
||||||
|
|
||||||
o "pq" indicates that this CPU has passed through a quiescent state
|
o "pq" indicates that this CPU has passed through a quiescent state
|
||||||
for the current grace period. It is possible for "pq" to be
|
for the current grace period. It is possible for "pq" to be
|
||||||
@@ -81,22 +94,16 @@ o "pqc" indicates which grace period the last-observed quiescent
|
|||||||
the next grace period!
|
the next grace period!
|
||||||
|
|
||||||
o "qp" indicates that RCU still expects a quiescent state from
|
o "qp" indicates that RCU still expects a quiescent state from
|
||||||
this CPU.
|
this CPU. Offlined CPUs and CPUs in dyntick idle mode might
|
||||||
|
well have qp=1, which is OK: RCU is still ignoring them.
|
||||||
|
|
||||||
o "dt" is the current value of the dyntick counter that is incremented
|
o "dt" is the current value of the dyntick counter that is incremented
|
||||||
when entering or leaving dynticks idle state, either by the
|
when entering or leaving dynticks idle state, either by the
|
||||||
scheduler or by irq. The number after the "/" is the interrupt
|
scheduler or by irq. This number is even if the CPU is in
|
||||||
nesting depth when in dyntick-idle state, or one greater than
|
dyntick idle mode and odd otherwise. The number after the first
|
||||||
the interrupt-nesting depth otherwise.
|
"/" is the interrupt nesting depth when in dyntick-idle state,
|
||||||
|
or one greater than the interrupt-nesting depth otherwise.
|
||||||
This field is displayed only for CONFIG_NO_HZ kernels.
|
The number after the second "/" is the NMI nesting depth.
|
||||||
|
|
||||||
o "dn" is the current value of the dyntick counter that is incremented
|
|
||||||
when entering or leaving dynticks idle state via NMI. If both
|
|
||||||
the "dt" and "dn" values are even, then this CPU is in dynticks
|
|
||||||
idle mode and may be ignored by RCU. If either of these two
|
|
||||||
counters is odd, then RCU must be alert to the possibility of
|
|
||||||
an RCU read-side critical section running on this CPU.
|
|
||||||
|
|
||||||
This field is displayed only for CONFIG_NO_HZ kernels.
|
This field is displayed only for CONFIG_NO_HZ kernels.
|
||||||
|
|
||||||
@@ -108,7 +115,7 @@ o "df" is the number of times that some other CPU has forced a
|
|||||||
|
|
||||||
o "of" is the number of times that some other CPU has forced a
|
o "of" is the number of times that some other CPU has forced a
|
||||||
quiescent state on behalf of this CPU due to this CPU being
|
quiescent state on behalf of this CPU due to this CPU being
|
||||||
offline. In a perfect world, this might neve happen, but it
|
offline. In a perfect world, this might never happen, but it
|
||||||
turns out that offlining and onlining a CPU can take several grace
|
turns out that offlining and onlining a CPU can take several grace
|
||||||
periods, and so there is likely to be an extended period of time
|
periods, and so there is likely to be an extended period of time
|
||||||
when RCU believes that the CPU is online when it really is not.
|
when RCU believes that the CPU is online when it really is not.
|
||||||
@@ -125,6 +132,62 @@ o "ql" is the number of RCU callbacks currently residing on
|
|||||||
of what state they are in (new, waiting for grace period to
|
of what state they are in (new, waiting for grace period to
|
||||||
start, waiting for grace period to end, ready to invoke).
|
start, waiting for grace period to end, ready to invoke).
|
||||||
|
|
||||||
|
o "qs" gives an indication of the state of the callback queue
|
||||||
|
with four characters:
|
||||||
|
|
||||||
|
"N" Indicates that there are callbacks queued that are not
|
||||||
|
ready to be handled by the next grace period, and thus
|
||||||
|
will be handled by the grace period following the next
|
||||||
|
one.
|
||||||
|
|
||||||
|
"R" Indicates that there are callbacks queued that are
|
||||||
|
ready to be handled by the next grace period.
|
||||||
|
|
||||||
|
"W" Indicates that there are callbacks queued that are
|
||||||
|
waiting on the current grace period.
|
||||||
|
|
||||||
|
"D" Indicates that there are callbacks queued that have
|
||||||
|
already been handled by a prior grace period, and are
|
||||||
|
thus waiting to be invoked. Note that callbacks in
|
||||||
|
the process of being invoked are not counted here.
|
||||||
|
Callbacks in the process of being invoked are those
|
||||||
|
that have been removed from the rcu_data structures
|
||||||
|
queues by rcu_do_batch(), but which have not yet been
|
||||||
|
invoked.
|
||||||
|
|
||||||
|
If there are no callbacks in a given one of the above states,
|
||||||
|
the corresponding character is replaced by ".".
|
||||||
|
|
||||||
|
o "kt" is the per-CPU kernel-thread state. The digit preceding
|
||||||
|
the first slash is zero if there is no work pending and 1
|
||||||
|
otherwise. The character between the first pair of slashes is
|
||||||
|
as follows:
|
||||||
|
|
||||||
|
"S" The kernel thread is stopped, in other words, all
|
||||||
|
CPUs corresponding to this rcu_node structure are
|
||||||
|
offline.
|
||||||
|
|
||||||
|
"R" The kernel thread is running.
|
||||||
|
|
||||||
|
"W" The kernel thread is waiting because there is no work
|
||||||
|
for it to do.
|
||||||
|
|
||||||
|
"O" The kernel thread is waiting because it has been
|
||||||
|
forced off of its designated CPU or because its
|
||||||
|
->cpus_allowed mask permits it to run on other than
|
||||||
|
its designated CPU.
|
||||||
|
|
||||||
|
"Y" The kernel thread is yielding to avoid hogging CPU.
|
||||||
|
|
||||||
|
"?" Unknown value, indicates a bug.
|
||||||
|
|
||||||
|
The number after the final slash is the CPU that the kthread
|
||||||
|
is actually running on.
|
||||||
|
|
||||||
|
o "ktl" is the low-order 16 bits (in hexadecimal) of the count of
|
||||||
|
the number of times that this CPU's per-CPU kthread has gone
|
||||||
|
through its loop servicing invoke_rcu_cpu_kthread() requests.
|
||||||
|
|
||||||
o "b" is the batch limit for this CPU. If more than this number
|
o "b" is the batch limit for this CPU. If more than this number
|
||||||
of RCU callbacks is ready to invoke, then the remainder will
|
of RCU callbacks is ready to invoke, then the remainder will
|
||||||
be deferred.
|
be deferred.
|
||||||
@@ -174,14 +237,14 @@ o "gpnum" is the number of grace periods that have started. It is
|
|||||||
The output of "cat rcu/rcuhier" looks as follows, with very long lines:
|
The output of "cat rcu/rcuhier" looks as follows, with very long lines:
|
||||||
|
|
||||||
c=6902 g=6903 s=2 jfq=3 j=72c7 nfqs=13142/nfqsng=0(13142) fqlh=6
|
c=6902 g=6903 s=2 jfq=3 j=72c7 nfqs=13142/nfqsng=0(13142) fqlh=6
|
||||||
1/1 .>. 0:127 ^0
|
1/1 ..>. 0:127 ^0
|
||||||
3/3 .>. 0:35 ^0 0/0 .>. 36:71 ^1 0/0 .>. 72:107 ^2 0/0 .>. 108:127 ^3
|
3/3 ..>. 0:35 ^0 0/0 ..>. 36:71 ^1 0/0 ..>. 72:107 ^2 0/0 ..>. 108:127 ^3
|
||||||
3/3f .>. 0:5 ^0 2/3 .>. 6:11 ^1 0/0 .>. 12:17 ^2 0/0 .>. 18:23 ^3 0/0 .>. 24:29 ^4 0/0 .>. 30:35 ^5 0/0 .>. 36:41 ^0 0/0 .>. 42:47 ^1 0/0 .>. 48:53 ^2 0/0 .>. 54:59 ^3 0/0 .>. 60:65 ^4 0/0 .>. 66:71 ^5 0/0 .>. 72:77 ^0 0/0 .>. 78:83 ^1 0/0 .>. 84:89 ^2 0/0 .>. 90:95 ^3 0/0 .>. 96:101 ^4 0/0 .>. 102:107 ^5 0/0 .>. 108:113 ^0 0/0 .>. 114:119 ^1 0/0 .>. 120:125 ^2 0/0 .>. 126:127 ^3
|
3/3f ..>. 0:5 ^0 2/3 ..>. 6:11 ^1 0/0 ..>. 12:17 ^2 0/0 ..>. 18:23 ^3 0/0 ..>. 24:29 ^4 0/0 ..>. 30:35 ^5 0/0 ..>. 36:41 ^0 0/0 ..>. 42:47 ^1 0/0 ..>. 48:53 ^2 0/0 ..>. 54:59 ^3 0/0 ..>. 60:65 ^4 0/0 ..>. 66:71 ^5 0/0 ..>. 72:77 ^0 0/0 ..>. 78:83 ^1 0/0 ..>. 84:89 ^2 0/0 ..>. 90:95 ^3 0/0 ..>. 96:101 ^4 0/0 ..>. 102:107 ^5 0/0 ..>. 108:113 ^0 0/0 ..>. 114:119 ^1 0/0 ..>. 120:125 ^2 0/0 ..>. 126:127 ^3
|
||||||
rcu_bh:
|
rcu_bh:
|
||||||
c=-226 g=-226 s=1 jfq=-5701 j=72c7 nfqs=88/nfqsng=0(88) fqlh=0
|
c=-226 g=-226 s=1 jfq=-5701 j=72c7 nfqs=88/nfqsng=0(88) fqlh=0
|
||||||
0/1 .>. 0:127 ^0
|
0/1 ..>. 0:127 ^0
|
||||||
0/3 .>. 0:35 ^0 0/0 .>. 36:71 ^1 0/0 .>. 72:107 ^2 0/0 .>. 108:127 ^3
|
0/3 ..>. 0:35 ^0 0/0 ..>. 36:71 ^1 0/0 ..>. 72:107 ^2 0/0 ..>. 108:127 ^3
|
||||||
0/3f .>. 0:5 ^0 0/3 .>. 6:11 ^1 0/0 .>. 12:17 ^2 0/0 .>. 18:23 ^3 0/0 .>. 24:29 ^4 0/0 .>. 30:35 ^5 0/0 .>. 36:41 ^0 0/0 .>. 42:47 ^1 0/0 .>. 48:53 ^2 0/0 .>. 54:59 ^3 0/0 .>. 60:65 ^4 0/0 .>. 66:71 ^5 0/0 .>. 72:77 ^0 0/0 .>. 78:83 ^1 0/0 .>. 84:89 ^2 0/0 .>. 90:95 ^3 0/0 .>. 96:101 ^4 0/0 .>. 102:107 ^5 0/0 .>. 108:113 ^0 0/0 .>. 114:119 ^1 0/0 .>. 120:125 ^2 0/0 .>. 126:127 ^3
|
0/3f ..>. 0:5 ^0 0/3 ..>. 6:11 ^1 0/0 ..>. 12:17 ^2 0/0 ..>. 18:23 ^3 0/0 ..>. 24:29 ^4 0/0 ..>. 30:35 ^5 0/0 ..>. 36:41 ^0 0/0 ..>. 42:47 ^1 0/0 ..>. 48:53 ^2 0/0 ..>. 54:59 ^3 0/0 ..>. 60:65 ^4 0/0 ..>. 66:71 ^5 0/0 ..>. 72:77 ^0 0/0 ..>. 78:83 ^1 0/0 ..>. 84:89 ^2 0/0 ..>. 90:95 ^3 0/0 ..>. 96:101 ^4 0/0 ..>. 102:107 ^5 0/0 ..>. 108:113 ^0 0/0 ..>. 114:119 ^1 0/0 ..>. 120:125 ^2 0/0 ..>. 126:127 ^3
|
||||||
|
|
||||||
This is once again split into "rcu_sched" and "rcu_bh" portions,
|
This is once again split into "rcu_sched" and "rcu_bh" portions,
|
||||||
and CONFIG_TREE_PREEMPT_RCU kernels will again have an additional
|
and CONFIG_TREE_PREEMPT_RCU kernels will again have an additional
|
||||||
@@ -240,13 +303,20 @@ o Each element of the form "1/1 0:127 ^0" represents one struct
|
|||||||
current grace period.
|
current grace period.
|
||||||
|
|
||||||
o The characters separated by the ">" indicate the state
|
o The characters separated by the ">" indicate the state
|
||||||
of the blocked-tasks lists. A "T" preceding the ">"
|
of the blocked-tasks lists. A "G" preceding the ">"
|
||||||
indicates that at least one task blocked in an RCU
|
indicates that at least one task blocked in an RCU
|
||||||
read-side critical section blocks the current grace
|
read-side critical section blocks the current grace
|
||||||
period, while a "." preceding the ">" indicates otherwise.
|
period, while a "E" preceding the ">" indicates that
|
||||||
The character following the ">" indicates similarly for
|
at least one task blocked in an RCU read-side critical
|
||||||
the next grace period. A "T" should appear in this
|
section blocks the current expedited grace period.
|
||||||
field only for rcu-preempt.
|
A "T" character following the ">" indicates that at
|
||||||
|
least one task is blocked within an RCU read-side
|
||||||
|
critical section, regardless of whether any current
|
||||||
|
grace period (expedited or normal) is inconvenienced.
|
||||||
|
A "." character appears if the corresponding condition
|
||||||
|
does not hold, so that "..>." indicates that no tasks
|
||||||
|
are blocked. In contrast, "GE>T" indicates maximal
|
||||||
|
inconvenience from blocked tasks.
|
||||||
|
|
||||||
o The numbers separated by the ":" are the range of CPUs
|
o The numbers separated by the ":" are the range of CPUs
|
||||||
served by this struct rcu_node. This can be helpful
|
served by this struct rcu_node. This can be helpful
|
||||||
@@ -328,6 +398,113 @@ o "nn" is the number of times that this CPU needed nothing. Alert
|
|||||||
is due to short-circuit evaluation in rcu_pending().
|
is due to short-circuit evaluation in rcu_pending().
|
||||||
|
|
||||||
|
|
||||||
|
The output of "cat rcu/rcutorture" looks as follows:
|
||||||
|
|
||||||
|
rcutorture test sequence: 0 (test in progress)
|
||||||
|
rcutorture update version number: 615
|
||||||
|
|
||||||
|
The first line shows the number of rcutorture tests that have completed
|
||||||
|
since boot. If a test is currently running, the "(test in progress)"
|
||||||
|
string will appear as shown above. The second line shows the number of
|
||||||
|
update cycles that the current test has started, or zero if there is
|
||||||
|
no test in progress.
|
||||||
|
|
||||||
|
|
||||||
|
The output of "cat rcu/rcuboost" looks as follows:
|
||||||
|
|
||||||
|
0:5 tasks=.... kt=W ntb=0 neb=0 nnb=0 j=2f95 bt=300f
|
||||||
|
balk: nt=0 egt=989 bt=0 nb=0 ny=0 nos=16
|
||||||
|
6:7 tasks=.... kt=W ntb=0 neb=0 nnb=0 j=2f95 bt=300f
|
||||||
|
balk: nt=0 egt=225 bt=0 nb=0 ny=0 nos=6
|
||||||
|
|
||||||
|
This information is output only for rcu_preempt. Each two-line entry
|
||||||
|
corresponds to a leaf rcu_node strcuture. The fields are as follows:
|
||||||
|
|
||||||
|
o "n:m" is the CPU-number range for the corresponding two-line
|
||||||
|
entry. In the sample output above, the first entry covers
|
||||||
|
CPUs zero through five and the second entry covers CPUs 6
|
||||||
|
and 7.
|
||||||
|
|
||||||
|
o "tasks=TNEB" gives the state of the various segments of the
|
||||||
|
rnp->blocked_tasks list:
|
||||||
|
|
||||||
|
"T" This indicates that there are some tasks that blocked
|
||||||
|
while running on one of the corresponding CPUs while
|
||||||
|
in an RCU read-side critical section.
|
||||||
|
|
||||||
|
"N" This indicates that some of the blocked tasks are preventing
|
||||||
|
the current normal (non-expedited) grace period from
|
||||||
|
completing.
|
||||||
|
|
||||||
|
"E" This indicates that some of the blocked tasks are preventing
|
||||||
|
the current expedited grace period from completing.
|
||||||
|
|
||||||
|
"B" This indicates that some of the blocked tasks are in
|
||||||
|
need of RCU priority boosting.
|
||||||
|
|
||||||
|
Each character is replaced with "." if the corresponding
|
||||||
|
condition does not hold.
|
||||||
|
|
||||||
|
o "kt" is the state of the RCU priority-boosting kernel
|
||||||
|
thread associated with the corresponding rcu_node structure.
|
||||||
|
The state can be one of the following:
|
||||||
|
|
||||||
|
"S" The kernel thread is stopped, in other words, all
|
||||||
|
CPUs corresponding to this rcu_node structure are
|
||||||
|
offline.
|
||||||
|
|
||||||
|
"R" The kernel thread is running.
|
||||||
|
|
||||||
|
"W" The kernel thread is waiting because there is no work
|
||||||
|
for it to do.
|
||||||
|
|
||||||
|
"Y" The kernel thread is yielding to avoid hogging CPU.
|
||||||
|
|
||||||
|
"?" Unknown value, indicates a bug.
|
||||||
|
|
||||||
|
o "ntb" is the number of tasks boosted.
|
||||||
|
|
||||||
|
o "neb" is the number of tasks boosted in order to complete an
|
||||||
|
expedited grace period.
|
||||||
|
|
||||||
|
o "nnb" is the number of tasks boosted in order to complete a
|
||||||
|
normal (non-expedited) grace period. When boosting a task
|
||||||
|
that was blocking both an expedited and a normal grace period,
|
||||||
|
it is counted against the expedited total above.
|
||||||
|
|
||||||
|
o "j" is the low-order 16 bits of the jiffies counter in
|
||||||
|
hexadecimal.
|
||||||
|
|
||||||
|
o "bt" is the low-order 16 bits of the value that the jiffies
|
||||||
|
counter will have when we next start boosting, assuming that
|
||||||
|
the current grace period does not end beforehand. This is
|
||||||
|
also in hexadecimal.
|
||||||
|
|
||||||
|
o "balk: nt" counts the number of times we didn't boost (in
|
||||||
|
other words, we balked) even though it was time to boost because
|
||||||
|
there were no blocked tasks to boost. This situation occurs
|
||||||
|
when there is one blocked task on one rcu_node structure and
|
||||||
|
none on some other rcu_node structure.
|
||||||
|
|
||||||
|
o "egt" counts the number of times we balked because although
|
||||||
|
there were blocked tasks, none of them were blocking the
|
||||||
|
current grace period, whether expedited or otherwise.
|
||||||
|
|
||||||
|
o "bt" counts the number of times we balked because boosting
|
||||||
|
had already been initiated for the current grace period.
|
||||||
|
|
||||||
|
o "nb" counts the number of times we balked because there
|
||||||
|
was at least one task blocking the current non-expedited grace
|
||||||
|
period that never had blocked. If it is already running, it
|
||||||
|
just won't help to boost its priority!
|
||||||
|
|
||||||
|
o "ny" counts the number of times we balked because it was
|
||||||
|
not yet time to start boosting.
|
||||||
|
|
||||||
|
o "nos" counts the number of times we balked for other
|
||||||
|
reasons, e.g., the grace period ended first.
|
||||||
|
|
||||||
|
|
||||||
CONFIG_TINY_RCU and CONFIG_TINY_PREEMPT_RCU debugfs Files and Formats
|
CONFIG_TINY_RCU and CONFIG_TINY_PREEMPT_RCU debugfs Files and Formats
|
||||||
|
|
||||||
These implementations of RCU provides a single debugfs file under the
|
These implementations of RCU provides a single debugfs file under the
|
||||||
@@ -394,9 +571,9 @@ o "neb" is the number of expedited grace periods that have had
|
|||||||
o "nnb" is the number of normal grace periods that have had
|
o "nnb" is the number of normal grace periods that have had
|
||||||
to resort to RCU priority boosting since boot.
|
to resort to RCU priority boosting since boot.
|
||||||
|
|
||||||
o "j" is the low-order 12 bits of the jiffies counter in hexadecimal.
|
o "j" is the low-order 16 bits of the jiffies counter in hexadecimal.
|
||||||
|
|
||||||
o "bt" is the low-order 12 bits of the value that the jiffies counter
|
o "bt" is the low-order 16 bits of the value that the jiffies counter
|
||||||
will have at the next time that boosting is scheduled to begin.
|
will have at the next time that boosting is scheduled to begin.
|
||||||
|
|
||||||
o In the line beginning with "normal balk", the fields are as follows:
|
o In the line beginning with "normal balk", the fields are as follows:
|
||||||
|
@@ -836,7 +836,6 @@ Provides counts of softirq handlers serviced since boot time, for each cpu.
|
|||||||
TASKLET: 0 0 0 290
|
TASKLET: 0 0 0 290
|
||||||
SCHED: 27035 26983 26971 26746
|
SCHED: 27035 26983 26971 26746
|
||||||
HRTIMER: 0 0 0 0
|
HRTIMER: 0 0 0 0
|
||||||
RCU: 1678 1769 2178 2250
|
|
||||||
|
|
||||||
|
|
||||||
1.3 IDE devices in /proc/ide
|
1.3 IDE devices in /proc/ide
|
||||||
|
@@ -414,7 +414,6 @@ enum
|
|||||||
TASKLET_SOFTIRQ,
|
TASKLET_SOFTIRQ,
|
||||||
SCHED_SOFTIRQ,
|
SCHED_SOFTIRQ,
|
||||||
HRTIMER_SOFTIRQ,
|
HRTIMER_SOFTIRQ,
|
||||||
RCU_SOFTIRQ, /* Preferable RCU should always be the last softirq */
|
|
||||||
|
|
||||||
NR_SOFTIRQS
|
NR_SOFTIRQS
|
||||||
};
|
};
|
||||||
|
@@ -47,6 +47,18 @@
|
|||||||
extern int rcutorture_runnable; /* for sysctl */
|
extern int rcutorture_runnable; /* for sysctl */
|
||||||
#endif /* #ifdef CONFIG_RCU_TORTURE_TEST */
|
#endif /* #ifdef CONFIG_RCU_TORTURE_TEST */
|
||||||
|
|
||||||
|
#if defined(CONFIG_TREE_RCU) || defined(CONFIG_TREE_PREEMPT_RCU)
|
||||||
|
extern void rcutorture_record_test_transition(void);
|
||||||
|
extern void rcutorture_record_progress(unsigned long vernum);
|
||||||
|
#else
|
||||||
|
static inline void rcutorture_record_test_transition(void)
|
||||||
|
{
|
||||||
|
}
|
||||||
|
static inline void rcutorture_record_progress(unsigned long vernum)
|
||||||
|
{
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
#define UINT_CMP_GE(a, b) (UINT_MAX / 2 >= (a) - (b))
|
#define UINT_CMP_GE(a, b) (UINT_MAX / 2 >= (a) - (b))
|
||||||
#define UINT_CMP_LT(a, b) (UINT_MAX / 2 < (a) - (b))
|
#define UINT_CMP_LT(a, b) (UINT_MAX / 2 < (a) - (b))
|
||||||
#define ULONG_CMP_GE(a, b) (ULONG_MAX / 2 >= (a) - (b))
|
#define ULONG_CMP_GE(a, b) (ULONG_MAX / 2 >= (a) - (b))
|
||||||
@@ -68,7 +80,6 @@ extern void call_rcu_sched(struct rcu_head *head,
|
|||||||
extern void synchronize_sched(void);
|
extern void synchronize_sched(void);
|
||||||
extern void rcu_barrier_bh(void);
|
extern void rcu_barrier_bh(void);
|
||||||
extern void rcu_barrier_sched(void);
|
extern void rcu_barrier_sched(void);
|
||||||
extern int sched_expedited_torture_stats(char *page);
|
|
||||||
|
|
||||||
static inline void __rcu_read_lock_bh(void)
|
static inline void __rcu_read_lock_bh(void)
|
||||||
{
|
{
|
||||||
@@ -774,6 +785,7 @@ extern struct debug_obj_descr rcuhead_debug_descr;
|
|||||||
|
|
||||||
static inline void debug_rcu_head_queue(struct rcu_head *head)
|
static inline void debug_rcu_head_queue(struct rcu_head *head)
|
||||||
{
|
{
|
||||||
|
WARN_ON_ONCE((unsigned long)head & 0x3);
|
||||||
debug_object_activate(head, &rcuhead_debug_descr);
|
debug_object_activate(head, &rcuhead_debug_descr);
|
||||||
debug_object_active_state(head, &rcuhead_debug_descr,
|
debug_object_active_state(head, &rcuhead_debug_descr,
|
||||||
STATE_RCU_HEAD_READY,
|
STATE_RCU_HEAD_READY,
|
||||||
@@ -797,4 +809,60 @@ static inline void debug_rcu_head_unqueue(struct rcu_head *head)
|
|||||||
}
|
}
|
||||||
#endif /* #else !CONFIG_DEBUG_OBJECTS_RCU_HEAD */
|
#endif /* #else !CONFIG_DEBUG_OBJECTS_RCU_HEAD */
|
||||||
|
|
||||||
|
static __always_inline bool __is_kfree_rcu_offset(unsigned long offset)
|
||||||
|
{
|
||||||
|
return offset < 4096;
|
||||||
|
}
|
||||||
|
|
||||||
|
static __always_inline
|
||||||
|
void __kfree_rcu(struct rcu_head *head, unsigned long offset)
|
||||||
|
{
|
||||||
|
typedef void (*rcu_callback)(struct rcu_head *);
|
||||||
|
|
||||||
|
BUILD_BUG_ON(!__builtin_constant_p(offset));
|
||||||
|
|
||||||
|
/* See the kfree_rcu() header comment. */
|
||||||
|
BUILD_BUG_ON(!__is_kfree_rcu_offset(offset));
|
||||||
|
|
||||||
|
call_rcu(head, (rcu_callback)offset);
|
||||||
|
}
|
||||||
|
|
||||||
|
extern void kfree(const void *);
|
||||||
|
|
||||||
|
static inline void __rcu_reclaim(struct rcu_head *head)
|
||||||
|
{
|
||||||
|
unsigned long offset = (unsigned long)head->func;
|
||||||
|
|
||||||
|
if (__is_kfree_rcu_offset(offset))
|
||||||
|
kfree((void *)head - offset);
|
||||||
|
else
|
||||||
|
head->func(head);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* kfree_rcu() - kfree an object after a grace period.
|
||||||
|
* @ptr: pointer to kfree
|
||||||
|
* @rcu_head: the name of the struct rcu_head within the type of @ptr.
|
||||||
|
*
|
||||||
|
* Many rcu callbacks functions just call kfree() on the base structure.
|
||||||
|
* These functions are trivial, but their size adds up, and furthermore
|
||||||
|
* when they are used in a kernel module, that module must invoke the
|
||||||
|
* high-latency rcu_barrier() function at module-unload time.
|
||||||
|
*
|
||||||
|
* The kfree_rcu() function handles this issue. Rather than encoding a
|
||||||
|
* function address in the embedded rcu_head structure, kfree_rcu() instead
|
||||||
|
* encodes the offset of the rcu_head structure within the base structure.
|
||||||
|
* Because the functions are not allowed in the low-order 4096 bytes of
|
||||||
|
* kernel virtual memory, offsets up to 4095 bytes can be accommodated.
|
||||||
|
* If the offset is larger than 4095 bytes, a compile-time error will
|
||||||
|
* be generated in __kfree_rcu(). If this error is triggered, you can
|
||||||
|
* either fall back to use of call_rcu() or rearrange the structure to
|
||||||
|
* position the rcu_head structure into the first 4096 bytes.
|
||||||
|
*
|
||||||
|
* Note that the allowable offset might decrease in the future, for example,
|
||||||
|
* to allow something like kmem_cache_free_rcu().
|
||||||
|
*/
|
||||||
|
#define kfree_rcu(ptr, rcu_head) \
|
||||||
|
__kfree_rcu(&((ptr)->rcu_head), offsetof(typeof(*(ptr)), rcu_head))
|
||||||
|
|
||||||
#endif /* __LINUX_RCUPDATE_H */
|
#endif /* __LINUX_RCUPDATE_H */
|
||||||
|
@@ -99,6 +99,14 @@ static inline void rcu_note_context_switch(int cpu)
|
|||||||
rcu_preempt_note_context_switch();
|
rcu_preempt_note_context_switch();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Take advantage of the fact that there is only one CPU, which
|
||||||
|
* allows us to ignore virtualization-based context switches.
|
||||||
|
*/
|
||||||
|
static inline void rcu_virt_note_context_switch(int cpu)
|
||||||
|
{
|
||||||
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Return the number of grace periods.
|
* Return the number of grace periods.
|
||||||
*/
|
*/
|
||||||
|
@@ -35,6 +35,16 @@ extern void rcu_note_context_switch(int cpu);
|
|||||||
extern int rcu_needs_cpu(int cpu);
|
extern int rcu_needs_cpu(int cpu);
|
||||||
extern void rcu_cpu_stall_reset(void);
|
extern void rcu_cpu_stall_reset(void);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Note a virtualization-based context switch. This is simply a
|
||||||
|
* wrapper around rcu_note_context_switch(), which allows TINY_RCU
|
||||||
|
* to save a few bytes.
|
||||||
|
*/
|
||||||
|
static inline void rcu_virt_note_context_switch(int cpu)
|
||||||
|
{
|
||||||
|
rcu_note_context_switch(cpu);
|
||||||
|
}
|
||||||
|
|
||||||
#ifdef CONFIG_TREE_PREEMPT_RCU
|
#ifdef CONFIG_TREE_PREEMPT_RCU
|
||||||
|
|
||||||
extern void exit_rcu(void);
|
extern void exit_rcu(void);
|
||||||
@@ -58,9 +68,12 @@ static inline void synchronize_rcu_bh_expedited(void)
|
|||||||
|
|
||||||
extern void rcu_barrier(void);
|
extern void rcu_barrier(void);
|
||||||
|
|
||||||
|
extern unsigned long rcutorture_testseq;
|
||||||
|
extern unsigned long rcutorture_vernum;
|
||||||
extern long rcu_batches_completed(void);
|
extern long rcu_batches_completed(void);
|
||||||
extern long rcu_batches_completed_bh(void);
|
extern long rcu_batches_completed_bh(void);
|
||||||
extern long rcu_batches_completed_sched(void);
|
extern long rcu_batches_completed_sched(void);
|
||||||
|
|
||||||
extern void rcu_force_quiescent_state(void);
|
extern void rcu_force_quiescent_state(void);
|
||||||
extern void rcu_bh_force_quiescent_state(void);
|
extern void rcu_bh_force_quiescent_state(void);
|
||||||
extern void rcu_sched_force_quiescent_state(void);
|
extern void rcu_sched_force_quiescent_state(void);
|
||||||
|
@@ -20,8 +20,7 @@ struct softirq_action;
|
|||||||
softirq_name(BLOCK_IOPOLL), \
|
softirq_name(BLOCK_IOPOLL), \
|
||||||
softirq_name(TASKLET), \
|
softirq_name(TASKLET), \
|
||||||
softirq_name(SCHED), \
|
softirq_name(SCHED), \
|
||||||
softirq_name(HRTIMER), \
|
softirq_name(HRTIMER))
|
||||||
softirq_name(RCU))
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* irq_handler_entry - called immediately before the irq action handler
|
* irq_handler_entry - called immediately before the irq action handler
|
||||||
|
@@ -485,7 +485,7 @@ config TREE_RCU_TRACE
|
|||||||
|
|
||||||
config RCU_BOOST
|
config RCU_BOOST
|
||||||
bool "Enable RCU priority boosting"
|
bool "Enable RCU priority boosting"
|
||||||
depends on RT_MUTEXES && TINY_PREEMPT_RCU
|
depends on RT_MUTEXES && PREEMPT_RCU
|
||||||
default n
|
default n
|
||||||
help
|
help
|
||||||
This option boosts the priority of preempted RCU readers that
|
This option boosts the priority of preempted RCU readers that
|
||||||
|
@@ -142,10 +142,17 @@ static int rcuhead_fixup_init(void *addr, enum debug_obj_state state)
|
|||||||
* Ensure that queued callbacks are all executed.
|
* Ensure that queued callbacks are all executed.
|
||||||
* If we detect that we are nested in a RCU read-side critical
|
* If we detect that we are nested in a RCU read-side critical
|
||||||
* section, we should simply fail, otherwise we would deadlock.
|
* section, we should simply fail, otherwise we would deadlock.
|
||||||
|
* In !PREEMPT configurations, there is no way to tell if we are
|
||||||
|
* in a RCU read-side critical section or not, so we never
|
||||||
|
* attempt any fixup and just print a warning.
|
||||||
*/
|
*/
|
||||||
|
#ifndef CONFIG_PREEMPT
|
||||||
|
WARN_ON_ONCE(1);
|
||||||
|
return 0;
|
||||||
|
#endif
|
||||||
if (rcu_preempt_depth() != 0 || preempt_count() != 0 ||
|
if (rcu_preempt_depth() != 0 || preempt_count() != 0 ||
|
||||||
irqs_disabled()) {
|
irqs_disabled()) {
|
||||||
WARN_ON(1);
|
WARN_ON_ONCE(1);
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
rcu_barrier();
|
rcu_barrier();
|
||||||
@@ -184,10 +191,17 @@ static int rcuhead_fixup_activate(void *addr, enum debug_obj_state state)
|
|||||||
* Ensure that queued callbacks are all executed.
|
* Ensure that queued callbacks are all executed.
|
||||||
* If we detect that we are nested in a RCU read-side critical
|
* If we detect that we are nested in a RCU read-side critical
|
||||||
* section, we should simply fail, otherwise we would deadlock.
|
* section, we should simply fail, otherwise we would deadlock.
|
||||||
|
* In !PREEMPT configurations, there is no way to tell if we are
|
||||||
|
* in a RCU read-side critical section or not, so we never
|
||||||
|
* attempt any fixup and just print a warning.
|
||||||
*/
|
*/
|
||||||
|
#ifndef CONFIG_PREEMPT
|
||||||
|
WARN_ON_ONCE(1);
|
||||||
|
return 0;
|
||||||
|
#endif
|
||||||
if (rcu_preempt_depth() != 0 || preempt_count() != 0 ||
|
if (rcu_preempt_depth() != 0 || preempt_count() != 0 ||
|
||||||
irqs_disabled()) {
|
irqs_disabled()) {
|
||||||
WARN_ON(1);
|
WARN_ON_ONCE(1);
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
rcu_barrier();
|
rcu_barrier();
|
||||||
@@ -214,15 +228,17 @@ static int rcuhead_fixup_free(void *addr, enum debug_obj_state state)
|
|||||||
* Ensure that queued callbacks are all executed.
|
* Ensure that queued callbacks are all executed.
|
||||||
* If we detect that we are nested in a RCU read-side critical
|
* If we detect that we are nested in a RCU read-side critical
|
||||||
* section, we should simply fail, otherwise we would deadlock.
|
* section, we should simply fail, otherwise we would deadlock.
|
||||||
* Note that the machinery to reliably determine whether
|
* In !PREEMPT configurations, there is no way to tell if we are
|
||||||
* or not we are in an RCU read-side critical section
|
* in a RCU read-side critical section or not, so we never
|
||||||
* exists only in the preemptible RCU implementations
|
* attempt any fixup and just print a warning.
|
||||||
* (TINY_PREEMPT_RCU and TREE_PREEMPT_RCU), which is why
|
|
||||||
* DEBUG_OBJECTS_RCU_HEAD is disallowed if !PREEMPT.
|
|
||||||
*/
|
*/
|
||||||
|
#ifndef CONFIG_PREEMPT
|
||||||
|
WARN_ON_ONCE(1);
|
||||||
|
return 0;
|
||||||
|
#endif
|
||||||
if (rcu_preempt_depth() != 0 || preempt_count() != 0 ||
|
if (rcu_preempt_depth() != 0 || preempt_count() != 0 ||
|
||||||
irqs_disabled()) {
|
irqs_disabled()) {
|
||||||
WARN_ON(1);
|
WARN_ON_ONCE(1);
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
rcu_barrier();
|
rcu_barrier();
|
||||||
|
@@ -40,10 +40,10 @@
|
|||||||
static struct task_struct *rcu_kthread_task;
|
static struct task_struct *rcu_kthread_task;
|
||||||
static DECLARE_WAIT_QUEUE_HEAD(rcu_kthread_wq);
|
static DECLARE_WAIT_QUEUE_HEAD(rcu_kthread_wq);
|
||||||
static unsigned long have_rcu_kthread_work;
|
static unsigned long have_rcu_kthread_work;
|
||||||
static void invoke_rcu_kthread(void);
|
|
||||||
|
|
||||||
/* Forward declarations for rcutiny_plugin.h. */
|
/* Forward declarations for rcutiny_plugin.h. */
|
||||||
struct rcu_ctrlblk;
|
struct rcu_ctrlblk;
|
||||||
|
static void invoke_rcu_kthread(void);
|
||||||
static void rcu_process_callbacks(struct rcu_ctrlblk *rcp);
|
static void rcu_process_callbacks(struct rcu_ctrlblk *rcp);
|
||||||
static int rcu_kthread(void *arg);
|
static int rcu_kthread(void *arg);
|
||||||
static void __call_rcu(struct rcu_head *head,
|
static void __call_rcu(struct rcu_head *head,
|
||||||
@@ -79,26 +79,31 @@ void rcu_exit_nohz(void)
|
|||||||
#endif /* #ifdef CONFIG_NO_HZ */
|
#endif /* #ifdef CONFIG_NO_HZ */
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Helper function for rcu_qsctr_inc() and rcu_bh_qsctr_inc().
|
* Helper function for rcu_sched_qs() and rcu_bh_qs().
|
||||||
* Also disable irqs to avoid confusion due to interrupt handlers
|
* Also irqs are disabled to avoid confusion due to interrupt handlers
|
||||||
* invoking call_rcu().
|
* invoking call_rcu().
|
||||||
*/
|
*/
|
||||||
static int rcu_qsctr_help(struct rcu_ctrlblk *rcp)
|
static int rcu_qsctr_help(struct rcu_ctrlblk *rcp)
|
||||||
{
|
{
|
||||||
unsigned long flags;
|
|
||||||
|
|
||||||
local_irq_save(flags);
|
|
||||||
if (rcp->rcucblist != NULL &&
|
if (rcp->rcucblist != NULL &&
|
||||||
rcp->donetail != rcp->curtail) {
|
rcp->donetail != rcp->curtail) {
|
||||||
rcp->donetail = rcp->curtail;
|
rcp->donetail = rcp->curtail;
|
||||||
local_irq_restore(flags);
|
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
local_irq_restore(flags);
|
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Wake up rcu_kthread() to process callbacks now eligible for invocation
|
||||||
|
* or to boost readers.
|
||||||
|
*/
|
||||||
|
static void invoke_rcu_kthread(void)
|
||||||
|
{
|
||||||
|
have_rcu_kthread_work = 1;
|
||||||
|
wake_up(&rcu_kthread_wq);
|
||||||
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Record an rcu quiescent state. And an rcu_bh quiescent state while we
|
* Record an rcu quiescent state. And an rcu_bh quiescent state while we
|
||||||
* are at it, given that any rcu quiescent state is also an rcu_bh
|
* are at it, given that any rcu quiescent state is also an rcu_bh
|
||||||
@@ -106,9 +111,13 @@ static int rcu_qsctr_help(struct rcu_ctrlblk *rcp)
|
|||||||
*/
|
*/
|
||||||
void rcu_sched_qs(int cpu)
|
void rcu_sched_qs(int cpu)
|
||||||
{
|
{
|
||||||
|
unsigned long flags;
|
||||||
|
|
||||||
|
local_irq_save(flags);
|
||||||
if (rcu_qsctr_help(&rcu_sched_ctrlblk) +
|
if (rcu_qsctr_help(&rcu_sched_ctrlblk) +
|
||||||
rcu_qsctr_help(&rcu_bh_ctrlblk))
|
rcu_qsctr_help(&rcu_bh_ctrlblk))
|
||||||
invoke_rcu_kthread();
|
invoke_rcu_kthread();
|
||||||
|
local_irq_restore(flags);
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@@ -116,8 +125,12 @@ void rcu_sched_qs(int cpu)
|
|||||||
*/
|
*/
|
||||||
void rcu_bh_qs(int cpu)
|
void rcu_bh_qs(int cpu)
|
||||||
{
|
{
|
||||||
|
unsigned long flags;
|
||||||
|
|
||||||
|
local_irq_save(flags);
|
||||||
if (rcu_qsctr_help(&rcu_bh_ctrlblk))
|
if (rcu_qsctr_help(&rcu_bh_ctrlblk))
|
||||||
invoke_rcu_kthread();
|
invoke_rcu_kthread();
|
||||||
|
local_irq_restore(flags);
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@@ -167,7 +180,7 @@ static void rcu_process_callbacks(struct rcu_ctrlblk *rcp)
|
|||||||
prefetch(next);
|
prefetch(next);
|
||||||
debug_rcu_head_unqueue(list);
|
debug_rcu_head_unqueue(list);
|
||||||
local_bh_disable();
|
local_bh_disable();
|
||||||
list->func(list);
|
__rcu_reclaim(list);
|
||||||
local_bh_enable();
|
local_bh_enable();
|
||||||
list = next;
|
list = next;
|
||||||
RCU_TRACE(cb_count++);
|
RCU_TRACE(cb_count++);
|
||||||
@@ -207,20 +220,6 @@ static int rcu_kthread(void *arg)
|
|||||||
return 0; /* Not reached, but needed to shut gcc up. */
|
return 0; /* Not reached, but needed to shut gcc up. */
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
|
||||||
* Wake up rcu_kthread() to process callbacks now eligible for invocation
|
|
||||||
* or to boost readers.
|
|
||||||
*/
|
|
||||||
static void invoke_rcu_kthread(void)
|
|
||||||
{
|
|
||||||
unsigned long flags;
|
|
||||||
|
|
||||||
local_irq_save(flags);
|
|
||||||
have_rcu_kthread_work = 1;
|
|
||||||
wake_up(&rcu_kthread_wq);
|
|
||||||
local_irq_restore(flags);
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Wait for a grace period to elapse. But it is illegal to invoke
|
* Wait for a grace period to elapse. But it is illegal to invoke
|
||||||
* synchronize_sched() from within an RCU read-side critical section.
|
* synchronize_sched() from within an RCU read-side critical section.
|
||||||
|
@@ -100,23 +100,28 @@ struct rcu_preempt_ctrlblk {
|
|||||||
u8 completed; /* Last grace period completed. */
|
u8 completed; /* Last grace period completed. */
|
||||||
/* If all three are equal, RCU is idle. */
|
/* If all three are equal, RCU is idle. */
|
||||||
#ifdef CONFIG_RCU_BOOST
|
#ifdef CONFIG_RCU_BOOST
|
||||||
s8 boosted_this_gp; /* Has boosting already happened? */
|
|
||||||
unsigned long boost_time; /* When to start boosting (jiffies) */
|
unsigned long boost_time; /* When to start boosting (jiffies) */
|
||||||
#endif /* #ifdef CONFIG_RCU_BOOST */
|
#endif /* #ifdef CONFIG_RCU_BOOST */
|
||||||
#ifdef CONFIG_RCU_TRACE
|
#ifdef CONFIG_RCU_TRACE
|
||||||
unsigned long n_grace_periods;
|
unsigned long n_grace_periods;
|
||||||
#ifdef CONFIG_RCU_BOOST
|
#ifdef CONFIG_RCU_BOOST
|
||||||
unsigned long n_tasks_boosted;
|
unsigned long n_tasks_boosted;
|
||||||
|
/* Total number of tasks boosted. */
|
||||||
unsigned long n_exp_boosts;
|
unsigned long n_exp_boosts;
|
||||||
|
/* Number of tasks boosted for expedited GP. */
|
||||||
unsigned long n_normal_boosts;
|
unsigned long n_normal_boosts;
|
||||||
unsigned long n_normal_balk_blkd_tasks;
|
/* Number of tasks boosted for normal GP. */
|
||||||
unsigned long n_normal_balk_gp_tasks;
|
unsigned long n_balk_blkd_tasks;
|
||||||
unsigned long n_normal_balk_boost_tasks;
|
/* Refused to boost: no blocked tasks. */
|
||||||
unsigned long n_normal_balk_boosted;
|
unsigned long n_balk_exp_gp_tasks;
|
||||||
unsigned long n_normal_balk_notyet;
|
/* Refused to boost: nothing blocking GP. */
|
||||||
unsigned long n_normal_balk_nos;
|
unsigned long n_balk_boost_tasks;
|
||||||
unsigned long n_exp_balk_blkd_tasks;
|
/* Refused to boost: already boosting. */
|
||||||
unsigned long n_exp_balk_nos;
|
unsigned long n_balk_notyet;
|
||||||
|
/* Refused to boost: not yet time. */
|
||||||
|
unsigned long n_balk_nos;
|
||||||
|
/* Refused to boost: not sure why, though. */
|
||||||
|
/* This can happen due to race conditions. */
|
||||||
#endif /* #ifdef CONFIG_RCU_BOOST */
|
#endif /* #ifdef CONFIG_RCU_BOOST */
|
||||||
#endif /* #ifdef CONFIG_RCU_TRACE */
|
#endif /* #ifdef CONFIG_RCU_TRACE */
|
||||||
};
|
};
|
||||||
@@ -201,7 +206,6 @@ static struct list_head *rcu_next_node_entry(struct task_struct *t)
|
|||||||
|
|
||||||
#ifdef CONFIG_RCU_BOOST
|
#ifdef CONFIG_RCU_BOOST
|
||||||
static void rcu_initiate_boost_trace(void);
|
static void rcu_initiate_boost_trace(void);
|
||||||
static void rcu_initiate_exp_boost_trace(void);
|
|
||||||
#endif /* #ifdef CONFIG_RCU_BOOST */
|
#endif /* #ifdef CONFIG_RCU_BOOST */
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@@ -219,41 +223,21 @@ static void show_tiny_preempt_stats(struct seq_file *m)
|
|||||||
"N."[!rcu_preempt_ctrlblk.gp_tasks],
|
"N."[!rcu_preempt_ctrlblk.gp_tasks],
|
||||||
"E."[!rcu_preempt_ctrlblk.exp_tasks]);
|
"E."[!rcu_preempt_ctrlblk.exp_tasks]);
|
||||||
#ifdef CONFIG_RCU_BOOST
|
#ifdef CONFIG_RCU_BOOST
|
||||||
seq_printf(m, " ttb=%c btg=",
|
seq_printf(m, "%sttb=%c ntb=%lu neb=%lu nnb=%lu j=%04x bt=%04x\n",
|
||||||
"B."[!rcu_preempt_ctrlblk.boost_tasks]);
|
" ",
|
||||||
switch (rcu_preempt_ctrlblk.boosted_this_gp) {
|
"B."[!rcu_preempt_ctrlblk.boost_tasks],
|
||||||
case -1:
|
|
||||||
seq_puts(m, "exp");
|
|
||||||
break;
|
|
||||||
case 0:
|
|
||||||
seq_puts(m, "no");
|
|
||||||
break;
|
|
||||||
case 1:
|
|
||||||
seq_puts(m, "begun");
|
|
||||||
break;
|
|
||||||
case 2:
|
|
||||||
seq_puts(m, "done");
|
|
||||||
break;
|
|
||||||
default:
|
|
||||||
seq_printf(m, "?%d?", rcu_preempt_ctrlblk.boosted_this_gp);
|
|
||||||
}
|
|
||||||
seq_printf(m, " ntb=%lu neb=%lu nnb=%lu j=%04x bt=%04x\n",
|
|
||||||
rcu_preempt_ctrlblk.n_tasks_boosted,
|
rcu_preempt_ctrlblk.n_tasks_boosted,
|
||||||
rcu_preempt_ctrlblk.n_exp_boosts,
|
rcu_preempt_ctrlblk.n_exp_boosts,
|
||||||
rcu_preempt_ctrlblk.n_normal_boosts,
|
rcu_preempt_ctrlblk.n_normal_boosts,
|
||||||
(int)(jiffies & 0xffff),
|
(int)(jiffies & 0xffff),
|
||||||
(int)(rcu_preempt_ctrlblk.boost_time & 0xffff));
|
(int)(rcu_preempt_ctrlblk.boost_time & 0xffff));
|
||||||
seq_printf(m, " %s: nt=%lu gt=%lu bt=%lu b=%lu ny=%lu nos=%lu\n",
|
seq_printf(m, "%s: nt=%lu egt=%lu bt=%lu ny=%lu nos=%lu\n",
|
||||||
"normal balk",
|
" balk",
|
||||||
rcu_preempt_ctrlblk.n_normal_balk_blkd_tasks,
|
rcu_preempt_ctrlblk.n_balk_blkd_tasks,
|
||||||
rcu_preempt_ctrlblk.n_normal_balk_gp_tasks,
|
rcu_preempt_ctrlblk.n_balk_exp_gp_tasks,
|
||||||
rcu_preempt_ctrlblk.n_normal_balk_boost_tasks,
|
rcu_preempt_ctrlblk.n_balk_boost_tasks,
|
||||||
rcu_preempt_ctrlblk.n_normal_balk_boosted,
|
rcu_preempt_ctrlblk.n_balk_notyet,
|
||||||
rcu_preempt_ctrlblk.n_normal_balk_notyet,
|
rcu_preempt_ctrlblk.n_balk_nos);
|
||||||
rcu_preempt_ctrlblk.n_normal_balk_nos);
|
|
||||||
seq_printf(m, " exp balk: bt=%lu nos=%lu\n",
|
|
||||||
rcu_preempt_ctrlblk.n_exp_balk_blkd_tasks,
|
|
||||||
rcu_preempt_ctrlblk.n_exp_balk_nos);
|
|
||||||
#endif /* #ifdef CONFIG_RCU_BOOST */
|
#endif /* #ifdef CONFIG_RCU_BOOST */
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -271,25 +255,59 @@ static int rcu_boost(void)
|
|||||||
{
|
{
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
struct rt_mutex mtx;
|
struct rt_mutex mtx;
|
||||||
struct list_head *np;
|
|
||||||
struct task_struct *t;
|
struct task_struct *t;
|
||||||
|
struct list_head *tb;
|
||||||
|
|
||||||
if (rcu_preempt_ctrlblk.boost_tasks == NULL)
|
if (rcu_preempt_ctrlblk.boost_tasks == NULL &&
|
||||||
|
rcu_preempt_ctrlblk.exp_tasks == NULL)
|
||||||
return 0; /* Nothing to boost. */
|
return 0; /* Nothing to boost. */
|
||||||
|
|
||||||
raw_local_irq_save(flags);
|
raw_local_irq_save(flags);
|
||||||
rcu_preempt_ctrlblk.boosted_this_gp++;
|
|
||||||
t = container_of(rcu_preempt_ctrlblk.boost_tasks, struct task_struct,
|
/*
|
||||||
rcu_node_entry);
|
* Recheck with irqs disabled: all tasks in need of boosting
|
||||||
np = rcu_next_node_entry(t);
|
* might exit their RCU read-side critical sections on their own
|
||||||
|
* if we are preempted just before disabling irqs.
|
||||||
|
*/
|
||||||
|
if (rcu_preempt_ctrlblk.boost_tasks == NULL &&
|
||||||
|
rcu_preempt_ctrlblk.exp_tasks == NULL) {
|
||||||
|
raw_local_irq_restore(flags);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Preferentially boost tasks blocking expedited grace periods.
|
||||||
|
* This cannot starve the normal grace periods because a second
|
||||||
|
* expedited grace period must boost all blocked tasks, including
|
||||||
|
* those blocking the pre-existing normal grace period.
|
||||||
|
*/
|
||||||
|
if (rcu_preempt_ctrlblk.exp_tasks != NULL) {
|
||||||
|
tb = rcu_preempt_ctrlblk.exp_tasks;
|
||||||
|
RCU_TRACE(rcu_preempt_ctrlblk.n_exp_boosts++);
|
||||||
|
} else {
|
||||||
|
tb = rcu_preempt_ctrlblk.boost_tasks;
|
||||||
|
RCU_TRACE(rcu_preempt_ctrlblk.n_normal_boosts++);
|
||||||
|
}
|
||||||
|
RCU_TRACE(rcu_preempt_ctrlblk.n_tasks_boosted++);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* We boost task t by manufacturing an rt_mutex that appears to
|
||||||
|
* be held by task t. We leave a pointer to that rt_mutex where
|
||||||
|
* task t can find it, and task t will release the mutex when it
|
||||||
|
* exits its outermost RCU read-side critical section. Then
|
||||||
|
* simply acquiring this artificial rt_mutex will boost task
|
||||||
|
* t's priority. (Thanks to tglx for suggesting this approach!)
|
||||||
|
*/
|
||||||
|
t = container_of(tb, struct task_struct, rcu_node_entry);
|
||||||
rt_mutex_init_proxy_locked(&mtx, t);
|
rt_mutex_init_proxy_locked(&mtx, t);
|
||||||
t->rcu_boost_mutex = &mtx;
|
t->rcu_boost_mutex = &mtx;
|
||||||
t->rcu_read_unlock_special |= RCU_READ_UNLOCK_BOOSTED;
|
t->rcu_read_unlock_special |= RCU_READ_UNLOCK_BOOSTED;
|
||||||
raw_local_irq_restore(flags);
|
raw_local_irq_restore(flags);
|
||||||
rt_mutex_lock(&mtx);
|
rt_mutex_lock(&mtx);
|
||||||
RCU_TRACE(rcu_preempt_ctrlblk.n_tasks_boosted++);
|
rt_mutex_unlock(&mtx); /* Keep lockdep happy. */
|
||||||
rcu_preempt_ctrlblk.boosted_this_gp++;
|
|
||||||
rt_mutex_unlock(&mtx);
|
return rcu_preempt_ctrlblk.boost_tasks != NULL ||
|
||||||
return rcu_preempt_ctrlblk.boost_tasks != NULL;
|
rcu_preempt_ctrlblk.exp_tasks != NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@@ -304,42 +322,25 @@ static int rcu_boost(void)
|
|||||||
*/
|
*/
|
||||||
static int rcu_initiate_boost(void)
|
static int rcu_initiate_boost(void)
|
||||||
{
|
{
|
||||||
if (!rcu_preempt_blocked_readers_cgp()) {
|
if (!rcu_preempt_blocked_readers_cgp() &&
|
||||||
RCU_TRACE(rcu_preempt_ctrlblk.n_normal_balk_blkd_tasks++);
|
rcu_preempt_ctrlblk.exp_tasks == NULL) {
|
||||||
|
RCU_TRACE(rcu_preempt_ctrlblk.n_balk_exp_gp_tasks++);
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
if (rcu_preempt_ctrlblk.gp_tasks != NULL &&
|
if (rcu_preempt_ctrlblk.exp_tasks != NULL ||
|
||||||
|
(rcu_preempt_ctrlblk.gp_tasks != NULL &&
|
||||||
rcu_preempt_ctrlblk.boost_tasks == NULL &&
|
rcu_preempt_ctrlblk.boost_tasks == NULL &&
|
||||||
rcu_preempt_ctrlblk.boosted_this_gp == 0 &&
|
ULONG_CMP_GE(jiffies, rcu_preempt_ctrlblk.boost_time))) {
|
||||||
ULONG_CMP_GE(jiffies, rcu_preempt_ctrlblk.boost_time)) {
|
if (rcu_preempt_ctrlblk.exp_tasks == NULL)
|
||||||
rcu_preempt_ctrlblk.boost_tasks = rcu_preempt_ctrlblk.gp_tasks;
|
rcu_preempt_ctrlblk.boost_tasks =
|
||||||
|
rcu_preempt_ctrlblk.gp_tasks;
|
||||||
invoke_rcu_kthread();
|
invoke_rcu_kthread();
|
||||||
RCU_TRACE(rcu_preempt_ctrlblk.n_normal_boosts++);
|
|
||||||
} else
|
} else
|
||||||
RCU_TRACE(rcu_initiate_boost_trace());
|
RCU_TRACE(rcu_initiate_boost_trace());
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
#define RCU_BOOST_DELAY_JIFFIES DIV_ROUND_UP(CONFIG_RCU_BOOST_DELAY * HZ, 1000)
|
||||||
* Initiate boosting for an expedited grace period.
|
|
||||||
*/
|
|
||||||
static void rcu_initiate_expedited_boost(void)
|
|
||||||
{
|
|
||||||
unsigned long flags;
|
|
||||||
|
|
||||||
raw_local_irq_save(flags);
|
|
||||||
if (!list_empty(&rcu_preempt_ctrlblk.blkd_tasks)) {
|
|
||||||
rcu_preempt_ctrlblk.boost_tasks =
|
|
||||||
rcu_preempt_ctrlblk.blkd_tasks.next;
|
|
||||||
rcu_preempt_ctrlblk.boosted_this_gp = -1;
|
|
||||||
invoke_rcu_kthread();
|
|
||||||
RCU_TRACE(rcu_preempt_ctrlblk.n_exp_boosts++);
|
|
||||||
} else
|
|
||||||
RCU_TRACE(rcu_initiate_exp_boost_trace());
|
|
||||||
raw_local_irq_restore(flags);
|
|
||||||
}
|
|
||||||
|
|
||||||
#define RCU_BOOST_DELAY_JIFFIES DIV_ROUND_UP(CONFIG_RCU_BOOST_DELAY * HZ, 1000);
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Do priority-boost accounting for the start of a new grace period.
|
* Do priority-boost accounting for the start of a new grace period.
|
||||||
@@ -347,8 +348,6 @@ static void rcu_initiate_expedited_boost(void)
|
|||||||
static void rcu_preempt_boost_start_gp(void)
|
static void rcu_preempt_boost_start_gp(void)
|
||||||
{
|
{
|
||||||
rcu_preempt_ctrlblk.boost_time = jiffies + RCU_BOOST_DELAY_JIFFIES;
|
rcu_preempt_ctrlblk.boost_time = jiffies + RCU_BOOST_DELAY_JIFFIES;
|
||||||
if (rcu_preempt_ctrlblk.boosted_this_gp > 0)
|
|
||||||
rcu_preempt_ctrlblk.boosted_this_gp = 0;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#else /* #ifdef CONFIG_RCU_BOOST */
|
#else /* #ifdef CONFIG_RCU_BOOST */
|
||||||
@@ -371,13 +370,6 @@ static int rcu_initiate_boost(void)
|
|||||||
return rcu_preempt_blocked_readers_cgp();
|
return rcu_preempt_blocked_readers_cgp();
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
|
||||||
* If there is no RCU priority boosting, we don't initiate expedited boosting.
|
|
||||||
*/
|
|
||||||
static void rcu_initiate_expedited_boost(void)
|
|
||||||
{
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* If there is no RCU priority boosting, nothing to do at grace-period start.
|
* If there is no RCU priority boosting, nothing to do at grace-period start.
|
||||||
*/
|
*/
|
||||||
@@ -418,7 +410,7 @@ static void rcu_preempt_cpu_qs(void)
|
|||||||
if (!rcu_preempt_gp_in_progress())
|
if (!rcu_preempt_gp_in_progress())
|
||||||
return;
|
return;
|
||||||
/*
|
/*
|
||||||
* Check up on boosting. If there are no readers blocking the
|
* Check up on boosting. If there are readers blocking the
|
||||||
* current grace period, leave.
|
* current grace period, leave.
|
||||||
*/
|
*/
|
||||||
if (rcu_initiate_boost())
|
if (rcu_initiate_boost())
|
||||||
@@ -578,7 +570,7 @@ static void rcu_read_unlock_special(struct task_struct *t)
|
|||||||
empty = !rcu_preempt_blocked_readers_cgp();
|
empty = !rcu_preempt_blocked_readers_cgp();
|
||||||
empty_exp = rcu_preempt_ctrlblk.exp_tasks == NULL;
|
empty_exp = rcu_preempt_ctrlblk.exp_tasks == NULL;
|
||||||
np = rcu_next_node_entry(t);
|
np = rcu_next_node_entry(t);
|
||||||
list_del(&t->rcu_node_entry);
|
list_del_init(&t->rcu_node_entry);
|
||||||
if (&t->rcu_node_entry == rcu_preempt_ctrlblk.gp_tasks)
|
if (&t->rcu_node_entry == rcu_preempt_ctrlblk.gp_tasks)
|
||||||
rcu_preempt_ctrlblk.gp_tasks = np;
|
rcu_preempt_ctrlblk.gp_tasks = np;
|
||||||
if (&t->rcu_node_entry == rcu_preempt_ctrlblk.exp_tasks)
|
if (&t->rcu_node_entry == rcu_preempt_ctrlblk.exp_tasks)
|
||||||
@@ -587,7 +579,6 @@ static void rcu_read_unlock_special(struct task_struct *t)
|
|||||||
if (&t->rcu_node_entry == rcu_preempt_ctrlblk.boost_tasks)
|
if (&t->rcu_node_entry == rcu_preempt_ctrlblk.boost_tasks)
|
||||||
rcu_preempt_ctrlblk.boost_tasks = np;
|
rcu_preempt_ctrlblk.boost_tasks = np;
|
||||||
#endif /* #ifdef CONFIG_RCU_BOOST */
|
#endif /* #ifdef CONFIG_RCU_BOOST */
|
||||||
INIT_LIST_HEAD(&t->rcu_node_entry);
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* If this was the last task on the current list, and if
|
* If this was the last task on the current list, and if
|
||||||
@@ -812,13 +803,16 @@ void synchronize_rcu_expedited(void)
|
|||||||
rpcp->exp_tasks = rpcp->blkd_tasks.next;
|
rpcp->exp_tasks = rpcp->blkd_tasks.next;
|
||||||
if (rpcp->exp_tasks == &rpcp->blkd_tasks)
|
if (rpcp->exp_tasks == &rpcp->blkd_tasks)
|
||||||
rpcp->exp_tasks = NULL;
|
rpcp->exp_tasks = NULL;
|
||||||
local_irq_restore(flags);
|
|
||||||
|
|
||||||
/* Wait for tail of ->blkd_tasks list to drain. */
|
/* Wait for tail of ->blkd_tasks list to drain. */
|
||||||
if (rcu_preempted_readers_exp())
|
if (!rcu_preempted_readers_exp())
|
||||||
rcu_initiate_expedited_boost();
|
local_irq_restore(flags);
|
||||||
|
else {
|
||||||
|
rcu_initiate_boost();
|
||||||
|
local_irq_restore(flags);
|
||||||
wait_event(sync_rcu_preempt_exp_wq,
|
wait_event(sync_rcu_preempt_exp_wq,
|
||||||
!rcu_preempted_readers_exp());
|
!rcu_preempted_readers_exp());
|
||||||
|
}
|
||||||
|
|
||||||
/* Clean up and exit. */
|
/* Clean up and exit. */
|
||||||
barrier(); /* ensure expedited GP seen before counter increment. */
|
barrier(); /* ensure expedited GP seen before counter increment. */
|
||||||
@@ -930,25 +924,18 @@ void __init rcu_scheduler_starting(void)
|
|||||||
#ifdef CONFIG_RCU_BOOST
|
#ifdef CONFIG_RCU_BOOST
|
||||||
|
|
||||||
static void rcu_initiate_boost_trace(void)
|
static void rcu_initiate_boost_trace(void)
|
||||||
{
|
|
||||||
if (rcu_preempt_ctrlblk.gp_tasks == NULL)
|
|
||||||
rcu_preempt_ctrlblk.n_normal_balk_gp_tasks++;
|
|
||||||
else if (rcu_preempt_ctrlblk.boost_tasks != NULL)
|
|
||||||
rcu_preempt_ctrlblk.n_normal_balk_boost_tasks++;
|
|
||||||
else if (rcu_preempt_ctrlblk.boosted_this_gp != 0)
|
|
||||||
rcu_preempt_ctrlblk.n_normal_balk_boosted++;
|
|
||||||
else if (!ULONG_CMP_GE(jiffies, rcu_preempt_ctrlblk.boost_time))
|
|
||||||
rcu_preempt_ctrlblk.n_normal_balk_notyet++;
|
|
||||||
else
|
|
||||||
rcu_preempt_ctrlblk.n_normal_balk_nos++;
|
|
||||||
}
|
|
||||||
|
|
||||||
static void rcu_initiate_exp_boost_trace(void)
|
|
||||||
{
|
{
|
||||||
if (list_empty(&rcu_preempt_ctrlblk.blkd_tasks))
|
if (list_empty(&rcu_preempt_ctrlblk.blkd_tasks))
|
||||||
rcu_preempt_ctrlblk.n_exp_balk_blkd_tasks++;
|
rcu_preempt_ctrlblk.n_balk_blkd_tasks++;
|
||||||
|
else if (rcu_preempt_ctrlblk.gp_tasks == NULL &&
|
||||||
|
rcu_preempt_ctrlblk.exp_tasks == NULL)
|
||||||
|
rcu_preempt_ctrlblk.n_balk_exp_gp_tasks++;
|
||||||
|
else if (rcu_preempt_ctrlblk.boost_tasks != NULL)
|
||||||
|
rcu_preempt_ctrlblk.n_balk_boost_tasks++;
|
||||||
|
else if (!ULONG_CMP_GE(jiffies, rcu_preempt_ctrlblk.boost_time))
|
||||||
|
rcu_preempt_ctrlblk.n_balk_notyet++;
|
||||||
else
|
else
|
||||||
rcu_preempt_ctrlblk.n_exp_balk_nos++;
|
rcu_preempt_ctrlblk.n_balk_nos++;
|
||||||
}
|
}
|
||||||
|
|
||||||
#endif /* #ifdef CONFIG_RCU_BOOST */
|
#endif /* #ifdef CONFIG_RCU_BOOST */
|
||||||
|
@@ -131,7 +131,7 @@ struct rcu_torture {
|
|||||||
|
|
||||||
static LIST_HEAD(rcu_torture_freelist);
|
static LIST_HEAD(rcu_torture_freelist);
|
||||||
static struct rcu_torture __rcu *rcu_torture_current;
|
static struct rcu_torture __rcu *rcu_torture_current;
|
||||||
static long rcu_torture_current_version;
|
static unsigned long rcu_torture_current_version;
|
||||||
static struct rcu_torture rcu_tortures[10 * RCU_TORTURE_PIPE_LEN];
|
static struct rcu_torture rcu_tortures[10 * RCU_TORTURE_PIPE_LEN];
|
||||||
static DEFINE_SPINLOCK(rcu_torture_lock);
|
static DEFINE_SPINLOCK(rcu_torture_lock);
|
||||||
static DEFINE_PER_CPU(long [RCU_TORTURE_PIPE_LEN + 1], rcu_torture_count) =
|
static DEFINE_PER_CPU(long [RCU_TORTURE_PIPE_LEN + 1], rcu_torture_count) =
|
||||||
@@ -146,8 +146,6 @@ static atomic_t n_rcu_torture_mberror;
|
|||||||
static atomic_t n_rcu_torture_error;
|
static atomic_t n_rcu_torture_error;
|
||||||
static long n_rcu_torture_boost_ktrerror;
|
static long n_rcu_torture_boost_ktrerror;
|
||||||
static long n_rcu_torture_boost_rterror;
|
static long n_rcu_torture_boost_rterror;
|
||||||
static long n_rcu_torture_boost_allocerror;
|
|
||||||
static long n_rcu_torture_boost_afferror;
|
|
||||||
static long n_rcu_torture_boost_failure;
|
static long n_rcu_torture_boost_failure;
|
||||||
static long n_rcu_torture_boosts;
|
static long n_rcu_torture_boosts;
|
||||||
static long n_rcu_torture_timers;
|
static long n_rcu_torture_timers;
|
||||||
@@ -163,11 +161,11 @@ static int stutter_pause_test;
|
|||||||
#endif
|
#endif
|
||||||
int rcutorture_runnable = RCUTORTURE_RUNNABLE_INIT;
|
int rcutorture_runnable = RCUTORTURE_RUNNABLE_INIT;
|
||||||
|
|
||||||
#ifdef CONFIG_RCU_BOOST
|
#if defined(CONFIG_RCU_BOOST) && !defined(CONFIG_HOTPLUG_CPU)
|
||||||
#define rcu_can_boost() 1
|
#define rcu_can_boost() 1
|
||||||
#else /* #ifdef CONFIG_RCU_BOOST */
|
#else /* #if defined(CONFIG_RCU_BOOST) && !defined(CONFIG_HOTPLUG_CPU) */
|
||||||
#define rcu_can_boost() 0
|
#define rcu_can_boost() 0
|
||||||
#endif /* #else #ifdef CONFIG_RCU_BOOST */
|
#endif /* #else #if defined(CONFIG_RCU_BOOST) && !defined(CONFIG_HOTPLUG_CPU) */
|
||||||
|
|
||||||
static unsigned long boost_starttime; /* jiffies of next boost test start. */
|
static unsigned long boost_starttime; /* jiffies of next boost test start. */
|
||||||
DEFINE_MUTEX(boost_mutex); /* protect setting boost_starttime */
|
DEFINE_MUTEX(boost_mutex); /* protect setting boost_starttime */
|
||||||
@@ -751,6 +749,7 @@ static int rcu_torture_boost(void *arg)
|
|||||||
n_rcu_torture_boost_rterror++;
|
n_rcu_torture_boost_rterror++;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
init_rcu_head_on_stack(&rbi.rcu);
|
||||||
/* Each pass through the following loop does one boost-test cycle. */
|
/* Each pass through the following loop does one boost-test cycle. */
|
||||||
do {
|
do {
|
||||||
/* Wait for the next test interval. */
|
/* Wait for the next test interval. */
|
||||||
@@ -810,6 +809,7 @@ checkwait: rcu_stutter_wait("rcu_torture_boost");
|
|||||||
|
|
||||||
/* Clean up and exit. */
|
/* Clean up and exit. */
|
||||||
VERBOSE_PRINTK_STRING("rcu_torture_boost task stopping");
|
VERBOSE_PRINTK_STRING("rcu_torture_boost task stopping");
|
||||||
|
destroy_rcu_head_on_stack(&rbi.rcu);
|
||||||
rcutorture_shutdown_absorb("rcu_torture_boost");
|
rcutorture_shutdown_absorb("rcu_torture_boost");
|
||||||
while (!kthread_should_stop() || rbi.inflight)
|
while (!kthread_should_stop() || rbi.inflight)
|
||||||
schedule_timeout_uninterruptible(1);
|
schedule_timeout_uninterruptible(1);
|
||||||
@@ -886,7 +886,7 @@ rcu_torture_writer(void *arg)
|
|||||||
old_rp->rtort_pipe_count++;
|
old_rp->rtort_pipe_count++;
|
||||||
cur_ops->deferred_free(old_rp);
|
cur_ops->deferred_free(old_rp);
|
||||||
}
|
}
|
||||||
rcu_torture_current_version++;
|
rcutorture_record_progress(++rcu_torture_current_version);
|
||||||
oldbatch = cur_ops->completed();
|
oldbatch = cur_ops->completed();
|
||||||
rcu_stutter_wait("rcu_torture_writer");
|
rcu_stutter_wait("rcu_torture_writer");
|
||||||
} while (!kthread_should_stop() && fullstop == FULLSTOP_DONTSTOP);
|
} while (!kthread_should_stop() && fullstop == FULLSTOP_DONTSTOP);
|
||||||
@@ -1066,8 +1066,8 @@ rcu_torture_printk(char *page)
|
|||||||
}
|
}
|
||||||
cnt += sprintf(&page[cnt], "%s%s ", torture_type, TORTURE_FLAG);
|
cnt += sprintf(&page[cnt], "%s%s ", torture_type, TORTURE_FLAG);
|
||||||
cnt += sprintf(&page[cnt],
|
cnt += sprintf(&page[cnt],
|
||||||
"rtc: %p ver: %ld tfle: %d rta: %d rtaf: %d rtf: %d "
|
"rtc: %p ver: %lu tfle: %d rta: %d rtaf: %d rtf: %d "
|
||||||
"rtmbe: %d rtbke: %ld rtbre: %ld rtbae: %ld rtbafe: %ld "
|
"rtmbe: %d rtbke: %ld rtbre: %ld "
|
||||||
"rtbf: %ld rtb: %ld nt: %ld",
|
"rtbf: %ld rtb: %ld nt: %ld",
|
||||||
rcu_torture_current,
|
rcu_torture_current,
|
||||||
rcu_torture_current_version,
|
rcu_torture_current_version,
|
||||||
@@ -1078,16 +1078,12 @@ rcu_torture_printk(char *page)
|
|||||||
atomic_read(&n_rcu_torture_mberror),
|
atomic_read(&n_rcu_torture_mberror),
|
||||||
n_rcu_torture_boost_ktrerror,
|
n_rcu_torture_boost_ktrerror,
|
||||||
n_rcu_torture_boost_rterror,
|
n_rcu_torture_boost_rterror,
|
||||||
n_rcu_torture_boost_allocerror,
|
|
||||||
n_rcu_torture_boost_afferror,
|
|
||||||
n_rcu_torture_boost_failure,
|
n_rcu_torture_boost_failure,
|
||||||
n_rcu_torture_boosts,
|
n_rcu_torture_boosts,
|
||||||
n_rcu_torture_timers);
|
n_rcu_torture_timers);
|
||||||
if (atomic_read(&n_rcu_torture_mberror) != 0 ||
|
if (atomic_read(&n_rcu_torture_mberror) != 0 ||
|
||||||
n_rcu_torture_boost_ktrerror != 0 ||
|
n_rcu_torture_boost_ktrerror != 0 ||
|
||||||
n_rcu_torture_boost_rterror != 0 ||
|
n_rcu_torture_boost_rterror != 0 ||
|
||||||
n_rcu_torture_boost_allocerror != 0 ||
|
|
||||||
n_rcu_torture_boost_afferror != 0 ||
|
|
||||||
n_rcu_torture_boost_failure != 0)
|
n_rcu_torture_boost_failure != 0)
|
||||||
cnt += sprintf(&page[cnt], " !!!");
|
cnt += sprintf(&page[cnt], " !!!");
|
||||||
cnt += sprintf(&page[cnt], "\n%s%s ", torture_type, TORTURE_FLAG);
|
cnt += sprintf(&page[cnt], "\n%s%s ", torture_type, TORTURE_FLAG);
|
||||||
@@ -1331,6 +1327,7 @@ rcu_torture_cleanup(void)
|
|||||||
int i;
|
int i;
|
||||||
|
|
||||||
mutex_lock(&fullstop_mutex);
|
mutex_lock(&fullstop_mutex);
|
||||||
|
rcutorture_record_test_transition();
|
||||||
if (fullstop == FULLSTOP_SHUTDOWN) {
|
if (fullstop == FULLSTOP_SHUTDOWN) {
|
||||||
printk(KERN_WARNING /* but going down anyway, so... */
|
printk(KERN_WARNING /* but going down anyway, so... */
|
||||||
"Concurrent 'rmmod rcutorture' and shutdown illegal!\n");
|
"Concurrent 'rmmod rcutorture' and shutdown illegal!\n");
|
||||||
@@ -1486,8 +1483,6 @@ rcu_torture_init(void)
|
|||||||
atomic_set(&n_rcu_torture_error, 0);
|
atomic_set(&n_rcu_torture_error, 0);
|
||||||
n_rcu_torture_boost_ktrerror = 0;
|
n_rcu_torture_boost_ktrerror = 0;
|
||||||
n_rcu_torture_boost_rterror = 0;
|
n_rcu_torture_boost_rterror = 0;
|
||||||
n_rcu_torture_boost_allocerror = 0;
|
|
||||||
n_rcu_torture_boost_afferror = 0;
|
|
||||||
n_rcu_torture_boost_failure = 0;
|
n_rcu_torture_boost_failure = 0;
|
||||||
n_rcu_torture_boosts = 0;
|
n_rcu_torture_boosts = 0;
|
||||||
for (i = 0; i < RCU_TORTURE_PIPE_LEN + 1; i++)
|
for (i = 0; i < RCU_TORTURE_PIPE_LEN + 1; i++)
|
||||||
@@ -1624,6 +1619,7 @@ rcu_torture_init(void)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
register_reboot_notifier(&rcutorture_shutdown_nb);
|
register_reboot_notifier(&rcutorture_shutdown_nb);
|
||||||
|
rcutorture_record_test_transition();
|
||||||
mutex_unlock(&fullstop_mutex);
|
mutex_unlock(&fullstop_mutex);
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
|
672
kernel/rcutree.c
672
kernel/rcutree.c
@@ -47,6 +47,8 @@
|
|||||||
#include <linux/mutex.h>
|
#include <linux/mutex.h>
|
||||||
#include <linux/time.h>
|
#include <linux/time.h>
|
||||||
#include <linux/kernel_stat.h>
|
#include <linux/kernel_stat.h>
|
||||||
|
#include <linux/wait.h>
|
||||||
|
#include <linux/kthread.h>
|
||||||
|
|
||||||
#include "rcutree.h"
|
#include "rcutree.h"
|
||||||
|
|
||||||
@@ -79,9 +81,40 @@ DEFINE_PER_CPU(struct rcu_data, rcu_sched_data);
|
|||||||
struct rcu_state rcu_bh_state = RCU_STATE_INITIALIZER(rcu_bh_state);
|
struct rcu_state rcu_bh_state = RCU_STATE_INITIALIZER(rcu_bh_state);
|
||||||
DEFINE_PER_CPU(struct rcu_data, rcu_bh_data);
|
DEFINE_PER_CPU(struct rcu_data, rcu_bh_data);
|
||||||
|
|
||||||
|
static struct rcu_state *rcu_state;
|
||||||
|
|
||||||
int rcu_scheduler_active __read_mostly;
|
int rcu_scheduler_active __read_mostly;
|
||||||
EXPORT_SYMBOL_GPL(rcu_scheduler_active);
|
EXPORT_SYMBOL_GPL(rcu_scheduler_active);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Control variables for per-CPU and per-rcu_node kthreads. These
|
||||||
|
* handle all flavors of RCU.
|
||||||
|
*/
|
||||||
|
static DEFINE_PER_CPU(struct task_struct *, rcu_cpu_kthread_task);
|
||||||
|
DEFINE_PER_CPU(unsigned int, rcu_cpu_kthread_status);
|
||||||
|
DEFINE_PER_CPU(int, rcu_cpu_kthread_cpu);
|
||||||
|
DEFINE_PER_CPU(unsigned int, rcu_cpu_kthread_loops);
|
||||||
|
static DEFINE_PER_CPU(wait_queue_head_t, rcu_cpu_wq);
|
||||||
|
DEFINE_PER_CPU(char, rcu_cpu_has_work);
|
||||||
|
static char rcu_kthreads_spawnable;
|
||||||
|
|
||||||
|
static void rcu_node_kthread_setaffinity(struct rcu_node *rnp, int outgoingcpu);
|
||||||
|
static void invoke_rcu_cpu_kthread(void);
|
||||||
|
|
||||||
|
#define RCU_KTHREAD_PRIO 1 /* RT priority for per-CPU kthreads. */
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Track the rcutorture test sequence number and the update version
|
||||||
|
* number within a given test. The rcutorture_testseq is incremented
|
||||||
|
* on every rcutorture module load and unload, so has an odd value
|
||||||
|
* when a test is running. The rcutorture_vernum is set to zero
|
||||||
|
* when rcutorture starts and is incremented on each rcutorture update.
|
||||||
|
* These variables enable correlating rcutorture output with the
|
||||||
|
* RCU tracing information.
|
||||||
|
*/
|
||||||
|
unsigned long rcutorture_testseq;
|
||||||
|
unsigned long rcutorture_vernum;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Return true if an RCU grace period is in progress. The ACCESS_ONCE()s
|
* Return true if an RCU grace period is in progress. The ACCESS_ONCE()s
|
||||||
* permit this function to be invoked without holding the root rcu_node
|
* permit this function to be invoked without holding the root rcu_node
|
||||||
@@ -124,11 +157,12 @@ void rcu_note_context_switch(int cpu)
|
|||||||
rcu_sched_qs(cpu);
|
rcu_sched_qs(cpu);
|
||||||
rcu_preempt_note_context_switch(cpu);
|
rcu_preempt_note_context_switch(cpu);
|
||||||
}
|
}
|
||||||
|
EXPORT_SYMBOL_GPL(rcu_note_context_switch);
|
||||||
|
|
||||||
#ifdef CONFIG_NO_HZ
|
#ifdef CONFIG_NO_HZ
|
||||||
DEFINE_PER_CPU(struct rcu_dynticks, rcu_dynticks) = {
|
DEFINE_PER_CPU(struct rcu_dynticks, rcu_dynticks) = {
|
||||||
.dynticks_nesting = 1,
|
.dynticks_nesting = 1,
|
||||||
.dynticks = 1,
|
.dynticks = ATOMIC_INIT(1),
|
||||||
};
|
};
|
||||||
#endif /* #ifdef CONFIG_NO_HZ */
|
#endif /* #ifdef CONFIG_NO_HZ */
|
||||||
|
|
||||||
@@ -140,10 +174,8 @@ module_param(blimit, int, 0);
|
|||||||
module_param(qhimark, int, 0);
|
module_param(qhimark, int, 0);
|
||||||
module_param(qlowmark, int, 0);
|
module_param(qlowmark, int, 0);
|
||||||
|
|
||||||
#ifdef CONFIG_RCU_CPU_STALL_DETECTOR
|
int rcu_cpu_stall_suppress __read_mostly;
|
||||||
int rcu_cpu_stall_suppress __read_mostly = RCU_CPU_STALL_SUPPRESS_INIT;
|
|
||||||
module_param(rcu_cpu_stall_suppress, int, 0644);
|
module_param(rcu_cpu_stall_suppress, int, 0644);
|
||||||
#endif /* #ifdef CONFIG_RCU_CPU_STALL_DETECTOR */
|
|
||||||
|
|
||||||
static void force_quiescent_state(struct rcu_state *rsp, int relaxed);
|
static void force_quiescent_state(struct rcu_state *rsp, int relaxed);
|
||||||
static int rcu_pending(int cpu);
|
static int rcu_pending(int cpu);
|
||||||
@@ -175,6 +207,31 @@ void rcu_bh_force_quiescent_state(void)
|
|||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(rcu_bh_force_quiescent_state);
|
EXPORT_SYMBOL_GPL(rcu_bh_force_quiescent_state);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Record the number of times rcutorture tests have been initiated and
|
||||||
|
* terminated. This information allows the debugfs tracing stats to be
|
||||||
|
* correlated to the rcutorture messages, even when the rcutorture module
|
||||||
|
* is being repeatedly loaded and unloaded. In other words, we cannot
|
||||||
|
* store this state in rcutorture itself.
|
||||||
|
*/
|
||||||
|
void rcutorture_record_test_transition(void)
|
||||||
|
{
|
||||||
|
rcutorture_testseq++;
|
||||||
|
rcutorture_vernum = 0;
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL_GPL(rcutorture_record_test_transition);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Record the number of writer passes through the current rcutorture test.
|
||||||
|
* This is also used to correlate debugfs tracing stats with the rcutorture
|
||||||
|
* messages.
|
||||||
|
*/
|
||||||
|
void rcutorture_record_progress(unsigned long vernum)
|
||||||
|
{
|
||||||
|
rcutorture_vernum++;
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL_GPL(rcutorture_record_progress);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Force a quiescent state for RCU-sched.
|
* Force a quiescent state for RCU-sched.
|
||||||
*/
|
*/
|
||||||
@@ -234,8 +291,8 @@ static int rcu_implicit_offline_qs(struct rcu_data *rdp)
|
|||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* If preemptable RCU, no point in sending reschedule IPI. */
|
/* If preemptible RCU, no point in sending reschedule IPI. */
|
||||||
if (rdp->preemptable)
|
if (rdp->preemptible)
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
/* The CPU is online, so send it a reschedule IPI. */
|
/* The CPU is online, so send it a reschedule IPI. */
|
||||||
@@ -264,13 +321,25 @@ void rcu_enter_nohz(void)
|
|||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
struct rcu_dynticks *rdtp;
|
struct rcu_dynticks *rdtp;
|
||||||
|
|
||||||
smp_mb(); /* CPUs seeing ++ must see prior RCU read-side crit sects */
|
|
||||||
local_irq_save(flags);
|
local_irq_save(flags);
|
||||||
rdtp = &__get_cpu_var(rcu_dynticks);
|
rdtp = &__get_cpu_var(rcu_dynticks);
|
||||||
rdtp->dynticks++;
|
if (--rdtp->dynticks_nesting) {
|
||||||
rdtp->dynticks_nesting--;
|
|
||||||
WARN_ON_ONCE(rdtp->dynticks & 0x1);
|
|
||||||
local_irq_restore(flags);
|
local_irq_restore(flags);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
/* CPUs seeing atomic_inc() must see prior RCU read-side crit sects */
|
||||||
|
smp_mb__before_atomic_inc(); /* See above. */
|
||||||
|
atomic_inc(&rdtp->dynticks);
|
||||||
|
smp_mb__after_atomic_inc(); /* Force ordering with next sojourn. */
|
||||||
|
WARN_ON_ONCE(atomic_read(&rdtp->dynticks) & 0x1);
|
||||||
|
local_irq_restore(flags);
|
||||||
|
|
||||||
|
/* If the interrupt queued a callback, get out of dyntick mode. */
|
||||||
|
if (in_irq() &&
|
||||||
|
(__get_cpu_var(rcu_sched_data).nxtlist ||
|
||||||
|
__get_cpu_var(rcu_bh_data).nxtlist ||
|
||||||
|
rcu_preempt_needs_cpu(smp_processor_id())))
|
||||||
|
set_need_resched();
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@@ -286,11 +355,16 @@ void rcu_exit_nohz(void)
|
|||||||
|
|
||||||
local_irq_save(flags);
|
local_irq_save(flags);
|
||||||
rdtp = &__get_cpu_var(rcu_dynticks);
|
rdtp = &__get_cpu_var(rcu_dynticks);
|
||||||
rdtp->dynticks++;
|
if (rdtp->dynticks_nesting++) {
|
||||||
rdtp->dynticks_nesting++;
|
local_irq_restore(flags);
|
||||||
WARN_ON_ONCE(!(rdtp->dynticks & 0x1));
|
return;
|
||||||
|
}
|
||||||
|
smp_mb__before_atomic_inc(); /* Force ordering w/previous sojourn. */
|
||||||
|
atomic_inc(&rdtp->dynticks);
|
||||||
|
/* CPUs seeing atomic_inc() must see later RCU read-side crit sects */
|
||||||
|
smp_mb__after_atomic_inc(); /* See above. */
|
||||||
|
WARN_ON_ONCE(!(atomic_read(&rdtp->dynticks) & 0x1));
|
||||||
local_irq_restore(flags);
|
local_irq_restore(flags);
|
||||||
smp_mb(); /* CPUs seeing ++ must see later RCU read-side crit sects */
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -304,11 +378,15 @@ void rcu_nmi_enter(void)
|
|||||||
{
|
{
|
||||||
struct rcu_dynticks *rdtp = &__get_cpu_var(rcu_dynticks);
|
struct rcu_dynticks *rdtp = &__get_cpu_var(rcu_dynticks);
|
||||||
|
|
||||||
if (rdtp->dynticks & 0x1)
|
if (rdtp->dynticks_nmi_nesting == 0 &&
|
||||||
|
(atomic_read(&rdtp->dynticks) & 0x1))
|
||||||
return;
|
return;
|
||||||
rdtp->dynticks_nmi++;
|
rdtp->dynticks_nmi_nesting++;
|
||||||
WARN_ON_ONCE(!(rdtp->dynticks_nmi & 0x1));
|
smp_mb__before_atomic_inc(); /* Force delay from prior write. */
|
||||||
smp_mb(); /* CPUs seeing ++ must see later RCU read-side crit sects */
|
atomic_inc(&rdtp->dynticks);
|
||||||
|
/* CPUs seeing atomic_inc() must see later RCU read-side crit sects */
|
||||||
|
smp_mb__after_atomic_inc(); /* See above. */
|
||||||
|
WARN_ON_ONCE(!(atomic_read(&rdtp->dynticks) & 0x1));
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -322,11 +400,14 @@ void rcu_nmi_exit(void)
|
|||||||
{
|
{
|
||||||
struct rcu_dynticks *rdtp = &__get_cpu_var(rcu_dynticks);
|
struct rcu_dynticks *rdtp = &__get_cpu_var(rcu_dynticks);
|
||||||
|
|
||||||
if (rdtp->dynticks & 0x1)
|
if (rdtp->dynticks_nmi_nesting == 0 ||
|
||||||
|
--rdtp->dynticks_nmi_nesting != 0)
|
||||||
return;
|
return;
|
||||||
smp_mb(); /* CPUs seeing ++ must see prior RCU read-side crit sects */
|
/* CPUs seeing atomic_inc() must see prior RCU read-side crit sects */
|
||||||
rdtp->dynticks_nmi++;
|
smp_mb__before_atomic_inc(); /* See above. */
|
||||||
WARN_ON_ONCE(rdtp->dynticks_nmi & 0x1);
|
atomic_inc(&rdtp->dynticks);
|
||||||
|
smp_mb__after_atomic_inc(); /* Force delay to next write. */
|
||||||
|
WARN_ON_ONCE(atomic_read(&rdtp->dynticks) & 0x1);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -337,13 +418,7 @@ void rcu_nmi_exit(void)
|
|||||||
*/
|
*/
|
||||||
void rcu_irq_enter(void)
|
void rcu_irq_enter(void)
|
||||||
{
|
{
|
||||||
struct rcu_dynticks *rdtp = &__get_cpu_var(rcu_dynticks);
|
rcu_exit_nohz();
|
||||||
|
|
||||||
if (rdtp->dynticks_nesting++)
|
|
||||||
return;
|
|
||||||
rdtp->dynticks++;
|
|
||||||
WARN_ON_ONCE(!(rdtp->dynticks & 0x1));
|
|
||||||
smp_mb(); /* CPUs seeing ++ must see later RCU read-side crit sects */
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -355,18 +430,7 @@ void rcu_irq_enter(void)
|
|||||||
*/
|
*/
|
||||||
void rcu_irq_exit(void)
|
void rcu_irq_exit(void)
|
||||||
{
|
{
|
||||||
struct rcu_dynticks *rdtp = &__get_cpu_var(rcu_dynticks);
|
rcu_enter_nohz();
|
||||||
|
|
||||||
if (--rdtp->dynticks_nesting)
|
|
||||||
return;
|
|
||||||
smp_mb(); /* CPUs seeing ++ must see prior RCU read-side crit sects */
|
|
||||||
rdtp->dynticks++;
|
|
||||||
WARN_ON_ONCE(rdtp->dynticks & 0x1);
|
|
||||||
|
|
||||||
/* If the interrupt queued a callback, get out of dyntick mode. */
|
|
||||||
if (__this_cpu_read(rcu_sched_data.nxtlist) ||
|
|
||||||
__this_cpu_read(rcu_bh_data.nxtlist))
|
|
||||||
set_need_resched();
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#ifdef CONFIG_SMP
|
#ifdef CONFIG_SMP
|
||||||
@@ -378,19 +442,8 @@ void rcu_irq_exit(void)
|
|||||||
*/
|
*/
|
||||||
static int dyntick_save_progress_counter(struct rcu_data *rdp)
|
static int dyntick_save_progress_counter(struct rcu_data *rdp)
|
||||||
{
|
{
|
||||||
int ret;
|
rdp->dynticks_snap = atomic_add_return(0, &rdp->dynticks->dynticks);
|
||||||
int snap;
|
return 0;
|
||||||
int snap_nmi;
|
|
||||||
|
|
||||||
snap = rdp->dynticks->dynticks;
|
|
||||||
snap_nmi = rdp->dynticks->dynticks_nmi;
|
|
||||||
smp_mb(); /* Order sampling of snap with end of grace period. */
|
|
||||||
rdp->dynticks_snap = snap;
|
|
||||||
rdp->dynticks_nmi_snap = snap_nmi;
|
|
||||||
ret = ((snap & 0x1) == 0) && ((snap_nmi & 0x1) == 0);
|
|
||||||
if (ret)
|
|
||||||
rdp->dynticks_fqs++;
|
|
||||||
return ret;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@@ -401,16 +454,11 @@ static int dyntick_save_progress_counter(struct rcu_data *rdp)
|
|||||||
*/
|
*/
|
||||||
static int rcu_implicit_dynticks_qs(struct rcu_data *rdp)
|
static int rcu_implicit_dynticks_qs(struct rcu_data *rdp)
|
||||||
{
|
{
|
||||||
long curr;
|
unsigned long curr;
|
||||||
long curr_nmi;
|
unsigned long snap;
|
||||||
long snap;
|
|
||||||
long snap_nmi;
|
|
||||||
|
|
||||||
curr = rdp->dynticks->dynticks;
|
curr = (unsigned long)atomic_add_return(0, &rdp->dynticks->dynticks);
|
||||||
snap = rdp->dynticks_snap;
|
snap = (unsigned long)rdp->dynticks_snap;
|
||||||
curr_nmi = rdp->dynticks->dynticks_nmi;
|
|
||||||
snap_nmi = rdp->dynticks_nmi_snap;
|
|
||||||
smp_mb(); /* force ordering with cpu entering/leaving dynticks. */
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* If the CPU passed through or entered a dynticks idle phase with
|
* If the CPU passed through or entered a dynticks idle phase with
|
||||||
@@ -420,8 +468,7 @@ static int rcu_implicit_dynticks_qs(struct rcu_data *rdp)
|
|||||||
* read-side critical section that started before the beginning
|
* read-side critical section that started before the beginning
|
||||||
* of the current RCU grace period.
|
* of the current RCU grace period.
|
||||||
*/
|
*/
|
||||||
if ((curr != snap || (curr & 0x1) == 0) &&
|
if ((curr & 0x1) == 0 || ULONG_CMP_GE(curr, snap + 2)) {
|
||||||
(curr_nmi != snap_nmi || (curr_nmi & 0x1) == 0)) {
|
|
||||||
rdp->dynticks_fqs++;
|
rdp->dynticks_fqs++;
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
@@ -450,8 +497,6 @@ static int rcu_implicit_dynticks_qs(struct rcu_data *rdp)
|
|||||||
|
|
||||||
#endif /* #else #ifdef CONFIG_NO_HZ */
|
#endif /* #else #ifdef CONFIG_NO_HZ */
|
||||||
|
|
||||||
#ifdef CONFIG_RCU_CPU_STALL_DETECTOR
|
|
||||||
|
|
||||||
int rcu_cpu_stall_suppress __read_mostly;
|
int rcu_cpu_stall_suppress __read_mostly;
|
||||||
|
|
||||||
static void record_gp_stall_check_time(struct rcu_state *rsp)
|
static void record_gp_stall_check_time(struct rcu_state *rsp)
|
||||||
@@ -537,21 +582,24 @@ static void print_cpu_stall(struct rcu_state *rsp)
|
|||||||
|
|
||||||
static void check_cpu_stall(struct rcu_state *rsp, struct rcu_data *rdp)
|
static void check_cpu_stall(struct rcu_state *rsp, struct rcu_data *rdp)
|
||||||
{
|
{
|
||||||
long delta;
|
unsigned long j;
|
||||||
|
unsigned long js;
|
||||||
struct rcu_node *rnp;
|
struct rcu_node *rnp;
|
||||||
|
|
||||||
if (rcu_cpu_stall_suppress)
|
if (rcu_cpu_stall_suppress)
|
||||||
return;
|
return;
|
||||||
delta = jiffies - ACCESS_ONCE(rsp->jiffies_stall);
|
j = ACCESS_ONCE(jiffies);
|
||||||
|
js = ACCESS_ONCE(rsp->jiffies_stall);
|
||||||
rnp = rdp->mynode;
|
rnp = rdp->mynode;
|
||||||
if ((ACCESS_ONCE(rnp->qsmask) & rdp->grpmask) && delta >= 0) {
|
if ((ACCESS_ONCE(rnp->qsmask) & rdp->grpmask) && ULONG_CMP_GE(j, js)) {
|
||||||
|
|
||||||
/* We haven't checked in, so go dump stack. */
|
/* We haven't checked in, so go dump stack. */
|
||||||
print_cpu_stall(rsp);
|
print_cpu_stall(rsp);
|
||||||
|
|
||||||
} else if (rcu_gp_in_progress(rsp) && delta >= RCU_STALL_RAT_DELAY) {
|
} else if (rcu_gp_in_progress(rsp) &&
|
||||||
|
ULONG_CMP_GE(j, js + RCU_STALL_RAT_DELAY)) {
|
||||||
|
|
||||||
/* They had two time units to dump stack, so complain. */
|
/* They had a few time units to dump stack, so complain. */
|
||||||
print_other_cpu_stall(rsp);
|
print_other_cpu_stall(rsp);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -587,26 +635,6 @@ static void __init check_cpu_stall_init(void)
|
|||||||
atomic_notifier_chain_register(&panic_notifier_list, &rcu_panic_block);
|
atomic_notifier_chain_register(&panic_notifier_list, &rcu_panic_block);
|
||||||
}
|
}
|
||||||
|
|
||||||
#else /* #ifdef CONFIG_RCU_CPU_STALL_DETECTOR */
|
|
||||||
|
|
||||||
static void record_gp_stall_check_time(struct rcu_state *rsp)
|
|
||||||
{
|
|
||||||
}
|
|
||||||
|
|
||||||
static void check_cpu_stall(struct rcu_state *rsp, struct rcu_data *rdp)
|
|
||||||
{
|
|
||||||
}
|
|
||||||
|
|
||||||
void rcu_cpu_stall_reset(void)
|
|
||||||
{
|
|
||||||
}
|
|
||||||
|
|
||||||
static void __init check_cpu_stall_init(void)
|
|
||||||
{
|
|
||||||
}
|
|
||||||
|
|
||||||
#endif /* #else #ifdef CONFIG_RCU_CPU_STALL_DETECTOR */
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Update CPU-local rcu_data state to record the newly noticed grace period.
|
* Update CPU-local rcu_data state to record the newly noticed grace period.
|
||||||
* This is used both when we started the grace period and when we notice
|
* This is used both when we started the grace period and when we notice
|
||||||
@@ -809,6 +837,7 @@ rcu_start_gp(struct rcu_state *rsp, unsigned long flags)
|
|||||||
rnp->completed = rsp->completed;
|
rnp->completed = rsp->completed;
|
||||||
rsp->signaled = RCU_SIGNAL_INIT; /* force_quiescent_state OK. */
|
rsp->signaled = RCU_SIGNAL_INIT; /* force_quiescent_state OK. */
|
||||||
rcu_start_gp_per_cpu(rsp, rnp, rdp);
|
rcu_start_gp_per_cpu(rsp, rnp, rdp);
|
||||||
|
rcu_preempt_boost_start_gp(rnp);
|
||||||
raw_spin_unlock_irqrestore(&rnp->lock, flags);
|
raw_spin_unlock_irqrestore(&rnp->lock, flags);
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
@@ -844,6 +873,7 @@ rcu_start_gp(struct rcu_state *rsp, unsigned long flags)
|
|||||||
rnp->completed = rsp->completed;
|
rnp->completed = rsp->completed;
|
||||||
if (rnp == rdp->mynode)
|
if (rnp == rdp->mynode)
|
||||||
rcu_start_gp_per_cpu(rsp, rnp, rdp);
|
rcu_start_gp_per_cpu(rsp, rnp, rdp);
|
||||||
|
rcu_preempt_boost_start_gp(rnp);
|
||||||
raw_spin_unlock(&rnp->lock); /* irqs remain disabled. */
|
raw_spin_unlock(&rnp->lock); /* irqs remain disabled. */
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -864,7 +894,18 @@ rcu_start_gp(struct rcu_state *rsp, unsigned long flags)
|
|||||||
static void rcu_report_qs_rsp(struct rcu_state *rsp, unsigned long flags)
|
static void rcu_report_qs_rsp(struct rcu_state *rsp, unsigned long flags)
|
||||||
__releases(rcu_get_root(rsp)->lock)
|
__releases(rcu_get_root(rsp)->lock)
|
||||||
{
|
{
|
||||||
|
unsigned long gp_duration;
|
||||||
|
|
||||||
WARN_ON_ONCE(!rcu_gp_in_progress(rsp));
|
WARN_ON_ONCE(!rcu_gp_in_progress(rsp));
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Ensure that all grace-period and pre-grace-period activity
|
||||||
|
* is seen before the assignment to rsp->completed.
|
||||||
|
*/
|
||||||
|
smp_mb(); /* See above block comment. */
|
||||||
|
gp_duration = jiffies - rsp->gp_start;
|
||||||
|
if (gp_duration > rsp->gp_max)
|
||||||
|
rsp->gp_max = gp_duration;
|
||||||
rsp->completed = rsp->gpnum;
|
rsp->completed = rsp->gpnum;
|
||||||
rsp->signaled = RCU_GP_IDLE;
|
rsp->signaled = RCU_GP_IDLE;
|
||||||
rcu_start_gp(rsp, flags); /* releases root node's rnp->lock. */
|
rcu_start_gp(rsp, flags); /* releases root node's rnp->lock. */
|
||||||
@@ -894,7 +935,7 @@ rcu_report_qs_rnp(unsigned long mask, struct rcu_state *rsp,
|
|||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
rnp->qsmask &= ~mask;
|
rnp->qsmask &= ~mask;
|
||||||
if (rnp->qsmask != 0 || rcu_preempted_readers(rnp)) {
|
if (rnp->qsmask != 0 || rcu_preempt_blocked_readers_cgp(rnp)) {
|
||||||
|
|
||||||
/* Other bits still set at this level, so done. */
|
/* Other bits still set at this level, so done. */
|
||||||
raw_spin_unlock_irqrestore(&rnp->lock, flags);
|
raw_spin_unlock_irqrestore(&rnp->lock, flags);
|
||||||
@@ -1037,6 +1078,8 @@ static void rcu_send_cbs_to_online(struct rcu_state *rsp)
|
|||||||
/*
|
/*
|
||||||
* Remove the outgoing CPU from the bitmasks in the rcu_node hierarchy
|
* Remove the outgoing CPU from the bitmasks in the rcu_node hierarchy
|
||||||
* and move all callbacks from the outgoing CPU to the current one.
|
* and move all callbacks from the outgoing CPU to the current one.
|
||||||
|
* There can only be one CPU hotplug operation at a time, so no other
|
||||||
|
* CPU can be attempting to update rcu_cpu_kthread_task.
|
||||||
*/
|
*/
|
||||||
static void __rcu_offline_cpu(int cpu, struct rcu_state *rsp)
|
static void __rcu_offline_cpu(int cpu, struct rcu_state *rsp)
|
||||||
{
|
{
|
||||||
@@ -1045,6 +1088,14 @@ static void __rcu_offline_cpu(int cpu, struct rcu_state *rsp)
|
|||||||
int need_report = 0;
|
int need_report = 0;
|
||||||
struct rcu_data *rdp = per_cpu_ptr(rsp->rda, cpu);
|
struct rcu_data *rdp = per_cpu_ptr(rsp->rda, cpu);
|
||||||
struct rcu_node *rnp;
|
struct rcu_node *rnp;
|
||||||
|
struct task_struct *t;
|
||||||
|
|
||||||
|
/* Stop the CPU's kthread. */
|
||||||
|
t = per_cpu(rcu_cpu_kthread_task, cpu);
|
||||||
|
if (t != NULL) {
|
||||||
|
per_cpu(rcu_cpu_kthread_task, cpu) = NULL;
|
||||||
|
kthread_stop(t);
|
||||||
|
}
|
||||||
|
|
||||||
/* Exclude any attempts to start a new grace period. */
|
/* Exclude any attempts to start a new grace period. */
|
||||||
raw_spin_lock_irqsave(&rsp->onofflock, flags);
|
raw_spin_lock_irqsave(&rsp->onofflock, flags);
|
||||||
@@ -1082,6 +1133,22 @@ static void __rcu_offline_cpu(int cpu, struct rcu_state *rsp)
|
|||||||
raw_spin_unlock_irqrestore(&rnp->lock, flags);
|
raw_spin_unlock_irqrestore(&rnp->lock, flags);
|
||||||
if (need_report & RCU_OFL_TASKS_EXP_GP)
|
if (need_report & RCU_OFL_TASKS_EXP_GP)
|
||||||
rcu_report_exp_rnp(rsp, rnp);
|
rcu_report_exp_rnp(rsp, rnp);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* If there are no more online CPUs for this rcu_node structure,
|
||||||
|
* kill the rcu_node structure's kthread. Otherwise, adjust its
|
||||||
|
* affinity.
|
||||||
|
*/
|
||||||
|
t = rnp->node_kthread_task;
|
||||||
|
if (t != NULL &&
|
||||||
|
rnp->qsmaskinit == 0) {
|
||||||
|
raw_spin_lock_irqsave(&rnp->lock, flags);
|
||||||
|
rnp->node_kthread_task = NULL;
|
||||||
|
raw_spin_unlock_irqrestore(&rnp->lock, flags);
|
||||||
|
kthread_stop(t);
|
||||||
|
rcu_stop_boost_kthread(rnp);
|
||||||
|
} else
|
||||||
|
rcu_node_kthread_setaffinity(rnp, -1);
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@@ -1143,7 +1210,7 @@ static void rcu_do_batch(struct rcu_state *rsp, struct rcu_data *rdp)
|
|||||||
next = list->next;
|
next = list->next;
|
||||||
prefetch(next);
|
prefetch(next);
|
||||||
debug_rcu_head_unqueue(list);
|
debug_rcu_head_unqueue(list);
|
||||||
list->func(list);
|
__rcu_reclaim(list);
|
||||||
list = next;
|
list = next;
|
||||||
if (++count >= rdp->blimit)
|
if (++count >= rdp->blimit)
|
||||||
break;
|
break;
|
||||||
@@ -1179,7 +1246,7 @@ static void rcu_do_batch(struct rcu_state *rsp, struct rcu_data *rdp)
|
|||||||
|
|
||||||
/* Re-raise the RCU softirq if there are callbacks remaining. */
|
/* Re-raise the RCU softirq if there are callbacks remaining. */
|
||||||
if (cpu_has_callbacks_ready_to_invoke(rdp))
|
if (cpu_has_callbacks_ready_to_invoke(rdp))
|
||||||
raise_softirq(RCU_SOFTIRQ);
|
invoke_rcu_cpu_kthread();
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@@ -1225,7 +1292,7 @@ void rcu_check_callbacks(int cpu, int user)
|
|||||||
}
|
}
|
||||||
rcu_preempt_check_callbacks(cpu);
|
rcu_preempt_check_callbacks(cpu);
|
||||||
if (rcu_pending(cpu))
|
if (rcu_pending(cpu))
|
||||||
raise_softirq(RCU_SOFTIRQ);
|
invoke_rcu_cpu_kthread();
|
||||||
}
|
}
|
||||||
|
|
||||||
#ifdef CONFIG_SMP
|
#ifdef CONFIG_SMP
|
||||||
@@ -1233,6 +1300,8 @@ void rcu_check_callbacks(int cpu, int user)
|
|||||||
/*
|
/*
|
||||||
* Scan the leaf rcu_node structures, processing dyntick state for any that
|
* Scan the leaf rcu_node structures, processing dyntick state for any that
|
||||||
* have not yet encountered a quiescent state, using the function specified.
|
* have not yet encountered a quiescent state, using the function specified.
|
||||||
|
* Also initiate boosting for any threads blocked on the root rcu_node.
|
||||||
|
*
|
||||||
* The caller must have suppressed start of new grace periods.
|
* The caller must have suppressed start of new grace periods.
|
||||||
*/
|
*/
|
||||||
static void force_qs_rnp(struct rcu_state *rsp, int (*f)(struct rcu_data *))
|
static void force_qs_rnp(struct rcu_state *rsp, int (*f)(struct rcu_data *))
|
||||||
@@ -1251,6 +1320,7 @@ static void force_qs_rnp(struct rcu_state *rsp, int (*f)(struct rcu_data *))
|
|||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
if (rnp->qsmask == 0) {
|
if (rnp->qsmask == 0) {
|
||||||
|
rcu_initiate_boost(rnp);
|
||||||
raw_spin_unlock_irqrestore(&rnp->lock, flags);
|
raw_spin_unlock_irqrestore(&rnp->lock, flags);
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
@@ -1269,6 +1339,11 @@ static void force_qs_rnp(struct rcu_state *rsp, int (*f)(struct rcu_data *))
|
|||||||
}
|
}
|
||||||
raw_spin_unlock_irqrestore(&rnp->lock, flags);
|
raw_spin_unlock_irqrestore(&rnp->lock, flags);
|
||||||
}
|
}
|
||||||
|
rnp = rcu_get_root(rsp);
|
||||||
|
raw_spin_lock_irqsave(&rnp->lock, flags);
|
||||||
|
if (rnp->qsmask == 0)
|
||||||
|
rcu_initiate_boost(rnp);
|
||||||
|
raw_spin_unlock_irqrestore(&rnp->lock, flags);
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@@ -1389,31 +1464,360 @@ __rcu_process_callbacks(struct rcu_state *rsp, struct rcu_data *rdp)
|
|||||||
/*
|
/*
|
||||||
* Do softirq processing for the current CPU.
|
* Do softirq processing for the current CPU.
|
||||||
*/
|
*/
|
||||||
static void rcu_process_callbacks(struct softirq_action *unused)
|
static void rcu_process_callbacks(void)
|
||||||
{
|
{
|
||||||
/*
|
|
||||||
* Memory references from any prior RCU read-side critical sections
|
|
||||||
* executed by the interrupted code must be seen before any RCU
|
|
||||||
* grace-period manipulations below.
|
|
||||||
*/
|
|
||||||
smp_mb(); /* See above block comment. */
|
|
||||||
|
|
||||||
__rcu_process_callbacks(&rcu_sched_state,
|
__rcu_process_callbacks(&rcu_sched_state,
|
||||||
&__get_cpu_var(rcu_sched_data));
|
&__get_cpu_var(rcu_sched_data));
|
||||||
__rcu_process_callbacks(&rcu_bh_state, &__get_cpu_var(rcu_bh_data));
|
__rcu_process_callbacks(&rcu_bh_state, &__get_cpu_var(rcu_bh_data));
|
||||||
rcu_preempt_process_callbacks();
|
rcu_preempt_process_callbacks();
|
||||||
|
|
||||||
/*
|
|
||||||
* Memory references from any later RCU read-side critical sections
|
|
||||||
* executed by the interrupted code must be seen after any RCU
|
|
||||||
* grace-period manipulations above.
|
|
||||||
*/
|
|
||||||
smp_mb(); /* See above block comment. */
|
|
||||||
|
|
||||||
/* If we are last CPU on way to dyntick-idle mode, accelerate it. */
|
/* If we are last CPU on way to dyntick-idle mode, accelerate it. */
|
||||||
rcu_needs_cpu_flush();
|
rcu_needs_cpu_flush();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Wake up the current CPU's kthread. This replaces raise_softirq()
|
||||||
|
* in earlier versions of RCU. Note that because we are running on
|
||||||
|
* the current CPU with interrupts disabled, the rcu_cpu_kthread_task
|
||||||
|
* cannot disappear out from under us.
|
||||||
|
*/
|
||||||
|
static void invoke_rcu_cpu_kthread(void)
|
||||||
|
{
|
||||||
|
unsigned long flags;
|
||||||
|
|
||||||
|
local_irq_save(flags);
|
||||||
|
__this_cpu_write(rcu_cpu_has_work, 1);
|
||||||
|
if (__this_cpu_read(rcu_cpu_kthread_task) == NULL) {
|
||||||
|
local_irq_restore(flags);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
wake_up(&__get_cpu_var(rcu_cpu_wq));
|
||||||
|
local_irq_restore(flags);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Wake up the specified per-rcu_node-structure kthread.
|
||||||
|
* The caller must hold ->lock.
|
||||||
|
*/
|
||||||
|
static void invoke_rcu_node_kthread(struct rcu_node *rnp)
|
||||||
|
{
|
||||||
|
struct task_struct *t;
|
||||||
|
|
||||||
|
t = rnp->node_kthread_task;
|
||||||
|
if (t != NULL)
|
||||||
|
wake_up_process(t);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Set the specified CPU's kthread to run RT or not, as specified by
|
||||||
|
* the to_rt argument. The CPU-hotplug locks are held, so the task
|
||||||
|
* is not going away.
|
||||||
|
*/
|
||||||
|
static void rcu_cpu_kthread_setrt(int cpu, int to_rt)
|
||||||
|
{
|
||||||
|
int policy;
|
||||||
|
struct sched_param sp;
|
||||||
|
struct task_struct *t;
|
||||||
|
|
||||||
|
t = per_cpu(rcu_cpu_kthread_task, cpu);
|
||||||
|
if (t == NULL)
|
||||||
|
return;
|
||||||
|
if (to_rt) {
|
||||||
|
policy = SCHED_FIFO;
|
||||||
|
sp.sched_priority = RCU_KTHREAD_PRIO;
|
||||||
|
} else {
|
||||||
|
policy = SCHED_NORMAL;
|
||||||
|
sp.sched_priority = 0;
|
||||||
|
}
|
||||||
|
sched_setscheduler_nocheck(t, policy, &sp);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Timer handler to initiate the waking up of per-CPU kthreads that
|
||||||
|
* have yielded the CPU due to excess numbers of RCU callbacks.
|
||||||
|
* We wake up the per-rcu_node kthread, which in turn will wake up
|
||||||
|
* the booster kthread.
|
||||||
|
*/
|
||||||
|
static void rcu_cpu_kthread_timer(unsigned long arg)
|
||||||
|
{
|
||||||
|
unsigned long flags;
|
||||||
|
struct rcu_data *rdp = per_cpu_ptr(rcu_state->rda, arg);
|
||||||
|
struct rcu_node *rnp = rdp->mynode;
|
||||||
|
|
||||||
|
raw_spin_lock_irqsave(&rnp->lock, flags);
|
||||||
|
rnp->wakemask |= rdp->grpmask;
|
||||||
|
invoke_rcu_node_kthread(rnp);
|
||||||
|
raw_spin_unlock_irqrestore(&rnp->lock, flags);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Drop to non-real-time priority and yield, but only after posting a
|
||||||
|
* timer that will cause us to regain our real-time priority if we
|
||||||
|
* remain preempted. Either way, we restore our real-time priority
|
||||||
|
* before returning.
|
||||||
|
*/
|
||||||
|
static void rcu_yield(void (*f)(unsigned long), unsigned long arg)
|
||||||
|
{
|
||||||
|
struct sched_param sp;
|
||||||
|
struct timer_list yield_timer;
|
||||||
|
|
||||||
|
setup_timer_on_stack(&yield_timer, f, arg);
|
||||||
|
mod_timer(&yield_timer, jiffies + 2);
|
||||||
|
sp.sched_priority = 0;
|
||||||
|
sched_setscheduler_nocheck(current, SCHED_NORMAL, &sp);
|
||||||
|
set_user_nice(current, 19);
|
||||||
|
schedule();
|
||||||
|
sp.sched_priority = RCU_KTHREAD_PRIO;
|
||||||
|
sched_setscheduler_nocheck(current, SCHED_FIFO, &sp);
|
||||||
|
del_timer(&yield_timer);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Handle cases where the rcu_cpu_kthread() ends up on the wrong CPU.
|
||||||
|
* This can happen while the corresponding CPU is either coming online
|
||||||
|
* or going offline. We cannot wait until the CPU is fully online
|
||||||
|
* before starting the kthread, because the various notifier functions
|
||||||
|
* can wait for RCU grace periods. So we park rcu_cpu_kthread() until
|
||||||
|
* the corresponding CPU is online.
|
||||||
|
*
|
||||||
|
* Return 1 if the kthread needs to stop, 0 otherwise.
|
||||||
|
*
|
||||||
|
* Caller must disable bh. This function can momentarily enable it.
|
||||||
|
*/
|
||||||
|
static int rcu_cpu_kthread_should_stop(int cpu)
|
||||||
|
{
|
||||||
|
while (cpu_is_offline(cpu) ||
|
||||||
|
!cpumask_equal(¤t->cpus_allowed, cpumask_of(cpu)) ||
|
||||||
|
smp_processor_id() != cpu) {
|
||||||
|
if (kthread_should_stop())
|
||||||
|
return 1;
|
||||||
|
per_cpu(rcu_cpu_kthread_status, cpu) = RCU_KTHREAD_OFFCPU;
|
||||||
|
per_cpu(rcu_cpu_kthread_cpu, cpu) = raw_smp_processor_id();
|
||||||
|
local_bh_enable();
|
||||||
|
schedule_timeout_uninterruptible(1);
|
||||||
|
if (!cpumask_equal(¤t->cpus_allowed, cpumask_of(cpu)))
|
||||||
|
set_cpus_allowed_ptr(current, cpumask_of(cpu));
|
||||||
|
local_bh_disable();
|
||||||
|
}
|
||||||
|
per_cpu(rcu_cpu_kthread_cpu, cpu) = cpu;
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Per-CPU kernel thread that invokes RCU callbacks. This replaces the
|
||||||
|
* earlier RCU softirq.
|
||||||
|
*/
|
||||||
|
static int rcu_cpu_kthread(void *arg)
|
||||||
|
{
|
||||||
|
int cpu = (int)(long)arg;
|
||||||
|
unsigned long flags;
|
||||||
|
int spincnt = 0;
|
||||||
|
unsigned int *statusp = &per_cpu(rcu_cpu_kthread_status, cpu);
|
||||||
|
wait_queue_head_t *wqp = &per_cpu(rcu_cpu_wq, cpu);
|
||||||
|
char work;
|
||||||
|
char *workp = &per_cpu(rcu_cpu_has_work, cpu);
|
||||||
|
|
||||||
|
for (;;) {
|
||||||
|
*statusp = RCU_KTHREAD_WAITING;
|
||||||
|
wait_event_interruptible(*wqp,
|
||||||
|
*workp != 0 || kthread_should_stop());
|
||||||
|
local_bh_disable();
|
||||||
|
if (rcu_cpu_kthread_should_stop(cpu)) {
|
||||||
|
local_bh_enable();
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
*statusp = RCU_KTHREAD_RUNNING;
|
||||||
|
per_cpu(rcu_cpu_kthread_loops, cpu)++;
|
||||||
|
local_irq_save(flags);
|
||||||
|
work = *workp;
|
||||||
|
*workp = 0;
|
||||||
|
local_irq_restore(flags);
|
||||||
|
if (work)
|
||||||
|
rcu_process_callbacks();
|
||||||
|
local_bh_enable();
|
||||||
|
if (*workp != 0)
|
||||||
|
spincnt++;
|
||||||
|
else
|
||||||
|
spincnt = 0;
|
||||||
|
if (spincnt > 10) {
|
||||||
|
*statusp = RCU_KTHREAD_YIELDING;
|
||||||
|
rcu_yield(rcu_cpu_kthread_timer, (unsigned long)cpu);
|
||||||
|
spincnt = 0;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
*statusp = RCU_KTHREAD_STOPPED;
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Spawn a per-CPU kthread, setting up affinity and priority.
|
||||||
|
* Because the CPU hotplug lock is held, no other CPU will be attempting
|
||||||
|
* to manipulate rcu_cpu_kthread_task. There might be another CPU
|
||||||
|
* attempting to access it during boot, but the locking in kthread_bind()
|
||||||
|
* will enforce sufficient ordering.
|
||||||
|
*/
|
||||||
|
static int __cpuinit rcu_spawn_one_cpu_kthread(int cpu)
|
||||||
|
{
|
||||||
|
struct sched_param sp;
|
||||||
|
struct task_struct *t;
|
||||||
|
|
||||||
|
if (!rcu_kthreads_spawnable ||
|
||||||
|
per_cpu(rcu_cpu_kthread_task, cpu) != NULL)
|
||||||
|
return 0;
|
||||||
|
t = kthread_create(rcu_cpu_kthread, (void *)(long)cpu, "rcuc%d", cpu);
|
||||||
|
if (IS_ERR(t))
|
||||||
|
return PTR_ERR(t);
|
||||||
|
kthread_bind(t, cpu);
|
||||||
|
per_cpu(rcu_cpu_kthread_cpu, cpu) = cpu;
|
||||||
|
WARN_ON_ONCE(per_cpu(rcu_cpu_kthread_task, cpu) != NULL);
|
||||||
|
per_cpu(rcu_cpu_kthread_task, cpu) = t;
|
||||||
|
wake_up_process(t);
|
||||||
|
sp.sched_priority = RCU_KTHREAD_PRIO;
|
||||||
|
sched_setscheduler_nocheck(t, SCHED_FIFO, &sp);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Per-rcu_node kthread, which is in charge of waking up the per-CPU
|
||||||
|
* kthreads when needed. We ignore requests to wake up kthreads
|
||||||
|
* for offline CPUs, which is OK because force_quiescent_state()
|
||||||
|
* takes care of this case.
|
||||||
|
*/
|
||||||
|
static int rcu_node_kthread(void *arg)
|
||||||
|
{
|
||||||
|
int cpu;
|
||||||
|
unsigned long flags;
|
||||||
|
unsigned long mask;
|
||||||
|
struct rcu_node *rnp = (struct rcu_node *)arg;
|
||||||
|
struct sched_param sp;
|
||||||
|
struct task_struct *t;
|
||||||
|
|
||||||
|
for (;;) {
|
||||||
|
rnp->node_kthread_status = RCU_KTHREAD_WAITING;
|
||||||
|
wait_event_interruptible(rnp->node_wq, rnp->wakemask != 0 ||
|
||||||
|
kthread_should_stop());
|
||||||
|
if (kthread_should_stop())
|
||||||
|
break;
|
||||||
|
rnp->node_kthread_status = RCU_KTHREAD_RUNNING;
|
||||||
|
raw_spin_lock_irqsave(&rnp->lock, flags);
|
||||||
|
mask = rnp->wakemask;
|
||||||
|
rnp->wakemask = 0;
|
||||||
|
rcu_initiate_boost(rnp);
|
||||||
|
raw_spin_unlock_irqrestore(&rnp->lock, flags);
|
||||||
|
for (cpu = rnp->grplo; cpu <= rnp->grphi; cpu++, mask >>= 1) {
|
||||||
|
if ((mask & 0x1) == 0)
|
||||||
|
continue;
|
||||||
|
preempt_disable();
|
||||||
|
t = per_cpu(rcu_cpu_kthread_task, cpu);
|
||||||
|
if (!cpu_online(cpu) || t == NULL) {
|
||||||
|
preempt_enable();
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
per_cpu(rcu_cpu_has_work, cpu) = 1;
|
||||||
|
sp.sched_priority = RCU_KTHREAD_PRIO;
|
||||||
|
sched_setscheduler_nocheck(t, SCHED_FIFO, &sp);
|
||||||
|
preempt_enable();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
rnp->node_kthread_status = RCU_KTHREAD_STOPPED;
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Set the per-rcu_node kthread's affinity to cover all CPUs that are
|
||||||
|
* served by the rcu_node in question. The CPU hotplug lock is still
|
||||||
|
* held, so the value of rnp->qsmaskinit will be stable.
|
||||||
|
*
|
||||||
|
* We don't include outgoingcpu in the affinity set, use -1 if there is
|
||||||
|
* no outgoing CPU. If there are no CPUs left in the affinity set,
|
||||||
|
* this function allows the kthread to execute on any CPU.
|
||||||
|
*/
|
||||||
|
static void rcu_node_kthread_setaffinity(struct rcu_node *rnp, int outgoingcpu)
|
||||||
|
{
|
||||||
|
cpumask_var_t cm;
|
||||||
|
int cpu;
|
||||||
|
unsigned long mask = rnp->qsmaskinit;
|
||||||
|
|
||||||
|
if (rnp->node_kthread_task == NULL || mask == 0)
|
||||||
|
return;
|
||||||
|
if (!alloc_cpumask_var(&cm, GFP_KERNEL))
|
||||||
|
return;
|
||||||
|
cpumask_clear(cm);
|
||||||
|
for (cpu = rnp->grplo; cpu <= rnp->grphi; cpu++, mask >>= 1)
|
||||||
|
if ((mask & 0x1) && cpu != outgoingcpu)
|
||||||
|
cpumask_set_cpu(cpu, cm);
|
||||||
|
if (cpumask_weight(cm) == 0) {
|
||||||
|
cpumask_setall(cm);
|
||||||
|
for (cpu = rnp->grplo; cpu <= rnp->grphi; cpu++)
|
||||||
|
cpumask_clear_cpu(cpu, cm);
|
||||||
|
WARN_ON_ONCE(cpumask_weight(cm) == 0);
|
||||||
|
}
|
||||||
|
set_cpus_allowed_ptr(rnp->node_kthread_task, cm);
|
||||||
|
rcu_boost_kthread_setaffinity(rnp, cm);
|
||||||
|
free_cpumask_var(cm);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Spawn a per-rcu_node kthread, setting priority and affinity.
|
||||||
|
* Called during boot before online/offline can happen, or, if
|
||||||
|
* during runtime, with the main CPU-hotplug locks held. So only
|
||||||
|
* one of these can be executing at a time.
|
||||||
|
*/
|
||||||
|
static int __cpuinit rcu_spawn_one_node_kthread(struct rcu_state *rsp,
|
||||||
|
struct rcu_node *rnp)
|
||||||
|
{
|
||||||
|
unsigned long flags;
|
||||||
|
int rnp_index = rnp - &rsp->node[0];
|
||||||
|
struct sched_param sp;
|
||||||
|
struct task_struct *t;
|
||||||
|
|
||||||
|
if (!rcu_kthreads_spawnable ||
|
||||||
|
rnp->qsmaskinit == 0)
|
||||||
|
return 0;
|
||||||
|
if (rnp->node_kthread_task == NULL) {
|
||||||
|
t = kthread_create(rcu_node_kthread, (void *)rnp,
|
||||||
|
"rcun%d", rnp_index);
|
||||||
|
if (IS_ERR(t))
|
||||||
|
return PTR_ERR(t);
|
||||||
|
raw_spin_lock_irqsave(&rnp->lock, flags);
|
||||||
|
rnp->node_kthread_task = t;
|
||||||
|
raw_spin_unlock_irqrestore(&rnp->lock, flags);
|
||||||
|
wake_up_process(t);
|
||||||
|
sp.sched_priority = 99;
|
||||||
|
sched_setscheduler_nocheck(t, SCHED_FIFO, &sp);
|
||||||
|
}
|
||||||
|
return rcu_spawn_one_boost_kthread(rsp, rnp, rnp_index);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Spawn all kthreads -- called as soon as the scheduler is running.
|
||||||
|
*/
|
||||||
|
static int __init rcu_spawn_kthreads(void)
|
||||||
|
{
|
||||||
|
int cpu;
|
||||||
|
struct rcu_node *rnp;
|
||||||
|
|
||||||
|
rcu_kthreads_spawnable = 1;
|
||||||
|
for_each_possible_cpu(cpu) {
|
||||||
|
init_waitqueue_head(&per_cpu(rcu_cpu_wq, cpu));
|
||||||
|
per_cpu(rcu_cpu_has_work, cpu) = 0;
|
||||||
|
if (cpu_online(cpu))
|
||||||
|
(void)rcu_spawn_one_cpu_kthread(cpu);
|
||||||
|
}
|
||||||
|
rnp = rcu_get_root(rcu_state);
|
||||||
|
init_waitqueue_head(&rnp->node_wq);
|
||||||
|
rcu_init_boost_waitqueue(rnp);
|
||||||
|
(void)rcu_spawn_one_node_kthread(rcu_state, rnp);
|
||||||
|
if (NUM_RCU_NODES > 1)
|
||||||
|
rcu_for_each_leaf_node(rcu_state, rnp) {
|
||||||
|
init_waitqueue_head(&rnp->node_wq);
|
||||||
|
rcu_init_boost_waitqueue(rnp);
|
||||||
|
(void)rcu_spawn_one_node_kthread(rcu_state, rnp);
|
||||||
|
}
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
early_initcall(rcu_spawn_kthreads);
|
||||||
|
|
||||||
static void
|
static void
|
||||||
__call_rcu(struct rcu_head *head, void (*func)(struct rcu_head *rcu),
|
__call_rcu(struct rcu_head *head, void (*func)(struct rcu_head *rcu),
|
||||||
struct rcu_state *rsp)
|
struct rcu_state *rsp)
|
||||||
@@ -1439,6 +1843,13 @@ __call_rcu(struct rcu_head *head, void (*func)(struct rcu_head *rcu),
|
|||||||
/* Add the callback to our list. */
|
/* Add the callback to our list. */
|
||||||
*rdp->nxttail[RCU_NEXT_TAIL] = head;
|
*rdp->nxttail[RCU_NEXT_TAIL] = head;
|
||||||
rdp->nxttail[RCU_NEXT_TAIL] = &head->next;
|
rdp->nxttail[RCU_NEXT_TAIL] = &head->next;
|
||||||
|
rdp->qlen++;
|
||||||
|
|
||||||
|
/* If interrupts were disabled, don't dive into RCU core. */
|
||||||
|
if (irqs_disabled_flags(flags)) {
|
||||||
|
local_irq_restore(flags);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Force the grace period if too many callbacks or too long waiting.
|
* Force the grace period if too many callbacks or too long waiting.
|
||||||
@@ -1447,7 +1858,7 @@ __call_rcu(struct rcu_head *head, void (*func)(struct rcu_head *rcu),
|
|||||||
* invoking force_quiescent_state() if the newly enqueued callback
|
* invoking force_quiescent_state() if the newly enqueued callback
|
||||||
* is the only one waiting for a grace period to complete.
|
* is the only one waiting for a grace period to complete.
|
||||||
*/
|
*/
|
||||||
if (unlikely(++rdp->qlen > rdp->qlen_last_fqs_check + qhimark)) {
|
if (unlikely(rdp->qlen > rdp->qlen_last_fqs_check + qhimark)) {
|
||||||
|
|
||||||
/* Are we ignoring a completed grace period? */
|
/* Are we ignoring a completed grace period? */
|
||||||
rcu_process_gp_end(rsp, rdp);
|
rcu_process_gp_end(rsp, rdp);
|
||||||
@@ -1583,7 +1994,7 @@ static int __rcu_pending(struct rcu_state *rsp, struct rcu_data *rdp)
|
|||||||
* or RCU-bh, force a local reschedule.
|
* or RCU-bh, force a local reschedule.
|
||||||
*/
|
*/
|
||||||
rdp->n_rp_qs_pending++;
|
rdp->n_rp_qs_pending++;
|
||||||
if (!rdp->preemptable &&
|
if (!rdp->preemptible &&
|
||||||
ULONG_CMP_LT(ACCESS_ONCE(rsp->jiffies_force_qs) - 1,
|
ULONG_CMP_LT(ACCESS_ONCE(rsp->jiffies_force_qs) - 1,
|
||||||
jiffies))
|
jiffies))
|
||||||
set_need_resched();
|
set_need_resched();
|
||||||
@@ -1760,7 +2171,7 @@ rcu_boot_init_percpu_data(int cpu, struct rcu_state *rsp)
|
|||||||
* that this CPU cannot possibly have any RCU callbacks in flight yet.
|
* that this CPU cannot possibly have any RCU callbacks in flight yet.
|
||||||
*/
|
*/
|
||||||
static void __cpuinit
|
static void __cpuinit
|
||||||
rcu_init_percpu_data(int cpu, struct rcu_state *rsp, int preemptable)
|
rcu_init_percpu_data(int cpu, struct rcu_state *rsp, int preemptible)
|
||||||
{
|
{
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
unsigned long mask;
|
unsigned long mask;
|
||||||
@@ -1772,7 +2183,7 @@ rcu_init_percpu_data(int cpu, struct rcu_state *rsp, int preemptable)
|
|||||||
rdp->passed_quiesc = 0; /* We could be racing with new GP, */
|
rdp->passed_quiesc = 0; /* We could be racing with new GP, */
|
||||||
rdp->qs_pending = 1; /* so set up to respond to current GP. */
|
rdp->qs_pending = 1; /* so set up to respond to current GP. */
|
||||||
rdp->beenonline = 1; /* We have now been online. */
|
rdp->beenonline = 1; /* We have now been online. */
|
||||||
rdp->preemptable = preemptable;
|
rdp->preemptible = preemptible;
|
||||||
rdp->qlen_last_fqs_check = 0;
|
rdp->qlen_last_fqs_check = 0;
|
||||||
rdp->n_force_qs_snap = rsp->n_force_qs;
|
rdp->n_force_qs_snap = rsp->n_force_qs;
|
||||||
rdp->blimit = blimit;
|
rdp->blimit = blimit;
|
||||||
@@ -1813,6 +2224,19 @@ static void __cpuinit rcu_online_cpu(int cpu)
|
|||||||
rcu_preempt_init_percpu_data(cpu);
|
rcu_preempt_init_percpu_data(cpu);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void __cpuinit rcu_online_kthreads(int cpu)
|
||||||
|
{
|
||||||
|
struct rcu_data *rdp = per_cpu_ptr(rcu_state->rda, cpu);
|
||||||
|
struct rcu_node *rnp = rdp->mynode;
|
||||||
|
|
||||||
|
/* Fire up the incoming CPU's kthread and leaf rcu_node kthread. */
|
||||||
|
if (rcu_kthreads_spawnable) {
|
||||||
|
(void)rcu_spawn_one_cpu_kthread(cpu);
|
||||||
|
if (rnp->node_kthread_task == NULL)
|
||||||
|
(void)rcu_spawn_one_node_kthread(rcu_state, rnp);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Handle CPU online/offline notification events.
|
* Handle CPU online/offline notification events.
|
||||||
*/
|
*/
|
||||||
@@ -1820,11 +2244,23 @@ static int __cpuinit rcu_cpu_notify(struct notifier_block *self,
|
|||||||
unsigned long action, void *hcpu)
|
unsigned long action, void *hcpu)
|
||||||
{
|
{
|
||||||
long cpu = (long)hcpu;
|
long cpu = (long)hcpu;
|
||||||
|
struct rcu_data *rdp = per_cpu_ptr(rcu_state->rda, cpu);
|
||||||
|
struct rcu_node *rnp = rdp->mynode;
|
||||||
|
|
||||||
switch (action) {
|
switch (action) {
|
||||||
case CPU_UP_PREPARE:
|
case CPU_UP_PREPARE:
|
||||||
case CPU_UP_PREPARE_FROZEN:
|
case CPU_UP_PREPARE_FROZEN:
|
||||||
rcu_online_cpu(cpu);
|
rcu_online_cpu(cpu);
|
||||||
|
rcu_online_kthreads(cpu);
|
||||||
|
break;
|
||||||
|
case CPU_ONLINE:
|
||||||
|
case CPU_DOWN_FAILED:
|
||||||
|
rcu_node_kthread_setaffinity(rnp, -1);
|
||||||
|
rcu_cpu_kthread_setrt(cpu, 1);
|
||||||
|
break;
|
||||||
|
case CPU_DOWN_PREPARE:
|
||||||
|
rcu_node_kthread_setaffinity(rnp, cpu);
|
||||||
|
rcu_cpu_kthread_setrt(cpu, 0);
|
||||||
break;
|
break;
|
||||||
case CPU_DYING:
|
case CPU_DYING:
|
||||||
case CPU_DYING_FROZEN:
|
case CPU_DYING_FROZEN:
|
||||||
@@ -1943,10 +2379,7 @@ static void __init rcu_init_one(struct rcu_state *rsp,
|
|||||||
j / rsp->levelspread[i - 1];
|
j / rsp->levelspread[i - 1];
|
||||||
}
|
}
|
||||||
rnp->level = i;
|
rnp->level = i;
|
||||||
INIT_LIST_HEAD(&rnp->blocked_tasks[0]);
|
INIT_LIST_HEAD(&rnp->blkd_tasks);
|
||||||
INIT_LIST_HEAD(&rnp->blocked_tasks[1]);
|
|
||||||
INIT_LIST_HEAD(&rnp->blocked_tasks[2]);
|
|
||||||
INIT_LIST_HEAD(&rnp->blocked_tasks[3]);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -1968,7 +2401,6 @@ void __init rcu_init(void)
|
|||||||
rcu_init_one(&rcu_sched_state, &rcu_sched_data);
|
rcu_init_one(&rcu_sched_state, &rcu_sched_data);
|
||||||
rcu_init_one(&rcu_bh_state, &rcu_bh_data);
|
rcu_init_one(&rcu_bh_state, &rcu_bh_data);
|
||||||
__rcu_init_preempt();
|
__rcu_init_preempt();
|
||||||
open_softirq(RCU_SOFTIRQ, rcu_process_callbacks);
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* We don't need protection against CPU-hotplug here because
|
* We don't need protection against CPU-hotplug here because
|
||||||
|
116
kernel/rcutree.h
116
kernel/rcutree.h
@@ -84,13 +84,19 @@
|
|||||||
* Dynticks per-CPU state.
|
* Dynticks per-CPU state.
|
||||||
*/
|
*/
|
||||||
struct rcu_dynticks {
|
struct rcu_dynticks {
|
||||||
int dynticks_nesting; /* Track nesting level, sort of. */
|
int dynticks_nesting; /* Track irq/process nesting level. */
|
||||||
int dynticks; /* Even value for dynticks-idle, else odd. */
|
int dynticks_nmi_nesting; /* Track NMI nesting level. */
|
||||||
int dynticks_nmi; /* Even value for either dynticks-idle or */
|
atomic_t dynticks; /* Even value for dynticks-idle, else odd. */
|
||||||
/* not in nmi handler, else odd. So this */
|
|
||||||
/* remains even for nmi from irq handler. */
|
|
||||||
};
|
};
|
||||||
|
|
||||||
|
/* RCU's kthread states for tracing. */
|
||||||
|
#define RCU_KTHREAD_STOPPED 0
|
||||||
|
#define RCU_KTHREAD_RUNNING 1
|
||||||
|
#define RCU_KTHREAD_WAITING 2
|
||||||
|
#define RCU_KTHREAD_OFFCPU 3
|
||||||
|
#define RCU_KTHREAD_YIELDING 4
|
||||||
|
#define RCU_KTHREAD_MAX 4
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Definition for node within the RCU grace-period-detection hierarchy.
|
* Definition for node within the RCU grace-period-detection hierarchy.
|
||||||
*/
|
*/
|
||||||
@@ -109,10 +115,11 @@ struct rcu_node {
|
|||||||
/* an rcu_data structure, otherwise, each */
|
/* an rcu_data structure, otherwise, each */
|
||||||
/* bit corresponds to a child rcu_node */
|
/* bit corresponds to a child rcu_node */
|
||||||
/* structure. */
|
/* structure. */
|
||||||
unsigned long expmask; /* Groups that have ->blocked_tasks[] */
|
unsigned long expmask; /* Groups that have ->blkd_tasks */
|
||||||
/* elements that need to drain to allow the */
|
/* elements that need to drain to allow the */
|
||||||
/* current expedited grace period to */
|
/* current expedited grace period to */
|
||||||
/* complete (only for TREE_PREEMPT_RCU). */
|
/* complete (only for TREE_PREEMPT_RCU). */
|
||||||
|
unsigned long wakemask; /* CPUs whose kthread needs to be awakened. */
|
||||||
unsigned long qsmaskinit;
|
unsigned long qsmaskinit;
|
||||||
/* Per-GP initial value for qsmask & expmask. */
|
/* Per-GP initial value for qsmask & expmask. */
|
||||||
unsigned long grpmask; /* Mask to apply to parent qsmask. */
|
unsigned long grpmask; /* Mask to apply to parent qsmask. */
|
||||||
@@ -122,11 +129,68 @@ struct rcu_node {
|
|||||||
u8 grpnum; /* CPU/group number for next level up. */
|
u8 grpnum; /* CPU/group number for next level up. */
|
||||||
u8 level; /* root is at level 0. */
|
u8 level; /* root is at level 0. */
|
||||||
struct rcu_node *parent;
|
struct rcu_node *parent;
|
||||||
struct list_head blocked_tasks[4];
|
struct list_head blkd_tasks;
|
||||||
/* Tasks blocked in RCU read-side critsect. */
|
/* Tasks blocked in RCU read-side critical */
|
||||||
/* Grace period number (->gpnum) x blocked */
|
/* section. Tasks are placed at the head */
|
||||||
/* by tasks on the (x & 0x1) element of the */
|
/* of this list and age towards the tail. */
|
||||||
/* blocked_tasks[] array. */
|
struct list_head *gp_tasks;
|
||||||
|
/* Pointer to the first task blocking the */
|
||||||
|
/* current grace period, or NULL if there */
|
||||||
|
/* is no such task. */
|
||||||
|
struct list_head *exp_tasks;
|
||||||
|
/* Pointer to the first task blocking the */
|
||||||
|
/* current expedited grace period, or NULL */
|
||||||
|
/* if there is no such task. If there */
|
||||||
|
/* is no current expedited grace period, */
|
||||||
|
/* then there can cannot be any such task. */
|
||||||
|
#ifdef CONFIG_RCU_BOOST
|
||||||
|
struct list_head *boost_tasks;
|
||||||
|
/* Pointer to first task that needs to be */
|
||||||
|
/* priority boosted, or NULL if no priority */
|
||||||
|
/* boosting is needed for this rcu_node */
|
||||||
|
/* structure. If there are no tasks */
|
||||||
|
/* queued on this rcu_node structure that */
|
||||||
|
/* are blocking the current grace period, */
|
||||||
|
/* there can be no such task. */
|
||||||
|
unsigned long boost_time;
|
||||||
|
/* When to start boosting (jiffies). */
|
||||||
|
struct task_struct *boost_kthread_task;
|
||||||
|
/* kthread that takes care of priority */
|
||||||
|
/* boosting for this rcu_node structure. */
|
||||||
|
wait_queue_head_t boost_wq;
|
||||||
|
/* Wait queue on which to park the boost */
|
||||||
|
/* kthread. */
|
||||||
|
unsigned int boost_kthread_status;
|
||||||
|
/* State of boost_kthread_task for tracing. */
|
||||||
|
unsigned long n_tasks_boosted;
|
||||||
|
/* Total number of tasks boosted. */
|
||||||
|
unsigned long n_exp_boosts;
|
||||||
|
/* Number of tasks boosted for expedited GP. */
|
||||||
|
unsigned long n_normal_boosts;
|
||||||
|
/* Number of tasks boosted for normal GP. */
|
||||||
|
unsigned long n_balk_blkd_tasks;
|
||||||
|
/* Refused to boost: no blocked tasks. */
|
||||||
|
unsigned long n_balk_exp_gp_tasks;
|
||||||
|
/* Refused to boost: nothing blocking GP. */
|
||||||
|
unsigned long n_balk_boost_tasks;
|
||||||
|
/* Refused to boost: already boosting. */
|
||||||
|
unsigned long n_balk_notblocked;
|
||||||
|
/* Refused to boost: RCU RS CS still running. */
|
||||||
|
unsigned long n_balk_notyet;
|
||||||
|
/* Refused to boost: not yet time. */
|
||||||
|
unsigned long n_balk_nos;
|
||||||
|
/* Refused to boost: not sure why, though. */
|
||||||
|
/* This can happen due to race conditions. */
|
||||||
|
#endif /* #ifdef CONFIG_RCU_BOOST */
|
||||||
|
struct task_struct *node_kthread_task;
|
||||||
|
/* kthread that takes care of this rcu_node */
|
||||||
|
/* structure, for example, awakening the */
|
||||||
|
/* per-CPU kthreads as needed. */
|
||||||
|
wait_queue_head_t node_wq;
|
||||||
|
/* Wait queue on which to park the per-node */
|
||||||
|
/* kthread. */
|
||||||
|
unsigned int node_kthread_status;
|
||||||
|
/* State of node_kthread_task for tracing. */
|
||||||
} ____cacheline_internodealigned_in_smp;
|
} ____cacheline_internodealigned_in_smp;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@@ -175,7 +239,7 @@ struct rcu_data {
|
|||||||
bool passed_quiesc; /* User-mode/idle loop etc. */
|
bool passed_quiesc; /* User-mode/idle loop etc. */
|
||||||
bool qs_pending; /* Core waits for quiesc state. */
|
bool qs_pending; /* Core waits for quiesc state. */
|
||||||
bool beenonline; /* CPU online at least once. */
|
bool beenonline; /* CPU online at least once. */
|
||||||
bool preemptable; /* Preemptable RCU? */
|
bool preemptible; /* Preemptible RCU? */
|
||||||
struct rcu_node *mynode; /* This CPU's leaf of hierarchy */
|
struct rcu_node *mynode; /* This CPU's leaf of hierarchy */
|
||||||
unsigned long grpmask; /* Mask to apply to leaf qsmask. */
|
unsigned long grpmask; /* Mask to apply to leaf qsmask. */
|
||||||
|
|
||||||
@@ -218,7 +282,6 @@ struct rcu_data {
|
|||||||
/* 3) dynticks interface. */
|
/* 3) dynticks interface. */
|
||||||
struct rcu_dynticks *dynticks; /* Shared per-CPU dynticks state. */
|
struct rcu_dynticks *dynticks; /* Shared per-CPU dynticks state. */
|
||||||
int dynticks_snap; /* Per-GP tracking for dynticks. */
|
int dynticks_snap; /* Per-GP tracking for dynticks. */
|
||||||
int dynticks_nmi_snap; /* Per-GP tracking for dynticks_nmi. */
|
|
||||||
#endif /* #ifdef CONFIG_NO_HZ */
|
#endif /* #ifdef CONFIG_NO_HZ */
|
||||||
|
|
||||||
/* 4) reasons this CPU needed to be kicked by force_quiescent_state */
|
/* 4) reasons this CPU needed to be kicked by force_quiescent_state */
|
||||||
@@ -254,7 +317,6 @@ struct rcu_data {
|
|||||||
#endif /* #else #ifdef CONFIG_NO_HZ */
|
#endif /* #else #ifdef CONFIG_NO_HZ */
|
||||||
|
|
||||||
#define RCU_JIFFIES_TILL_FORCE_QS 3 /* for rsp->jiffies_force_qs */
|
#define RCU_JIFFIES_TILL_FORCE_QS 3 /* for rsp->jiffies_force_qs */
|
||||||
#ifdef CONFIG_RCU_CPU_STALL_DETECTOR
|
|
||||||
|
|
||||||
#ifdef CONFIG_PROVE_RCU
|
#ifdef CONFIG_PROVE_RCU
|
||||||
#define RCU_STALL_DELAY_DELTA (5 * HZ)
|
#define RCU_STALL_DELAY_DELTA (5 * HZ)
|
||||||
@@ -272,13 +334,6 @@ struct rcu_data {
|
|||||||
/* scheduling clock irq */
|
/* scheduling clock irq */
|
||||||
/* before ratting on them. */
|
/* before ratting on them. */
|
||||||
|
|
||||||
#ifdef CONFIG_RCU_CPU_STALL_DETECTOR_RUNNABLE
|
|
||||||
#define RCU_CPU_STALL_SUPPRESS_INIT 0
|
|
||||||
#else
|
|
||||||
#define RCU_CPU_STALL_SUPPRESS_INIT 1
|
|
||||||
#endif
|
|
||||||
|
|
||||||
#endif /* #ifdef CONFIG_RCU_CPU_STALL_DETECTOR */
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* RCU global state, including node hierarchy. This hierarchy is
|
* RCU global state, including node hierarchy. This hierarchy is
|
||||||
@@ -325,12 +380,12 @@ struct rcu_state {
|
|||||||
/* due to lock unavailable. */
|
/* due to lock unavailable. */
|
||||||
unsigned long n_force_qs_ngp; /* Number of calls leaving */
|
unsigned long n_force_qs_ngp; /* Number of calls leaving */
|
||||||
/* due to no GP active. */
|
/* due to no GP active. */
|
||||||
#ifdef CONFIG_RCU_CPU_STALL_DETECTOR
|
|
||||||
unsigned long gp_start; /* Time at which GP started, */
|
unsigned long gp_start; /* Time at which GP started, */
|
||||||
/* but in jiffies. */
|
/* but in jiffies. */
|
||||||
unsigned long jiffies_stall; /* Time at which to check */
|
unsigned long jiffies_stall; /* Time at which to check */
|
||||||
/* for CPU stalls. */
|
/* for CPU stalls. */
|
||||||
#endif /* #ifdef CONFIG_RCU_CPU_STALL_DETECTOR */
|
unsigned long gp_max; /* Maximum GP duration in */
|
||||||
|
/* jiffies. */
|
||||||
char *name; /* Name of structure. */
|
char *name; /* Name of structure. */
|
||||||
};
|
};
|
||||||
|
|
||||||
@@ -361,16 +416,14 @@ DECLARE_PER_CPU(struct rcu_data, rcu_preempt_data);
|
|||||||
static void rcu_bootup_announce(void);
|
static void rcu_bootup_announce(void);
|
||||||
long rcu_batches_completed(void);
|
long rcu_batches_completed(void);
|
||||||
static void rcu_preempt_note_context_switch(int cpu);
|
static void rcu_preempt_note_context_switch(int cpu);
|
||||||
static int rcu_preempted_readers(struct rcu_node *rnp);
|
static int rcu_preempt_blocked_readers_cgp(struct rcu_node *rnp);
|
||||||
#ifdef CONFIG_HOTPLUG_CPU
|
#ifdef CONFIG_HOTPLUG_CPU
|
||||||
static void rcu_report_unblock_qs_rnp(struct rcu_node *rnp,
|
static void rcu_report_unblock_qs_rnp(struct rcu_node *rnp,
|
||||||
unsigned long flags);
|
unsigned long flags);
|
||||||
#endif /* #ifdef CONFIG_HOTPLUG_CPU */
|
#endif /* #ifdef CONFIG_HOTPLUG_CPU */
|
||||||
#ifdef CONFIG_RCU_CPU_STALL_DETECTOR
|
|
||||||
static void rcu_print_detail_task_stall(struct rcu_state *rsp);
|
static void rcu_print_detail_task_stall(struct rcu_state *rsp);
|
||||||
static void rcu_print_task_stall(struct rcu_node *rnp);
|
static void rcu_print_task_stall(struct rcu_node *rnp);
|
||||||
static void rcu_preempt_stall_reset(void);
|
static void rcu_preempt_stall_reset(void);
|
||||||
#endif /* #ifdef CONFIG_RCU_CPU_STALL_DETECTOR */
|
|
||||||
static void rcu_preempt_check_blocked_tasks(struct rcu_node *rnp);
|
static void rcu_preempt_check_blocked_tasks(struct rcu_node *rnp);
|
||||||
#ifdef CONFIG_HOTPLUG_CPU
|
#ifdef CONFIG_HOTPLUG_CPU
|
||||||
static int rcu_preempt_offline_tasks(struct rcu_state *rsp,
|
static int rcu_preempt_offline_tasks(struct rcu_state *rsp,
|
||||||
@@ -390,5 +443,16 @@ static void __cpuinit rcu_preempt_init_percpu_data(int cpu);
|
|||||||
static void rcu_preempt_send_cbs_to_online(void);
|
static void rcu_preempt_send_cbs_to_online(void);
|
||||||
static void __init __rcu_init_preempt(void);
|
static void __init __rcu_init_preempt(void);
|
||||||
static void rcu_needs_cpu_flush(void);
|
static void rcu_needs_cpu_flush(void);
|
||||||
|
static void __init rcu_init_boost_waitqueue(struct rcu_node *rnp);
|
||||||
|
static void rcu_initiate_boost(struct rcu_node *rnp);
|
||||||
|
static void rcu_boost_kthread_setaffinity(struct rcu_node *rnp,
|
||||||
|
cpumask_var_t cm);
|
||||||
|
static void rcu_preempt_boost_start_gp(struct rcu_node *rnp);
|
||||||
|
static int __cpuinit rcu_spawn_one_boost_kthread(struct rcu_state *rsp,
|
||||||
|
struct rcu_node *rnp,
|
||||||
|
int rnp_index);
|
||||||
|
#ifdef CONFIG_HOTPLUG_CPU
|
||||||
|
static void rcu_stop_boost_kthread(struct rcu_node *rnp);
|
||||||
|
#endif /* #ifdef CONFIG_HOTPLUG_CPU */
|
||||||
|
|
||||||
#endif /* #ifndef RCU_TREE_NONCORE */
|
#endif /* #ifndef RCU_TREE_NONCORE */
|
||||||
|
@@ -1,7 +1,7 @@
|
|||||||
/*
|
/*
|
||||||
* Read-Copy Update mechanism for mutual exclusion (tree-based version)
|
* Read-Copy Update mechanism for mutual exclusion (tree-based version)
|
||||||
* Internal non-public definitions that provide either classic
|
* Internal non-public definitions that provide either classic
|
||||||
* or preemptable semantics.
|
* or preemptible semantics.
|
||||||
*
|
*
|
||||||
* This program is free software; you can redistribute it and/or modify
|
* This program is free software; you can redistribute it and/or modify
|
||||||
* it under the terms of the GNU General Public License as published by
|
* it under the terms of the GNU General Public License as published by
|
||||||
@@ -54,10 +54,6 @@ static void __init rcu_bootup_announce_oddness(void)
|
|||||||
#ifdef CONFIG_RCU_TORTURE_TEST_RUNNABLE
|
#ifdef CONFIG_RCU_TORTURE_TEST_RUNNABLE
|
||||||
printk(KERN_INFO "\tRCU torture testing starts during boot.\n");
|
printk(KERN_INFO "\tRCU torture testing starts during boot.\n");
|
||||||
#endif
|
#endif
|
||||||
#ifndef CONFIG_RCU_CPU_STALL_DETECTOR
|
|
||||||
printk(KERN_INFO
|
|
||||||
"\tRCU-based detection of stalled CPUs is disabled.\n");
|
|
||||||
#endif
|
|
||||||
#if defined(CONFIG_TREE_PREEMPT_RCU) && !defined(CONFIG_RCU_CPU_STALL_VERBOSE)
|
#if defined(CONFIG_TREE_PREEMPT_RCU) && !defined(CONFIG_RCU_CPU_STALL_VERBOSE)
|
||||||
printk(KERN_INFO "\tVerbose stalled-CPUs detection is disabled.\n");
|
printk(KERN_INFO "\tVerbose stalled-CPUs detection is disabled.\n");
|
||||||
#endif
|
#endif
|
||||||
@@ -70,6 +66,7 @@ static void __init rcu_bootup_announce_oddness(void)
|
|||||||
|
|
||||||
struct rcu_state rcu_preempt_state = RCU_STATE_INITIALIZER(rcu_preempt_state);
|
struct rcu_state rcu_preempt_state = RCU_STATE_INITIALIZER(rcu_preempt_state);
|
||||||
DEFINE_PER_CPU(struct rcu_data, rcu_preempt_data);
|
DEFINE_PER_CPU(struct rcu_data, rcu_preempt_data);
|
||||||
|
static struct rcu_state *rcu_state = &rcu_preempt_state;
|
||||||
|
|
||||||
static int rcu_preempted_readers_exp(struct rcu_node *rnp);
|
static int rcu_preempted_readers_exp(struct rcu_node *rnp);
|
||||||
|
|
||||||
@@ -78,7 +75,7 @@ static int rcu_preempted_readers_exp(struct rcu_node *rnp);
|
|||||||
*/
|
*/
|
||||||
static void __init rcu_bootup_announce(void)
|
static void __init rcu_bootup_announce(void)
|
||||||
{
|
{
|
||||||
printk(KERN_INFO "Preemptable hierarchical RCU implementation.\n");
|
printk(KERN_INFO "Preemptible hierarchical RCU implementation.\n");
|
||||||
rcu_bootup_announce_oddness();
|
rcu_bootup_announce_oddness();
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -111,7 +108,7 @@ void rcu_force_quiescent_state(void)
|
|||||||
EXPORT_SYMBOL_GPL(rcu_force_quiescent_state);
|
EXPORT_SYMBOL_GPL(rcu_force_quiescent_state);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Record a preemptable-RCU quiescent state for the specified CPU. Note
|
* Record a preemptible-RCU quiescent state for the specified CPU. Note
|
||||||
* that this just means that the task currently running on the CPU is
|
* that this just means that the task currently running on the CPU is
|
||||||
* not in a quiescent state. There might be any number of tasks blocked
|
* not in a quiescent state. There might be any number of tasks blocked
|
||||||
* while in an RCU read-side critical section.
|
* while in an RCU read-side critical section.
|
||||||
@@ -134,12 +131,12 @@ static void rcu_preempt_qs(int cpu)
|
|||||||
* We have entered the scheduler, and the current task might soon be
|
* We have entered the scheduler, and the current task might soon be
|
||||||
* context-switched away from. If this task is in an RCU read-side
|
* context-switched away from. If this task is in an RCU read-side
|
||||||
* critical section, we will no longer be able to rely on the CPU to
|
* critical section, we will no longer be able to rely on the CPU to
|
||||||
* record that fact, so we enqueue the task on the appropriate entry
|
* record that fact, so we enqueue the task on the blkd_tasks list.
|
||||||
* of the blocked_tasks[] array. The task will dequeue itself when
|
* The task will dequeue itself when it exits the outermost enclosing
|
||||||
* it exits the outermost enclosing RCU read-side critical section.
|
* RCU read-side critical section. Therefore, the current grace period
|
||||||
* Therefore, the current grace period cannot be permitted to complete
|
* cannot be permitted to complete until the blkd_tasks list entries
|
||||||
* until the blocked_tasks[] entry indexed by the low-order bit of
|
* predating the current grace period drain, in other words, until
|
||||||
* rnp->gpnum empties.
|
* rnp->gp_tasks becomes NULL.
|
||||||
*
|
*
|
||||||
* Caller must disable preemption.
|
* Caller must disable preemption.
|
||||||
*/
|
*/
|
||||||
@@ -147,7 +144,6 @@ static void rcu_preempt_note_context_switch(int cpu)
|
|||||||
{
|
{
|
||||||
struct task_struct *t = current;
|
struct task_struct *t = current;
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
int phase;
|
|
||||||
struct rcu_data *rdp;
|
struct rcu_data *rdp;
|
||||||
struct rcu_node *rnp;
|
struct rcu_node *rnp;
|
||||||
|
|
||||||
@@ -169,15 +165,30 @@ static void rcu_preempt_note_context_switch(int cpu)
|
|||||||
* (i.e., this CPU has not yet passed through a quiescent
|
* (i.e., this CPU has not yet passed through a quiescent
|
||||||
* state for the current grace period), then as long
|
* state for the current grace period), then as long
|
||||||
* as that task remains queued, the current grace period
|
* as that task remains queued, the current grace period
|
||||||
* cannot end.
|
* cannot end. Note that there is some uncertainty as
|
||||||
|
* to exactly when the current grace period started.
|
||||||
|
* We take a conservative approach, which can result
|
||||||
|
* in unnecessarily waiting on tasks that started very
|
||||||
|
* slightly after the current grace period began. C'est
|
||||||
|
* la vie!!!
|
||||||
*
|
*
|
||||||
* But first, note that the current CPU must still be
|
* But first, note that the current CPU must still be
|
||||||
* on line!
|
* on line!
|
||||||
*/
|
*/
|
||||||
WARN_ON_ONCE((rdp->grpmask & rnp->qsmaskinit) == 0);
|
WARN_ON_ONCE((rdp->grpmask & rnp->qsmaskinit) == 0);
|
||||||
WARN_ON_ONCE(!list_empty(&t->rcu_node_entry));
|
WARN_ON_ONCE(!list_empty(&t->rcu_node_entry));
|
||||||
phase = (rnp->gpnum + !(rnp->qsmask & rdp->grpmask)) & 0x1;
|
if ((rnp->qsmask & rdp->grpmask) && rnp->gp_tasks != NULL) {
|
||||||
list_add(&t->rcu_node_entry, &rnp->blocked_tasks[phase]);
|
list_add(&t->rcu_node_entry, rnp->gp_tasks->prev);
|
||||||
|
rnp->gp_tasks = &t->rcu_node_entry;
|
||||||
|
#ifdef CONFIG_RCU_BOOST
|
||||||
|
if (rnp->boost_tasks != NULL)
|
||||||
|
rnp->boost_tasks = rnp->gp_tasks;
|
||||||
|
#endif /* #ifdef CONFIG_RCU_BOOST */
|
||||||
|
} else {
|
||||||
|
list_add(&t->rcu_node_entry, &rnp->blkd_tasks);
|
||||||
|
if (rnp->qsmask & rdp->grpmask)
|
||||||
|
rnp->gp_tasks = &t->rcu_node_entry;
|
||||||
|
}
|
||||||
raw_spin_unlock_irqrestore(&rnp->lock, flags);
|
raw_spin_unlock_irqrestore(&rnp->lock, flags);
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -196,7 +207,7 @@ static void rcu_preempt_note_context_switch(int cpu)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Tree-preemptable RCU implementation for rcu_read_lock().
|
* Tree-preemptible RCU implementation for rcu_read_lock().
|
||||||
* Just increment ->rcu_read_lock_nesting, shared state will be updated
|
* Just increment ->rcu_read_lock_nesting, shared state will be updated
|
||||||
* if we block.
|
* if we block.
|
||||||
*/
|
*/
|
||||||
@@ -212,12 +223,9 @@ EXPORT_SYMBOL_GPL(__rcu_read_lock);
|
|||||||
* for the specified rcu_node structure. If the caller needs a reliable
|
* for the specified rcu_node structure. If the caller needs a reliable
|
||||||
* answer, it must hold the rcu_node's ->lock.
|
* answer, it must hold the rcu_node's ->lock.
|
||||||
*/
|
*/
|
||||||
static int rcu_preempted_readers(struct rcu_node *rnp)
|
static int rcu_preempt_blocked_readers_cgp(struct rcu_node *rnp)
|
||||||
{
|
{
|
||||||
int phase = rnp->gpnum & 0x1;
|
return rnp->gp_tasks != NULL;
|
||||||
|
|
||||||
return !list_empty(&rnp->blocked_tasks[phase]) ||
|
|
||||||
!list_empty(&rnp->blocked_tasks[phase + 2]);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@@ -233,7 +241,7 @@ static void rcu_report_unblock_qs_rnp(struct rcu_node *rnp, unsigned long flags)
|
|||||||
unsigned long mask;
|
unsigned long mask;
|
||||||
struct rcu_node *rnp_p;
|
struct rcu_node *rnp_p;
|
||||||
|
|
||||||
if (rnp->qsmask != 0 || rcu_preempted_readers(rnp)) {
|
if (rnp->qsmask != 0 || rcu_preempt_blocked_readers_cgp(rnp)) {
|
||||||
raw_spin_unlock_irqrestore(&rnp->lock, flags);
|
raw_spin_unlock_irqrestore(&rnp->lock, flags);
|
||||||
return; /* Still need more quiescent states! */
|
return; /* Still need more quiescent states! */
|
||||||
}
|
}
|
||||||
@@ -256,6 +264,21 @@ static void rcu_report_unblock_qs_rnp(struct rcu_node *rnp, unsigned long flags)
|
|||||||
rcu_report_qs_rnp(mask, &rcu_preempt_state, rnp_p, flags);
|
rcu_report_qs_rnp(mask, &rcu_preempt_state, rnp_p, flags);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Advance a ->blkd_tasks-list pointer to the next entry, instead
|
||||||
|
* returning NULL if at the end of the list.
|
||||||
|
*/
|
||||||
|
static struct list_head *rcu_next_node_entry(struct task_struct *t,
|
||||||
|
struct rcu_node *rnp)
|
||||||
|
{
|
||||||
|
struct list_head *np;
|
||||||
|
|
||||||
|
np = t->rcu_node_entry.next;
|
||||||
|
if (np == &rnp->blkd_tasks)
|
||||||
|
np = NULL;
|
||||||
|
return np;
|
||||||
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Handle special cases during rcu_read_unlock(), such as needing to
|
* Handle special cases during rcu_read_unlock(), such as needing to
|
||||||
* notify RCU core processing or task having blocked during the RCU
|
* notify RCU core processing or task having blocked during the RCU
|
||||||
@@ -266,6 +289,7 @@ static void rcu_read_unlock_special(struct task_struct *t)
|
|||||||
int empty;
|
int empty;
|
||||||
int empty_exp;
|
int empty_exp;
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
|
struct list_head *np;
|
||||||
struct rcu_node *rnp;
|
struct rcu_node *rnp;
|
||||||
int special;
|
int special;
|
||||||
|
|
||||||
@@ -306,10 +330,19 @@ static void rcu_read_unlock_special(struct task_struct *t)
|
|||||||
break;
|
break;
|
||||||
raw_spin_unlock(&rnp->lock); /* irqs remain disabled. */
|
raw_spin_unlock(&rnp->lock); /* irqs remain disabled. */
|
||||||
}
|
}
|
||||||
empty = !rcu_preempted_readers(rnp);
|
empty = !rcu_preempt_blocked_readers_cgp(rnp);
|
||||||
empty_exp = !rcu_preempted_readers_exp(rnp);
|
empty_exp = !rcu_preempted_readers_exp(rnp);
|
||||||
smp_mb(); /* ensure expedited fastpath sees end of RCU c-s. */
|
smp_mb(); /* ensure expedited fastpath sees end of RCU c-s. */
|
||||||
|
np = rcu_next_node_entry(t, rnp);
|
||||||
list_del_init(&t->rcu_node_entry);
|
list_del_init(&t->rcu_node_entry);
|
||||||
|
if (&t->rcu_node_entry == rnp->gp_tasks)
|
||||||
|
rnp->gp_tasks = np;
|
||||||
|
if (&t->rcu_node_entry == rnp->exp_tasks)
|
||||||
|
rnp->exp_tasks = np;
|
||||||
|
#ifdef CONFIG_RCU_BOOST
|
||||||
|
if (&t->rcu_node_entry == rnp->boost_tasks)
|
||||||
|
rnp->boost_tasks = np;
|
||||||
|
#endif /* #ifdef CONFIG_RCU_BOOST */
|
||||||
t->rcu_blocked_node = NULL;
|
t->rcu_blocked_node = NULL;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@@ -322,6 +355,15 @@ static void rcu_read_unlock_special(struct task_struct *t)
|
|||||||
else
|
else
|
||||||
rcu_report_unblock_qs_rnp(rnp, flags);
|
rcu_report_unblock_qs_rnp(rnp, flags);
|
||||||
|
|
||||||
|
#ifdef CONFIG_RCU_BOOST
|
||||||
|
/* Unboost if we were boosted. */
|
||||||
|
if (special & RCU_READ_UNLOCK_BOOSTED) {
|
||||||
|
t->rcu_read_unlock_special &= ~RCU_READ_UNLOCK_BOOSTED;
|
||||||
|
rt_mutex_unlock(t->rcu_boost_mutex);
|
||||||
|
t->rcu_boost_mutex = NULL;
|
||||||
|
}
|
||||||
|
#endif /* #ifdef CONFIG_RCU_BOOST */
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* If this was the last task on the expedited lists,
|
* If this was the last task on the expedited lists,
|
||||||
* then we need to report up the rcu_node hierarchy.
|
* then we need to report up the rcu_node hierarchy.
|
||||||
@@ -334,7 +376,7 @@ static void rcu_read_unlock_special(struct task_struct *t)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Tree-preemptable RCU implementation for rcu_read_unlock().
|
* Tree-preemptible RCU implementation for rcu_read_unlock().
|
||||||
* Decrement ->rcu_read_lock_nesting. If the result is zero (outermost
|
* Decrement ->rcu_read_lock_nesting. If the result is zero (outermost
|
||||||
* rcu_read_unlock()) and ->rcu_read_unlock_special is non-zero, then
|
* rcu_read_unlock()) and ->rcu_read_unlock_special is non-zero, then
|
||||||
* invoke rcu_read_unlock_special() to clean up after a context switch
|
* invoke rcu_read_unlock_special() to clean up after a context switch
|
||||||
@@ -356,8 +398,6 @@ void __rcu_read_unlock(void)
|
|||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(__rcu_read_unlock);
|
EXPORT_SYMBOL_GPL(__rcu_read_unlock);
|
||||||
|
|
||||||
#ifdef CONFIG_RCU_CPU_STALL_DETECTOR
|
|
||||||
|
|
||||||
#ifdef CONFIG_RCU_CPU_STALL_VERBOSE
|
#ifdef CONFIG_RCU_CPU_STALL_VERBOSE
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@@ -367,18 +407,16 @@ EXPORT_SYMBOL_GPL(__rcu_read_unlock);
|
|||||||
static void rcu_print_detail_task_stall_rnp(struct rcu_node *rnp)
|
static void rcu_print_detail_task_stall_rnp(struct rcu_node *rnp)
|
||||||
{
|
{
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
struct list_head *lp;
|
|
||||||
int phase;
|
|
||||||
struct task_struct *t;
|
struct task_struct *t;
|
||||||
|
|
||||||
if (rcu_preempted_readers(rnp)) {
|
if (!rcu_preempt_blocked_readers_cgp(rnp))
|
||||||
|
return;
|
||||||
raw_spin_lock_irqsave(&rnp->lock, flags);
|
raw_spin_lock_irqsave(&rnp->lock, flags);
|
||||||
phase = rnp->gpnum & 0x1;
|
t = list_entry(rnp->gp_tasks,
|
||||||
lp = &rnp->blocked_tasks[phase];
|
struct task_struct, rcu_node_entry);
|
||||||
list_for_each_entry(t, lp, rcu_node_entry)
|
list_for_each_entry_continue(t, &rnp->blkd_tasks, rcu_node_entry)
|
||||||
sched_show_task(t);
|
sched_show_task(t);
|
||||||
raw_spin_unlock_irqrestore(&rnp->lock, flags);
|
raw_spin_unlock_irqrestore(&rnp->lock, flags);
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@@ -408,16 +446,14 @@ static void rcu_print_detail_task_stall(struct rcu_state *rsp)
|
|||||||
*/
|
*/
|
||||||
static void rcu_print_task_stall(struct rcu_node *rnp)
|
static void rcu_print_task_stall(struct rcu_node *rnp)
|
||||||
{
|
{
|
||||||
struct list_head *lp;
|
|
||||||
int phase;
|
|
||||||
struct task_struct *t;
|
struct task_struct *t;
|
||||||
|
|
||||||
if (rcu_preempted_readers(rnp)) {
|
if (!rcu_preempt_blocked_readers_cgp(rnp))
|
||||||
phase = rnp->gpnum & 0x1;
|
return;
|
||||||
lp = &rnp->blocked_tasks[phase];
|
t = list_entry(rnp->gp_tasks,
|
||||||
list_for_each_entry(t, lp, rcu_node_entry)
|
struct task_struct, rcu_node_entry);
|
||||||
|
list_for_each_entry_continue(t, &rnp->blkd_tasks, rcu_node_entry)
|
||||||
printk(" P%d", t->pid);
|
printk(" P%d", t->pid);
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@@ -430,18 +466,21 @@ static void rcu_preempt_stall_reset(void)
|
|||||||
rcu_preempt_state.jiffies_stall = jiffies + ULONG_MAX / 2;
|
rcu_preempt_state.jiffies_stall = jiffies + ULONG_MAX / 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
#endif /* #ifdef CONFIG_RCU_CPU_STALL_DETECTOR */
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Check that the list of blocked tasks for the newly completed grace
|
* Check that the list of blocked tasks for the newly completed grace
|
||||||
* period is in fact empty. It is a serious bug to complete a grace
|
* period is in fact empty. It is a serious bug to complete a grace
|
||||||
* period that still has RCU readers blocked! This function must be
|
* period that still has RCU readers blocked! This function must be
|
||||||
* invoked -before- updating this rnp's ->gpnum, and the rnp's ->lock
|
* invoked -before- updating this rnp's ->gpnum, and the rnp's ->lock
|
||||||
* must be held by the caller.
|
* must be held by the caller.
|
||||||
|
*
|
||||||
|
* Also, if there are blocked tasks on the list, they automatically
|
||||||
|
* block the newly created grace period, so set up ->gp_tasks accordingly.
|
||||||
*/
|
*/
|
||||||
static void rcu_preempt_check_blocked_tasks(struct rcu_node *rnp)
|
static void rcu_preempt_check_blocked_tasks(struct rcu_node *rnp)
|
||||||
{
|
{
|
||||||
WARN_ON_ONCE(rcu_preempted_readers(rnp));
|
WARN_ON_ONCE(rcu_preempt_blocked_readers_cgp(rnp));
|
||||||
|
if (!list_empty(&rnp->blkd_tasks))
|
||||||
|
rnp->gp_tasks = rnp->blkd_tasks.next;
|
||||||
WARN_ON_ONCE(rnp->qsmask);
|
WARN_ON_ONCE(rnp->qsmask);
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -465,50 +504,68 @@ static int rcu_preempt_offline_tasks(struct rcu_state *rsp,
|
|||||||
struct rcu_node *rnp,
|
struct rcu_node *rnp,
|
||||||
struct rcu_data *rdp)
|
struct rcu_data *rdp)
|
||||||
{
|
{
|
||||||
int i;
|
|
||||||
struct list_head *lp;
|
struct list_head *lp;
|
||||||
struct list_head *lp_root;
|
struct list_head *lp_root;
|
||||||
int retval = 0;
|
int retval = 0;
|
||||||
struct rcu_node *rnp_root = rcu_get_root(rsp);
|
struct rcu_node *rnp_root = rcu_get_root(rsp);
|
||||||
struct task_struct *tp;
|
struct task_struct *t;
|
||||||
|
|
||||||
if (rnp == rnp_root) {
|
if (rnp == rnp_root) {
|
||||||
WARN_ONCE(1, "Last CPU thought to be offlined?");
|
WARN_ONCE(1, "Last CPU thought to be offlined?");
|
||||||
return 0; /* Shouldn't happen: at least one CPU online. */
|
return 0; /* Shouldn't happen: at least one CPU online. */
|
||||||
}
|
}
|
||||||
WARN_ON_ONCE(rnp != rdp->mynode &&
|
|
||||||
(!list_empty(&rnp->blocked_tasks[0]) ||
|
/* If we are on an internal node, complain bitterly. */
|
||||||
!list_empty(&rnp->blocked_tasks[1]) ||
|
WARN_ON_ONCE(rnp != rdp->mynode);
|
||||||
!list_empty(&rnp->blocked_tasks[2]) ||
|
|
||||||
!list_empty(&rnp->blocked_tasks[3])));
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Move tasks up to root rcu_node. Rely on the fact that the
|
* Move tasks up to root rcu_node. Don't try to get fancy for
|
||||||
* root rcu_node can be at most one ahead of the rest of the
|
* this corner-case operation -- just put this node's tasks
|
||||||
* rcu_nodes in terms of gp_num value. This fact allows us to
|
* at the head of the root node's list, and update the root node's
|
||||||
* move the blocked_tasks[] array directly, element by element.
|
* ->gp_tasks and ->exp_tasks pointers to those of this node's,
|
||||||
|
* if non-NULL. This might result in waiting for more tasks than
|
||||||
|
* absolutely necessary, but this is a good performance/complexity
|
||||||
|
* tradeoff.
|
||||||
*/
|
*/
|
||||||
if (rcu_preempted_readers(rnp))
|
if (rcu_preempt_blocked_readers_cgp(rnp))
|
||||||
retval |= RCU_OFL_TASKS_NORM_GP;
|
retval |= RCU_OFL_TASKS_NORM_GP;
|
||||||
if (rcu_preempted_readers_exp(rnp))
|
if (rcu_preempted_readers_exp(rnp))
|
||||||
retval |= RCU_OFL_TASKS_EXP_GP;
|
retval |= RCU_OFL_TASKS_EXP_GP;
|
||||||
for (i = 0; i < 4; i++) {
|
lp = &rnp->blkd_tasks;
|
||||||
lp = &rnp->blocked_tasks[i];
|
lp_root = &rnp_root->blkd_tasks;
|
||||||
lp_root = &rnp_root->blocked_tasks[i];
|
|
||||||
while (!list_empty(lp)) {
|
while (!list_empty(lp)) {
|
||||||
tp = list_entry(lp->next, typeof(*tp), rcu_node_entry);
|
t = list_entry(lp->next, typeof(*t), rcu_node_entry);
|
||||||
raw_spin_lock(&rnp_root->lock); /* irqs already disabled */
|
raw_spin_lock(&rnp_root->lock); /* irqs already disabled */
|
||||||
list_del(&tp->rcu_node_entry);
|
list_del(&t->rcu_node_entry);
|
||||||
tp->rcu_blocked_node = rnp_root;
|
t->rcu_blocked_node = rnp_root;
|
||||||
list_add(&tp->rcu_node_entry, lp_root);
|
list_add(&t->rcu_node_entry, lp_root);
|
||||||
raw_spin_unlock(&rnp_root->lock); /* irqs remain disabled */
|
if (&t->rcu_node_entry == rnp->gp_tasks)
|
||||||
}
|
rnp_root->gp_tasks = rnp->gp_tasks;
|
||||||
|
if (&t->rcu_node_entry == rnp->exp_tasks)
|
||||||
|
rnp_root->exp_tasks = rnp->exp_tasks;
|
||||||
|
#ifdef CONFIG_RCU_BOOST
|
||||||
|
if (&t->rcu_node_entry == rnp->boost_tasks)
|
||||||
|
rnp_root->boost_tasks = rnp->boost_tasks;
|
||||||
|
#endif /* #ifdef CONFIG_RCU_BOOST */
|
||||||
|
raw_spin_unlock(&rnp_root->lock); /* irqs still disabled */
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#ifdef CONFIG_RCU_BOOST
|
||||||
|
/* In case root is being boosted and leaf is not. */
|
||||||
|
raw_spin_lock(&rnp_root->lock); /* irqs already disabled */
|
||||||
|
if (rnp_root->boost_tasks != NULL &&
|
||||||
|
rnp_root->boost_tasks != rnp_root->gp_tasks)
|
||||||
|
rnp_root->boost_tasks = rnp_root->gp_tasks;
|
||||||
|
raw_spin_unlock(&rnp_root->lock); /* irqs still disabled */
|
||||||
|
#endif /* #ifdef CONFIG_RCU_BOOST */
|
||||||
|
|
||||||
|
rnp->gp_tasks = NULL;
|
||||||
|
rnp->exp_tasks = NULL;
|
||||||
return retval;
|
return retval;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Do CPU-offline processing for preemptable RCU.
|
* Do CPU-offline processing for preemptible RCU.
|
||||||
*/
|
*/
|
||||||
static void rcu_preempt_offline_cpu(int cpu)
|
static void rcu_preempt_offline_cpu(int cpu)
|
||||||
{
|
{
|
||||||
@@ -537,7 +594,7 @@ static void rcu_preempt_check_callbacks(int cpu)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Process callbacks for preemptable RCU.
|
* Process callbacks for preemptible RCU.
|
||||||
*/
|
*/
|
||||||
static void rcu_preempt_process_callbacks(void)
|
static void rcu_preempt_process_callbacks(void)
|
||||||
{
|
{
|
||||||
@@ -546,7 +603,7 @@ static void rcu_preempt_process_callbacks(void)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Queue a preemptable-RCU callback for invocation after a grace period.
|
* Queue a preemptible-RCU callback for invocation after a grace period.
|
||||||
*/
|
*/
|
||||||
void call_rcu(struct rcu_head *head, void (*func)(struct rcu_head *rcu))
|
void call_rcu(struct rcu_head *head, void (*func)(struct rcu_head *rcu))
|
||||||
{
|
{
|
||||||
@@ -594,8 +651,7 @@ static DEFINE_MUTEX(sync_rcu_preempt_exp_mutex);
|
|||||||
*/
|
*/
|
||||||
static int rcu_preempted_readers_exp(struct rcu_node *rnp)
|
static int rcu_preempted_readers_exp(struct rcu_node *rnp)
|
||||||
{
|
{
|
||||||
return !list_empty(&rnp->blocked_tasks[2]) ||
|
return rnp->exp_tasks != NULL;
|
||||||
!list_empty(&rnp->blocked_tasks[3]);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@@ -655,12 +711,14 @@ static void rcu_report_exp_rnp(struct rcu_state *rsp, struct rcu_node *rnp)
|
|||||||
static void
|
static void
|
||||||
sync_rcu_preempt_exp_init(struct rcu_state *rsp, struct rcu_node *rnp)
|
sync_rcu_preempt_exp_init(struct rcu_state *rsp, struct rcu_node *rnp)
|
||||||
{
|
{
|
||||||
int must_wait;
|
int must_wait = 0;
|
||||||
|
|
||||||
raw_spin_lock(&rnp->lock); /* irqs already disabled */
|
raw_spin_lock(&rnp->lock); /* irqs already disabled */
|
||||||
list_splice_init(&rnp->blocked_tasks[0], &rnp->blocked_tasks[2]);
|
if (!list_empty(&rnp->blkd_tasks)) {
|
||||||
list_splice_init(&rnp->blocked_tasks[1], &rnp->blocked_tasks[3]);
|
rnp->exp_tasks = rnp->blkd_tasks.next;
|
||||||
must_wait = rcu_preempted_readers_exp(rnp);
|
rcu_initiate_boost(rnp);
|
||||||
|
must_wait = 1;
|
||||||
|
}
|
||||||
raw_spin_unlock(&rnp->lock); /* irqs remain disabled */
|
raw_spin_unlock(&rnp->lock); /* irqs remain disabled */
|
||||||
if (!must_wait)
|
if (!must_wait)
|
||||||
rcu_report_exp_rnp(rsp, rnp);
|
rcu_report_exp_rnp(rsp, rnp);
|
||||||
@@ -669,9 +727,7 @@ sync_rcu_preempt_exp_init(struct rcu_state *rsp, struct rcu_node *rnp)
|
|||||||
/*
|
/*
|
||||||
* Wait for an rcu-preempt grace period, but expedite it. The basic idea
|
* Wait for an rcu-preempt grace period, but expedite it. The basic idea
|
||||||
* is to invoke synchronize_sched_expedited() to push all the tasks to
|
* is to invoke synchronize_sched_expedited() to push all the tasks to
|
||||||
* the ->blocked_tasks[] lists, move all entries from the first set of
|
* the ->blkd_tasks lists and wait for this list to drain.
|
||||||
* ->blocked_tasks[] lists to the second set, and finally wait for this
|
|
||||||
* second set to drain.
|
|
||||||
*/
|
*/
|
||||||
void synchronize_rcu_expedited(void)
|
void synchronize_rcu_expedited(void)
|
||||||
{
|
{
|
||||||
@@ -703,7 +759,7 @@ void synchronize_rcu_expedited(void)
|
|||||||
if ((ACCESS_ONCE(sync_rcu_preempt_exp_count) - snap) > 0)
|
if ((ACCESS_ONCE(sync_rcu_preempt_exp_count) - snap) > 0)
|
||||||
goto unlock_mb_ret; /* Others did our work for us. */
|
goto unlock_mb_ret; /* Others did our work for us. */
|
||||||
|
|
||||||
/* force all RCU readers onto blocked_tasks[]. */
|
/* force all RCU readers onto ->blkd_tasks lists. */
|
||||||
synchronize_sched_expedited();
|
synchronize_sched_expedited();
|
||||||
|
|
||||||
raw_spin_lock_irqsave(&rsp->onofflock, flags);
|
raw_spin_lock_irqsave(&rsp->onofflock, flags);
|
||||||
@@ -715,7 +771,7 @@ void synchronize_rcu_expedited(void)
|
|||||||
raw_spin_unlock(&rnp->lock); /* irqs remain disabled. */
|
raw_spin_unlock(&rnp->lock); /* irqs remain disabled. */
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Snapshot current state of ->blocked_tasks[] lists. */
|
/* Snapshot current state of ->blkd_tasks lists. */
|
||||||
rcu_for_each_leaf_node(rsp, rnp)
|
rcu_for_each_leaf_node(rsp, rnp)
|
||||||
sync_rcu_preempt_exp_init(rsp, rnp);
|
sync_rcu_preempt_exp_init(rsp, rnp);
|
||||||
if (NUM_RCU_NODES > 1)
|
if (NUM_RCU_NODES > 1)
|
||||||
@@ -723,7 +779,7 @@ void synchronize_rcu_expedited(void)
|
|||||||
|
|
||||||
raw_spin_unlock_irqrestore(&rsp->onofflock, flags);
|
raw_spin_unlock_irqrestore(&rsp->onofflock, flags);
|
||||||
|
|
||||||
/* Wait for snapshotted ->blocked_tasks[] lists to drain. */
|
/* Wait for snapshotted ->blkd_tasks lists to drain. */
|
||||||
rnp = rcu_get_root(rsp);
|
rnp = rcu_get_root(rsp);
|
||||||
wait_event(sync_rcu_preempt_exp_wq,
|
wait_event(sync_rcu_preempt_exp_wq,
|
||||||
sync_rcu_preempt_exp_done(rnp));
|
sync_rcu_preempt_exp_done(rnp));
|
||||||
@@ -739,7 +795,7 @@ mb_ret:
|
|||||||
EXPORT_SYMBOL_GPL(synchronize_rcu_expedited);
|
EXPORT_SYMBOL_GPL(synchronize_rcu_expedited);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Check to see if there is any immediate preemptable-RCU-related work
|
* Check to see if there is any immediate preemptible-RCU-related work
|
||||||
* to be done.
|
* to be done.
|
||||||
*/
|
*/
|
||||||
static int rcu_preempt_pending(int cpu)
|
static int rcu_preempt_pending(int cpu)
|
||||||
@@ -749,7 +805,7 @@ static int rcu_preempt_pending(int cpu)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Does preemptable RCU need the CPU to stay out of dynticks mode?
|
* Does preemptible RCU need the CPU to stay out of dynticks mode?
|
||||||
*/
|
*/
|
||||||
static int rcu_preempt_needs_cpu(int cpu)
|
static int rcu_preempt_needs_cpu(int cpu)
|
||||||
{
|
{
|
||||||
@@ -766,7 +822,7 @@ void rcu_barrier(void)
|
|||||||
EXPORT_SYMBOL_GPL(rcu_barrier);
|
EXPORT_SYMBOL_GPL(rcu_barrier);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Initialize preemptable RCU's per-CPU data.
|
* Initialize preemptible RCU's per-CPU data.
|
||||||
*/
|
*/
|
||||||
static void __cpuinit rcu_preempt_init_percpu_data(int cpu)
|
static void __cpuinit rcu_preempt_init_percpu_data(int cpu)
|
||||||
{
|
{
|
||||||
@@ -774,7 +830,7 @@ static void __cpuinit rcu_preempt_init_percpu_data(int cpu)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Move preemptable RCU's callbacks from dying CPU to other online CPU.
|
* Move preemptible RCU's callbacks from dying CPU to other online CPU.
|
||||||
*/
|
*/
|
||||||
static void rcu_preempt_send_cbs_to_online(void)
|
static void rcu_preempt_send_cbs_to_online(void)
|
||||||
{
|
{
|
||||||
@@ -782,7 +838,7 @@ static void rcu_preempt_send_cbs_to_online(void)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Initialize preemptable RCU's state structures.
|
* Initialize preemptible RCU's state structures.
|
||||||
*/
|
*/
|
||||||
static void __init __rcu_init_preempt(void)
|
static void __init __rcu_init_preempt(void)
|
||||||
{
|
{
|
||||||
@@ -790,7 +846,7 @@ static void __init __rcu_init_preempt(void)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Check for a task exiting while in a preemptable-RCU read-side
|
* Check for a task exiting while in a preemptible-RCU read-side
|
||||||
* critical section, clean up if so. No need to issue warnings,
|
* critical section, clean up if so. No need to issue warnings,
|
||||||
* as debug_check_no_locks_held() already does this if lockdep
|
* as debug_check_no_locks_held() already does this if lockdep
|
||||||
* is enabled.
|
* is enabled.
|
||||||
@@ -802,11 +858,13 @@ void exit_rcu(void)
|
|||||||
if (t->rcu_read_lock_nesting == 0)
|
if (t->rcu_read_lock_nesting == 0)
|
||||||
return;
|
return;
|
||||||
t->rcu_read_lock_nesting = 1;
|
t->rcu_read_lock_nesting = 1;
|
||||||
rcu_read_unlock();
|
__rcu_read_unlock();
|
||||||
}
|
}
|
||||||
|
|
||||||
#else /* #ifdef CONFIG_TREE_PREEMPT_RCU */
|
#else /* #ifdef CONFIG_TREE_PREEMPT_RCU */
|
||||||
|
|
||||||
|
static struct rcu_state *rcu_state = &rcu_sched_state;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Tell them what RCU they are running.
|
* Tell them what RCU they are running.
|
||||||
*/
|
*/
|
||||||
@@ -836,7 +894,7 @@ void rcu_force_quiescent_state(void)
|
|||||||
EXPORT_SYMBOL_GPL(rcu_force_quiescent_state);
|
EXPORT_SYMBOL_GPL(rcu_force_quiescent_state);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Because preemptable RCU does not exist, we never have to check for
|
* Because preemptible RCU does not exist, we never have to check for
|
||||||
* CPUs being in quiescent states.
|
* CPUs being in quiescent states.
|
||||||
*/
|
*/
|
||||||
static void rcu_preempt_note_context_switch(int cpu)
|
static void rcu_preempt_note_context_switch(int cpu)
|
||||||
@@ -844,10 +902,10 @@ static void rcu_preempt_note_context_switch(int cpu)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Because preemptable RCU does not exist, there are never any preempted
|
* Because preemptible RCU does not exist, there are never any preempted
|
||||||
* RCU readers.
|
* RCU readers.
|
||||||
*/
|
*/
|
||||||
static int rcu_preempted_readers(struct rcu_node *rnp)
|
static int rcu_preempt_blocked_readers_cgp(struct rcu_node *rnp)
|
||||||
{
|
{
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
@@ -862,10 +920,8 @@ static void rcu_report_unblock_qs_rnp(struct rcu_node *rnp, unsigned long flags)
|
|||||||
|
|
||||||
#endif /* #ifdef CONFIG_HOTPLUG_CPU */
|
#endif /* #ifdef CONFIG_HOTPLUG_CPU */
|
||||||
|
|
||||||
#ifdef CONFIG_RCU_CPU_STALL_DETECTOR
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Because preemptable RCU does not exist, we never have to check for
|
* Because preemptible RCU does not exist, we never have to check for
|
||||||
* tasks blocked within RCU read-side critical sections.
|
* tasks blocked within RCU read-side critical sections.
|
||||||
*/
|
*/
|
||||||
static void rcu_print_detail_task_stall(struct rcu_state *rsp)
|
static void rcu_print_detail_task_stall(struct rcu_state *rsp)
|
||||||
@@ -873,7 +929,7 @@ static void rcu_print_detail_task_stall(struct rcu_state *rsp)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Because preemptable RCU does not exist, we never have to check for
|
* Because preemptible RCU does not exist, we never have to check for
|
||||||
* tasks blocked within RCU read-side critical sections.
|
* tasks blocked within RCU read-side critical sections.
|
||||||
*/
|
*/
|
||||||
static void rcu_print_task_stall(struct rcu_node *rnp)
|
static void rcu_print_task_stall(struct rcu_node *rnp)
|
||||||
@@ -888,10 +944,8 @@ static void rcu_preempt_stall_reset(void)
|
|||||||
{
|
{
|
||||||
}
|
}
|
||||||
|
|
||||||
#endif /* #ifdef CONFIG_RCU_CPU_STALL_DETECTOR */
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Because there is no preemptable RCU, there can be no readers blocked,
|
* Because there is no preemptible RCU, there can be no readers blocked,
|
||||||
* so there is no need to check for blocked tasks. So check only for
|
* so there is no need to check for blocked tasks. So check only for
|
||||||
* bogus qsmask values.
|
* bogus qsmask values.
|
||||||
*/
|
*/
|
||||||
@@ -903,7 +957,7 @@ static void rcu_preempt_check_blocked_tasks(struct rcu_node *rnp)
|
|||||||
#ifdef CONFIG_HOTPLUG_CPU
|
#ifdef CONFIG_HOTPLUG_CPU
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Because preemptable RCU does not exist, it never needs to migrate
|
* Because preemptible RCU does not exist, it never needs to migrate
|
||||||
* tasks that were blocked within RCU read-side critical sections, and
|
* tasks that were blocked within RCU read-side critical sections, and
|
||||||
* such non-existent tasks cannot possibly have been blocking the current
|
* such non-existent tasks cannot possibly have been blocking the current
|
||||||
* grace period.
|
* grace period.
|
||||||
@@ -916,7 +970,7 @@ static int rcu_preempt_offline_tasks(struct rcu_state *rsp,
|
|||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Because preemptable RCU does not exist, it never needs CPU-offline
|
* Because preemptible RCU does not exist, it never needs CPU-offline
|
||||||
* processing.
|
* processing.
|
||||||
*/
|
*/
|
||||||
static void rcu_preempt_offline_cpu(int cpu)
|
static void rcu_preempt_offline_cpu(int cpu)
|
||||||
@@ -926,7 +980,7 @@ static void rcu_preempt_offline_cpu(int cpu)
|
|||||||
#endif /* #ifdef CONFIG_HOTPLUG_CPU */
|
#endif /* #ifdef CONFIG_HOTPLUG_CPU */
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Because preemptable RCU does not exist, it never has any callbacks
|
* Because preemptible RCU does not exist, it never has any callbacks
|
||||||
* to check.
|
* to check.
|
||||||
*/
|
*/
|
||||||
static void rcu_preempt_check_callbacks(int cpu)
|
static void rcu_preempt_check_callbacks(int cpu)
|
||||||
@@ -934,7 +988,7 @@ static void rcu_preempt_check_callbacks(int cpu)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Because preemptable RCU does not exist, it never has any callbacks
|
* Because preemptible RCU does not exist, it never has any callbacks
|
||||||
* to process.
|
* to process.
|
||||||
*/
|
*/
|
||||||
static void rcu_preempt_process_callbacks(void)
|
static void rcu_preempt_process_callbacks(void)
|
||||||
@@ -943,7 +997,7 @@ static void rcu_preempt_process_callbacks(void)
|
|||||||
|
|
||||||
/*
|
/*
|
||||||
* Wait for an rcu-preempt grace period, but make it happen quickly.
|
* Wait for an rcu-preempt grace period, but make it happen quickly.
|
||||||
* But because preemptable RCU does not exist, map to rcu-sched.
|
* But because preemptible RCU does not exist, map to rcu-sched.
|
||||||
*/
|
*/
|
||||||
void synchronize_rcu_expedited(void)
|
void synchronize_rcu_expedited(void)
|
||||||
{
|
{
|
||||||
@@ -954,7 +1008,7 @@ EXPORT_SYMBOL_GPL(synchronize_rcu_expedited);
|
|||||||
#ifdef CONFIG_HOTPLUG_CPU
|
#ifdef CONFIG_HOTPLUG_CPU
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Because preemptable RCU does not exist, there is never any need to
|
* Because preemptible RCU does not exist, there is never any need to
|
||||||
* report on tasks preempted in RCU read-side critical sections during
|
* report on tasks preempted in RCU read-side critical sections during
|
||||||
* expedited RCU grace periods.
|
* expedited RCU grace periods.
|
||||||
*/
|
*/
|
||||||
@@ -966,7 +1020,7 @@ static void rcu_report_exp_rnp(struct rcu_state *rsp, struct rcu_node *rnp)
|
|||||||
#endif /* #ifdef CONFIG_HOTPLUG_CPU */
|
#endif /* #ifdef CONFIG_HOTPLUG_CPU */
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Because preemptable RCU does not exist, it never has any work to do.
|
* Because preemptible RCU does not exist, it never has any work to do.
|
||||||
*/
|
*/
|
||||||
static int rcu_preempt_pending(int cpu)
|
static int rcu_preempt_pending(int cpu)
|
||||||
{
|
{
|
||||||
@@ -974,7 +1028,7 @@ static int rcu_preempt_pending(int cpu)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Because preemptable RCU does not exist, it never needs any CPU.
|
* Because preemptible RCU does not exist, it never needs any CPU.
|
||||||
*/
|
*/
|
||||||
static int rcu_preempt_needs_cpu(int cpu)
|
static int rcu_preempt_needs_cpu(int cpu)
|
||||||
{
|
{
|
||||||
@@ -982,7 +1036,7 @@ static int rcu_preempt_needs_cpu(int cpu)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Because preemptable RCU does not exist, rcu_barrier() is just
|
* Because preemptible RCU does not exist, rcu_barrier() is just
|
||||||
* another name for rcu_barrier_sched().
|
* another name for rcu_barrier_sched().
|
||||||
*/
|
*/
|
||||||
void rcu_barrier(void)
|
void rcu_barrier(void)
|
||||||
@@ -992,7 +1046,7 @@ void rcu_barrier(void)
|
|||||||
EXPORT_SYMBOL_GPL(rcu_barrier);
|
EXPORT_SYMBOL_GPL(rcu_barrier);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Because preemptable RCU does not exist, there is no per-CPU
|
* Because preemptible RCU does not exist, there is no per-CPU
|
||||||
* data to initialize.
|
* data to initialize.
|
||||||
*/
|
*/
|
||||||
static void __cpuinit rcu_preempt_init_percpu_data(int cpu)
|
static void __cpuinit rcu_preempt_init_percpu_data(int cpu)
|
||||||
@@ -1000,14 +1054,14 @@ static void __cpuinit rcu_preempt_init_percpu_data(int cpu)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Because there is no preemptable RCU, there are no callbacks to move.
|
* Because there is no preemptible RCU, there are no callbacks to move.
|
||||||
*/
|
*/
|
||||||
static void rcu_preempt_send_cbs_to_online(void)
|
static void rcu_preempt_send_cbs_to_online(void)
|
||||||
{
|
{
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Because preemptable RCU does not exist, it need not be initialized.
|
* Because preemptible RCU does not exist, it need not be initialized.
|
||||||
*/
|
*/
|
||||||
static void __init __rcu_init_preempt(void)
|
static void __init __rcu_init_preempt(void)
|
||||||
{
|
{
|
||||||
@@ -1015,6 +1069,302 @@ static void __init __rcu_init_preempt(void)
|
|||||||
|
|
||||||
#endif /* #else #ifdef CONFIG_TREE_PREEMPT_RCU */
|
#endif /* #else #ifdef CONFIG_TREE_PREEMPT_RCU */
|
||||||
|
|
||||||
|
#ifdef CONFIG_RCU_BOOST
|
||||||
|
|
||||||
|
#include "rtmutex_common.h"
|
||||||
|
|
||||||
|
#ifdef CONFIG_RCU_TRACE
|
||||||
|
|
||||||
|
static void rcu_initiate_boost_trace(struct rcu_node *rnp)
|
||||||
|
{
|
||||||
|
if (list_empty(&rnp->blkd_tasks))
|
||||||
|
rnp->n_balk_blkd_tasks++;
|
||||||
|
else if (rnp->exp_tasks == NULL && rnp->gp_tasks == NULL)
|
||||||
|
rnp->n_balk_exp_gp_tasks++;
|
||||||
|
else if (rnp->gp_tasks != NULL && rnp->boost_tasks != NULL)
|
||||||
|
rnp->n_balk_boost_tasks++;
|
||||||
|
else if (rnp->gp_tasks != NULL && rnp->qsmask != 0)
|
||||||
|
rnp->n_balk_notblocked++;
|
||||||
|
else if (rnp->gp_tasks != NULL &&
|
||||||
|
ULONG_CMP_LT(jiffies, rnp->boost_time))
|
||||||
|
rnp->n_balk_notyet++;
|
||||||
|
else
|
||||||
|
rnp->n_balk_nos++;
|
||||||
|
}
|
||||||
|
|
||||||
|
#else /* #ifdef CONFIG_RCU_TRACE */
|
||||||
|
|
||||||
|
static void rcu_initiate_boost_trace(struct rcu_node *rnp)
|
||||||
|
{
|
||||||
|
}
|
||||||
|
|
||||||
|
#endif /* #else #ifdef CONFIG_RCU_TRACE */
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Carry out RCU priority boosting on the task indicated by ->exp_tasks
|
||||||
|
* or ->boost_tasks, advancing the pointer to the next task in the
|
||||||
|
* ->blkd_tasks list.
|
||||||
|
*
|
||||||
|
* Note that irqs must be enabled: boosting the task can block.
|
||||||
|
* Returns 1 if there are more tasks needing to be boosted.
|
||||||
|
*/
|
||||||
|
static int rcu_boost(struct rcu_node *rnp)
|
||||||
|
{
|
||||||
|
unsigned long flags;
|
||||||
|
struct rt_mutex mtx;
|
||||||
|
struct task_struct *t;
|
||||||
|
struct list_head *tb;
|
||||||
|
|
||||||
|
if (rnp->exp_tasks == NULL && rnp->boost_tasks == NULL)
|
||||||
|
return 0; /* Nothing left to boost. */
|
||||||
|
|
||||||
|
raw_spin_lock_irqsave(&rnp->lock, flags);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Recheck under the lock: all tasks in need of boosting
|
||||||
|
* might exit their RCU read-side critical sections on their own.
|
||||||
|
*/
|
||||||
|
if (rnp->exp_tasks == NULL && rnp->boost_tasks == NULL) {
|
||||||
|
raw_spin_unlock_irqrestore(&rnp->lock, flags);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Preferentially boost tasks blocking expedited grace periods.
|
||||||
|
* This cannot starve the normal grace periods because a second
|
||||||
|
* expedited grace period must boost all blocked tasks, including
|
||||||
|
* those blocking the pre-existing normal grace period.
|
||||||
|
*/
|
||||||
|
if (rnp->exp_tasks != NULL) {
|
||||||
|
tb = rnp->exp_tasks;
|
||||||
|
rnp->n_exp_boosts++;
|
||||||
|
} else {
|
||||||
|
tb = rnp->boost_tasks;
|
||||||
|
rnp->n_normal_boosts++;
|
||||||
|
}
|
||||||
|
rnp->n_tasks_boosted++;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* We boost task t by manufacturing an rt_mutex that appears to
|
||||||
|
* be held by task t. We leave a pointer to that rt_mutex where
|
||||||
|
* task t can find it, and task t will release the mutex when it
|
||||||
|
* exits its outermost RCU read-side critical section. Then
|
||||||
|
* simply acquiring this artificial rt_mutex will boost task
|
||||||
|
* t's priority. (Thanks to tglx for suggesting this approach!)
|
||||||
|
*
|
||||||
|
* Note that task t must acquire rnp->lock to remove itself from
|
||||||
|
* the ->blkd_tasks list, which it will do from exit() if from
|
||||||
|
* nowhere else. We therefore are guaranteed that task t will
|
||||||
|
* stay around at least until we drop rnp->lock. Note that
|
||||||
|
* rnp->lock also resolves races between our priority boosting
|
||||||
|
* and task t's exiting its outermost RCU read-side critical
|
||||||
|
* section.
|
||||||
|
*/
|
||||||
|
t = container_of(tb, struct task_struct, rcu_node_entry);
|
||||||
|
rt_mutex_init_proxy_locked(&mtx, t);
|
||||||
|
t->rcu_boost_mutex = &mtx;
|
||||||
|
t->rcu_read_unlock_special |= RCU_READ_UNLOCK_BOOSTED;
|
||||||
|
raw_spin_unlock_irqrestore(&rnp->lock, flags);
|
||||||
|
rt_mutex_lock(&mtx); /* Side effect: boosts task t's priority. */
|
||||||
|
rt_mutex_unlock(&mtx); /* Keep lockdep happy. */
|
||||||
|
|
||||||
|
return rnp->exp_tasks != NULL || rnp->boost_tasks != NULL;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Timer handler to initiate waking up of boost kthreads that
|
||||||
|
* have yielded the CPU due to excessive numbers of tasks to
|
||||||
|
* boost. We wake up the per-rcu_node kthread, which in turn
|
||||||
|
* will wake up the booster kthread.
|
||||||
|
*/
|
||||||
|
static void rcu_boost_kthread_timer(unsigned long arg)
|
||||||
|
{
|
||||||
|
unsigned long flags;
|
||||||
|
struct rcu_node *rnp = (struct rcu_node *)arg;
|
||||||
|
|
||||||
|
raw_spin_lock_irqsave(&rnp->lock, flags);
|
||||||
|
invoke_rcu_node_kthread(rnp);
|
||||||
|
raw_spin_unlock_irqrestore(&rnp->lock, flags);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Priority-boosting kthread. One per leaf rcu_node and one for the
|
||||||
|
* root rcu_node.
|
||||||
|
*/
|
||||||
|
static int rcu_boost_kthread(void *arg)
|
||||||
|
{
|
||||||
|
struct rcu_node *rnp = (struct rcu_node *)arg;
|
||||||
|
int spincnt = 0;
|
||||||
|
int more2boost;
|
||||||
|
|
||||||
|
for (;;) {
|
||||||
|
rnp->boost_kthread_status = RCU_KTHREAD_WAITING;
|
||||||
|
wait_event_interruptible(rnp->boost_wq, rnp->boost_tasks ||
|
||||||
|
rnp->exp_tasks ||
|
||||||
|
kthread_should_stop());
|
||||||
|
if (kthread_should_stop())
|
||||||
|
break;
|
||||||
|
rnp->boost_kthread_status = RCU_KTHREAD_RUNNING;
|
||||||
|
more2boost = rcu_boost(rnp);
|
||||||
|
if (more2boost)
|
||||||
|
spincnt++;
|
||||||
|
else
|
||||||
|
spincnt = 0;
|
||||||
|
if (spincnt > 10) {
|
||||||
|
rcu_yield(rcu_boost_kthread_timer, (unsigned long)rnp);
|
||||||
|
spincnt = 0;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
rnp->boost_kthread_status = RCU_KTHREAD_STOPPED;
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Check to see if it is time to start boosting RCU readers that are
|
||||||
|
* blocking the current grace period, and, if so, tell the per-rcu_node
|
||||||
|
* kthread to start boosting them. If there is an expedited grace
|
||||||
|
* period in progress, it is always time to boost.
|
||||||
|
*
|
||||||
|
* The caller must hold rnp->lock.
|
||||||
|
*/
|
||||||
|
static void rcu_initiate_boost(struct rcu_node *rnp)
|
||||||
|
{
|
||||||
|
struct task_struct *t;
|
||||||
|
|
||||||
|
if (!rcu_preempt_blocked_readers_cgp(rnp) && rnp->exp_tasks == NULL) {
|
||||||
|
rnp->n_balk_exp_gp_tasks++;
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
if (rnp->exp_tasks != NULL ||
|
||||||
|
(rnp->gp_tasks != NULL &&
|
||||||
|
rnp->boost_tasks == NULL &&
|
||||||
|
rnp->qsmask == 0 &&
|
||||||
|
ULONG_CMP_GE(jiffies, rnp->boost_time))) {
|
||||||
|
if (rnp->exp_tasks == NULL)
|
||||||
|
rnp->boost_tasks = rnp->gp_tasks;
|
||||||
|
t = rnp->boost_kthread_task;
|
||||||
|
if (t != NULL)
|
||||||
|
wake_up_process(t);
|
||||||
|
} else
|
||||||
|
rcu_initiate_boost_trace(rnp);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Set the affinity of the boost kthread. The CPU-hotplug locks are
|
||||||
|
* held, so no one should be messing with the existence of the boost
|
||||||
|
* kthread.
|
||||||
|
*/
|
||||||
|
static void rcu_boost_kthread_setaffinity(struct rcu_node *rnp,
|
||||||
|
cpumask_var_t cm)
|
||||||
|
{
|
||||||
|
struct task_struct *t;
|
||||||
|
|
||||||
|
t = rnp->boost_kthread_task;
|
||||||
|
if (t != NULL)
|
||||||
|
set_cpus_allowed_ptr(rnp->boost_kthread_task, cm);
|
||||||
|
}
|
||||||
|
|
||||||
|
#define RCU_BOOST_DELAY_JIFFIES DIV_ROUND_UP(CONFIG_RCU_BOOST_DELAY * HZ, 1000)
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Do priority-boost accounting for the start of a new grace period.
|
||||||
|
*/
|
||||||
|
static void rcu_preempt_boost_start_gp(struct rcu_node *rnp)
|
||||||
|
{
|
||||||
|
rnp->boost_time = jiffies + RCU_BOOST_DELAY_JIFFIES;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Initialize the RCU-boost waitqueue.
|
||||||
|
*/
|
||||||
|
static void __init rcu_init_boost_waitqueue(struct rcu_node *rnp)
|
||||||
|
{
|
||||||
|
init_waitqueue_head(&rnp->boost_wq);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Create an RCU-boost kthread for the specified node if one does not
|
||||||
|
* already exist. We only create this kthread for preemptible RCU.
|
||||||
|
* Returns zero if all is well, a negated errno otherwise.
|
||||||
|
*/
|
||||||
|
static int __cpuinit rcu_spawn_one_boost_kthread(struct rcu_state *rsp,
|
||||||
|
struct rcu_node *rnp,
|
||||||
|
int rnp_index)
|
||||||
|
{
|
||||||
|
unsigned long flags;
|
||||||
|
struct sched_param sp;
|
||||||
|
struct task_struct *t;
|
||||||
|
|
||||||
|
if (&rcu_preempt_state != rsp)
|
||||||
|
return 0;
|
||||||
|
if (rnp->boost_kthread_task != NULL)
|
||||||
|
return 0;
|
||||||
|
t = kthread_create(rcu_boost_kthread, (void *)rnp,
|
||||||
|
"rcub%d", rnp_index);
|
||||||
|
if (IS_ERR(t))
|
||||||
|
return PTR_ERR(t);
|
||||||
|
raw_spin_lock_irqsave(&rnp->lock, flags);
|
||||||
|
rnp->boost_kthread_task = t;
|
||||||
|
raw_spin_unlock_irqrestore(&rnp->lock, flags);
|
||||||
|
wake_up_process(t);
|
||||||
|
sp.sched_priority = RCU_KTHREAD_PRIO;
|
||||||
|
sched_setscheduler_nocheck(t, SCHED_FIFO, &sp);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
#ifdef CONFIG_HOTPLUG_CPU
|
||||||
|
|
||||||
|
static void rcu_stop_boost_kthread(struct rcu_node *rnp)
|
||||||
|
{
|
||||||
|
unsigned long flags;
|
||||||
|
struct task_struct *t;
|
||||||
|
|
||||||
|
raw_spin_lock_irqsave(&rnp->lock, flags);
|
||||||
|
t = rnp->boost_kthread_task;
|
||||||
|
rnp->boost_kthread_task = NULL;
|
||||||
|
raw_spin_unlock_irqrestore(&rnp->lock, flags);
|
||||||
|
if (t != NULL)
|
||||||
|
kthread_stop(t);
|
||||||
|
}
|
||||||
|
|
||||||
|
#endif /* #ifdef CONFIG_HOTPLUG_CPU */
|
||||||
|
|
||||||
|
#else /* #ifdef CONFIG_RCU_BOOST */
|
||||||
|
|
||||||
|
static void rcu_initiate_boost(struct rcu_node *rnp)
|
||||||
|
{
|
||||||
|
}
|
||||||
|
|
||||||
|
static void rcu_boost_kthread_setaffinity(struct rcu_node *rnp,
|
||||||
|
cpumask_var_t cm)
|
||||||
|
{
|
||||||
|
}
|
||||||
|
|
||||||
|
static void rcu_preempt_boost_start_gp(struct rcu_node *rnp)
|
||||||
|
{
|
||||||
|
}
|
||||||
|
|
||||||
|
static void __init rcu_init_boost_waitqueue(struct rcu_node *rnp)
|
||||||
|
{
|
||||||
|
}
|
||||||
|
|
||||||
|
static int __cpuinit rcu_spawn_one_boost_kthread(struct rcu_state *rsp,
|
||||||
|
struct rcu_node *rnp,
|
||||||
|
int rnp_index)
|
||||||
|
{
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
#ifdef CONFIG_HOTPLUG_CPU
|
||||||
|
|
||||||
|
static void rcu_stop_boost_kthread(struct rcu_node *rnp)
|
||||||
|
{
|
||||||
|
}
|
||||||
|
|
||||||
|
#endif /* #ifdef CONFIG_HOTPLUG_CPU */
|
||||||
|
|
||||||
|
#endif /* #else #ifdef CONFIG_RCU_BOOST */
|
||||||
|
|
||||||
#ifndef CONFIG_SMP
|
#ifndef CONFIG_SMP
|
||||||
|
|
||||||
void synchronize_sched_expedited(void)
|
void synchronize_sched_expedited(void)
|
||||||
@@ -1187,14 +1537,13 @@ static DEFINE_PER_CPU(unsigned long, rcu_dyntick_holdoff);
|
|||||||
*
|
*
|
||||||
* Because it is not legal to invoke rcu_process_callbacks() with irqs
|
* Because it is not legal to invoke rcu_process_callbacks() with irqs
|
||||||
* disabled, we do one pass of force_quiescent_state(), then do a
|
* disabled, we do one pass of force_quiescent_state(), then do a
|
||||||
* raise_softirq() to cause rcu_process_callbacks() to be invoked later.
|
* invoke_rcu_cpu_kthread() to cause rcu_process_callbacks() to be invoked
|
||||||
* The per-cpu rcu_dyntick_drain variable controls the sequencing.
|
* later. The per-cpu rcu_dyntick_drain variable controls the sequencing.
|
||||||
*/
|
*/
|
||||||
int rcu_needs_cpu(int cpu)
|
int rcu_needs_cpu(int cpu)
|
||||||
{
|
{
|
||||||
int c = 0;
|
int c = 0;
|
||||||
int snap;
|
int snap;
|
||||||
int snap_nmi;
|
|
||||||
int thatcpu;
|
int thatcpu;
|
||||||
|
|
||||||
/* Check for being in the holdoff period. */
|
/* Check for being in the holdoff period. */
|
||||||
@@ -1205,10 +1554,10 @@ int rcu_needs_cpu(int cpu)
|
|||||||
for_each_online_cpu(thatcpu) {
|
for_each_online_cpu(thatcpu) {
|
||||||
if (thatcpu == cpu)
|
if (thatcpu == cpu)
|
||||||
continue;
|
continue;
|
||||||
snap = per_cpu(rcu_dynticks, thatcpu).dynticks;
|
snap = atomic_add_return(0, &per_cpu(rcu_dynticks,
|
||||||
snap_nmi = per_cpu(rcu_dynticks, thatcpu).dynticks_nmi;
|
thatcpu).dynticks);
|
||||||
smp_mb(); /* Order sampling of snap with end of grace period. */
|
smp_mb(); /* Order sampling of snap with end of grace period. */
|
||||||
if (((snap & 0x1) != 0) || ((snap_nmi & 0x1) != 0)) {
|
if ((snap & 0x1) != 0) {
|
||||||
per_cpu(rcu_dyntick_drain, cpu) = 0;
|
per_cpu(rcu_dyntick_drain, cpu) = 0;
|
||||||
per_cpu(rcu_dyntick_holdoff, cpu) = jiffies - 1;
|
per_cpu(rcu_dyntick_holdoff, cpu) = jiffies - 1;
|
||||||
return rcu_needs_cpu_quick_check(cpu);
|
return rcu_needs_cpu_quick_check(cpu);
|
||||||
@@ -1239,7 +1588,7 @@ int rcu_needs_cpu(int cpu)
|
|||||||
|
|
||||||
/* If RCU callbacks are still pending, RCU still needs this CPU. */
|
/* If RCU callbacks are still pending, RCU still needs this CPU. */
|
||||||
if (c)
|
if (c)
|
||||||
raise_softirq(RCU_SOFTIRQ);
|
invoke_rcu_cpu_kthread();
|
||||||
return c;
|
return c;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@@ -46,6 +46,18 @@
|
|||||||
#define RCU_TREE_NONCORE
|
#define RCU_TREE_NONCORE
|
||||||
#include "rcutree.h"
|
#include "rcutree.h"
|
||||||
|
|
||||||
|
DECLARE_PER_CPU(unsigned int, rcu_cpu_kthread_status);
|
||||||
|
DECLARE_PER_CPU(unsigned int, rcu_cpu_kthread_cpu);
|
||||||
|
DECLARE_PER_CPU(unsigned int, rcu_cpu_kthread_loops);
|
||||||
|
DECLARE_PER_CPU(char, rcu_cpu_has_work);
|
||||||
|
|
||||||
|
static char convert_kthread_status(unsigned int kthread_status)
|
||||||
|
{
|
||||||
|
if (kthread_status > RCU_KTHREAD_MAX)
|
||||||
|
return '?';
|
||||||
|
return "SRWOY"[kthread_status];
|
||||||
|
}
|
||||||
|
|
||||||
static void print_one_rcu_data(struct seq_file *m, struct rcu_data *rdp)
|
static void print_one_rcu_data(struct seq_file *m, struct rcu_data *rdp)
|
||||||
{
|
{
|
||||||
if (!rdp->beenonline)
|
if (!rdp->beenonline)
|
||||||
@@ -57,14 +69,28 @@ static void print_one_rcu_data(struct seq_file *m, struct rcu_data *rdp)
|
|||||||
rdp->passed_quiesc, rdp->passed_quiesc_completed,
|
rdp->passed_quiesc, rdp->passed_quiesc_completed,
|
||||||
rdp->qs_pending);
|
rdp->qs_pending);
|
||||||
#ifdef CONFIG_NO_HZ
|
#ifdef CONFIG_NO_HZ
|
||||||
seq_printf(m, " dt=%d/%d dn=%d df=%lu",
|
seq_printf(m, " dt=%d/%d/%d df=%lu",
|
||||||
rdp->dynticks->dynticks,
|
atomic_read(&rdp->dynticks->dynticks),
|
||||||
rdp->dynticks->dynticks_nesting,
|
rdp->dynticks->dynticks_nesting,
|
||||||
rdp->dynticks->dynticks_nmi,
|
rdp->dynticks->dynticks_nmi_nesting,
|
||||||
rdp->dynticks_fqs);
|
rdp->dynticks_fqs);
|
||||||
#endif /* #ifdef CONFIG_NO_HZ */
|
#endif /* #ifdef CONFIG_NO_HZ */
|
||||||
seq_printf(m, " of=%lu ri=%lu", rdp->offline_fqs, rdp->resched_ipi);
|
seq_printf(m, " of=%lu ri=%lu", rdp->offline_fqs, rdp->resched_ipi);
|
||||||
seq_printf(m, " ql=%ld b=%ld", rdp->qlen, rdp->blimit);
|
seq_printf(m, " ql=%ld qs=%c%c%c%c kt=%d/%c/%d ktl=%x b=%ld",
|
||||||
|
rdp->qlen,
|
||||||
|
".N"[rdp->nxttail[RCU_NEXT_READY_TAIL] !=
|
||||||
|
rdp->nxttail[RCU_NEXT_TAIL]],
|
||||||
|
".R"[rdp->nxttail[RCU_WAIT_TAIL] !=
|
||||||
|
rdp->nxttail[RCU_NEXT_READY_TAIL]],
|
||||||
|
".W"[rdp->nxttail[RCU_DONE_TAIL] !=
|
||||||
|
rdp->nxttail[RCU_WAIT_TAIL]],
|
||||||
|
".D"[&rdp->nxtlist != rdp->nxttail[RCU_DONE_TAIL]],
|
||||||
|
per_cpu(rcu_cpu_has_work, rdp->cpu),
|
||||||
|
convert_kthread_status(per_cpu(rcu_cpu_kthread_status,
|
||||||
|
rdp->cpu)),
|
||||||
|
per_cpu(rcu_cpu_kthread_cpu, rdp->cpu),
|
||||||
|
per_cpu(rcu_cpu_kthread_loops, rdp->cpu) & 0xffff,
|
||||||
|
rdp->blimit);
|
||||||
seq_printf(m, " ci=%lu co=%lu ca=%lu\n",
|
seq_printf(m, " ci=%lu co=%lu ca=%lu\n",
|
||||||
rdp->n_cbs_invoked, rdp->n_cbs_orphaned, rdp->n_cbs_adopted);
|
rdp->n_cbs_invoked, rdp->n_cbs_orphaned, rdp->n_cbs_adopted);
|
||||||
}
|
}
|
||||||
@@ -115,13 +141,24 @@ static void print_one_rcu_data_csv(struct seq_file *m, struct rcu_data *rdp)
|
|||||||
rdp->qs_pending);
|
rdp->qs_pending);
|
||||||
#ifdef CONFIG_NO_HZ
|
#ifdef CONFIG_NO_HZ
|
||||||
seq_printf(m, ",%d,%d,%d,%lu",
|
seq_printf(m, ",%d,%d,%d,%lu",
|
||||||
rdp->dynticks->dynticks,
|
atomic_read(&rdp->dynticks->dynticks),
|
||||||
rdp->dynticks->dynticks_nesting,
|
rdp->dynticks->dynticks_nesting,
|
||||||
rdp->dynticks->dynticks_nmi,
|
rdp->dynticks->dynticks_nmi_nesting,
|
||||||
rdp->dynticks_fqs);
|
rdp->dynticks_fqs);
|
||||||
#endif /* #ifdef CONFIG_NO_HZ */
|
#endif /* #ifdef CONFIG_NO_HZ */
|
||||||
seq_printf(m, ",%lu,%lu", rdp->offline_fqs, rdp->resched_ipi);
|
seq_printf(m, ",%lu,%lu", rdp->offline_fqs, rdp->resched_ipi);
|
||||||
seq_printf(m, ",%ld,%ld", rdp->qlen, rdp->blimit);
|
seq_printf(m, ",%ld,\"%c%c%c%c\",%d,\"%c\",%ld", rdp->qlen,
|
||||||
|
".N"[rdp->nxttail[RCU_NEXT_READY_TAIL] !=
|
||||||
|
rdp->nxttail[RCU_NEXT_TAIL]],
|
||||||
|
".R"[rdp->nxttail[RCU_WAIT_TAIL] !=
|
||||||
|
rdp->nxttail[RCU_NEXT_READY_TAIL]],
|
||||||
|
".W"[rdp->nxttail[RCU_DONE_TAIL] !=
|
||||||
|
rdp->nxttail[RCU_WAIT_TAIL]],
|
||||||
|
".D"[&rdp->nxtlist != rdp->nxttail[RCU_DONE_TAIL]],
|
||||||
|
per_cpu(rcu_cpu_has_work, rdp->cpu),
|
||||||
|
convert_kthread_status(per_cpu(rcu_cpu_kthread_status,
|
||||||
|
rdp->cpu)),
|
||||||
|
rdp->blimit);
|
||||||
seq_printf(m, ",%lu,%lu,%lu\n",
|
seq_printf(m, ",%lu,%lu,%lu\n",
|
||||||
rdp->n_cbs_invoked, rdp->n_cbs_orphaned, rdp->n_cbs_adopted);
|
rdp->n_cbs_invoked, rdp->n_cbs_orphaned, rdp->n_cbs_adopted);
|
||||||
}
|
}
|
||||||
@@ -130,7 +167,7 @@ static int show_rcudata_csv(struct seq_file *m, void *unused)
|
|||||||
{
|
{
|
||||||
seq_puts(m, "\"CPU\",\"Online?\",\"c\",\"g\",\"pq\",\"pqc\",\"pq\",");
|
seq_puts(m, "\"CPU\",\"Online?\",\"c\",\"g\",\"pq\",\"pqc\",\"pq\",");
|
||||||
#ifdef CONFIG_NO_HZ
|
#ifdef CONFIG_NO_HZ
|
||||||
seq_puts(m, "\"dt\",\"dt nesting\",\"dn\",\"df\",");
|
seq_puts(m, "\"dt\",\"dt nesting\",\"dt NMI nesting\",\"df\",");
|
||||||
#endif /* #ifdef CONFIG_NO_HZ */
|
#endif /* #ifdef CONFIG_NO_HZ */
|
||||||
seq_puts(m, "\"of\",\"ri\",\"ql\",\"b\",\"ci\",\"co\",\"ca\"\n");
|
seq_puts(m, "\"of\",\"ri\",\"ql\",\"b\",\"ci\",\"co\",\"ca\"\n");
|
||||||
#ifdef CONFIG_TREE_PREEMPT_RCU
|
#ifdef CONFIG_TREE_PREEMPT_RCU
|
||||||
@@ -157,11 +194,76 @@ static const struct file_operations rcudata_csv_fops = {
|
|||||||
.release = single_release,
|
.release = single_release,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
#ifdef CONFIG_RCU_BOOST
|
||||||
|
|
||||||
|
static void print_one_rcu_node_boost(struct seq_file *m, struct rcu_node *rnp)
|
||||||
|
{
|
||||||
|
seq_printf(m, "%d:%d tasks=%c%c%c%c kt=%c ntb=%lu neb=%lu nnb=%lu "
|
||||||
|
"j=%04x bt=%04x\n",
|
||||||
|
rnp->grplo, rnp->grphi,
|
||||||
|
"T."[list_empty(&rnp->blkd_tasks)],
|
||||||
|
"N."[!rnp->gp_tasks],
|
||||||
|
"E."[!rnp->exp_tasks],
|
||||||
|
"B."[!rnp->boost_tasks],
|
||||||
|
convert_kthread_status(rnp->boost_kthread_status),
|
||||||
|
rnp->n_tasks_boosted, rnp->n_exp_boosts,
|
||||||
|
rnp->n_normal_boosts,
|
||||||
|
(int)(jiffies & 0xffff),
|
||||||
|
(int)(rnp->boost_time & 0xffff));
|
||||||
|
seq_printf(m, "%s: nt=%lu egt=%lu bt=%lu nb=%lu ny=%lu nos=%lu\n",
|
||||||
|
" balk",
|
||||||
|
rnp->n_balk_blkd_tasks,
|
||||||
|
rnp->n_balk_exp_gp_tasks,
|
||||||
|
rnp->n_balk_boost_tasks,
|
||||||
|
rnp->n_balk_notblocked,
|
||||||
|
rnp->n_balk_notyet,
|
||||||
|
rnp->n_balk_nos);
|
||||||
|
}
|
||||||
|
|
||||||
|
static int show_rcu_node_boost(struct seq_file *m, void *unused)
|
||||||
|
{
|
||||||
|
struct rcu_node *rnp;
|
||||||
|
|
||||||
|
rcu_for_each_leaf_node(&rcu_preempt_state, rnp)
|
||||||
|
print_one_rcu_node_boost(m, rnp);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
static int rcu_node_boost_open(struct inode *inode, struct file *file)
|
||||||
|
{
|
||||||
|
return single_open(file, show_rcu_node_boost, NULL);
|
||||||
|
}
|
||||||
|
|
||||||
|
static const struct file_operations rcu_node_boost_fops = {
|
||||||
|
.owner = THIS_MODULE,
|
||||||
|
.open = rcu_node_boost_open,
|
||||||
|
.read = seq_read,
|
||||||
|
.llseek = seq_lseek,
|
||||||
|
.release = single_release,
|
||||||
|
};
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Create the rcuboost debugfs entry. Standard error return.
|
||||||
|
*/
|
||||||
|
static int rcu_boost_trace_create_file(struct dentry *rcudir)
|
||||||
|
{
|
||||||
|
return !debugfs_create_file("rcuboost", 0444, rcudir, NULL,
|
||||||
|
&rcu_node_boost_fops);
|
||||||
|
}
|
||||||
|
|
||||||
|
#else /* #ifdef CONFIG_RCU_BOOST */
|
||||||
|
|
||||||
|
static int rcu_boost_trace_create_file(struct dentry *rcudir)
|
||||||
|
{
|
||||||
|
return 0; /* There cannot be an error if we didn't create it! */
|
||||||
|
}
|
||||||
|
|
||||||
|
#endif /* #else #ifdef CONFIG_RCU_BOOST */
|
||||||
|
|
||||||
static void print_one_rcu_state(struct seq_file *m, struct rcu_state *rsp)
|
static void print_one_rcu_state(struct seq_file *m, struct rcu_state *rsp)
|
||||||
{
|
{
|
||||||
unsigned long gpnum;
|
unsigned long gpnum;
|
||||||
int level = 0;
|
int level = 0;
|
||||||
int phase;
|
|
||||||
struct rcu_node *rnp;
|
struct rcu_node *rnp;
|
||||||
|
|
||||||
gpnum = rsp->gpnum;
|
gpnum = rsp->gpnum;
|
||||||
@@ -178,13 +280,11 @@ static void print_one_rcu_state(struct seq_file *m, struct rcu_state *rsp)
|
|||||||
seq_puts(m, "\n");
|
seq_puts(m, "\n");
|
||||||
level = rnp->level;
|
level = rnp->level;
|
||||||
}
|
}
|
||||||
phase = gpnum & 0x1;
|
seq_printf(m, "%lx/%lx %c%c>%c %d:%d ^%d ",
|
||||||
seq_printf(m, "%lx/%lx %c%c>%c%c %d:%d ^%d ",
|
|
||||||
rnp->qsmask, rnp->qsmaskinit,
|
rnp->qsmask, rnp->qsmaskinit,
|
||||||
"T."[list_empty(&rnp->blocked_tasks[phase])],
|
".G"[rnp->gp_tasks != NULL],
|
||||||
"E."[list_empty(&rnp->blocked_tasks[phase + 2])],
|
".E"[rnp->exp_tasks != NULL],
|
||||||
"T."[list_empty(&rnp->blocked_tasks[!phase])],
|
".T"[!list_empty(&rnp->blkd_tasks)],
|
||||||
"E."[list_empty(&rnp->blocked_tasks[!phase + 2])],
|
|
||||||
rnp->grplo, rnp->grphi, rnp->grpnum);
|
rnp->grplo, rnp->grphi, rnp->grpnum);
|
||||||
}
|
}
|
||||||
seq_puts(m, "\n");
|
seq_puts(m, "\n");
|
||||||
@@ -216,16 +316,35 @@ static const struct file_operations rcuhier_fops = {
|
|||||||
.release = single_release,
|
.release = single_release,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
static void show_one_rcugp(struct seq_file *m, struct rcu_state *rsp)
|
||||||
|
{
|
||||||
|
unsigned long flags;
|
||||||
|
unsigned long completed;
|
||||||
|
unsigned long gpnum;
|
||||||
|
unsigned long gpage;
|
||||||
|
unsigned long gpmax;
|
||||||
|
struct rcu_node *rnp = &rsp->node[0];
|
||||||
|
|
||||||
|
raw_spin_lock_irqsave(&rnp->lock, flags);
|
||||||
|
completed = rsp->completed;
|
||||||
|
gpnum = rsp->gpnum;
|
||||||
|
if (rsp->completed == rsp->gpnum)
|
||||||
|
gpage = 0;
|
||||||
|
else
|
||||||
|
gpage = jiffies - rsp->gp_start;
|
||||||
|
gpmax = rsp->gp_max;
|
||||||
|
raw_spin_unlock_irqrestore(&rnp->lock, flags);
|
||||||
|
seq_printf(m, "%s: completed=%ld gpnum=%lu age=%ld max=%ld\n",
|
||||||
|
rsp->name, completed, gpnum, gpage, gpmax);
|
||||||
|
}
|
||||||
|
|
||||||
static int show_rcugp(struct seq_file *m, void *unused)
|
static int show_rcugp(struct seq_file *m, void *unused)
|
||||||
{
|
{
|
||||||
#ifdef CONFIG_TREE_PREEMPT_RCU
|
#ifdef CONFIG_TREE_PREEMPT_RCU
|
||||||
seq_printf(m, "rcu_preempt: completed=%ld gpnum=%lu\n",
|
show_one_rcugp(m, &rcu_preempt_state);
|
||||||
rcu_preempt_state.completed, rcu_preempt_state.gpnum);
|
|
||||||
#endif /* #ifdef CONFIG_TREE_PREEMPT_RCU */
|
#endif /* #ifdef CONFIG_TREE_PREEMPT_RCU */
|
||||||
seq_printf(m, "rcu_sched: completed=%ld gpnum=%lu\n",
|
show_one_rcugp(m, &rcu_sched_state);
|
||||||
rcu_sched_state.completed, rcu_sched_state.gpnum);
|
show_one_rcugp(m, &rcu_bh_state);
|
||||||
seq_printf(m, "rcu_bh: completed=%ld gpnum=%lu\n",
|
|
||||||
rcu_bh_state.completed, rcu_bh_state.gpnum);
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -298,6 +417,29 @@ static const struct file_operations rcu_pending_fops = {
|
|||||||
.release = single_release,
|
.release = single_release,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
static int show_rcutorture(struct seq_file *m, void *unused)
|
||||||
|
{
|
||||||
|
seq_printf(m, "rcutorture test sequence: %lu %s\n",
|
||||||
|
rcutorture_testseq >> 1,
|
||||||
|
(rcutorture_testseq & 0x1) ? "(test in progress)" : "");
|
||||||
|
seq_printf(m, "rcutorture update version number: %lu\n",
|
||||||
|
rcutorture_vernum);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
static int rcutorture_open(struct inode *inode, struct file *file)
|
||||||
|
{
|
||||||
|
return single_open(file, show_rcutorture, NULL);
|
||||||
|
}
|
||||||
|
|
||||||
|
static const struct file_operations rcutorture_fops = {
|
||||||
|
.owner = THIS_MODULE,
|
||||||
|
.open = rcutorture_open,
|
||||||
|
.read = seq_read,
|
||||||
|
.llseek = seq_lseek,
|
||||||
|
.release = single_release,
|
||||||
|
};
|
||||||
|
|
||||||
static struct dentry *rcudir;
|
static struct dentry *rcudir;
|
||||||
|
|
||||||
static int __init rcutree_trace_init(void)
|
static int __init rcutree_trace_init(void)
|
||||||
@@ -318,6 +460,9 @@ static int __init rcutree_trace_init(void)
|
|||||||
if (!retval)
|
if (!retval)
|
||||||
goto free_out;
|
goto free_out;
|
||||||
|
|
||||||
|
if (rcu_boost_trace_create_file(rcudir))
|
||||||
|
goto free_out;
|
||||||
|
|
||||||
retval = debugfs_create_file("rcugp", 0444, rcudir, NULL, &rcugp_fops);
|
retval = debugfs_create_file("rcugp", 0444, rcudir, NULL, &rcugp_fops);
|
||||||
if (!retval)
|
if (!retval)
|
||||||
goto free_out;
|
goto free_out;
|
||||||
@@ -331,6 +476,11 @@ static int __init rcutree_trace_init(void)
|
|||||||
NULL, &rcu_pending_fops);
|
NULL, &rcu_pending_fops);
|
||||||
if (!retval)
|
if (!retval)
|
||||||
goto free_out;
|
goto free_out;
|
||||||
|
|
||||||
|
retval = debugfs_create_file("rcutorture", 0444, rcudir,
|
||||||
|
NULL, &rcutorture_fops);
|
||||||
|
if (!retval)
|
||||||
|
goto free_out;
|
||||||
return 0;
|
return 0;
|
||||||
free_out:
|
free_out:
|
||||||
debugfs_remove_recursive(rcudir);
|
debugfs_remove_recursive(rcudir);
|
||||||
|
@@ -58,7 +58,7 @@ DEFINE_PER_CPU(struct task_struct *, ksoftirqd);
|
|||||||
|
|
||||||
char *softirq_to_name[NR_SOFTIRQS] = {
|
char *softirq_to_name[NR_SOFTIRQS] = {
|
||||||
"HI", "TIMER", "NET_TX", "NET_RX", "BLOCK", "BLOCK_IOPOLL",
|
"HI", "TIMER", "NET_TX", "NET_RX", "BLOCK", "BLOCK_IOPOLL",
|
||||||
"TASKLET", "SCHED", "HRTIMER", "RCU"
|
"TASKLET", "SCHED", "HRTIMER"
|
||||||
};
|
};
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@@ -337,7 +337,7 @@ config DEBUG_OBJECTS_WORK
|
|||||||
|
|
||||||
config DEBUG_OBJECTS_RCU_HEAD
|
config DEBUG_OBJECTS_RCU_HEAD
|
||||||
bool "Debug RCU callbacks objects"
|
bool "Debug RCU callbacks objects"
|
||||||
depends on DEBUG_OBJECTS && PREEMPT
|
depends on DEBUG_OBJECTS
|
||||||
help
|
help
|
||||||
Enable this to turn on debugging of RCU list heads (call_rcu() usage).
|
Enable this to turn on debugging of RCU list heads (call_rcu() usage).
|
||||||
|
|
||||||
@@ -875,22 +875,9 @@ config RCU_TORTURE_TEST_RUNNABLE
|
|||||||
Say N here if you want the RCU torture tests to start only
|
Say N here if you want the RCU torture tests to start only
|
||||||
after being manually enabled via /proc.
|
after being manually enabled via /proc.
|
||||||
|
|
||||||
config RCU_CPU_STALL_DETECTOR
|
|
||||||
bool "Check for stalled CPUs delaying RCU grace periods"
|
|
||||||
depends on TREE_RCU || TREE_PREEMPT_RCU
|
|
||||||
default y
|
|
||||||
help
|
|
||||||
This option causes RCU to printk information on which
|
|
||||||
CPUs are delaying the current grace period, but only when
|
|
||||||
the grace period extends for excessive time periods.
|
|
||||||
|
|
||||||
Say N if you want to disable such checks.
|
|
||||||
|
|
||||||
Say Y if you are unsure.
|
|
||||||
|
|
||||||
config RCU_CPU_STALL_TIMEOUT
|
config RCU_CPU_STALL_TIMEOUT
|
||||||
int "RCU CPU stall timeout in seconds"
|
int "RCU CPU stall timeout in seconds"
|
||||||
depends on RCU_CPU_STALL_DETECTOR
|
depends on TREE_RCU || TREE_PREEMPT_RCU
|
||||||
range 3 300
|
range 3 300
|
||||||
default 60
|
default 60
|
||||||
help
|
help
|
||||||
@@ -899,22 +886,9 @@ config RCU_CPU_STALL_TIMEOUT
|
|||||||
RCU grace period persists, additional CPU stall warnings are
|
RCU grace period persists, additional CPU stall warnings are
|
||||||
printed at more widely spaced intervals.
|
printed at more widely spaced intervals.
|
||||||
|
|
||||||
config RCU_CPU_STALL_DETECTOR_RUNNABLE
|
|
||||||
bool "RCU CPU stall checking starts automatically at boot"
|
|
||||||
depends on RCU_CPU_STALL_DETECTOR
|
|
||||||
default y
|
|
||||||
help
|
|
||||||
If set, start checking for RCU CPU stalls immediately on
|
|
||||||
boot. Otherwise, RCU CPU stall checking must be manually
|
|
||||||
enabled.
|
|
||||||
|
|
||||||
Say Y if you are unsure.
|
|
||||||
|
|
||||||
Say N if you wish to suppress RCU CPU stall checking during boot.
|
|
||||||
|
|
||||||
config RCU_CPU_STALL_VERBOSE
|
config RCU_CPU_STALL_VERBOSE
|
||||||
bool "Print additional per-task information for RCU_CPU_STALL_DETECTOR"
|
bool "Print additional per-task information for RCU_CPU_STALL_DETECTOR"
|
||||||
depends on RCU_CPU_STALL_DETECTOR && TREE_PREEMPT_RCU
|
depends on TREE_PREEMPT_RCU
|
||||||
default y
|
default y
|
||||||
help
|
help
|
||||||
This option causes RCU to printk detailed per-task information
|
This option causes RCU to printk detailed per-task information
|
||||||
|
@@ -2187,7 +2187,6 @@ static const struct flag flags[] = {
|
|||||||
{ "TASKLET_SOFTIRQ", 6 },
|
{ "TASKLET_SOFTIRQ", 6 },
|
||||||
{ "SCHED_SOFTIRQ", 7 },
|
{ "SCHED_SOFTIRQ", 7 },
|
||||||
{ "HRTIMER_SOFTIRQ", 8 },
|
{ "HRTIMER_SOFTIRQ", 8 },
|
||||||
{ "RCU_SOFTIRQ", 9 },
|
|
||||||
|
|
||||||
{ "HRTIMER_NORESTART", 0 },
|
{ "HRTIMER_NORESTART", 0 },
|
||||||
{ "HRTIMER_RESTART", 1 },
|
{ "HRTIMER_RESTART", 1 },
|
||||||
|
Reference in New Issue
Block a user