2761 lines
133 KiB
ReStructuredText
2761 lines
133 KiB
ReStructuredText
|
=================================
|
|||
|
A Tour Through RCU's Requirements
|
|||
|
=================================
|
|||
|
|
|||
|
Copyright IBM Corporation, 2015
|
|||
|
|
|||
|
Author: Paul E. McKenney
|
|||
|
|
|||
|
The initial version of this document appeared in the
|
|||
|
`LWN <https://lwn.net/>`_ on those articles:
|
|||
|
`part 1 <https://lwn.net/Articles/652156/>`_,
|
|||
|
`part 2 <https://lwn.net/Articles/652677/>`_, and
|
|||
|
`part 3 <https://lwn.net/Articles/653326/>`_.
|
|||
|
|
|||
|
Introduction
|
|||
|
------------
|
|||
|
|
|||
|
Read-copy update (RCU) is a synchronization mechanism that is often used
|
|||
|
as a replacement for reader-writer locking. RCU is unusual in that
|
|||
|
updaters do not block readers, which means that RCU's read-side
|
|||
|
primitives can be exceedingly fast and scalable. In addition, updaters
|
|||
|
can make useful forward progress concurrently with readers. However, all
|
|||
|
this concurrency between RCU readers and updaters does raise the
|
|||
|
question of exactly what RCU readers are doing, which in turn raises the
|
|||
|
question of exactly what RCU's requirements are.
|
|||
|
|
|||
|
This document therefore summarizes RCU's requirements, and can be
|
|||
|
thought of as an informal, high-level specification for RCU. It is
|
|||
|
important to understand that RCU's specification is primarily empirical
|
|||
|
in nature; in fact, I learned about many of these requirements the hard
|
|||
|
way. This situation might cause some consternation, however, not only
|
|||
|
has this learning process been a lot of fun, but it has also been a
|
|||
|
great privilege to work with so many people willing to apply
|
|||
|
technologies in interesting new ways.
|
|||
|
|
|||
|
All that aside, here are the categories of currently known RCU
|
|||
|
requirements:
|
|||
|
|
|||
|
#. `Fundamental Requirements`_
|
|||
|
#. `Fundamental Non-Requirements`_
|
|||
|
#. `Parallelism Facts of Life`_
|
|||
|
#. `Quality-of-Implementation Requirements`_
|
|||
|
#. `Linux Kernel Complications`_
|
|||
|
#. `Software-Engineering Requirements`_
|
|||
|
#. `Other RCU Flavors`_
|
|||
|
#. `Possible Future Changes`_
|
|||
|
|
|||
|
This is followed by a summary_, however, the answers to
|
|||
|
each quick quiz immediately follows the quiz. Select the big white space
|
|||
|
with your mouse to see the answer.
|
|||
|
|
|||
|
Fundamental Requirements
|
|||
|
------------------------
|
|||
|
|
|||
|
RCU's fundamental requirements are the closest thing RCU has to hard
|
|||
|
mathematical requirements. These are:
|
|||
|
|
|||
|
#. `Grace-Period Guarantee`_
|
|||
|
#. `Publish/Subscribe Guarantee`_
|
|||
|
#. `Memory-Barrier Guarantees`_
|
|||
|
#. `RCU Primitives Guaranteed to Execute Unconditionally`_
|
|||
|
#. `Guaranteed Read-to-Write Upgrade`_
|
|||
|
|
|||
|
Grace-Period Guarantee
|
|||
|
~~~~~~~~~~~~~~~~~~~~~~
|
|||
|
|
|||
|
RCU's grace-period guarantee is unusual in being premeditated: Jack
|
|||
|
Slingwine and I had this guarantee firmly in mind when we started work
|
|||
|
on RCU (then called “rclock”) in the early 1990s. That said, the past
|
|||
|
two decades of experience with RCU have produced a much more detailed
|
|||
|
understanding of this guarantee.
|
|||
|
|
|||
|
RCU's grace-period guarantee allows updaters to wait for the completion
|
|||
|
of all pre-existing RCU read-side critical sections. An RCU read-side
|
|||
|
critical section begins with the marker rcu_read_lock() and ends
|
|||
|
with the marker rcu_read_unlock(). These markers may be nested, and
|
|||
|
RCU treats a nested set as one big RCU read-side critical section.
|
|||
|
Production-quality implementations of rcu_read_lock() and
|
|||
|
rcu_read_unlock() are extremely lightweight, and in fact have
|
|||
|
exactly zero overhead in Linux kernels built for production use with
|
|||
|
``CONFIG_PREEMPTION=n``.
|
|||
|
|
|||
|
This guarantee allows ordering to be enforced with extremely low
|
|||
|
overhead to readers, for example:
|
|||
|
|
|||
|
::
|
|||
|
|
|||
|
1 int x, y;
|
|||
|
2
|
|||
|
3 void thread0(void)
|
|||
|
4 {
|
|||
|
5 rcu_read_lock();
|
|||
|
6 r1 = READ_ONCE(x);
|
|||
|
7 r2 = READ_ONCE(y);
|
|||
|
8 rcu_read_unlock();
|
|||
|
9 }
|
|||
|
10
|
|||
|
11 void thread1(void)
|
|||
|
12 {
|
|||
|
13 WRITE_ONCE(x, 1);
|
|||
|
14 synchronize_rcu();
|
|||
|
15 WRITE_ONCE(y, 1);
|
|||
|
16 }
|
|||
|
|
|||
|
Because the synchronize_rcu() on line 14 waits for all pre-existing
|
|||
|
readers, any instance of thread0() that loads a value of zero from
|
|||
|
``x`` must complete before thread1() stores to ``y``, so that
|
|||
|
instance must also load a value of zero from ``y``. Similarly, any
|
|||
|
instance of thread0() that loads a value of one from ``y`` must have
|
|||
|
started after the synchronize_rcu() started, and must therefore also
|
|||
|
load a value of one from ``x``. Therefore, the outcome:
|
|||
|
|
|||
|
::
|
|||
|
|
|||
|
(r1 == 0 && r2 == 1)
|
|||
|
|
|||
|
cannot happen.
|
|||
|
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Quick Quiz**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| Wait a minute! You said that updaters can make useful forward |
|
|||
|
| progress concurrently with readers, but pre-existing readers will |
|
|||
|
| block synchronize_rcu()!!! |
|
|||
|
| Just who are you trying to fool??? |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Answer**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| First, if updaters do not wish to be blocked by readers, they can use |
|
|||
|
| call_rcu() or kfree_rcu(), which will be discussed later. |
|
|||
|
| Second, even when using synchronize_rcu(), the other update-side |
|
|||
|
| code does run concurrently with readers, whether pre-existing or not. |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
|
|||
|
This scenario resembles one of the first uses of RCU in
|
|||
|
`DYNIX/ptx <https://en.wikipedia.org/wiki/DYNIX>`__, which managed a
|
|||
|
distributed lock manager's transition into a state suitable for handling
|
|||
|
recovery from node failure, more or less as follows:
|
|||
|
|
|||
|
::
|
|||
|
|
|||
|
1 #define STATE_NORMAL 0
|
|||
|
2 #define STATE_WANT_RECOVERY 1
|
|||
|
3 #define STATE_RECOVERING 2
|
|||
|
4 #define STATE_WANT_NORMAL 3
|
|||
|
5
|
|||
|
6 int state = STATE_NORMAL;
|
|||
|
7
|
|||
|
8 void do_something_dlm(void)
|
|||
|
9 {
|
|||
|
10 int state_snap;
|
|||
|
11
|
|||
|
12 rcu_read_lock();
|
|||
|
13 state_snap = READ_ONCE(state);
|
|||
|
14 if (state_snap == STATE_NORMAL)
|
|||
|
15 do_something();
|
|||
|
16 else
|
|||
|
17 do_something_carefully();
|
|||
|
18 rcu_read_unlock();
|
|||
|
19 }
|
|||
|
20
|
|||
|
21 void start_recovery(void)
|
|||
|
22 {
|
|||
|
23 WRITE_ONCE(state, STATE_WANT_RECOVERY);
|
|||
|
24 synchronize_rcu();
|
|||
|
25 WRITE_ONCE(state, STATE_RECOVERING);
|
|||
|
26 recovery();
|
|||
|
27 WRITE_ONCE(state, STATE_WANT_NORMAL);
|
|||
|
28 synchronize_rcu();
|
|||
|
29 WRITE_ONCE(state, STATE_NORMAL);
|
|||
|
30 }
|
|||
|
|
|||
|
The RCU read-side critical section in do_something_dlm() works with
|
|||
|
the synchronize_rcu() in start_recovery() to guarantee that
|
|||
|
do_something() never runs concurrently with recovery(), but with
|
|||
|
little or no synchronization overhead in do_something_dlm().
|
|||
|
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Quick Quiz**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| Why is the synchronize_rcu() on line 28 needed? |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Answer**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| Without that extra grace period, memory reordering could result in |
|
|||
|
| do_something_dlm() executing do_something() concurrently with |
|
|||
|
| the last bits of recovery(). |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
|
|||
|
In order to avoid fatal problems such as deadlocks, an RCU read-side
|
|||
|
critical section must not contain calls to synchronize_rcu().
|
|||
|
Similarly, an RCU read-side critical section must not contain anything
|
|||
|
that waits, directly or indirectly, on completion of an invocation of
|
|||
|
synchronize_rcu().
|
|||
|
|
|||
|
Although RCU's grace-period guarantee is useful in and of itself, with
|
|||
|
`quite a few use cases <https://lwn.net/Articles/573497/>`__, it would
|
|||
|
be good to be able to use RCU to coordinate read-side access to linked
|
|||
|
data structures. For this, the grace-period guarantee is not sufficient,
|
|||
|
as can be seen in function add_gp_buggy() below. We will look at the
|
|||
|
reader's code later, but in the meantime, just think of the reader as
|
|||
|
locklessly picking up the ``gp`` pointer, and, if the value loaded is
|
|||
|
non-\ ``NULL``, locklessly accessing the ``->a`` and ``->b`` fields.
|
|||
|
|
|||
|
::
|
|||
|
|
|||
|
1 bool add_gp_buggy(int a, int b)
|
|||
|
2 {
|
|||
|
3 p = kmalloc(sizeof(*p), GFP_KERNEL);
|
|||
|
4 if (!p)
|
|||
|
5 return -ENOMEM;
|
|||
|
6 spin_lock(&gp_lock);
|
|||
|
7 if (rcu_access_pointer(gp)) {
|
|||
|
8 spin_unlock(&gp_lock);
|
|||
|
9 return false;
|
|||
|
10 }
|
|||
|
11 p->a = a;
|
|||
|
12 p->b = a;
|
|||
|
13 gp = p; /* ORDERING BUG */
|
|||
|
14 spin_unlock(&gp_lock);
|
|||
|
15 return true;
|
|||
|
16 }
|
|||
|
|
|||
|
The problem is that both the compiler and weakly ordered CPUs are within
|
|||
|
their rights to reorder this code as follows:
|
|||
|
|
|||
|
::
|
|||
|
|
|||
|
1 bool add_gp_buggy_optimized(int a, int b)
|
|||
|
2 {
|
|||
|
3 p = kmalloc(sizeof(*p), GFP_KERNEL);
|
|||
|
4 if (!p)
|
|||
|
5 return -ENOMEM;
|
|||
|
6 spin_lock(&gp_lock);
|
|||
|
7 if (rcu_access_pointer(gp)) {
|
|||
|
8 spin_unlock(&gp_lock);
|
|||
|
9 return false;
|
|||
|
10 }
|
|||
|
11 gp = p; /* ORDERING BUG */
|
|||
|
12 p->a = a;
|
|||
|
13 p->b = a;
|
|||
|
14 spin_unlock(&gp_lock);
|
|||
|
15 return true;
|
|||
|
16 }
|
|||
|
|
|||
|
If an RCU reader fetches ``gp`` just after ``add_gp_buggy_optimized``
|
|||
|
executes line 11, it will see garbage in the ``->a`` and ``->b`` fields.
|
|||
|
And this is but one of many ways in which compiler and hardware
|
|||
|
optimizations could cause trouble. Therefore, we clearly need some way
|
|||
|
to prevent the compiler and the CPU from reordering in this manner,
|
|||
|
which brings us to the publish-subscribe guarantee discussed in the next
|
|||
|
section.
|
|||
|
|
|||
|
Publish/Subscribe Guarantee
|
|||
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|||
|
|
|||
|
RCU's publish-subscribe guarantee allows data to be inserted into a
|
|||
|
linked data structure without disrupting RCU readers. The updater uses
|
|||
|
rcu_assign_pointer() to insert the new data, and readers use
|
|||
|
rcu_dereference() to access data, whether new or old. The following
|
|||
|
shows an example of insertion:
|
|||
|
|
|||
|
::
|
|||
|
|
|||
|
1 bool add_gp(int a, int b)
|
|||
|
2 {
|
|||
|
3 p = kmalloc(sizeof(*p), GFP_KERNEL);
|
|||
|
4 if (!p)
|
|||
|
5 return -ENOMEM;
|
|||
|
6 spin_lock(&gp_lock);
|
|||
|
7 if (rcu_access_pointer(gp)) {
|
|||
|
8 spin_unlock(&gp_lock);
|
|||
|
9 return false;
|
|||
|
10 }
|
|||
|
11 p->a = a;
|
|||
|
12 p->b = a;
|
|||
|
13 rcu_assign_pointer(gp, p);
|
|||
|
14 spin_unlock(&gp_lock);
|
|||
|
15 return true;
|
|||
|
16 }
|
|||
|
|
|||
|
The rcu_assign_pointer() on line 13 is conceptually equivalent to a
|
|||
|
simple assignment statement, but also guarantees that its assignment
|
|||
|
will happen after the two assignments in lines 11 and 12, similar to the
|
|||
|
C11 ``memory_order_release`` store operation. It also prevents any
|
|||
|
number of “interesting” compiler optimizations, for example, the use of
|
|||
|
``gp`` as a scratch location immediately preceding the assignment.
|
|||
|
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Quick Quiz**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| But rcu_assign_pointer() does nothing to prevent the two |
|
|||
|
| assignments to ``p->a`` and ``p->b`` from being reordered. Can't that |
|
|||
|
| also cause problems? |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Answer**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| No, it cannot. The readers cannot see either of these two fields |
|
|||
|
| until the assignment to ``gp``, by which time both fields are fully |
|
|||
|
| initialized. So reordering the assignments to ``p->a`` and ``p->b`` |
|
|||
|
| cannot possibly cause any problems. |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
|
|||
|
It is tempting to assume that the reader need not do anything special to
|
|||
|
control its accesses to the RCU-protected data, as shown in
|
|||
|
do_something_gp_buggy() below:
|
|||
|
|
|||
|
::
|
|||
|
|
|||
|
1 bool do_something_gp_buggy(void)
|
|||
|
2 {
|
|||
|
3 rcu_read_lock();
|
|||
|
4 p = gp; /* OPTIMIZATIONS GALORE!!! */
|
|||
|
5 if (p) {
|
|||
|
6 do_something(p->a, p->b);
|
|||
|
7 rcu_read_unlock();
|
|||
|
8 return true;
|
|||
|
9 }
|
|||
|
10 rcu_read_unlock();
|
|||
|
11 return false;
|
|||
|
12 }
|
|||
|
|
|||
|
However, this temptation must be resisted because there are a
|
|||
|
surprisingly large number of ways that the compiler (or weak ordering
|
|||
|
CPUs like the DEC Alpha) can trip this code up. For but one example, if
|
|||
|
the compiler were short of registers, it might choose to refetch from
|
|||
|
``gp`` rather than keeping a separate copy in ``p`` as follows:
|
|||
|
|
|||
|
::
|
|||
|
|
|||
|
1 bool do_something_gp_buggy_optimized(void)
|
|||
|
2 {
|
|||
|
3 rcu_read_lock();
|
|||
|
4 if (gp) { /* OPTIMIZATIONS GALORE!!! */
|
|||
|
5 do_something(gp->a, gp->b);
|
|||
|
6 rcu_read_unlock();
|
|||
|
7 return true;
|
|||
|
8 }
|
|||
|
9 rcu_read_unlock();
|
|||
|
10 return false;
|
|||
|
11 }
|
|||
|
|
|||
|
If this function ran concurrently with a series of updates that replaced
|
|||
|
the current structure with a new one, the fetches of ``gp->a`` and
|
|||
|
``gp->b`` might well come from two different structures, which could
|
|||
|
cause serious confusion. To prevent this (and much else besides),
|
|||
|
do_something_gp() uses rcu_dereference() to fetch from ``gp``:
|
|||
|
|
|||
|
::
|
|||
|
|
|||
|
1 bool do_something_gp(void)
|
|||
|
2 {
|
|||
|
3 rcu_read_lock();
|
|||
|
4 p = rcu_dereference(gp);
|
|||
|
5 if (p) {
|
|||
|
6 do_something(p->a, p->b);
|
|||
|
7 rcu_read_unlock();
|
|||
|
8 return true;
|
|||
|
9 }
|
|||
|
10 rcu_read_unlock();
|
|||
|
11 return false;
|
|||
|
12 }
|
|||
|
|
|||
|
The rcu_dereference() uses volatile casts and (for DEC Alpha) memory
|
|||
|
barriers in the Linux kernel. Should a |high-quality implementation of
|
|||
|
C11 memory_order_consume [PDF]|_
|
|||
|
ever appear, then rcu_dereference() could be implemented as a
|
|||
|
``memory_order_consume`` load. Regardless of the exact implementation, a
|
|||
|
pointer fetched by rcu_dereference() may not be used outside of the
|
|||
|
outermost RCU read-side critical section containing that
|
|||
|
rcu_dereference(), unless protection of the corresponding data
|
|||
|
element has been passed from RCU to some other synchronization
|
|||
|
mechanism, most commonly locking or reference counting
|
|||
|
(see ../../rcuref.rst).
|
|||
|
|
|||
|
.. |high-quality implementation of C11 memory_order_consume [PDF]| replace:: high-quality implementation of C11 ``memory_order_consume`` [PDF]
|
|||
|
.. _high-quality implementation of C11 memory_order_consume [PDF]: http://www.rdrop.com/users/paulmck/RCU/consume.2015.07.13a.pdf
|
|||
|
|
|||
|
In short, updaters use rcu_assign_pointer() and readers use
|
|||
|
rcu_dereference(), and these two RCU API elements work together to
|
|||
|
ensure that readers have a consistent view of newly added data elements.
|
|||
|
|
|||
|
Of course, it is also necessary to remove elements from RCU-protected
|
|||
|
data structures, for example, using the following process:
|
|||
|
|
|||
|
#. Remove the data element from the enclosing structure.
|
|||
|
#. Wait for all pre-existing RCU read-side critical sections to complete
|
|||
|
(because only pre-existing readers can possibly have a reference to
|
|||
|
the newly removed data element).
|
|||
|
#. At this point, only the updater has a reference to the newly removed
|
|||
|
data element, so it can safely reclaim the data element, for example,
|
|||
|
by passing it to kfree().
|
|||
|
|
|||
|
This process is implemented by remove_gp_synchronous():
|
|||
|
|
|||
|
::
|
|||
|
|
|||
|
1 bool remove_gp_synchronous(void)
|
|||
|
2 {
|
|||
|
3 struct foo *p;
|
|||
|
4
|
|||
|
5 spin_lock(&gp_lock);
|
|||
|
6 p = rcu_access_pointer(gp);
|
|||
|
7 if (!p) {
|
|||
|
8 spin_unlock(&gp_lock);
|
|||
|
9 return false;
|
|||
|
10 }
|
|||
|
11 rcu_assign_pointer(gp, NULL);
|
|||
|
12 spin_unlock(&gp_lock);
|
|||
|
13 synchronize_rcu();
|
|||
|
14 kfree(p);
|
|||
|
15 return true;
|
|||
|
16 }
|
|||
|
|
|||
|
This function is straightforward, with line 13 waiting for a grace
|
|||
|
period before line 14 frees the old data element. This waiting ensures
|
|||
|
that readers will reach line 7 of do_something_gp() before the data
|
|||
|
element referenced by ``p`` is freed. The rcu_access_pointer() on
|
|||
|
line 6 is similar to rcu_dereference(), except that:
|
|||
|
|
|||
|
#. The value returned by rcu_access_pointer() cannot be
|
|||
|
dereferenced. If you want to access the value pointed to as well as
|
|||
|
the pointer itself, use rcu_dereference() instead of
|
|||
|
rcu_access_pointer().
|
|||
|
#. The call to rcu_access_pointer() need not be protected. In
|
|||
|
contrast, rcu_dereference() must either be within an RCU
|
|||
|
read-side critical section or in a code segment where the pointer
|
|||
|
cannot change, for example, in code protected by the corresponding
|
|||
|
update-side lock.
|
|||
|
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Quick Quiz**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| Without the rcu_dereference() or the rcu_access_pointer(), |
|
|||
|
| what destructive optimizations might the compiler make use of? |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Answer**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| Let's start with what happens to do_something_gp() if it fails to |
|
|||
|
| use rcu_dereference(). It could reuse a value formerly fetched |
|
|||
|
| from this same pointer. It could also fetch the pointer from ``gp`` |
|
|||
|
| in a byte-at-a-time manner, resulting in *load tearing*, in turn |
|
|||
|
| resulting a bytewise mash-up of two distinct pointer values. It might |
|
|||
|
| even use value-speculation optimizations, where it makes a wrong |
|
|||
|
| guess, but by the time it gets around to checking the value, an |
|
|||
|
| update has changed the pointer to match the wrong guess. Too bad |
|
|||
|
| about any dereferences that returned pre-initialization garbage in |
|
|||
|
| the meantime! |
|
|||
|
| For remove_gp_synchronous(), as long as all modifications to |
|
|||
|
| ``gp`` are carried out while holding ``gp_lock``, the above |
|
|||
|
| optimizations are harmless. However, ``sparse`` will complain if you |
|
|||
|
| define ``gp`` with ``__rcu`` and then access it without using either |
|
|||
|
| rcu_access_pointer() or rcu_dereference(). |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
|
|||
|
In short, RCU's publish-subscribe guarantee is provided by the
|
|||
|
combination of rcu_assign_pointer() and rcu_dereference(). This
|
|||
|
guarantee allows data elements to be safely added to RCU-protected
|
|||
|
linked data structures without disrupting RCU readers. This guarantee
|
|||
|
can be used in combination with the grace-period guarantee to also allow
|
|||
|
data elements to be removed from RCU-protected linked data structures,
|
|||
|
again without disrupting RCU readers.
|
|||
|
|
|||
|
This guarantee was only partially premeditated. DYNIX/ptx used an
|
|||
|
explicit memory barrier for publication, but had nothing resembling
|
|||
|
rcu_dereference() for subscription, nor did it have anything
|
|||
|
resembling the dependency-ordering barrier that was later subsumed
|
|||
|
into rcu_dereference() and later still into READ_ONCE(). The
|
|||
|
need for these operations made itself known quite suddenly at a
|
|||
|
late-1990s meeting with the DEC Alpha architects, back in the days when
|
|||
|
DEC was still a free-standing company. It took the Alpha architects a
|
|||
|
good hour to convince me that any sort of barrier would ever be needed,
|
|||
|
and it then took me a good *two* hours to convince them that their
|
|||
|
documentation did not make this point clear. More recent work with the C
|
|||
|
and C++ standards committees have provided much education on tricks and
|
|||
|
traps from the compiler. In short, compilers were much less tricky in
|
|||
|
the early 1990s, but in 2015, don't even think about omitting
|
|||
|
rcu_dereference()!
|
|||
|
|
|||
|
Memory-Barrier Guarantees
|
|||
|
~~~~~~~~~~~~~~~~~~~~~~~~~
|
|||
|
|
|||
|
The previous section's simple linked-data-structure scenario clearly
|
|||
|
demonstrates the need for RCU's stringent memory-ordering guarantees on
|
|||
|
systems with more than one CPU:
|
|||
|
|
|||
|
#. Each CPU that has an RCU read-side critical section that begins
|
|||
|
before synchronize_rcu() starts is guaranteed to execute a full
|
|||
|
memory barrier between the time that the RCU read-side critical
|
|||
|
section ends and the time that synchronize_rcu() returns. Without
|
|||
|
this guarantee, a pre-existing RCU read-side critical section might
|
|||
|
hold a reference to the newly removed ``struct foo`` after the
|
|||
|
kfree() on line 14 of remove_gp_synchronous().
|
|||
|
#. Each CPU that has an RCU read-side critical section that ends after
|
|||
|
synchronize_rcu() returns is guaranteed to execute a full memory
|
|||
|
barrier between the time that synchronize_rcu() begins and the
|
|||
|
time that the RCU read-side critical section begins. Without this
|
|||
|
guarantee, a later RCU read-side critical section running after the
|
|||
|
kfree() on line 14 of remove_gp_synchronous() might later run
|
|||
|
do_something_gp() and find the newly deleted ``struct foo``.
|
|||
|
#. If the task invoking synchronize_rcu() remains on a given CPU,
|
|||
|
then that CPU is guaranteed to execute a full memory barrier sometime
|
|||
|
during the execution of synchronize_rcu(). This guarantee ensures
|
|||
|
that the kfree() on line 14 of remove_gp_synchronous() really
|
|||
|
does execute after the removal on line 11.
|
|||
|
#. If the task invoking synchronize_rcu() migrates among a group of
|
|||
|
CPUs during that invocation, then each of the CPUs in that group is
|
|||
|
guaranteed to execute a full memory barrier sometime during the
|
|||
|
execution of synchronize_rcu(). This guarantee also ensures that
|
|||
|
the kfree() on line 14 of remove_gp_synchronous() really does
|
|||
|
execute after the removal on line 11, but also in the case where the
|
|||
|
thread executing the synchronize_rcu() migrates in the meantime.
|
|||
|
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Quick Quiz**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| Given that multiple CPUs can start RCU read-side critical sections at |
|
|||
|
| any time without any ordering whatsoever, how can RCU possibly tell |
|
|||
|
| whether or not a given RCU read-side critical section starts before a |
|
|||
|
| given instance of synchronize_rcu()? |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Answer**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| If RCU cannot tell whether or not a given RCU read-side critical |
|
|||
|
| section starts before a given instance of synchronize_rcu(), then |
|
|||
|
| it must assume that the RCU read-side critical section started first. |
|
|||
|
| In other words, a given instance of synchronize_rcu() can avoid |
|
|||
|
| waiting on a given RCU read-side critical section only if it can |
|
|||
|
| prove that synchronize_rcu() started first. |
|
|||
|
| A related question is “When rcu_read_lock() doesn't generate any |
|
|||
|
| code, why does it matter how it relates to a grace period?” The |
|
|||
|
| answer is that it is not the relationship of rcu_read_lock() |
|
|||
|
| itself that is important, but rather the relationship of the code |
|
|||
|
| within the enclosed RCU read-side critical section to the code |
|
|||
|
| preceding and following the grace period. If we take this viewpoint, |
|
|||
|
| then a given RCU read-side critical section begins before a given |
|
|||
|
| grace period when some access preceding the grace period observes the |
|
|||
|
| effect of some access within the critical section, in which case none |
|
|||
|
| of the accesses within the critical section may observe the effects |
|
|||
|
| of any access following the grace period. |
|
|||
|
| |
|
|||
|
| As of late 2016, mathematical models of RCU take this viewpoint, for |
|
|||
|
| example, see slides 62 and 63 of the `2016 LinuxCon |
|
|||
|
| EU <http://www2.rdrop.com/users/paulmck/scalability/paper/LinuxMM.201 |
|
|||
|
| 6.10.04c.LCE.pdf>`__ |
|
|||
|
| presentation. |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Quick Quiz**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| The first and second guarantees require unbelievably strict ordering! |
|
|||
|
| Are all these memory barriers *really* required? |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Answer**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| Yes, they really are required. To see why the first guarantee is |
|
|||
|
| required, consider the following sequence of events: |
|
|||
|
| |
|
|||
|
| #. CPU 1: rcu_read_lock() |
|
|||
|
| #. CPU 1: ``q = rcu_dereference(gp); /* Very likely to return p. */`` |
|
|||
|
| #. CPU 0: ``list_del_rcu(p);`` |
|
|||
|
| #. CPU 0: synchronize_rcu() starts. |
|
|||
|
| #. CPU 1: ``do_something_with(q->a);`` |
|
|||
|
| ``/* No smp_mb(), so might happen after kfree(). */`` |
|
|||
|
| #. CPU 1: rcu_read_unlock() |
|
|||
|
| #. CPU 0: synchronize_rcu() returns. |
|
|||
|
| #. CPU 0: ``kfree(p);`` |
|
|||
|
| |
|
|||
|
| Therefore, there absolutely must be a full memory barrier between the |
|
|||
|
| end of the RCU read-side critical section and the end of the grace |
|
|||
|
| period. |
|
|||
|
| |
|
|||
|
| The sequence of events demonstrating the necessity of the second rule |
|
|||
|
| is roughly similar: |
|
|||
|
| |
|
|||
|
| #. CPU 0: ``list_del_rcu(p);`` |
|
|||
|
| #. CPU 0: synchronize_rcu() starts. |
|
|||
|
| #. CPU 1: rcu_read_lock() |
|
|||
|
| #. CPU 1: ``q = rcu_dereference(gp);`` |
|
|||
|
| ``/* Might return p if no memory barrier. */`` |
|
|||
|
| #. CPU 0: synchronize_rcu() returns. |
|
|||
|
| #. CPU 0: ``kfree(p);`` |
|
|||
|
| #. CPU 1: ``do_something_with(q->a); /* Boom!!! */`` |
|
|||
|
| #. CPU 1: rcu_read_unlock() |
|
|||
|
| |
|
|||
|
| And similarly, without a memory barrier between the beginning of the |
|
|||
|
| grace period and the beginning of the RCU read-side critical section, |
|
|||
|
| CPU 1 might end up accessing the freelist. |
|
|||
|
| |
|
|||
|
| The “as if” rule of course applies, so that any implementation that |
|
|||
|
| acts as if the appropriate memory barriers were in place is a correct |
|
|||
|
| implementation. That said, it is much easier to fool yourself into |
|
|||
|
| believing that you have adhered to the as-if rule than it is to |
|
|||
|
| actually adhere to it! |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Quick Quiz**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| You claim that rcu_read_lock() and rcu_read_unlock() generate |
|
|||
|
| absolutely no code in some kernel builds. This means that the |
|
|||
|
| compiler might arbitrarily rearrange consecutive RCU read-side |
|
|||
|
| critical sections. Given such rearrangement, if a given RCU read-side |
|
|||
|
| critical section is done, how can you be sure that all prior RCU |
|
|||
|
| read-side critical sections are done? Won't the compiler |
|
|||
|
| rearrangements make that impossible to determine? |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Answer**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| In cases where rcu_read_lock() and rcu_read_unlock() generate |
|
|||
|
| absolutely no code, RCU infers quiescent states only at special |
|
|||
|
| locations, for example, within the scheduler. Because calls to |
|
|||
|
| schedule() had better prevent calling-code accesses to shared |
|
|||
|
| variables from being rearranged across the call to schedule(), if |
|
|||
|
| RCU detects the end of a given RCU read-side critical section, it |
|
|||
|
| will necessarily detect the end of all prior RCU read-side critical |
|
|||
|
| sections, no matter how aggressively the compiler scrambles the code. |
|
|||
|
| Again, this all assumes that the compiler cannot scramble code across |
|
|||
|
| calls to the scheduler, out of interrupt handlers, into the idle |
|
|||
|
| loop, into user-mode code, and so on. But if your kernel build allows |
|
|||
|
| that sort of scrambling, you have broken far more than just RCU! |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
|
|||
|
Note that these memory-barrier requirements do not replace the
|
|||
|
fundamental RCU requirement that a grace period wait for all
|
|||
|
pre-existing readers. On the contrary, the memory barriers called out in
|
|||
|
this section must operate in such a way as to *enforce* this fundamental
|
|||
|
requirement. Of course, different implementations enforce this
|
|||
|
requirement in different ways, but enforce it they must.
|
|||
|
|
|||
|
RCU Primitives Guaranteed to Execute Unconditionally
|
|||
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|||
|
|
|||
|
The common-case RCU primitives are unconditional. They are invoked, they
|
|||
|
do their job, and they return, with no possibility of error, and no need
|
|||
|
to retry. This is a key RCU design philosophy.
|
|||
|
|
|||
|
However, this philosophy is pragmatic rather than pigheaded. If someone
|
|||
|
comes up with a good justification for a particular conditional RCU
|
|||
|
primitive, it might well be implemented and added. After all, this
|
|||
|
guarantee was reverse-engineered, not premeditated. The unconditional
|
|||
|
nature of the RCU primitives was initially an accident of
|
|||
|
implementation, and later experience with synchronization primitives
|
|||
|
with conditional primitives caused me to elevate this accident to a
|
|||
|
guarantee. Therefore, the justification for adding a conditional
|
|||
|
primitive to RCU would need to be based on detailed and compelling use
|
|||
|
cases.
|
|||
|
|
|||
|
Guaranteed Read-to-Write Upgrade
|
|||
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|||
|
|
|||
|
As far as RCU is concerned, it is always possible to carry out an update
|
|||
|
within an RCU read-side critical section. For example, that RCU
|
|||
|
read-side critical section might search for a given data element, and
|
|||
|
then might acquire the update-side spinlock in order to update that
|
|||
|
element, all while remaining in that RCU read-side critical section. Of
|
|||
|
course, it is necessary to exit the RCU read-side critical section
|
|||
|
before invoking synchronize_rcu(), however, this inconvenience can
|
|||
|
be avoided through use of the call_rcu() and kfree_rcu() API
|
|||
|
members described later in this document.
|
|||
|
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Quick Quiz**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| But how does the upgrade-to-write operation exclude other readers? |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Answer**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| It doesn't, just like normal RCU updates, which also do not exclude |
|
|||
|
| RCU readers. |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
|
|||
|
This guarantee allows lookup code to be shared between read-side and
|
|||
|
update-side code, and was premeditated, appearing in the earliest
|
|||
|
DYNIX/ptx RCU documentation.
|
|||
|
|
|||
|
Fundamental Non-Requirements
|
|||
|
----------------------------
|
|||
|
|
|||
|
RCU provides extremely lightweight readers, and its read-side
|
|||
|
guarantees, though quite useful, are correspondingly lightweight. It is
|
|||
|
therefore all too easy to assume that RCU is guaranteeing more than it
|
|||
|
really is. Of course, the list of things that RCU does not guarantee is
|
|||
|
infinitely long, however, the following sections list a few
|
|||
|
non-guarantees that have caused confusion. Except where otherwise noted,
|
|||
|
these non-guarantees were premeditated.
|
|||
|
|
|||
|
#. `Readers Impose Minimal Ordering`_
|
|||
|
#. `Readers Do Not Exclude Updaters`_
|
|||
|
#. `Updaters Only Wait For Old Readers`_
|
|||
|
#. `Grace Periods Don't Partition Read-Side Critical Sections`_
|
|||
|
#. `Read-Side Critical Sections Don't Partition Grace Periods`_
|
|||
|
|
|||
|
Readers Impose Minimal Ordering
|
|||
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|||
|
|
|||
|
Reader-side markers such as rcu_read_lock() and
|
|||
|
rcu_read_unlock() provide absolutely no ordering guarantees except
|
|||
|
through their interaction with the grace-period APIs such as
|
|||
|
synchronize_rcu(). To see this, consider the following pair of
|
|||
|
threads:
|
|||
|
|
|||
|
::
|
|||
|
|
|||
|
1 void thread0(void)
|
|||
|
2 {
|
|||
|
3 rcu_read_lock();
|
|||
|
4 WRITE_ONCE(x, 1);
|
|||
|
5 rcu_read_unlock();
|
|||
|
6 rcu_read_lock();
|
|||
|
7 WRITE_ONCE(y, 1);
|
|||
|
8 rcu_read_unlock();
|
|||
|
9 }
|
|||
|
10
|
|||
|
11 void thread1(void)
|
|||
|
12 {
|
|||
|
13 rcu_read_lock();
|
|||
|
14 r1 = READ_ONCE(y);
|
|||
|
15 rcu_read_unlock();
|
|||
|
16 rcu_read_lock();
|
|||
|
17 r2 = READ_ONCE(x);
|
|||
|
18 rcu_read_unlock();
|
|||
|
19 }
|
|||
|
|
|||
|
After thread0() and thread1() execute concurrently, it is quite
|
|||
|
possible to have
|
|||
|
|
|||
|
::
|
|||
|
|
|||
|
(r1 == 1 && r2 == 0)
|
|||
|
|
|||
|
(that is, ``y`` appears to have been assigned before ``x``), which would
|
|||
|
not be possible if rcu_read_lock() and rcu_read_unlock() had
|
|||
|
much in the way of ordering properties. But they do not, so the CPU is
|
|||
|
within its rights to do significant reordering. This is by design: Any
|
|||
|
significant ordering constraints would slow down these fast-path APIs.
|
|||
|
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Quick Quiz**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| Can't the compiler also reorder this code? |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Answer**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| No, the volatile casts in READ_ONCE() and WRITE_ONCE() |
|
|||
|
| prevent the compiler from reordering in this particular case. |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
|
|||
|
Readers Do Not Exclude Updaters
|
|||
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|||
|
|
|||
|
Neither rcu_read_lock() nor rcu_read_unlock() exclude updates.
|
|||
|
All they do is to prevent grace periods from ending. The following
|
|||
|
example illustrates this:
|
|||
|
|
|||
|
::
|
|||
|
|
|||
|
1 void thread0(void)
|
|||
|
2 {
|
|||
|
3 rcu_read_lock();
|
|||
|
4 r1 = READ_ONCE(y);
|
|||
|
5 if (r1) {
|
|||
|
6 do_something_with_nonzero_x();
|
|||
|
7 r2 = READ_ONCE(x);
|
|||
|
8 WARN_ON(!r2); /* BUG!!! */
|
|||
|
9 }
|
|||
|
10 rcu_read_unlock();
|
|||
|
11 }
|
|||
|
12
|
|||
|
13 void thread1(void)
|
|||
|
14 {
|
|||
|
15 spin_lock(&my_lock);
|
|||
|
16 WRITE_ONCE(x, 1);
|
|||
|
17 WRITE_ONCE(y, 1);
|
|||
|
18 spin_unlock(&my_lock);
|
|||
|
19 }
|
|||
|
|
|||
|
If the thread0() function's rcu_read_lock() excluded the
|
|||
|
thread1() function's update, the WARN_ON() could never fire. But
|
|||
|
the fact is that rcu_read_lock() does not exclude much of anything
|
|||
|
aside from subsequent grace periods, of which thread1() has none, so
|
|||
|
the WARN_ON() can and does fire.
|
|||
|
|
|||
|
Updaters Only Wait For Old Readers
|
|||
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|||
|
|
|||
|
It might be tempting to assume that after synchronize_rcu()
|
|||
|
completes, there are no readers executing. This temptation must be
|
|||
|
avoided because new readers can start immediately after
|
|||
|
synchronize_rcu() starts, and synchronize_rcu() is under no
|
|||
|
obligation to wait for these new readers.
|
|||
|
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Quick Quiz**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| Suppose that synchronize_rcu() did wait until *all* readers had |
|
|||
|
| completed instead of waiting only on pre-existing readers. For how |
|
|||
|
| long would the updater be able to rely on there being no readers? |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Answer**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| For no time at all. Even if synchronize_rcu() were to wait until |
|
|||
|
| all readers had completed, a new reader might start immediately after |
|
|||
|
| synchronize_rcu() completed. Therefore, the code following |
|
|||
|
| synchronize_rcu() can *never* rely on there being no readers. |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
|
|||
|
Grace Periods Don't Partition Read-Side Critical Sections
|
|||
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|||
|
|
|||
|
It is tempting to assume that if any part of one RCU read-side critical
|
|||
|
section precedes a given grace period, and if any part of another RCU
|
|||
|
read-side critical section follows that same grace period, then all of
|
|||
|
the first RCU read-side critical section must precede all of the second.
|
|||
|
However, this just isn't the case: A single grace period does not
|
|||
|
partition the set of RCU read-side critical sections. An example of this
|
|||
|
situation can be illustrated as follows, where ``x``, ``y``, and ``z``
|
|||
|
are initially all zero:
|
|||
|
|
|||
|
::
|
|||
|
|
|||
|
1 void thread0(void)
|
|||
|
2 {
|
|||
|
3 rcu_read_lock();
|
|||
|
4 WRITE_ONCE(a, 1);
|
|||
|
5 WRITE_ONCE(b, 1);
|
|||
|
6 rcu_read_unlock();
|
|||
|
7 }
|
|||
|
8
|
|||
|
9 void thread1(void)
|
|||
|
10 {
|
|||
|
11 r1 = READ_ONCE(a);
|
|||
|
12 synchronize_rcu();
|
|||
|
13 WRITE_ONCE(c, 1);
|
|||
|
14 }
|
|||
|
15
|
|||
|
16 void thread2(void)
|
|||
|
17 {
|
|||
|
18 rcu_read_lock();
|
|||
|
19 r2 = READ_ONCE(b);
|
|||
|
20 r3 = READ_ONCE(c);
|
|||
|
21 rcu_read_unlock();
|
|||
|
22 }
|
|||
|
|
|||
|
It turns out that the outcome:
|
|||
|
|
|||
|
::
|
|||
|
|
|||
|
(r1 == 1 && r2 == 0 && r3 == 1)
|
|||
|
|
|||
|
is entirely possible. The following figure show how this can happen,
|
|||
|
with each circled ``QS`` indicating the point at which RCU recorded a
|
|||
|
*quiescent state* for each thread, that is, a state in which RCU knows
|
|||
|
that the thread cannot be in the midst of an RCU read-side critical
|
|||
|
section that started before the current grace period:
|
|||
|
|
|||
|
.. kernel-figure:: GPpartitionReaders1.svg
|
|||
|
|
|||
|
If it is necessary to partition RCU read-side critical sections in this
|
|||
|
manner, it is necessary to use two grace periods, where the first grace
|
|||
|
period is known to end before the second grace period starts:
|
|||
|
|
|||
|
::
|
|||
|
|
|||
|
1 void thread0(void)
|
|||
|
2 {
|
|||
|
3 rcu_read_lock();
|
|||
|
4 WRITE_ONCE(a, 1);
|
|||
|
5 WRITE_ONCE(b, 1);
|
|||
|
6 rcu_read_unlock();
|
|||
|
7 }
|
|||
|
8
|
|||
|
9 void thread1(void)
|
|||
|
10 {
|
|||
|
11 r1 = READ_ONCE(a);
|
|||
|
12 synchronize_rcu();
|
|||
|
13 WRITE_ONCE(c, 1);
|
|||
|
14 }
|
|||
|
15
|
|||
|
16 void thread2(void)
|
|||
|
17 {
|
|||
|
18 r2 = READ_ONCE(c);
|
|||
|
19 synchronize_rcu();
|
|||
|
20 WRITE_ONCE(d, 1);
|
|||
|
21 }
|
|||
|
22
|
|||
|
23 void thread3(void)
|
|||
|
24 {
|
|||
|
25 rcu_read_lock();
|
|||
|
26 r3 = READ_ONCE(b);
|
|||
|
27 r4 = READ_ONCE(d);
|
|||
|
28 rcu_read_unlock();
|
|||
|
29 }
|
|||
|
|
|||
|
Here, if ``(r1 == 1)``, then thread0()'s write to ``b`` must happen
|
|||
|
before the end of thread1()'s grace period. If in addition
|
|||
|
``(r4 == 1)``, then thread3()'s read from ``b`` must happen after
|
|||
|
the beginning of thread2()'s grace period. If it is also the case
|
|||
|
that ``(r2 == 1)``, then the end of thread1()'s grace period must
|
|||
|
precede the beginning of thread2()'s grace period. This mean that
|
|||
|
the two RCU read-side critical sections cannot overlap, guaranteeing
|
|||
|
that ``(r3 == 1)``. As a result, the outcome:
|
|||
|
|
|||
|
::
|
|||
|
|
|||
|
(r1 == 1 && r2 == 1 && r3 == 0 && r4 == 1)
|
|||
|
|
|||
|
cannot happen.
|
|||
|
|
|||
|
This non-requirement was also non-premeditated, but became apparent when
|
|||
|
studying RCU's interaction with memory ordering.
|
|||
|
|
|||
|
Read-Side Critical Sections Don't Partition Grace Periods
|
|||
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|||
|
|
|||
|
It is also tempting to assume that if an RCU read-side critical section
|
|||
|
happens between a pair of grace periods, then those grace periods cannot
|
|||
|
overlap. However, this temptation leads nowhere good, as can be
|
|||
|
illustrated by the following, with all variables initially zero:
|
|||
|
|
|||
|
::
|
|||
|
|
|||
|
1 void thread0(void)
|
|||
|
2 {
|
|||
|
3 rcu_read_lock();
|
|||
|
4 WRITE_ONCE(a, 1);
|
|||
|
5 WRITE_ONCE(b, 1);
|
|||
|
6 rcu_read_unlock();
|
|||
|
7 }
|
|||
|
8
|
|||
|
9 void thread1(void)
|
|||
|
10 {
|
|||
|
11 r1 = READ_ONCE(a);
|
|||
|
12 synchronize_rcu();
|
|||
|
13 WRITE_ONCE(c, 1);
|
|||
|
14 }
|
|||
|
15
|
|||
|
16 void thread2(void)
|
|||
|
17 {
|
|||
|
18 rcu_read_lock();
|
|||
|
19 WRITE_ONCE(d, 1);
|
|||
|
20 r2 = READ_ONCE(c);
|
|||
|
21 rcu_read_unlock();
|
|||
|
22 }
|
|||
|
23
|
|||
|
24 void thread3(void)
|
|||
|
25 {
|
|||
|
26 r3 = READ_ONCE(d);
|
|||
|
27 synchronize_rcu();
|
|||
|
28 WRITE_ONCE(e, 1);
|
|||
|
29 }
|
|||
|
30
|
|||
|
31 void thread4(void)
|
|||
|
32 {
|
|||
|
33 rcu_read_lock();
|
|||
|
34 r4 = READ_ONCE(b);
|
|||
|
35 r5 = READ_ONCE(e);
|
|||
|
36 rcu_read_unlock();
|
|||
|
37 }
|
|||
|
|
|||
|
In this case, the outcome:
|
|||
|
|
|||
|
::
|
|||
|
|
|||
|
(r1 == 1 && r2 == 1 && r3 == 1 && r4 == 0 && r5 == 1)
|
|||
|
|
|||
|
is entirely possible, as illustrated below:
|
|||
|
|
|||
|
.. kernel-figure:: ReadersPartitionGP1.svg
|
|||
|
|
|||
|
Again, an RCU read-side critical section can overlap almost all of a
|
|||
|
given grace period, just so long as it does not overlap the entire grace
|
|||
|
period. As a result, an RCU read-side critical section cannot partition
|
|||
|
a pair of RCU grace periods.
|
|||
|
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Quick Quiz**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| How long a sequence of grace periods, each separated by an RCU |
|
|||
|
| read-side critical section, would be required to partition the RCU |
|
|||
|
| read-side critical sections at the beginning and end of the chain? |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Answer**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| In theory, an infinite number. In practice, an unknown number that is |
|
|||
|
| sensitive to both implementation details and timing considerations. |
|
|||
|
| Therefore, even in practice, RCU users must abide by the theoretical |
|
|||
|
| rather than the practical answer. |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
|
|||
|
Parallelism Facts of Life
|
|||
|
-------------------------
|
|||
|
|
|||
|
These parallelism facts of life are by no means specific to RCU, but the
|
|||
|
RCU implementation must abide by them. They therefore bear repeating:
|
|||
|
|
|||
|
#. Any CPU or task may be delayed at any time, and any attempts to avoid
|
|||
|
these delays by disabling preemption, interrupts, or whatever are
|
|||
|
completely futile. This is most obvious in preemptible user-level
|
|||
|
environments and in virtualized environments (where a given guest
|
|||
|
OS's VCPUs can be preempted at any time by the underlying
|
|||
|
hypervisor), but can also happen in bare-metal environments due to
|
|||
|
ECC errors, NMIs, and other hardware events. Although a delay of more
|
|||
|
than about 20 seconds can result in splats, the RCU implementation is
|
|||
|
obligated to use algorithms that can tolerate extremely long delays,
|
|||
|
but where “extremely long” is not long enough to allow wrap-around
|
|||
|
when incrementing a 64-bit counter.
|
|||
|
#. Both the compiler and the CPU can reorder memory accesses. Where it
|
|||
|
matters, RCU must use compiler directives and memory-barrier
|
|||
|
instructions to preserve ordering.
|
|||
|
#. Conflicting writes to memory locations in any given cache line will
|
|||
|
result in expensive cache misses. Greater numbers of concurrent
|
|||
|
writes and more-frequent concurrent writes will result in more
|
|||
|
dramatic slowdowns. RCU is therefore obligated to use algorithms that
|
|||
|
have sufficient locality to avoid significant performance and
|
|||
|
scalability problems.
|
|||
|
#. As a rough rule of thumb, only one CPU's worth of processing may be
|
|||
|
carried out under the protection of any given exclusive lock. RCU
|
|||
|
must therefore use scalable locking designs.
|
|||
|
#. Counters are finite, especially on 32-bit systems. RCU's use of
|
|||
|
counters must therefore tolerate counter wrap, or be designed such
|
|||
|
that counter wrap would take way more time than a single system is
|
|||
|
likely to run. An uptime of ten years is quite possible, a runtime of
|
|||
|
a century much less so. As an example of the latter, RCU's
|
|||
|
dyntick-idle nesting counter allows 54 bits for interrupt nesting
|
|||
|
level (this counter is 64 bits even on a 32-bit system). Overflowing
|
|||
|
this counter requires 2\ :sup:`54` half-interrupts on a given CPU
|
|||
|
without that CPU ever going idle. If a half-interrupt happened every
|
|||
|
microsecond, it would take 570 years of runtime to overflow this
|
|||
|
counter, which is currently believed to be an acceptably long time.
|
|||
|
#. Linux systems can have thousands of CPUs running a single Linux
|
|||
|
kernel in a single shared-memory environment. RCU must therefore pay
|
|||
|
close attention to high-end scalability.
|
|||
|
|
|||
|
This last parallelism fact of life means that RCU must pay special
|
|||
|
attention to the preceding facts of life. The idea that Linux might
|
|||
|
scale to systems with thousands of CPUs would have been met with some
|
|||
|
skepticism in the 1990s, but these requirements would have otherwise
|
|||
|
have been unsurprising, even in the early 1990s.
|
|||
|
|
|||
|
Quality-of-Implementation Requirements
|
|||
|
--------------------------------------
|
|||
|
|
|||
|
These sections list quality-of-implementation requirements. Although an
|
|||
|
RCU implementation that ignores these requirements could still be used,
|
|||
|
it would likely be subject to limitations that would make it
|
|||
|
inappropriate for industrial-strength production use. Classes of
|
|||
|
quality-of-implementation requirements are as follows:
|
|||
|
|
|||
|
#. `Specialization`_
|
|||
|
#. `Performance and Scalability`_
|
|||
|
#. `Forward Progress`_
|
|||
|
#. `Composability`_
|
|||
|
#. `Corner Cases`_
|
|||
|
|
|||
|
These classes is covered in the following sections.
|
|||
|
|
|||
|
Specialization
|
|||
|
~~~~~~~~~~~~~~
|
|||
|
|
|||
|
RCU is and always has been intended primarily for read-mostly
|
|||
|
situations, which means that RCU's read-side primitives are optimized,
|
|||
|
often at the expense of its update-side primitives. Experience thus far
|
|||
|
is captured by the following list of situations:
|
|||
|
|
|||
|
#. Read-mostly data, where stale and inconsistent data is not a problem:
|
|||
|
RCU works great!
|
|||
|
#. Read-mostly data, where data must be consistent: RCU works well.
|
|||
|
#. Read-write data, where data must be consistent: RCU *might* work OK.
|
|||
|
Or not.
|
|||
|
#. Write-mostly data, where data must be consistent: RCU is very
|
|||
|
unlikely to be the right tool for the job, with the following
|
|||
|
exceptions, where RCU can provide:
|
|||
|
|
|||
|
a. Existence guarantees for update-friendly mechanisms.
|
|||
|
b. Wait-free read-side primitives for real-time use.
|
|||
|
|
|||
|
This focus on read-mostly situations means that RCU must interoperate
|
|||
|
with other synchronization primitives. For example, the add_gp() and
|
|||
|
remove_gp_synchronous() examples discussed earlier use RCU to
|
|||
|
protect readers and locking to coordinate updaters. However, the need
|
|||
|
extends much farther, requiring that a variety of synchronization
|
|||
|
primitives be legal within RCU read-side critical sections, including
|
|||
|
spinlocks, sequence locks, atomic operations, reference counters, and
|
|||
|
memory barriers.
|
|||
|
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Quick Quiz**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| What about sleeping locks? |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Answer**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| These are forbidden within Linux-kernel RCU read-side critical |
|
|||
|
| sections because it is not legal to place a quiescent state (in this |
|
|||
|
| case, voluntary context switch) within an RCU read-side critical |
|
|||
|
| section. However, sleeping locks may be used within userspace RCU |
|
|||
|
| read-side critical sections, and also within Linux-kernel sleepable |
|
|||
|
| RCU `(SRCU) <Sleepable RCU_>`__ read-side critical sections. In |
|
|||
|
| addition, the -rt patchset turns spinlocks into a sleeping locks so |
|
|||
|
| that the corresponding critical sections can be preempted, which also |
|
|||
|
| means that these sleeplockified spinlocks (but not other sleeping |
|
|||
|
| locks!) may be acquire within -rt-Linux-kernel RCU read-side critical |
|
|||
|
| sections. |
|
|||
|
| Note that it *is* legal for a normal RCU read-side critical section |
|
|||
|
| to conditionally acquire a sleeping locks (as in |
|
|||
|
| mutex_trylock()), but only as long as it does not loop |
|
|||
|
| indefinitely attempting to conditionally acquire that sleeping locks. |
|
|||
|
| The key point is that things like mutex_trylock() either return |
|
|||
|
| with the mutex held, or return an error indication if the mutex was |
|
|||
|
| not immediately available. Either way, mutex_trylock() returns |
|
|||
|
| immediately without sleeping. |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
|
|||
|
It often comes as a surprise that many algorithms do not require a
|
|||
|
consistent view of data, but many can function in that mode, with
|
|||
|
network routing being the poster child. Internet routing algorithms take
|
|||
|
significant time to propagate updates, so that by the time an update
|
|||
|
arrives at a given system, that system has been sending network traffic
|
|||
|
the wrong way for a considerable length of time. Having a few threads
|
|||
|
continue to send traffic the wrong way for a few more milliseconds is
|
|||
|
clearly not a problem: In the worst case, TCP retransmissions will
|
|||
|
eventually get the data where it needs to go. In general, when tracking
|
|||
|
the state of the universe outside of the computer, some level of
|
|||
|
inconsistency must be tolerated due to speed-of-light delays if nothing
|
|||
|
else.
|
|||
|
|
|||
|
Furthermore, uncertainty about external state is inherent in many cases.
|
|||
|
For example, a pair of veterinarians might use heartbeat to determine
|
|||
|
whether or not a given cat was alive. But how long should they wait
|
|||
|
after the last heartbeat to decide that the cat is in fact dead? Waiting
|
|||
|
less than 400 milliseconds makes no sense because this would mean that a
|
|||
|
relaxed cat would be considered to cycle between death and life more
|
|||
|
than 100 times per minute. Moreover, just as with human beings, a cat's
|
|||
|
heart might stop for some period of time, so the exact wait period is a
|
|||
|
judgment call. One of our pair of veterinarians might wait 30 seconds
|
|||
|
before pronouncing the cat dead, while the other might insist on waiting
|
|||
|
a full minute. The two veterinarians would then disagree on the state of
|
|||
|
the cat during the final 30 seconds of the minute following the last
|
|||
|
heartbeat.
|
|||
|
|
|||
|
Interestingly enough, this same situation applies to hardware. When push
|
|||
|
comes to shove, how do we tell whether or not some external server has
|
|||
|
failed? We send messages to it periodically, and declare it failed if we
|
|||
|
don't receive a response within a given period of time. Policy decisions
|
|||
|
can usually tolerate short periods of inconsistency. The policy was
|
|||
|
decided some time ago, and is only now being put into effect, so a few
|
|||
|
milliseconds of delay is normally inconsequential.
|
|||
|
|
|||
|
However, there are algorithms that absolutely must see consistent data.
|
|||
|
For example, the translation between a user-level SystemV semaphore ID
|
|||
|
to the corresponding in-kernel data structure is protected by RCU, but
|
|||
|
it is absolutely forbidden to update a semaphore that has just been
|
|||
|
removed. In the Linux kernel, this need for consistency is accommodated
|
|||
|
by acquiring spinlocks located in the in-kernel data structure from
|
|||
|
within the RCU read-side critical section, and this is indicated by the
|
|||
|
green box in the figure above. Many other techniques may be used, and
|
|||
|
are in fact used within the Linux kernel.
|
|||
|
|
|||
|
In short, RCU is not required to maintain consistency, and other
|
|||
|
mechanisms may be used in concert with RCU when consistency is required.
|
|||
|
RCU's specialization allows it to do its job extremely well, and its
|
|||
|
ability to interoperate with other synchronization mechanisms allows the
|
|||
|
right mix of synchronization tools to be used for a given job.
|
|||
|
|
|||
|
Performance and Scalability
|
|||
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|||
|
|
|||
|
Energy efficiency is a critical component of performance today, and
|
|||
|
Linux-kernel RCU implementations must therefore avoid unnecessarily
|
|||
|
awakening idle CPUs. I cannot claim that this requirement was
|
|||
|
premeditated. In fact, I learned of it during a telephone conversation
|
|||
|
in which I was given “frank and open” feedback on the importance of
|
|||
|
energy efficiency in battery-powered systems and on specific
|
|||
|
energy-efficiency shortcomings of the Linux-kernel RCU implementation.
|
|||
|
In my experience, the battery-powered embedded community will consider
|
|||
|
any unnecessary wakeups to be extremely unfriendly acts. So much so that
|
|||
|
mere Linux-kernel-mailing-list posts are insufficient to vent their ire.
|
|||
|
|
|||
|
Memory consumption is not particularly important for in most situations,
|
|||
|
and has become decreasingly so as memory sizes have expanded and memory
|
|||
|
costs have plummeted. However, as I learned from Matt Mackall's
|
|||
|
`bloatwatch <http://elinux.org/Linux_Tiny-FAQ>`__ efforts, memory
|
|||
|
footprint is critically important on single-CPU systems with
|
|||
|
non-preemptible (``CONFIG_PREEMPTION=n``) kernels, and thus `tiny
|
|||
|
RCU <https://lore.kernel.org/r/20090113221724.GA15307@linux.vnet.ibm.com>`__
|
|||
|
was born. Josh Triplett has since taken over the small-memory banner
|
|||
|
with his `Linux kernel tinification <https://tiny.wiki.kernel.org/>`__
|
|||
|
project, which resulted in `SRCU <Sleepable RCU_>`__ becoming optional
|
|||
|
for those kernels not needing it.
|
|||
|
|
|||
|
The remaining performance requirements are, for the most part,
|
|||
|
unsurprising. For example, in keeping with RCU's read-side
|
|||
|
specialization, rcu_dereference() should have negligible overhead
|
|||
|
(for example, suppression of a few minor compiler optimizations).
|
|||
|
Similarly, in non-preemptible environments, rcu_read_lock() and
|
|||
|
rcu_read_unlock() should have exactly zero overhead.
|
|||
|
|
|||
|
In preemptible environments, in the case where the RCU read-side
|
|||
|
critical section was not preempted (as will be the case for the
|
|||
|
highest-priority real-time process), rcu_read_lock() and
|
|||
|
rcu_read_unlock() should have minimal overhead. In particular, they
|
|||
|
should not contain atomic read-modify-write operations, memory-barrier
|
|||
|
instructions, preemption disabling, interrupt disabling, or backwards
|
|||
|
branches. However, in the case where the RCU read-side critical section
|
|||
|
was preempted, rcu_read_unlock() may acquire spinlocks and disable
|
|||
|
interrupts. This is why it is better to nest an RCU read-side critical
|
|||
|
section within a preempt-disable region than vice versa, at least in
|
|||
|
cases where that critical section is short enough to avoid unduly
|
|||
|
degrading real-time latencies.
|
|||
|
|
|||
|
The synchronize_rcu() grace-period-wait primitive is optimized for
|
|||
|
throughput. It may therefore incur several milliseconds of latency in
|
|||
|
addition to the duration of the longest RCU read-side critical section.
|
|||
|
On the other hand, multiple concurrent invocations of
|
|||
|
synchronize_rcu() are required to use batching optimizations so that
|
|||
|
they can be satisfied by a single underlying grace-period-wait
|
|||
|
operation. For example, in the Linux kernel, it is not unusual for a
|
|||
|
single grace-period-wait operation to serve more than `1,000 separate
|
|||
|
invocations <https://www.usenix.org/conference/2004-usenix-annual-technical-conference/making-rcu-safe-deep-sub-millisecond-response>`__
|
|||
|
of synchronize_rcu(), thus amortizing the per-invocation overhead
|
|||
|
down to nearly zero. However, the grace-period optimization is also
|
|||
|
required to avoid measurable degradation of real-time scheduling and
|
|||
|
interrupt latencies.
|
|||
|
|
|||
|
In some cases, the multi-millisecond synchronize_rcu() latencies are
|
|||
|
unacceptable. In these cases, synchronize_rcu_expedited() may be
|
|||
|
used instead, reducing the grace-period latency down to a few tens of
|
|||
|
microseconds on small systems, at least in cases where the RCU read-side
|
|||
|
critical sections are short. There are currently no special latency
|
|||
|
requirements for synchronize_rcu_expedited() on large systems, but,
|
|||
|
consistent with the empirical nature of the RCU specification, that is
|
|||
|
subject to change. However, there most definitely are scalability
|
|||
|
requirements: A storm of synchronize_rcu_expedited() invocations on
|
|||
|
4096 CPUs should at least make reasonable forward progress. In return
|
|||
|
for its shorter latencies, synchronize_rcu_expedited() is permitted
|
|||
|
to impose modest degradation of real-time latency on non-idle online
|
|||
|
CPUs. Here, “modest” means roughly the same latency degradation as a
|
|||
|
scheduling-clock interrupt.
|
|||
|
|
|||
|
There are a number of situations where even
|
|||
|
synchronize_rcu_expedited()'s reduced grace-period latency is
|
|||
|
unacceptable. In these situations, the asynchronous call_rcu() can
|
|||
|
be used in place of synchronize_rcu() as follows:
|
|||
|
|
|||
|
::
|
|||
|
|
|||
|
1 struct foo {
|
|||
|
2 int a;
|
|||
|
3 int b;
|
|||
|
4 struct rcu_head rh;
|
|||
|
5 };
|
|||
|
6
|
|||
|
7 static void remove_gp_cb(struct rcu_head *rhp)
|
|||
|
8 {
|
|||
|
9 struct foo *p = container_of(rhp, struct foo, rh);
|
|||
|
10
|
|||
|
11 kfree(p);
|
|||
|
12 }
|
|||
|
13
|
|||
|
14 bool remove_gp_asynchronous(void)
|
|||
|
15 {
|
|||
|
16 struct foo *p;
|
|||
|
17
|
|||
|
18 spin_lock(&gp_lock);
|
|||
|
19 p = rcu_access_pointer(gp);
|
|||
|
20 if (!p) {
|
|||
|
21 spin_unlock(&gp_lock);
|
|||
|
22 return false;
|
|||
|
23 }
|
|||
|
24 rcu_assign_pointer(gp, NULL);
|
|||
|
25 call_rcu(&p->rh, remove_gp_cb);
|
|||
|
26 spin_unlock(&gp_lock);
|
|||
|
27 return true;
|
|||
|
28 }
|
|||
|
|
|||
|
A definition of ``struct foo`` is finally needed, and appears on
|
|||
|
lines 1-5. The function remove_gp_cb() is passed to call_rcu()
|
|||
|
on line 25, and will be invoked after the end of a subsequent grace
|
|||
|
period. This gets the same effect as remove_gp_synchronous(), but
|
|||
|
without forcing the updater to wait for a grace period to elapse. The
|
|||
|
call_rcu() function may be used in a number of situations where
|
|||
|
neither synchronize_rcu() nor synchronize_rcu_expedited() would
|
|||
|
be legal, including within preempt-disable code, local_bh_disable()
|
|||
|
code, interrupt-disable code, and interrupt handlers. However, even
|
|||
|
call_rcu() is illegal within NMI handlers and from idle and offline
|
|||
|
CPUs. The callback function (remove_gp_cb() in this case) will be
|
|||
|
executed within softirq (software interrupt) environment within the
|
|||
|
Linux kernel, either within a real softirq handler or under the
|
|||
|
protection of local_bh_disable(). In both the Linux kernel and in
|
|||
|
userspace, it is bad practice to write an RCU callback function that
|
|||
|
takes too long. Long-running operations should be relegated to separate
|
|||
|
threads or (in the Linux kernel) workqueues.
|
|||
|
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Quick Quiz**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| Why does line 19 use rcu_access_pointer()? After all, |
|
|||
|
| call_rcu() on line 25 stores into the structure, which would |
|
|||
|
| interact badly with concurrent insertions. Doesn't this mean that |
|
|||
|
| rcu_dereference() is required? |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Answer**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| Presumably the ``->gp_lock`` acquired on line 18 excludes any |
|
|||
|
| changes, including any insertions that rcu_dereference() would |
|
|||
|
| protect against. Therefore, any insertions will be delayed until |
|
|||
|
| after ``->gp_lock`` is released on line 25, which in turn means that |
|
|||
|
| rcu_access_pointer() suffices. |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
|
|||
|
However, all that remove_gp_cb() is doing is invoking kfree() on
|
|||
|
the data element. This is a common idiom, and is supported by
|
|||
|
kfree_rcu(), which allows “fire and forget” operation as shown
|
|||
|
below:
|
|||
|
|
|||
|
::
|
|||
|
|
|||
|
1 struct foo {
|
|||
|
2 int a;
|
|||
|
3 int b;
|
|||
|
4 struct rcu_head rh;
|
|||
|
5 };
|
|||
|
6
|
|||
|
7 bool remove_gp_faf(void)
|
|||
|
8 {
|
|||
|
9 struct foo *p;
|
|||
|
10
|
|||
|
11 spin_lock(&gp_lock);
|
|||
|
12 p = rcu_dereference(gp);
|
|||
|
13 if (!p) {
|
|||
|
14 spin_unlock(&gp_lock);
|
|||
|
15 return false;
|
|||
|
16 }
|
|||
|
17 rcu_assign_pointer(gp, NULL);
|
|||
|
18 kfree_rcu(p, rh);
|
|||
|
19 spin_unlock(&gp_lock);
|
|||
|
20 return true;
|
|||
|
21 }
|
|||
|
|
|||
|
Note that remove_gp_faf() simply invokes kfree_rcu() and
|
|||
|
proceeds, without any need to pay any further attention to the
|
|||
|
subsequent grace period and kfree(). It is permissible to invoke
|
|||
|
kfree_rcu() from the same environments as for call_rcu().
|
|||
|
Interestingly enough, DYNIX/ptx had the equivalents of call_rcu()
|
|||
|
and kfree_rcu(), but not synchronize_rcu(). This was due to the
|
|||
|
fact that RCU was not heavily used within DYNIX/ptx, so the very few
|
|||
|
places that needed something like synchronize_rcu() simply
|
|||
|
open-coded it.
|
|||
|
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Quick Quiz**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| Earlier it was claimed that call_rcu() and kfree_rcu() |
|
|||
|
| allowed updaters to avoid being blocked by readers. But how can that |
|
|||
|
| be correct, given that the invocation of the callback and the freeing |
|
|||
|
| of the memory (respectively) must still wait for a grace period to |
|
|||
|
| elapse? |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Answer**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| We could define things this way, but keep in mind that this sort of |
|
|||
|
| definition would say that updates in garbage-collected languages |
|
|||
|
| cannot complete until the next time the garbage collector runs, which |
|
|||
|
| does not seem at all reasonable. The key point is that in most cases, |
|
|||
|
| an updater using either call_rcu() or kfree_rcu() can proceed |
|
|||
|
| to the next update as soon as it has invoked call_rcu() or |
|
|||
|
| kfree_rcu(), without having to wait for a subsequent grace |
|
|||
|
| period. |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
|
|||
|
But what if the updater must wait for the completion of code to be
|
|||
|
executed after the end of the grace period, but has other tasks that can
|
|||
|
be carried out in the meantime? The polling-style
|
|||
|
get_state_synchronize_rcu() and cond_synchronize_rcu() functions
|
|||
|
may be used for this purpose, as shown below:
|
|||
|
|
|||
|
::
|
|||
|
|
|||
|
1 bool remove_gp_poll(void)
|
|||
|
2 {
|
|||
|
3 struct foo *p;
|
|||
|
4 unsigned long s;
|
|||
|
5
|
|||
|
6 spin_lock(&gp_lock);
|
|||
|
7 p = rcu_access_pointer(gp);
|
|||
|
8 if (!p) {
|
|||
|
9 spin_unlock(&gp_lock);
|
|||
|
10 return false;
|
|||
|
11 }
|
|||
|
12 rcu_assign_pointer(gp, NULL);
|
|||
|
13 spin_unlock(&gp_lock);
|
|||
|
14 s = get_state_synchronize_rcu();
|
|||
|
15 do_something_while_waiting();
|
|||
|
16 cond_synchronize_rcu(s);
|
|||
|
17 kfree(p);
|
|||
|
18 return true;
|
|||
|
19 }
|
|||
|
|
|||
|
On line 14, get_state_synchronize_rcu() obtains a “cookie” from RCU,
|
|||
|
then line 15 carries out other tasks, and finally, line 16 returns
|
|||
|
immediately if a grace period has elapsed in the meantime, but otherwise
|
|||
|
waits as required. The need for ``get_state_synchronize_rcu`` and
|
|||
|
cond_synchronize_rcu() has appeared quite recently, so it is too
|
|||
|
early to tell whether they will stand the test of time.
|
|||
|
|
|||
|
RCU thus provides a range of tools to allow updaters to strike the
|
|||
|
required tradeoff between latency, flexibility and CPU overhead.
|
|||
|
|
|||
|
Forward Progress
|
|||
|
~~~~~~~~~~~~~~~~
|
|||
|
|
|||
|
In theory, delaying grace-period completion and callback invocation is
|
|||
|
harmless. In practice, not only are memory sizes finite but also
|
|||
|
callbacks sometimes do wakeups, and sufficiently deferred wakeups can be
|
|||
|
difficult to distinguish from system hangs. Therefore, RCU must provide
|
|||
|
a number of mechanisms to promote forward progress.
|
|||
|
|
|||
|
These mechanisms are not foolproof, nor can they be. For one simple
|
|||
|
example, an infinite loop in an RCU read-side critical section must by
|
|||
|
definition prevent later grace periods from ever completing. For a more
|
|||
|
involved example, consider a 64-CPU system built with
|
|||
|
``CONFIG_RCU_NOCB_CPU=y`` and booted with ``rcu_nocbs=1-63``, where
|
|||
|
CPUs 1 through 63 spin in tight loops that invoke call_rcu(). Even
|
|||
|
if these tight loops also contain calls to cond_resched() (thus
|
|||
|
allowing grace periods to complete), CPU 0 simply will not be able to
|
|||
|
invoke callbacks as fast as the other 63 CPUs can register them, at
|
|||
|
least not until the system runs out of memory. In both of these
|
|||
|
examples, the Spiderman principle applies: With great power comes great
|
|||
|
responsibility. However, short of this level of abuse, RCU is required
|
|||
|
to ensure timely completion of grace periods and timely invocation of
|
|||
|
callbacks.
|
|||
|
|
|||
|
RCU takes the following steps to encourage timely completion of grace
|
|||
|
periods:
|
|||
|
|
|||
|
#. If a grace period fails to complete within 100 milliseconds, RCU
|
|||
|
causes future invocations of cond_resched() on the holdout CPUs
|
|||
|
to provide an RCU quiescent state. RCU also causes those CPUs'
|
|||
|
need_resched() invocations to return ``true``, but only after the
|
|||
|
corresponding CPU's next scheduling-clock.
|
|||
|
#. CPUs mentioned in the ``nohz_full`` kernel boot parameter can run
|
|||
|
indefinitely in the kernel without scheduling-clock interrupts, which
|
|||
|
defeats the above need_resched() strategem. RCU will therefore
|
|||
|
invoke resched_cpu() on any ``nohz_full`` CPUs still holding out
|
|||
|
after 109 milliseconds.
|
|||
|
#. In kernels built with ``CONFIG_RCU_BOOST=y``, if a given task that
|
|||
|
has been preempted within an RCU read-side critical section is
|
|||
|
holding out for more than 500 milliseconds, RCU will resort to
|
|||
|
priority boosting.
|
|||
|
#. If a CPU is still holding out 10 seconds into the grace period, RCU
|
|||
|
will invoke resched_cpu() on it regardless of its ``nohz_full``
|
|||
|
state.
|
|||
|
|
|||
|
The above values are defaults for systems running with ``HZ=1000``. They
|
|||
|
will vary as the value of ``HZ`` varies, and can also be changed using
|
|||
|
the relevant Kconfig options and kernel boot parameters. RCU currently
|
|||
|
does not do much sanity checking of these parameters, so please use
|
|||
|
caution when changing them. Note that these forward-progress measures
|
|||
|
are provided only for RCU, not for `SRCU <Sleepable RCU_>`__ or `Tasks
|
|||
|
RCU`_.
|
|||
|
|
|||
|
RCU takes the following steps in call_rcu() to encourage timely
|
|||
|
invocation of callbacks when any given non-\ ``rcu_nocbs`` CPU has
|
|||
|
10,000 callbacks, or has 10,000 more callbacks than it had the last time
|
|||
|
encouragement was provided:
|
|||
|
|
|||
|
#. Starts a grace period, if one is not already in progress.
|
|||
|
#. Forces immediate checking for quiescent states, rather than waiting
|
|||
|
for three milliseconds to have elapsed since the beginning of the
|
|||
|
grace period.
|
|||
|
#. Immediately tags the CPU's callbacks with their grace period
|
|||
|
completion numbers, rather than waiting for the ``RCU_SOFTIRQ``
|
|||
|
handler to get around to it.
|
|||
|
#. Lifts callback-execution batch limits, which speeds up callback
|
|||
|
invocation at the expense of degrading realtime response.
|
|||
|
|
|||
|
Again, these are default values when running at ``HZ=1000``, and can be
|
|||
|
overridden. Again, these forward-progress measures are provided only for
|
|||
|
RCU, not for `SRCU <Sleepable RCU_>`__ or `Tasks
|
|||
|
RCU`_. Even for RCU, callback-invocation forward
|
|||
|
progress for ``rcu_nocbs`` CPUs is much less well-developed, in part
|
|||
|
because workloads benefiting from ``rcu_nocbs`` CPUs tend to invoke
|
|||
|
call_rcu() relatively infrequently. If workloads emerge that need
|
|||
|
both ``rcu_nocbs`` CPUs and high call_rcu() invocation rates, then
|
|||
|
additional forward-progress work will be required.
|
|||
|
|
|||
|
Composability
|
|||
|
~~~~~~~~~~~~~
|
|||
|
|
|||
|
Composability has received much attention in recent years, perhaps in
|
|||
|
part due to the collision of multicore hardware with object-oriented
|
|||
|
techniques designed in single-threaded environments for single-threaded
|
|||
|
use. And in theory, RCU read-side critical sections may be composed, and
|
|||
|
in fact may be nested arbitrarily deeply. In practice, as with all
|
|||
|
real-world implementations of composable constructs, there are
|
|||
|
limitations.
|
|||
|
|
|||
|
Implementations of RCU for which rcu_read_lock() and
|
|||
|
rcu_read_unlock() generate no code, such as Linux-kernel RCU when
|
|||
|
``CONFIG_PREEMPTION=n``, can be nested arbitrarily deeply. After all, there
|
|||
|
is no overhead. Except that if all these instances of
|
|||
|
rcu_read_lock() and rcu_read_unlock() are visible to the
|
|||
|
compiler, compilation will eventually fail due to exhausting memory,
|
|||
|
mass storage, or user patience, whichever comes first. If the nesting is
|
|||
|
not visible to the compiler, as is the case with mutually recursive
|
|||
|
functions each in its own translation unit, stack overflow will result.
|
|||
|
If the nesting takes the form of loops, perhaps in the guise of tail
|
|||
|
recursion, either the control variable will overflow or (in the Linux
|
|||
|
kernel) you will get an RCU CPU stall warning. Nevertheless, this class
|
|||
|
of RCU implementations is one of the most composable constructs in
|
|||
|
existence.
|
|||
|
|
|||
|
RCU implementations that explicitly track nesting depth are limited by
|
|||
|
the nesting-depth counter. For example, the Linux kernel's preemptible
|
|||
|
RCU limits nesting to ``INT_MAX``. This should suffice for almost all
|
|||
|
practical purposes. That said, a consecutive pair of RCU read-side
|
|||
|
critical sections between which there is an operation that waits for a
|
|||
|
grace period cannot be enclosed in another RCU read-side critical
|
|||
|
section. This is because it is not legal to wait for a grace period
|
|||
|
within an RCU read-side critical section: To do so would result either
|
|||
|
in deadlock or in RCU implicitly splitting the enclosing RCU read-side
|
|||
|
critical section, neither of which is conducive to a long-lived and
|
|||
|
prosperous kernel.
|
|||
|
|
|||
|
It is worth noting that RCU is not alone in limiting composability. For
|
|||
|
example, many transactional-memory implementations prohibit composing a
|
|||
|
pair of transactions separated by an irrevocable operation (for example,
|
|||
|
a network receive operation). For another example, lock-based critical
|
|||
|
sections can be composed surprisingly freely, but only if deadlock is
|
|||
|
avoided.
|
|||
|
|
|||
|
In short, although RCU read-side critical sections are highly
|
|||
|
composable, care is required in some situations, just as is the case for
|
|||
|
any other composable synchronization mechanism.
|
|||
|
|
|||
|
Corner Cases
|
|||
|
~~~~~~~~~~~~
|
|||
|
|
|||
|
A given RCU workload might have an endless and intense stream of RCU
|
|||
|
read-side critical sections, perhaps even so intense that there was
|
|||
|
never a point in time during which there was not at least one RCU
|
|||
|
read-side critical section in flight. RCU cannot allow this situation to
|
|||
|
block grace periods: As long as all the RCU read-side critical sections
|
|||
|
are finite, grace periods must also be finite.
|
|||
|
|
|||
|
That said, preemptible RCU implementations could potentially result in
|
|||
|
RCU read-side critical sections being preempted for long durations,
|
|||
|
which has the effect of creating a long-duration RCU read-side critical
|
|||
|
section. This situation can arise only in heavily loaded systems, but
|
|||
|
systems using real-time priorities are of course more vulnerable.
|
|||
|
Therefore, RCU priority boosting is provided to help deal with this
|
|||
|
case. That said, the exact requirements on RCU priority boosting will
|
|||
|
likely evolve as more experience accumulates.
|
|||
|
|
|||
|
Other workloads might have very high update rates. Although one can
|
|||
|
argue that such workloads should instead use something other than RCU,
|
|||
|
the fact remains that RCU must handle such workloads gracefully. This
|
|||
|
requirement is another factor driving batching of grace periods, but it
|
|||
|
is also the driving force behind the checks for large numbers of queued
|
|||
|
RCU callbacks in the call_rcu() code path. Finally, high update
|
|||
|
rates should not delay RCU read-side critical sections, although some
|
|||
|
small read-side delays can occur when using
|
|||
|
synchronize_rcu_expedited(), courtesy of this function's use of
|
|||
|
smp_call_function_single().
|
|||
|
|
|||
|
Although all three of these corner cases were understood in the early
|
|||
|
1990s, a simple user-level test consisting of ``close(open(path))`` in a
|
|||
|
tight loop in the early 2000s suddenly provided a much deeper
|
|||
|
appreciation of the high-update-rate corner case. This test also
|
|||
|
motivated addition of some RCU code to react to high update rates, for
|
|||
|
example, if a given CPU finds itself with more than 10,000 RCU callbacks
|
|||
|
queued, it will cause RCU to take evasive action by more aggressively
|
|||
|
starting grace periods and more aggressively forcing completion of
|
|||
|
grace-period processing. This evasive action causes the grace period to
|
|||
|
complete more quickly, but at the cost of restricting RCU's batching
|
|||
|
optimizations, thus increasing the CPU overhead incurred by that grace
|
|||
|
period.
|
|||
|
|
|||
|
Software-Engineering Requirements
|
|||
|
---------------------------------
|
|||
|
|
|||
|
Between Murphy's Law and “To err is human”, it is necessary to guard
|
|||
|
against mishaps and misuse:
|
|||
|
|
|||
|
#. It is all too easy to forget to use rcu_read_lock() everywhere
|
|||
|
that it is needed, so kernels built with ``CONFIG_PROVE_RCU=y`` will
|
|||
|
splat if rcu_dereference() is used outside of an RCU read-side
|
|||
|
critical section. Update-side code can use
|
|||
|
rcu_dereference_protected(), which takes a `lockdep
|
|||
|
expression <https://lwn.net/Articles/371986/>`__ to indicate what is
|
|||
|
providing the protection. If the indicated protection is not
|
|||
|
provided, a lockdep splat is emitted.
|
|||
|
Code shared between readers and updaters can use
|
|||
|
rcu_dereference_check(), which also takes a lockdep expression,
|
|||
|
and emits a lockdep splat if neither rcu_read_lock() nor the
|
|||
|
indicated protection is in place. In addition,
|
|||
|
rcu_dereference_raw() is used in those (hopefully rare) cases
|
|||
|
where the required protection cannot be easily described. Finally,
|
|||
|
rcu_read_lock_held() is provided to allow a function to verify
|
|||
|
that it has been invoked within an RCU read-side critical section. I
|
|||
|
was made aware of this set of requirements shortly after Thomas
|
|||
|
Gleixner audited a number of RCU uses.
|
|||
|
#. A given function might wish to check for RCU-related preconditions
|
|||
|
upon entry, before using any other RCU API. The
|
|||
|
rcu_lockdep_assert() does this job, asserting the expression in
|
|||
|
kernels having lockdep enabled and doing nothing otherwise.
|
|||
|
#. It is also easy to forget to use rcu_assign_pointer() and
|
|||
|
rcu_dereference(), perhaps (incorrectly) substituting a simple
|
|||
|
assignment. To catch this sort of error, a given RCU-protected
|
|||
|
pointer may be tagged with ``__rcu``, after which sparse will
|
|||
|
complain about simple-assignment accesses to that pointer. Arnd
|
|||
|
Bergmann made me aware of this requirement, and also supplied the
|
|||
|
needed `patch series <https://lwn.net/Articles/376011/>`__.
|
|||
|
#. Kernels built with ``CONFIG_DEBUG_OBJECTS_RCU_HEAD=y`` will splat if
|
|||
|
a data element is passed to call_rcu() twice in a row, without a
|
|||
|
grace period in between. (This error is similar to a double free.)
|
|||
|
The corresponding ``rcu_head`` structures that are dynamically
|
|||
|
allocated are automatically tracked, but ``rcu_head`` structures
|
|||
|
allocated on the stack must be initialized with
|
|||
|
init_rcu_head_on_stack() and cleaned up with
|
|||
|
destroy_rcu_head_on_stack(). Similarly, statically allocated
|
|||
|
non-stack ``rcu_head`` structures must be initialized with
|
|||
|
init_rcu_head() and cleaned up with destroy_rcu_head().
|
|||
|
Mathieu Desnoyers made me aware of this requirement, and also
|
|||
|
supplied the needed
|
|||
|
`patch <https://lore.kernel.org/r/20100319013024.GA28456@Krystal>`__.
|
|||
|
#. An infinite loop in an RCU read-side critical section will eventually
|
|||
|
trigger an RCU CPU stall warning splat, with the duration of
|
|||
|
“eventually” being controlled by the ``RCU_CPU_STALL_TIMEOUT``
|
|||
|
``Kconfig`` option, or, alternatively, by the
|
|||
|
``rcupdate.rcu_cpu_stall_timeout`` boot/sysfs parameter. However, RCU
|
|||
|
is not obligated to produce this splat unless there is a grace period
|
|||
|
waiting on that particular RCU read-side critical section.
|
|||
|
|
|||
|
Some extreme workloads might intentionally delay RCU grace periods,
|
|||
|
and systems running those workloads can be booted with
|
|||
|
``rcupdate.rcu_cpu_stall_suppress`` to suppress the splats. This
|
|||
|
kernel parameter may also be set via ``sysfs``. Furthermore, RCU CPU
|
|||
|
stall warnings are counter-productive during sysrq dumps and during
|
|||
|
panics. RCU therefore supplies the rcu_sysrq_start() and
|
|||
|
rcu_sysrq_end() API members to be called before and after long
|
|||
|
sysrq dumps. RCU also supplies the rcu_panic() notifier that is
|
|||
|
automatically invoked at the beginning of a panic to suppress further
|
|||
|
RCU CPU stall warnings.
|
|||
|
|
|||
|
This requirement made itself known in the early 1990s, pretty much
|
|||
|
the first time that it was necessary to debug a CPU stall. That said,
|
|||
|
the initial implementation in DYNIX/ptx was quite generic in
|
|||
|
comparison with that of Linux.
|
|||
|
|
|||
|
#. Although it would be very good to detect pointers leaking out of RCU
|
|||
|
read-side critical sections, there is currently no good way of doing
|
|||
|
this. One complication is the need to distinguish between pointers
|
|||
|
leaking and pointers that have been handed off from RCU to some other
|
|||
|
synchronization mechanism, for example, reference counting.
|
|||
|
#. In kernels built with ``CONFIG_RCU_TRACE=y``, RCU-related information
|
|||
|
is provided via event tracing.
|
|||
|
#. Open-coded use of rcu_assign_pointer() and rcu_dereference()
|
|||
|
to create typical linked data structures can be surprisingly
|
|||
|
error-prone. Therefore, RCU-protected `linked
|
|||
|
lists <https://lwn.net/Articles/609973/#RCU%20List%20APIs>`__ and,
|
|||
|
more recently, RCU-protected `hash
|
|||
|
tables <https://lwn.net/Articles/612100/>`__ are available. Many
|
|||
|
other special-purpose RCU-protected data structures are available in
|
|||
|
the Linux kernel and the userspace RCU library.
|
|||
|
#. Some linked structures are created at compile time, but still require
|
|||
|
``__rcu`` checking. The RCU_POINTER_INITIALIZER() macro serves
|
|||
|
this purpose.
|
|||
|
#. It is not necessary to use rcu_assign_pointer() when creating
|
|||
|
linked structures that are to be published via a single external
|
|||
|
pointer. The RCU_INIT_POINTER() macro is provided for this task.
|
|||
|
|
|||
|
This not a hard-and-fast list: RCU's diagnostic capabilities will
|
|||
|
continue to be guided by the number and type of usage bugs found in
|
|||
|
real-world RCU usage.
|
|||
|
|
|||
|
Linux Kernel Complications
|
|||
|
--------------------------
|
|||
|
|
|||
|
The Linux kernel provides an interesting environment for all kinds of
|
|||
|
software, including RCU. Some of the relevant points of interest are as
|
|||
|
follows:
|
|||
|
|
|||
|
#. `Configuration`_
|
|||
|
#. `Firmware Interface`_
|
|||
|
#. `Early Boot`_
|
|||
|
#. `Interrupts and NMIs`_
|
|||
|
#. `Loadable Modules`_
|
|||
|
#. `Hotplug CPU`_
|
|||
|
#. `Scheduler and RCU`_
|
|||
|
#. `Tracing and RCU`_
|
|||
|
#. `Accesses to User Memory and RCU`_
|
|||
|
#. `Energy Efficiency`_
|
|||
|
#. `Scheduling-Clock Interrupts and RCU`_
|
|||
|
#. `Memory Efficiency`_
|
|||
|
#. `Performance, Scalability, Response Time, and Reliability`_
|
|||
|
|
|||
|
This list is probably incomplete, but it does give a feel for the most
|
|||
|
notable Linux-kernel complications. Each of the following sections
|
|||
|
covers one of the above topics.
|
|||
|
|
|||
|
Configuration
|
|||
|
~~~~~~~~~~~~~
|
|||
|
|
|||
|
RCU's goal is automatic configuration, so that almost nobody needs to
|
|||
|
worry about RCU's ``Kconfig`` options. And for almost all users, RCU
|
|||
|
does in fact work well “out of the box.”
|
|||
|
|
|||
|
However, there are specialized use cases that are handled by kernel boot
|
|||
|
parameters and ``Kconfig`` options. Unfortunately, the ``Kconfig``
|
|||
|
system will explicitly ask users about new ``Kconfig`` options, which
|
|||
|
requires almost all of them be hidden behind a ``CONFIG_RCU_EXPERT``
|
|||
|
``Kconfig`` option.
|
|||
|
|
|||
|
This all should be quite obvious, but the fact remains that Linus
|
|||
|
Torvalds recently had to
|
|||
|
`remind <https://lore.kernel.org/r/CA+55aFy4wcCwaL4okTs8wXhGZ5h-ibecy_Meg9C4MNQrUnwMcg@mail.gmail.com>`__
|
|||
|
me of this requirement.
|
|||
|
|
|||
|
Firmware Interface
|
|||
|
~~~~~~~~~~~~~~~~~~
|
|||
|
|
|||
|
In many cases, kernel obtains information about the system from the
|
|||
|
firmware, and sometimes things are lost in translation. Or the
|
|||
|
translation is accurate, but the original message is bogus.
|
|||
|
|
|||
|
For example, some systems' firmware overreports the number of CPUs,
|
|||
|
sometimes by a large factor. If RCU naively believed the firmware, as it
|
|||
|
used to do, it would create too many per-CPU kthreads. Although the
|
|||
|
resulting system will still run correctly, the extra kthreads needlessly
|
|||
|
consume memory and can cause confusion when they show up in ``ps``
|
|||
|
listings.
|
|||
|
|
|||
|
RCU must therefore wait for a given CPU to actually come online before
|
|||
|
it can allow itself to believe that the CPU actually exists. The
|
|||
|
resulting “ghost CPUs” (which are never going to come online) cause a
|
|||
|
number of `interesting
|
|||
|
complications <https://paulmck.livejournal.com/37494.html>`__.
|
|||
|
|
|||
|
Early Boot
|
|||
|
~~~~~~~~~~
|
|||
|
|
|||
|
The Linux kernel's boot sequence is an interesting process, and RCU is
|
|||
|
used early, even before rcu_init() is invoked. In fact, a number of
|
|||
|
RCU's primitives can be used as soon as the initial task's
|
|||
|
``task_struct`` is available and the boot CPU's per-CPU variables are
|
|||
|
set up. The read-side primitives (rcu_read_lock(),
|
|||
|
rcu_read_unlock(), rcu_dereference(), and
|
|||
|
rcu_access_pointer()) will operate normally very early on, as will
|
|||
|
rcu_assign_pointer().
|
|||
|
|
|||
|
Although call_rcu() may be invoked at any time during boot,
|
|||
|
callbacks are not guaranteed to be invoked until after all of RCU's
|
|||
|
kthreads have been spawned, which occurs at early_initcall() time.
|
|||
|
This delay in callback invocation is due to the fact that RCU does not
|
|||
|
invoke callbacks until it is fully initialized, and this full
|
|||
|
initialization cannot occur until after the scheduler has initialized
|
|||
|
itself to the point where RCU can spawn and run its kthreads. In theory,
|
|||
|
it would be possible to invoke callbacks earlier, however, this is not a
|
|||
|
panacea because there would be severe restrictions on what operations
|
|||
|
those callbacks could invoke.
|
|||
|
|
|||
|
Perhaps surprisingly, synchronize_rcu() and
|
|||
|
synchronize_rcu_expedited(), will operate normally during very early
|
|||
|
boot, the reason being that there is only one CPU and preemption is
|
|||
|
disabled. This means that the call synchronize_rcu() (or friends)
|
|||
|
itself is a quiescent state and thus a grace period, so the early-boot
|
|||
|
implementation can be a no-op.
|
|||
|
|
|||
|
However, once the scheduler has spawned its first kthread, this early
|
|||
|
boot trick fails for synchronize_rcu() (as well as for
|
|||
|
synchronize_rcu_expedited()) in ``CONFIG_PREEMPTION=y`` kernels. The
|
|||
|
reason is that an RCU read-side critical section might be preempted,
|
|||
|
which means that a subsequent synchronize_rcu() really does have to
|
|||
|
wait for something, as opposed to simply returning immediately.
|
|||
|
Unfortunately, synchronize_rcu() can't do this until all of its
|
|||
|
kthreads are spawned, which doesn't happen until some time during
|
|||
|
early_initcalls() time. But this is no excuse: RCU is nevertheless
|
|||
|
required to correctly handle synchronous grace periods during this time
|
|||
|
period. Once all of its kthreads are up and running, RCU starts running
|
|||
|
normally.
|
|||
|
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Quick Quiz**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| How can RCU possibly handle grace periods before all of its kthreads |
|
|||
|
| have been spawned??? |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Answer**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| Very carefully! |
|
|||
|
| During the “dead zone” between the time that the scheduler spawns the |
|
|||
|
| first task and the time that all of RCU's kthreads have been spawned, |
|
|||
|
| all synchronous grace periods are handled by the expedited |
|
|||
|
| grace-period mechanism. At runtime, this expedited mechanism relies |
|
|||
|
| on workqueues, but during the dead zone the requesting task itself |
|
|||
|
| drives the desired expedited grace period. Because dead-zone |
|
|||
|
| execution takes place within task context, everything works. Once the |
|
|||
|
| dead zone ends, expedited grace periods go back to using workqueues, |
|
|||
|
| as is required to avoid problems that would otherwise occur when a |
|
|||
|
| user task received a POSIX signal while driving an expedited grace |
|
|||
|
| period. |
|
|||
|
| |
|
|||
|
| And yes, this does mean that it is unhelpful to send POSIX signals to |
|
|||
|
| random tasks between the time that the scheduler spawns its first |
|
|||
|
| kthread and the time that RCU's kthreads have all been spawned. If |
|
|||
|
| there ever turns out to be a good reason for sending POSIX signals |
|
|||
|
| during that time, appropriate adjustments will be made. (If it turns |
|
|||
|
| out that POSIX signals are sent during this time for no good reason, |
|
|||
|
| other adjustments will be made, appropriate or otherwise.) |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
|
|||
|
I learned of these boot-time requirements as a result of a series of
|
|||
|
system hangs.
|
|||
|
|
|||
|
Interrupts and NMIs
|
|||
|
~~~~~~~~~~~~~~~~~~~
|
|||
|
|
|||
|
The Linux kernel has interrupts, and RCU read-side critical sections are
|
|||
|
legal within interrupt handlers and within interrupt-disabled regions of
|
|||
|
code, as are invocations of call_rcu().
|
|||
|
|
|||
|
Some Linux-kernel architectures can enter an interrupt handler from
|
|||
|
non-idle process context, and then just never leave it, instead
|
|||
|
stealthily transitioning back to process context. This trick is
|
|||
|
sometimes used to invoke system calls from inside the kernel. These
|
|||
|
“half-interrupts” mean that RCU has to be very careful about how it
|
|||
|
counts interrupt nesting levels. I learned of this requirement the hard
|
|||
|
way during a rewrite of RCU's dyntick-idle code.
|
|||
|
|
|||
|
The Linux kernel has non-maskable interrupts (NMIs), and RCU read-side
|
|||
|
critical sections are legal within NMI handlers. Thankfully, RCU
|
|||
|
update-side primitives, including call_rcu(), are prohibited within
|
|||
|
NMI handlers.
|
|||
|
|
|||
|
The name notwithstanding, some Linux-kernel architectures can have
|
|||
|
nested NMIs, which RCU must handle correctly. Andy Lutomirski `surprised
|
|||
|
me <https://lore.kernel.org/r/CALCETrXLq1y7e_dKFPgou-FKHB6Pu-r8+t-6Ds+8=va7anBWDA@mail.gmail.com>`__
|
|||
|
with this requirement; he also kindly surprised me with `an
|
|||
|
algorithm <https://lore.kernel.org/r/CALCETrXSY9JpW3uE6H8WYk81sg56qasA2aqmjMPsq5dOtzso=g@mail.gmail.com>`__
|
|||
|
that meets this requirement.
|
|||
|
|
|||
|
Furthermore, NMI handlers can be interrupted by what appear to RCU to be
|
|||
|
normal interrupts. One way that this can happen is for code that
|
|||
|
directly invokes ct_irq_enter() and ct_irq_exit() to be called
|
|||
|
from an NMI handler. This astonishing fact of life prompted the current
|
|||
|
code structure, which has ct_irq_enter() invoking
|
|||
|
ct_nmi_enter() and ct_irq_exit() invoking ct_nmi_exit().
|
|||
|
And yes, I also learned of this requirement the hard way.
|
|||
|
|
|||
|
Loadable Modules
|
|||
|
~~~~~~~~~~~~~~~~
|
|||
|
|
|||
|
The Linux kernel has loadable modules, and these modules can also be
|
|||
|
unloaded. After a given module has been unloaded, any attempt to call
|
|||
|
one of its functions results in a segmentation fault. The module-unload
|
|||
|
functions must therefore cancel any delayed calls to loadable-module
|
|||
|
functions, for example, any outstanding mod_timer() must be dealt
|
|||
|
with via del_timer_sync() or similar.
|
|||
|
|
|||
|
Unfortunately, there is no way to cancel an RCU callback; once you
|
|||
|
invoke call_rcu(), the callback function is eventually going to be
|
|||
|
invoked, unless the system goes down first. Because it is normally
|
|||
|
considered socially irresponsible to crash the system in response to a
|
|||
|
module unload request, we need some other way to deal with in-flight RCU
|
|||
|
callbacks.
|
|||
|
|
|||
|
RCU therefore provides rcu_barrier(), which waits until all
|
|||
|
in-flight RCU callbacks have been invoked. If a module uses
|
|||
|
call_rcu(), its exit function should therefore prevent any future
|
|||
|
invocation of call_rcu(), then invoke rcu_barrier(). In theory,
|
|||
|
the underlying module-unload code could invoke rcu_barrier()
|
|||
|
unconditionally, but in practice this would incur unacceptable
|
|||
|
latencies.
|
|||
|
|
|||
|
Nikita Danilov noted this requirement for an analogous
|
|||
|
filesystem-unmount situation, and Dipankar Sarma incorporated
|
|||
|
rcu_barrier() into RCU. The need for rcu_barrier() for module
|
|||
|
unloading became apparent later.
|
|||
|
|
|||
|
.. important::
|
|||
|
|
|||
|
The rcu_barrier() function is not, repeat,
|
|||
|
*not*, obligated to wait for a grace period. It is instead only required
|
|||
|
to wait for RCU callbacks that have already been posted. Therefore, if
|
|||
|
there are no RCU callbacks posted anywhere in the system,
|
|||
|
rcu_barrier() is within its rights to return immediately. Even if
|
|||
|
there are callbacks posted, rcu_barrier() does not necessarily need
|
|||
|
to wait for a grace period.
|
|||
|
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Quick Quiz**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| Wait a minute! Each RCU callbacks must wait for a grace period to |
|
|||
|
| complete, and rcu_barrier() must wait for each pre-existing |
|
|||
|
| callback to be invoked. Doesn't rcu_barrier() therefore need to |
|
|||
|
| wait for a full grace period if there is even one callback posted |
|
|||
|
| anywhere in the system? |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Answer**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| Absolutely not!!! |
|
|||
|
| Yes, each RCU callbacks must wait for a grace period to complete, but |
|
|||
|
| it might well be partly (or even completely) finished waiting by the |
|
|||
|
| time rcu_barrier() is invoked. In that case, rcu_barrier() |
|
|||
|
| need only wait for the remaining portion of the grace period to |
|
|||
|
| elapse. So even if there are quite a few callbacks posted, |
|
|||
|
| rcu_barrier() might well return quite quickly. |
|
|||
|
| |
|
|||
|
| So if you need to wait for a grace period as well as for all |
|
|||
|
| pre-existing callbacks, you will need to invoke both |
|
|||
|
| synchronize_rcu() and rcu_barrier(). If latency is a concern, |
|
|||
|
| you can always use workqueues to invoke them concurrently. |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
|
|||
|
Hotplug CPU
|
|||
|
~~~~~~~~~~~
|
|||
|
|
|||
|
The Linux kernel supports CPU hotplug, which means that CPUs can come
|
|||
|
and go. It is of course illegal to use any RCU API member from an
|
|||
|
offline CPU, with the exception of `SRCU <Sleepable RCU_>`__ read-side
|
|||
|
critical sections. This requirement was present from day one in
|
|||
|
DYNIX/ptx, but on the other hand, the Linux kernel's CPU-hotplug
|
|||
|
implementation is “interesting.”
|
|||
|
|
|||
|
The Linux-kernel CPU-hotplug implementation has notifiers that are used
|
|||
|
to allow the various kernel subsystems (including RCU) to respond
|
|||
|
appropriately to a given CPU-hotplug operation. Most RCU operations may
|
|||
|
be invoked from CPU-hotplug notifiers, including even synchronous
|
|||
|
grace-period operations such as (synchronize_rcu() and
|
|||
|
synchronize_rcu_expedited()). However, these synchronous operations
|
|||
|
do block and therefore cannot be invoked from notifiers that execute via
|
|||
|
stop_machine(), specifically those between the ``CPUHP_AP_OFFLINE``
|
|||
|
and ``CPUHP_AP_ONLINE`` states.
|
|||
|
|
|||
|
In addition, all-callback-wait operations such as rcu_barrier() may
|
|||
|
not be invoked from any CPU-hotplug notifier. This restriction is due
|
|||
|
to the fact that there are phases of CPU-hotplug operations where the
|
|||
|
outgoing CPU's callbacks will not be invoked until after the CPU-hotplug
|
|||
|
operation ends, which could also result in deadlock. Furthermore,
|
|||
|
rcu_barrier() blocks CPU-hotplug operations during its execution,
|
|||
|
which results in another type of deadlock when invoked from a CPU-hotplug
|
|||
|
notifier.
|
|||
|
|
|||
|
Finally, RCU must avoid deadlocks due to interaction between hotplug,
|
|||
|
timers and grace period processing. It does so by maintaining its own set
|
|||
|
of books that duplicate the centrally maintained ``cpu_online_mask``,
|
|||
|
and also by reporting quiescent states explicitly when a CPU goes
|
|||
|
offline. This explicit reporting of quiescent states avoids any need
|
|||
|
for the force-quiescent-state loop (FQS) to report quiescent states for
|
|||
|
offline CPUs. However, as a debugging measure, the FQS loop does splat
|
|||
|
if offline CPUs block an RCU grace period for too long.
|
|||
|
|
|||
|
An offline CPU's quiescent state will be reported either:
|
|||
|
|
|||
|
1. As the CPU goes offline using RCU's hotplug notifier (rcu_report_dead()).
|
|||
|
2. When grace period initialization (rcu_gp_init()) detects a
|
|||
|
race either with CPU offlining or with a task unblocking on a leaf
|
|||
|
``rcu_node`` structure whose CPUs are all offline.
|
|||
|
|
|||
|
The CPU-online path (rcu_cpu_starting()) should never need to report
|
|||
|
a quiescent state for an offline CPU. However, as a debugging measure,
|
|||
|
it does emit a warning if a quiescent state was not already reported
|
|||
|
for that CPU.
|
|||
|
|
|||
|
During the checking/modification of RCU's hotplug bookkeeping, the
|
|||
|
corresponding CPU's leaf node lock is held. This avoids race conditions
|
|||
|
between RCU's hotplug notifier hooks, the grace period initialization
|
|||
|
code, and the FQS loop, all of which refer to or modify this bookkeeping.
|
|||
|
|
|||
|
Scheduler and RCU
|
|||
|
~~~~~~~~~~~~~~~~~
|
|||
|
|
|||
|
RCU makes use of kthreads, and it is necessary to avoid excessive CPU-time
|
|||
|
accumulation by these kthreads. This requirement was no surprise, but
|
|||
|
RCU's violation of it when running context-switch-heavy workloads when
|
|||
|
built with ``CONFIG_NO_HZ_FULL=y`` `did come as a surprise
|
|||
|
[PDF] <http://www.rdrop.com/users/paulmck/scalability/paper/BareMetal.2015.01.15b.pdf>`__.
|
|||
|
RCU has made good progress towards meeting this requirement, even for
|
|||
|
context-switch-heavy ``CONFIG_NO_HZ_FULL=y`` workloads, but there is
|
|||
|
room for further improvement.
|
|||
|
|
|||
|
There is no longer any prohibition against holding any of
|
|||
|
scheduler's runqueue or priority-inheritance spinlocks across an
|
|||
|
rcu_read_unlock(), even if interrupts and preemption were enabled
|
|||
|
somewhere within the corresponding RCU read-side critical section.
|
|||
|
Therefore, it is now perfectly legal to execute rcu_read_lock()
|
|||
|
with preemption enabled, acquire one of the scheduler locks, and hold
|
|||
|
that lock across the matching rcu_read_unlock().
|
|||
|
|
|||
|
Similarly, the RCU flavor consolidation has removed the need for negative
|
|||
|
nesting. The fact that interrupt-disabled regions of code act as RCU
|
|||
|
read-side critical sections implicitly avoids earlier issues that used
|
|||
|
to result in destructive recursion via interrupt handler's use of RCU.
|
|||
|
|
|||
|
Tracing and RCU
|
|||
|
~~~~~~~~~~~~~~~
|
|||
|
|
|||
|
It is possible to use tracing on RCU code, but tracing itself uses RCU.
|
|||
|
For this reason, rcu_dereference_raw_check() is provided for use
|
|||
|
by tracing, which avoids the destructive recursion that could otherwise
|
|||
|
ensue. This API is also used by virtualization in some architectures,
|
|||
|
where RCU readers execute in environments in which tracing cannot be
|
|||
|
used. The tracing folks both located the requirement and provided the
|
|||
|
needed fix, so this surprise requirement was relatively painless.
|
|||
|
|
|||
|
Accesses to User Memory and RCU
|
|||
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|||
|
|
|||
|
The kernel needs to access user-space memory, for example, to access data
|
|||
|
referenced by system-call parameters. The get_user() macro does this job.
|
|||
|
|
|||
|
However, user-space memory might well be paged out, which means that
|
|||
|
get_user() might well page-fault and thus block while waiting for the
|
|||
|
resulting I/O to complete. It would be a very bad thing for the compiler to
|
|||
|
reorder a get_user() invocation into an RCU read-side critical section.
|
|||
|
|
|||
|
For example, suppose that the source code looked like this:
|
|||
|
|
|||
|
::
|
|||
|
|
|||
|
1 rcu_read_lock();
|
|||
|
2 p = rcu_dereference(gp);
|
|||
|
3 v = p->value;
|
|||
|
4 rcu_read_unlock();
|
|||
|
5 get_user(user_v, user_p);
|
|||
|
6 do_something_with(v, user_v);
|
|||
|
|
|||
|
The compiler must not be permitted to transform this source code into
|
|||
|
the following:
|
|||
|
|
|||
|
::
|
|||
|
|
|||
|
1 rcu_read_lock();
|
|||
|
2 p = rcu_dereference(gp);
|
|||
|
3 get_user(user_v, user_p); // BUG: POSSIBLE PAGE FAULT!!!
|
|||
|
4 v = p->value;
|
|||
|
5 rcu_read_unlock();
|
|||
|
6 do_something_with(v, user_v);
|
|||
|
|
|||
|
If the compiler did make this transformation in a ``CONFIG_PREEMPTION=n`` kernel
|
|||
|
build, and if get_user() did page fault, the result would be a quiescent
|
|||
|
state in the middle of an RCU read-side critical section. This misplaced
|
|||
|
quiescent state could result in line 4 being a use-after-free access,
|
|||
|
which could be bad for your kernel's actuarial statistics. Similar examples
|
|||
|
can be constructed with the call to get_user() preceding the
|
|||
|
rcu_read_lock().
|
|||
|
|
|||
|
Unfortunately, get_user() doesn't have any particular ordering properties,
|
|||
|
and in some architectures the underlying ``asm`` isn't even marked
|
|||
|
``volatile``. And even if it was marked ``volatile``, the above access to
|
|||
|
``p->value`` is not volatile, so the compiler would not have any reason to keep
|
|||
|
those two accesses in order.
|
|||
|
|
|||
|
Therefore, the Linux-kernel definitions of rcu_read_lock() and
|
|||
|
rcu_read_unlock() must act as compiler barriers, at least for outermost
|
|||
|
instances of rcu_read_lock() and rcu_read_unlock() within a nested set
|
|||
|
of RCU read-side critical sections.
|
|||
|
|
|||
|
Energy Efficiency
|
|||
|
~~~~~~~~~~~~~~~~~
|
|||
|
|
|||
|
Interrupting idle CPUs is considered socially unacceptable, especially
|
|||
|
by people with battery-powered embedded systems. RCU therefore conserves
|
|||
|
energy by detecting which CPUs are idle, including tracking CPUs that
|
|||
|
have been interrupted from idle. This is a large part of the
|
|||
|
energy-efficiency requirement, so I learned of this via an irate phone
|
|||
|
call.
|
|||
|
|
|||
|
Because RCU avoids interrupting idle CPUs, it is illegal to execute an
|
|||
|
RCU read-side critical section on an idle CPU. (Kernels built with
|
|||
|
``CONFIG_PROVE_RCU=y`` will splat if you try it.) The RCU_NONIDLE()
|
|||
|
macro and ``_rcuidle`` event tracing is provided to work around this
|
|||
|
restriction. In addition, rcu_is_watching() may be used to test
|
|||
|
whether or not it is currently legal to run RCU read-side critical
|
|||
|
sections on this CPU. I learned of the need for diagnostics on the one
|
|||
|
hand and RCU_NONIDLE() on the other while inspecting idle-loop code.
|
|||
|
Steven Rostedt supplied ``_rcuidle`` event tracing, which is used quite
|
|||
|
heavily in the idle loop. However, there are some restrictions on the
|
|||
|
code placed within RCU_NONIDLE():
|
|||
|
|
|||
|
#. Blocking is prohibited. In practice, this is not a serious
|
|||
|
restriction given that idle tasks are prohibited from blocking to
|
|||
|
begin with.
|
|||
|
#. Although nesting RCU_NONIDLE() is permitted, they cannot nest
|
|||
|
indefinitely deeply. However, given that they can be nested on the
|
|||
|
order of a million deep, even on 32-bit systems, this should not be a
|
|||
|
serious restriction. This nesting limit would probably be reached
|
|||
|
long after the compiler OOMed or the stack overflowed.
|
|||
|
#. Any code path that enters RCU_NONIDLE() must sequence out of that
|
|||
|
same RCU_NONIDLE(). For example, the following is grossly
|
|||
|
illegal:
|
|||
|
|
|||
|
::
|
|||
|
|
|||
|
1 RCU_NONIDLE({
|
|||
|
2 do_something();
|
|||
|
3 goto bad_idea; /* BUG!!! */
|
|||
|
4 do_something_else();});
|
|||
|
5 bad_idea:
|
|||
|
|
|||
|
|
|||
|
It is just as illegal to transfer control into the middle of
|
|||
|
RCU_NONIDLE()'s argument. Yes, in theory, you could transfer in
|
|||
|
as long as you also transferred out, but in practice you could also
|
|||
|
expect to get sharply worded review comments.
|
|||
|
|
|||
|
It is similarly socially unacceptable to interrupt an ``nohz_full`` CPU
|
|||
|
running in userspace. RCU must therefore track ``nohz_full`` userspace
|
|||
|
execution. RCU must therefore be able to sample state at two points in
|
|||
|
time, and be able to determine whether or not some other CPU spent any
|
|||
|
time idle and/or executing in userspace.
|
|||
|
|
|||
|
These energy-efficiency requirements have proven quite difficult to
|
|||
|
understand and to meet, for example, there have been more than five
|
|||
|
clean-sheet rewrites of RCU's energy-efficiency code, the last of which
|
|||
|
was finally able to demonstrate `real energy savings running on real
|
|||
|
hardware
|
|||
|
[PDF] <http://www.rdrop.com/users/paulmck/realtime/paper/AMPenergy.2013.04.19a.pdf>`__.
|
|||
|
As noted earlier, I learned of many of these requirements via angry
|
|||
|
phone calls: Flaming me on the Linux-kernel mailing list was apparently
|
|||
|
not sufficient to fully vent their ire at RCU's energy-efficiency bugs!
|
|||
|
|
|||
|
Scheduling-Clock Interrupts and RCU
|
|||
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|||
|
|
|||
|
The kernel transitions between in-kernel non-idle execution, userspace
|
|||
|
execution, and the idle loop. Depending on kernel configuration, RCU
|
|||
|
handles these states differently:
|
|||
|
|
|||
|
+-----------------+------------------+------------------+-----------------+
|
|||
|
| ``HZ`` Kconfig | In-Kernel | Usermode | Idle |
|
|||
|
+=================+==================+==================+=================+
|
|||
|
| ``HZ_PERIODIC`` | Can rely on | Can rely on | Can rely on |
|
|||
|
| | scheduling-clock | scheduling-clock | RCU's |
|
|||
|
| | interrupt. | interrupt and | dyntick-idle |
|
|||
|
| | | its detection | detection. |
|
|||
|
| | | of interrupt | |
|
|||
|
| | | from usermode. | |
|
|||
|
+-----------------+------------------+------------------+-----------------+
|
|||
|
| ``NO_HZ_IDLE`` | Can rely on | Can rely on | Can rely on |
|
|||
|
| | scheduling-clock | scheduling-clock | RCU's |
|
|||
|
| | interrupt. | interrupt and | dyntick-idle |
|
|||
|
| | | its detection | detection. |
|
|||
|
| | | of interrupt | |
|
|||
|
| | | from usermode. | |
|
|||
|
+-----------------+------------------+------------------+-----------------+
|
|||
|
| ``NO_HZ_FULL`` | Can only | Can rely on | Can rely on |
|
|||
|
| | sometimes rely | RCU's | RCU's |
|
|||
|
| | on | dyntick-idle | dyntick-idle |
|
|||
|
| | scheduling-clock | detection. | detection. |
|
|||
|
| | interrupt. In | | |
|
|||
|
| | other cases, it | | |
|
|||
|
| | is necessary to | | |
|
|||
|
| | bound kernel | | |
|
|||
|
| | execution times | | |
|
|||
|
| | and/or use | | |
|
|||
|
| | IPIs. | | |
|
|||
|
+-----------------+------------------+------------------+-----------------+
|
|||
|
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Quick Quiz**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| Why can't ``NO_HZ_FULL`` in-kernel execution rely on the |
|
|||
|
| scheduling-clock interrupt, just like ``HZ_PERIODIC`` and |
|
|||
|
| ``NO_HZ_IDLE`` do? |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Answer**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| Because, as a performance optimization, ``NO_HZ_FULL`` does not |
|
|||
|
| necessarily re-enable the scheduling-clock interrupt on entry to each |
|
|||
|
| and every system call. |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
|
|||
|
However, RCU must be reliably informed as to whether any given CPU is
|
|||
|
currently in the idle loop, and, for ``NO_HZ_FULL``, also whether that
|
|||
|
CPU is executing in usermode, as discussed
|
|||
|
`earlier <Energy Efficiency_>`__. It also requires that the
|
|||
|
scheduling-clock interrupt be enabled when RCU needs it to be:
|
|||
|
|
|||
|
#. If a CPU is either idle or executing in usermode, and RCU believes it
|
|||
|
is non-idle, the scheduling-clock tick had better be running.
|
|||
|
Otherwise, you will get RCU CPU stall warnings. Or at best, very long
|
|||
|
(11-second) grace periods, with a pointless IPI waking the CPU from
|
|||
|
time to time.
|
|||
|
#. If a CPU is in a portion of the kernel that executes RCU read-side
|
|||
|
critical sections, and RCU believes this CPU to be idle, you will get
|
|||
|
random memory corruption. **DON'T DO THIS!!!**
|
|||
|
This is one reason to test with lockdep, which will complain about
|
|||
|
this sort of thing.
|
|||
|
#. If a CPU is in a portion of the kernel that is absolutely positively
|
|||
|
no-joking guaranteed to never execute any RCU read-side critical
|
|||
|
sections, and RCU believes this CPU to be idle, no problem. This
|
|||
|
sort of thing is used by some architectures for light-weight
|
|||
|
exception handlers, which can then avoid the overhead of
|
|||
|
ct_irq_enter() and ct_irq_exit() at exception entry and
|
|||
|
exit, respectively. Some go further and avoid the entireties of
|
|||
|
irq_enter() and irq_exit().
|
|||
|
Just make very sure you are running some of your tests with
|
|||
|
``CONFIG_PROVE_RCU=y``, just in case one of your code paths was in
|
|||
|
fact joking about not doing RCU read-side critical sections.
|
|||
|
#. If a CPU is executing in the kernel with the scheduling-clock
|
|||
|
interrupt disabled and RCU believes this CPU to be non-idle, and if
|
|||
|
the CPU goes idle (from an RCU perspective) every few jiffies, no
|
|||
|
problem. It is usually OK for there to be the occasional gap between
|
|||
|
idle periods of up to a second or so.
|
|||
|
If the gap grows too long, you get RCU CPU stall warnings.
|
|||
|
#. If a CPU is either idle or executing in usermode, and RCU believes it
|
|||
|
to be idle, of course no problem.
|
|||
|
#. If a CPU is executing in the kernel, the kernel code path is passing
|
|||
|
through quiescent states at a reasonable frequency (preferably about
|
|||
|
once per few jiffies, but the occasional excursion to a second or so
|
|||
|
is usually OK) and the scheduling-clock interrupt is enabled, of
|
|||
|
course no problem.
|
|||
|
If the gap between a successive pair of quiescent states grows too
|
|||
|
long, you get RCU CPU stall warnings.
|
|||
|
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Quick Quiz**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| But what if my driver has a hardware interrupt handler that can run |
|
|||
|
| for many seconds? I cannot invoke schedule() from an hardware |
|
|||
|
| interrupt handler, after all! |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| **Answer**: |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
| One approach is to do ``ct_irq_exit();ct_irq_enter();`` every so |
|
|||
|
| often. But given that long-running interrupt handlers can cause other |
|
|||
|
| problems, not least for response time, shouldn't you work to keep |
|
|||
|
| your interrupt handler's runtime within reasonable bounds? |
|
|||
|
+-----------------------------------------------------------------------+
|
|||
|
|
|||
|
But as long as RCU is properly informed of kernel state transitions
|
|||
|
between in-kernel execution, usermode execution, and idle, and as long
|
|||
|
as the scheduling-clock interrupt is enabled when RCU needs it to be,
|
|||
|
you can rest assured that the bugs you encounter will be in some other
|
|||
|
part of RCU or some other part of the kernel!
|
|||
|
|
|||
|
Memory Efficiency
|
|||
|
~~~~~~~~~~~~~~~~~
|
|||
|
|
|||
|
Although small-memory non-realtime systems can simply use Tiny RCU, code
|
|||
|
size is only one aspect of memory efficiency. Another aspect is the size
|
|||
|
of the ``rcu_head`` structure used by call_rcu() and
|
|||
|
kfree_rcu(). Although this structure contains nothing more than a
|
|||
|
pair of pointers, it does appear in many RCU-protected data structures,
|
|||
|
including some that are size critical. The ``page`` structure is a case
|
|||
|
in point, as evidenced by the many occurrences of the ``union`` keyword
|
|||
|
within that structure.
|
|||
|
|
|||
|
This need for memory efficiency is one reason that RCU uses hand-crafted
|
|||
|
singly linked lists to track the ``rcu_head`` structures that are
|
|||
|
waiting for a grace period to elapse. It is also the reason why
|
|||
|
``rcu_head`` structures do not contain debug information, such as fields
|
|||
|
tracking the file and line of the call_rcu() or kfree_rcu() that
|
|||
|
posted them. Although this information might appear in debug-only kernel
|
|||
|
builds at some point, in the meantime, the ``->func`` field will often
|
|||
|
provide the needed debug information.
|
|||
|
|
|||
|
However, in some cases, the need for memory efficiency leads to even
|
|||
|
more extreme measures. Returning to the ``page`` structure, the
|
|||
|
``rcu_head`` field shares storage with a great many other structures
|
|||
|
that are used at various points in the corresponding page's lifetime. In
|
|||
|
order to correctly resolve certain `race
|
|||
|
conditions <https://lore.kernel.org/r/1439976106-137226-1-git-send-email-kirill.shutemov@linux.intel.com>`__,
|
|||
|
the Linux kernel's memory-management subsystem needs a particular bit to
|
|||
|
remain zero during all phases of grace-period processing, and that bit
|
|||
|
happens to map to the bottom bit of the ``rcu_head`` structure's
|
|||
|
``->next`` field. RCU makes this guarantee as long as call_rcu() is
|
|||
|
used to post the callback, as opposed to kfree_rcu() or some future
|
|||
|
“lazy” variant of call_rcu() that might one day be created for
|
|||
|
energy-efficiency purposes.
|
|||
|
|
|||
|
That said, there are limits. RCU requires that the ``rcu_head``
|
|||
|
structure be aligned to a two-byte boundary, and passing a misaligned
|
|||
|
``rcu_head`` structure to one of the call_rcu() family of functions
|
|||
|
will result in a splat. It is therefore necessary to exercise caution
|
|||
|
when packing structures containing fields of type ``rcu_head``. Why not
|
|||
|
a four-byte or even eight-byte alignment requirement? Because the m68k
|
|||
|
architecture provides only two-byte alignment, and thus acts as
|
|||
|
alignment's least common denominator.
|
|||
|
|
|||
|
The reason for reserving the bottom bit of pointers to ``rcu_head``
|
|||
|
structures is to leave the door open to “lazy” callbacks whose
|
|||
|
invocations can safely be deferred. Deferring invocation could
|
|||
|
potentially have energy-efficiency benefits, but only if the rate of
|
|||
|
non-lazy callbacks decreases significantly for some important workload.
|
|||
|
In the meantime, reserving the bottom bit keeps this option open in case
|
|||
|
it one day becomes useful.
|
|||
|
|
|||
|
Performance, Scalability, Response Time, and Reliability
|
|||
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|||
|
|
|||
|
Expanding on the `earlier
|
|||
|
discussion <Performance and Scalability_>`__, RCU is used heavily by
|
|||
|
hot code paths in performance-critical portions of the Linux kernel's
|
|||
|
networking, security, virtualization, and scheduling code paths. RCU
|
|||
|
must therefore use efficient implementations, especially in its
|
|||
|
read-side primitives. To that end, it would be good if preemptible RCU's
|
|||
|
implementation of rcu_read_lock() could be inlined, however, doing
|
|||
|
this requires resolving ``#include`` issues with the ``task_struct``
|
|||
|
structure.
|
|||
|
|
|||
|
The Linux kernel supports hardware configurations with up to 4096 CPUs,
|
|||
|
which means that RCU must be extremely scalable. Algorithms that involve
|
|||
|
frequent acquisitions of global locks or frequent atomic operations on
|
|||
|
global variables simply cannot be tolerated within the RCU
|
|||
|
implementation. RCU therefore makes heavy use of a combining tree based
|
|||
|
on the ``rcu_node`` structure. RCU is required to tolerate all CPUs
|
|||
|
continuously invoking any combination of RCU's runtime primitives with
|
|||
|
minimal per-operation overhead. In fact, in many cases, increasing load
|
|||
|
must *decrease* the per-operation overhead, witness the batching
|
|||
|
optimizations for synchronize_rcu(), call_rcu(),
|
|||
|
synchronize_rcu_expedited(), and rcu_barrier(). As a general
|
|||
|
rule, RCU must cheerfully accept whatever the rest of the Linux kernel
|
|||
|
decides to throw at it.
|
|||
|
|
|||
|
The Linux kernel is used for real-time workloads, especially in
|
|||
|
conjunction with the `-rt
|
|||
|
patchset <https://wiki.linuxfoundation.org/realtime/>`__. The
|
|||
|
real-time-latency response requirements are such that the traditional
|
|||
|
approach of disabling preemption across RCU read-side critical sections
|
|||
|
is inappropriate. Kernels built with ``CONFIG_PREEMPTION=y`` therefore use
|
|||
|
an RCU implementation that allows RCU read-side critical sections to be
|
|||
|
preempted. This requirement made its presence known after users made it
|
|||
|
clear that an earlier `real-time
|
|||
|
patch <https://lwn.net/Articles/107930/>`__ did not meet their needs, in
|
|||
|
conjunction with some `RCU
|
|||
|
issues <https://lore.kernel.org/r/20050318002026.GA2693@us.ibm.com>`__
|
|||
|
encountered by a very early version of the -rt patchset.
|
|||
|
|
|||
|
In addition, RCU must make do with a sub-100-microsecond real-time
|
|||
|
latency budget. In fact, on smaller systems with the -rt patchset, the
|
|||
|
Linux kernel provides sub-20-microsecond real-time latencies for the
|
|||
|
whole kernel, including RCU. RCU's scalability and latency must
|
|||
|
therefore be sufficient for these sorts of configurations. To my
|
|||
|
surprise, the sub-100-microsecond real-time latency budget `applies to
|
|||
|
even the largest systems
|
|||
|
[PDF] <http://www.rdrop.com/users/paulmck/realtime/paper/bigrt.2013.01.31a.LCA.pdf>`__,
|
|||
|
up to and including systems with 4096 CPUs. This real-time requirement
|
|||
|
motivated the grace-period kthread, which also simplified handling of a
|
|||
|
number of race conditions.
|
|||
|
|
|||
|
RCU must avoid degrading real-time response for CPU-bound threads,
|
|||
|
whether executing in usermode (which is one use case for
|
|||
|
``CONFIG_NO_HZ_FULL=y``) or in the kernel. That said, CPU-bound loops in
|
|||
|
the kernel must execute cond_resched() at least once per few tens of
|
|||
|
milliseconds in order to avoid receiving an IPI from RCU.
|
|||
|
|
|||
|
Finally, RCU's status as a synchronization primitive means that any RCU
|
|||
|
failure can result in arbitrary memory corruption that can be extremely
|
|||
|
difficult to debug. This means that RCU must be extremely reliable,
|
|||
|
which in practice also means that RCU must have an aggressive
|
|||
|
stress-test suite. This stress-test suite is called ``rcutorture``.
|
|||
|
|
|||
|
Although the need for ``rcutorture`` was no surprise, the current
|
|||
|
immense popularity of the Linux kernel is posing interesting—and perhaps
|
|||
|
unprecedented—validation challenges. To see this, keep in mind that
|
|||
|
there are well over one billion instances of the Linux kernel running
|
|||
|
today, given Android smartphones, Linux-powered televisions, and
|
|||
|
servers. This number can be expected to increase sharply with the advent
|
|||
|
of the celebrated Internet of Things.
|
|||
|
|
|||
|
Suppose that RCU contains a race condition that manifests on average
|
|||
|
once per million years of runtime. This bug will be occurring about
|
|||
|
three times per *day* across the installed base. RCU could simply hide
|
|||
|
behind hardware error rates, given that no one should really expect
|
|||
|
their smartphone to last for a million years. However, anyone taking too
|
|||
|
much comfort from this thought should consider the fact that in most
|
|||
|
jurisdictions, a successful multi-year test of a given mechanism, which
|
|||
|
might include a Linux kernel, suffices for a number of types of
|
|||
|
safety-critical certifications. In fact, rumor has it that the Linux
|
|||
|
kernel is already being used in production for safety-critical
|
|||
|
applications. I don't know about you, but I would feel quite bad if a
|
|||
|
bug in RCU killed someone. Which might explain my recent focus on
|
|||
|
validation and verification.
|
|||
|
|
|||
|
Other RCU Flavors
|
|||
|
-----------------
|
|||
|
|
|||
|
One of the more surprising things about RCU is that there are now no
|
|||
|
fewer than five *flavors*, or API families. In addition, the primary
|
|||
|
flavor that has been the sole focus up to this point has two different
|
|||
|
implementations, non-preemptible and preemptible. The other four flavors
|
|||
|
are listed below, with requirements for each described in a separate
|
|||
|
section.
|
|||
|
|
|||
|
#. `Bottom-Half Flavor (Historical)`_
|
|||
|
#. `Sched Flavor (Historical)`_
|
|||
|
#. `Sleepable RCU`_
|
|||
|
#. `Tasks RCU`_
|
|||
|
|
|||
|
Bottom-Half Flavor (Historical)
|
|||
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|||
|
|
|||
|
The RCU-bh flavor of RCU has since been expressed in terms of the other
|
|||
|
RCU flavors as part of a consolidation of the three flavors into a
|
|||
|
single flavor. The read-side API remains, and continues to disable
|
|||
|
softirq and to be accounted for by lockdep. Much of the material in this
|
|||
|
section is therefore strictly historical in nature.
|
|||
|
|
|||
|
The softirq-disable (AKA “bottom-half”, hence the “_bh” abbreviations)
|
|||
|
flavor of RCU, or *RCU-bh*, was developed by Dipankar Sarma to provide a
|
|||
|
flavor of RCU that could withstand the network-based denial-of-service
|
|||
|
attacks researched by Robert Olsson. These attacks placed so much
|
|||
|
networking load on the system that some of the CPUs never exited softirq
|
|||
|
execution, which in turn prevented those CPUs from ever executing a
|
|||
|
context switch, which, in the RCU implementation of that time, prevented
|
|||
|
grace periods from ever ending. The result was an out-of-memory
|
|||
|
condition and a system hang.
|
|||
|
|
|||
|
The solution was the creation of RCU-bh, which does
|
|||
|
local_bh_disable() across its read-side critical sections, and which
|
|||
|
uses the transition from one type of softirq processing to another as a
|
|||
|
quiescent state in addition to context switch, idle, user mode, and
|
|||
|
offline. This means that RCU-bh grace periods can complete even when
|
|||
|
some of the CPUs execute in softirq indefinitely, thus allowing
|
|||
|
algorithms based on RCU-bh to withstand network-based denial-of-service
|
|||
|
attacks.
|
|||
|
|
|||
|
Because rcu_read_lock_bh() and rcu_read_unlock_bh() disable and
|
|||
|
re-enable softirq handlers, any attempt to start a softirq handlers
|
|||
|
during the RCU-bh read-side critical section will be deferred. In this
|
|||
|
case, rcu_read_unlock_bh() will invoke softirq processing, which can
|
|||
|
take considerable time. One can of course argue that this softirq
|
|||
|
overhead should be associated with the code following the RCU-bh
|
|||
|
read-side critical section rather than rcu_read_unlock_bh(), but the
|
|||
|
fact is that most profiling tools cannot be expected to make this sort
|
|||
|
of fine distinction. For example, suppose that a three-millisecond-long
|
|||
|
RCU-bh read-side critical section executes during a time of heavy
|
|||
|
networking load. There will very likely be an attempt to invoke at least
|
|||
|
one softirq handler during that three milliseconds, but any such
|
|||
|
invocation will be delayed until the time of the
|
|||
|
rcu_read_unlock_bh(). This can of course make it appear at first
|
|||
|
glance as if rcu_read_unlock_bh() was executing very slowly.
|
|||
|
|
|||
|
The `RCU-bh
|
|||
|
API <https://lwn.net/Articles/609973/#RCU%20Per-Flavor%20API%20Table>`__
|
|||
|
includes rcu_read_lock_bh(), rcu_read_unlock_bh(), rcu_dereference_bh(),
|
|||
|
rcu_dereference_bh_check(), and rcu_read_lock_bh_held(). However, the
|
|||
|
old RCU-bh update-side APIs are now gone, replaced by synchronize_rcu(),
|
|||
|
synchronize_rcu_expedited(), call_rcu(), and rcu_barrier(). In addition,
|
|||
|
anything that disables bottom halves also marks an RCU-bh read-side
|
|||
|
critical section, including local_bh_disable() and local_bh_enable(),
|
|||
|
local_irq_save() and local_irq_restore(), and so on.
|
|||
|
|
|||
|
Sched Flavor (Historical)
|
|||
|
~~~~~~~~~~~~~~~~~~~~~~~~~
|
|||
|
|
|||
|
The RCU-sched flavor of RCU has since been expressed in terms of the
|
|||
|
other RCU flavors as part of a consolidation of the three flavors into a
|
|||
|
single flavor. The read-side API remains, and continues to disable
|
|||
|
preemption and to be accounted for by lockdep. Much of the material in
|
|||
|
this section is therefore strictly historical in nature.
|
|||
|
|
|||
|
Before preemptible RCU, waiting for an RCU grace period had the side
|
|||
|
effect of also waiting for all pre-existing interrupt and NMI handlers.
|
|||
|
However, there are legitimate preemptible-RCU implementations that do
|
|||
|
not have this property, given that any point in the code outside of an
|
|||
|
RCU read-side critical section can be a quiescent state. Therefore,
|
|||
|
*RCU-sched* was created, which follows “classic” RCU in that an
|
|||
|
RCU-sched grace period waits for pre-existing interrupt and NMI
|
|||
|
handlers. In kernels built with ``CONFIG_PREEMPTION=n``, the RCU and
|
|||
|
RCU-sched APIs have identical implementations, while kernels built with
|
|||
|
``CONFIG_PREEMPTION=y`` provide a separate implementation for each.
|
|||
|
|
|||
|
Note well that in ``CONFIG_PREEMPTION=y`` kernels,
|
|||
|
rcu_read_lock_sched() and rcu_read_unlock_sched() disable and
|
|||
|
re-enable preemption, respectively. This means that if there was a
|
|||
|
preemption attempt during the RCU-sched read-side critical section,
|
|||
|
rcu_read_unlock_sched() will enter the scheduler, with all the
|
|||
|
latency and overhead entailed. Just as with rcu_read_unlock_bh(),
|
|||
|
this can make it look as if rcu_read_unlock_sched() was executing
|
|||
|
very slowly. However, the highest-priority task won't be preempted, so
|
|||
|
that task will enjoy low-overhead rcu_read_unlock_sched()
|
|||
|
invocations.
|
|||
|
|
|||
|
The `RCU-sched
|
|||
|
API <https://lwn.net/Articles/609973/#RCU%20Per-Flavor%20API%20Table>`__
|
|||
|
includes rcu_read_lock_sched(), rcu_read_unlock_sched(),
|
|||
|
rcu_read_lock_sched_notrace(), rcu_read_unlock_sched_notrace(),
|
|||
|
rcu_dereference_sched(), rcu_dereference_sched_check(), and
|
|||
|
rcu_read_lock_sched_held(). However, the old RCU-sched update-side APIs
|
|||
|
are now gone, replaced by synchronize_rcu(), synchronize_rcu_expedited(),
|
|||
|
call_rcu(), and rcu_barrier(). In addition, anything that disables
|
|||
|
preemption also marks an RCU-sched read-side critical section,
|
|||
|
including preempt_disable() and preempt_enable(), local_irq_save()
|
|||
|
and local_irq_restore(), and so on.
|
|||
|
|
|||
|
Sleepable RCU
|
|||
|
~~~~~~~~~~~~~
|
|||
|
|
|||
|
For well over a decade, someone saying “I need to block within an RCU
|
|||
|
read-side critical section” was a reliable indication that this someone
|
|||
|
did not understand RCU. After all, if you are always blocking in an RCU
|
|||
|
read-side critical section, you can probably afford to use a
|
|||
|
higher-overhead synchronization mechanism. However, that changed with
|
|||
|
the advent of the Linux kernel's notifiers, whose RCU read-side critical
|
|||
|
sections almost never sleep, but sometimes need to. This resulted in the
|
|||
|
introduction of `sleepable RCU <https://lwn.net/Articles/202847/>`__, or
|
|||
|
*SRCU*.
|
|||
|
|
|||
|
SRCU allows different domains to be defined, with each such domain
|
|||
|
defined by an instance of an ``srcu_struct`` structure. A pointer to
|
|||
|
this structure must be passed in to each SRCU function, for example,
|
|||
|
``synchronize_srcu(&ss)``, where ``ss`` is the ``srcu_struct``
|
|||
|
structure. The key benefit of these domains is that a slow SRCU reader
|
|||
|
in one domain does not delay an SRCU grace period in some other domain.
|
|||
|
That said, one consequence of these domains is that read-side code must
|
|||
|
pass a “cookie” from srcu_read_lock() to srcu_read_unlock(), for
|
|||
|
example, as follows:
|
|||
|
|
|||
|
::
|
|||
|
|
|||
|
1 int idx;
|
|||
|
2
|
|||
|
3 idx = srcu_read_lock(&ss);
|
|||
|
4 do_something();
|
|||
|
5 srcu_read_unlock(&ss, idx);
|
|||
|
|
|||
|
As noted above, it is legal to block within SRCU read-side critical
|
|||
|
sections, however, with great power comes great responsibility. If you
|
|||
|
block forever in one of a given domain's SRCU read-side critical
|
|||
|
sections, then that domain's grace periods will also be blocked forever.
|
|||
|
Of course, one good way to block forever is to deadlock, which can
|
|||
|
happen if any operation in a given domain's SRCU read-side critical
|
|||
|
section can wait, either directly or indirectly, for that domain's grace
|
|||
|
period to elapse. For example, this results in a self-deadlock:
|
|||
|
|
|||
|
::
|
|||
|
|
|||
|
1 int idx;
|
|||
|
2
|
|||
|
3 idx = srcu_read_lock(&ss);
|
|||
|
4 do_something();
|
|||
|
5 synchronize_srcu(&ss);
|
|||
|
6 srcu_read_unlock(&ss, idx);
|
|||
|
|
|||
|
However, if line 5 acquired a mutex that was held across a
|
|||
|
synchronize_srcu() for domain ``ss``, deadlock would still be
|
|||
|
possible. Furthermore, if line 5 acquired a mutex that was held across a
|
|||
|
synchronize_srcu() for some other domain ``ss1``, and if an
|
|||
|
``ss1``-domain SRCU read-side critical section acquired another mutex
|
|||
|
that was held across as ``ss``-domain synchronize_srcu(), deadlock
|
|||
|
would again be possible. Such a deadlock cycle could extend across an
|
|||
|
arbitrarily large number of different SRCU domains. Again, with great
|
|||
|
power comes great responsibility.
|
|||
|
|
|||
|
Unlike the other RCU flavors, SRCU read-side critical sections can run
|
|||
|
on idle and even offline CPUs. This ability requires that
|
|||
|
srcu_read_lock() and srcu_read_unlock() contain memory barriers,
|
|||
|
which means that SRCU readers will run a bit slower than would RCU
|
|||
|
readers. It also motivates the smp_mb__after_srcu_read_unlock() API,
|
|||
|
which, in combination with srcu_read_unlock(), guarantees a full
|
|||
|
memory barrier.
|
|||
|
|
|||
|
Also unlike other RCU flavors, synchronize_srcu() may **not** be
|
|||
|
invoked from CPU-hotplug notifiers, due to the fact that SRCU grace
|
|||
|
periods make use of timers and the possibility of timers being
|
|||
|
temporarily “stranded” on the outgoing CPU. This stranding of timers
|
|||
|
means that timers posted to the outgoing CPU will not fire until late in
|
|||
|
the CPU-hotplug process. The problem is that if a notifier is waiting on
|
|||
|
an SRCU grace period, that grace period is waiting on a timer, and that
|
|||
|
timer is stranded on the outgoing CPU, then the notifier will never be
|
|||
|
awakened, in other words, deadlock has occurred. This same situation of
|
|||
|
course also prohibits srcu_barrier() from being invoked from
|
|||
|
CPU-hotplug notifiers.
|
|||
|
|
|||
|
SRCU also differs from other RCU flavors in that SRCU's expedited and
|
|||
|
non-expedited grace periods are implemented by the same mechanism. This
|
|||
|
means that in the current SRCU implementation, expediting a future grace
|
|||
|
period has the side effect of expediting all prior grace periods that
|
|||
|
have not yet completed. (But please note that this is a property of the
|
|||
|
current implementation, not necessarily of future implementations.) In
|
|||
|
addition, if SRCU has been idle for longer than the interval specified
|
|||
|
by the ``srcutree.exp_holdoff`` kernel boot parameter (25 microseconds
|
|||
|
by default), and if a synchronize_srcu() invocation ends this idle
|
|||
|
period, that invocation will be automatically expedited.
|
|||
|
|
|||
|
As of v4.12, SRCU's callbacks are maintained per-CPU, eliminating a
|
|||
|
locking bottleneck present in prior kernel versions. Although this will
|
|||
|
allow users to put much heavier stress on call_srcu(), it is
|
|||
|
important to note that SRCU does not yet take any special steps to deal
|
|||
|
with callback flooding. So if you are posting (say) 10,000 SRCU
|
|||
|
callbacks per second per CPU, you are probably totally OK, but if you
|
|||
|
intend to post (say) 1,000,000 SRCU callbacks per second per CPU, please
|
|||
|
run some tests first. SRCU just might need a few adjustment to deal with
|
|||
|
that sort of load. Of course, your mileage may vary based on the speed
|
|||
|
of your CPUs and the size of your memory.
|
|||
|
|
|||
|
The `SRCU
|
|||
|
API <https://lwn.net/Articles/609973/#RCU%20Per-Flavor%20API%20Table>`__
|
|||
|
includes srcu_read_lock(), srcu_read_unlock(),
|
|||
|
srcu_dereference(), srcu_dereference_check(),
|
|||
|
synchronize_srcu(), synchronize_srcu_expedited(),
|
|||
|
call_srcu(), srcu_barrier(), and srcu_read_lock_held(). It
|
|||
|
also includes DEFINE_SRCU(), DEFINE_STATIC_SRCU(), and
|
|||
|
init_srcu_struct() APIs for defining and initializing
|
|||
|
``srcu_struct`` structures.
|
|||
|
|
|||
|
More recently, the SRCU API has added polling interfaces:
|
|||
|
|
|||
|
#. start_poll_synchronize_srcu() returns a cookie identifying
|
|||
|
the completion of a future SRCU grace period and ensures
|
|||
|
that this grace period will be started.
|
|||
|
#. poll_state_synchronize_srcu() returns ``true`` iff the
|
|||
|
specified cookie corresponds to an already-completed
|
|||
|
SRCU grace period.
|
|||
|
#. get_state_synchronize_srcu() returns a cookie just like
|
|||
|
start_poll_synchronize_srcu() does, but differs in that
|
|||
|
it does nothing to ensure that any future SRCU grace period
|
|||
|
will be started.
|
|||
|
|
|||
|
These functions are used to avoid unnecessary SRCU grace periods in
|
|||
|
certain types of buffer-cache algorithms having multi-stage age-out
|
|||
|
mechanisms. The idea is that by the time the block has aged completely
|
|||
|
from the cache, an SRCU grace period will be very likely to have elapsed.
|
|||
|
|
|||
|
Tasks RCU
|
|||
|
~~~~~~~~~
|
|||
|
|
|||
|
Some forms of tracing use “trampolines” to handle the binary rewriting
|
|||
|
required to install different types of probes. It would be good to be
|
|||
|
able to free old trampolines, which sounds like a job for some form of
|
|||
|
RCU. However, because it is necessary to be able to install a trace
|
|||
|
anywhere in the code, it is not possible to use read-side markers such
|
|||
|
as rcu_read_lock() and rcu_read_unlock(). In addition, it does
|
|||
|
not work to have these markers in the trampoline itself, because there
|
|||
|
would need to be instructions following rcu_read_unlock(). Although
|
|||
|
synchronize_rcu() would guarantee that execution reached the
|
|||
|
rcu_read_unlock(), it would not be able to guarantee that execution
|
|||
|
had completely left the trampoline. Worse yet, in some situations
|
|||
|
the trampoline's protection must extend a few instructions *prior* to
|
|||
|
execution reaching the trampoline. For example, these few instructions
|
|||
|
might calculate the address of the trampoline, so that entering the
|
|||
|
trampoline would be pre-ordained a surprisingly long time before execution
|
|||
|
actually reached the trampoline itself.
|
|||
|
|
|||
|
The solution, in the form of `Tasks
|
|||
|
RCU <https://lwn.net/Articles/607117/>`__, is to have implicit read-side
|
|||
|
critical sections that are delimited by voluntary context switches, that
|
|||
|
is, calls to schedule(), cond_resched(), and
|
|||
|
synchronize_rcu_tasks(). In addition, transitions to and from
|
|||
|
userspace execution also delimit tasks-RCU read-side critical sections.
|
|||
|
|
|||
|
The tasks-RCU API is quite compact, consisting only of
|
|||
|
call_rcu_tasks(), synchronize_rcu_tasks(), and
|
|||
|
rcu_barrier_tasks(). In ``CONFIG_PREEMPTION=n`` kernels, trampolines
|
|||
|
cannot be preempted, so these APIs map to call_rcu(),
|
|||
|
synchronize_rcu(), and rcu_barrier(), respectively. In
|
|||
|
``CONFIG_PREEMPTION=y`` kernels, trampolines can be preempted, and these
|
|||
|
three APIs are therefore implemented by separate functions that check
|
|||
|
for voluntary context switches.
|
|||
|
|
|||
|
Tasks Rude RCU
|
|||
|
~~~~~~~~~~~~~~
|
|||
|
|
|||
|
Some forms of tracing need to wait for all preemption-disabled regions
|
|||
|
of code running on any online CPU, including those executed when RCU is
|
|||
|
not watching. This means that synchronize_rcu() is insufficient, and
|
|||
|
Tasks Rude RCU must be used instead. This flavor of RCU does its work by
|
|||
|
forcing a workqueue to be scheduled on each online CPU, hence the "Rude"
|
|||
|
moniker. And this operation is considered to be quite rude by real-time
|
|||
|
workloads that don't want their ``nohz_full`` CPUs receiving IPIs and
|
|||
|
by battery-powered systems that don't want their idle CPUs to be awakened.
|
|||
|
|
|||
|
The tasks-rude-RCU API is also reader-marking-free and thus quite compact,
|
|||
|
consisting of call_rcu_tasks_rude(), synchronize_rcu_tasks_rude(),
|
|||
|
and rcu_barrier_tasks_rude().
|
|||
|
|
|||
|
Tasks Trace RCU
|
|||
|
~~~~~~~~~~~~~~~
|
|||
|
|
|||
|
Some forms of tracing need to sleep in readers, but cannot tolerate
|
|||
|
SRCU's read-side overhead, which includes a full memory barrier in both
|
|||
|
srcu_read_lock() and srcu_read_unlock(). This need is handled by a
|
|||
|
Tasks Trace RCU that uses scheduler locking and IPIs to synchronize with
|
|||
|
readers. Real-time systems that cannot tolerate IPIs may build their
|
|||
|
kernels with ``CONFIG_TASKS_TRACE_RCU_READ_MB=y``, which avoids the IPIs at
|
|||
|
the expense of adding full memory barriers to the read-side primitives.
|
|||
|
|
|||
|
The tasks-trace-RCU API is also reasonably compact,
|
|||
|
consisting of rcu_read_lock_trace(), rcu_read_unlock_trace(),
|
|||
|
rcu_read_lock_trace_held(), call_rcu_tasks_trace(),
|
|||
|
synchronize_rcu_tasks_trace(), and rcu_barrier_tasks_trace().
|
|||
|
|
|||
|
Possible Future Changes
|
|||
|
-----------------------
|
|||
|
|
|||
|
One of the tricks that RCU uses to attain update-side scalability is to
|
|||
|
increase grace-period latency with increasing numbers of CPUs. If this
|
|||
|
becomes a serious problem, it will be necessary to rework the
|
|||
|
grace-period state machine so as to avoid the need for the additional
|
|||
|
latency.
|
|||
|
|
|||
|
RCU disables CPU hotplug in a few places, perhaps most notably in the
|
|||
|
rcu_barrier() operations. If there is a strong reason to use
|
|||
|
rcu_barrier() in CPU-hotplug notifiers, it will be necessary to
|
|||
|
avoid disabling CPU hotplug. This would introduce some complexity, so
|
|||
|
there had better be a *very* good reason.
|
|||
|
|
|||
|
The tradeoff between grace-period latency on the one hand and
|
|||
|
interruptions of other CPUs on the other hand may need to be
|
|||
|
re-examined. The desire is of course for zero grace-period latency as
|
|||
|
well as zero interprocessor interrupts undertaken during an expedited
|
|||
|
grace period operation. While this ideal is unlikely to be achievable,
|
|||
|
it is quite possible that further improvements can be made.
|
|||
|
|
|||
|
The multiprocessor implementations of RCU use a combining tree that
|
|||
|
groups CPUs so as to reduce lock contention and increase cache locality.
|
|||
|
However, this combining tree does not spread its memory across NUMA
|
|||
|
nodes nor does it align the CPU groups with hardware features such as
|
|||
|
sockets or cores. Such spreading and alignment is currently believed to
|
|||
|
be unnecessary because the hotpath read-side primitives do not access
|
|||
|
the combining tree, nor does call_rcu() in the common case. If you
|
|||
|
believe that your architecture needs such spreading and alignment, then
|
|||
|
your architecture should also benefit from the
|
|||
|
``rcutree.rcu_fanout_leaf`` boot parameter, which can be set to the
|
|||
|
number of CPUs in a socket, NUMA node, or whatever. If the number of
|
|||
|
CPUs is too large, use a fraction of the number of CPUs. If the number
|
|||
|
of CPUs is a large prime number, well, that certainly is an
|
|||
|
“interesting” architectural choice! More flexible arrangements might be
|
|||
|
considered, but only if ``rcutree.rcu_fanout_leaf`` has proven
|
|||
|
inadequate, and only if the inadequacy has been demonstrated by a
|
|||
|
carefully run and realistic system-level workload.
|
|||
|
|
|||
|
Please note that arrangements that require RCU to remap CPU numbers will
|
|||
|
require extremely good demonstration of need and full exploration of
|
|||
|
alternatives.
|
|||
|
|
|||
|
RCU's various kthreads are reasonably recent additions. It is quite
|
|||
|
likely that adjustments will be required to more gracefully handle
|
|||
|
extreme loads. It might also be necessary to be able to relate CPU
|
|||
|
utilization by RCU's kthreads and softirq handlers to the code that
|
|||
|
instigated this CPU utilization. For example, RCU callback overhead
|
|||
|
might be charged back to the originating call_rcu() instance, though
|
|||
|
probably not in production kernels.
|
|||
|
|
|||
|
Additional work may be required to provide reasonable forward-progress
|
|||
|
guarantees under heavy load for grace periods and for callback
|
|||
|
invocation.
|
|||
|
|
|||
|
Summary
|
|||
|
-------
|
|||
|
|
|||
|
This document has presented more than two decade's worth of RCU
|
|||
|
requirements. Given that the requirements keep changing, this will not
|
|||
|
be the last word on this subject, but at least it serves to get an
|
|||
|
important subset of the requirements set forth.
|
|||
|
|
|||
|
Acknowledgments
|
|||
|
---------------
|
|||
|
|
|||
|
I am grateful to Steven Rostedt, Lai Jiangshan, Ingo Molnar, Oleg
|
|||
|
Nesterov, Borislav Petkov, Peter Zijlstra, Boqun Feng, and Andy
|
|||
|
Lutomirski for their help in rendering this article human readable, and
|
|||
|
to Michelle Rankin for her support of this effort. Other contributions
|
|||
|
are acknowledged in the Linux kernel's git archive.
|