124 lines
4.2 KiB
ReStructuredText
124 lines
4.2 KiB
ReStructuredText
|
.. _NMI_rcu_doc:
|
||
|
|
||
|
Using RCU to Protect Dynamic NMI Handlers
|
||
|
=========================================
|
||
|
|
||
|
|
||
|
Although RCU is usually used to protect read-mostly data structures,
|
||
|
it is possible to use RCU to provide dynamic non-maskable interrupt
|
||
|
handlers, as well as dynamic irq handlers. This document describes
|
||
|
how to do this, drawing loosely from Zwane Mwaikambo's NMI-timer
|
||
|
work in "arch/x86/kernel/traps.c".
|
||
|
|
||
|
The relevant pieces of code are listed below, each followed by a
|
||
|
brief explanation::
|
||
|
|
||
|
static int dummy_nmi_callback(struct pt_regs *regs, int cpu)
|
||
|
{
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
The dummy_nmi_callback() function is a "dummy" NMI handler that does
|
||
|
nothing, but returns zero, thus saying that it did nothing, allowing
|
||
|
the NMI handler to take the default machine-specific action::
|
||
|
|
||
|
static nmi_callback_t nmi_callback = dummy_nmi_callback;
|
||
|
|
||
|
This nmi_callback variable is a global function pointer to the current
|
||
|
NMI handler::
|
||
|
|
||
|
void do_nmi(struct pt_regs * regs, long error_code)
|
||
|
{
|
||
|
int cpu;
|
||
|
|
||
|
nmi_enter();
|
||
|
|
||
|
cpu = smp_processor_id();
|
||
|
++nmi_count(cpu);
|
||
|
|
||
|
if (!rcu_dereference_sched(nmi_callback)(regs, cpu))
|
||
|
default_do_nmi(regs);
|
||
|
|
||
|
nmi_exit();
|
||
|
}
|
||
|
|
||
|
The do_nmi() function processes each NMI. It first disables preemption
|
||
|
in the same way that a hardware irq would, then increments the per-CPU
|
||
|
count of NMIs. It then invokes the NMI handler stored in the nmi_callback
|
||
|
function pointer. If this handler returns zero, do_nmi() invokes the
|
||
|
default_do_nmi() function to handle a machine-specific NMI. Finally,
|
||
|
preemption is restored.
|
||
|
|
||
|
In theory, rcu_dereference_sched() is not needed, since this code runs
|
||
|
only on i386, which in theory does not need rcu_dereference_sched()
|
||
|
anyway. However, in practice it is a good documentation aid, particularly
|
||
|
for anyone attempting to do something similar on Alpha or on systems
|
||
|
with aggressive optimizing compilers.
|
||
|
|
||
|
Quick Quiz:
|
||
|
Why might the rcu_dereference_sched() be necessary on Alpha, given that the code referenced by the pointer is read-only?
|
||
|
|
||
|
:ref:`Answer to Quick Quiz <answer_quick_quiz_NMI>`
|
||
|
|
||
|
Back to the discussion of NMI and RCU::
|
||
|
|
||
|
void set_nmi_callback(nmi_callback_t callback)
|
||
|
{
|
||
|
rcu_assign_pointer(nmi_callback, callback);
|
||
|
}
|
||
|
|
||
|
The set_nmi_callback() function registers an NMI handler. Note that any
|
||
|
data that is to be used by the callback must be initialized up -before-
|
||
|
the call to set_nmi_callback(). On architectures that do not order
|
||
|
writes, the rcu_assign_pointer() ensures that the NMI handler sees the
|
||
|
initialized values::
|
||
|
|
||
|
void unset_nmi_callback(void)
|
||
|
{
|
||
|
rcu_assign_pointer(nmi_callback, dummy_nmi_callback);
|
||
|
}
|
||
|
|
||
|
This function unregisters an NMI handler, restoring the original
|
||
|
dummy_nmi_handler(). However, there may well be an NMI handler
|
||
|
currently executing on some other CPU. We therefore cannot free
|
||
|
up any data structures used by the old NMI handler until execution
|
||
|
of it completes on all other CPUs.
|
||
|
|
||
|
One way to accomplish this is via synchronize_rcu(), perhaps as
|
||
|
follows::
|
||
|
|
||
|
unset_nmi_callback();
|
||
|
synchronize_rcu();
|
||
|
kfree(my_nmi_data);
|
||
|
|
||
|
This works because (as of v4.20) synchronize_rcu() blocks until all
|
||
|
CPUs complete any preemption-disabled segments of code that they were
|
||
|
executing.
|
||
|
Since NMI handlers disable preemption, synchronize_rcu() is guaranteed
|
||
|
not to return until all ongoing NMI handlers exit. It is therefore safe
|
||
|
to free up the handler's data as soon as synchronize_rcu() returns.
|
||
|
|
||
|
Important note: for this to work, the architecture in question must
|
||
|
invoke nmi_enter() and nmi_exit() on NMI entry and exit, respectively.
|
||
|
|
||
|
.. _answer_quick_quiz_NMI:
|
||
|
|
||
|
Answer to Quick Quiz:
|
||
|
Why might the rcu_dereference_sched() be necessary on Alpha, given that the code referenced by the pointer is read-only?
|
||
|
|
||
|
The caller to set_nmi_callback() might well have
|
||
|
initialized some data that is to be used by the new NMI
|
||
|
handler. In this case, the rcu_dereference_sched() would
|
||
|
be needed, because otherwise a CPU that received an NMI
|
||
|
just after the new handler was set might see the pointer
|
||
|
to the new NMI handler, but the old pre-initialized
|
||
|
version of the handler's data.
|
||
|
|
||
|
This same sad story can happen on other CPUs when using
|
||
|
a compiler with aggressive pointer-value speculation
|
||
|
optimizations.
|
||
|
|
||
|
More important, the rcu_dereference_sched() makes it
|
||
|
clear to someone reading the code that the pointer is
|
||
|
being protected by RCU-sched.
|