timers/migration: Move hierarchy setup into cpuhotplug prepare callback
When a CPU comes online the first time, it is possible that a new top level
group will be created. In general all propagation is done from the bottom
to top. This minimizes complexity and prevents possible races. But when a
new top level group is created, the formely top level group needs to be
connected to the new level. This is the only time, when the direction to
propagate changes is changed: the changes are propagated from top (new top
level group) to bottom (formerly top level group).
This introduces two races (see (A) and (B)) as reported by Frederic:
(A) This race happens, when marking the formely top level group as active,
but the last active CPU of the formerly top level group goes idle. Then
it's likely that formerly group is no longer active, but marked
nevertheless as active in new top level group:
[GRP0:0]
migrator = 0
active = 0
nextevt = KTIME_MAX
/ \
0 1 .. 7
active idle
0) Hierarchy has for now only 8 CPUs and CPU 0 is the only active CPU.
[GRP1:0]
migrator = TMIGR_NONE
active = NONE
nextevt = KTIME_MAX
\
[GRP0:0] [GRP0:1]
migrator = 0 migrator = TMIGR_NONE
active = 0 active = NONE
nextevt = KTIME_MAX nextevt = KTIME_MAX
/ \
0 1 .. 7 8
active idle !online
1) CPU 8 is booting and creates a new group in first level GRP0:1 and
therefore also a new top group GRP1:0. For now the setup code proceeded
only until the connected between GRP0:1 to the new top group. The
connection between CPU8 and GRP0:1 is not yet established and CPU 8 is
still !online.
[GRP1:0]
migrator = TMIGR_NONE
active = NONE
nextevt = KTIME_MAX
/ \
[GRP0:0] [GRP0:1]
migrator = 0 migrator = TMIGR_NONE
active = 0 active = NONE
nextevt = KTIME_MAX nextevt = KTIME_MAX
/ \
0 1 .. 7 8
active idle !online
2) Setup code now connects GRP0:0 to GRP1:0 and observes while in
tmigr_connect_child_parent() that GRP0:0 is not TMIGR_NONE. So it
prepares to call tmigr_active_up() on it. It hasn't done it yet.
[GRP1:0]
migrator = TMIGR_NONE
active = NONE
nextevt = KTIME_MAX
/ \
[GRP0:0] [GRP0:1]
migrator = TMIGR_NONE migrator = TMIGR_NONE
active = NONE active = NONE
nextevt = KTIME_MAX nextevt = KTIME_MAX
/ \
0 1 .. 7 8
idle idle !online
3) CPU 0 goes idle. Since GRP0:0->parent has been updated by CPU 8 with
GRP0:0->lock held, CPU 0 observes GRP1:0 after calling
tmigr_update_events() and it propagates the change to the top (no change
there and no wakeup programmed since there is no timer).
[GRP1:0]
migrator = GRP0:0
active = GRP0:0
nextevt = KTIME_MAX
/ \
[GRP0:0] [GRP0:1]
migrator = TMIGR_NONE migrator = TMIGR_NONE
active = NONE active = NONE
nextevt = KTIME_MAX nextevt = KTIME_MAX
/ \
0 1 .. 7 8
idle idle !online
4) Now the setup code finally calls tmigr_active_up() to and sets GRP0:0
active in GRP1:0
[GRP1:0]
migrator = GRP0:0
active = GRP0:0, GRP0:1
nextevt = KTIME_MAX
/ \
[GRP0:0] [GRP0:1]
migrator = TMIGR_NONE migrator = 8
active = NONE active = 8
nextevt = KTIME_MAX nextevt = KTIME_MAX
/ \ |
0 1 .. 7 8
idle idle active
5) Now CPU 8 is connected with GRP0:1 and CPU 8 calls tmigr_active_up() out
of tmigr_cpu_online().
[GRP1:0]
migrator = GRP0:0
active = GRP0:0
nextevt = T8
/ \
[GRP0:0] [GRP0:1]
migrator = TMIGR_NONE migrator = TMIGR_NONE
active = NONE active = NONE
nextevt = KTIME_MAX nextevt = T8
/ \ |
0 1 .. 7 8
idle idle idle
5) CPU 8 goes idle with a timer T8 and relies on GRP0:0 as the migrator.
But it's not really active, so T8 gets ignored.
--> The update which is done in third step is not noticed by setup code. So
a wrong migrator is set to top level group and a timer could get
ignored.
(B) Reading group->parent and group->childmask when an hierarchy update is
ongoing and reaches the formerly top level group is racy as those values
could be inconsistent. (The notation of migrator and active now slightly
changes in contrast to the above example, as now the childmasks are used.)
[GRP1:0]
migrator = TMIGR_NONE
active = 0x00
nextevt = KTIME_MAX
\
[GRP0:0] [GRP0:1]
migrator = TMIGR_NONE migrator = TMIGR_NONE
active = 0x00 active = 0x00
nextevt = KTIME_MAX nextevt = KTIME_MAX
childmask= 0 childmask= 1
parent = NULL parent = GRP1:0
/ \
0 1 .. 7 8
idle idle !online
childmask=1
1) Hierarchy has 8 CPUs. CPU 8 is at the moment in the process of onlining
but did not yet connect GRP0:0 to GRP1:0.
[GRP1:0]
migrator = TMIGR_NONE
active = 0x00
nextevt = KTIME_MAX
/ \
[GRP0:0] [GRP0:1]
migrator = TMIGR_NONE migrator = TMIGR_NONE
active = 0x00 active = 0x00
nextevt = KTIME_MAX nextevt = KTIME_MAX
childmask= 0 childmask= 1
parent = GRP1:0 parent = GRP1:0
/ \
0 1 .. 7 8
idle idle !online
childmask=1
2) Setup code (running on CPU 8) now connects GRP0:0 to GRP1:0, updates
parent pointer of GRP0:0 and ...
[GRP1:0]
migrator = TMIGR_NONE
active = 0x00
nextevt = KTIME_MAX
/ \
[GRP0:0] [GRP0:1]
migrator = 0x01 migrator = TMIGR_NONE
active = 0x01 active = 0x00
nextevt = KTIME_MAX nextevt = KTIME_MAX
childmask= 0 childmask= 1
parent = GRP1:0 parent = GRP1:0
/ \
0 1 .. 7 8
active idle !online
childmask=1
tmigr_walk.childmask = 0
3) ... CPU 0 comes active in the same time. As migrator in GRP0:0 was
TMIGR_NONE, childmask of GRP0:0 is stored in update propagation data
structure tmigr_walk (as update of childmask is not yet
visible/updated). And now ...
[GRP1:0]
migrator = TMIGR_NONE
active = 0x00
nextevt = KTIME_MAX
/ \
[GRP0:0] [GRP0:1]
migrator = 0x01 migrator = TMIGR_NONE
active = 0x01 active = 0x00
nextevt = KTIME_MAX nextevt = KTIME_MAX
childmask= 2 childmask= 1
parent = GRP1:0 parent = GRP1:0
/ \
0 1 .. 7 8
active idle !online
childmask=1
tmigr_walk.childmask = 0
4) ... childmask of GRP0:0 is updated by CPU 8 (still part of setup
code).
[GRP1:0]
migrator = 0x00
active = 0x00
nextevt = KTIME_MAX
/ \
[GRP0:0] [GRP0:1]
migrator = 0x01 migrator = TMIGR_NONE
active = 0x01 active = 0x00
nextevt = KTIME_MAX nextevt = KTIME_MAX
childmask= 2 childmask= 1
parent = GRP1:0 parent = GRP1:0
/ \
0 1 .. 7 8
active idle !online
childmask=1
tmigr_walk.childmask = 0
5) CPU 0 sees the connection to GRP1:0 and now propagates active state to
GRP1:0 but with childmask = 0 as stored in propagation data structure.
--> Now GRP1:0 always has a migrator as 0x00 != TMIGR_NONE and for all CPUs
it looks like GRP1:0 is always active.
To prevent those races, the setup of the hierarchy is moved into the
cpuhotplug prepare callback. The prepare callback is not executed by the
CPU which will come online, it is executed by the CPU which prepares
onlining of the other CPU. This CPU is active while it is connecting the
formerly top level to the new one. This prevents from (A) to happen and it
also prevents from any further walk above the formerly top level until that
active CPU becomes inactive, releasing the new ->parent and ->childmask
updates to be visible by any subsequent walk up above the formerly top
level hierarchy. This prevents from (B) to happen. The direction for the
updates is now forced to look like "from bottom to top".
However if the active CPU prevents from tmigr_cpu_(in)active() to walk up
with the update not-or-half visible, nothing prevents walking up to the new
top with a 0 childmask in tmigr_handle_remote_up() or
tmigr_requires_handle_remote_up() if the active CPU doing the prepare is
not the migrator. But then it looks fine because:
* tmigr_check_migrator() should just return false
* The migrator is active and should eventually observe the new childmask
at some point in a future tick.
Split setup functionality of online callback into the cpuhotplug prepare
callback and setup hotplug state. Change init call into early_initcall() to
make sure an already active CPU prepares everything for newly upcoming
CPUs. Reorder the code, that all prepare related functions are close to
each other and online and offline callbacks are also close together.
Fixes: 7ee9887703
("timers: Implement the hierarchical pull model")
Signed-off-by: Anna-Maria Behnsen <anna-maria@linutronix.de>
Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
Reviewed-by: Frederic Weisbecker <frederic@kernel.org>
Link: https://lore.kernel.org/r/20240717094940.18687-1-anna-maria@linutronix.de
This commit is contained in:
parent
facd40aa5c
commit
10a0e6f3d3
@ -122,6 +122,7 @@ enum cpuhp_state {
|
||||
CPUHP_KVM_PPC_BOOK3S_PREPARE,
|
||||
CPUHP_ZCOMP_PREPARE,
|
||||
CPUHP_TIMERS_PREPARE,
|
||||
CPUHP_TMIGR_PREPARE,
|
||||
CPUHP_MIPS_SOC_PREPARE,
|
||||
CPUHP_BP_PREPARE_DYN,
|
||||
CPUHP_BP_PREPARE_DYN_END = CPUHP_BP_PREPARE_DYN + 20,
|
||||
|
@ -1438,6 +1438,66 @@ u64 tmigr_quick_check(u64 nextevt)
|
||||
return KTIME_MAX;
|
||||
}
|
||||
|
||||
/*
|
||||
* tmigr_trigger_active() - trigger a CPU to become active again
|
||||
*
|
||||
* This function is executed on a CPU which is part of cpu_online_mask, when the
|
||||
* last active CPU in the hierarchy is offlining. With this, it is ensured that
|
||||
* the other CPU is active and takes over the migrator duty.
|
||||
*/
|
||||
static long tmigr_trigger_active(void *unused)
|
||||
{
|
||||
struct tmigr_cpu *tmc = this_cpu_ptr(&tmigr_cpu);
|
||||
|
||||
WARN_ON_ONCE(!tmc->online || tmc->idle);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int tmigr_cpu_offline(unsigned int cpu)
|
||||
{
|
||||
struct tmigr_cpu *tmc = this_cpu_ptr(&tmigr_cpu);
|
||||
int migrator;
|
||||
u64 firstexp;
|
||||
|
||||
raw_spin_lock_irq(&tmc->lock);
|
||||
tmc->online = false;
|
||||
WRITE_ONCE(tmc->wakeup, KTIME_MAX);
|
||||
|
||||
/*
|
||||
* CPU has to handle the local events on his own, when on the way to
|
||||
* offline; Therefore nextevt value is set to KTIME_MAX
|
||||
*/
|
||||
firstexp = __tmigr_cpu_deactivate(tmc, KTIME_MAX);
|
||||
trace_tmigr_cpu_offline(tmc);
|
||||
raw_spin_unlock_irq(&tmc->lock);
|
||||
|
||||
if (firstexp != KTIME_MAX) {
|
||||
migrator = cpumask_any_but(cpu_online_mask, cpu);
|
||||
work_on_cpu(migrator, tmigr_trigger_active, NULL);
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int tmigr_cpu_online(unsigned int cpu)
|
||||
{
|
||||
struct tmigr_cpu *tmc = this_cpu_ptr(&tmigr_cpu);
|
||||
|
||||
/* Check whether CPU data was successfully initialized */
|
||||
if (WARN_ON_ONCE(!tmc->tmgroup))
|
||||
return -EINVAL;
|
||||
|
||||
raw_spin_lock_irq(&tmc->lock);
|
||||
trace_tmigr_cpu_online(tmc);
|
||||
tmc->idle = timer_base_is_idle();
|
||||
if (!tmc->idle)
|
||||
__tmigr_cpu_activate(tmc);
|
||||
tmc->online = true;
|
||||
raw_spin_unlock_irq(&tmc->lock);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void tmigr_init_group(struct tmigr_group *group, unsigned int lvl,
|
||||
int node)
|
||||
{
|
||||
@ -1510,9 +1570,10 @@ static struct tmigr_group *tmigr_get_group(unsigned int cpu, int node,
|
||||
}
|
||||
|
||||
static void tmigr_connect_child_parent(struct tmigr_group *child,
|
||||
struct tmigr_group *parent)
|
||||
struct tmigr_group *parent,
|
||||
bool activate)
|
||||
{
|
||||
union tmigr_state childstate;
|
||||
struct tmigr_walk data;
|
||||
|
||||
raw_spin_lock_irq(&child->lock);
|
||||
raw_spin_lock_nested(&parent->lock, SINGLE_DEPTH_NESTING);
|
||||
@ -1525,6 +1586,9 @@ static void tmigr_connect_child_parent(struct tmigr_group *child,
|
||||
|
||||
trace_tmigr_connect_child_parent(child);
|
||||
|
||||
if (!activate)
|
||||
return;
|
||||
|
||||
/*
|
||||
* To prevent inconsistent states, active children need to be active in
|
||||
* the new parent as well. Inactive children are already marked inactive
|
||||
@ -1540,22 +1604,24 @@ static void tmigr_connect_child_parent(struct tmigr_group *child,
|
||||
* child to the new parent. So tmigr_connect_child_parent() is
|
||||
* executed with the formerly top level group (child) and the newly
|
||||
* created group (parent).
|
||||
*
|
||||
* * It is ensured that the child is active, as this setup path is
|
||||
* executed in hotplug prepare callback. This is exectued by an
|
||||
* already connected and !idle CPU. Even if all other CPUs go idle,
|
||||
* the CPU executing the setup will be responsible up to current top
|
||||
* level group. And the next time it goes inactive, it will release
|
||||
* the new childmask and parent to subsequent walkers through this
|
||||
* @child. Therefore propagate active state unconditionally.
|
||||
*/
|
||||
childstate.state = atomic_read(&child->migr_state);
|
||||
if (childstate.migrator != TMIGR_NONE) {
|
||||
struct tmigr_walk data;
|
||||
data.childmask = child->childmask;
|
||||
|
||||
data.childmask = child->childmask;
|
||||
|
||||
/*
|
||||
* There is only one new level per time (which is protected by
|
||||
* tmigr_mutex). When connecting the child and the parent and
|
||||
* set the child active when the parent is inactive, the parent
|
||||
* needs to be the uppermost level. Otherwise there went
|
||||
* something wrong!
|
||||
*/
|
||||
WARN_ON(!tmigr_active_up(parent, child, &data) && parent->parent);
|
||||
}
|
||||
/*
|
||||
* There is only one new level per time (which is protected by
|
||||
* tmigr_mutex). When connecting the child and the parent and set the
|
||||
* child active when the parent is inactive, the parent needs to be the
|
||||
* uppermost level. Otherwise there went something wrong!
|
||||
*/
|
||||
WARN_ON(!tmigr_active_up(parent, child, &data) && parent->parent);
|
||||
}
|
||||
|
||||
static int tmigr_setup_groups(unsigned int cpu, unsigned int node)
|
||||
@ -1608,7 +1674,7 @@ static int tmigr_setup_groups(unsigned int cpu, unsigned int node)
|
||||
* Update tmc -> group / child -> group connection
|
||||
*/
|
||||
if (i == 0) {
|
||||
struct tmigr_cpu *tmc = this_cpu_ptr(&tmigr_cpu);
|
||||
struct tmigr_cpu *tmc = per_cpu_ptr(&tmigr_cpu, cpu);
|
||||
|
||||
raw_spin_lock_irq(&group->lock);
|
||||
|
||||
@ -1623,7 +1689,8 @@ static int tmigr_setup_groups(unsigned int cpu, unsigned int node)
|
||||
continue;
|
||||
} else {
|
||||
child = stack[i - 1];
|
||||
tmigr_connect_child_parent(child, group);
|
||||
/* Will be activated at online time */
|
||||
tmigr_connect_child_parent(child, group, false);
|
||||
}
|
||||
|
||||
/* check if uppermost level was newly created */
|
||||
@ -1634,12 +1701,21 @@ static int tmigr_setup_groups(unsigned int cpu, unsigned int node)
|
||||
|
||||
lvllist = &tmigr_level_list[top];
|
||||
if (group->num_children == 1 && list_is_singular(lvllist)) {
|
||||
/*
|
||||
* The target CPU must never do the prepare work, except
|
||||
* on early boot when the boot CPU is the target. Otherwise
|
||||
* it may spuriously activate the old top level group inside
|
||||
* the new one (nevertheless whether old top level group is
|
||||
* active or not) and/or release an uninitialized childmask.
|
||||
*/
|
||||
WARN_ON_ONCE(cpu == raw_smp_processor_id());
|
||||
|
||||
lvllist = &tmigr_level_list[top - 1];
|
||||
list_for_each_entry(child, lvllist, list) {
|
||||
if (child->parent)
|
||||
continue;
|
||||
|
||||
tmigr_connect_child_parent(child, group);
|
||||
tmigr_connect_child_parent(child, group, true);
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1661,80 +1737,31 @@ static int tmigr_add_cpu(unsigned int cpu)
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int tmigr_cpu_online(unsigned int cpu)
|
||||
static int tmigr_cpu_prepare(unsigned int cpu)
|
||||
{
|
||||
struct tmigr_cpu *tmc = this_cpu_ptr(&tmigr_cpu);
|
||||
int ret;
|
||||
struct tmigr_cpu *tmc = per_cpu_ptr(&tmigr_cpu, cpu);
|
||||
int ret = 0;
|
||||
|
||||
/* First online attempt? Initialize CPU data */
|
||||
if (!tmc->tmgroup) {
|
||||
raw_spin_lock_init(&tmc->lock);
|
||||
/* Not first online attempt? */
|
||||
if (tmc->tmgroup)
|
||||
return ret;
|
||||
|
||||
ret = tmigr_add_cpu(cpu);
|
||||
if (ret < 0)
|
||||
return ret;
|
||||
|
||||
if (tmc->childmask == 0)
|
||||
return -EINVAL;
|
||||
|
||||
timerqueue_init(&tmc->cpuevt.nextevt);
|
||||
tmc->cpuevt.nextevt.expires = KTIME_MAX;
|
||||
tmc->cpuevt.ignore = true;
|
||||
tmc->cpuevt.cpu = cpu;
|
||||
|
||||
tmc->remote = false;
|
||||
WRITE_ONCE(tmc->wakeup, KTIME_MAX);
|
||||
}
|
||||
raw_spin_lock_irq(&tmc->lock);
|
||||
trace_tmigr_cpu_online(tmc);
|
||||
tmc->idle = timer_base_is_idle();
|
||||
if (!tmc->idle)
|
||||
__tmigr_cpu_activate(tmc);
|
||||
tmc->online = true;
|
||||
raw_spin_unlock_irq(&tmc->lock);
|
||||
return 0;
|
||||
}
|
||||
|
||||
/*
|
||||
* tmigr_trigger_active() - trigger a CPU to become active again
|
||||
*
|
||||
* This function is executed on a CPU which is part of cpu_online_mask, when the
|
||||
* last active CPU in the hierarchy is offlining. With this, it is ensured that
|
||||
* the other CPU is active and takes over the migrator duty.
|
||||
*/
|
||||
static long tmigr_trigger_active(void *unused)
|
||||
{
|
||||
struct tmigr_cpu *tmc = this_cpu_ptr(&tmigr_cpu);
|
||||
|
||||
WARN_ON_ONCE(!tmc->online || tmc->idle);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int tmigr_cpu_offline(unsigned int cpu)
|
||||
{
|
||||
struct tmigr_cpu *tmc = this_cpu_ptr(&tmigr_cpu);
|
||||
int migrator;
|
||||
u64 firstexp;
|
||||
|
||||
raw_spin_lock_irq(&tmc->lock);
|
||||
tmc->online = false;
|
||||
raw_spin_lock_init(&tmc->lock);
|
||||
timerqueue_init(&tmc->cpuevt.nextevt);
|
||||
tmc->cpuevt.nextevt.expires = KTIME_MAX;
|
||||
tmc->cpuevt.ignore = true;
|
||||
tmc->cpuevt.cpu = cpu;
|
||||
tmc->remote = false;
|
||||
WRITE_ONCE(tmc->wakeup, KTIME_MAX);
|
||||
|
||||
/*
|
||||
* CPU has to handle the local events on his own, when on the way to
|
||||
* offline; Therefore nextevt value is set to KTIME_MAX
|
||||
*/
|
||||
firstexp = __tmigr_cpu_deactivate(tmc, KTIME_MAX);
|
||||
trace_tmigr_cpu_offline(tmc);
|
||||
raw_spin_unlock_irq(&tmc->lock);
|
||||
ret = tmigr_add_cpu(cpu);
|
||||
if (ret < 0)
|
||||
return ret;
|
||||
|
||||
if (firstexp != KTIME_MAX) {
|
||||
migrator = cpumask_any_but(cpu_online_mask, cpu);
|
||||
work_on_cpu(migrator, tmigr_trigger_active, NULL);
|
||||
}
|
||||
if (tmc->childmask == 0)
|
||||
return -EINVAL;
|
||||
|
||||
return 0;
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int __init tmigr_init(void)
|
||||
@ -1793,6 +1820,11 @@ static int __init tmigr_init(void)
|
||||
tmigr_hierarchy_levels, TMIGR_CHILDREN_PER_GROUP,
|
||||
tmigr_crossnode_level);
|
||||
|
||||
ret = cpuhp_setup_state(CPUHP_TMIGR_PREPARE, "tmigr:prepare",
|
||||
tmigr_cpu_prepare, NULL);
|
||||
if (ret)
|
||||
goto err;
|
||||
|
||||
ret = cpuhp_setup_state(CPUHP_AP_TMIGR_ONLINE, "tmigr:online",
|
||||
tmigr_cpu_online, tmigr_cpu_offline);
|
||||
if (ret)
|
||||
@ -1804,4 +1836,4 @@ err:
|
||||
pr_err("Timer migration setup failed\n");
|
||||
return ret;
|
||||
}
|
||||
late_initcall(tmigr_init);
|
||||
early_initcall(tmigr_init);
|
||||
|
Loading…
Reference in New Issue
Block a user