* [PATCH 1/2] lazy tlb: fix hotplug exit race with MMU_LAZY_TLB_SHOOTDOWN
@ 2023-05-24 6:04 Nicholas Piggin
2023-05-24 6:04 ` [PATCH 2/2] lazy tlb: consolidate lazy tlb mm switching Nicholas Piggin
0 siblings, 1 reply; 2+ messages in thread
From: Nicholas Piggin @ 2023-05-24 6:04 UTC (permalink / raw)
To: Andrew Morton
Cc: Peter Zijlstra, linuxppc-dev, Linus Torvalds, Nicholas Piggin,
linux-mm
CPU unplug first calls __cpu_disable(), and that's where powerpc calls
cleanup_cpu_mmu_context(), which clears this CPU from mm_cpumask() of
all mms in the system.
However this CPU may still be using a lazy tlb mm, and its mm_cpumask
bit will be cleared from it. The CPU does not switch away from the lazy
tlb mm until arch_cpu_idle_dead() calls idle_task_exit().
If that user mm exits in this window, it will not be subject to the lazy
tlb mm shootdown and may be freed while in use as a lazy mm by the CPU
that is being unplugged.
cleanup_cpu_mmu_context() could be moved later, but it looks better to
move the lazy tlb mm switching earlier. The problem with doing the lazy
mm switching in idle_task_exit() is explained in commit bf2c59fce4074
("sched/core: Fix illegal RCU from offline CPUs"), which added a wart to
switch away from the mm but leave it set in active_mm to be cleaned up
later.
So instead, switch away from the lazy tlb mm on the stopper kthread
before the CPU is taken down. This CPU will never switch to a user
thread from this point, so it has no chance to pick up a new lazy tlb
mm. This removes the lazy tlb mm handling wart in CPU unplug.
idle_task_exit() remains to reduce churn in the patch. It could be
removed entirely after this because finish_cpu() makes a similar check.
finish_cpu() itself is not strictly needed because init_mm will never
have its refcount drop to zero. But it is conceptually nicer to keep it
rather than have the idle thread drop the reference on the mm it is
using.
Fixes: 2655421ae69fa ("lazy tlb: shoot lazies, non-refcounting lazy tlb mm reference handling scheme")
Signed-off-by: Nicholas Piggin <npiggin@gmail.com>
---
include/linux/sched/hotplug.h | 2 ++
kernel/cpu.c | 11 +++++++----
kernel/sched/core.c | 24 +++++++++++++++++++-----
3 files changed, 28 insertions(+), 9 deletions(-)
diff --git a/include/linux/sched/hotplug.h b/include/linux/sched/hotplug.h
index 412cdaba33eb..cb447d8e3f9a 100644
--- a/include/linux/sched/hotplug.h
+++ b/include/linux/sched/hotplug.h
@@ -19,8 +19,10 @@ extern int sched_cpu_dying(unsigned int cpu);
#endif
#ifdef CONFIG_HOTPLUG_CPU
+extern void idle_task_prepare_exit(void);
extern void idle_task_exit(void);
#else
+static inline void idle_task_prepare_exit(void) {}
static inline void idle_task_exit(void) {}
#endif
diff --git a/kernel/cpu.c b/kernel/cpu.c
index f4a2c5845bcb..584def27ff24 100644
--- a/kernel/cpu.c
+++ b/kernel/cpu.c
@@ -618,12 +618,13 @@ static int finish_cpu(unsigned int cpu)
struct mm_struct *mm = idle->active_mm;
/*
- * idle_task_exit() will have switched to &init_mm, now
- * clean up any remaining active_mm state.
+ * idle_task_prepare_exit() ensured the idle task was using
+ * &init_mm. Now that the CPU has stopped, drop that refcount.
*/
- if (mm != &init_mm)
- idle->active_mm = &init_mm;
+ WARN_ON(mm != &init_mm);
+ idle->active_mm = NULL;
mmdrop_lazy_tlb(mm);
+
return 0;
}
@@ -1030,6 +1031,8 @@ static int take_cpu_down(void *_param)
enum cpuhp_state target = max((int)st->target, CPUHP_AP_OFFLINE);
int err, cpu = smp_processor_id();
+ idle_task_prepare_exit();
+
/* Ensure this CPU doesn't handle any more interrupts. */
err = __cpu_disable();
if (err < 0)
diff --git a/kernel/sched/core.c b/kernel/sched/core.c
index a68d1276bab0..bc4ef1f3394b 100644
--- a/kernel/sched/core.c
+++ b/kernel/sched/core.c
@@ -9373,19 +9373,33 @@ void sched_setnuma(struct task_struct *p, int nid)
* Ensure that the idle task is using init_mm right before its CPU goes
* offline.
*/
-void idle_task_exit(void)
+void idle_task_prepare_exit(void)
{
struct mm_struct *mm = current->active_mm;
- BUG_ON(cpu_online(smp_processor_id()));
- BUG_ON(current != this_rq()->idle);
+ WARN_ON(!irqs_disabled());
if (mm != &init_mm) {
- switch_mm(mm, &init_mm, current);
+ mmgrab_lazy_tlb(&init_mm);
+ current->active_mm = &init_mm;
+ switch_mm_irqs_off(mm, &init_mm, current);
finish_arch_post_lock_switch();
+ mmdrop_lazy_tlb(mm);
}
+ /* finish_cpu() will mmdrop the init_mm ref after this CPU stops */
+}
+
+/*
+ * After the CPU is offline, double check that it was previously switched to
+ * init_mm. This call can be removed because the condition is caught in
+ * finish_cpu() as well.
+ */
+void idle_task_exit(void)
+{
+ BUG_ON(cpu_online(smp_processor_id()));
+ BUG_ON(current != this_rq()->idle);
- /* finish_cpu(), as ran on the BP, will clean up the active_mm state */
+ WARN_ON_ONCE(current->active_mm != &init_mm);
}
static int __balance_push_cpu_stop(void *arg)
--
2.40.1
^ permalink raw reply related [flat|nested] 2+ messages in thread* [PATCH 2/2] lazy tlb: consolidate lazy tlb mm switching
2023-05-24 6:04 [PATCH 1/2] lazy tlb: fix hotplug exit race with MMU_LAZY_TLB_SHOOTDOWN Nicholas Piggin
@ 2023-05-24 6:04 ` Nicholas Piggin
0 siblings, 0 replies; 2+ messages in thread
From: Nicholas Piggin @ 2023-05-24 6:04 UTC (permalink / raw)
To: Andrew Morton
Cc: Peter Zijlstra, linuxppc-dev, Linus Torvalds, Nicholas Piggin,
linux-mm
Switching a kernel thread using a "lazy tlb mm" to init_mm is a
relatively common sequence that is not quite trivial. Consolidate this
into a function.
This fixes a bug in do_shoot_lazy_tlb() for any arch that implements
finish_arch_post_lock_switch(). None select MMU_LAZY_TLB_SHOOTDOWN at
the moment.
Fixes: 2655421ae69fa ("lazy tlb: shoot lazies, non-refcounting lazy tlb mm reference handling scheme")
Signed-off-by: Nicholas Piggin <npiggin@gmail.com>
---
arch/powerpc/mm/book3s64/radix_tlb.c | 6 +----
include/linux/sched/task.h | 2 ++
kernel/fork.c | 7 ++----
kernel/sched/core.c | 34 ++++++++++++++++++++--------
4 files changed, 29 insertions(+), 20 deletions(-)
diff --git a/arch/powerpc/mm/book3s64/radix_tlb.c b/arch/powerpc/mm/book3s64/radix_tlb.c
index ce804b7bf84e..90953cf9f648 100644
--- a/arch/powerpc/mm/book3s64/radix_tlb.c
+++ b/arch/powerpc/mm/book3s64/radix_tlb.c
@@ -795,12 +795,8 @@ void exit_lazy_flush_tlb(struct mm_struct *mm, bool always_flush)
goto out;
if (current->active_mm == mm) {
- WARN_ON_ONCE(current->mm != NULL);
/* Is a kernel thread and is using mm as the lazy tlb */
- mmgrab_lazy_tlb(&init_mm);
- current->active_mm = &init_mm;
- switch_mm_irqs_off(mm, &init_mm, current);
- mmdrop_lazy_tlb(mm);
+ kthread_end_lazy_tlb_mm();
}
/*
diff --git a/include/linux/sched/task.h b/include/linux/sched/task.h
index 537cbf9a2ade..23693b94a09b 100644
--- a/include/linux/sched/task.h
+++ b/include/linux/sched/task.h
@@ -61,6 +61,8 @@ extern int lockdep_tasklist_lock_is_held(void);
extern asmlinkage void schedule_tail(struct task_struct *prev);
extern void init_idle(struct task_struct *idle, int cpu);
+extern void kthread_end_lazy_tlb_mm(void);
+
extern int sched_fork(unsigned long clone_flags, struct task_struct *p);
extern void sched_cgroup_fork(struct task_struct *p, struct kernel_clone_args *kargs);
extern void sched_post_fork(struct task_struct *p);
diff --git a/kernel/fork.c b/kernel/fork.c
index ed4e01daccaa..8b005c2c7c3c 100644
--- a/kernel/fork.c
+++ b/kernel/fork.c
@@ -854,11 +854,8 @@ static void do_shoot_lazy_tlb(void *arg)
{
struct mm_struct *mm = arg;
- if (current->active_mm == mm) {
- WARN_ON_ONCE(current->mm);
- current->active_mm = &init_mm;
- switch_mm(mm, &init_mm, current);
- }
+ if (current->active_mm == mm)
+ kthread_end_lazy_tlb_mm();
}
static void cleanup_lazy_tlbs(struct mm_struct *mm)
diff --git a/kernel/sched/core.c b/kernel/sched/core.c
index bc4ef1f3394b..71706df22b41 100644
--- a/kernel/sched/core.c
+++ b/kernel/sched/core.c
@@ -5346,6 +5346,29 @@ context_switch(struct rq *rq, struct task_struct *prev,
return finish_task_switch(prev);
}
+/*
+ * If this kthread has a user process's mm for its active_mm (aka lazy tlb mm)
+ * then switch away from it, to init_mm. Must not be called while using an
+ * mm with kthread_use_mm().
+ */
+void kthread_end_lazy_tlb_mm(void)
+{
+ struct mm_struct *mm = current->active_mm;
+
+ WARN_ON_ONCE(!irqs_disabled());
+
+ if (WARN_ON_ONCE(current->mm))
+ return; /* Not a kthread or doing kthread_use_mm */
+
+ if (mm != &init_mm) {
+ mmgrab_lazy_tlb(&init_mm);
+ current->active_mm = &init_mm;
+ switch_mm_irqs_off(mm, &init_mm, current);
+ finish_arch_post_lock_switch();
+ mmdrop_lazy_tlb(mm);
+ }
+}
+
/*
* nr_running and nr_context_switches:
*
@@ -9375,17 +9398,8 @@ void sched_setnuma(struct task_struct *p, int nid)
*/
void idle_task_prepare_exit(void)
{
- struct mm_struct *mm = current->active_mm;
-
WARN_ON(!irqs_disabled());
-
- if (mm != &init_mm) {
- mmgrab_lazy_tlb(&init_mm);
- current->active_mm = &init_mm;
- switch_mm_irqs_off(mm, &init_mm, current);
- finish_arch_post_lock_switch();
- mmdrop_lazy_tlb(mm);
- }
+ kthread_end_lazy_tlb_mm();
/* finish_cpu() will mmdrop the init_mm ref after this CPU stops */
}
--
2.40.1
^ permalink raw reply related [flat|nested] 2+ messages in thread
end of thread, other threads:[~2023-05-24 6:06 UTC | newest]
Thread overview: 2+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2023-05-24 6:04 [PATCH 1/2] lazy tlb: fix hotplug exit race with MMU_LAZY_TLB_SHOOTDOWN Nicholas Piggin
2023-05-24 6:04 ` [PATCH 2/2] lazy tlb: consolidate lazy tlb mm switching Nicholas Piggin
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).