* [PATCH v4 01/31] ARM: Prepare includes for generic entry
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
@ 2025-02-12 11:22 ` Linus Walleij
2025-02-12 11:22 ` [PATCH v4 02/31] ARM: ptrace: Split report_syscall() Linus Walleij
` (30 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:22 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
The generic entry code needs a static inline function for
regs_irq_disabled() so add this, in further requires the signature
of on_thread_stack() to be __always_inline so add this as
well.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/include/asm/ptrace.h | 5 +++++
arch/arm/include/asm/stacktrace.h | 2 +-
2 files changed, 6 insertions(+), 1 deletion(-)
diff --git a/arch/arm/include/asm/ptrace.h b/arch/arm/include/asm/ptrace.h
index 6eb311fb2da06fa393f8be7caec8d997637a88b2..a01c66f0ad907882af2a383e53de8aca68c551e5 100644
--- a/arch/arm/include/asm/ptrace.h
+++ b/arch/arm/include/asm/ptrace.h
@@ -52,6 +52,11 @@ struct svc_pt_regs {
#define fast_interrupts_enabled(regs) \
(!((regs)->ARM_cpsr & PSR_F_BIT))
+static inline int regs_irqs_disabled(struct pt_regs *regs)
+{
+ return !interrupts_enabled(regs);
+}
+
/* Are the current registers suitable for user mode?
* (used to maintain security in signal handlers)
*/
diff --git a/arch/arm/include/asm/stacktrace.h b/arch/arm/include/asm/stacktrace.h
index f80a85b091d6c4ff365e15ae7100af1c3aed597f..815b5f256af164daa053f3bfd6cb2f54042a8d2b 100644
--- a/arch/arm/include/asm/stacktrace.h
+++ b/arch/arm/include/asm/stacktrace.h
@@ -26,7 +26,7 @@ struct stackframe {
#endif
};
-static inline bool on_thread_stack(void)
+static __always_inline bool on_thread_stack(void)
{
unsigned long delta = current_stack_pointer ^ (unsigned long)current->stack;
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 02/31] ARM: ptrace: Split report_syscall()
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
2025-02-12 11:22 ` [PATCH v4 01/31] ARM: Prepare includes for " Linus Walleij
@ 2025-02-12 11:22 ` Linus Walleij
2025-02-12 11:22 ` [PATCH v4 03/31] ARM: entry: Skip ret_slow_syscall label Linus Walleij
` (29 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:22 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
The generic entry code requires that report_syscall() be provided
in two parts: report_syscall_enter() and report_syscall_exit()
so split the combined function in two.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/kernel/ptrace.c | 30 ++++++++++++++++++++++--------
1 file changed, 22 insertions(+), 8 deletions(-)
diff --git a/arch/arm/kernel/ptrace.c b/arch/arm/kernel/ptrace.c
index c421a899fc84c476a672cbfe1c8ece8355d8512d..07b0daf47441f1f76a8af416acc74fa5ed770403 100644
--- a/arch/arm/kernel/ptrace.c
+++ b/arch/arm/kernel/ptrace.c
@@ -827,31 +827,45 @@ enum ptrace_syscall_dir {
PTRACE_SYSCALL_EXIT,
};
-static void report_syscall(struct pt_regs *regs, enum ptrace_syscall_dir dir)
+static void report_syscall_enter(struct pt_regs *regs)
{
unsigned long ip;
/*
* IP is used to denote syscall entry/exit:
- * IP = 0 -> entry, =1 -> exit
+ * IP = 0 -> entry
*/
ip = regs->ARM_ip;
- regs->ARM_ip = dir;
+ regs->ARM_ip = PTRACE_SYSCALL_ENTER;
- if (dir == PTRACE_SYSCALL_EXIT)
- ptrace_report_syscall_exit(regs, 0);
- else if (ptrace_report_syscall_entry(regs))
+ if (ptrace_report_syscall_entry(regs))
current_thread_info()->abi_syscall = -1;
regs->ARM_ip = ip;
}
+static void report_syscall_exit(struct pt_regs *regs)
+{
+ unsigned long ip;
+
+ /*
+ * IP is used to denote syscall entry/exit:
+ * IP = 1 -> exit
+ */
+ ip = regs->ARM_ip;
+ regs->ARM_ip = PTRACE_SYSCALL_EXIT;
+
+ ptrace_report_syscall_exit(regs, 0);
+
+ regs->ARM_ip = ip;
+}
+
asmlinkage int syscall_trace_enter(struct pt_regs *regs)
{
int scno;
if (test_thread_flag(TIF_SYSCALL_TRACE))
- report_syscall(regs, PTRACE_SYSCALL_ENTER);
+ report_syscall_enter(regs);
/* Do seccomp after ptrace; syscall may have changed. */
#ifdef CONFIG_HAVE_ARCH_SECCOMP_FILTER
@@ -892,5 +906,5 @@ asmlinkage void syscall_trace_exit(struct pt_regs *regs)
trace_sys_exit(regs, regs_return_value(regs));
if (test_thread_flag(TIF_SYSCALL_TRACE))
- report_syscall(regs, PTRACE_SYSCALL_EXIT);
+ report_syscall_exit(regs);
}
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 03/31] ARM: entry: Skip ret_slow_syscall label
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
2025-02-12 11:22 ` [PATCH v4 01/31] ARM: Prepare includes for " Linus Walleij
2025-02-12 11:22 ` [PATCH v4 02/31] ARM: ptrace: Split report_syscall() Linus Walleij
@ 2025-02-12 11:22 ` Linus Walleij
2025-02-12 11:22 ` [PATCH v4 04/31] ARM: process: Rewrite ret_from_fork i C Linus Walleij
` (28 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:22 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
Don't introduce this extra label anymore now that we have
a clearly defined entry point for ret_to_user.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/kernel/entry-common.S | 7 +++----
1 file changed, 3 insertions(+), 4 deletions(-)
diff --git a/arch/arm/kernel/entry-common.S b/arch/arm/kernel/entry-common.S
index f379c852dcb72fa5e960bba5621cc2152f0df0d3..2f93c2f6d5b8fdf6aaf3dda0192bde144fcf238d 100644
--- a/arch/arm/kernel/entry-common.S
+++ b/arch/arm/kernel/entry-common.S
@@ -102,7 +102,6 @@ ENDPROC(ret_fast_syscall)
* do_work_pending() will update this state if necessary.
*/
ENTRY(ret_to_user)
-ret_slow_syscall:
#if IS_ENABLED(CONFIG_DEBUG_RSEQ)
/* do_rseq_syscall needs interrupts enabled. */
enable_irq_notrace @ enable interrupts
@@ -136,7 +135,7 @@ ENTRY(ret_from_fork)
badrne lr, 1f
retne r5
1: get_thread_info tsk
- b ret_slow_syscall
+ b ret_to_user
ENDPROC(ret_from_fork)
/*=============================================================================
@@ -310,13 +309,13 @@ __sys_trace_return_nosave:
enable_irq_notrace
mov r0, sp
bl syscall_trace_exit
- b ret_slow_syscall
+ b ret_to_user
__sys_trace_return:
str r0, [sp, #S_R0 + S_OFF]! @ save returned r0
mov r0, sp
bl syscall_trace_exit
- b ret_slow_syscall
+ b ret_to_user
.macro syscall_table_start, sym
.equ __sys_nr, 0
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 04/31] ARM: process: Rewrite ret_from_fork i C
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (2 preceding siblings ...)
2025-02-12 11:22 ` [PATCH v4 03/31] ARM: entry: Skip ret_slow_syscall label Linus Walleij
@ 2025-02-12 11:22 ` Linus Walleij
2025-02-12 11:22 ` [PATCH v4 05/31] ARM: process: Remove local restart Linus Walleij
` (27 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:22 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
Move the direct assembly call to schedule_tail() and the
call starting a kernel thread into a C version of ret_from_fork()
in process.c and rename the old ret_from_fork to ret_from_fork_asm
following the pattern in x86.
Leave a comment on what we will do when we get to switching to
generic entry with this as a base.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/include/asm/switch_to.h | 4 ++++
arch/arm/kernel/entry-common.S | 22 +++++++++++++++-------
arch/arm/kernel/process.c | 24 ++++++++++++++++++++++--
3 files changed, 41 insertions(+), 9 deletions(-)
diff --git a/arch/arm/include/asm/switch_to.h b/arch/arm/include/asm/switch_to.h
index 9372348516ce9b80fa713966943d0bc622e86066..18d084aff5a4faa7cc5e40d48fbc4aa97a89b0c6 100644
--- a/arch/arm/include/asm/switch_to.h
+++ b/arch/arm/include/asm/switch_to.h
@@ -17,6 +17,10 @@
#define __complete_pending_tlbi()
#endif
+asmlinkage void ret_from_fork_asm(void);
+__visible void ret_from_fork(struct task_struct *prev, struct pt_regs *regs,
+ int (*fn)(void *), void *fn_arg);
+
/*
* switch_to(prev, next) should switch from task `prev' to `next'
* `prev' will never be the same as `next'. schedule() itself
diff --git a/arch/arm/kernel/entry-common.S b/arch/arm/kernel/entry-common.S
index 2f93c2f6d5b8fdf6aaf3dda0192bde144fcf238d..df6961a1006b7a0f3522728812bf52ccb849e511 100644
--- a/arch/arm/kernel/entry-common.S
+++ b/arch/arm/kernel/entry-common.S
@@ -5,6 +5,7 @@
* Copyright (C) 2000 Russell King
*/
+#include <linux/cfi_types.h>
#include <asm/assembler.h>
#include <asm/unistd.h>
#include <asm/ftrace.h>
@@ -128,15 +129,22 @@ ENDPROC(ret_to_user)
/*
* This is how we return from a fork.
*/
-ENTRY(ret_from_fork)
- bl schedule_tail
- cmp r5, #0
- movne r0, r4
- badrne lr, 1f
- retne r5
+SYM_TYPED_FUNC_START(ret_from_fork_asm)
+ /*
+ * ret_from_fork() expects:
+ * r0: prev
+ * r1: regs
+ * r2: fn() for kernel process or NULL
+ * r3: fn_arg for kernel process or N/A
+ */
+ mov r1, sp
+ mov r2, r5
+ mov r3, r4
+ bl ret_from_fork
+ mov r0, #0
1: get_thread_info tsk
b ret_to_user
-ENDPROC(ret_from_fork)
+SYM_FUNC_END(ret_from_fork_asm)
/*=============================================================================
* SWI handler
diff --git a/arch/arm/kernel/process.c b/arch/arm/kernel/process.c
index e16ed102960cb01d625e2ccc484f238b824b03b3..81c1c3f988344185917cf4e53d0d8ee47ae912f4 100644
--- a/arch/arm/kernel/process.c
+++ b/arch/arm/kernel/process.c
@@ -29,6 +29,7 @@
#include <asm/processor.h>
#include <asm/thread_notify.h>
#include <asm/stacktrace.h>
+#include <asm/switch_to.h>
#include <asm/system_misc.h>
#include <asm/mach/time.h>
#include <asm/tls.h>
@@ -230,7 +231,26 @@ void flush_thread(void)
thread_notify(THREAD_NOTIFY_FLUSH, thread);
}
-asmlinkage void ret_from_fork(void) __asm__("ret_from_fork");
+/* The assembly stub sets up the parameters and calls ret_from_fork() */
+asmlinkage void ret_from_fork_asm(void) __asm__("ret_from_fork_asm");
+
+__visible void ret_from_fork(struct task_struct *prev, struct pt_regs *regs,
+ int (*fn)(void *), void *fn_arg)
+{
+ schedule_tail(prev);
+
+ /* Is this a kernel thread? */
+ if (unlikely(fn)) {
+ fn(fn_arg);
+ /*
+ * A kernel thread is allowed to return here after successfully
+ * calling kernel_execve(). Exit to userspace to complete the
+ * execve() syscall.
+ */
+ }
+
+ /* syscall_exit_to_user_mode(regs); here once we switch to generic entry */
+}
int copy_thread(struct task_struct *p, const struct kernel_clone_args *args)
{
@@ -263,7 +283,7 @@ int copy_thread(struct task_struct *p, const struct kernel_clone_args *args)
thread->cpu_context.r5 = (unsigned long)args->fn;
childregs->ARM_cpsr = SVC_MODE;
}
- thread->cpu_context.pc = (unsigned long)ret_from_fork;
+ thread->cpu_context.pc = (unsigned long)ret_from_fork_asm;
thread->cpu_context.sp = (unsigned long)childregs;
clear_ptrace_hw_breakpoint(p);
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 05/31] ARM: process: Remove local restart
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (3 preceding siblings ...)
2025-02-12 11:22 ` [PATCH v4 04/31] ARM: process: Rewrite ret_from_fork i C Linus Walleij
@ 2025-02-12 11:22 ` Linus Walleij
2025-02-12 11:23 ` [PATCH v4 06/31] ARM: entry: Invoke syscalls using C Linus Walleij
` (26 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:22 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
The ARM kernel contains a quirk to handle syscall restarts
inside the kernel without exiting to userspace. The generic
entry cannot handle this.
Rename do_signal() to arch_do_signal_or_restart() to fit
with the upcoming generic entry conversion.
This is essentially a revert of commit 81783786d5cf
"ARM: 7473/1: deal with handlerless restarts without leaving the kernel"
from 2012.
Other solutions may be possible, such as checking the PC after
do_work_pending and assume it is a local restart if that address
is inside the kernel, or modifying the generic entry code to
track local restarts like ARM does and pass that information
back.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/include/asm/signal.h | 3 +--
arch/arm/kernel/entry-common.S | 9 +--------
arch/arm/kernel/signal.c | 28 +++++++++-------------------
3 files changed, 11 insertions(+), 29 deletions(-)
diff --git a/arch/arm/include/asm/signal.h b/arch/arm/include/asm/signal.h
index 8b84092d151800cb4076672d1c31f0a83777d0d8..7acccc96840c8a17744cc3f2894f19d142aba4fd 100644
--- a/arch/arm/include/asm/signal.h
+++ b/arch/arm/include/asm/signal.h
@@ -24,7 +24,6 @@ typedef struct {
#include <asm/sigcontext.h>
void do_rseq_syscall(struct pt_regs *regs);
-int do_work_pending(struct pt_regs *regs, unsigned int thread_flags,
- int syscall);
+void do_work_pending(struct pt_regs *regs, unsigned int thread_flags);
#endif
diff --git a/arch/arm/kernel/entry-common.S b/arch/arm/kernel/entry-common.S
index df6961a1006b7a0f3522728812bf52ccb849e511..da5c2d4b62e56cce644b15e329ab4149d69c5b79 100644
--- a/arch/arm/kernel/entry-common.S
+++ b/arch/arm/kernel/entry-common.S
@@ -86,14 +86,8 @@ ENDPROC(ret_fast_syscall)
bne __sys_trace_return_nosave
slow_work_pending:
mov r0, sp @ 'regs'
- mov r2, why @ 'syscall'
bl do_work_pending
- cmp r0, #0
- beq no_work_pending
- movlt scno, #(__NR_restart_syscall - __NR_SYSCALL_BASE)
- str scno, [tsk, #TI_ABI_SYSCALL] @ make sure tracers see update
- ldmia sp, {r0 - r6} @ have to reload r0 - r6
- b local_restart @ ... and off we go
+ b no_work_pending
ENDPROC(ret_fast_syscall)
/*
@@ -266,7 +260,6 @@ ENTRY(vector_swi)
*/
TRACE( ldmia sp, {r0 - r3} )
-local_restart:
ldr r10, [tsk, #TI_FLAGS] @ check for syscall tracing
stmdb sp!, {r4, r5} @ push fifth and sixth args
diff --git a/arch/arm/kernel/signal.c b/arch/arm/kernel/signal.c
index 79a6730fa0eb7d6581343bdbbeb6834dcaabb9f8..7b1a16e86b236575efdc29cda9b751e8e2a3f64e 100644
--- a/arch/arm/kernel/signal.c
+++ b/arch/arm/kernel/signal.c
@@ -18,6 +18,7 @@
#include <asm/traps.h>
#include <asm/unistd.h>
#include <asm/vfp.h>
+#include <asm/syscall.h>
#include <asm/syscalls.h>
#include "signal.h"
@@ -534,9 +535,10 @@ static void handle_signal(struct ksignal *ksig, struct pt_regs *regs)
* the kernel can handle, and then we build all the user-level signal handling
* stack-frames in one go after that.
*/
-static int do_signal(struct pt_regs *regs, int syscall)
+static void arch_do_signal_or_restart(struct pt_regs *regs)
{
unsigned int retval = 0, continue_addr = 0, restart_addr = 0;
+ bool syscall = (syscall_get_nr(current, regs) != -1);
struct ksignal ksig;
int restart = 0;
@@ -590,16 +592,14 @@ static int do_signal(struct pt_regs *regs, int syscall)
} else {
/* no handler */
restore_saved_sigmask();
- if (unlikely(restart) && regs->ARM_pc == restart_addr) {
+ if (unlikely(restart) && regs->ARM_pc == restart_addr)
regs->ARM_pc = continue_addr;
- return restart;
- }
}
- return 0;
+ return;
}
-asmlinkage int
-do_work_pending(struct pt_regs *regs, unsigned int thread_flags, int syscall)
+asmlinkage void
+do_work_pending(struct pt_regs *regs, unsigned int thread_flags)
{
/*
* The assembly code enters us with IRQs off, but it hasn't
@@ -612,19 +612,10 @@ do_work_pending(struct pt_regs *regs, unsigned int thread_flags, int syscall)
schedule();
} else {
if (unlikely(!user_mode(regs)))
- return 0;
+ return;
local_irq_enable();
if (thread_flags & (_TIF_SIGPENDING | _TIF_NOTIFY_SIGNAL)) {
- int restart = do_signal(regs, syscall);
- if (unlikely(restart)) {
- /*
- * Restart without handlers.
- * Deal with it without leaving
- * the kernel space.
- */
- return restart;
- }
- syscall = 0;
+ arch_do_signal_or_restart(regs);
} else if (thread_flags & _TIF_UPROBE) {
uprobe_notify_resume(regs);
} else {
@@ -634,7 +625,6 @@ do_work_pending(struct pt_regs *regs, unsigned int thread_flags, int syscall)
local_irq_disable();
thread_flags = read_thread_flags();
} while (thread_flags & _TIF_WORK_MASK);
- return 0;
}
struct page *get_signal_page(void)
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 06/31] ARM: entry: Invoke syscalls using C
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (4 preceding siblings ...)
2025-02-12 11:22 ` [PATCH v4 05/31] ARM: process: Remove local restart Linus Walleij
@ 2025-02-12 11:23 ` Linus Walleij
2025-02-12 11:23 ` [PATCH v4 07/31] ARM: entry: Rewrite two asm calls in C Linus Walleij
` (25 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:23 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
This makes the assembly invoking syscalls switch over to
doing this from C with a small assembly stub to actually
jump into the syscall.
Split the syscall invocation into two paths: one for plain
invocation and one for tracing (also known as "reload"
as it was reloading the registers from regs). We rename
this path with the infix "trace" as that code will be
trace-specific as we move code over to C.
Some registers such as r1 and lr get cobbled during the
C calls and need to be restored when we return.
Right now the part in C doesn't do much more than check the
syscall number to be valid (a test previously done with
a cmp r #NR_syscalls inside the invoke_syscall macro)
but we will gradually factor over more assembly to C that can
then be switched to the generic entry code so the exercise
gets a point.
Since both functions are called from a C file that is in
turn only called from assembly, the functions have to be
tagged with __ADDRESSABLE() in order not to upset CFI.
Tested with a full system boot and by issuing some command
line tools with strace to make sure the tracing path still
works.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/include/asm/syscall.h | 3 ++
arch/arm/kernel/Makefile | 3 +-
arch/arm/kernel/entry-common.S | 65 ++++++++++++++++++++++++++++++++++++------
| 25 ----------------
arch/arm/kernel/syscall.c | 28 ++++++++++++++++++
5 files changed, 90 insertions(+), 34 deletions(-)
diff --git a/arch/arm/include/asm/syscall.h b/arch/arm/include/asm/syscall.h
index fe4326d938c18efb54c96930e6b52ab7da9b2942..1d21f26ecf510ac00a878b4d51a5753ee49475f5 100644
--- a/arch/arm/include/asm/syscall.h
+++ b/arch/arm/include/asm/syscall.h
@@ -19,6 +19,9 @@
extern const unsigned long sys_call_table[];
+int invoke_syscall(void *table, struct pt_regs *regs, int scno, void *retp);
+int invoke_syscall_trace(void *table, struct pt_regs *regs, int scno, void *retp);
+
static inline int syscall_get_nr(struct task_struct *task,
struct pt_regs *regs)
{
diff --git a/arch/arm/kernel/Makefile b/arch/arm/kernel/Makefile
index b3333d070390a8ff366a306931614b1b260647bd..69ddf51081f4791982518d37df60c11211b8955a 100644
--- a/arch/arm/kernel/Makefile
+++ b/arch/arm/kernel/Makefile
@@ -20,7 +20,8 @@ CFLAGS_REMOVE_return_address.o = -pg
obj-y := elf.o entry-common.o irq.o opcodes.o \
process.o ptrace.o reboot.o io.o \
setup.o signal.o sigreturn_codes.o \
- stacktrace.o sys_arm.o time.o traps.o
+ stacktrace.o sys_arm.o time.o traps.o \
+ syscall.o
KASAN_SANITIZE_stacktrace.o := n
KASAN_SANITIZE_traps.o := n
diff --git a/arch/arm/kernel/entry-common.S b/arch/arm/kernel/entry-common.S
index da5c2d4b62e56cce644b15e329ab4149d69c5b79..57aa1084a0476e9c218cd100ce4fdf4aaa748234 100644
--- a/arch/arm/kernel/entry-common.S
+++ b/arch/arm/kernel/entry-common.S
@@ -254,21 +254,23 @@ ENTRY(vector_swi)
#else
str scno, [tsk, #TI_ABI_SYSCALL]
#endif
- /*
- * Reload the registers that may have been corrupted on entry to
- * the syscall assembly (by tracing or context tracking.)
- */
- TRACE( ldmia sp, {r0 - r3} )
-
+ mov r1, sp @ put regs into r1
ldr r10, [tsk, #TI_FLAGS] @ check for syscall tracing
stmdb sp!, {r4, r5} @ push fifth and sixth args
tst r10, #_TIF_SYSCALL_WORK @ are we tracing syscalls?
bne __sys_trace
- invoke_syscall tbl, scno, r10, __ret_fast_syscall
+ mov r0, tbl
+ /* r1 already contains regs */
+ mov r2, scno @ syscall number from r7
+ badr r3, __ret_fast_syscall
+ bl invoke_syscall
+ /* Restore regs into r1 and lr after C call */
+ badr lr, __ret_fast_syscall
add r1, sp, #S_OFF
+
2: cmp scno, #(__ARM_NR_BASE - __NR_SYSCALL_BASE)
eor r0, scno, #__NR_SYSCALL_BASE @ put OS number back
bcs arm_syscall
@@ -301,7 +303,16 @@ __sys_trace:
add r0, sp, #S_OFF
bl syscall_trace_enter
mov scno, r0
- invoke_syscall tbl, scno, r10, __sys_trace_return, reload=1
+ mov r2, r0 @ scno into r2
+ add r1, sp, #S_R0 + S_OFF @ pointer to regs
+ mov r0, tbl
+ badr r3, __sys_trace_return
+ bl invoke_syscall_trace
+
+ /* Restore regs into r1 and lr after C call */
+ add r1, sp, #S_R0 + S_OFF @ pointer to regs
+ badr lr, __sys_trace_return
+
cmp scno, #-1 @ skip the syscall?
bne 2b
add sp, sp, #S_OFF @ restore stack
@@ -415,6 +426,44 @@ sys_mmap2:
b sys_mmap_pgoff
ENDPROC(sys_mmap2)
+/*
+ * This call wants:
+ * r0: syscall table
+ * r1: regs
+ * r2: syscall number
+ * r3: pointer to return function
+ */
+SYM_TYPED_FUNC_START(invoke_syscall_asm)
+#ifdef CONFIG_CPU_SPECTRE
+ csdb
+#endif
+ mov tbl, r0
+ mov scno, r2
+ mov lr, r3 @ return address
+ ldmia r1, {r0 - r3} @ reload r0-r3
+ /* Arguments 5 and 6 are (hopefully) on the stack */
+ ldr pc, [tbl, scno, lsl #2] @ call sys_* routine
+SYM_FUNC_END(invoke_syscall_asm)
+
+/*
+ * This call wants:
+ * r0: syscall table
+ * r1: regs
+ * r2: syscall number
+ * r3: pointer to return function
+ */
+SYM_TYPED_FUNC_START(invoke_syscall_trace_asm)
+#ifdef CONFIG_CPU_SPECTRE
+ csdb
+#endif
+ mov tbl, r0
+ mov scno, r2
+ mov lr, r3 @ return address
+ ldmia r1, {r0 - r6} @ reload r0-r6
+ stmia sp, {r4, r5} @ update stack arguments
+ ldr pc, [tbl, scno, lsl #2] @ call sys_* routine
+SYM_FUNC_END(invoke_syscall_trace_asm)
+
#ifdef CONFIG_OABI_COMPAT
/*
--git a/arch/arm/kernel/entry-header.S b/arch/arm/kernel/entry-header.S
index 99411fa913501cad8b55ef04a2b8ab3d44f3e39e..52b4fa97226dbfa4b55aca8eaf74ae24e1e220f8 100644
--- a/arch/arm/kernel/entry-header.S
+++ b/arch/arm/kernel/entry-header.S
@@ -389,31 +389,6 @@ ALT_UP_B(.L1_\@)
#endif
.endm
- .macro invoke_syscall, table, nr, tmp, ret, reload=0
-#ifdef CONFIG_CPU_SPECTRE
- mov \tmp, \nr
- cmp \tmp, #NR_syscalls @ check upper syscall limit
- movcs \tmp, #0
- csdb
- badr lr, \ret @ return address
- .if \reload
- add r1, sp, #S_R0 + S_OFF @ pointer to regs
- ldmiacc r1, {r0 - r6} @ reload r0-r6
- stmiacc sp, {r4, r5} @ update stack arguments
- .endif
- ldrcc pc, [\table, \tmp, lsl #2] @ call sys_* routine
-#else
- cmp \nr, #NR_syscalls @ check upper syscall limit
- badr lr, \ret @ return address
- .if \reload
- add r1, sp, #S_R0 + S_OFF @ pointer to regs
- ldmiacc r1, {r0 - r6} @ reload r0-r6
- stmiacc sp, {r4, r5} @ update stack arguments
- .endif
- ldrcc pc, [\table, \nr, lsl #2] @ call sys_* routine
-#endif
- .endm
-
/*
* These are the registers used in the syscall handler, and allow us to
* have in theory up to 7 arguments to a function - r0 to r6.
diff --git a/arch/arm/kernel/syscall.c b/arch/arm/kernel/syscall.c
new file mode 100644
index 0000000000000000000000000000000000000000..edc0ac88ec60ce3f23149a526de7dc1205906552
--- /dev/null
+++ b/arch/arm/kernel/syscall.c
@@ -0,0 +1,28 @@
+// SPDX-License-Identifier: GPL-2.0
+
+#include <linux/syscalls.h>
+#include <asm/syscall.h>
+
+int invoke_syscall_asm(void *table, struct pt_regs *regs, int scno, void *retp);
+__ADDRESSABLE(invoke_syscall_asm);
+
+__visible int invoke_syscall(void *table, struct pt_regs *regs, int scno, void *retp)
+{
+ if (scno < NR_syscalls)
+ /* Doing this with return makes sure the stack gets pop:ed */
+ return invoke_syscall_asm(table, regs, scno, retp);
+
+ return 0;
+}
+
+int invoke_syscall_trace_asm(void *table, struct pt_regs *regs, int scno, void *retp);
+__ADDRESSABLE(invoke_syscall_trace_asm);
+
+__visible int invoke_syscall_trace(void *table, struct pt_regs *regs, int scno, void *retp)
+{
+ if (scno < NR_syscalls)
+ /* Doing this with return makes sure the stack gets pop:ed */
+ return invoke_syscall_trace_asm(table, regs, scno, retp);
+
+ return 0;
+}
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 07/31] ARM: entry: Rewrite two asm calls in C
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (5 preceding siblings ...)
2025-02-12 11:23 ` [PATCH v4 06/31] ARM: entry: Invoke syscalls using C Linus Walleij
@ 2025-02-12 11:23 ` Linus Walleij
2025-02-12 11:23 ` [PATCH v4 08/31] ARM: entry: Move trace entry to C function Linus Walleij
` (24 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:23 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
The normal and trace entry code calls out to arm_syscall()
and sys_ni_syscall() from assembly, but these calls can
be moved over to the new C implementation.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/include/asm/traps.h | 2 +-
arch/arm/kernel/entry-common.S | 23 +++++------------------
arch/arm/kernel/syscall.c | 13 +++++++++++--
arch/arm/kernel/traps.c | 2 +-
4 files changed, 18 insertions(+), 22 deletions(-)
diff --git a/arch/arm/include/asm/traps.h b/arch/arm/include/asm/traps.h
index 2621b9fb9b19b064aa40a8ab4a3a89067b98a010..b888912c2450e0e8eec8139166c7292f283545c7 100644
--- a/arch/arm/include/asm/traps.h
+++ b/arch/arm/include/asm/traps.h
@@ -40,7 +40,7 @@ asmlinkage void dump_backtrace_stm(u32 *stack, u32 instruction, const char *logl
asmlinkage void do_undefinstr(struct pt_regs *regs);
asmlinkage void handle_fiq_as_nmi(struct pt_regs *regs);
asmlinkage void bad_mode(struct pt_regs *regs, int reason);
-asmlinkage int arm_syscall(int no, struct pt_regs *regs);
+int arm_syscall(int no, struct pt_regs *regs);
asmlinkage void baddataabort(int code, unsigned long instr, struct pt_regs *regs);
asmlinkage void __div0(void);
asmlinkage void handle_bad_stack(struct pt_regs *regs);
diff --git a/arch/arm/kernel/entry-common.S b/arch/arm/kernel/entry-common.S
index 57aa1084a0476e9c218cd100ce4fdf4aaa748234..77801d039b5f6562b231a6fd4979e318c7e736a6 100644
--- a/arch/arm/kernel/entry-common.S
+++ b/arch/arm/kernel/entry-common.S
@@ -264,18 +264,10 @@ ENTRY(vector_swi)
mov r0, tbl
/* r1 already contains regs */
mov r2, scno @ syscall number from r7
- badr r3, __ret_fast_syscall
- bl invoke_syscall
-
- /* Restore regs into r1 and lr after C call */
+ /* We return here no matter what, also pass this as an argument */
badr lr, __ret_fast_syscall
- add r1, sp, #S_OFF
-
-2: cmp scno, #(__ARM_NR_BASE - __NR_SYSCALL_BASE)
- eor r0, scno, #__NR_SYSCALL_BASE @ put OS number back
- bcs arm_syscall
- mov why, #0 @ no longer a real syscall
- b sys_ni_syscall @ not private func
+ mov r3, lr
+ b invoke_syscall
#if defined(CONFIG_OABI_COMPAT) || !defined(CONFIG_AEABI)
/*
@@ -308,13 +300,8 @@ __sys_trace:
mov r0, tbl
badr r3, __sys_trace_return
bl invoke_syscall_trace
-
- /* Restore regs into r1 and lr after C call */
- add r1, sp, #S_R0 + S_OFF @ pointer to regs
- badr lr, __sys_trace_return
-
- cmp scno, #-1 @ skip the syscall?
- bne 2b
+ cmp r0, #-1
+ bne __sys_trace_return
add sp, sp, #S_OFF @ restore stack
__sys_trace_return_nosave:
diff --git a/arch/arm/kernel/syscall.c b/arch/arm/kernel/syscall.c
index edc0ac88ec60ce3f23149a526de7dc1205906552..d637dc74b1c70415933898fbcadbedc71df5f654 100644
--- a/arch/arm/kernel/syscall.c
+++ b/arch/arm/kernel/syscall.c
@@ -12,7 +12,10 @@ __visible int invoke_syscall(void *table, struct pt_regs *regs, int scno, void *
/* Doing this with return makes sure the stack gets pop:ed */
return invoke_syscall_asm(table, regs, scno, retp);
- return 0;
+ if (scno >= __ARM_NR_BASE)
+ return arm_syscall(scno, regs);
+
+ return sys_ni_syscall();
}
int invoke_syscall_trace_asm(void *table, struct pt_regs *regs, int scno, void *retp);
@@ -20,9 +23,15 @@ __ADDRESSABLE(invoke_syscall_trace_asm);
__visible int invoke_syscall_trace(void *table, struct pt_regs *regs, int scno, void *retp)
{
+ if (scno == -1)
+ return -1;
+
if (scno < NR_syscalls)
/* Doing this with return makes sure the stack gets pop:ed */
return invoke_syscall_trace_asm(table, regs, scno, retp);
- return 0;
+ if (scno >= __ARM_NR_BASE)
+ return arm_syscall(scno, regs);
+
+ return sys_ni_syscall();
}
diff --git a/arch/arm/kernel/traps.c b/arch/arm/kernel/traps.c
index 6ea645939573fb65ed36f5435a9c94e98828f45b..942c77767919466d5b89d83d8eb77032b119473e 100644
--- a/arch/arm/kernel/traps.c
+++ b/arch/arm/kernel/traps.c
@@ -609,7 +609,7 @@ do_cache_op(unsigned long start, unsigned long end, int flags)
* 0x9f0000 - 0x9fffff are some more esoteric system calls
*/
#define NR(x) ((__ARM_NR_##x) - __ARM_NR_BASE)
-asmlinkage int arm_syscall(int no, struct pt_regs *regs)
+int arm_syscall(int no, struct pt_regs *regs)
{
if ((no >> 16) != (__ARM_NR_BASE>> 16))
return bad_syscall(no, regs);
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 08/31] ARM: entry: Move trace entry to C function
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (6 preceding siblings ...)
2025-02-12 11:23 ` [PATCH v4 07/31] ARM: entry: Rewrite two asm calls in C Linus Walleij
@ 2025-02-12 11:23 ` Linus Walleij
2025-02-12 11:23 ` [PATCH v4 09/31] ARM: entry: save the syscall sp in thread_info Linus Walleij
` (23 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:23 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
This moves over the code entering into tracing into the
C syscall invocation code. As syscall_trace_enter() will
provide the current syscall number from the regs we need
not provide the syscall number to the invocation function
any more.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/include/asm/syscall.h | 2 +-
arch/arm/kernel/entry-common.S | 6 +-----
arch/arm/kernel/syscall.c | 5 ++++-
3 files changed, 6 insertions(+), 7 deletions(-)
diff --git a/arch/arm/include/asm/syscall.h b/arch/arm/include/asm/syscall.h
index 1d21f26ecf510ac00a878b4d51a5753ee49475f5..66067d165ba3fbbe0f840a89ae396eb1a311bacc 100644
--- a/arch/arm/include/asm/syscall.h
+++ b/arch/arm/include/asm/syscall.h
@@ -20,7 +20,7 @@
extern const unsigned long sys_call_table[];
int invoke_syscall(void *table, struct pt_regs *regs, int scno, void *retp);
-int invoke_syscall_trace(void *table, struct pt_regs *regs, int scno, void *retp);
+int invoke_syscall_trace(void *table, struct pt_regs *regs, void *retp);
static inline int syscall_get_nr(struct task_struct *task,
struct pt_regs *regs)
diff --git a/arch/arm/kernel/entry-common.S b/arch/arm/kernel/entry-common.S
index 77801d039b5f6562b231a6fd4979e318c7e736a6..3cfc6d952ff99be9c4c1be4481ac3039260e3e57 100644
--- a/arch/arm/kernel/entry-common.S
+++ b/arch/arm/kernel/entry-common.S
@@ -292,13 +292,9 @@ ENDPROC(vector_swi)
* context switches, and waiting for our parent to respond.
*/
__sys_trace:
- add r0, sp, #S_OFF
- bl syscall_trace_enter
- mov scno, r0
- mov r2, r0 @ scno into r2
add r1, sp, #S_R0 + S_OFF @ pointer to regs
mov r0, tbl
- badr r3, __sys_trace_return
+ badr r2, __sys_trace_return
bl invoke_syscall_trace
cmp r0, #-1
bne __sys_trace_return
diff --git a/arch/arm/kernel/syscall.c b/arch/arm/kernel/syscall.c
index d637dc74b1c70415933898fbcadbedc71df5f654..377e16c8c53c8e809e3f8b157a889ef115d5c043 100644
--- a/arch/arm/kernel/syscall.c
+++ b/arch/arm/kernel/syscall.c
@@ -21,8 +21,11 @@ __visible int invoke_syscall(void *table, struct pt_regs *regs, int scno, void *
int invoke_syscall_trace_asm(void *table, struct pt_regs *regs, int scno, void *retp);
__ADDRESSABLE(invoke_syscall_trace_asm);
-__visible int invoke_syscall_trace(void *table, struct pt_regs *regs, int scno, void *retp)
+__visible int invoke_syscall_trace(void *table, struct pt_regs *regs, void *retp)
{
+ int scno;
+
+ scno = syscall_trace_enter(regs);
if (scno == -1)
return -1;
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 09/31] ARM: entry: save the syscall sp in thread_info
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (7 preceding siblings ...)
2025-02-12 11:23 ` [PATCH v4 08/31] ARM: entry: Move trace entry to C function Linus Walleij
@ 2025-02-12 11:23 ` Linus Walleij
2025-02-12 11:23 ` [PATCH v4 10/31] ARM: entry: move all tracing invocation to C Linus Walleij
` (22 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:23 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
We are going to rewrite the syscall handling in C, which
means that the stack used by the call code is no longer
predicatably 8 bytes (for syscall arguments r4 and r5)
but a varying number of bytes depending on how nested the
C code is.
However the current code is just assuming it can rewind
the stack by adding 8 to sp if a syscall is interrupted by
a sigreturn call.
Solve this by storing the entry sp in the per-task
struct thread_info and use that in the sigreturn wrapper
instead. We already have the thread info available in
the SWI entry and sigreturn is probably not so common
that retrieveing a pointer to thread_info should affect
anything very much.
Storing this per-task in thread_info makes the solution
SMP robust.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/include/asm/thread_info.h | 1 +
arch/arm/kernel/asm-offsets.c | 1 +
arch/arm/kernel/entry-common.S | 8 ++++++--
3 files changed, 8 insertions(+), 2 deletions(-)
diff --git a/arch/arm/include/asm/thread_info.h b/arch/arm/include/asm/thread_info.h
index 943ffcf069d29cf4a035964d20d56f7ebdd6d602..d8a45c5a10496aaf806bfeaa0353d5e8985bd6f5 100644
--- a/arch/arm/include/asm/thread_info.h
+++ b/arch/arm/include/asm/thread_info.h
@@ -67,6 +67,7 @@ struct thread_info {
__u32 cpu_domain; /* cpu domain */
struct cpu_context_save cpu_context; /* cpu context */
__u32 abi_syscall; /* ABI type and syscall nr */
+ __u32 sp_syscall; /* SP when entering syscall */
unsigned long tp_value[2]; /* TLS registers */
union fp_state fpstate __attribute__((aligned(8)));
union vfp_state vfpstate;
diff --git a/arch/arm/kernel/asm-offsets.c b/arch/arm/kernel/asm-offsets.c
index 4853875740d0fe61c6bbc32ddd9a16fa8d1fb530..c9525cbb26b73827821aa746030e56b037f49556 100644
--- a/arch/arm/kernel/asm-offsets.c
+++ b/arch/arm/kernel/asm-offsets.c
@@ -49,6 +49,7 @@ int main(void)
DEFINE(TI_CPU_DOMAIN, offsetof(struct thread_info, cpu_domain));
DEFINE(TI_CPU_SAVE, offsetof(struct thread_info, cpu_context));
DEFINE(TI_ABI_SYSCALL, offsetof(struct thread_info, abi_syscall));
+ DEFINE(TI_SP_SYSCALL, offsetof(struct thread_info, sp_syscall));
DEFINE(TI_TP_VALUE, offsetof(struct thread_info, tp_value));
DEFINE(TI_FPSTATE, offsetof(struct thread_info, fpstate));
#ifdef CONFIG_VFP
diff --git a/arch/arm/kernel/entry-common.S b/arch/arm/kernel/entry-common.S
index 3cfc6d952ff99be9c4c1be4481ac3039260e3e57..8baab7f97f59c434396f30b08ddd3029c5f9c0e5 100644
--- a/arch/arm/kernel/entry-common.S
+++ b/arch/arm/kernel/entry-common.S
@@ -232,6 +232,8 @@ ENTRY(vector_swi)
uaccess_disable tbl
get_thread_info tsk
+ /* Save a per-task copy of SP for sigreturn */
+ str sp, [tsk, #TI_SP_SYSCALL]
adr tbl, sys_call_table @ load syscall table pointer
@@ -377,13 +379,15 @@ sys_syscall:
ENDPROC(sys_syscall)
sys_sigreturn_wrapper:
- add r0, sp, #S_OFF
+ get_thread_info tsk
+ ldr r0, [tsk, #TI_SP_SYSCALL] @ read back SP
mov why, #0 @ prevent syscall restart handling
b sys_sigreturn
ENDPROC(sys_sigreturn_wrapper)
sys_rt_sigreturn_wrapper:
- add r0, sp, #S_OFF
+ get_thread_info tsk
+ ldr r0, [tsk, #TI_SP_SYSCALL] @ read back SP
mov why, #0 @ prevent syscall restart handling
b sys_rt_sigreturn
ENDPROC(sys_rt_sigreturn_wrapper)
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 10/31] ARM: entry: move all tracing invocation to C
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (8 preceding siblings ...)
2025-02-12 11:23 ` [PATCH v4 09/31] ARM: entry: save the syscall sp in thread_info Linus Walleij
@ 2025-02-12 11:23 ` Linus Walleij
2025-02-12 11:23 ` [PATCH v4 11/31] ARM: entry: Merge the common and trace entry code Linus Walleij
` (21 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:23 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
The traced invocation of syscall is rewritten in C, moving
over also the call to syscall_trace_exit() to C.
To do this we cannot have invoke_syscall_trace_asm() be a
leaf call, which is the mechanism we hithereto relied on
to make sure the stack is pulled back to the state we were
at before we called out to the code written in C.
Push the registers potentially used by C on the stack, and
make a copy of the two stack-pushed syscall arguments
on the top of the stack before invoking the syscall and
then drop these copied arguments and pop back the
registers before returning from invoke_syscall_trace_asm().
As we are now calling from and returning to C, we no
longer need to pass the regs pointer around so drop it.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/include/asm/syscall.h | 2 +-
arch/arm/kernel/entry-common.S | 32 +++++++++++++++-----------------
arch/arm/kernel/ptrace.c | 2 +-
arch/arm/kernel/syscall.c | 30 +++++++++++++++++++++---------
4 files changed, 38 insertions(+), 28 deletions(-)
diff --git a/arch/arm/include/asm/syscall.h b/arch/arm/include/asm/syscall.h
index 66067d165ba3fbbe0f840a89ae396eb1a311bacc..cb0073c4151bf70a82d15e17a95b6b6f48b245d6 100644
--- a/arch/arm/include/asm/syscall.h
+++ b/arch/arm/include/asm/syscall.h
@@ -20,7 +20,7 @@
extern const unsigned long sys_call_table[];
int invoke_syscall(void *table, struct pt_regs *regs, int scno, void *retp);
-int invoke_syscall_trace(void *table, struct pt_regs *regs, void *retp);
+void invoke_syscall_trace(void *table, struct pt_regs *regs);
static inline int syscall_get_nr(struct task_struct *task,
struct pt_regs *regs)
diff --git a/arch/arm/kernel/entry-common.S b/arch/arm/kernel/entry-common.S
index 8baab7f97f59c434396f30b08ddd3029c5f9c0e5..dbc947d301ec11bb13007cb4bb161c035ede5c10 100644
--- a/arch/arm/kernel/entry-common.S
+++ b/arch/arm/kernel/entry-common.S
@@ -296,22 +296,11 @@ ENDPROC(vector_swi)
__sys_trace:
add r1, sp, #S_R0 + S_OFF @ pointer to regs
mov r0, tbl
- badr r2, __sys_trace_return
bl invoke_syscall_trace
- cmp r0, #-1
- bne __sys_trace_return
- add sp, sp, #S_OFF @ restore stack
-
-__sys_trace_return_nosave:
- enable_irq_notrace
- mov r0, sp
- bl syscall_trace_exit
+ add sp, sp, #S_OFF @ restore stack pointer
b ret_to_user
-__sys_trace_return:
- str r0, [sp, #S_R0 + S_OFF]! @ save returned r0
- mov r0, sp
- bl syscall_trace_exit
+__sys_trace_return_nosave:
b ret_to_user
.macro syscall_table_start, sym
@@ -437,18 +426,27 @@ SYM_FUNC_END(invoke_syscall_asm)
* r0: syscall table
* r1: regs
* r2: syscall number
- * r3: pointer to return function
*/
SYM_TYPED_FUNC_START(invoke_syscall_trace_asm)
#ifdef CONFIG_CPU_SPECTRE
csdb
#endif
+ /* Save registers because we are being called from C */
+ push {r4 - r10, lr}
mov tbl, r0
+ /* Make space to copy the two syscall stack arguments */
+ sub sp, sp, #S_OFF
mov scno, r2
- mov lr, r3 @ return address
- ldmia r1, {r0 - r6} @ reload r0-r6
- stmia sp, {r4, r5} @ update stack arguments
+ badr lr, __invoke_syscall_trace_ret @ return right here
+ ldmia r1, {r0 - r6} @ reload r0-r6 from regs
+ stmia sp, {r4, r5} @ copy stack arguments
ldr pc, [tbl, scno, lsl #2] @ call sys_* routine
+__invoke_syscall_trace_ret:
+ /* Drop the copied stack arguments */
+ add sp, sp, #S_OFF
+ pop {r4 - r10, lr}
+ ARM( mov pc, lr )
+ THUMB( bx lr )
SYM_FUNC_END(invoke_syscall_trace_asm)
#ifdef CONFIG_OABI_COMPAT
diff --git a/arch/arm/kernel/ptrace.c b/arch/arm/kernel/ptrace.c
index 07b0daf47441f1f76a8af416acc74fa5ed770403..ac7b98ae47249b4a00fc1cb871bb2d309bfd8c88 100644
--- a/arch/arm/kernel/ptrace.c
+++ b/arch/arm/kernel/ptrace.c
@@ -888,7 +888,7 @@ asmlinkage int syscall_trace_enter(struct pt_regs *regs)
return scno;
}
-asmlinkage void syscall_trace_exit(struct pt_regs *regs)
+void syscall_trace_exit(struct pt_regs *regs)
{
/*
* Audit the syscall before anything else, as a debugger may
diff --git a/arch/arm/kernel/syscall.c b/arch/arm/kernel/syscall.c
index 377e16c8c53c8e809e3f8b157a889ef115d5c043..e90f04f35b3485968a1e9046cf0889f56cca92e5 100644
--- a/arch/arm/kernel/syscall.c
+++ b/arch/arm/kernel/syscall.c
@@ -18,23 +18,35 @@ __visible int invoke_syscall(void *table, struct pt_regs *regs, int scno, void *
return sys_ni_syscall();
}
-int invoke_syscall_trace_asm(void *table, struct pt_regs *regs, int scno, void *retp);
+int invoke_syscall_trace_asm(void *table, struct pt_regs *regs, int scno);
__ADDRESSABLE(invoke_syscall_trace_asm);
-__visible int invoke_syscall_trace(void *table, struct pt_regs *regs, void *retp)
+__visible void invoke_syscall_trace(void *table, struct pt_regs *regs)
{
int scno;
+ int ret;
scno = syscall_trace_enter(regs);
if (scno == -1)
- return -1;
+ goto trace_exit_nosave;
- if (scno < NR_syscalls)
- /* Doing this with return makes sure the stack gets pop:ed */
- return invoke_syscall_trace_asm(table, regs, scno, retp);
+ if (scno < NR_syscalls) {
+ ret = invoke_syscall_trace_asm(table, regs, scno);
+ goto trace_exit_save;
+ }
- if (scno >= __ARM_NR_BASE)
- return arm_syscall(scno, regs);
+ if (scno >= __ARM_NR_BASE) {
+ ret = arm_syscall(scno, regs);
+ goto trace_exit_save;
+ }
- return sys_ni_syscall();
+ ret = sys_ni_syscall();
+
+trace_exit_save:
+ /* Save return value from syscall */
+ regs->ARM_r0 = ret;
+
+trace_exit_nosave:
+ local_irq_enable();
+ syscall_trace_exit(regs);
}
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 11/31] ARM: entry: Merge the common and trace entry code
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (9 preceding siblings ...)
2025-02-12 11:23 ` [PATCH v4 10/31] ARM: entry: move all tracing invocation to C Linus Walleij
@ 2025-02-12 11:23 ` Linus Walleij
2025-02-12 11:23 ` [PATCH v4 12/31] ARM: entry: Rename syscall invocation Linus Walleij
` (20 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:23 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
The trace entry code now can handle recursive and complex
calls in C using stack.
Move the common code over to using that approach.
We now use the ret_fast_syscall return path also when tracing,
which appears to work just fine.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/include/asm/syscall.h | 3 +--
arch/arm/kernel/entry-common.S | 61 +++++++++---------------------------------
arch/arm/kernel/syscall.c | 35 +++++++++++-------------
3 files changed, 28 insertions(+), 71 deletions(-)
diff --git a/arch/arm/include/asm/syscall.h b/arch/arm/include/asm/syscall.h
index cb0073c4151bf70a82d15e17a95b6b6f48b245d6..9c664d8c5718f58034f442c20a030b59faf65dd9 100644
--- a/arch/arm/include/asm/syscall.h
+++ b/arch/arm/include/asm/syscall.h
@@ -19,8 +19,7 @@
extern const unsigned long sys_call_table[];
-int invoke_syscall(void *table, struct pt_regs *regs, int scno, void *retp);
-void invoke_syscall_trace(void *table, struct pt_regs *regs);
+int invoke_syscall_trace(void *table, struct pt_regs *regs, int scno);
static inline int syscall_get_nr(struct task_struct *task,
struct pt_regs *regs)
diff --git a/arch/arm/kernel/entry-common.S b/arch/arm/kernel/entry-common.S
index dbc947d301ec11bb13007cb4bb161c035ede5c10..f0f1f8723965313b287158fb2a238908db72fcd2 100644
--- a/arch/arm/kernel/entry-common.S
+++ b/arch/arm/kernel/entry-common.S
@@ -39,7 +39,6 @@ saved_pc .req lr
* from those features make this path too inefficient.
*/
ret_fast_syscall:
-__ret_fast_syscall:
UNWIND(.fnstart )
UNWIND(.cantunwind )
disable_irq_notrace @ disable interrupts
@@ -47,13 +46,13 @@ __ret_fast_syscall:
movs r1, r1, lsl #16
bne fast_work_pending
- restore_user_regs fast = 1, offset = S_OFF
+ restore_user_regs fast = 0, offset = S_OFF
UNWIND(.fnend )
ENDPROC(ret_fast_syscall)
/* Ok, we need to do extra processing, enter the slow path. */
fast_work_pending:
- str r0, [sp, #S_R0+S_OFF]! @ returned r0
+ add sp, sp, #(S_R0 + S_OFF)
/* fall through to work_pending */
#else
/*
@@ -63,10 +62,9 @@ fast_work_pending:
* call.
*/
ret_fast_syscall:
-__ret_fast_syscall:
UNWIND(.fnstart )
UNWIND(.cantunwind )
- str r0, [sp, #S_R0 + S_OFF]! @ save returned r0
+ add sp, sp, #(S_R0 + S_OFF)
#if IS_ENABLED(CONFIG_DEBUG_RSEQ)
/* do_rseq_syscall needs interrupts enabled. */
mov r0, sp @ 'regs'
@@ -83,7 +81,9 @@ ENDPROC(ret_fast_syscall)
#endif
tst r1, #_TIF_SYSCALL_WORK
- bne __sys_trace_return_nosave
+ beq slow_work_pending
+ b ret_to_user
+
slow_work_pending:
mov r0, sp @ 'regs'
bl do_work_pending
@@ -257,19 +257,15 @@ ENTRY(vector_swi)
str scno, [tsk, #TI_ABI_SYSCALL]
#endif
mov r1, sp @ put regs into r1
- ldr r10, [tsk, #TI_FLAGS] @ check for syscall tracing
stmdb sp!, {r4, r5} @ push fifth and sixth args
-
- tst r10, #_TIF_SYSCALL_WORK @ are we tracing syscalls?
- bne __sys_trace
-
mov r0, tbl
- /* r1 already contains regs */
mov r2, scno @ syscall number from r7
- /* We return here no matter what, also pass this as an argument */
- badr lr, __ret_fast_syscall
- mov r3, lr
- b invoke_syscall
+ bl invoke_syscall_trace
+ cmp r0, #0
+ beq ret_fast_syscall
+ /* This path taken when tracing */
+ add sp, sp, #(S_R0 + S_OFF)
+ b ret_to_user
#if defined(CONFIG_OABI_COMPAT) || !defined(CONFIG_AEABI)
/*
@@ -289,20 +285,6 @@ ENTRY(vector_swi)
ENDPROC(vector_swi)
.ltorg
- /*
- * This is the really slow path. We're going to be doing
- * context switches, and waiting for our parent to respond.
- */
-__sys_trace:
- add r1, sp, #S_R0 + S_OFF @ pointer to regs
- mov r0, tbl
- bl invoke_syscall_trace
- add sp, sp, #S_OFF @ restore stack pointer
- b ret_to_user
-
-__sys_trace_return_nosave:
- b ret_to_user
-
.macro syscall_table_start, sym
.equ __sys_nr, 0
.type \sym, #object
@@ -402,25 +384,6 @@ sys_mmap2:
b sys_mmap_pgoff
ENDPROC(sys_mmap2)
-/*
- * This call wants:
- * r0: syscall table
- * r1: regs
- * r2: syscall number
- * r3: pointer to return function
- */
-SYM_TYPED_FUNC_START(invoke_syscall_asm)
-#ifdef CONFIG_CPU_SPECTRE
- csdb
-#endif
- mov tbl, r0
- mov scno, r2
- mov lr, r3 @ return address
- ldmia r1, {r0 - r3} @ reload r0-r3
- /* Arguments 5 and 6 are (hopefully) on the stack */
- ldr pc, [tbl, scno, lsl #2] @ call sys_* routine
-SYM_FUNC_END(invoke_syscall_asm)
-
/*
* This call wants:
* r0: syscall table
diff --git a/arch/arm/kernel/syscall.c b/arch/arm/kernel/syscall.c
index e90f04f35b3485968a1e9046cf0889f56cca92e5..36330023b5484399a4ca2fdb5727b410e2c74b6f 100644
--- a/arch/arm/kernel/syscall.c
+++ b/arch/arm/kernel/syscall.c
@@ -3,50 +3,45 @@
#include <linux/syscalls.h>
#include <asm/syscall.h>
-int invoke_syscall_asm(void *table, struct pt_regs *regs, int scno, void *retp);
-__ADDRESSABLE(invoke_syscall_asm);
-
-__visible int invoke_syscall(void *table, struct pt_regs *regs, int scno, void *retp)
+static inline bool has_syscall_work(unsigned long flags)
{
- if (scno < NR_syscalls)
- /* Doing this with return makes sure the stack gets pop:ed */
- return invoke_syscall_asm(table, regs, scno, retp);
-
- if (scno >= __ARM_NR_BASE)
- return arm_syscall(scno, regs);
-
- return sys_ni_syscall();
+ return unlikely(flags & _TIF_SYSCALL_WORK);
}
int invoke_syscall_trace_asm(void *table, struct pt_regs *regs, int scno);
__ADDRESSABLE(invoke_syscall_trace_asm);
-__visible void invoke_syscall_trace(void *table, struct pt_regs *regs)
+__visible int invoke_syscall_trace(void *table, struct pt_regs *regs, int scno)
{
- int scno;
+ unsigned long flags = read_thread_flags();
int ret;
- scno = syscall_trace_enter(regs);
- if (scno == -1)
- goto trace_exit_nosave;
+ if (has_syscall_work(flags)) {
+ scno = syscall_trace_enter(regs);
+ if (scno == -1)
+ goto trace_exit_nosave;
+ }
if (scno < NR_syscalls) {
ret = invoke_syscall_trace_asm(table, regs, scno);
- goto trace_exit_save;
+ goto exit_save;
}
if (scno >= __ARM_NR_BASE) {
ret = arm_syscall(scno, regs);
- goto trace_exit_save;
+ goto exit_save;
}
ret = sys_ni_syscall();
-trace_exit_save:
+exit_save:
/* Save return value from syscall */
regs->ARM_r0 = ret;
+ if (!has_syscall_work(flags))
+ return 0;
trace_exit_nosave:
local_irq_enable();
syscall_trace_exit(regs);
+ return 1;
}
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 12/31] ARM: entry: Rename syscall invocation
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (10 preceding siblings ...)
2025-02-12 11:23 ` [PATCH v4 11/31] ARM: entry: Merge the common and trace entry code Linus Walleij
@ 2025-02-12 11:23 ` Linus Walleij
2025-02-12 11:23 ` [PATCH v4 13/31] ARM: entry: Create user_mode_enter/exit Linus Walleij
` (19 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:23 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
Now that we do not have a separate trace syscall invocation
path, rename the invocation functions to be generic.
This can be squashed into the previous patch, it is just done
here for the previous patch to be easier to get a cleaner
diff and be easier to read.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/include/asm/syscall.h | 2 +-
arch/arm/kernel/entry-common.S | 10 +++++-----
arch/arm/kernel/syscall.c | 8 ++++----
3 files changed, 10 insertions(+), 10 deletions(-)
diff --git a/arch/arm/include/asm/syscall.h b/arch/arm/include/asm/syscall.h
index 9c664d8c5718f58034f442c20a030b59faf65dd9..b94fd7ce17f82f0baf06b74800245112e1f13002 100644
--- a/arch/arm/include/asm/syscall.h
+++ b/arch/arm/include/asm/syscall.h
@@ -19,7 +19,7 @@
extern const unsigned long sys_call_table[];
-int invoke_syscall_trace(void *table, struct pt_regs *regs, int scno);
+int invoke_syscall(void *table, struct pt_regs *regs, int scno);
static inline int syscall_get_nr(struct task_struct *task,
struct pt_regs *regs)
diff --git a/arch/arm/kernel/entry-common.S b/arch/arm/kernel/entry-common.S
index f0f1f8723965313b287158fb2a238908db72fcd2..f1e48002bd3075ea67b5883178583127fa0055c6 100644
--- a/arch/arm/kernel/entry-common.S
+++ b/arch/arm/kernel/entry-common.S
@@ -260,7 +260,7 @@ ENTRY(vector_swi)
stmdb sp!, {r4, r5} @ push fifth and sixth args
mov r0, tbl
mov r2, scno @ syscall number from r7
- bl invoke_syscall_trace
+ bl invoke_syscall
cmp r0, #0
beq ret_fast_syscall
/* This path taken when tracing */
@@ -390,7 +390,7 @@ ENDPROC(sys_mmap2)
* r1: regs
* r2: syscall number
*/
-SYM_TYPED_FUNC_START(invoke_syscall_trace_asm)
+SYM_TYPED_FUNC_START(invoke_syscall_asm)
#ifdef CONFIG_CPU_SPECTRE
csdb
#endif
@@ -400,17 +400,17 @@ SYM_TYPED_FUNC_START(invoke_syscall_trace_asm)
/* Make space to copy the two syscall stack arguments */
sub sp, sp, #S_OFF
mov scno, r2
- badr lr, __invoke_syscall_trace_ret @ return right here
+ badr lr, __invoke_syscall_ret @ return right here
ldmia r1, {r0 - r6} @ reload r0-r6 from regs
stmia sp, {r4, r5} @ copy stack arguments
ldr pc, [tbl, scno, lsl #2] @ call sys_* routine
-__invoke_syscall_trace_ret:
+__invoke_syscall_ret:
/* Drop the copied stack arguments */
add sp, sp, #S_OFF
pop {r4 - r10, lr}
ARM( mov pc, lr )
THUMB( bx lr )
-SYM_FUNC_END(invoke_syscall_trace_asm)
+SYM_FUNC_END(invoke_syscall_asm)
#ifdef CONFIG_OABI_COMPAT
diff --git a/arch/arm/kernel/syscall.c b/arch/arm/kernel/syscall.c
index 36330023b5484399a4ca2fdb5727b410e2c74b6f..6f518ec3870fb36a64d1b671b20f96d63a52bf78 100644
--- a/arch/arm/kernel/syscall.c
+++ b/arch/arm/kernel/syscall.c
@@ -8,10 +8,10 @@ static inline bool has_syscall_work(unsigned long flags)
return unlikely(flags & _TIF_SYSCALL_WORK);
}
-int invoke_syscall_trace_asm(void *table, struct pt_regs *regs, int scno);
-__ADDRESSABLE(invoke_syscall_trace_asm);
+int invoke_syscall_asm(void *table, struct pt_regs *regs, int scno);
+__ADDRESSABLE(invoke_syscall_asm);
-__visible int invoke_syscall_trace(void *table, struct pt_regs *regs, int scno)
+__visible int invoke_syscall(void *table, struct pt_regs *regs, int scno)
{
unsigned long flags = read_thread_flags();
int ret;
@@ -23,7 +23,7 @@ __visible int invoke_syscall_trace(void *table, struct pt_regs *regs, int scno)
}
if (scno < NR_syscalls) {
- ret = invoke_syscall_trace_asm(table, regs, scno);
+ ret = invoke_syscall_asm(table, regs, scno);
goto exit_save;
}
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 13/31] ARM: entry: Create user_mode_enter/exit
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (11 preceding siblings ...)
2025-02-12 11:23 ` [PATCH v4 12/31] ARM: entry: Rename syscall invocation Linus Walleij
@ 2025-02-12 11:23 ` Linus Walleij
2025-02-12 11:23 ` [PATCH v4 14/31] ARM: entry: Drop trace argument from usr_entry macro Linus Walleij
` (18 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:23 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
The callbacks to the context tracking will be repurposed
for several uses that are needed on the IRQ transition
to/from userspace.
Rename the macro, establish call sites in C calling into
the context tracking following the corresponding generic
entry function prototypes, despite the assembly macro
names become a bit long this makes it clear to readers
exactly what is going on and where this call will go.
Drop the ifdefs pertaining to context tracking from
the macro. The C calls we will use have stubs that will
compile these out anyway.
The inversion of the signature of the context tracking
calls are especially confusing since the generic entry
uses the reverse semantics: *enter from* user mode (to
kernel mode) and *exit to* user mode (from kernel mode)
instead of the other way around as the old context tracker
code user_exit_callable() and user_enter_callable()
which have inverted semantics.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/include/asm/entry.h | 14 ++++++++++++++
arch/arm/kernel/Makefile | 2 +-
arch/arm/kernel/entry-armv.S | 2 +-
arch/arm/kernel/entry-common.S | 4 ++--
| 24 ++++++++++++------------
arch/arm/kernel/entry.c | 15 +++++++++++++++
6 files changed, 45 insertions(+), 16 deletions(-)
diff --git a/arch/arm/include/asm/entry.h b/arch/arm/include/asm/entry.h
new file mode 100644
index 0000000000000000000000000000000000000000..e26f369375ca3cf762f92fb499657a666b223ca2
--- /dev/null
+++ b/arch/arm/include/asm/entry.h
@@ -0,0 +1,14 @@
+/* SPDX-License-Identifier: GPL-2.0-only */
+#ifndef __ASM_ENTRY_H__
+#define __ASM_ENTRY_H__
+
+struct pt_regs;
+
+/*
+ * These are copies of generic entry headers so we can transition
+ * to generic entry once they are semantically equivalent.
+ */
+void irqentry_enter_from_user_mode(struct pt_regs *regs);
+void irqentry_exit_to_user_mode(struct pt_regs *regs);
+
+#endif /* __ASM_ENTRY_H__ */
diff --git a/arch/arm/kernel/Makefile b/arch/arm/kernel/Makefile
index 69ddf51081f4791982518d37df60c11211b8955a..dd45c6eb8cbdd2595edc4f26026c12b25dd49717 100644
--- a/arch/arm/kernel/Makefile
+++ b/arch/arm/kernel/Makefile
@@ -17,7 +17,7 @@ CFLAGS_REMOVE_return_address.o = -pg
# Object file lists.
-obj-y := elf.o entry-common.o irq.o opcodes.o \
+obj-y := elf.o entry.o entry-common.o irq.o opcodes.o \
process.o ptrace.o reboot.o io.o \
setup.o signal.o sigreturn_codes.o \
stacktrace.o sys_arm.o time.o traps.o \
diff --git a/arch/arm/kernel/entry-armv.S b/arch/arm/kernel/entry-armv.S
index ef6a657c8d130663e68acd07226815ce8a3fff7d..93a37437a9520118eff411f009e8f6bd729412fc 100644
--- a/arch/arm/kernel/entry-armv.S
+++ b/arch/arm/kernel/entry-armv.S
@@ -406,7 +406,7 @@ ENDPROC(__fiq_abt)
#ifdef CONFIG_TRACE_IRQFLAGS
bl trace_hardirqs_off
#endif
- ct_user_exit save = 0
+ asm_irqentry_enter_from_user_mode save = 0
.endif
.endm
diff --git a/arch/arm/kernel/entry-common.S b/arch/arm/kernel/entry-common.S
index f1e48002bd3075ea67b5883178583127fa0055c6..ff1dd3169346f3770cad6b7e218f5d74ffc646fe 100644
--- a/arch/arm/kernel/entry-common.S
+++ b/arch/arm/kernel/entry-common.S
@@ -111,7 +111,7 @@ ENTRY(ret_to_user_from_irq)
no_work_pending:
asm_trace_hardirqs_on save = 0
- ct_user_enter save = 0
+ asm_irqentry_exit_to_user_mode save = 0
#ifdef CONFIG_GCC_PLUGIN_STACKLEAK
bl stackleak_erase_on_task_stack
@@ -191,7 +191,7 @@ ENTRY(vector_swi)
alignment_trap r10, ip, cr_alignment
asm_trace_hardirqs_on save=0
enable_irq_notrace
- ct_user_exit save=0
+ asm_irqentry_enter_from_user_mode save = 0
/*
* Get the system call number.
--git a/arch/arm/kernel/entry-header.S b/arch/arm/kernel/entry-header.S
index 52b4fa97226dbfa4b55aca8eaf74ae24e1e220f8..fb5bb019199b2871e29e306a29bea8fdf47dd7f3 100644
--- a/arch/arm/kernel/entry-header.S
+++ b/arch/arm/kernel/entry-header.S
@@ -362,31 +362,31 @@ ALT_UP_B(.L1_\@)
.endm
/*
- * Context tracking subsystem. Used to instrument transitions
+ * Context tracking and other mode transitions. Used to instrument transitions
* between user and kernel mode.
- */
- .macro ct_user_exit, save = 1
-#ifdef CONFIG_CONTEXT_TRACKING_USER
+*/
+ .macro asm_irqentry_enter_from_user_mode, save = 1
.if \save
stmdb sp!, {r0-r3, ip, lr}
- bl user_exit_callable
+ mov r0, sp @ regs
+ bl irqentry_enter_from_user_mode
ldmia sp!, {r0-r3, ip, lr}
.else
- bl user_exit_callable
+ mov r0, sp @ regs
+ bl irqentry_enter_from_user_mode
.endif
-#endif
.endm
- .macro ct_user_enter, save = 1
-#ifdef CONFIG_CONTEXT_TRACKING_USER
+ .macro asm_irqentry_exit_to_user_mode, save = 1
.if \save
stmdb sp!, {r0-r3, ip, lr}
- bl user_enter_callable
+ mov r0, sp @ regs
+ bl irqentry_exit_to_user_mode
ldmia sp!, {r0-r3, ip, lr}
.else
- bl user_enter_callable
+ mov r0, sp @ regs
+ bl irqentry_exit_to_user_mode
.endif
-#endif
.endm
/*
diff --git a/arch/arm/kernel/entry.c b/arch/arm/kernel/entry.c
new file mode 100644
index 0000000000000000000000000000000000000000..8b2e8ea66c1376759d6c0c14aad8728895b3ff1e
--- /dev/null
+++ b/arch/arm/kernel/entry.c
@@ -0,0 +1,15 @@
+// SPDX-License-Identifier: GPL-2.0
+#include <asm/entry.h>
+#include <linux/context_tracking.h>
+
+noinstr void irqentry_enter_from_user_mode(struct pt_regs *regs)
+{
+ /* This context tracking call has inverse naming */
+ user_exit_callable();
+}
+
+noinstr void irqentry_exit_to_user_mode(struct pt_regs *regs)
+{
+ /* This context tracking call has inverse naming */
+ user_enter_callable();
+}
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 14/31] ARM: entry: Drop trace argument from usr_entry macro
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (12 preceding siblings ...)
2025-02-12 11:23 ` [PATCH v4 13/31] ARM: entry: Create user_mode_enter/exit Linus Walleij
@ 2025-02-12 11:23 ` Linus Walleij
2025-02-12 11:23 ` [PATCH v4 15/31] ARM: entry: Separate call path for syscall SWI entry Linus Walleij
` (17 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:23 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
The trace argument select whether to look into tracing of
the hardirqs for the exceptions from userspace, but also
selects whether to call further down to establish context
for RCU.
Split off a second macro that is called on everything
except the FIQ since this is the only outlier.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/kernel/entry-armv.S | 13 +++++++++----
1 file changed, 9 insertions(+), 4 deletions(-)
diff --git a/arch/arm/kernel/entry-armv.S b/arch/arm/kernel/entry-armv.S
index 93a37437a9520118eff411f009e8f6bd729412fc..c71110126fc105fc6ac2d6cb0f5f399b4c8b1548 100644
--- a/arch/arm/kernel/entry-armv.S
+++ b/arch/arm/kernel/entry-armv.S
@@ -357,7 +357,7 @@ ENDPROC(__fiq_abt)
#error "sizeof(struct pt_regs) must be a multiple of 8"
#endif
- .macro usr_entry, trace=1, uaccess=1
+ .macro usr_entry, uaccess=1
UNWIND(.fnstart )
UNWIND(.cantunwind ) @ don't unwind the user space
sub sp, sp, #PT_REGS_SIZE
@@ -401,13 +401,14 @@ ENDPROC(__fiq_abt)
@ Clear FP to mark the first stack frame
@
zero_fp
+ .endm
- .if \trace
+ /* Called after usr_entry for everything except FIQ */
+ .macro usr_entry_enter
#ifdef CONFIG_TRACE_IRQFLAGS
bl trace_hardirqs_off
#endif
asm_irqentry_enter_from_user_mode save = 0
- .endif
.endm
.macro kuser_cmpxchg_check
@@ -429,6 +430,7 @@ ENDPROC(__fiq_abt)
.align 5
__dabt_usr:
usr_entry uaccess=0
+ usr_entry_enter
kuser_cmpxchg_check
mov r2, sp
dabt_helper
@@ -439,6 +441,7 @@ ENDPROC(__dabt_usr)
.align 5
__irq_usr:
usr_entry
+ usr_entry_enter
kuser_cmpxchg_check
irq_handler from_user=1
get_thread_info tsk
@@ -452,6 +455,7 @@ ENDPROC(__irq_usr)
.align 5
__und_usr:
usr_entry uaccess=0
+ usr_entry_enter
@ IRQs must be enabled before attempting to read the instruction from
@ user space since that could cause a page/translation fault if the
@@ -476,6 +480,7 @@ ENDPROC(__und_usr)
.align 5
__pabt_usr:
usr_entry
+ usr_entry_enter
mov r2, sp @ regs
pabt_helper
UNWIND(.fnend )
@@ -495,7 +500,7 @@ ENDPROC(ret_from_exception)
.align 5
__fiq_usr:
- usr_entry trace=0
+ usr_entry
kuser_cmpxchg_check
mov r0, sp @ struct pt_regs *regs
bl handle_fiq_as_nmi
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 15/31] ARM: entry: Separate call path for syscall SWI entry
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (13 preceding siblings ...)
2025-02-12 11:23 ` [PATCH v4 14/31] ARM: entry: Drop trace argument from usr_entry macro Linus Walleij
@ 2025-02-12 11:23 ` Linus Walleij
2025-02-12 11:23 ` [PATCH v4 16/31] ARM: entry: Drop argument to asm_irqentry macros Linus Walleij
` (16 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:23 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
The syscalls (SWIs, software interrupts) are deviating from how
any other interrupts are handled as they enable the IRQs again
while processing the syscall, while "hard" IRQs disable all
interrupts until they are handled.
Break out syscall_enter_from_user_mode() into its own function
and call it instead of irqentry_enter_from_user_mode().
As we are moving toward generic entry, we use the signature
from the generic function.
As the generic function requires the syscall number to be
determined, we move the call down below the code that
figures out the syscall number, the only practical effect
should be that interrupts are re-enabled a few instructions
later.
As we move the trace_hardirqs_on/off() calls into C, we can
just get rid of the helper macro usr_entry_enter again and
call asm_irqentry_enter_from_user_mode directly.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/include/asm/entry.h | 1 +
arch/arm/kernel/entry-armv.S | 16 ++++------------
arch/arm/kernel/entry-common.S | 18 +++++++++++++-----
arch/arm/kernel/entry.c | 14 ++++++++++++++
4 files changed, 32 insertions(+), 17 deletions(-)
diff --git a/arch/arm/include/asm/entry.h b/arch/arm/include/asm/entry.h
index e26f369375ca3cf762f92fb499657a666b223ca2..e259b074caef75c7f777b18199623f07bebee5b4 100644
--- a/arch/arm/include/asm/entry.h
+++ b/arch/arm/include/asm/entry.h
@@ -8,6 +8,7 @@ struct pt_regs;
* These are copies of generic entry headers so we can transition
* to generic entry once they are semantically equivalent.
*/
+long syscall_enter_from_user_mode(struct pt_regs *regs, long);
void irqentry_enter_from_user_mode(struct pt_regs *regs);
void irqentry_exit_to_user_mode(struct pt_regs *regs);
diff --git a/arch/arm/kernel/entry-armv.S b/arch/arm/kernel/entry-armv.S
index c71110126fc105fc6ac2d6cb0f5f399b4c8b1548..6edf362ab1e1035dafebf6fb7c55db71462c1eae 100644
--- a/arch/arm/kernel/entry-armv.S
+++ b/arch/arm/kernel/entry-armv.S
@@ -403,14 +403,6 @@ ENDPROC(__fiq_abt)
zero_fp
.endm
- /* Called after usr_entry for everything except FIQ */
- .macro usr_entry_enter
-#ifdef CONFIG_TRACE_IRQFLAGS
- bl trace_hardirqs_off
-#endif
- asm_irqentry_enter_from_user_mode save = 0
- .endm
-
.macro kuser_cmpxchg_check
#if !defined(CONFIG_CPU_32v6K) && defined(CONFIG_KUSER_HELPERS)
#ifndef CONFIG_MMU
@@ -430,7 +422,7 @@ ENDPROC(__fiq_abt)
.align 5
__dabt_usr:
usr_entry uaccess=0
- usr_entry_enter
+ asm_irqentry_enter_from_user_mode save = 0
kuser_cmpxchg_check
mov r2, sp
dabt_helper
@@ -441,7 +433,7 @@ ENDPROC(__dabt_usr)
.align 5
__irq_usr:
usr_entry
- usr_entry_enter
+ asm_irqentry_enter_from_user_mode save = 0
kuser_cmpxchg_check
irq_handler from_user=1
get_thread_info tsk
@@ -455,7 +447,7 @@ ENDPROC(__irq_usr)
.align 5
__und_usr:
usr_entry uaccess=0
- usr_entry_enter
+ asm_irqentry_enter_from_user_mode save = 0
@ IRQs must be enabled before attempting to read the instruction from
@ user space since that could cause a page/translation fault if the
@@ -480,7 +472,7 @@ ENDPROC(__und_usr)
.align 5
__pabt_usr:
usr_entry
- usr_entry_enter
+ asm_irqentry_enter_from_user_mode save = 0
mov r2, sp @ regs
pabt_helper
UNWIND(.fnend )
diff --git a/arch/arm/kernel/entry-common.S b/arch/arm/kernel/entry-common.S
index ff1dd3169346f3770cad6b7e218f5d74ffc646fe..14b2495cae3c2f95b0dfecd849b4e16ec143dbe9 100644
--- a/arch/arm/kernel/entry-common.S
+++ b/arch/arm/kernel/entry-common.S
@@ -109,8 +109,6 @@ ENTRY(ret_to_user_from_irq)
movs r1, r1, lsl #16
bne slow_work_pending
no_work_pending:
- asm_trace_hardirqs_on save = 0
-
asm_irqentry_exit_to_user_mode save = 0
#ifdef CONFIG_GCC_PLUGIN_STACKLEAK
@@ -189,9 +187,6 @@ ENTRY(vector_swi)
reload_current r10, ip
zero_fp
alignment_trap r10, ip, cr_alignment
- asm_trace_hardirqs_on save=0
- enable_irq_notrace
- asm_irqentry_enter_from_user_mode save = 0
/*
* Get the system call number.
@@ -256,6 +251,19 @@ ENTRY(vector_swi)
#else
str scno, [tsk, #TI_ABI_SYSCALL]
#endif
+
+ /*
+ * Calling out to C to be careful to save and restore registers.
+ * This call could modify the syscall number. scno is r7 so we
+ * do not save and restore r7.
+ */
+ mov r0, sp @ regs
+ mov r1, scno
+ push {r4 - r6, r8 - r10, lr}
+ bl syscall_enter_from_user_mode
+ pop {r4 - r6, r8 - r10, lr}
+ mov scno, r0
+
mov r1, sp @ put regs into r1
stmdb sp!, {r4, r5} @ push fifth and sixth args
mov r0, tbl
diff --git a/arch/arm/kernel/entry.c b/arch/arm/kernel/entry.c
index 8b2e8ea66c1376759d6c0c14aad8728895b3ff1e..1973947c7ad753fccd694b3ef334fba1326f58b6 100644
--- a/arch/arm/kernel/entry.c
+++ b/arch/arm/kernel/entry.c
@@ -1,15 +1,29 @@
// SPDX-License-Identifier: GPL-2.0
#include <asm/entry.h>
#include <linux/context_tracking.h>
+#include <linux/irqflags.h>
+
+long syscall_enter_from_user_mode(struct pt_regs *regs, long syscall)
+{
+ trace_hardirqs_on();
+ local_irq_enable();
+ /* This context tracking call has inverse naming */
+ user_exit_callable();
+
+ /* This will optionally be modified later */
+ return syscall;
+}
noinstr void irqentry_enter_from_user_mode(struct pt_regs *regs)
{
+ trace_hardirqs_off();
/* This context tracking call has inverse naming */
user_exit_callable();
}
noinstr void irqentry_exit_to_user_mode(struct pt_regs *regs)
{
+ trace_hardirqs_on();
/* This context tracking call has inverse naming */
user_enter_callable();
}
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 16/31] ARM: entry: Drop argument to asm_irqentry macros
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (14 preceding siblings ...)
2025-02-12 11:23 ` [PATCH v4 15/31] ARM: entry: Separate call path for syscall SWI entry Linus Walleij
@ 2025-02-12 11:23 ` Linus Walleij
2025-02-12 11:23 ` [PATCH v4 17/31] ARM: entry: Implement syscall_exit_to_user_mode() Linus Walleij
` (15 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:23 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
asm_irqentry_enter_from_user_mode and asm_irqentry_exit_to_user_mode
have a "save" argument that will save and restore registers before
the call to the C function.
Now all invocations set this argument to 0 so drop the surplus
code.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/kernel/entry-armv.S | 8 ++++----
arch/arm/kernel/entry-common.S | 2 +-
| 18 ++----------------
3 files changed, 7 insertions(+), 21 deletions(-)
diff --git a/arch/arm/kernel/entry-armv.S b/arch/arm/kernel/entry-armv.S
index 6edf362ab1e1035dafebf6fb7c55db71462c1eae..460aa92c3c1f50de905550acf363c58f509bfe0a 100644
--- a/arch/arm/kernel/entry-armv.S
+++ b/arch/arm/kernel/entry-armv.S
@@ -422,7 +422,7 @@ ENDPROC(__fiq_abt)
.align 5
__dabt_usr:
usr_entry uaccess=0
- asm_irqentry_enter_from_user_mode save = 0
+ asm_irqentry_enter_from_user_mode
kuser_cmpxchg_check
mov r2, sp
dabt_helper
@@ -433,7 +433,7 @@ ENDPROC(__dabt_usr)
.align 5
__irq_usr:
usr_entry
- asm_irqentry_enter_from_user_mode save = 0
+ asm_irqentry_enter_from_user_mode
kuser_cmpxchg_check
irq_handler from_user=1
get_thread_info tsk
@@ -447,7 +447,7 @@ ENDPROC(__irq_usr)
.align 5
__und_usr:
usr_entry uaccess=0
- asm_irqentry_enter_from_user_mode save = 0
+ asm_irqentry_enter_from_user_mode
@ IRQs must be enabled before attempting to read the instruction from
@ user space since that could cause a page/translation fault if the
@@ -472,7 +472,7 @@ ENDPROC(__und_usr)
.align 5
__pabt_usr:
usr_entry
- asm_irqentry_enter_from_user_mode save = 0
+ asm_irqentry_enter_from_user_mode
mov r2, sp @ regs
pabt_helper
UNWIND(.fnend )
diff --git a/arch/arm/kernel/entry-common.S b/arch/arm/kernel/entry-common.S
index 14b2495cae3c2f95b0dfecd849b4e16ec143dbe9..df564388905ee019cd5553f8b37e678da59e3222 100644
--- a/arch/arm/kernel/entry-common.S
+++ b/arch/arm/kernel/entry-common.S
@@ -109,7 +109,7 @@ ENTRY(ret_to_user_from_irq)
movs r1, r1, lsl #16
bne slow_work_pending
no_work_pending:
- asm_irqentry_exit_to_user_mode save = 0
+ asm_irqentry_exit_to_user_mode
#ifdef CONFIG_GCC_PLUGIN_STACKLEAK
bl stackleak_erase_on_task_stack
--git a/arch/arm/kernel/entry-header.S b/arch/arm/kernel/entry-header.S
index fb5bb019199b2871e29e306a29bea8fdf47dd7f3..50c0b55adc7421e7be123c9d00f94b1ebb93ff9e 100644
--- a/arch/arm/kernel/entry-header.S
+++ b/arch/arm/kernel/entry-header.S
@@ -365,28 +365,14 @@ ALT_UP_B(.L1_\@)
* Context tracking and other mode transitions. Used to instrument transitions
* between user and kernel mode.
*/
- .macro asm_irqentry_enter_from_user_mode, save = 1
- .if \save
- stmdb sp!, {r0-r3, ip, lr}
+ .macro asm_irqentry_enter_from_user_mode
mov r0, sp @ regs
bl irqentry_enter_from_user_mode
- ldmia sp!, {r0-r3, ip, lr}
- .else
- mov r0, sp @ regs
- bl irqentry_enter_from_user_mode
- .endif
.endm
- .macro asm_irqentry_exit_to_user_mode, save = 1
- .if \save
- stmdb sp!, {r0-r3, ip, lr}
+ .macro asm_irqentry_exit_to_user_mode
mov r0, sp @ regs
bl irqentry_exit_to_user_mode
- ldmia sp!, {r0-r3, ip, lr}
- .else
- mov r0, sp @ regs
- bl irqentry_exit_to_user_mode
- .endif
.endm
/*
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 17/31] ARM: entry: Implement syscall_exit_to_user_mode()
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (15 preceding siblings ...)
2025-02-12 11:23 ` [PATCH v4 16/31] ARM: entry: Drop argument to asm_irqentry macros Linus Walleij
@ 2025-02-12 11:23 ` Linus Walleij
2025-02-12 11:23 ` [PATCH v4 18/31] ARM: entry: Drop the superfast ret_fast_syscall Linus Walleij
` (14 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:23 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
Mirroring syscall_enter_from_user_mode() we create the sibling
call syscall_exit_to_user_mode() and start to move some stuff
over from assembly, first the rseq call and the interrupt
disablement.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/include/asm/entry.h | 1 +
arch/arm/include/asm/signal.h | 1 -
arch/arm/kernel/entry-common.S | 10 ++--------
arch/arm/kernel/entry.c | 8 ++++++++
arch/arm/kernel/signal.c | 7 -------
5 files changed, 11 insertions(+), 16 deletions(-)
diff --git a/arch/arm/include/asm/entry.h b/arch/arm/include/asm/entry.h
index e259b074caef75c7f777b18199623f07bebee5b4..8b8cc9a1eefbd00dd75d4e344576fc184a413768 100644
--- a/arch/arm/include/asm/entry.h
+++ b/arch/arm/include/asm/entry.h
@@ -9,6 +9,7 @@ struct pt_regs;
* to generic entry once they are semantically equivalent.
*/
long syscall_enter_from_user_mode(struct pt_regs *regs, long);
+void syscall_exit_to_user_mode(struct pt_regs *regs);
void irqentry_enter_from_user_mode(struct pt_regs *regs);
void irqentry_exit_to_user_mode(struct pt_regs *regs);
diff --git a/arch/arm/include/asm/signal.h b/arch/arm/include/asm/signal.h
index 7acccc96840c8a17744cc3f2894f19d142aba4fd..bb753cd3e29fc54186543c48b2aee7b221301a37 100644
--- a/arch/arm/include/asm/signal.h
+++ b/arch/arm/include/asm/signal.h
@@ -23,7 +23,6 @@ typedef struct {
#include <asm/sigcontext.h>
-void do_rseq_syscall(struct pt_regs *regs);
void do_work_pending(struct pt_regs *regs, unsigned int thread_flags);
#endif
diff --git a/arch/arm/kernel/entry-common.S b/arch/arm/kernel/entry-common.S
index df564388905ee019cd5553f8b37e678da59e3222..3559e0a9f5e079f1af91392bf000d39f9b45906c 100644
--- a/arch/arm/kernel/entry-common.S
+++ b/arch/arm/kernel/entry-common.S
@@ -65,12 +65,9 @@ ret_fast_syscall:
UNWIND(.fnstart )
UNWIND(.cantunwind )
add sp, sp, #(S_R0 + S_OFF)
-#if IS_ENABLED(CONFIG_DEBUG_RSEQ)
/* do_rseq_syscall needs interrupts enabled. */
mov r0, sp @ 'regs'
- bl do_rseq_syscall
-#endif
- disable_irq_notrace @ disable interrupts
+ bl syscall_exit_to_user_mode
ldr r1, [tsk, #TI_FLAGS] @ re-check for syscall tracing
movs r1, r1, lsl #16
beq no_work_pending
@@ -97,13 +94,10 @@ ENDPROC(ret_fast_syscall)
* do_work_pending() will update this state if necessary.
*/
ENTRY(ret_to_user)
-#if IS_ENABLED(CONFIG_DEBUG_RSEQ)
/* do_rseq_syscall needs interrupts enabled. */
enable_irq_notrace @ enable interrupts
mov r0, sp @ 'regs'
- bl do_rseq_syscall
-#endif
- disable_irq_notrace @ disable interrupts
+ bl syscall_exit_to_user_mode
ENTRY(ret_to_user_from_irq)
ldr r1, [tsk, #TI_FLAGS]
movs r1, r1, lsl #16
diff --git a/arch/arm/kernel/entry.c b/arch/arm/kernel/entry.c
index 1973947c7ad753fccd694b3ef334fba1326f58b6..47233804733752aef0f2fec6f54badeab18e98d0 100644
--- a/arch/arm/kernel/entry.c
+++ b/arch/arm/kernel/entry.c
@@ -1,7 +1,9 @@
// SPDX-License-Identifier: GPL-2.0
#include <asm/entry.h>
+#include <asm/signal.h>
#include <linux/context_tracking.h>
#include <linux/irqflags.h>
+#include <linux/rseq.h>
long syscall_enter_from_user_mode(struct pt_regs *regs, long syscall)
{
@@ -14,6 +16,12 @@ long syscall_enter_from_user_mode(struct pt_regs *regs, long syscall)
return syscall;
}
+void syscall_exit_to_user_mode(struct pt_regs *regs)
+{
+ rseq_syscall(regs);
+ local_irq_disable();
+}
+
noinstr void irqentry_enter_from_user_mode(struct pt_regs *regs)
{
trace_hardirqs_off();
diff --git a/arch/arm/kernel/signal.c b/arch/arm/kernel/signal.c
index 7b1a16e86b236575efdc29cda9b751e8e2a3f64e..03b20637a2e1ca6f327223c9132915a97a9c4697 100644
--- a/arch/arm/kernel/signal.c
+++ b/arch/arm/kernel/signal.c
@@ -659,13 +659,6 @@ struct page *get_signal_page(void)
return page;
}
-#ifdef CONFIG_DEBUG_RSEQ
-asmlinkage void do_rseq_syscall(struct pt_regs *regs)
-{
- rseq_syscall(regs);
-}
-#endif
-
/*
* Compile-time assertions for siginfo_t offsets. Check NSIG* as well, as
* changes likely come with new fields that should be added below.
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 18/31] ARM: entry: Drop the superfast ret_fast_syscall
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (16 preceding siblings ...)
2025-02-12 11:23 ` [PATCH v4 17/31] ARM: entry: Implement syscall_exit_to_user_mode() Linus Walleij
@ 2025-02-12 11:23 ` Linus Walleij
2025-02-12 11:23 ` [PATCH v4 19/31] ARM: entry: Remove fast and offset register restore Linus Walleij
` (13 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:23 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
When not using any irqflags tracing, context tracking or rseq
debugging, the ARM kernel uses a "superfast" ret_fast_syscall
with a very few assembly instructions.
With generic entry, such local optimizations are not really
possible, we need to call C unconditionally and take the
(hopefully small) overhead.
With generic entry in C, most of the tracing (etc) calls are
replaced by stubs and optimized out by the compiler.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/kernel/entry-common.S | 29 +----------------------------
1 file changed, 1 insertion(+), 28 deletions(-)
diff --git a/arch/arm/kernel/entry-common.S b/arch/arm/kernel/entry-common.S
index 3559e0a9f5e079f1af91392bf000d39f9b45906c..dddbd6ba1eb9fcbc2c0d81f24e1a9b3231558d7f 100644
--- a/arch/arm/kernel/entry-common.S
+++ b/arch/arm/kernel/entry-common.S
@@ -30,34 +30,8 @@ saved_pc .req lr
.section .entry.text,"ax",%progbits
.align 5
-#if !(IS_ENABLED(CONFIG_TRACE_IRQFLAGS) || IS_ENABLED(CONFIG_CONTEXT_TRACKING_USER) || \
- IS_ENABLED(CONFIG_DEBUG_RSEQ))
/*
- * This is the fast syscall return path. We do as little as possible here,
- * such as avoiding writing r0 to the stack. We only use this path if we
- * have tracing, context tracking and rseq debug disabled - the overheads
- * from those features make this path too inefficient.
- */
-ret_fast_syscall:
- UNWIND(.fnstart )
- UNWIND(.cantunwind )
- disable_irq_notrace @ disable interrupts
- ldr r1, [tsk, #TI_FLAGS] @ re-check for syscall tracing
- movs r1, r1, lsl #16
- bne fast_work_pending
-
- restore_user_regs fast = 0, offset = S_OFF
- UNWIND(.fnend )
-ENDPROC(ret_fast_syscall)
-
- /* Ok, we need to do extra processing, enter the slow path. */
-fast_work_pending:
- add sp, sp, #(S_R0 + S_OFF)
- /* fall through to work_pending */
-#else
-/*
- * The "replacement" ret_fast_syscall for when tracing, context tracking,
- * or rseq debug is enabled. As we will need to call out to some C functions,
+ * As we will need to call out to some C functions,
* we save r0 first to avoid needing to save registers around each C function
* call.
*/
@@ -75,7 +49,6 @@ ret_fast_syscall:
ENDPROC(ret_fast_syscall)
/* Slower path - fall through to work_pending */
-#endif
tst r1, #_TIF_SYSCALL_WORK
beq slow_work_pending
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 19/31] ARM: entry: Remove fast and offset register restore
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (17 preceding siblings ...)
2025-02-12 11:23 ` [PATCH v4 18/31] ARM: entry: Drop the superfast ret_fast_syscall Linus Walleij
@ 2025-02-12 11:23 ` Linus Walleij
2025-02-12 11:23 ` [PATCH v4 20/31] ARM: entry: Untangle ret_fast_syscall/to_user Linus Walleij
` (12 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:23 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
Now all remaining invocations of restore_user_regs has
fast = 0 and offset = 0 set, so drop these parameters
to the macro.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/kernel/entry-armv.S | 2 +-
arch/arm/kernel/entry-common.S | 2 +-
| 29 +++++++++--------------------
3 files changed, 11 insertions(+), 22 deletions(-)
diff --git a/arch/arm/kernel/entry-armv.S b/arch/arm/kernel/entry-armv.S
index 460aa92c3c1f50de905550acf363c58f509bfe0a..3cd5a80b47758867d1db6c3971030370bca1551a 100644
--- a/arch/arm/kernel/entry-armv.S
+++ b/arch/arm/kernel/entry-armv.S
@@ -497,7 +497,7 @@ __fiq_usr:
mov r0, sp @ struct pt_regs *regs
bl handle_fiq_as_nmi
get_thread_info tsk
- restore_user_regs fast = 0, offset = 0
+ restore_user_regs
UNWIND(.fnend )
ENDPROC(__fiq_usr)
diff --git a/arch/arm/kernel/entry-common.S b/arch/arm/kernel/entry-common.S
index dddbd6ba1eb9fcbc2c0d81f24e1a9b3231558d7f..d27e34d7e97a079d62e7b778cf7cf97c302e4763 100644
--- a/arch/arm/kernel/entry-common.S
+++ b/arch/arm/kernel/entry-common.S
@@ -81,7 +81,7 @@ no_work_pending:
#ifdef CONFIG_GCC_PLUGIN_STACKLEAK
bl stackleak_erase_on_task_stack
#endif
- restore_user_regs fast = 0, offset = 0
+ restore_user_regs
ENDPROC(ret_to_user_from_irq)
ENDPROC(ret_to_user)
--git a/arch/arm/kernel/entry-header.S b/arch/arm/kernel/entry-header.S
index 50c0b55adc7421e7be123c9d00f94b1ebb93ff9e..51c7f93c1ded1db56f173c7aaed8094ef011ff5a 100644
--- a/arch/arm/kernel/entry-header.S
+++ b/arch/arm/kernel/entry-header.S
@@ -290,7 +290,7 @@
.endm
- .macro restore_user_regs, fast = 0, offset = 0
+ .macro restore_user_regs
#if defined(CONFIG_CPU_32v6K) && \
(!defined(CONFIG_CPU_V6) || defined(CONFIG_SMP))
#ifdef CONFIG_CPU_V6
@@ -309,8 +309,8 @@ ALT_UP_B(.L1_\@)
#ifndef CONFIG_THUMB2_KERNEL
@ ARM mode restore
mov r2, sp
- ldr r1, [r2, #\offset + S_PSR] @ get calling cpsr
- ldr lr, [r2, #\offset + S_PC]! @ get pc
+ ldr r1, [r2, S_PSR] @ get calling cpsr
+ ldr lr, [r2, S_PC]! @ get pc
tst r1, #PSR_I_BIT | 0x0f
bne 1f
msr spsr_cxsf, r1 @ save in spsr_svc
@@ -318,31 +318,24 @@ ALT_UP_B(.L1_\@)
@ We must avoid clrex due to Cortex-A15 erratum #830321
strex r1, r2, [r2] @ clear the exclusive monitor
#endif
- .if \fast
- ldmdb r2, {r1 - lr}^ @ get calling r1 - lr
- .else
ldmdb r2, {r0 - lr}^ @ get calling r0 - lr
- .endif
mov r0, r0 @ ARMv5T and earlier require a nop
@ after ldm {}^
- add sp, sp, #\offset + PT_REGS_SIZE
+ add sp, sp, #PT_REGS_SIZE
movs pc, lr @ return & move spsr_svc into cpsr
1: bug "Returning to usermode but unexpected PSR bits set?", \@
#elif defined(CONFIG_CPU_V7M)
@ V7M restore.
@ Note that we don't need to do clrex here as clearing the local
@ monitor is part of the exception entry and exit sequence.
- .if \offset
- add sp, #\offset
- .endif
- v7m_exception_slow_exit ret_r0 = \fast
+ v7m_exception_slow_exit ret_r0 = 0
#else
@ Thumb mode restore
mov r2, sp
- load_user_sp_lr r2, r3, \offset + S_SP @ calling sp, lr
- ldr r1, [sp, #\offset + S_PSR] @ get calling cpsr
- ldr lr, [sp, #\offset + S_PC] @ get pc
- add sp, sp, #\offset + S_SP
+ load_user_sp_lr r2, r3, S_SP @ calling sp, lr
+ ldr r1, [sp, S_PSR] @ get calling cpsr
+ ldr lr, [sp, S_PC] @ get pc
+ add sp, sp, S_SP
tst r1, #PSR_I_BIT | 0x0f
bne 1f
msr spsr_cxsf, r1 @ save in spsr_svc
@@ -350,11 +343,7 @@ ALT_UP_B(.L1_\@)
@ We must avoid clrex due to Cortex-A15 erratum #830321
strex r1, r2, [sp] @ clear the exclusive monitor
- .if \fast
- ldmdb sp, {r1 - r12} @ get calling r1 - r12
- .else
ldmdb sp, {r0 - r12} @ get calling r0 - r12
- .endif
add sp, sp, #PT_REGS_SIZE - S_SP
movs pc, lr @ return & move spsr_svc into cpsr
1: bug "Returning to usermode but unexpected PSR bits set?", \@
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 20/31] ARM: entry: Untangle ret_fast_syscall/to_user
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (18 preceding siblings ...)
2025-02-12 11:23 ` [PATCH v4 19/31] ARM: entry: Remove fast and offset register restore Linus Walleij
@ 2025-02-12 11:23 ` Linus Walleij
2025-02-12 11:23 ` [PATCH v4 21/31] ARM: entry: Do not double-call exit functions Linus Walleij
` (11 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:23 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
The return to userspace is a bit hard to follow. To prepare
the ground for further refactorings we rewrite the functions
to ret_fast_syscall and ret_to_user are mostly stand-alone
symbols.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/kernel/entry-common.S | 25 +++++++++++++++----------
1 file changed, 15 insertions(+), 10 deletions(-)
diff --git a/arch/arm/kernel/entry-common.S b/arch/arm/kernel/entry-common.S
index d27e34d7e97a079d62e7b778cf7cf97c302e4763..ece9215717143427de7572a18a6dbe620a9193e9 100644
--- a/arch/arm/kernel/entry-common.S
+++ b/arch/arm/kernel/entry-common.S
@@ -44,20 +44,23 @@ ret_fast_syscall:
bl syscall_exit_to_user_mode
ldr r1, [tsk, #TI_FLAGS] @ re-check for syscall tracing
movs r1, r1, lsl #16
- beq no_work_pending
- UNWIND(.fnend )
-ENDPROC(ret_fast_syscall)
-
- /* Slower path - fall through to work_pending */
+ beq 2f
tst r1, #_TIF_SYSCALL_WORK
- beq slow_work_pending
+ beq 1f
+
b ret_to_user
-slow_work_pending:
- mov r0, sp @ 'regs'
+1: mov r0, sp @ 'regs'
bl do_work_pending
- b no_work_pending
+
+2: asm_irqentry_exit_to_user_mode
+
+#ifdef CONFIG_GCC_PLUGIN_STACKLEAK
+ bl stackleak_erase_on_task_stack
+#endif
+ restore_user_regs
+ UNWIND(.fnend )
ENDPROC(ret_fast_syscall)
/*
@@ -74,7 +77,9 @@ ENTRY(ret_to_user)
ENTRY(ret_to_user_from_irq)
ldr r1, [tsk, #TI_FLAGS]
movs r1, r1, lsl #16
- bne slow_work_pending
+ beq no_work_pending
+ mov r0, sp @ 'regs'
+ bl do_work_pending
no_work_pending:
asm_irqentry_exit_to_user_mode
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 21/31] ARM: entry: Do not double-call exit functions
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (19 preceding siblings ...)
2025-02-12 11:23 ` [PATCH v4 20/31] ARM: entry: Untangle ret_fast_syscall/to_user Linus Walleij
@ 2025-02-12 11:23 ` Linus Walleij
2025-02-12 11:23 ` [PATCH v4 22/31] ARM: entry: Move work processing to C Linus Walleij
` (10 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:23 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
This is a semantic change to ret_fast_syscall: if there is no
work pending the function used to jump to ret_to_user, instead
we proceed to return from the syscall.
If we jump to ret_to_user, IRQs are enabled and we call
syscall_exit_to_user_mode a second time; at this point a
this just means a second call to rseq_syscall(), but it
seems wrong and we want syscall_exit_to_user_mode to be called
exactly once.
Apparently the ARM kernel has under some circumstances called
rseq_syscall() twice on the exit to userspace without side effects.
ret_to_user also checks for pending work a second time under
ret_to_user_from_irq, but this isn't even needed any more on
the fast return path: this was needed because the syscall
invocation enabled interrupts when doing syscall tracing, and
all tracing invocation has been moved over to C in earlier
patches.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/kernel/entry-common.S | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff --git a/arch/arm/kernel/entry-common.S b/arch/arm/kernel/entry-common.S
index ece9215717143427de7572a18a6dbe620a9193e9..33bc9e7c5b32f62474724f1d47f97af24dba53bf 100644
--- a/arch/arm/kernel/entry-common.S
+++ b/arch/arm/kernel/entry-common.S
@@ -49,7 +49,7 @@ ret_fast_syscall:
tst r1, #_TIF_SYSCALL_WORK
beq 1f
- b ret_to_user
+ b 2f
1: mov r0, sp @ 'regs'
bl do_work_pending
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 22/31] ARM: entry: Move work processing to C
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (20 preceding siblings ...)
2025-02-12 11:23 ` [PATCH v4 21/31] ARM: entry: Do not double-call exit functions Linus Walleij
@ 2025-02-12 11:23 ` Linus Walleij
2025-02-12 11:23 ` [PATCH v4 23/31] ARM: entry: Stop exiting syscalls like IRQs Linus Walleij
` (9 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:23 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
At the end of a syscall software interrupt and every time we exit
an interrupt handler we check if there is work to do, i.e. if the
current thread (userspace or kernel daemon) has set the
_TIF_SYSCALL_WORK flag.
Move this processing over to the C entry code.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/kernel/entry-common.S | 14 +-------------
arch/arm/kernel/entry.c | 10 ++++++++++
2 files changed, 11 insertions(+), 13 deletions(-)
diff --git a/arch/arm/kernel/entry-common.S b/arch/arm/kernel/entry-common.S
index 33bc9e7c5b32f62474724f1d47f97af24dba53bf..5385574c4339714f3d8b5ab6894f27466b89f932 100644
--- a/arch/arm/kernel/entry-common.S
+++ b/arch/arm/kernel/entry-common.S
@@ -42,19 +42,7 @@ ret_fast_syscall:
/* do_rseq_syscall needs interrupts enabled. */
mov r0, sp @ 'regs'
bl syscall_exit_to_user_mode
- ldr r1, [tsk, #TI_FLAGS] @ re-check for syscall tracing
- movs r1, r1, lsl #16
- beq 2f
-
- tst r1, #_TIF_SYSCALL_WORK
- beq 1f
-
- b 2f
-
-1: mov r0, sp @ 'regs'
- bl do_work_pending
-
-2: asm_irqentry_exit_to_user_mode
+ asm_irqentry_exit_to_user_mode
#ifdef CONFIG_GCC_PLUGIN_STACKLEAK
bl stackleak_erase_on_task_stack
diff --git a/arch/arm/kernel/entry.c b/arch/arm/kernel/entry.c
index 47233804733752aef0f2fec6f54badeab18e98d0..0e3960844b94b14325eb00ce7dd791b36c43ac3c 100644
--- a/arch/arm/kernel/entry.c
+++ b/arch/arm/kernel/entry.c
@@ -1,10 +1,16 @@
// SPDX-License-Identifier: GPL-2.0
#include <asm/entry.h>
+#include <asm/ptrace.h>
#include <asm/signal.h>
#include <linux/context_tracking.h>
#include <linux/irqflags.h>
#include <linux/rseq.h>
+static inline bool has_syscall_work(unsigned long flags)
+{
+ return unlikely(flags & _TIF_SYSCALL_WORK);
+}
+
long syscall_enter_from_user_mode(struct pt_regs *regs, long syscall)
{
trace_hardirqs_on();
@@ -18,8 +24,12 @@ long syscall_enter_from_user_mode(struct pt_regs *regs, long syscall)
void syscall_exit_to_user_mode(struct pt_regs *regs)
{
+ unsigned long flags = read_thread_flags();
+
rseq_syscall(regs);
local_irq_disable();
+ if (has_syscall_work(flags))
+ do_work_pending(regs, flags);
}
noinstr void irqentry_enter_from_user_mode(struct pt_regs *regs)
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 23/31] ARM: entry: Stop exiting syscalls like IRQs
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (21 preceding siblings ...)
2025-02-12 11:23 ` [PATCH v4 22/31] ARM: entry: Move work processing to C Linus Walleij
@ 2025-02-12 11:23 ` Linus Walleij
2025-02-12 11:23 ` [PATCH v4 24/31] ARM: entry: Complete syscall and IRQ transition to C Linus Walleij
` (8 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:23 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
The syscalls are issued as software interrupts, and to this point
they were sharing code with the other exception handlers, but
this does not work with generic entry.
Make syscall_exit_to_user_mode do what irqentry_exit_to_user_mode
does at it's tail, and drop the invocations of
irqentry_exit_to_user_mode from the syscall path so that these
are now exception-exclusive.
Split ret_to_user and ret_to_user_from_irq into two distinct
execution paths.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/kernel/entry-common.S | 14 ++++++++++++--
arch/arm/kernel/entry.c | 4 ++++
2 files changed, 16 insertions(+), 2 deletions(-)
diff --git a/arch/arm/kernel/entry-common.S b/arch/arm/kernel/entry-common.S
index 5385574c4339714f3d8b5ab6894f27466b89f932..e2ac6d3216b6828d948d60c424ff95ccdad12331 100644
--- a/arch/arm/kernel/entry-common.S
+++ b/arch/arm/kernel/entry-common.S
@@ -42,7 +42,6 @@ ret_fast_syscall:
/* do_rseq_syscall needs interrupts enabled. */
mov r0, sp @ 'regs'
bl syscall_exit_to_user_mode
- asm_irqentry_exit_to_user_mode
#ifdef CONFIG_GCC_PLUGIN_STACKLEAK
bl stackleak_erase_on_task_stack
@@ -62,6 +61,18 @@ ENTRY(ret_to_user)
enable_irq_notrace @ enable interrupts
mov r0, sp @ 'regs'
bl syscall_exit_to_user_mode
+ ldr r1, [tsk, #TI_FLAGS]
+ movs r1, r1, lsl #16
+ beq 1f
+ mov r0, sp @ 'regs'
+ bl do_work_pending
+1:
+#ifdef CONFIG_GCC_PLUGIN_STACKLEAK
+ bl stackleak_erase_on_task_stack
+#endif
+ restore_user_regs
+ENDPROC(ret_to_user)
+
ENTRY(ret_to_user_from_irq)
ldr r1, [tsk, #TI_FLAGS]
movs r1, r1, lsl #16
@@ -76,7 +87,6 @@ no_work_pending:
#endif
restore_user_regs
ENDPROC(ret_to_user_from_irq)
-ENDPROC(ret_to_user)
/*
* This is how we return from a fork.
diff --git a/arch/arm/kernel/entry.c b/arch/arm/kernel/entry.c
index 0e3960844b94b14325eb00ce7dd791b36c43ac3c..88a7a699306ab61153d57333c26266b9161f5466 100644
--- a/arch/arm/kernel/entry.c
+++ b/arch/arm/kernel/entry.c
@@ -30,6 +30,10 @@ void syscall_exit_to_user_mode(struct pt_regs *regs)
local_irq_disable();
if (has_syscall_work(flags))
do_work_pending(regs, flags);
+
+ trace_hardirqs_on();
+ /* This context tracking call has inverse naming */
+ user_enter_callable();
}
noinstr void irqentry_enter_from_user_mode(struct pt_regs *regs)
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 24/31] ARM: entry: Complete syscall and IRQ transition to C
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (22 preceding siblings ...)
2025-02-12 11:23 ` [PATCH v4 23/31] ARM: entry: Stop exiting syscalls like IRQs Linus Walleij
@ 2025-02-12 11:23 ` Linus Walleij
2025-02-12 11:23 ` [PATCH v4 25/31] ARM: entry: Create irqentry calls from kernel mode Linus Walleij
` (7 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:23 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
This moves over the last few lines of assembly to C. The
subtle change is that in return to userspace from syscall
(SWI) or interrupt, we need to call do_work_pending()
as soon as the thread flags are != 0, just checking for
work with _TIF_SYSCALL_WORK is not enough (the machine will
freeze if we do that).
This is because do_work_pending() does not just handle
work: it handles _TIF_NEED_RESCHED, _TIF_SIGPENDING,
_TIF_NOTIFY_SIGNAL and _TIF_UPROBE as well.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/kernel/entry-common.S | 13 +------------
arch/arm/kernel/entry.c | 19 +++++++++++++------
arch/arm/kernel/signal.c | 3 +--
3 files changed, 15 insertions(+), 20 deletions(-)
diff --git a/arch/arm/kernel/entry-common.S b/arch/arm/kernel/entry-common.S
index e2ac6d3216b6828d948d60c424ff95ccdad12331..6b0f86786a7d9e656ecbe8c741f2edbcbafd2915 100644
--- a/arch/arm/kernel/entry-common.S
+++ b/arch/arm/kernel/entry-common.S
@@ -61,12 +61,7 @@ ENTRY(ret_to_user)
enable_irq_notrace @ enable interrupts
mov r0, sp @ 'regs'
bl syscall_exit_to_user_mode
- ldr r1, [tsk, #TI_FLAGS]
- movs r1, r1, lsl #16
- beq 1f
- mov r0, sp @ 'regs'
- bl do_work_pending
-1:
+
#ifdef CONFIG_GCC_PLUGIN_STACKLEAK
bl stackleak_erase_on_task_stack
#endif
@@ -74,12 +69,6 @@ ENTRY(ret_to_user)
ENDPROC(ret_to_user)
ENTRY(ret_to_user_from_irq)
- ldr r1, [tsk, #TI_FLAGS]
- movs r1, r1, lsl #16
- beq no_work_pending
- mov r0, sp @ 'regs'
- bl do_work_pending
-no_work_pending:
asm_irqentry_exit_to_user_mode
#ifdef CONFIG_GCC_PLUGIN_STACKLEAK
diff --git a/arch/arm/kernel/entry.c b/arch/arm/kernel/entry.c
index 88a7a699306ab61153d57333c26266b9161f5466..d7fdb9df33312dd2a914b59512b701748f33b597 100644
--- a/arch/arm/kernel/entry.c
+++ b/arch/arm/kernel/entry.c
@@ -6,11 +6,6 @@
#include <linux/irqflags.h>
#include <linux/rseq.h>
-static inline bool has_syscall_work(unsigned long flags)
-{
- return unlikely(flags & _TIF_SYSCALL_WORK);
-}
-
long syscall_enter_from_user_mode(struct pt_regs *regs, long syscall)
{
trace_hardirqs_on();
@@ -28,7 +23,11 @@ void syscall_exit_to_user_mode(struct pt_regs *regs)
rseq_syscall(regs);
local_irq_disable();
- if (has_syscall_work(flags))
+ /*
+ * It really matters that we check for flags != 0 and not
+ * just for pending work here!
+ */
+ if (flags)
do_work_pending(regs, flags);
trace_hardirqs_on();
@@ -45,6 +44,14 @@ noinstr void irqentry_enter_from_user_mode(struct pt_regs *regs)
noinstr void irqentry_exit_to_user_mode(struct pt_regs *regs)
{
+ unsigned long flags = read_thread_flags();
+
+ /*
+ * It really matters that we check for flags != 0 and not
+ * just for pending work here!
+ */
+ if (flags)
+ do_work_pending(regs, flags);
trace_hardirqs_on();
/* This context tracking call has inverse naming */
user_enter_callable();
diff --git a/arch/arm/kernel/signal.c b/arch/arm/kernel/signal.c
index 03b20637a2e1ca6f327223c9132915a97a9c4697..a39ee14ec5b70b3738ecef537709326ef1e8217c 100644
--- a/arch/arm/kernel/signal.c
+++ b/arch/arm/kernel/signal.c
@@ -598,8 +598,7 @@ static void arch_do_signal_or_restart(struct pt_regs *regs)
return;
}
-asmlinkage void
-do_work_pending(struct pt_regs *regs, unsigned int thread_flags)
+void do_work_pending(struct pt_regs *regs, unsigned int thread_flags)
{
/*
* The assembly code enters us with IRQs off, but it hasn't
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 25/31] ARM: entry: Create irqentry calls from kernel mode
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (23 preceding siblings ...)
2025-02-12 11:23 ` [PATCH v4 24/31] ARM: entry: Complete syscall and IRQ transition to C Linus Walleij
@ 2025-02-12 11:23 ` Linus Walleij
2025-02-12 11:23 ` [PATCH v4 26/31] ARM: entry: Move in-kernel hardirq tracing to C Linus Walleij
` (6 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:23 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
This adds a few calls to C when exceptions enter/exit from
the kernel itself, so these can be used for generic entry.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/include/asm/entry.h | 2 ++
arch/arm/kernel/entry-armv.S | 3 +++
| 5 +++++
arch/arm/kernel/entry.c | 8 ++++++++
4 files changed, 18 insertions(+)
diff --git a/arch/arm/include/asm/entry.h b/arch/arm/include/asm/entry.h
index 8b8cc9a1eefbd00dd75d4e344576fc184a413768..a78bc5054b093120e903f08369fb8b1a9feb7f81 100644
--- a/arch/arm/include/asm/entry.h
+++ b/arch/arm/include/asm/entry.h
@@ -12,5 +12,7 @@ long syscall_enter_from_user_mode(struct pt_regs *regs, long);
void syscall_exit_to_user_mode(struct pt_regs *regs);
void irqentry_enter_from_user_mode(struct pt_regs *regs);
void irqentry_exit_to_user_mode(struct pt_regs *regs);
+void irqentry_enter_from_kernel_mode(struct pt_regs *regs);
+void irqentry_exit_to_kernel_mode(struct pt_regs *regs);
#endif /* __ASM_ENTRY_H__ */
diff --git a/arch/arm/kernel/entry-armv.S b/arch/arm/kernel/entry-armv.S
index 3cd5a80b47758867d1db6c3971030370bca1551a..f64c4cc8beda6bcd469e6ff1a1f337d52dbbaf9c 100644
--- a/arch/arm/kernel/entry-armv.S
+++ b/arch/arm/kernel/entry-armv.S
@@ -205,6 +205,9 @@ ENDPROC(__und_invalid)
get_thread_info tsk
uaccess_entry tsk, r0, r1, r2, \uaccess
+ mov r0, sp @ 'regs'
+ bl irqentry_enter_from_kernel_mode
+
.if \trace
#ifdef CONFIG_TRACE_IRQFLAGS
bl trace_hardirqs_off
--git a/arch/arm/kernel/entry-header.S b/arch/arm/kernel/entry-header.S
index 51c7f93c1ded1db56f173c7aaed8094ef011ff5a..49a9c5cf6fd5fbb917f2ada6c0d6cc400b7d3fb3 100644
--- a/arch/arm/kernel/entry-header.S
+++ b/arch/arm/kernel/entry-header.S
@@ -210,6 +210,7 @@
.else
@ IRQs off again before pulling preserved data off the stack
disable_irq_notrace
+
#ifdef CONFIG_TRACE_IRQFLAGS
tst \rpsr, #PSR_I_BIT
bleq trace_hardirqs_on
@@ -217,6 +218,10 @@
blne trace_hardirqs_off
#endif
.endif
+
+ mov r0, sp @ 'regs'
+ bl irqentry_exit_to_kernel_mode
+
uaccess_exit tsk, r0, r1
#ifndef CONFIG_THUMB2_KERNEL
diff --git a/arch/arm/kernel/entry.c b/arch/arm/kernel/entry.c
index d7fdb9df33312dd2a914b59512b701748f33b597..674b5adcec0001b7d075d6936bfb4e318cb7ce74 100644
--- a/arch/arm/kernel/entry.c
+++ b/arch/arm/kernel/entry.c
@@ -56,3 +56,11 @@ noinstr void irqentry_exit_to_user_mode(struct pt_regs *regs)
/* This context tracking call has inverse naming */
user_enter_callable();
}
+
+noinstr void irqentry_enter_from_kernel_mode(struct pt_regs *regs)
+{
+}
+
+noinstr void irqentry_exit_to_kernel_mode(struct pt_regs *regs)
+{
+}
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 26/31] ARM: entry: Move in-kernel hardirq tracing to C
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (24 preceding siblings ...)
2025-02-12 11:23 ` [PATCH v4 25/31] ARM: entry: Create irqentry calls from kernel mode Linus Walleij
@ 2025-02-12 11:23 ` Linus Walleij
2025-02-12 11:23 ` [PATCH v4 27/31] ARM: irq: Add irqstack helper Linus Walleij
` (5 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:23 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
Move the code tracing hardirqs on/off into the C callbacks
for irqentry_enter_from_kernel_mode() and
irqentry_exit_to_kernel_mode().
The semantic difference occurred is that we alsways check
the PSR_I_BIT to determine if (hard) interrupts were enabled
or not. The assembly has a tweak to avoid this if we are
exiting an IRQ since it is obvious that IRQs must have been
enabled to get there, but for simplicity we just check it
for all exceptions.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/kernel/entry-armv.S | 13 ++++---------
| 19 ++-----------------
arch/arm/kernel/entry.c | 5 +++++
3 files changed, 11 insertions(+), 26 deletions(-)
diff --git a/arch/arm/kernel/entry-armv.S b/arch/arm/kernel/entry-armv.S
index f64c4cc8beda6bcd469e6ff1a1f337d52dbbaf9c..2a789c8834b93475c32dcb6ba5854e24ddd8d6e9 100644
--- a/arch/arm/kernel/entry-armv.S
+++ b/arch/arm/kernel/entry-armv.S
@@ -158,7 +158,7 @@ ENDPROC(__und_invalid)
#define SPFIX(code...)
#endif
- .macro svc_entry, stack_hole=0, trace=1, uaccess=1, overflow_check=1
+ .macro svc_entry, stack_hole=0 uaccess=1, overflow_check=1
UNWIND(.fnstart )
sub sp, sp, #(SVC_REGS_SIZE + \stack_hole)
THUMB( add sp, r1 ) @ get SP in a GPR without
@@ -208,11 +208,6 @@ ENDPROC(__und_invalid)
mov r0, sp @ 'regs'
bl irqentry_enter_from_kernel_mode
- .if \trace
-#ifdef CONFIG_TRACE_IRQFLAGS
- bl trace_hardirqs_off
-#endif
- .endif
.endm
.align 5
@@ -239,7 +234,7 @@ __irq_svc:
blne svc_preempt
#endif
- svc_exit r5, irq = 1 @ return from exception
+ svc_exit r5 @ return from exception
UNWIND(.fnend )
ENDPROC(__irq_svc)
@@ -303,7 +298,7 @@ ENDPROC(__pabt_svc)
.align 5
__fiq_svc:
- svc_entry trace=0
+ svc_entry
mov r0, sp @ struct pt_regs *regs
bl handle_fiq_as_nmi
svc_exit_via_fiq
@@ -321,7 +316,7 @@ ENDPROC(__fiq_svc)
@
.align 5
__fiq_abt:
- svc_entry trace=0
+ svc_entry
ARM( msr cpsr_c, #ABT_MODE | PSR_I_BIT | PSR_F_BIT )
THUMB( mov r0, #ABT_MODE | PSR_I_BIT | PSR_F_BIT )
--git a/arch/arm/kernel/entry-header.S b/arch/arm/kernel/entry-header.S
index 49a9c5cf6fd5fbb917f2ada6c0d6cc400b7d3fb3..cfaf14d71378ba14bbb2a42cd36d48a23838eee1 100644
--- a/arch/arm/kernel/entry-header.S
+++ b/arch/arm/kernel/entry-header.S
@@ -199,26 +199,11 @@
.endm
- .macro svc_exit, rpsr, irq = 0
- .if \irq != 0
- @ IRQs already off
-#ifdef CONFIG_TRACE_IRQFLAGS
- @ The parent context IRQs must have been enabled to get here in
- @ the first place, so there's no point checking the PSR I bit.
- bl trace_hardirqs_on
-#endif
- .else
+ .macro svc_exit, rpsr
+
@ IRQs off again before pulling preserved data off the stack
disable_irq_notrace
-#ifdef CONFIG_TRACE_IRQFLAGS
- tst \rpsr, #PSR_I_BIT
- bleq trace_hardirqs_on
- tst \rpsr, #PSR_I_BIT
- blne trace_hardirqs_off
-#endif
- .endif
-
mov r0, sp @ 'regs'
bl irqentry_exit_to_kernel_mode
diff --git a/arch/arm/kernel/entry.c b/arch/arm/kernel/entry.c
index 674b5adcec0001b7d075d6936bfb4e318cb7ce74..1e1284cc4caed6e602ce36e812d535e6fe324f34 100644
--- a/arch/arm/kernel/entry.c
+++ b/arch/arm/kernel/entry.c
@@ -59,8 +59,13 @@ noinstr void irqentry_exit_to_user_mode(struct pt_regs *regs)
noinstr void irqentry_enter_from_kernel_mode(struct pt_regs *regs)
{
+ trace_hardirqs_off();
}
noinstr void irqentry_exit_to_kernel_mode(struct pt_regs *regs)
{
+ if (interrupts_enabled(regs))
+ trace_hardirqs_on();
+ else
+ trace_hardirqs_off();
}
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 27/31] ARM: irq: Add irqstack helper
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (25 preceding siblings ...)
2025-02-12 11:23 ` [PATCH v4 26/31] ARM: entry: Move in-kernel hardirq tracing to C Linus Walleij
@ 2025-02-12 11:23 ` Linus Walleij
2025-02-12 11:23 ` [PATCH v4 28/31] ARM: entry: Convert to generic entry Linus Walleij
` (4 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:23 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
Add a helper to dispatch IRQ execution to the IRQ stack:
call_on_irq_stack() so we can explicitly issue handle_irq()
on the IRQ stack from a C program.
Cc: Ard Biesheuvel <ardb@kernel.org>
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/kernel/irq.c | 6 ++++++
arch/arm/kernel/irq.h | 2 ++
2 files changed, 8 insertions(+)
diff --git a/arch/arm/kernel/irq.c b/arch/arm/kernel/irq.c
index e1993e28a9ecfd80b55b2677253ac582467e6c14..f99d6b24d8ff56f901e4a817c61a1e57b7e532cc 100644
--- a/arch/arm/kernel/irq.c
+++ b/arch/arm/kernel/irq.c
@@ -43,6 +43,7 @@
#include <asm/mach/irq.h>
#include <asm/mach/time.h>
+#include "irq.h"
#include "reboot.h"
unsigned long irq_err_count;
@@ -71,6 +72,11 @@ static void __init init_irq_stacks(void)
}
}
+void call_on_irq_stack(void (*fn)(void *), void *arg)
+{
+ call_with_stack(fn, arg, __this_cpu_read(irq_stack_ptr));
+}
+
#ifdef CONFIG_SOFTIRQ_ON_OWN_STACK
static void ____do_softirq(void *arg)
{
diff --git a/arch/arm/kernel/irq.h b/arch/arm/kernel/irq.h
new file mode 100644
index 0000000000000000000000000000000000000000..80dd5bfe6403d68fc62fab30666d5645748d329f
--- /dev/null
+++ b/arch/arm/kernel/irq.h
@@ -0,0 +1,2 @@
+/* SPDX-License-Identifier: GPL-2.0 */
+void call_on_irq_stack(void (*fn)(void *), void *arg);
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 28/31] ARM: entry: Convert to generic entry
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (26 preceding siblings ...)
2025-02-12 11:23 ` [PATCH v4 27/31] ARM: irq: Add irqstack helper Linus Walleij
@ 2025-02-12 11:23 ` Linus Walleij
2025-02-12 11:23 ` [PATCH v4 29/31] ARM: entry: Handle dabt, pabt, and und as interrupts Linus Walleij
` (3 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:23 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
This rewrites ARM to use the generic entry.
All of the irqentry_* callbacks are collapsed and made to
call into the generic entry code that handle the interrupts
storing regs context before entry and restoring it after.
The syscall_enter_from_user_mode(), syscall_exit_to_user_mode(),
do_work_pending() constructs are replaced with generic code.
The syscall invocation now enables syscall tracing "work" using
the generic callbacks and no extra actions are needed to
enable/disable syscall tracing.
The old context tracking with user_exit_callable()
user_enter_callable() is no longer used. Only CSKY and Xtensa
use this deprecated mechanism now.
handle_fiq_as_nmi() is retired as the generic entry provides
the same functionality in irqentry_nmi_enter() and
irqentry_nmi_exit().
Like with the ARM64 proposed patch for generic entry, it is
necessary to create a small call "arm_exit_to_user_mode()"
that is used for things such as ret_from_fork, i.e. when
we create a completely new execution context with a botched
return from interrupt.
The elaborate checking and switching to IRQ stack in
the irq_handler macro was reimplemented in C for this
exercise: this was easiest and there was already code
making use of the IRQ stack from C for soft interrupts.
I have dropped the context tracking (i.e. calling
irqentry_enter() and looking for trace conditions) on
dabt/pabt i.e. do_DataAbort and do_PrefetchAbort code paths.
These are addressed in a separate patch for development/review
reasons.
I added stubs for PTRACE_SYSEMU and PTRACE_SYSEMU_SINGLESTEP
returning -EIO like the UM arch does, since those syscalls
are required by generic entry. If someone wants to implement
them, the placeholders are there.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/Kconfig | 1 +
arch/arm/include/asm/entry-common.h | 66 ++++++++++++++++++++++++++
arch/arm/include/asm/entry.h | 13 ++---
arch/arm/include/asm/ptrace.h | 3 --
arch/arm/include/asm/signal.h | 2 -
arch/arm/include/asm/syscall.h | 7 ++-
arch/arm/include/asm/thread_info.h | 17 +------
arch/arm/include/asm/traps.h | 1 -
arch/arm/include/uapi/asm/ptrace.h | 2 +
arch/arm/kernel/entry-armv.S | 50 ++++---------------
arch/arm/kernel/entry-common.S | 34 +++----------
| 18 -------
arch/arm/kernel/entry.c | 94 ++++++++++++++++++------------------
arch/arm/kernel/process.c | 5 +-
arch/arm/kernel/ptrace.c | 95 ++-----------------------------------
arch/arm/kernel/signal.c | 31 +-----------
arch/arm/kernel/syscall.c | 30 ++++--------
arch/arm/kernel/traps.c | 25 ----------
18 files changed, 161 insertions(+), 333 deletions(-)
diff --git a/arch/arm/Kconfig b/arch/arm/Kconfig
index 835b5f100e926e54bab4be7802aeada4af37a145..564575f52cf7ec8296938944fa5296a4acc2a598 100644
--- a/arch/arm/Kconfig
+++ b/arch/arm/Kconfig
@@ -69,6 +69,7 @@ config ARM
select GENERIC_CPU_AUTOPROBE
select GENERIC_CPU_DEVICES
select GENERIC_EARLY_IOREMAP
+ select GENERIC_ENTRY
select GENERIC_IDLE_POLL_SETUP
select GENERIC_IRQ_MULTI_HANDLER
select GENERIC_IRQ_PROBE
diff --git a/arch/arm/include/asm/entry-common.h b/arch/arm/include/asm/entry-common.h
new file mode 100644
index 0000000000000000000000000000000000000000..7b1f108512db0001a65911af0df4f48197d99e75
--- /dev/null
+++ b/arch/arm/include/asm/entry-common.h
@@ -0,0 +1,66 @@
+/* SPDX-License-Identifier: GPL-2.0 */
+
+#ifndef _ASM_ARM_ENTRY_COMMON_H
+#define _ASM_ARM_ENTRY_COMMON_H
+
+#include <linux/thread_info.h>
+
+#include <asm/stacktrace.h>
+
+enum ptrace_syscall_dir {
+ PTRACE_SYSCALL_ENTER = 0,
+ PTRACE_SYSCALL_EXIT,
+};
+
+static inline unsigned long
+arch_prepare_report_syscall_entry(struct pt_regs *regs)
+{
+ unsigned long ip;
+
+ /*
+ * IP is used to denote syscall entry/exit:
+ * IP = 0 -> entry
+ */
+ ip = regs->ARM_ip;
+ regs->ARM_ip = PTRACE_SYSCALL_ENTER;
+
+ return ip;
+}
+#define arch_prepare_report_syscall_entry arch_prepare_report_syscall_entry
+
+static inline void
+arch_post_report_syscall_entry(struct pt_regs *regs,
+ unsigned long saved_reg, long ret)
+{
+ regs->ARM_ip = saved_reg;
+}
+#define arch_post_report_syscall_entry arch_post_report_syscall_entry
+
+
+static inline unsigned long
+arch_prepare_report_syscall_exit(struct pt_regs *regs,
+ unsigned long work)
+{
+ unsigned long ip;
+
+ /*
+ * IP is used to denote syscall entry/exit:
+ * IP = 1 -> exit
+ */
+ ip = regs->ARM_ip;
+ regs->ARM_ip = PTRACE_SYSCALL_EXIT;
+
+ return ip;
+}
+#define arch_prepare_report_syscall_exit arch_prepare_report_syscall_exit
+
+static inline void
+arch_post_report_syscall_exit(struct pt_regs *regs,
+ unsigned long saved_reg,
+ unsigned long work)
+{
+ regs->ARM_ip = saved_reg;
+}
+#define arch_post_report_syscall_exit arch_post_report_syscall_exit
+
+#endif /* _ASM_ARM_ENTRY_COMMON_H */
diff --git a/arch/arm/include/asm/entry.h b/arch/arm/include/asm/entry.h
index a78bc5054b093120e903f08369fb8b1a9feb7f81..df2dd14ab51a586b83b6da6bbf9bd99858c1ddf4 100644
--- a/arch/arm/include/asm/entry.h
+++ b/arch/arm/include/asm/entry.h
@@ -4,15 +4,8 @@
struct pt_regs;
-/*
- * These are copies of generic entry headers so we can transition
- * to generic entry once they are semantically equivalent.
- */
-long syscall_enter_from_user_mode(struct pt_regs *regs, long);
-void syscall_exit_to_user_mode(struct pt_regs *regs);
-void irqentry_enter_from_user_mode(struct pt_regs *regs);
-void irqentry_exit_to_user_mode(struct pt_regs *regs);
-void irqentry_enter_from_kernel_mode(struct pt_regs *regs);
-void irqentry_exit_to_kernel_mode(struct pt_regs *regs);
+void arm_irq_handler(struct pt_regs *regs, int mode);
+void arm_fiq_handler(struct pt_regs *regs);
+void arm_exit_to_user_mode(struct pt_regs *regs);
#endif /* __ASM_ENTRY_H__ */
diff --git a/arch/arm/include/asm/ptrace.h b/arch/arm/include/asm/ptrace.h
index a01c66f0ad907882af2a383e53de8aca68c551e5..8a7de4020cbc016f743dcf515b9fa83b9bd3e3f0 100644
--- a/arch/arm/include/asm/ptrace.h
+++ b/arch/arm/include/asm/ptrace.h
@@ -200,8 +200,5 @@ static inline unsigned long it_advance(unsigned long cpsr)
return cpsr;
}
-int syscall_trace_enter(struct pt_regs *regs);
-void syscall_trace_exit(struct pt_regs *regs);
-
#endif /* __ASSEMBLY__ */
#endif
diff --git a/arch/arm/include/asm/signal.h b/arch/arm/include/asm/signal.h
index bb753cd3e29fc54186543c48b2aee7b221301a37..45885383646c9f166b0d7ee80080b78776b515e7 100644
--- a/arch/arm/include/asm/signal.h
+++ b/arch/arm/include/asm/signal.h
@@ -23,6 +23,4 @@ typedef struct {
#include <asm/sigcontext.h>
-void do_work_pending(struct pt_regs *regs, unsigned int thread_flags);
-
#endif
diff --git a/arch/arm/include/asm/syscall.h b/arch/arm/include/asm/syscall.h
index b94fd7ce17f82f0baf06b74800245112e1f13002..a6c8f0e799bb8bf3833161dec70fc7a1aaa513f6 100644
--- a/arch/arm/include/asm/syscall.h
+++ b/arch/arm/include/asm/syscall.h
@@ -19,7 +19,12 @@
extern const unsigned long sys_call_table[];
-int invoke_syscall(void *table, struct pt_regs *regs, int scno);
+void invoke_syscall(void *table, struct pt_regs *regs, int scno);
+
+static inline bool arch_syscall_is_vdso_sigreturn(struct pt_regs *regs)
+{
+ return false;
+}
static inline int syscall_get_nr(struct task_struct *task,
struct pt_regs *regs)
diff --git a/arch/arm/include/asm/thread_info.h b/arch/arm/include/asm/thread_info.h
index d8a45c5a10496aaf806bfeaa0353d5e8985bd6f5..84e58a9cdab63ad264c2cd2bad64239d1912cbe7 100644
--- a/arch/arm/include/asm/thread_info.h
+++ b/arch/arm/include/asm/thread_info.h
@@ -74,6 +74,7 @@ struct thread_info {
#ifdef CONFIG_ARM_THUMBEE
unsigned long thumbee_state; /* ThumbEE Handler Base register */
#endif
+ unsigned long syscall_work; /* SYSCALL_WORK_ flags */
};
#define INIT_THREAD_INFO(tsk) \
@@ -149,30 +150,14 @@ extern int vfp_restore_user_hwstate(struct user_vfp *,
#define TIF_USING_IWMMXT 17
#define TIF_MEMDIE 18 /* is terminating due to OOM killer */
#define TIF_RESTORE_SIGMASK 19
-#define TIF_SYSCALL_TRACE 20 /* syscall trace active */
-#define TIF_SYSCALL_AUDIT 21 /* syscall auditing active */
-#define TIF_SYSCALL_TRACEPOINT 22 /* syscall tracepoint instrumentation */
-#define TIF_SECCOMP 23 /* seccomp syscall filtering active */
-
#define _TIF_SIGPENDING (1 << TIF_SIGPENDING)
#define _TIF_NEED_RESCHED (1 << TIF_NEED_RESCHED)
#define _TIF_NOTIFY_RESUME (1 << TIF_NOTIFY_RESUME)
#define _TIF_UPROBE (1 << TIF_UPROBE)
-#define _TIF_SYSCALL_TRACE (1 << TIF_SYSCALL_TRACE)
-#define _TIF_SYSCALL_AUDIT (1 << TIF_SYSCALL_AUDIT)
-#define _TIF_SYSCALL_TRACEPOINT (1 << TIF_SYSCALL_TRACEPOINT)
-#define _TIF_SECCOMP (1 << TIF_SECCOMP)
#define _TIF_NOTIFY_SIGNAL (1 << TIF_NOTIFY_SIGNAL)
#define _TIF_USING_IWMMXT (1 << TIF_USING_IWMMXT)
-/* Checks for any syscall work in entry-common.S */
-#define _TIF_SYSCALL_WORK (_TIF_SYSCALL_TRACE | _TIF_SYSCALL_AUDIT | \
- _TIF_SYSCALL_TRACEPOINT | _TIF_SECCOMP)
-
-/*
- * Change these and you break ASM code in entry-common.S
- */
#define _TIF_WORK_MASK (_TIF_NEED_RESCHED | _TIF_SIGPENDING | \
_TIF_NOTIFY_RESUME | _TIF_UPROBE | \
_TIF_NOTIFY_SIGNAL)
diff --git a/arch/arm/include/asm/traps.h b/arch/arm/include/asm/traps.h
index b888912c2450e0e8eec8139166c7292f283545c7..f8695104c72f2f450092839c7c50920e540e594b 100644
--- a/arch/arm/include/asm/traps.h
+++ b/arch/arm/include/asm/traps.h
@@ -38,7 +38,6 @@ extern void *vectors_page;
asmlinkage void dump_backtrace_stm(u32 *stack, u32 instruction, const char *loglvl);
asmlinkage void do_undefinstr(struct pt_regs *regs);
-asmlinkage void handle_fiq_as_nmi(struct pt_regs *regs);
asmlinkage void bad_mode(struct pt_regs *regs, int reason);
int arm_syscall(int no, struct pt_regs *regs);
asmlinkage void baddataabort(int code, unsigned long instr, struct pt_regs *regs);
diff --git a/arch/arm/include/uapi/asm/ptrace.h b/arch/arm/include/uapi/asm/ptrace.h
index 8896c23ccba78ed7e10c7d1b9955124314496dc6..336a5beb282f360e875703045ee833974bbd68b2 100644
--- a/arch/arm/include/uapi/asm/ptrace.h
+++ b/arch/arm/include/uapi/asm/ptrace.h
@@ -33,6 +33,8 @@
#define PTRACE_GETHBPREGS 29
#define PTRACE_SETHBPREGS 30
#define PTRACE_GETFDPIC 31
+#define PTRACE_SYSEMU 32
+#define PTRACE_SYSEMU_SINGLESTEP 33
#define PTRACE_GETFDPIC_EXEC 0
#define PTRACE_GETFDPIC_INTERP 1
diff --git a/arch/arm/kernel/entry-armv.S b/arch/arm/kernel/entry-armv.S
index 2a789c8834b93475c32dcb6ba5854e24ddd8d6e9..498a22f5fa8a6516d1eab584f0523c3d6c6e4926 100644
--- a/arch/arm/kernel/entry-armv.S
+++ b/arch/arm/kernel/entry-armv.S
@@ -36,35 +36,6 @@
#define RELOC_TEXT_NONE
#endif
-/*
- * Interrupt handling.
- */
- .macro irq_handler, from_user:req
- mov r1, sp
- ldr_this_cpu r2, irq_stack_ptr, r2, r3
- .if \from_user == 0
- @
- @ If we took the interrupt while running in the kernel, we may already
- @ be using the IRQ stack, so revert to the original value in that case.
- @
- subs r3, r2, r1 @ SP above bottom of IRQ stack?
- rsbscs r3, r3, #THREAD_SIZE @ ... and below the top?
-#ifdef CONFIG_VMAP_STACK
- ldr_va r3, high_memory, cc @ End of the linear region
- cmpcc r3, r1 @ Stack pointer was below it?
-#endif
- bcc 0f @ If not, switch to the IRQ stack
- mov r0, r1
- bl generic_handle_arch_irq
- b 1f
-0:
- .endif
-
- mov_l r0, generic_handle_arch_irq
- bl call_with_stack
-1:
- .endm
-
.macro pabt_helper
@ PABORT handler takes pt_regs in r2, fault address in r4 and psr in r5
#ifdef MULTI_PABORT
@@ -205,9 +176,6 @@ ENDPROC(__und_invalid)
get_thread_info tsk
uaccess_entry tsk, r0, r1, r2, \uaccess
- mov r0, sp @ 'regs'
- bl irqentry_enter_from_kernel_mode
-
.endm
.align 5
@@ -223,7 +191,9 @@ ENDPROC(__dabt_svc)
.align 5
__irq_svc:
svc_entry
- irq_handler from_user=0
+ mov r0, sp @ regs
+ mov r1, #0 @ from kernel mode
+ bl arm_irq_handler
#ifdef CONFIG_PREEMPTION
ldr r8, [tsk, #TI_PREEMPT] @ get preempt count
@@ -300,7 +270,7 @@ ENDPROC(__pabt_svc)
__fiq_svc:
svc_entry
mov r0, sp @ struct pt_regs *regs
- bl handle_fiq_as_nmi
+ bl arm_fiq_handler
svc_exit_via_fiq
UNWIND(.fnend )
ENDPROC(__fiq_svc)
@@ -329,7 +299,7 @@ __fiq_abt:
stmfd sp!, {r1 - r2}
add r0, sp, #8 @ struct pt_regs *regs
- bl handle_fiq_as_nmi
+ bl arm_fiq_handler
ldmfd sp!, {r1 - r2}
ARM( msr cpsr_c, #ABT_MODE | PSR_I_BIT | PSR_F_BIT )
@@ -420,7 +390,6 @@ ENDPROC(__fiq_abt)
.align 5
__dabt_usr:
usr_entry uaccess=0
- asm_irqentry_enter_from_user_mode
kuser_cmpxchg_check
mov r2, sp
dabt_helper
@@ -431,9 +400,10 @@ ENDPROC(__dabt_usr)
.align 5
__irq_usr:
usr_entry
- asm_irqentry_enter_from_user_mode
kuser_cmpxchg_check
- irq_handler from_user=1
+ mov r0, sp @ regs
+ mov r1, #1 @ from user mode
+ bl arm_irq_handler
get_thread_info tsk
mov why, #0
b ret_to_user_from_irq
@@ -445,7 +415,6 @@ ENDPROC(__irq_usr)
.align 5
__und_usr:
usr_entry uaccess=0
- asm_irqentry_enter_from_user_mode
@ IRQs must be enabled before attempting to read the instruction from
@ user space since that could cause a page/translation fault if the
@@ -470,7 +439,6 @@ ENDPROC(__und_usr)
.align 5
__pabt_usr:
usr_entry
- asm_irqentry_enter_from_user_mode
mov r2, sp @ regs
pabt_helper
UNWIND(.fnend )
@@ -493,7 +461,7 @@ __fiq_usr:
usr_entry
kuser_cmpxchg_check
mov r0, sp @ struct pt_regs *regs
- bl handle_fiq_as_nmi
+ bl arm_fiq_handler
get_thread_info tsk
restore_user_regs
UNWIND(.fnend )
diff --git a/arch/arm/kernel/entry-common.S b/arch/arm/kernel/entry-common.S
index 6b0f86786a7d9e656ecbe8c741f2edbcbafd2915..a5a0e757ab23d151485092a078af98966e039869 100644
--- a/arch/arm/kernel/entry-common.S
+++ b/arch/arm/kernel/entry-common.S
@@ -39,10 +39,6 @@ ret_fast_syscall:
UNWIND(.fnstart )
UNWIND(.cantunwind )
add sp, sp, #(S_R0 + S_OFF)
- /* do_rseq_syscall needs interrupts enabled. */
- mov r0, sp @ 'regs'
- bl syscall_exit_to_user_mode
-
#ifdef CONFIG_GCC_PLUGIN_STACKLEAK
bl stackleak_erase_on_task_stack
#endif
@@ -60,7 +56,7 @@ ENTRY(ret_to_user)
/* do_rseq_syscall needs interrupts enabled. */
enable_irq_notrace @ enable interrupts
mov r0, sp @ 'regs'
- bl syscall_exit_to_user_mode
+ bl arm_exit_to_user_mode
#ifdef CONFIG_GCC_PLUGIN_STACKLEAK
bl stackleak_erase_on_task_stack
@@ -69,8 +65,6 @@ ENTRY(ret_to_user)
ENDPROC(ret_to_user)
ENTRY(ret_to_user_from_irq)
- asm_irqentry_exit_to_user_mode
-
#ifdef CONFIG_GCC_PLUGIN_STACKLEAK
bl stackleak_erase_on_task_stack
#endif
@@ -92,9 +86,10 @@ SYM_TYPED_FUNC_START(ret_from_fork_asm)
mov r2, r5
mov r3, r4
bl ret_from_fork
- mov r0, #0
-1: get_thread_info tsk
- b ret_to_user
+#ifdef CONFIG_GCC_PLUGIN_STACKLEAK
+ bl stackleak_erase_on_task_stack
+#endif
+ restore_user_regs
SYM_FUNC_END(ret_from_fork_asm)
/*=============================================================================
@@ -210,29 +205,12 @@ ENTRY(vector_swi)
#else
str scno, [tsk, #TI_ABI_SYSCALL]
#endif
-
- /*
- * Calling out to C to be careful to save and restore registers.
- * This call could modify the syscall number. scno is r7 so we
- * do not save and restore r7.
- */
- mov r0, sp @ regs
- mov r1, scno
- push {r4 - r6, r8 - r10, lr}
- bl syscall_enter_from_user_mode
- pop {r4 - r6, r8 - r10, lr}
- mov scno, r0
-
mov r1, sp @ put regs into r1
stmdb sp!, {r4, r5} @ push fifth and sixth args
mov r0, tbl
mov r2, scno @ syscall number from r7
bl invoke_syscall
- cmp r0, #0
- beq ret_fast_syscall
- /* This path taken when tracing */
- add sp, sp, #(S_R0 + S_OFF)
- b ret_to_user
+ b ret_fast_syscall
#if defined(CONFIG_OABI_COMPAT) || !defined(CONFIG_AEABI)
/*
--git a/arch/arm/kernel/entry-header.S b/arch/arm/kernel/entry-header.S
index cfaf14d71378ba14bbb2a42cd36d48a23838eee1..abd927f1cc3a4d813cc48bc046191f49f3c0e533 100644
--- a/arch/arm/kernel/entry-header.S
+++ b/arch/arm/kernel/entry-header.S
@@ -203,10 +203,6 @@
@ IRQs off again before pulling preserved data off the stack
disable_irq_notrace
-
- mov r0, sp @ 'regs'
- bl irqentry_exit_to_kernel_mode
-
uaccess_exit tsk, r0, r1
#ifndef CONFIG_THUMB2_KERNEL
@@ -340,20 +336,6 @@ ALT_UP_B(.L1_\@)
#endif /* !CONFIG_THUMB2_KERNEL */
.endm
-/*
- * Context tracking and other mode transitions. Used to instrument transitions
- * between user and kernel mode.
-*/
- .macro asm_irqentry_enter_from_user_mode
- mov r0, sp @ regs
- bl irqentry_enter_from_user_mode
- .endm
-
- .macro asm_irqentry_exit_to_user_mode
- mov r0, sp @ regs
- bl irqentry_exit_to_user_mode
- .endm
-
/*
* These are the registers used in the syscall handler, and allow us to
* have in theory up to 7 arguments to a function - r0 to r6.
diff --git a/arch/arm/kernel/entry.c b/arch/arm/kernel/entry.c
index 1e1284cc4caed6e602ce36e812d535e6fe324f34..08b8680233e5aae6e477f9bf15d7f06d7fd3e5af 100644
--- a/arch/arm/kernel/entry.c
+++ b/arch/arm/kernel/entry.c
@@ -1,71 +1,71 @@
// SPDX-License-Identifier: GPL-2.0
#include <asm/entry.h>
-#include <asm/ptrace.h>
-#include <asm/signal.h>
#include <linux/context_tracking.h>
+#include <linux/entry-common.h>
+#include <linux/hardirq.h>
+#include <linux/irq.h>
#include <linux/irqflags.h>
+#include <linux/percpu.h>
#include <linux/rseq.h>
+#include <asm/stacktrace.h>
-long syscall_enter_from_user_mode(struct pt_regs *regs, long syscall)
-{
- trace_hardirqs_on();
- local_irq_enable();
- /* This context tracking call has inverse naming */
- user_exit_callable();
-
- /* This will optionally be modified later */
- return syscall;
-}
+#include "irq.h"
-void syscall_exit_to_user_mode(struct pt_regs *regs)
+static void noinstr handle_arm_irq(void *data)
{
- unsigned long flags = read_thread_flags();
+ struct pt_regs *regs = data;
+ struct pt_regs *old_regs;
- rseq_syscall(regs);
- local_irq_disable();
- /*
- * It really matters that we check for flags != 0 and not
- * just for pending work here!
- */
- if (flags)
- do_work_pending(regs, flags);
+ irq_enter_rcu();
+ old_regs = set_irq_regs(regs);
- trace_hardirqs_on();
- /* This context tracking call has inverse naming */
- user_enter_callable();
-}
+ handle_arch_irq(regs);
-noinstr void irqentry_enter_from_user_mode(struct pt_regs *regs)
-{
- trace_hardirqs_off();
- /* This context tracking call has inverse naming */
- user_exit_callable();
+ set_irq_regs(old_regs);
+ irq_exit_rcu();
}
-noinstr void irqentry_exit_to_user_mode(struct pt_regs *regs)
+noinstr void arm_irq_handler(struct pt_regs *regs, int mode)
{
- unsigned long flags = read_thread_flags();
+ irqentry_state_t state = irqentry_enter(regs);
/*
- * It really matters that we check for flags != 0 and not
- * just for pending work here!
+ * If we are executing in usermode, or kernel process context
+ * (on the thread stack) then switch to the IRQ stack. Else we
+ * are already on the IRQ stack (or the overflow stack) and we
+ * can just proceed to handle the IRQ.
*/
- if (flags)
- do_work_pending(regs, flags);
- trace_hardirqs_on();
- /* This context tracking call has inverse naming */
- user_enter_callable();
+ if (mode == 1)
+ call_on_irq_stack(handle_arm_irq, regs);
+ else if (on_thread_stack())
+ call_on_irq_stack(handle_arm_irq, regs);
+ else
+ handle_arm_irq(regs);
+
+ irqentry_exit(regs, state);
}
-noinstr void irqentry_enter_from_kernel_mode(struct pt_regs *regs)
+/*
+ * Handle FIQ similarly to NMI on x86 systems.
+ *
+ * The runtime environment for NMIs is extremely restrictive
+ * (NMIs can pre-empt critical sections meaning almost all locking is
+ * forbidden) meaning this default FIQ handling must only be used in
+ * circumstances where non-maskability improves robustness, such as
+ * watchdog or debug logic.
+ *
+ * This handler is not appropriate for general purpose use in drivers
+ * platform code and can be overrideen using set_fiq_handler.
+ */
+noinstr void arm_fiq_handler(struct pt_regs *regs)
{
- trace_hardirqs_off();
+ irqentry_state_t state = irqentry_nmi_enter(regs);
+
+ irqentry_nmi_exit(regs, state);
}
-noinstr void irqentry_exit_to_kernel_mode(struct pt_regs *regs)
+asmlinkage void arm_exit_to_user_mode(struct pt_regs *regs)
{
- if (interrupts_enabled(regs))
- trace_hardirqs_on();
- else
- trace_hardirqs_off();
+ local_irq_disable();
+ irqentry_exit_to_user_mode(regs);
}
diff --git a/arch/arm/kernel/process.c b/arch/arm/kernel/process.c
index 81c1c3f988344185917cf4e53d0d8ee47ae912f4..a59ad5fdaaf51f653c352884bbdd10d1211b2669 100644
--- a/arch/arm/kernel/process.c
+++ b/arch/arm/kernel/process.c
@@ -25,6 +25,7 @@
#include <linux/random.h>
#include <linux/hw_breakpoint.h>
#include <linux/leds.h>
+#include <linux/entry-common.h>
#include <asm/processor.h>
#include <asm/thread_notify.h>
@@ -248,8 +249,8 @@ __visible void ret_from_fork(struct task_struct *prev, struct pt_regs *regs,
* execve() syscall.
*/
}
-
- /* syscall_exit_to_user_mode(regs); here once we switch to generic entry */
+ local_irq_enable();
+ syscall_exit_to_user_mode(regs);
}
int copy_thread(struct task_struct *p, const struct kernel_clone_args *args)
diff --git a/arch/arm/kernel/ptrace.c b/arch/arm/kernel/ptrace.c
index ac7b98ae47249b4a00fc1cb871bb2d309bfd8c88..d1e5a78c616b0e71f9466c201b74e06ce0d7a08f 100644
--- a/arch/arm/kernel/ptrace.c
+++ b/arch/arm/kernel/ptrace.c
@@ -27,9 +27,6 @@
#include <asm/syscall.h>
#include <asm/traps.h>
-#define CREATE_TRACE_POINTS
-#include <trace/events/syscalls.h>
-
#define REG_PC 15
#define REG_PSR 16
/*
@@ -737,6 +734,11 @@ long arch_ptrace(struct task_struct *child, long request,
ret = ptrace_write_user(child, addr, data);
break;
+ case PTRACE_SYSEMU:
+ case PTRACE_SYSEMU_SINGLESTEP:
+ ret = -EIO;
+ break;
+
case PTRACE_GETREGS:
ret = copy_regset_to_user(child,
&user_arm_view, REGSET_GPR,
@@ -821,90 +823,3 @@ long arch_ptrace(struct task_struct *child, long request,
return ret;
}
-
-enum ptrace_syscall_dir {
- PTRACE_SYSCALL_ENTER = 0,
- PTRACE_SYSCALL_EXIT,
-};
-
-static void report_syscall_enter(struct pt_regs *regs)
-{
- unsigned long ip;
-
- /*
- * IP is used to denote syscall entry/exit:
- * IP = 0 -> entry
- */
- ip = regs->ARM_ip;
- regs->ARM_ip = PTRACE_SYSCALL_ENTER;
-
- if (ptrace_report_syscall_entry(regs))
- current_thread_info()->abi_syscall = -1;
-
- regs->ARM_ip = ip;
-}
-
-static void report_syscall_exit(struct pt_regs *regs)
-{
- unsigned long ip;
-
- /*
- * IP is used to denote syscall entry/exit:
- * IP = 1 -> exit
- */
- ip = regs->ARM_ip;
- regs->ARM_ip = PTRACE_SYSCALL_EXIT;
-
- ptrace_report_syscall_exit(regs, 0);
-
- regs->ARM_ip = ip;
-}
-
-asmlinkage int syscall_trace_enter(struct pt_regs *regs)
-{
- int scno;
-
- if (test_thread_flag(TIF_SYSCALL_TRACE))
- report_syscall_enter(regs);
-
- /* Do seccomp after ptrace; syscall may have changed. */
-#ifdef CONFIG_HAVE_ARCH_SECCOMP_FILTER
- if (secure_computing() == -1)
- return -1;
-#else
- /* XXX: remove this once OABI gets fixed */
- secure_computing_strict(syscall_get_nr(current, regs));
-#endif
-
- /* Tracer or seccomp may have changed syscall. */
- scno = syscall_get_nr(current, regs);
-
- if (test_thread_flag(TIF_SYSCALL_TRACEPOINT))
- trace_sys_enter(regs, scno);
-
- audit_syscall_entry(scno, regs->ARM_r0, regs->ARM_r1, regs->ARM_r2,
- regs->ARM_r3);
-
- return scno;
-}
-
-void syscall_trace_exit(struct pt_regs *regs)
-{
- /*
- * Audit the syscall before anything else, as a debugger may
- * come in and change the current registers.
- */
- audit_syscall_exit(regs);
-
- /*
- * Note that we haven't updated the ->syscall field for the
- * current thread. This isn't a problem because it will have
- * been set on syscall entry and there hasn't been an opportunity
- * for a PTRACE_SET_SYSCALL since then.
- */
- if (test_thread_flag(TIF_SYSCALL_TRACEPOINT))
- trace_sys_exit(regs, regs_return_value(regs));
-
- if (test_thread_flag(TIF_SYSCALL_TRACE))
- report_syscall_exit(regs);
-}
diff --git a/arch/arm/kernel/signal.c b/arch/arm/kernel/signal.c
index a39ee14ec5b70b3738ecef537709326ef1e8217c..35d2bb3cd2b442dac164548037262e065fbfe12a 100644
--- a/arch/arm/kernel/signal.c
+++ b/arch/arm/kernel/signal.c
@@ -4,6 +4,7 @@
*
* Copyright (C) 1995-2009 Russell King
*/
+#include <linux/entry-common.h>
#include <linux/errno.h>
#include <linux/random.h>
#include <linux/signal.h>
@@ -535,7 +536,7 @@ static void handle_signal(struct ksignal *ksig, struct pt_regs *regs)
* the kernel can handle, and then we build all the user-level signal handling
* stack-frames in one go after that.
*/
-static void arch_do_signal_or_restart(struct pt_regs *regs)
+void arch_do_signal_or_restart(struct pt_regs *regs)
{
unsigned int retval = 0, continue_addr = 0, restart_addr = 0;
bool syscall = (syscall_get_nr(current, regs) != -1);
@@ -598,34 +599,6 @@ static void arch_do_signal_or_restart(struct pt_regs *regs)
return;
}
-void do_work_pending(struct pt_regs *regs, unsigned int thread_flags)
-{
- /*
- * The assembly code enters us with IRQs off, but it hasn't
- * informed the tracing code of that for efficiency reasons.
- * Update the trace code with the current status.
- */
- trace_hardirqs_off();
- do {
- if (likely(thread_flags & _TIF_NEED_RESCHED)) {
- schedule();
- } else {
- if (unlikely(!user_mode(regs)))
- return;
- local_irq_enable();
- if (thread_flags & (_TIF_SIGPENDING | _TIF_NOTIFY_SIGNAL)) {
- arch_do_signal_or_restart(regs);
- } else if (thread_flags & _TIF_UPROBE) {
- uprobe_notify_resume(regs);
- } else {
- resume_user_mode_work(regs);
- }
- }
- local_irq_disable();
- thread_flags = read_thread_flags();
- } while (thread_flags & _TIF_WORK_MASK);
-}
-
struct page *get_signal_page(void)
{
unsigned long ptr;
diff --git a/arch/arm/kernel/syscall.c b/arch/arm/kernel/syscall.c
index 6f518ec3870fb36a64d1b671b20f96d63a52bf78..ed3ab51283c06c1398ece2ad3ee1fae16cd03ee8 100644
--- a/arch/arm/kernel/syscall.c
+++ b/arch/arm/kernel/syscall.c
@@ -1,25 +1,21 @@
// SPDX-License-Identifier: GPL-2.0
+#include <linux/entry-common.h>
#include <linux/syscalls.h>
#include <asm/syscall.h>
-static inline bool has_syscall_work(unsigned long flags)
-{
- return unlikely(flags & _TIF_SYSCALL_WORK);
-}
-
int invoke_syscall_asm(void *table, struct pt_regs *regs, int scno);
__ADDRESSABLE(invoke_syscall_asm);
-__visible int invoke_syscall(void *table, struct pt_regs *regs, int scno)
+__visible void invoke_syscall(void *table, struct pt_regs *regs, int scno)
{
- unsigned long flags = read_thread_flags();
int ret;
- if (has_syscall_work(flags)) {
- scno = syscall_trace_enter(regs);
- if (scno == -1)
- goto trace_exit_nosave;
+ scno = syscall_enter_from_user_mode(regs, scno);
+ /* When tracing syscall -1 means "skip syscall" */
+ if (scno < 0) {
+ ret = 0;
+ goto exit_save;
}
if (scno < NR_syscalls) {
@@ -35,13 +31,7 @@ __visible int invoke_syscall(void *table, struct pt_regs *regs, int scno)
ret = sys_ni_syscall();
exit_save:
- /* Save return value from syscall */
- regs->ARM_r0 = ret;
- if (!has_syscall_work(flags))
- return 0;
-
-trace_exit_nosave:
- local_irq_enable();
- syscall_trace_exit(regs);
- return 1;
+ syscall_set_return_value(current, regs, 0, ret);
+
+ syscall_exit_to_user_mode(regs);
}
diff --git a/arch/arm/kernel/traps.c b/arch/arm/kernel/traps.c
index 942c77767919466d5b89d83d8eb77032b119473e..57e74b49c52533babe177a8a2e0160a83c9bde05 100644
--- a/arch/arm/kernel/traps.c
+++ b/arch/arm/kernel/traps.c
@@ -502,31 +502,6 @@ asmlinkage void do_undefinstr(struct pt_regs *regs)
}
NOKPROBE_SYMBOL(do_undefinstr)
-/*
- * Handle FIQ similarly to NMI on x86 systems.
- *
- * The runtime environment for NMIs is extremely restrictive
- * (NMIs can pre-empt critical sections meaning almost all locking is
- * forbidden) meaning this default FIQ handling must only be used in
- * circumstances where non-maskability improves robustness, such as
- * watchdog or debug logic.
- *
- * This handler is not appropriate for general purpose use in drivers
- * platform code and can be overrideen using set_fiq_handler.
- */
-asmlinkage void __exception_irq_entry handle_fiq_as_nmi(struct pt_regs *regs)
-{
- struct pt_regs *old_regs = set_irq_regs(regs);
-
- nmi_enter();
-
- /* nop. FIQ handlers for special arch/arm features can be added here. */
-
- nmi_exit();
-
- set_irq_regs(old_regs);
-}
-
/*
* bad_mode handles the impossible case in the vectors. If you see one of
* these, then it's extremely serious, and could mean you have buggy hardware.
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 29/31] ARM: entry: Handle dabt, pabt, and und as interrupts
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (27 preceding siblings ...)
2025-02-12 11:23 ` [PATCH v4 28/31] ARM: entry: Convert to generic entry Linus Walleij
@ 2025-02-12 11:23 ` Linus Walleij
2025-02-12 11:23 ` [PATCH v4 30/31] ARM: entry: Block IRQs in early IRQ context Linus Walleij
` (2 subsequent siblings)
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:23 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
While it isn't entirely intuitive, it appears that any kind
of exception such as data or prefetch abort ("page faults")
need to be handled as some kind of "interrupts" when using
generic entry. At least this is what other platforms are doing.
The same goes for undefined instruction handling, i.e.
floating point emulation in the kernel (und-exceptions).
This is necessary for the context checking to pass: without
this patch, a whole slew of warnings start to trigger
from syscall_exit_to_user_mode_prepare()
CT_WARN_ON(ct_state() != CT_STATE_KERNEL), i.e. syscalls
seems to exit from user mode to user mode (not good),
because the page faults screws up the context tracker.
This patch restores the order.
If this seems like the previous patch introduces a
regression that is then fixed in this patch, it can simply
be squashed into the former: having this rewrite separately
surely makes development and review easier.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/include/asm/entry.h | 3 +++
arch/arm/include/asm/traps.h | 2 +-
arch/arm/kernel/entry-armv.S | 7 +++----
arch/arm/kernel/entry.c | 31 +++++++++++++++++++++++++++++++
arch/arm/kernel/traps.c | 2 +-
arch/arm/mm/abort-ev4.S | 2 +-
arch/arm/mm/abort-ev4t.S | 2 +-
arch/arm/mm/abort-ev5t.S | 4 ++--
arch/arm/mm/abort-ev5tj.S | 6 +++---
arch/arm/mm/abort-ev6.S | 2 +-
arch/arm/mm/abort-ev7.S | 2 +-
arch/arm/mm/abort-lv4t.S | 36 ++++++++++++++++++------------------
arch/arm/mm/abort-macro.S | 2 +-
arch/arm/mm/abort-nommu.S | 2 +-
arch/arm/mm/fault.c | 4 ++--
arch/arm/mm/fault.h | 8 ++++----
arch/arm/mm/pabort-legacy.S | 2 +-
arch/arm/mm/pabort-v6.S | 2 +-
arch/arm/mm/pabort-v7.S | 2 +-
19 files changed, 77 insertions(+), 44 deletions(-)
diff --git a/arch/arm/include/asm/entry.h b/arch/arm/include/asm/entry.h
index df2dd14ab51a586b83b6da6bbf9bd99858c1ddf4..de722084d5309194e02cb87db64ec32b9e6cd627 100644
--- a/arch/arm/include/asm/entry.h
+++ b/arch/arm/include/asm/entry.h
@@ -4,6 +4,9 @@
struct pt_regs;
+void arm_und_handler(struct pt_regs *regs);
+void arm_dabt_handler(unsigned long addr, unsigned int fsr, struct pt_regs *regs);
+void arm_pabt_handler(unsigned long addr, unsigned int ifsr, struct pt_regs *regs);
void arm_irq_handler(struct pt_regs *regs, int mode);
void arm_fiq_handler(struct pt_regs *regs);
void arm_exit_to_user_mode(struct pt_regs *regs);
diff --git a/arch/arm/include/asm/traps.h b/arch/arm/include/asm/traps.h
index f8695104c72f2f450092839c7c50920e540e594b..0ca98271e10a210225c1714f5b24668cebab2def 100644
--- a/arch/arm/include/asm/traps.h
+++ b/arch/arm/include/asm/traps.h
@@ -37,7 +37,7 @@ extern void ptrace_break(struct pt_regs *regs);
extern void *vectors_page;
asmlinkage void dump_backtrace_stm(u32 *stack, u32 instruction, const char *loglvl);
-asmlinkage void do_undefinstr(struct pt_regs *regs);
+void do_undefinstr(struct pt_regs *regs);
asmlinkage void bad_mode(struct pt_regs *regs, int reason);
int arm_syscall(int no, struct pt_regs *regs);
asmlinkage void baddataabort(int code, unsigned long instr, struct pt_regs *regs);
diff --git a/arch/arm/kernel/entry-armv.S b/arch/arm/kernel/entry-armv.S
index 498a22f5fa8a6516d1eab584f0523c3d6c6e4926..ae2f952beea7611f0abc7bd299fc944335a21219 100644
--- a/arch/arm/kernel/entry-armv.S
+++ b/arch/arm/kernel/entry-armv.S
@@ -230,7 +230,7 @@ __und_fault:
ldr r2, [r0, #S_PC]
sub r2, r2, r1
str r2, [r0, #S_PC]
- b do_undefinstr
+ b arm_und_handler
ENDPROC(__und_fault)
.align 5
@@ -449,9 +449,8 @@ __pabt_usr:
ENTRY(ret_from_exception)
UNWIND(.fnstart )
UNWIND(.cantunwind )
- get_thread_info tsk
- mov why, #0
- b ret_to_user
+ disable_irq_notrace
+ b ret_to_user_from_irq
UNWIND(.fnend )
ENDPROC(__pabt_usr)
ENDPROC(ret_from_exception)
diff --git a/arch/arm/kernel/entry.c b/arch/arm/kernel/entry.c
index 08b8680233e5aae6e477f9bf15d7f06d7fd3e5af..01e4339ccdb4951e04a89fb91ad7c6e1991e09af 100644
--- a/arch/arm/kernel/entry.c
+++ b/arch/arm/kernel/entry.c
@@ -8,8 +8,39 @@
#include <linux/percpu.h>
#include <linux/rseq.h>
#include <asm/stacktrace.h>
+#include <asm/traps.h>
#include "irq.h"
+#include "../mm/fault.h"
+
+noinstr asmlinkage void arm_und_handler(struct pt_regs *regs)
+{
+ irqentry_state_t state = irqentry_enter(regs);
+
+ do_undefinstr(regs);
+
+ irqentry_exit(regs, state);
+}
+
+noinstr asmlinkage void arm_dabt_handler(unsigned long addr, unsigned int fsr,
+ struct pt_regs *regs)
+{
+ irqentry_state_t state = irqentry_enter(regs);
+
+ do_DataAbort(addr, fsr, regs);
+
+ irqentry_exit(regs, state);
+}
+
+noinstr asmlinkage void arm_pabt_handler(unsigned long addr, unsigned int ifsr,
+ struct pt_regs *regs)
+{
+ irqentry_state_t state = irqentry_enter(regs);
+
+ do_PrefetchAbort(addr, ifsr, regs);
+
+ irqentry_exit(regs, state);
+}
static void noinstr handle_arm_irq(void *data)
{
diff --git a/arch/arm/kernel/traps.c b/arch/arm/kernel/traps.c
index 57e74b49c52533babe177a8a2e0160a83c9bde05..d3a689bd05955f1ae46a6341e456bb097e831311 100644
--- a/arch/arm/kernel/traps.c
+++ b/arch/arm/kernel/traps.c
@@ -449,7 +449,7 @@ int call_undef_hook(struct pt_regs *regs, unsigned int instr)
return fn ? fn(regs, instr) : 1;
}
-asmlinkage void do_undefinstr(struct pt_regs *regs)
+noinstr void do_undefinstr(struct pt_regs *regs)
{
unsigned int instr;
void __user *pc;
diff --git a/arch/arm/mm/abort-ev4.S b/arch/arm/mm/abort-ev4.S
index a10bcb89594dd38ce31ca30bf97d68cf421afc42..68d8ae2f4e9c0955c9f08112b4ff1887867d550a 100644
--- a/arch/arm/mm/abort-ev4.S
+++ b/arch/arm/mm/abort-ev4.S
@@ -24,4 +24,4 @@ ENTRY(v4_early_abort)
bic r1, r1, #1 << 11 | 1 << 10 @ clear bits 11 and 10 of FSR
tst r3, #1 << 20 @ L = 1 -> write?
orreq r1, r1, #1 << 11 @ yes.
- b do_DataAbort
+ b arm_dabt_handler
diff --git a/arch/arm/mm/abort-ev4t.S b/arch/arm/mm/abort-ev4t.S
index 14743a2f6997fcae0eeb55f53dec07b7b2ec227c..76c6cfa79e2a32e34b6f35a15c71c1af407956f3 100644
--- a/arch/arm/mm/abort-ev4t.S
+++ b/arch/arm/mm/abort-ev4t.S
@@ -25,4 +25,4 @@ ENTRY(v4t_early_abort)
bic r1, r1, #1 << 11 | 1 << 10 @ clear bits 11 and 10 of FSR
tst r3, #1 << 20 @ check write
orreq r1, r1, #1 << 11
- b do_DataAbort
+ b arm_dabt_handler
diff --git a/arch/arm/mm/abort-ev5t.S b/arch/arm/mm/abort-ev5t.S
index 98c523118820798668bf04d065ad86ea05fb6d2c..fda1e54debc44b380fb3f3af95c1f198a9cc008a 100644
--- a/arch/arm/mm/abort-ev5t.S
+++ b/arch/arm/mm/abort-ev5t.S
@@ -25,7 +25,7 @@ ENTRY(v5t_early_abort)
uaccess_disable ip @ disable user access
bic r1, r1, #1 << 11 @ clear bits 11 of FSR
teq_ldrd tmp=ip, insn=r3 @ insn was LDRD?
- beq do_DataAbort @ yes
+ beq arm_dabt_handler @ yes
tst r3, #1 << 20 @ check write
orreq r1, r1, #1 << 11
- b do_DataAbort
+ b arm_dabt_handler
diff --git a/arch/arm/mm/abort-ev5tj.S b/arch/arm/mm/abort-ev5tj.S
index fec72f4fbaf508597d826e58d0dc084ee6e58dd0..a786a7d69e6ae23aa1125ccd41deaaaa743601e1 100644
--- a/arch/arm/mm/abort-ev5tj.S
+++ b/arch/arm/mm/abort-ev5tj.S
@@ -22,12 +22,12 @@ ENTRY(v5tj_early_abort)
mrc p15, 0, r0, c6, c0, 0 @ get FAR
bic r1, r1, #1 << 11 | 1 << 10 @ clear bits 11 and 10 of FSR
tst r5, #PSR_J_BIT @ Java?
- bne do_DataAbort
+ bne arm_dabt_handler
do_thumb_abort fsr=r1, pc=r4, psr=r5, tmp=r3
ldreq r3, [r4] @ read aborted ARM instruction
uaccess_disable ip @ disable userspace access
teq_ldrd tmp=ip, insn=r3 @ insn was LDRD?
- beq do_DataAbort @ yes
+ beq arm_dabt_handler @ yes
tst r3, #1 << 20 @ L = 0 -> write
orreq r1, r1, #1 << 11 @ yes.
- b do_DataAbort
+ b arm_dabt_handler
diff --git a/arch/arm/mm/abort-ev6.S b/arch/arm/mm/abort-ev6.S
index 836dc1299243415faaac1a93dce35ac46ae1a7da..2366236053eb86484cb179ff572f06db6559426b 100644
--- a/arch/arm/mm/abort-ev6.S
+++ b/arch/arm/mm/abort-ev6.S
@@ -42,4 +42,4 @@ ENTRY(v6_early_abort)
orreq r1, r1, #1 << 11 @ yes.
#endif
1: uaccess_disable ip @ disable userspace access
- b do_DataAbort
+ b arm_dabt_handler
diff --git a/arch/arm/mm/abort-ev7.S b/arch/arm/mm/abort-ev7.S
index 53fb41c24774db985eddc370904aa7666527f538..72b6eba101555651d7afd749fd5d7715a2964372 100644
--- a/arch/arm/mm/abort-ev7.S
+++ b/arch/arm/mm/abort-ev7.S
@@ -18,5 +18,5 @@ ENTRY(v7_early_abort)
mrc p15, 0, r1, c5, c0, 0 @ get FSR
mrc p15, 0, r0, c6, c0, 0 @ get FAR
uaccess_disable ip @ disable userspace access
- b do_DataAbort
+ b arm_dabt_handler
ENDPROC(v7_early_abort)
diff --git a/arch/arm/mm/abort-lv4t.S b/arch/arm/mm/abort-lv4t.S
index fbd60a120f6684c56c63cea10b00200765473f1d..3fef9c1479c17bb07652e43e0b8b0683e00c4b2f 100644
--- a/arch/arm/mm/abort-lv4t.S
+++ b/arch/arm/mm/abort-lv4t.S
@@ -46,8 +46,8 @@ ENTRY(v4t_late_abort)
/* 9 */ b .data_arm_ldmstm @ ldm*b rn, <rlist>
/* a */ b .data_unknown
/* b */ b .data_unknown
-/* c */ b do_DataAbort @ ldc rd, [rn], #m @ Same as ldr rd, [rn], #m
-/* d */ b do_DataAbort @ ldc rd, [rn, #m]
+/* c */ b arm_dabt_handler @ ldc rd, [rn], #m @ Same as ldr rd, [rn], #m
+/* d */ b arm_dabt_handler @ ldc rd, [rn, #m]
/* e */ b .data_unknown
/* f */ b .data_unknown
@@ -60,7 +60,7 @@ ENTRY(v4t_late_abort)
.data_arm_ldmstm:
tst r8, #1 << 21 @ check writeback bit
- beq do_DataAbort @ no writeback -> no fixup
+ beq arm_dabt_handler @ no writeback -> no fixup
str r9, [sp, #-4]!
mov r7, #0x11
orr r7, r7, #0x1100
@@ -81,11 +81,11 @@ ENTRY(v4t_late_abort)
addeq r7, r7, r6, lsl #2 @ Undo decrement
str r7, [r2, r9, lsr #14] @ Put register 'Rn'
ldr r9, [sp], #4
- b do_DataAbort
+ b arm_dabt_handler
.data_arm_lateldrhpre:
tst r8, #1 << 21 @ Check writeback bit
- beq do_DataAbort @ No writeback -> no fixup
+ beq arm_dabt_handler @ No writeback -> no fixup
.data_arm_lateldrhpost:
str r9, [sp, #-4]!
and r9, r8, #0x00f @ get Rm / low nibble of immediate value
@@ -101,14 +101,14 @@ ENTRY(v4t_late_abort)
addeq r7, r7, r6 @ Undo decrement
str r7, [r2, r9, lsr #14] @ Put register 'Rn'
ldr r9, [sp], #4
- b do_DataAbort
+ b arm_dabt_handler
.data_arm_lateldrpreconst:
tst r8, #1 << 21 @ check writeback bit
- beq do_DataAbort @ no writeback -> no fixup
+ beq arm_dabt_handler @ no writeback -> no fixup
.data_arm_lateldrpostconst:
movs r6, r8, lsl #20 @ Get offset
- beq do_DataAbort @ zero -> no fixup
+ beq arm_dabt_handler @ zero -> no fixup
str r9, [sp, #-4]!
and r9, r8, #15 << 16 @ Extract 'n' from instruction
ldr r7, [r2, r9, lsr #14] @ Get register 'Rn'
@@ -117,11 +117,11 @@ ENTRY(v4t_late_abort)
addeq r7, r7, r6, lsr #20 @ Undo decrement
str r7, [r2, r9, lsr #14] @ Put register 'Rn'
ldr r9, [sp], #4
- b do_DataAbort
+ b arm_dabt_handler
.data_arm_lateldrprereg:
tst r8, #1 << 21 @ check writeback bit
- beq do_DataAbort @ no writeback -> no fixup
+ beq arm_dabt_handler @ no writeback -> no fixup
.data_arm_lateldrpostreg:
and r7, r8, #15 @ Extract 'm' from instruction
ldr r6, [r2, r7, lsl #2] @ Get register 'Rm'
@@ -180,10 +180,10 @@ ENTRY(v4t_late_abort)
/* 3 */ b .data_unknown
/* 4 */ b .data_unknown
/* 5 */ b .data_thumb_reg
-/* 6 */ b do_DataAbort
-/* 7 */ b do_DataAbort
-/* 8 */ b do_DataAbort
-/* 9 */ b do_DataAbort
+/* 6 */ b arm_dabt_handler
+/* 7 */ b arm_dabt_handler
+/* 8 */ b arm_dabt_handler
+/* 9 */ b arm_dabt_handler
/* A */ b .data_unknown
/* B */ b .data_thumb_pushpop
/* C */ b .data_thumb_ldmstm
@@ -193,10 +193,10 @@ ENTRY(v4t_late_abort)
.data_thumb_reg:
tst r8, #1 << 9
- beq do_DataAbort
+ beq arm_dabt_handler
tst r8, #1 << 10 @ If 'S' (signed) bit is set
movne r1, #0 @ it must be a load instr
- b do_DataAbort
+ b arm_dabt_handler
.data_thumb_pushpop:
tst r8, #1 << 10
@@ -217,7 +217,7 @@ ENTRY(v4t_late_abort)
subne r7, r7, r6, lsl #2 @ decrement SP if POP
str r7, [r2, #13 << 2]
ldr r9, [sp], #4
- b do_DataAbort
+ b arm_dabt_handler
.data_thumb_ldmstm:
str r9, [sp, #-4]!
@@ -234,4 +234,4 @@ ENTRY(v4t_late_abort)
sub r7, r7, r6, lsl #2 @ always decrement
str r7, [r2, r9, lsr #6]
ldr r9, [sp], #4
- b do_DataAbort
+ b arm_dabt_handler
diff --git a/arch/arm/mm/abort-macro.S b/arch/arm/mm/abort-macro.S
index bacf53fd0b70c6307e74ef8601d8dcc7db292700..30bb5e2135862902951b5077b0b860fd12ed8c28 100644
--- a/arch/arm/mm/abort-macro.S
+++ b/arch/arm/mm/abort-macro.S
@@ -20,7 +20,7 @@
orreq \tmp, \tmp, #1 << 11 @ Set L-bit if yes
tst \tmp, #1 << 11 @ L = 0 -> write
orreq \fsr, \fsr, #1 << 11 @ yes.
- b do_DataAbort
+ b arm_dabt_handler
not_thumb:
.endm
diff --git a/arch/arm/mm/abort-nommu.S b/arch/arm/mm/abort-nommu.S
index 6e2366a263219b379f4fdb43cc8e5413fb36e52a..573a09529ac7794f8004680ab350750de7c8e15d 100644
--- a/arch/arm/mm/abort-nommu.S
+++ b/arch/arm/mm/abort-nommu.S
@@ -17,5 +17,5 @@
ENTRY(nommu_early_abort)
mov r0, #0 @ clear r0, r1 (no FSR/FAR)
mov r1, #0
- b do_DataAbort
+ b arm_dabt_handler
ENDPROC(nommu_early_abort)
diff --git a/arch/arm/mm/fault.c b/arch/arm/mm/fault.c
index ab01b51de5590bff6c2acaf446d01c589a5f7987..78afb701b34a17d8d841762944eb5f8af6843dc4 100644
--- a/arch/arm/mm/fault.c
+++ b/arch/arm/mm/fault.c
@@ -588,7 +588,7 @@ hook_fault_code(int nr, int (*fn)(unsigned long, unsigned int, struct pt_regs *)
/*
* Dispatch a data abort to the relevant handler.
*/
-asmlinkage void
+void
do_DataAbort(unsigned long addr, unsigned int fsr, struct pt_regs *regs)
{
const struct fsr_info *inf = fsr_info + fsr_fs(fsr);
@@ -618,7 +618,7 @@ hook_ifault_code(int nr, int (*fn)(unsigned long, unsigned int, struct pt_regs *
ifsr_info[nr].name = name;
}
-asmlinkage void
+void
do_PrefetchAbort(unsigned long addr, unsigned int ifsr, struct pt_regs *regs)
{
const struct fsr_info *inf = ifsr_info + fsr_fs(ifsr);
diff --git a/arch/arm/mm/fault.h b/arch/arm/mm/fault.h
index e8f8c190254442a015a6cbc179602589cfddb0ea..36a7e7138c310a8311458207d2b694cb44c63b93 100644
--- a/arch/arm/mm/fault.h
+++ b/arch/arm/mm/fault.h
@@ -37,9 +37,9 @@ static inline int fsr_fs(unsigned int fsr)
void do_bad_area(unsigned long addr, unsigned int fsr, struct pt_regs *regs);
void early_abt_enable(void);
-asmlinkage void do_DataAbort(unsigned long addr, unsigned int fsr,
- struct pt_regs *regs);
-asmlinkage void do_PrefetchAbort(unsigned long addr, unsigned int ifsr,
- struct pt_regs *regs);
+void do_DataAbort(unsigned long addr, unsigned int fsr,
+ struct pt_regs *regs);
+void do_PrefetchAbort(unsigned long addr, unsigned int ifsr,
+ struct pt_regs *regs);
#endif /* __ARCH_ARM_FAULT_H */
diff --git a/arch/arm/mm/pabort-legacy.S b/arch/arm/mm/pabort-legacy.S
index b2ffce4201062e3ec2045364ddc454cf706bab8d..26c62e568638d37133661f4f5b8dec430593f8eb 100644
--- a/arch/arm/mm/pabort-legacy.S
+++ b/arch/arm/mm/pabort-legacy.S
@@ -18,5 +18,5 @@
ENTRY(legacy_pabort)
mov r0, r4
mov r1, #5
- b do_PrefetchAbort
+ b arm_pabt_handler
ENDPROC(legacy_pabort)
diff --git a/arch/arm/mm/pabort-v6.S b/arch/arm/mm/pabort-v6.S
index 8686265dc9418b29381942bfd87a937a3234d46e..25abd11a35253bf464fe9fd7fc14fd11bc564dcd 100644
--- a/arch/arm/mm/pabort-v6.S
+++ b/arch/arm/mm/pabort-v6.S
@@ -18,5 +18,5 @@
ENTRY(v6_pabort)
mov r0, r4
mrc p15, 0, r1, c5, c0, 1 @ get IFSR
- b do_PrefetchAbort
+ b arm_pabt_handler
ENDPROC(v6_pabort)
diff --git a/arch/arm/mm/pabort-v7.S b/arch/arm/mm/pabort-v7.S
index 9c70b1a21dc9204f24524df9905fbc077a82f2dc..e05c7d44d307adeba6759213374b2aa328c693da 100644
--- a/arch/arm/mm/pabort-v7.S
+++ b/arch/arm/mm/pabort-v7.S
@@ -18,5 +18,5 @@
ENTRY(v7_pabort)
mrc p15, 0, r0, c6, c0, 2 @ get IFAR
mrc p15, 0, r1, c5, c0, 1 @ get IFSR
- b do_PrefetchAbort
+ b arm_pabt_handler
ENDPROC(v7_pabort)
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 30/31] ARM: entry: Block IRQs in early IRQ context
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (28 preceding siblings ...)
2025-02-12 11:23 ` [PATCH v4 29/31] ARM: entry: Handle dabt, pabt, and und as interrupts Linus Walleij
@ 2025-02-12 11:23 ` Linus Walleij
2025-02-12 11:23 ` [PATCH v4 31/31] ARM: entry: Straighten syscall returns Linus Walleij
2025-02-14 1:03 ` [PATCH v4 00/31] ARM: Switch to generic entry Dmitry V. Levin
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:23 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
When dabt, pabt or und exceptions occur on ARM, ordinary
interrupts (IRQs) can still happen. This isn't nice for the
kernels context tracker, which expect (when using generic
entry at least) that any nested IRQs happens
between irqentry_enter() and irqentry_exit(), else it
thinks something is fishy.
This change blocks interrupts in the pabt, dabt, und and
abt exception paths (all of them really) by unconditionally
setting PSR_I_BIT in the early exception handler, until
after context has been established with irqentry_enter()
and before it is exited with irqentry_exit(). Inside the
context-tracked exception handler we enable IRQs again,
and once we leave it we disable them while exiting the
exception.
The local_irq_disable() in bad_mode() can be dropped
since we are now disabling IRQs in the early assembly
exception handler for all exceptions.
This seems like not perfect: it seems an interrupt
could still occur right before CPSR is set, or right after
the userspace registers are restored in ret_from_exception.
I would like to know if there is some
way to set up these exceptions to inherently block IRQs
when handled, until we explicitly allow them between
irqentry_enter() and irqentry_exit() or if this is simply
the best we can do on ARM for these exceptions to make the
context tracker happy.
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/kernel/entry-armv.S | 6 +-----
arch/arm/kernel/entry.c | 18 ++++++++++++++++++
arch/arm/kernel/traps.c | 1 -
3 files changed, 19 insertions(+), 6 deletions(-)
diff --git a/arch/arm/kernel/entry-armv.S b/arch/arm/kernel/entry-armv.S
index ae2f952beea7611f0abc7bd299fc944335a21219..3dae35b0bb3f440ecaf157a45687bf4690fb8f88 100644
--- a/arch/arm/kernel/entry-armv.S
+++ b/arch/arm/kernel/entry-armv.S
@@ -416,11 +416,6 @@ ENDPROC(__irq_usr)
__und_usr:
usr_entry uaccess=0
- @ IRQs must be enabled before attempting to read the instruction from
- @ user space since that could cause a page/translation fault if the
- @ page table was modified by another CPU.
- enable_irq
-
tst r5, #PSR_T_BIT @ Thumb mode?
mov r1, #2 @ set insn size to 2 for Thumb
bne 0f @ handle as Thumb undef exception
@@ -847,6 +842,7 @@ vector_\name:
@
mrs r0, cpsr
eor r0, r0, #(\mode ^ SVC_MODE | PSR_ISETSTATE)
+ orr r0, r0, #PSR_I_BIT
msr spsr_cxsf, r0
@
diff --git a/arch/arm/kernel/entry.c b/arch/arm/kernel/entry.c
index 01e4339ccdb4951e04a89fb91ad7c6e1991e09af..3881670e0987ee40be6fff32f412edcf4f3ec80a 100644
--- a/arch/arm/kernel/entry.c
+++ b/arch/arm/kernel/entry.c
@@ -17,8 +17,18 @@ noinstr asmlinkage void arm_und_handler(struct pt_regs *regs)
{
irqentry_state_t state = irqentry_enter(regs);
+ /*
+ * IRQs must be enabled before attempting to read the instruction from
+ * user space since that could cause a page/translation fault if the
+ * page table was modified by another CPU.
+ */
+
+ local_irq_enable();
+
do_undefinstr(regs);
+ local_irq_disable();
+
irqentry_exit(regs, state);
}
@@ -27,8 +37,12 @@ noinstr asmlinkage void arm_dabt_handler(unsigned long addr, unsigned int fsr,
{
irqentry_state_t state = irqentry_enter(regs);
+ local_irq_enable();
+
do_DataAbort(addr, fsr, regs);
+ local_irq_disable();
+
irqentry_exit(regs, state);
}
@@ -37,8 +51,12 @@ noinstr asmlinkage void arm_pabt_handler(unsigned long addr, unsigned int ifsr,
{
irqentry_state_t state = irqentry_enter(regs);
+ local_irq_enable();
+
do_PrefetchAbort(addr, ifsr, regs);
+ local_irq_disable();
+
irqentry_exit(regs, state);
}
diff --git a/arch/arm/kernel/traps.c b/arch/arm/kernel/traps.c
index d3a689bd05955f1ae46a6341e456bb097e831311..b2eaab9cc4977c448f77faa4e6bc22a0749d1f9c 100644
--- a/arch/arm/kernel/traps.c
+++ b/arch/arm/kernel/traps.c
@@ -515,7 +515,6 @@ asmlinkage void bad_mode(struct pt_regs *regs, int reason)
pr_crit("Bad mode in %s handler detected\n", handler[reason]);
die("Oops - bad mode", regs, 0);
- local_irq_disable();
panic("bad mode");
}
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* [PATCH v4 31/31] ARM: entry: Straighten syscall returns
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (29 preceding siblings ...)
2025-02-12 11:23 ` [PATCH v4 30/31] ARM: entry: Block IRQs in early IRQ context Linus Walleij
@ 2025-02-12 11:23 ` Linus Walleij
2025-02-14 1:03 ` [PATCH v4 00/31] ARM: Switch to generic entry Dmitry V. Levin
31 siblings, 0 replies; 37+ messages in thread
From: Linus Walleij @ 2025-02-12 11:23 UTC (permalink / raw)
To: Dmitry Vyukov, Oleg Nesterov, Russell King, Kees Cook,
Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro
Cc: linux-arm-kernel, linux-kernel, Linus Walleij
Since ret_fast_syscall() is now only called in two spots in
vector_swi, we can inline the function into vector_swi, and
skip the separate call, placing the code from ret_fast_syscall
at the end of vector_swi and one of them inside an ifdef that
gets compiled out for !EABI.
TODO: is the UNWIND(.fnstart/.cantunwind/.fnend) really needed
on vector_swi after this change or can I just drop it?
I do not understand unwind well enough :/
Signed-off-by: Linus Walleij <linus.walleij@linaro.org>
---
arch/arm/kernel/entry-common.S | 32 ++++++++++++++------------------
1 file changed, 14 insertions(+), 18 deletions(-)
diff --git a/arch/arm/kernel/entry-common.S b/arch/arm/kernel/entry-common.S
index a5a0e757ab23d151485092a078af98966e039869..f9c38f5311851e299234997796cd78de06a5c531 100644
--- a/arch/arm/kernel/entry-common.S
+++ b/arch/arm/kernel/entry-common.S
@@ -30,21 +30,6 @@ saved_pc .req lr
.section .entry.text,"ax",%progbits
.align 5
-/*
- * As we will need to call out to some C functions,
- * we save r0 first to avoid needing to save registers around each C function
- * call.
- */
-ret_fast_syscall:
- UNWIND(.fnstart )
- UNWIND(.cantunwind )
- add sp, sp, #(S_R0 + S_OFF)
-#ifdef CONFIG_GCC_PLUGIN_STACKLEAK
- bl stackleak_erase_on_task_stack
-#endif
- restore_user_regs
- UNWIND(.fnend )
-ENDPROC(ret_fast_syscall)
/*
* "slow" syscall return path. "why" tells us if this was a real syscall.
@@ -122,6 +107,8 @@ ENDPROC(vector_bhb_bpiall_swi)
#endif
.align 5
ENTRY(vector_swi)
+ UNWIND(.fnstart )
+ UNWIND(.cantunwind )
#ifdef CONFIG_CPU_V7M
v7m_exception_entry
#else
@@ -210,9 +197,9 @@ ENTRY(vector_swi)
mov r0, tbl
mov r2, scno @ syscall number from r7
bl invoke_syscall
- b ret_fast_syscall
-
#if defined(CONFIG_OABI_COMPAT) || !defined(CONFIG_AEABI)
+ b 9002
+
/*
* We failed to handle a fault trying to access the page
* containing the swi instruction, but we're not really in a
@@ -225,8 +212,17 @@ ENTRY(vector_swi)
sub lr, saved_pc, #4
str lr, [sp, #S_PC]
get_thread_info tsk
- b ret_fast_syscall
+9002:
+ /* Exit !AEABI clause */
+#endif
+
+ /* Return from syscall */
+ add sp, sp, #(S_R0 + S_OFF)
+#ifdef CONFIG_GCC_PLUGIN_STACKLEAK
+ bl stackleak_erase_on_task_stack
#endif
+ restore_user_regs
+ UNWIND(.fnend )
ENDPROC(vector_swi)
.ltorg
--
2.48.1
^ permalink raw reply related [flat|nested] 37+ messages in thread* Re: [PATCH v4 00/31] ARM: Switch to generic entry
2025-02-12 11:22 [PATCH v4 00/31] ARM: Switch to generic entry Linus Walleij
` (30 preceding siblings ...)
2025-02-12 11:23 ` [PATCH v4 31/31] ARM: entry: Straighten syscall returns Linus Walleij
@ 2025-02-14 1:03 ` Dmitry V. Levin
2025-02-14 10:47 ` Linus Walleij
31 siblings, 1 reply; 37+ messages in thread
From: Dmitry V. Levin @ 2025-02-14 1:03 UTC (permalink / raw)
To: Linus Walleij
Cc: Dmitry Vyukov, Oleg Nesterov, Eugene Syromyatnikov, Russell King,
Kees Cook, Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro, linux-arm-kernel, linux-kernel
On Wed, Feb 12, 2025 at 12:22:54PM +0100, Linus Walleij wrote:
[...]
> - Tested some ptrace/strace obviously, such as issuing
> several instances of "ptrace find /" and let this scroll
> by in the terminal over some 10 minutes or so.
Could you also run the strace test suite, please? Given that it used to
catch quite a few regressions in the past, it could be useful in this case
as well.
--
ldv
^ permalink raw reply [flat|nested] 37+ messages in thread* Re: [PATCH v4 00/31] ARM: Switch to generic entry
2025-02-14 1:03 ` [PATCH v4 00/31] ARM: Switch to generic entry Dmitry V. Levin
@ 2025-02-14 10:47 ` Linus Walleij
2025-02-14 11:05 ` Dmitry V. Levin
0 siblings, 1 reply; 37+ messages in thread
From: Linus Walleij @ 2025-02-14 10:47 UTC (permalink / raw)
To: Dmitry V. Levin
Cc: Dmitry Vyukov, Oleg Nesterov, Eugene Syromyatnikov, Russell King,
Kees Cook, Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro, linux-arm-kernel, linux-kernel
On Fri, Feb 14, 2025 at 2:03 AM Dmitry V. Levin <ldv@strace.io> wrote:
> On Wed, Feb 12, 2025 at 12:22:54PM +0100, Linus Walleij wrote:
> [...]
> > - Tested some ptrace/strace obviously, such as issuing
> > several instances of "ptrace find /" and let this scroll
> > by in the terminal over some 10 minutes or so.
>
> Could you also run the strace test suite, please? Given that it used to
> catch quite a few regressions in the past, it could be useful in this case
> as well.
Sure, where can I find this test suite?
Yours,
Linus Walleij
^ permalink raw reply [flat|nested] 37+ messages in thread
* Re: [PATCH v4 00/31] ARM: Switch to generic entry
2025-02-14 10:47 ` Linus Walleij
@ 2025-02-14 11:05 ` Dmitry V. Levin
2025-02-20 14:04 ` Linus Walleij
0 siblings, 1 reply; 37+ messages in thread
From: Dmitry V. Levin @ 2025-02-14 11:05 UTC (permalink / raw)
To: Linus Walleij
Cc: Dmitry Vyukov, Oleg Nesterov, Eugene Syromyatnikov, Russell King,
Kees Cook, Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro, linux-arm-kernel, linux-kernel
On Fri, Feb 14, 2025 at 11:47:30AM +0100, Linus Walleij wrote:
> On Fri, Feb 14, 2025 at 2:03 AM Dmitry V. Levin <ldv@strace.io> wrote:
> > On Wed, Feb 12, 2025 at 12:22:54PM +0100, Linus Walleij wrote:
> > [...]
> > > - Tested some ptrace/strace obviously, such as issuing
> > > several instances of "ptrace find /" and let this scroll
> > > by in the terminal over some 10 minutes or so.
> >
> > Could you also run the strace test suite, please? Given that it used to
> > catch quite a few regressions in the past, it could be useful in this case
> > as well.
>
> Sure, where can I find this test suite?
It's a part of strace, you can find it e.g. at
https://github.com/strace/strace
To build and run it one can roughly do
./bootstrap && ./configure && make -j`nproc` && make -j`nproc check
--
ldv
^ permalink raw reply [flat|nested] 37+ messages in thread
* Re: [PATCH v4 00/31] ARM: Switch to generic entry
2025-02-14 11:05 ` Dmitry V. Levin
@ 2025-02-20 14:04 ` Linus Walleij
2025-02-20 14:30 ` Dmitry V. Levin
0 siblings, 1 reply; 37+ messages in thread
From: Linus Walleij @ 2025-02-20 14:04 UTC (permalink / raw)
To: Dmitry V. Levin
Cc: Dmitry Vyukov, Oleg Nesterov, Eugene Syromyatnikov, Russell King,
Kees Cook, Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro, linux-arm-kernel, linux-kernel
On Fri, Feb 14, 2025 at 12:05 PM Dmitry V. Levin <ldv@strace.io> wrote:
> > Sure, where can I find this test suite?
>
> It's a part of strace, you can find it e.g. at
> https://github.com/strace/strace
>
> To build and run it one can roughly do
> ./bootstrap && ./configure && make -j`nproc` && make -j`nproc check
make check produces some test failures on v6.14-rc1 on ARM
even before I apply the generic entry:
FAIL: filtering_syscall-syntax.test
FAIL: qual_fault-syscall.test
FAIL: qual_fault.test
FAIL: strace--tips-full.test
FAIL: strace-r.test
============================================================================
Testsuite summary for strace 6.13.0.27.bbda4
============================================================================
# TOTAL: 1409
# PASS: 1106
# SKIP: 298
# XFAIL: 0
# FAIL: 5
# XPASS: 0
# ERROR: 0
============================================================================
But I create more fails after my patch set ...
Some have to do with fast syscall restart (I need to look into this).
Then there is the fact that I had to add the PTRACE_SYSEMU and
PTRACE_SYSEMU_SINGLESTEP as stubs returning -EIO
(modeled after UM) to use generic entry. Do you think this affects
the results?
Is there a way to run a single test? I tried to check the docs but
nothing obvious to me ... I guess I'm not used to this unit-tester.
Yours,
Linus Walleij
^ permalink raw reply [flat|nested] 37+ messages in thread
* Re: [PATCH v4 00/31] ARM: Switch to generic entry
2025-02-20 14:04 ` Linus Walleij
@ 2025-02-20 14:30 ` Dmitry V. Levin
0 siblings, 0 replies; 37+ messages in thread
From: Dmitry V. Levin @ 2025-02-20 14:30 UTC (permalink / raw)
To: Linus Walleij
Cc: Dmitry Vyukov, Oleg Nesterov, Eugene Syromyatnikov, Russell King,
Kees Cook, Andy Lutomirski, Will Drewry, Frederic Weisbecker,
Paul E. McKenney, Jinjie Ruan, Arnd Bergmann, Ard Biesheuvel,
Al Viro, linux-arm-kernel, linux-kernel
On Thu, Feb 20, 2025 at 03:04:22PM +0100, Linus Walleij wrote:
> On Fri, Feb 14, 2025 at 12:05 PM Dmitry V. Levin <ldv@strace.io> wrote:
>
> > > Sure, where can I find this test suite?
> >
> > It's a part of strace, you can find it e.g. at
> > https://github.com/strace/strace
> >
> > To build and run it one can roughly do
> > ./bootstrap && ./configure && make -j`nproc` && make -j`nproc check
>
> make check produces some test failures on v6.14-rc1 on ARM
> even before I apply the generic entry:
>
> FAIL: filtering_syscall-syntax.test
> FAIL: qual_fault-syscall.test
> FAIL: qual_fault.test
> FAIL: strace--tips-full.test
> FAIL: strace-r.test
This is surprising, no tests are currently known to fail on arm.
> Then there is the fact that I had to add the PTRACE_SYSEMU and
> PTRACE_SYSEMU_SINGLESTEP as stubs returning -EIO
> (modeled after UM) to use generic entry. Do you think this affects
> the results?
No, strace doesn't use PTRACE_SYSEMU* and doesn't test it either.
> Is there a way to run a single test? I tried to check the docs but
> nothing obvious to me ... I guess I'm not used to this unit-tester.
Sure, it's a regular automake-based test suite, so you can do something like
$ make -s -k check VERBOSE=1 TESTS='filtering_syscall-syntax.test qual_fault-syscall.test qual_fault.test'
--
ldv
^ permalink raw reply [flat|nested] 37+ messages in thread