Commit f37099b6 authored by Mark Rutland's avatar Mark Rutland Committed by Will Deacon

arm64: convert syscall trace logic to C

Currently syscall tracing is a tricky assembly state machine, which can
be rather difficult to follow, and even harder to modify. Before we
start fiddling with it for pt_regs syscalls, let's convert it to C.

This is not intended to have any functional change.
Signed-off-by: default avatarMark Rutland <mark.rutland@arm.com>
Reviewed-by: default avatarCatalin Marinas <catalin.marinas@arm.com>
Cc: Will Deacon <will.deacon@arm.com>
Signed-off-by: default avatarWill Deacon <will.deacon@arm.com>
parent 4141c857
...@@ -896,26 +896,6 @@ el0_error_naked: ...@@ -896,26 +896,6 @@ el0_error_naked:
b ret_to_user b ret_to_user
ENDPROC(el0_error) ENDPROC(el0_error)
/*
* This is the fast syscall return path. We do as little as possible here,
* and this includes saving x0 back into the kernel stack.
*/
ret_fast_syscall:
disable_daif
#ifndef CONFIG_DEBUG_RSEQ
ldr x1, [tsk, #TSK_TI_FLAGS] // re-check for syscall tracing
and x2, x1, #_TIF_SYSCALL_WORK
cbnz x2, ret_fast_syscall_trace
and x2, x1, #_TIF_WORK_MASK
cbnz x2, work_pending
enable_step_tsk x1, x2
kernel_exit 0
#endif
ret_fast_syscall_trace:
enable_daif
b __sys_trace_return_skipped // we already saved x0
/* /*
* Ok, we need to do extra processing, enter the slow path. * Ok, we need to do extra processing, enter the slow path.
*/ */
...@@ -971,44 +951,13 @@ alternative_else_nop_endif ...@@ -971,44 +951,13 @@ alternative_else_nop_endif
#endif #endif
el0_svc_naked: // compat entry point el0_svc_naked: // compat entry point
stp x0, xscno, [sp, #S_ORIG_X0] // save the original x0 and syscall number
enable_daif
ct_user_exit 1
tst x16, #_TIF_SYSCALL_WORK // check for syscall hooks
b.ne __sys_trace
mov x0, sp mov x0, sp
mov w1, wscno mov w1, wscno
mov w2, wsc_nr mov w2, wsc_nr
mov x3, stbl mov x3, stbl
bl invoke_syscall bl el0_svc_common
b ret_fast_syscall
ENDPROC(el0_svc)
/*
* This is the really slow path. We're going to be doing context
* switches, and waiting for our parent to respond.
*/
__sys_trace:
cmp wscno, #NO_SYSCALL // user-issued syscall(-1)?
b.ne 1f
mov x0, #-ENOSYS // set default errno if so
str x0, [sp, #S_X0]
1: mov x0, sp
bl syscall_trace_enter
cmp w0, #NO_SYSCALL // skip the syscall?
b.eq __sys_trace_return_skipped
mov x0, sp
mov w1, wscno
mov w2, wsc_nr
mov x3, stbl
bl invoke_syscall
__sys_trace_return_skipped:
mov x0, sp
bl syscall_trace_exit
b ret_to_user b ret_to_user
ENDPROC(el0_svc)
.popsection // .entry.text .popsection // .entry.text
......
// SPDX-License-Identifier: GPL-2.0 // SPDX-License-Identifier: GPL-2.0
#include <linux/compiler.h>
#include <linux/context_tracking.h>
#include <linux/errno.h> #include <linux/errno.h>
#include <linux/nospec.h> #include <linux/nospec.h>
#include <linux/ptrace.h> #include <linux/ptrace.h>
#include <linux/syscalls.h> #include <linux/syscalls.h>
#include <asm/daifflags.h>
#include <asm/syscall.h> #include <asm/syscall.h>
#include <asm/thread_info.h>
long compat_arm_syscall(struct pt_regs *regs); long compat_arm_syscall(struct pt_regs *regs);
...@@ -29,9 +33,9 @@ static long __invoke_syscall(struct pt_regs *regs, syscall_fn_t syscall_fn) ...@@ -29,9 +33,9 @@ static long __invoke_syscall(struct pt_regs *regs, syscall_fn_t syscall_fn)
regs->regs[3], regs->regs[4], regs->regs[5]); regs->regs[3], regs->regs[4], regs->regs[5]);
} }
asmlinkage void invoke_syscall(struct pt_regs *regs, unsigned int scno, static void invoke_syscall(struct pt_regs *regs, unsigned int scno,
unsigned int sc_nr, unsigned int sc_nr,
const syscall_fn_t syscall_table[]) const syscall_fn_t syscall_table[])
{ {
long ret; long ret;
...@@ -45,3 +49,50 @@ asmlinkage void invoke_syscall(struct pt_regs *regs, unsigned int scno, ...@@ -45,3 +49,50 @@ asmlinkage void invoke_syscall(struct pt_regs *regs, unsigned int scno,
regs->regs[0] = ret; regs->regs[0] = ret;
} }
static inline bool has_syscall_work(unsigned long flags)
{
return unlikely(flags & _TIF_SYSCALL_WORK);
}
int syscall_trace_enter(struct pt_regs *regs);
void syscall_trace_exit(struct pt_regs *regs);
asmlinkage void el0_svc_common(struct pt_regs *regs, int scno, int sc_nr,
const syscall_fn_t syscall_table[])
{
unsigned long flags = current_thread_info()->flags;
regs->orig_x0 = regs->regs[0];
regs->syscallno = scno;
local_daif_restore(DAIF_PROCCTX);
user_exit();
if (has_syscall_work(flags)) {
/* set default errno for user-issued syscall(-1) */
if (scno == NO_SYSCALL)
regs->regs[0] = -ENOSYS;
scno = syscall_trace_enter(regs);
if (scno == NO_SYSCALL)
goto trace_exit;
}
invoke_syscall(regs, scno, sc_nr, syscall_table);
/*
* The tracing status may have changed under our feet, so we have to
* check again. However, if we were tracing entry, then we always trace
* exit regardless, as the old entry assembly did.
*/
if (!has_syscall_work(flags) && !IS_ENABLED(CONFIG_DEBUG_RSEQ)) {
local_daif_mask();
flags = current_thread_info()->flags;
if (!has_syscall_work(flags))
return;
local_daif_restore(DAIF_PROCCTX);
}
trace_exit:
syscall_trace_exit(regs);
}
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment