Commit 29701d69 authored by Jens Axboe's avatar Jens Axboe

Merge tag 'core-entry-notify-signal' of...

Merge tag 'core-entry-notify-signal' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip into tif-task_work.arch

Core changes to support TASK_NOTIFY_SIGNAL

* tag 'core-entry-notify-signal' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip:
  task_work: Use TIF_NOTIFY_SIGNAL if available
  entry: Add support for TIF_NOTIFY_SIGNAL
  signal: Add task_sigpending() helper
parents f8394f23 114518eb
...@@ -804,11 +804,11 @@ static inline unsigned long get_nr_restart_syscall(const struct pt_regs *regs) ...@@ -804,11 +804,11 @@ static inline unsigned long get_nr_restart_syscall(const struct pt_regs *regs)
* want to handle. Thus you cannot kill init even with a SIGKILL even by * want to handle. Thus you cannot kill init even with a SIGKILL even by
* mistake. * mistake.
*/ */
void arch_do_signal(struct pt_regs *regs) void arch_do_signal_or_restart(struct pt_regs *regs, bool has_signal)
{ {
struct ksignal ksig; struct ksignal ksig;
if (get_signal(&ksig)) { if (has_signal && get_signal(&ksig)) {
/* Whee! Actually deliver the signal. */ /* Whee! Actually deliver the signal. */
handle_signal(&ksig, regs); handle_signal(&ksig, regs);
return; return;
......
...@@ -37,6 +37,10 @@ ...@@ -37,6 +37,10 @@
# define _TIF_UPROBE (0) # define _TIF_UPROBE (0)
#endif #endif
#ifndef _TIF_NOTIFY_SIGNAL
# define _TIF_NOTIFY_SIGNAL (0)
#endif
/* /*
* TIF flags handled in syscall_enter_from_user_mode() * TIF flags handled in syscall_enter_from_user_mode()
*/ */
...@@ -69,7 +73,7 @@ ...@@ -69,7 +73,7 @@
#define EXIT_TO_USER_MODE_WORK \ #define EXIT_TO_USER_MODE_WORK \
(_TIF_SIGPENDING | _TIF_NOTIFY_RESUME | _TIF_UPROBE | \ (_TIF_SIGPENDING | _TIF_NOTIFY_RESUME | _TIF_UPROBE | \
_TIF_NEED_RESCHED | _TIF_PATCH_PENDING | \ _TIF_NEED_RESCHED | _TIF_PATCH_PENDING | _TIF_NOTIFY_SIGNAL | \
ARCH_EXIT_TO_USER_MODE_WORK) ARCH_EXIT_TO_USER_MODE_WORK)
/** /**
...@@ -259,12 +263,13 @@ static __always_inline void arch_exit_to_user_mode(void) { } ...@@ -259,12 +263,13 @@ static __always_inline void arch_exit_to_user_mode(void) { }
#endif #endif
/** /**
* arch_do_signal - Architecture specific signal delivery function * arch_do_signal_or_restart - Architecture specific signal delivery function
* @regs: Pointer to currents pt_regs * @regs: Pointer to currents pt_regs
* @has_signal: actual signal to handle
* *
* Invoked from exit_to_user_mode_loop(). * Invoked from exit_to_user_mode_loop().
*/ */
void arch_do_signal(struct pt_regs *regs); void arch_do_signal_or_restart(struct pt_regs *regs, bool has_signal);
/** /**
* arch_syscall_exit_tracehook - Wrapper around tracehook_report_syscall_exit() * arch_syscall_exit_tracehook - Wrapper around tracehook_report_syscall_exit()
......
...@@ -11,8 +11,8 @@ ...@@ -11,8 +11,8 @@
# define ARCH_XFER_TO_GUEST_MODE_WORK (0) # define ARCH_XFER_TO_GUEST_MODE_WORK (0)
#endif #endif
#define XFER_TO_GUEST_MODE_WORK \ #define XFER_TO_GUEST_MODE_WORK \
(_TIF_NEED_RESCHED | _TIF_SIGPENDING | \ (_TIF_NEED_RESCHED | _TIF_SIGPENDING | _TIF_NOTIFY_SIGNAL | \
_TIF_NOTIFY_RESUME | ARCH_XFER_TO_GUEST_MODE_WORK) _TIF_NOTIFY_RESUME | ARCH_XFER_TO_GUEST_MODE_WORK)
struct kvm_vcpu; struct kvm_vcpu;
......
...@@ -353,11 +353,25 @@ static inline int restart_syscall(void) ...@@ -353,11 +353,25 @@ static inline int restart_syscall(void)
return -ERESTARTNOINTR; return -ERESTARTNOINTR;
} }
static inline int signal_pending(struct task_struct *p) static inline int task_sigpending(struct task_struct *p)
{ {
return unlikely(test_tsk_thread_flag(p,TIF_SIGPENDING)); return unlikely(test_tsk_thread_flag(p,TIF_SIGPENDING));
} }
static inline int signal_pending(struct task_struct *p)
{
#if defined(TIF_NOTIFY_SIGNAL)
/*
* TIF_NOTIFY_SIGNAL isn't really a signal, but it requires the same
* behavior in terms of ensuring that we break out of wait loops
* so that notify signal callbacks can be processed.
*/
if (unlikely(test_tsk_thread_flag(p, TIF_NOTIFY_SIGNAL)))
return 1;
#endif
return task_sigpending(p);
}
static inline int __fatal_signal_pending(struct task_struct *p) static inline int __fatal_signal_pending(struct task_struct *p)
{ {
return unlikely(sigismember(&p->pending.signal, SIGKILL)); return unlikely(sigismember(&p->pending.signal, SIGKILL));
...@@ -365,7 +379,7 @@ static inline int __fatal_signal_pending(struct task_struct *p) ...@@ -365,7 +379,7 @@ static inline int __fatal_signal_pending(struct task_struct *p)
static inline int fatal_signal_pending(struct task_struct *p) static inline int fatal_signal_pending(struct task_struct *p)
{ {
return signal_pending(p) && __fatal_signal_pending(p); return task_sigpending(p) && __fatal_signal_pending(p);
} }
static inline int signal_pending_state(long state, struct task_struct *p) static inline int signal_pending_state(long state, struct task_struct *p)
...@@ -502,7 +516,7 @@ extern int set_user_sigmask(const sigset_t __user *umask, size_t sigsetsize); ...@@ -502,7 +516,7 @@ extern int set_user_sigmask(const sigset_t __user *umask, size_t sigsetsize);
static inline void restore_saved_sigmask_unless(bool interrupted) static inline void restore_saved_sigmask_unless(bool interrupted)
{ {
if (interrupted) if (interrupted)
WARN_ON(!test_thread_flag(TIF_SIGPENDING)); WARN_ON(!signal_pending(current));
else else
restore_saved_sigmask(); restore_saved_sigmask();
} }
......
...@@ -198,4 +198,31 @@ static inline void tracehook_notify_resume(struct pt_regs *regs) ...@@ -198,4 +198,31 @@ static inline void tracehook_notify_resume(struct pt_regs *regs)
blkcg_maybe_throttle_current(); blkcg_maybe_throttle_current();
} }
/*
* called by exit_to_user_mode_loop() if ti_work & _TIF_NOTIFY_SIGNAL. This
* is currently used by TWA_SIGNAL based task_work, which requires breaking
* wait loops to ensure that task_work is noticed and run.
*/
static inline void tracehook_notify_signal(void)
{
#if defined(TIF_NOTIFY_SIGNAL)
clear_thread_flag(TIF_NOTIFY_SIGNAL);
smp_mb__after_atomic();
if (current->task_works)
task_work_run();
#endif
}
/*
* Called when we have work to process from exit_to_user_mode_loop()
*/
static inline void set_notify_signal(struct task_struct *task)
{
#if defined(TIF_NOTIFY_SIGNAL)
if (!test_and_set_tsk_thread_flag(task, TIF_NOTIFY_SIGNAL) &&
!wake_up_state(task, TASK_INTERRUPTIBLE))
kick_process(task);
#endif
}
#endif /* <linux/tracehook.h> */ #endif /* <linux/tracehook.h> */
...@@ -135,7 +135,15 @@ static __always_inline void exit_to_user_mode(void) ...@@ -135,7 +135,15 @@ static __always_inline void exit_to_user_mode(void)
} }
/* Workaround to allow gradual conversion of architecture code */ /* Workaround to allow gradual conversion of architecture code */
void __weak arch_do_signal(struct pt_regs *regs) { } void __weak arch_do_signal_or_restart(struct pt_regs *regs, bool has_signal) { }
static void handle_signal_work(struct pt_regs *regs, unsigned long ti_work)
{
if (ti_work & _TIF_NOTIFY_SIGNAL)
tracehook_notify_signal();
arch_do_signal_or_restart(regs, ti_work & _TIF_SIGPENDING);
}
static unsigned long exit_to_user_mode_loop(struct pt_regs *regs, static unsigned long exit_to_user_mode_loop(struct pt_regs *regs,
unsigned long ti_work) unsigned long ti_work)
...@@ -157,8 +165,8 @@ static unsigned long exit_to_user_mode_loop(struct pt_regs *regs, ...@@ -157,8 +165,8 @@ static unsigned long exit_to_user_mode_loop(struct pt_regs *regs,
if (ti_work & _TIF_PATCH_PENDING) if (ti_work & _TIF_PATCH_PENDING)
klp_update_patch_state(current); klp_update_patch_state(current);
if (ti_work & _TIF_SIGPENDING) if (ti_work & (_TIF_SIGPENDING | _TIF_NOTIFY_SIGNAL))
arch_do_signal(regs); handle_signal_work(regs, ti_work);
if (ti_work & _TIF_NOTIFY_RESUME) { if (ti_work & _TIF_NOTIFY_RESUME) {
tracehook_notify_resume(regs); tracehook_notify_resume(regs);
......
...@@ -8,6 +8,9 @@ static int xfer_to_guest_mode_work(struct kvm_vcpu *vcpu, unsigned long ti_work) ...@@ -8,6 +8,9 @@ static int xfer_to_guest_mode_work(struct kvm_vcpu *vcpu, unsigned long ti_work)
do { do {
int ret; int ret;
if (ti_work & _TIF_NOTIFY_SIGNAL)
tracehook_notify_signal();
if (ti_work & _TIF_SIGPENDING) { if (ti_work & _TIF_SIGPENDING) {
kvm_handle_signal_exit(vcpu); kvm_handle_signal_exit(vcpu);
return -EINTR; return -EINTR;
......
...@@ -1973,7 +1973,7 @@ bool uprobe_deny_signal(void) ...@@ -1973,7 +1973,7 @@ bool uprobe_deny_signal(void)
WARN_ON_ONCE(utask->state != UTASK_SSTEP); WARN_ON_ONCE(utask->state != UTASK_SSTEP);
if (signal_pending(t)) { if (task_sigpending(t)) {
spin_lock_irq(&t->sighand->siglock); spin_lock_irq(&t->sighand->siglock);
clear_tsk_thread_flag(t, TIF_SIGPENDING); clear_tsk_thread_flag(t, TIF_SIGPENDING);
spin_unlock_irq(&t->sighand->siglock); spin_unlock_irq(&t->sighand->siglock);
......
...@@ -984,7 +984,7 @@ static inline bool wants_signal(int sig, struct task_struct *p) ...@@ -984,7 +984,7 @@ static inline bool wants_signal(int sig, struct task_struct *p)
if (task_is_stopped_or_traced(p)) if (task_is_stopped_or_traced(p))
return false; return false;
return task_curr(p) || !signal_pending(p); return task_curr(p) || !task_sigpending(p);
} }
static void complete_signal(int sig, struct task_struct *p, enum pid_type type) static void complete_signal(int sig, struct task_struct *p, enum pid_type type)
...@@ -2530,6 +2530,20 @@ bool get_signal(struct ksignal *ksig) ...@@ -2530,6 +2530,20 @@ bool get_signal(struct ksignal *ksig)
struct signal_struct *signal = current->signal; struct signal_struct *signal = current->signal;
int signr; int signr;
/*
* For non-generic architectures, check for TIF_NOTIFY_SIGNAL so
* that the arch handlers don't all have to do it. If we get here
* without TIF_SIGPENDING, just exit after running signal work.
*/
#ifdef TIF_NOTIFY_SIGNAL
if (!IS_ENABLED(CONFIG_GENERIC_ENTRY)) {
if (test_thread_flag(TIF_NOTIFY_SIGNAL))
tracehook_notify_signal();
if (!task_sigpending(current))
return false;
}
#endif
if (unlikely(uprobe_deny_signal())) if (unlikely(uprobe_deny_signal()))
return false; return false;
...@@ -2823,7 +2837,7 @@ static void retarget_shared_pending(struct task_struct *tsk, sigset_t *which) ...@@ -2823,7 +2837,7 @@ static void retarget_shared_pending(struct task_struct *tsk, sigset_t *which)
/* Remove the signals this thread can handle. */ /* Remove the signals this thread can handle. */
sigandsets(&retarget, &retarget, &t->blocked); sigandsets(&retarget, &retarget, &t->blocked);
if (!signal_pending(t)) if (!task_sigpending(t))
signal_wake_up(t, 0); signal_wake_up(t, 0);
if (sigisemptyset(&retarget)) if (sigisemptyset(&retarget))
...@@ -2857,7 +2871,7 @@ void exit_signals(struct task_struct *tsk) ...@@ -2857,7 +2871,7 @@ void exit_signals(struct task_struct *tsk)
cgroup_threadgroup_change_end(tsk); cgroup_threadgroup_change_end(tsk);
if (!signal_pending(tsk)) if (!task_sigpending(tsk))
goto out; goto out;
unblocked = tsk->blocked; unblocked = tsk->blocked;
...@@ -2901,7 +2915,7 @@ long do_no_restart_syscall(struct restart_block *param) ...@@ -2901,7 +2915,7 @@ long do_no_restart_syscall(struct restart_block *param)
static void __set_task_blocked(struct task_struct *tsk, const sigset_t *newset) static void __set_task_blocked(struct task_struct *tsk, const sigset_t *newset)
{ {
if (signal_pending(tsk) && !thread_group_empty(tsk)) { if (task_sigpending(tsk) && !thread_group_empty(tsk)) {
sigset_t newblocked; sigset_t newblocked;
/* A set of now blocked but previously unblocked signals. */ /* A set of now blocked but previously unblocked signals. */
sigandnsets(&newblocked, newset, &current->blocked); sigandnsets(&newblocked, newset, &current->blocked);
......
...@@ -5,6 +5,34 @@ ...@@ -5,6 +5,34 @@
static struct callback_head work_exited; /* all we need is ->next == NULL */ static struct callback_head work_exited; /* all we need is ->next == NULL */
/*
* TWA_SIGNAL signaling - use TIF_NOTIFY_SIGNAL, if available, as it's faster
* than TIF_SIGPENDING as there's no dependency on ->sighand. The latter is
* shared for threads, and can cause contention on sighand->lock. Even for
* the non-threaded case TIF_NOTIFY_SIGNAL is more efficient, as no locking
* or IRQ disabling is involved for notification (or running) purposes.
*/
static void task_work_notify_signal(struct task_struct *task)
{
#if defined(TIF_NOTIFY_SIGNAL)
set_notify_signal(task);
#else
unsigned long flags;
/*
* Only grab the sighand lock if we don't already have some
* task_work pending. This pairs with the smp_store_mb()
* in get_signal(), see comment there.
*/
if (!(READ_ONCE(task->jobctl) & JOBCTL_TASK_WORK) &&
lock_task_sighand(task, &flags)) {
task->jobctl |= JOBCTL_TASK_WORK;
signal_wake_up(task, 0);
unlock_task_sighand(task, &flags);
}
#endif
}
/** /**
* task_work_add - ask the @task to execute @work->func() * task_work_add - ask the @task to execute @work->func()
* @task: the task which should run the callback * @task: the task which should run the callback
...@@ -33,7 +61,6 @@ int task_work_add(struct task_struct *task, struct callback_head *work, ...@@ -33,7 +61,6 @@ int task_work_add(struct task_struct *task, struct callback_head *work,
enum task_work_notify_mode notify) enum task_work_notify_mode notify)
{ {
struct callback_head *head; struct callback_head *head;
unsigned long flags;
do { do {
head = READ_ONCE(task->task_works); head = READ_ONCE(task->task_works);
...@@ -49,17 +76,7 @@ int task_work_add(struct task_struct *task, struct callback_head *work, ...@@ -49,17 +76,7 @@ int task_work_add(struct task_struct *task, struct callback_head *work,
set_notify_resume(task); set_notify_resume(task);
break; break;
case TWA_SIGNAL: case TWA_SIGNAL:
/* task_work_notify_signal(task);
* Only grab the sighand lock if we don't already have some
* task_work pending. This pairs with the smp_store_mb()
* in get_signal(), see comment there.
*/
if (!(READ_ONCE(task->jobctl) & JOBCTL_TASK_WORK) &&
lock_task_sighand(task, &flags)) {
task->jobctl |= JOBCTL_TASK_WORK;
signal_wake_up(task, 0);
unlock_task_sighand(task, &flags);
}
break; break;
default: default:
WARN_ON_ONCE(1); WARN_ON_ONCE(1);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment