Commit b538ed27 authored by Jan Beulich's avatar Jan Beulich Committed by Linus Torvalds

[PATCH] x86_64: reliable stack trace support (x86-64)

These are the x86_64-specific pieces to enable reliable stack traces. The
only restriction with this is that it currently cannot unwind across the
interrupt->normal stack boundary, as that transition is lacking proper
annotation.
Signed-off-by: default avatarJan Beulich <jbeulich@novell.com>
Signed-off-by: default avatarAndi Kleen <ak@suse.de>
Signed-off-by: default avatarLinus Torvalds <torvalds@osdl.org>
parent 4552d5dc
...@@ -1089,3 +1089,36 @@ ENTRY(call_softirq) ...@@ -1089,3 +1089,36 @@ ENTRY(call_softirq)
ret ret
CFI_ENDPROC CFI_ENDPROC
ENDPROC(call_softirq) ENDPROC(call_softirq)
#ifdef CONFIG_STACK_UNWIND
ENTRY(arch_unwind_init_running)
CFI_STARTPROC
movq %r15, R15(%rdi)
movq %r14, R14(%rdi)
xchgq %rsi, %rdx
movq %r13, R13(%rdi)
movq %r12, R12(%rdi)
xorl %eax, %eax
movq %rbp, RBP(%rdi)
movq %rbx, RBX(%rdi)
movq (%rsp), %rcx
movq %rax, R11(%rdi)
movq %rax, R10(%rdi)
movq %rax, R9(%rdi)
movq %rax, R8(%rdi)
movq %rax, RAX(%rdi)
movq %rax, RCX(%rdi)
movq %rax, RDX(%rdi)
movq %rax, RSI(%rdi)
movq %rax, RDI(%rdi)
movq %rax, ORIG_RAX(%rdi)
movq %rcx, RIP(%rdi)
leaq 8(%rsp), %rcx
movq $__KERNEL_CS, CS(%rdi)
movq %rax, EFLAGS(%rdi)
movq %rcx, RSP(%rdi)
movq $__KERNEL_DS, SS(%rdi)
jmpq *%rdx
CFI_ENDPROC
ENDPROC(arch_unwind_init_running)
#endif
...@@ -334,7 +334,7 @@ void show_regs(struct pt_regs *regs) ...@@ -334,7 +334,7 @@ void show_regs(struct pt_regs *regs)
{ {
printk("CPU %d:", smp_processor_id()); printk("CPU %d:", smp_processor_id());
__show_regs(regs); __show_regs(regs);
show_trace(&regs->rsp); show_trace(NULL, regs, (void *)(regs + 1));
} }
/* /*
......
...@@ -29,6 +29,7 @@ ...@@ -29,6 +29,7 @@
#include <linux/nmi.h> #include <linux/nmi.h>
#include <linux/kprobes.h> #include <linux/kprobes.h>
#include <linux/kexec.h> #include <linux/kexec.h>
#include <linux/unwind.h>
#include <asm/system.h> #include <asm/system.h>
#include <asm/uaccess.h> #include <asm/uaccess.h>
...@@ -39,7 +40,7 @@ ...@@ -39,7 +40,7 @@
#include <asm/i387.h> #include <asm/i387.h>
#include <asm/kdebug.h> #include <asm/kdebug.h>
#include <asm/processor.h> #include <asm/processor.h>
#include <asm/unwind.h>
#include <asm/smp.h> #include <asm/smp.h>
#include <asm/pgalloc.h> #include <asm/pgalloc.h>
#include <asm/pda.h> #include <asm/pda.h>
...@@ -189,6 +190,23 @@ static unsigned long *in_exception_stack(unsigned cpu, unsigned long stack, ...@@ -189,6 +190,23 @@ static unsigned long *in_exception_stack(unsigned cpu, unsigned long stack,
return NULL; return NULL;
} }
static void show_trace_unwind(struct unwind_frame_info *info, void *context)
{
int i = 11;
while (unwind(info) == 0 && UNW_PC(info)) {
if (i > 50) {
printk("\n ");
i = 7;
} else
i += printk(" ");
i += printk_address(UNW_PC(info));
if (arch_unw_user_mode(info))
break;
}
printk("\n");
}
/* /*
* x86-64 can have upto three kernel stacks: * x86-64 can have upto three kernel stacks:
* process stack * process stack
...@@ -196,15 +214,34 @@ static unsigned long *in_exception_stack(unsigned cpu, unsigned long stack, ...@@ -196,15 +214,34 @@ static unsigned long *in_exception_stack(unsigned cpu, unsigned long stack,
* severe exception (double fault, nmi, stack fault, debug, mce) hardware stack * severe exception (double fault, nmi, stack fault, debug, mce) hardware stack
*/ */
void show_trace(unsigned long *stack) void show_trace(struct task_struct *tsk, struct pt_regs *regs, unsigned long * stack)
{ {
const unsigned cpu = safe_smp_processor_id(); const unsigned cpu = safe_smp_processor_id();
unsigned long *irqstack_end = (unsigned long *)cpu_pda(cpu)->irqstackptr; unsigned long *irqstack_end = (unsigned long *)cpu_pda(cpu)->irqstackptr;
int i; int i;
unsigned used = 0; unsigned used = 0;
struct unwind_frame_info info;
printk("\nCall Trace:"); printk("\nCall Trace:");
if (!tsk)
tsk = current;
if (regs) {
if (unwind_init_frame_info(&info, tsk, regs) == 0) {
show_trace_unwind(&info, NULL);
return;
}
} else if (tsk == current) {
if (unwind_init_running(&info, show_trace_unwind, NULL) == 0)
return;
} else {
if (unwind_init_blocked(&info, tsk) == 0) {
show_trace_unwind(&info, NULL);
return;
}
}
#define HANDLE_STACK(cond) \ #define HANDLE_STACK(cond) \
do while (cond) { \ do while (cond) { \
unsigned long addr = *stack++; \ unsigned long addr = *stack++; \
...@@ -262,7 +299,7 @@ void show_trace(unsigned long *stack) ...@@ -262,7 +299,7 @@ void show_trace(unsigned long *stack)
printk("\n"); printk("\n");
} }
void show_stack(struct task_struct *tsk, unsigned long * rsp) static void _show_stack(struct task_struct *tsk, struct pt_regs *regs, unsigned long * rsp)
{ {
unsigned long *stack; unsigned long *stack;
int i; int i;
...@@ -296,7 +333,12 @@ void show_stack(struct task_struct *tsk, unsigned long * rsp) ...@@ -296,7 +333,12 @@ void show_stack(struct task_struct *tsk, unsigned long * rsp)
printk("%016lx ", *stack++); printk("%016lx ", *stack++);
touch_nmi_watchdog(); touch_nmi_watchdog();
} }
show_trace((unsigned long *)rsp); show_trace(tsk, regs, rsp);
}
void show_stack(struct task_struct *tsk, unsigned long * rsp)
{
_show_stack(tsk, NULL, rsp);
} }
/* /*
...@@ -305,7 +347,7 @@ void show_stack(struct task_struct *tsk, unsigned long * rsp) ...@@ -305,7 +347,7 @@ void show_stack(struct task_struct *tsk, unsigned long * rsp)
void dump_stack(void) void dump_stack(void)
{ {
unsigned long dummy; unsigned long dummy;
show_trace(&dummy); show_trace(NULL, NULL, &dummy);
} }
EXPORT_SYMBOL(dump_stack); EXPORT_SYMBOL(dump_stack);
...@@ -332,7 +374,7 @@ void show_registers(struct pt_regs *regs) ...@@ -332,7 +374,7 @@ void show_registers(struct pt_regs *regs)
if (in_kernel) { if (in_kernel) {
printk("Stack: "); printk("Stack: ");
show_stack(NULL, (unsigned long*)rsp); _show_stack(NULL, regs, (unsigned long*)rsp);
printk("\nCode: "); printk("\nCode: ");
if (regs->rip < PAGE_OFFSET) if (regs->rip < PAGE_OFFSET)
......
...@@ -45,6 +45,15 @@ SECTIONS ...@@ -45,6 +45,15 @@ SECTIONS
RODATA RODATA
#ifdef CONFIG_STACK_UNWIND
. = ALIGN(8);
.eh_frame : AT(ADDR(.eh_frame) - LOAD_OFFSET) {
__start_unwind = .;
*(.eh_frame)
__end_unwind = .;
}
#endif
/* Data */ /* Data */
.data : AT(ADDR(.data) - LOAD_OFFSET) { .data : AT(ADDR(.data) - LOAD_OFFSET) {
*(.data) *(.data)
......
...@@ -75,7 +75,7 @@ extern void main_timer_handler(struct pt_regs *regs); ...@@ -75,7 +75,7 @@ extern void main_timer_handler(struct pt_regs *regs);
extern unsigned long end_pfn_map; extern unsigned long end_pfn_map;
extern void show_trace(unsigned long * rsp); extern void show_trace(struct task_struct *, struct pt_regs *, unsigned long * rsp);
extern void show_registers(struct pt_regs *regs); extern void show_registers(struct pt_regs *regs);
extern void exception_table_check(void); extern void exception_table_check(void);
......
#ifndef _ASM_X86_64_UNWIND_H
#define _ASM_X86_64_UNWIND_H
/*
* Copyright (C) 2002-2006 Novell, Inc.
* Jan Beulich <jbeulich@novell.com>
* This code is released under version 2 of the GNU GPL.
*/
#ifdef CONFIG_STACK_UNWIND
#include <linux/sched.h>
#include <asm/ptrace.h>
#include <asm/uaccess.h>
#include <asm/vsyscall.h>
struct unwind_frame_info
{
struct pt_regs regs;
struct task_struct *task;
};
#define UNW_PC(frame) (frame)->regs.rip
#define UNW_SP(frame) (frame)->regs.rsp
#ifdef CONFIG_FRAME_POINTER
#define UNW_FP(frame) (frame)->regs.rbp
#define FRAME_RETADDR_OFFSET 8
#define FRAME_LINK_OFFSET 0
#define STACK_BOTTOM(tsk) (((tsk)->thread.rsp0 - 1) & ~(THREAD_SIZE - 1))
#define STACK_TOP(tsk) ((tsk)->thread.rsp0)
#endif
/* Might need to account for the special exception and interrupt handling
stacks here, since normally
EXCEPTION_STACK_ORDER < THREAD_ORDER < IRQSTACK_ORDER,
but the construct is needed only for getting across the stack switch to
the interrupt stack - thus considering the IRQ stack itself is unnecessary,
and the overhead of comparing against all exception handling stacks seems
not desirable. */
#define STACK_LIMIT(ptr) (((ptr) - 1) & ~(THREAD_SIZE - 1))
#define UNW_REGISTER_INFO \
PTREGS_INFO(rax), \
PTREGS_INFO(rdx), \
PTREGS_INFO(rcx), \
PTREGS_INFO(rbx), \
PTREGS_INFO(rsi), \
PTREGS_INFO(rdi), \
PTREGS_INFO(rbp), \
PTREGS_INFO(rsp), \
PTREGS_INFO(r8), \
PTREGS_INFO(r9), \
PTREGS_INFO(r10), \
PTREGS_INFO(r11), \
PTREGS_INFO(r12), \
PTREGS_INFO(r13), \
PTREGS_INFO(r14), \
PTREGS_INFO(r15), \
PTREGS_INFO(rip)
static inline void arch_unw_init_frame_info(struct unwind_frame_info *info,
/*const*/ struct pt_regs *regs)
{
info->regs = *regs;
}
static inline void arch_unw_init_blocked(struct unwind_frame_info *info)
{
extern const char thread_return[];
memset(&info->regs, 0, sizeof(info->regs));
info->regs.rip = (unsigned long)thread_return;
info->regs.cs = __KERNEL_CS;
__get_user(info->regs.rbp, (unsigned long *)info->task->thread.rsp);
info->regs.rsp = info->task->thread.rsp;
info->regs.ss = __KERNEL_DS;
}
extern void arch_unwind_init_running(struct unwind_frame_info *,
void (*callback)(struct unwind_frame_info *,
void *arg),
void *arg);
static inline int arch_unw_user_mode(const struct unwind_frame_info *info)
{
#if 0 /* This can only work when selector register saves/restores
are properly annotated (and tracked in UNW_REGISTER_INFO). */
return user_mode(&info->regs);
#else
return (long)info->regs.rip >= 0
|| (info->regs.rip >= VSYSCALL_START && info->regs.rip < VSYSCALL_END)
|| (long)info->regs.rsp >= 0;
#endif
}
#else
#define UNW_PC(frame) ((void)(frame), 0)
static inline int arch_unw_user_mode(const void *info)
{
return 0;
}
#endif
#endif /* _ASM_X86_64_UNWIND_H */
...@@ -199,7 +199,7 @@ config UNWIND_INFO ...@@ -199,7 +199,7 @@ config UNWIND_INFO
config STACK_UNWIND config STACK_UNWIND
bool "Stack unwind support" bool "Stack unwind support"
depends on UNWIND_INFO depends on UNWIND_INFO
depends on n depends on X86_64
help help
This enables more precise stack traces, omitting all unrelated This enables more precise stack traces, omitting all unrelated
occurrences of pointers into kernel code from the dump. occurrences of pointers into kernel code from the dump.
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment