Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
L
linux
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
Analytics
Analytics
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Commits
Issue Boards
Open sidebar
nexedi
linux
Commits
ec1bb60b
Commit
ec1bb60b
authored
Jul 10, 2008
by
Ingo Molnar
Browse files
Options
Browse Files
Download
Plain Diff
Merge branch 'tracing/sysprof' into auto-ftrace-next
parents
5373fdbd
37f5d732
Changes
6
Show whitespace changes
Inline
Side-by-side
Showing
6 changed files
with
404 additions
and
0 deletions
+404
-0
kernel/trace/Kconfig
kernel/trace/Kconfig
+8
-0
kernel/trace/Makefile
kernel/trace/Makefile
+1
-0
kernel/trace/trace.c
kernel/trace/trace.c
+3
-0
kernel/trace/trace.h
kernel/trace/trace.h
+6
-0
kernel/trace/trace_selftest.c
kernel/trace/trace_selftest.c
+23
-0
kernel/trace/trace_sysprof.c
kernel/trace/trace_sysprof.c
+363
-0
No files found.
kernel/trace/Kconfig
View file @
ec1bb60b
...
@@ -75,6 +75,14 @@ config PREEMPT_TRACER
...
@@ -75,6 +75,14 @@ config PREEMPT_TRACER
enabled. This option and the irqs-off timing option can be
enabled. This option and the irqs-off timing option can be
used together or separately.)
used together or separately.)
config SYSPROF_TRACER
bool "Sysprof Tracer"
depends on X86
select TRACING
help
This tracer provides the trace needed by the 'Sysprof' userspace
tool.
config SCHED_TRACER
config SCHED_TRACER
bool "Scheduling Latency Tracer"
bool "Scheduling Latency Tracer"
depends on HAVE_FTRACE
depends on HAVE_FTRACE
...
...
kernel/trace/Makefile
View file @
ec1bb60b
...
@@ -14,6 +14,7 @@ obj-$(CONFIG_FTRACE) += libftrace.o
...
@@ -14,6 +14,7 @@ obj-$(CONFIG_FTRACE) += libftrace.o
obj-$(CONFIG_TRACING)
+=
trace.o
obj-$(CONFIG_TRACING)
+=
trace.o
obj-$(CONFIG_CONTEXT_SWITCH_TRACER)
+=
trace_sched_switch.o
obj-$(CONFIG_CONTEXT_SWITCH_TRACER)
+=
trace_sched_switch.o
obj-$(CONFIG_SYSPROF_TRACER)
+=
trace_sysprof.o
obj-$(CONFIG_FTRACE)
+=
trace_functions.o
obj-$(CONFIG_FTRACE)
+=
trace_functions.o
obj-$(CONFIG_IRQSOFF_TRACER)
+=
trace_irqsoff.o
obj-$(CONFIG_IRQSOFF_TRACER)
+=
trace_irqsoff.o
obj-$(CONFIG_PREEMPT_TRACER)
+=
trace_irqsoff.o
obj-$(CONFIG_PREEMPT_TRACER)
+=
trace_irqsoff.o
...
...
kernel/trace/trace.c
View file @
ec1bb60b
...
@@ -2913,6 +2913,9 @@ static __init void tracer_init_debugfs(void)
...
@@ -2913,6 +2913,9 @@ static __init void tracer_init_debugfs(void)
pr_warning
(
"Could not create debugfs "
pr_warning
(
"Could not create debugfs "
"'dyn_ftrace_total_info' entry
\n
"
);
"'dyn_ftrace_total_info' entry
\n
"
);
#endif
#endif
#ifdef CONFIG_SYSPROF_TRACER
init_tracer_sysprof_debugfs
(
d_tracer
);
#endif
}
}
static
int
trace_alloc_page
(
void
)
static
int
trace_alloc_page
(
void
)
...
...
kernel/trace/trace.h
View file @
ec1bb60b
...
@@ -193,6 +193,8 @@ struct trace_iterator {
...
@@ -193,6 +193,8 @@ struct trace_iterator {
void
tracing_reset
(
struct
trace_array_cpu
*
data
);
void
tracing_reset
(
struct
trace_array_cpu
*
data
);
int
tracing_open_generic
(
struct
inode
*
inode
,
struct
file
*
filp
);
int
tracing_open_generic
(
struct
inode
*
inode
,
struct
file
*
filp
);
struct
dentry
*
tracing_init_dentry
(
void
);
struct
dentry
*
tracing_init_dentry
(
void
);
void
init_tracer_sysprof_debugfs
(
struct
dentry
*
d_tracer
);
void
ftrace
(
struct
trace_array
*
tr
,
void
ftrace
(
struct
trace_array
*
tr
,
struct
trace_array_cpu
*
data
,
struct
trace_array_cpu
*
data
,
unsigned
long
ip
,
unsigned
long
ip
,
...
@@ -294,6 +296,10 @@ extern int trace_selftest_startup_wakeup(struct tracer *trace,
...
@@ -294,6 +296,10 @@ extern int trace_selftest_startup_wakeup(struct tracer *trace,
extern
int
trace_selftest_startup_sched_switch
(
struct
tracer
*
trace
,
extern
int
trace_selftest_startup_sched_switch
(
struct
tracer
*
trace
,
struct
trace_array
*
tr
);
struct
trace_array
*
tr
);
#endif
#endif
#ifdef CONFIG_SYSPROF_TRACER
extern
int
trace_selftest_startup_sysprof
(
struct
tracer
*
trace
,
struct
trace_array
*
tr
);
#endif
#endif
/* CONFIG_FTRACE_STARTUP_TEST */
#endif
/* CONFIG_FTRACE_STARTUP_TEST */
extern
void
*
head_page
(
struct
trace_array_cpu
*
data
);
extern
void
*
head_page
(
struct
trace_array_cpu
*
data
);
...
...
kernel/trace/trace_selftest.c
View file @
ec1bb60b
...
@@ -538,3 +538,26 @@ trace_selftest_startup_sched_switch(struct tracer *trace, struct trace_array *tr
...
@@ -538,3 +538,26 @@ trace_selftest_startup_sched_switch(struct tracer *trace, struct trace_array *tr
return
ret
;
return
ret
;
}
}
#endif
/* CONFIG_CONTEXT_SWITCH_TRACER */
#endif
/* CONFIG_CONTEXT_SWITCH_TRACER */
#ifdef CONFIG_SYSPROF_TRACER
int
trace_selftest_startup_sysprof
(
struct
tracer
*
trace
,
struct
trace_array
*
tr
)
{
unsigned
long
count
;
int
ret
;
/* start the tracing */
tr
->
ctrl
=
1
;
trace
->
init
(
tr
);
/* Sleep for a 1/10 of a second */
msleep
(
100
);
/* stop the tracing. */
tr
->
ctrl
=
0
;
trace
->
ctrl_update
(
tr
);
/* check the trace buffer */
ret
=
trace_test_buffer
(
tr
,
&
count
);
trace
->
reset
(
tr
);
return
ret
;
}
#endif
/* CONFIG_SYSPROF_TRACER */
kernel/trace/trace_sysprof.c
0 → 100644
View file @
ec1bb60b
/*
* trace stack traces
*
* Copyright (C) 2004-2008, Soeren Sandmann
* Copyright (C) 2007 Steven Rostedt <srostedt@redhat.com>
* Copyright (C) 2008 Ingo Molnar <mingo@redhat.com>
*/
#include <linux/kallsyms.h>
#include <linux/debugfs.h>
#include <linux/hrtimer.h>
#include <linux/uaccess.h>
#include <linux/ftrace.h>
#include <linux/module.h>
#include <linux/irq.h>
#include <linux/fs.h>
#include <asm/stacktrace.h>
#include "trace.h"
static
struct
trace_array
*
sysprof_trace
;
static
int
__read_mostly
tracer_enabled
;
/*
* 1 msec sample interval by default:
*/
static
unsigned
long
sample_period
=
1000000
;
static
const
unsigned
int
sample_max_depth
=
512
;
static
DEFINE_MUTEX
(
sample_timer_lock
);
/*
* Per CPU hrtimers that do the profiling:
*/
static
DEFINE_PER_CPU
(
struct
hrtimer
,
stack_trace_hrtimer
);
struct
stack_frame
{
const
void
__user
*
next_fp
;
unsigned
long
return_address
;
};
static
int
copy_stack_frame
(
const
void
__user
*
fp
,
struct
stack_frame
*
frame
)
{
int
ret
;
if
(
!
access_ok
(
VERIFY_READ
,
fp
,
sizeof
(
*
frame
)))
return
0
;
ret
=
1
;
pagefault_disable
();
if
(
__copy_from_user_inatomic
(
frame
,
fp
,
sizeof
(
*
frame
)))
ret
=
0
;
pagefault_enable
();
return
ret
;
}
struct
backtrace_info
{
struct
trace_array_cpu
*
data
;
struct
trace_array
*
tr
;
int
pos
;
};
static
void
backtrace_warning_symbol
(
void
*
data
,
char
*
msg
,
unsigned
long
symbol
)
{
/* Ignore warnings */
}
static
void
backtrace_warning
(
void
*
data
,
char
*
msg
)
{
/* Ignore warnings */
}
static
int
backtrace_stack
(
void
*
data
,
char
*
name
)
{
/* Don't bother with IRQ stacks for now */
return
-
1
;
}
static
void
backtrace_address
(
void
*
data
,
unsigned
long
addr
,
int
reliable
)
{
struct
backtrace_info
*
info
=
data
;
if
(
info
->
pos
<
sample_max_depth
&&
reliable
)
{
__trace_special
(
info
->
tr
,
info
->
data
,
1
,
addr
,
0
);
info
->
pos
++
;
}
}
const
static
struct
stacktrace_ops
backtrace_ops
=
{
.
warning
=
backtrace_warning
,
.
warning_symbol
=
backtrace_warning_symbol
,
.
stack
=
backtrace_stack
,
.
address
=
backtrace_address
,
};
static
int
trace_kernel
(
struct
pt_regs
*
regs
,
struct
trace_array
*
tr
,
struct
trace_array_cpu
*
data
)
{
struct
backtrace_info
info
;
unsigned
long
bp
;
char
*
stack
;
info
.
tr
=
tr
;
info
.
data
=
data
;
info
.
pos
=
1
;
__trace_special
(
info
.
tr
,
info
.
data
,
1
,
regs
->
ip
,
0
);
stack
=
((
char
*
)
regs
+
sizeof
(
struct
pt_regs
));
#ifdef CONFIG_FRAME_POINTER
bp
=
regs
->
bp
;
#else
bp
=
0
;
#endif
dump_trace
(
NULL
,
regs
,
(
void
*
)
stack
,
bp
,
&
backtrace_ops
,
&
info
);
return
info
.
pos
;
}
static
void
timer_notify
(
struct
pt_regs
*
regs
,
int
cpu
)
{
struct
trace_array_cpu
*
data
;
struct
stack_frame
frame
;
struct
trace_array
*
tr
;
const
void
__user
*
fp
;
int
is_user
;
int
i
;
if
(
!
regs
)
return
;
tr
=
sysprof_trace
;
data
=
tr
->
data
[
cpu
];
is_user
=
user_mode
(
regs
);
if
(
!
current
||
current
->
pid
==
0
)
return
;
if
(
is_user
&&
current
->
state
!=
TASK_RUNNING
)
return
;
__trace_special
(
tr
,
data
,
0
,
0
,
current
->
pid
);
if
(
!
is_user
)
i
=
trace_kernel
(
regs
,
tr
,
data
);
else
i
=
0
;
/*
* Trace user stack if we are not a kernel thread
*/
if
(
current
->
mm
&&
i
<
sample_max_depth
)
{
regs
=
(
struct
pt_regs
*
)
current
->
thread
.
sp0
-
1
;
fp
=
(
void
__user
*
)
regs
->
bp
;
__trace_special
(
tr
,
data
,
2
,
regs
->
ip
,
0
);
while
(
i
<
sample_max_depth
)
{
frame
.
next_fp
=
0
;
frame
.
return_address
=
0
;
if
(
!
copy_stack_frame
(
fp
,
&
frame
))
break
;
if
((
unsigned
long
)
fp
<
regs
->
sp
)
break
;
__trace_special
(
tr
,
data
,
2
,
frame
.
return_address
,
(
unsigned
long
)
fp
);
fp
=
frame
.
next_fp
;
i
++
;
}
}
/*
* Special trace entry if we overflow the max depth:
*/
if
(
i
==
sample_max_depth
)
__trace_special
(
tr
,
data
,
-
1
,
-
1
,
-
1
);
__trace_special
(
tr
,
data
,
3
,
current
->
pid
,
i
);
}
static
enum
hrtimer_restart
stack_trace_timer_fn
(
struct
hrtimer
*
hrtimer
)
{
/* trace here */
timer_notify
(
get_irq_regs
(),
smp_processor_id
());
hrtimer_forward_now
(
hrtimer
,
ns_to_ktime
(
sample_period
));
return
HRTIMER_RESTART
;
}
static
void
start_stack_timer
(
int
cpu
)
{
struct
hrtimer
*
hrtimer
=
&
per_cpu
(
stack_trace_hrtimer
,
cpu
);
hrtimer_init
(
hrtimer
,
CLOCK_MONOTONIC
,
HRTIMER_MODE_REL
);
hrtimer
->
function
=
stack_trace_timer_fn
;
hrtimer
->
cb_mode
=
HRTIMER_CB_IRQSAFE_NO_SOFTIRQ
;
hrtimer_start
(
hrtimer
,
ns_to_ktime
(
sample_period
),
HRTIMER_MODE_REL
);
}
static
void
start_stack_timers
(
void
)
{
cpumask_t
saved_mask
=
current
->
cpus_allowed
;
int
cpu
;
for_each_online_cpu
(
cpu
)
{
set_cpus_allowed_ptr
(
current
,
&
cpumask_of_cpu
(
cpu
));
start_stack_timer
(
cpu
);
}
set_cpus_allowed_ptr
(
current
,
&
saved_mask
);
}
static
void
stop_stack_timer
(
int
cpu
)
{
struct
hrtimer
*
hrtimer
=
&
per_cpu
(
stack_trace_hrtimer
,
cpu
);
hrtimer_cancel
(
hrtimer
);
}
static
void
stop_stack_timers
(
void
)
{
int
cpu
;
for_each_online_cpu
(
cpu
)
stop_stack_timer
(
cpu
);
}
static
void
stack_reset
(
struct
trace_array
*
tr
)
{
int
cpu
;
tr
->
time_start
=
ftrace_now
(
tr
->
cpu
);
for_each_online_cpu
(
cpu
)
tracing_reset
(
tr
->
data
[
cpu
]);
}
static
void
start_stack_trace
(
struct
trace_array
*
tr
)
{
mutex_lock
(
&
sample_timer_lock
);
stack_reset
(
tr
);
start_stack_timers
();
tracer_enabled
=
1
;
mutex_unlock
(
&
sample_timer_lock
);
}
static
void
stop_stack_trace
(
struct
trace_array
*
tr
)
{
mutex_lock
(
&
sample_timer_lock
);
stop_stack_timers
();
tracer_enabled
=
0
;
mutex_unlock
(
&
sample_timer_lock
);
}
static
void
stack_trace_init
(
struct
trace_array
*
tr
)
{
sysprof_trace
=
tr
;
if
(
tr
->
ctrl
)
start_stack_trace
(
tr
);
}
static
void
stack_trace_reset
(
struct
trace_array
*
tr
)
{
if
(
tr
->
ctrl
)
stop_stack_trace
(
tr
);
}
static
void
stack_trace_ctrl_update
(
struct
trace_array
*
tr
)
{
/* When starting a new trace, reset the buffers */
if
(
tr
->
ctrl
)
start_stack_trace
(
tr
);
else
stop_stack_trace
(
tr
);
}
static
struct
tracer
stack_trace
__read_mostly
=
{
.
name
=
"sysprof"
,
.
init
=
stack_trace_init
,
.
reset
=
stack_trace_reset
,
.
ctrl_update
=
stack_trace_ctrl_update
,
#ifdef CONFIG_FTRACE_SELFTEST
.
selftest
=
trace_selftest_startup_sysprof
,
#endif
};
__init
static
int
init_stack_trace
(
void
)
{
return
register_tracer
(
&
stack_trace
);
}
device_initcall
(
init_stack_trace
);
#define MAX_LONG_DIGITS 22
static
ssize_t
sysprof_sample_read
(
struct
file
*
filp
,
char
__user
*
ubuf
,
size_t
cnt
,
loff_t
*
ppos
)
{
char
buf
[
MAX_LONG_DIGITS
];
int
r
;
r
=
sprintf
(
buf
,
"%ld
\n
"
,
nsecs_to_usecs
(
sample_period
));
return
simple_read_from_buffer
(
ubuf
,
cnt
,
ppos
,
buf
,
r
);
}
static
ssize_t
sysprof_sample_write
(
struct
file
*
filp
,
const
char
__user
*
ubuf
,
size_t
cnt
,
loff_t
*
ppos
)
{
char
buf
[
MAX_LONG_DIGITS
];
unsigned
long
val
;
if
(
cnt
>
MAX_LONG_DIGITS
-
1
)
cnt
=
MAX_LONG_DIGITS
-
1
;
if
(
copy_from_user
(
&
buf
,
ubuf
,
cnt
))
return
-
EFAULT
;
buf
[
cnt
]
=
0
;
val
=
simple_strtoul
(
buf
,
NULL
,
10
);
/*
* Enforce a minimum sample period of 100 usecs:
*/
if
(
val
<
100
)
val
=
100
;
mutex_lock
(
&
sample_timer_lock
);
stop_stack_timers
();
sample_period
=
val
*
1000
;
start_stack_timers
();
mutex_unlock
(
&
sample_timer_lock
);
return
cnt
;
}
static
struct
file_operations
sysprof_sample_fops
=
{
.
read
=
sysprof_sample_read
,
.
write
=
sysprof_sample_write
,
};
void
init_tracer_sysprof_debugfs
(
struct
dentry
*
d_tracer
)
{
struct
dentry
*
entry
;
entry
=
debugfs_create_file
(
"sysprof_sample_period"
,
0644
,
d_tracer
,
NULL
,
&
sysprof_sample_fops
);
if
(
entry
)
return
;
pr_warning
(
"Could not create debugfs 'dyn_ftrace_total_info' entry
\n
"
);
}
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment