Commit fd49e80d authored by Anton Blanchard's avatar Anton Blanchard

Merge samba.org:/scratch/anton/linux-2.5

into samba.org:/scratch/anton/sfr
parents 57af183b 23f3c02f
...@@ -58,7 +58,7 @@ static void nmi_save_registers(struct op_msrs * msrs) ...@@ -58,7 +58,7 @@ static void nmi_save_registers(struct op_msrs * msrs)
unsigned int const nr_ctrls = model->num_controls; unsigned int const nr_ctrls = model->num_controls;
struct op_msr_group * counters = &msrs->counters; struct op_msr_group * counters = &msrs->counters;
struct op_msr_group * controls = &msrs->controls; struct op_msr_group * controls = &msrs->controls;
int i; unsigned int i;
for (i = 0; i < nr_ctrs; ++i) { for (i = 0; i < nr_ctrs; ++i) {
rdmsr(counters->addrs[i], rdmsr(counters->addrs[i],
...@@ -108,7 +108,7 @@ static void nmi_restore_registers(struct op_msrs * msrs) ...@@ -108,7 +108,7 @@ static void nmi_restore_registers(struct op_msrs * msrs)
unsigned int const nr_ctrls = model->num_controls; unsigned int const nr_ctrls = model->num_controls;
struct op_msr_group * counters = &msrs->counters; struct op_msr_group * counters = &msrs->counters;
struct op_msr_group * controls = &msrs->controls; struct op_msr_group * controls = &msrs->controls;
int i; unsigned int i;
for (i = 0; i < nr_ctrls; ++i) { for (i = 0; i < nr_ctrls; ++i) {
wrmsr(controls->addrs[i], wrmsr(controls->addrs[i],
...@@ -182,7 +182,7 @@ struct op_counter_config counter_config[OP_MAX_COUNTER]; ...@@ -182,7 +182,7 @@ struct op_counter_config counter_config[OP_MAX_COUNTER];
static int nmi_create_files(struct super_block * sb, struct dentry * root) static int nmi_create_files(struct super_block * sb, struct dentry * root)
{ {
int i; unsigned int i;
for (i = 0; i < model->num_counters; ++i) { for (i = 0; i < model->num_counters; ++i) {
struct dentry * dir; struct dentry * dir;
......
...@@ -389,7 +389,7 @@ static unsigned long reset_value[NUM_COUNTERS_NON_HT]; ...@@ -389,7 +389,7 @@ static unsigned long reset_value[NUM_COUNTERS_NON_HT];
static void p4_fill_in_addresses(struct op_msrs * const msrs) static void p4_fill_in_addresses(struct op_msrs * const msrs)
{ {
int i; unsigned int i;
unsigned int addr, stag; unsigned int addr, stag;
setup_num_counters(); setup_num_counters();
......
...@@ -59,12 +59,9 @@ void platform_device_unregister(struct platform_device * pdev) ...@@ -59,12 +59,9 @@ void platform_device_unregister(struct platform_device * pdev)
static int platform_match(struct device * dev, struct device_driver * drv) static int platform_match(struct device * dev, struct device_driver * drv)
{ {
char name[BUS_ID_SIZE]; struct platform_device *pdev = container_of(dev, struct platform_device, dev);
if (sscanf(dev->bus_id,"%s",name)) return (strncmp(pdev->name, drv->name, BUS_ID_SIZE) == 0);
return (strcmp(name,drv->name) == 0);
return 0;
} }
struct bus_type platform_bus_type = { struct bus_type platform_bus_type = {
......
...@@ -511,7 +511,7 @@ static int cpufreq_governor_userspace(struct cpufreq_policy *policy, ...@@ -511,7 +511,7 @@ static int cpufreq_governor_userspace(struct cpufreq_policy *policy,
cpu_min_freq[cpu] = policy->min; cpu_min_freq[cpu] = policy->min;
cpu_max_freq[cpu] = policy->max; cpu_max_freq[cpu] = policy->max;
cpu_cur_freq[cpu] = policy->cur; cpu_cur_freq[cpu] = policy->cur;
device_create_file (policy->intf.dev, &dev_attr_scaling_setspeed); device_create_file (policy->dev, &dev_attr_scaling_setspeed);
memcpy (&current_policy[cpu], policy, sizeof(struct cpufreq_policy)); memcpy (&current_policy[cpu], policy, sizeof(struct cpufreq_policy));
up(&userspace_sem); up(&userspace_sem);
break; break;
...@@ -520,7 +520,7 @@ static int cpufreq_governor_userspace(struct cpufreq_policy *policy, ...@@ -520,7 +520,7 @@ static int cpufreq_governor_userspace(struct cpufreq_policy *policy,
cpu_is_managed[cpu] = 0; cpu_is_managed[cpu] = 0;
cpu_min_freq[cpu] = 0; cpu_min_freq[cpu] = 0;
cpu_max_freq[cpu] = 0; cpu_max_freq[cpu] = 0;
device_remove_file (policy->intf.dev, &dev_attr_scaling_setspeed); device_remove_file (policy->dev, &dev_attr_scaling_setspeed);
up(&userspace_sem); up(&userspace_sem);
module_put(THIS_MODULE); module_put(THIS_MODULE);
break; break;
......
...@@ -277,10 +277,7 @@ static void release_mm(struct mm_struct * mm) ...@@ -277,10 +277,7 @@ static void release_mm(struct mm_struct * mm)
*/ */
static struct mm_struct * take_task_mm(struct task_struct * task) static struct mm_struct * take_task_mm(struct task_struct * task)
{ {
struct mm_struct * mm; struct mm_struct * mm = task->mm;
task_lock(task);
mm = task->mm;
task_unlock(task);
/* if task->mm !NULL, mm_count must be at least 1. It cannot /* if task->mm !NULL, mm_count must be at least 1. It cannot
* drop to 0 without the task exiting, which will have to sleep * drop to 0 without the task exiting, which will have to sleep
...@@ -310,6 +307,32 @@ static inline int is_ctx_switch(unsigned long val) ...@@ -310,6 +307,32 @@ static inline int is_ctx_switch(unsigned long val)
} }
/* compute number of filled slots in cpu_buffer queue */
static unsigned long nr_filled_slots(struct oprofile_cpu_buffer * b)
{
unsigned long head = b->head_pos;
unsigned long tail = b->tail_pos;
if (head >= tail)
return head - tail;
return head + (b->buffer_size - tail);
}
static void increment_tail(struct oprofile_cpu_buffer * b)
{
unsigned long new_tail = b->tail_pos + 1;
rmb();
if (new_tail < (b->buffer_size))
b->tail_pos = new_tail;
else
b->tail_pos = 0;
}
/* Sync one of the CPU's buffers into the global event buffer. /* Sync one of the CPU's buffers into the global event buffer.
* Here we need to go through each batch of samples punctuated * Here we need to go through each batch of samples punctuated
* by context switch notes, taking the task's mmap_sem and doing * by context switch notes, taking the task's mmap_sem and doing
...@@ -322,10 +345,14 @@ static void sync_buffer(struct oprofile_cpu_buffer * cpu_buf) ...@@ -322,10 +345,14 @@ static void sync_buffer(struct oprofile_cpu_buffer * cpu_buf)
struct task_struct * new; struct task_struct * new;
unsigned long cookie; unsigned long cookie;
int in_kernel = 1; int in_kernel = 1;
int i; unsigned int i;
for (i=0; i < cpu_buf->pos; ++i) { /* Remember, only we can modify tail_pos */
struct op_sample * s = &cpu_buf->buffer[i];
unsigned long const available_elements = nr_filled_slots(cpu_buf);
for (i=0; i < available_elements; ++i) {
struct op_sample * s = &cpu_buf->buffer[cpu_buf->tail_pos];
if (is_ctx_switch(s->eip)) { if (is_ctx_switch(s->eip)) {
if (s->event <= 1) { if (s->event <= 1) {
...@@ -345,6 +372,8 @@ static void sync_buffer(struct oprofile_cpu_buffer * cpu_buf) ...@@ -345,6 +372,8 @@ static void sync_buffer(struct oprofile_cpu_buffer * cpu_buf)
} else { } else {
add_sample(mm, s, in_kernel); add_sample(mm, s, in_kernel);
} }
increment_tail(cpu_buf);
} }
release_mm(mm); release_mm(mm);
...@@ -369,17 +398,8 @@ static void sync_cpu_buffers(void) ...@@ -369,17 +398,8 @@ static void sync_cpu_buffers(void)
cpu_buf = &cpu_buffer[i]; cpu_buf = &cpu_buffer[i];
/* We take a spin lock even though we might add_cpu_switch(i);
* sleep. It's OK because other users are try sync_buffer(cpu_buf);
* lockers only, and this region is already
* protected by buffer_sem. It's raw to prevent
* the preempt bogometer firing. Fruity, huh ? */
if (cpu_buf->pos > 0) {
_raw_spin_lock(&cpu_buf->int_lock);
add_cpu_switch(i);
sync_buffer(cpu_buf);
_raw_spin_unlock(&cpu_buf->int_lock);
}
} }
up(&buffer_sem); up(&buffer_sem);
......
...@@ -26,8 +26,6 @@ ...@@ -26,8 +26,6 @@
struct oprofile_cpu_buffer cpu_buffer[NR_CPUS] __cacheline_aligned; struct oprofile_cpu_buffer cpu_buffer[NR_CPUS] __cacheline_aligned;
static unsigned long buffer_size;
static void __free_cpu_buffers(int num) static void __free_cpu_buffers(int num)
{ {
int i; int i;
...@@ -47,7 +45,7 @@ int alloc_cpu_buffers(void) ...@@ -47,7 +45,7 @@ int alloc_cpu_buffers(void)
{ {
int i; int i;
buffer_size = fs_cpu_buffer_size; unsigned long buffer_size = fs_cpu_buffer_size;
for (i=0; i < NR_CPUS; ++i) { for (i=0; i < NR_CPUS; ++i) {
struct oprofile_cpu_buffer * b = &cpu_buffer[i]; struct oprofile_cpu_buffer * b = &cpu_buffer[i];
...@@ -59,12 +57,12 @@ int alloc_cpu_buffers(void) ...@@ -59,12 +57,12 @@ int alloc_cpu_buffers(void)
if (!b->buffer) if (!b->buffer)
goto fail; goto fail;
spin_lock_init(&b->int_lock);
b->pos = 0;
b->last_task = 0; b->last_task = 0;
b->last_is_kernel = -1; b->last_is_kernel = -1;
b->buffer_size = buffer_size;
b->tail_pos = 0;
b->head_pos = 0;
b->sample_received = 0; b->sample_received = 0;
b->sample_lost_locked = 0;
b->sample_lost_overflow = 0; b->sample_lost_overflow = 0;
b->sample_lost_task_exit = 0; b->sample_lost_task_exit = 0;
} }
...@@ -80,11 +78,41 @@ void free_cpu_buffers(void) ...@@ -80,11 +78,41 @@ void free_cpu_buffers(void)
__free_cpu_buffers(NR_CPUS); __free_cpu_buffers(NR_CPUS);
} }
/* Note we can't use a semaphore here as this is supposed to /* compute number of available slots in cpu_buffer queue */
* be safe from any context. Instead we trylock the CPU's int_lock. static unsigned long nr_available_slots(struct oprofile_cpu_buffer const * b)
* int_lock is taken by the processing code in sync_cpu_buffers() {
* so we avoid disturbing that. unsigned long head = b->head_pos;
unsigned long tail = b->tail_pos;
if (tail == head)
return b->buffer_size;
if (tail > head)
return tail - head;
return tail + (b->buffer_size - head);
}
static void increment_head(struct oprofile_cpu_buffer * b)
{
unsigned long new_head = b->head_pos + 1;
/* Ensure anything written to the slot before we
* increment is visible */
wmb();
if (new_head < (b->buffer_size))
b->head_pos = new_head;
else
b->head_pos = 0;
}
/* This must be safe from any context. It's safe writing here
* because of the head/tail separation of the writer and reader
* of the CPU buffer.
* *
* is_kernel is needed because on some architectures you cannot * is_kernel is needed because on some architectures you cannot
* tell if you are in kernel or user space simply by looking at * tell if you are in kernel or user space simply by looking at
...@@ -101,14 +129,10 @@ void oprofile_add_sample(unsigned long eip, unsigned int is_kernel, ...@@ -101,14 +129,10 @@ void oprofile_add_sample(unsigned long eip, unsigned int is_kernel,
cpu_buf->sample_received++; cpu_buf->sample_received++;
if (!spin_trylock(&cpu_buf->int_lock)) {
cpu_buf->sample_lost_locked++;
return;
}
if (cpu_buf->pos > buffer_size - 2) { if (nr_available_slots(cpu_buf) < 3) {
cpu_buf->sample_lost_overflow++; cpu_buf->sample_lost_overflow++;
goto out; return;
} }
task = current; task = current;
...@@ -116,18 +140,18 @@ void oprofile_add_sample(unsigned long eip, unsigned int is_kernel, ...@@ -116,18 +140,18 @@ void oprofile_add_sample(unsigned long eip, unsigned int is_kernel,
/* notice a switch from user->kernel or vice versa */ /* notice a switch from user->kernel or vice versa */
if (cpu_buf->last_is_kernel != is_kernel) { if (cpu_buf->last_is_kernel != is_kernel) {
cpu_buf->last_is_kernel = is_kernel; cpu_buf->last_is_kernel = is_kernel;
cpu_buf->buffer[cpu_buf->pos].eip = ~0UL; cpu_buf->buffer[cpu_buf->head_pos].eip = ~0UL;
cpu_buf->buffer[cpu_buf->pos].event = is_kernel; cpu_buf->buffer[cpu_buf->head_pos].event = is_kernel;
cpu_buf->pos++; increment_head(cpu_buf);
} }
/* notice a task switch */ /* notice a task switch */
if (cpu_buf->last_task != task) { if (cpu_buf->last_task != task) {
cpu_buf->last_task = task; cpu_buf->last_task = task;
if (!(task->flags & PF_EXITING)) { if (!(task->flags & PF_EXITING)) {
cpu_buf->buffer[cpu_buf->pos].eip = ~0UL; cpu_buf->buffer[cpu_buf->head_pos].eip = ~0UL;
cpu_buf->buffer[cpu_buf->pos].event = (unsigned long)task; cpu_buf->buffer[cpu_buf->head_pos].event = (unsigned long)task;
cpu_buf->pos++; increment_head(cpu_buf);
} }
} }
...@@ -138,23 +162,20 @@ void oprofile_add_sample(unsigned long eip, unsigned int is_kernel, ...@@ -138,23 +162,20 @@ void oprofile_add_sample(unsigned long eip, unsigned int is_kernel,
*/ */
if (task->flags & PF_EXITING) { if (task->flags & PF_EXITING) {
cpu_buf->sample_lost_task_exit++; cpu_buf->sample_lost_task_exit++;
goto out; return;
} }
cpu_buf->buffer[cpu_buf->pos].eip = eip; cpu_buf->buffer[cpu_buf->head_pos].eip = eip;
cpu_buf->buffer[cpu_buf->pos].event = event; cpu_buf->buffer[cpu_buf->head_pos].event = event;
cpu_buf->pos++; increment_head(cpu_buf);
out:
spin_unlock(&cpu_buf->int_lock);
} }
/* resets the cpu buffer to a sane state - should be called with /* resets the cpu buffer to a sane state - should be called with
* cpu_buf->int_lock held * cpu_buf->int_lock held
*/ */
void cpu_buffer_reset(struct oprofile_cpu_buffer *cpu_buf) void cpu_buffer_reset(struct oprofile_cpu_buffer *cpu_buf)
{ {
cpu_buf->pos = 0;
/* reset these to invalid values; the next sample /* reset these to invalid values; the next sample
* collected will populate the buffer with proper * collected will populate the buffer with proper
* values to initialize the buffer * values to initialize the buffer
...@@ -162,4 +183,3 @@ void cpu_buffer_reset(struct oprofile_cpu_buffer *cpu_buf) ...@@ -162,4 +183,3 @@ void cpu_buffer_reset(struct oprofile_cpu_buffer *cpu_buf)
cpu_buf->last_is_kernel = -1; cpu_buf->last_is_kernel = -1;
cpu_buf->last_task = 0; cpu_buf->last_task = 0;
} }
...@@ -30,14 +30,13 @@ struct op_sample { ...@@ -30,14 +30,13 @@ struct op_sample {
}; };
struct oprofile_cpu_buffer { struct oprofile_cpu_buffer {
spinlock_t int_lock; volatile unsigned long head_pos;
/* protected by int_lock */ volatile unsigned long tail_pos;
unsigned long pos; unsigned long buffer_size;
struct task_struct * last_task; struct task_struct * last_task;
int last_is_kernel; int last_is_kernel;
struct op_sample * buffer; struct op_sample * buffer;
unsigned long sample_received; unsigned long sample_received;
unsigned long sample_lost_locked;
unsigned long sample_lost_overflow; unsigned long sample_lost_overflow;
unsigned long sample_lost_task_exit; unsigned long sample_lost_task_exit;
} ____cacheline_aligned; } ____cacheline_aligned;
......
...@@ -27,7 +27,6 @@ void oprofile_reset_stats(void) ...@@ -27,7 +27,6 @@ void oprofile_reset_stats(void)
cpu_buf = &cpu_buffer[i]; cpu_buf = &cpu_buffer[i];
cpu_buf->sample_received = 0; cpu_buf->sample_received = 0;
cpu_buf->sample_lost_locked = 0;
cpu_buf->sample_lost_overflow = 0; cpu_buf->sample_lost_overflow = 0;
cpu_buf->sample_lost_task_exit = 0; cpu_buf->sample_lost_task_exit = 0;
} }
...@@ -63,8 +62,6 @@ void oprofile_create_stats_files(struct super_block * sb, struct dentry * root) ...@@ -63,8 +62,6 @@ void oprofile_create_stats_files(struct super_block * sb, struct dentry * root)
*/ */
oprofilefs_create_ro_ulong(sb, cpudir, "sample_received", oprofilefs_create_ro_ulong(sb, cpudir, "sample_received",
&cpu_buf->sample_received); &cpu_buf->sample_received);
oprofilefs_create_ro_ulong(sb, cpudir, "sample_lost_locked",
&cpu_buf->sample_lost_locked);
oprofilefs_create_ro_ulong(sb, cpudir, "sample_lost_overflow", oprofilefs_create_ro_ulong(sb, cpudir, "sample_lost_overflow",
&cpu_buf->sample_lost_overflow); &cpu_buf->sample_lost_overflow);
oprofilefs_create_ro_ulong(sb, cpudir, "sample_lost_task_exit", oprofilefs_create_ro_ulong(sb, cpudir, "sample_lost_task_exit",
......
...@@ -125,9 +125,7 @@ int register_filesystem(struct file_system_type * fs) ...@@ -125,9 +125,7 @@ int register_filesystem(struct file_system_type * fs)
if (!res) { if (!res) {
/* we implicitly possess reference to @fs during registration, /* we implicitly possess reference to @fs during registration,
* so it cannot be unregister from under us. */ * so it cannot be unregister from under us. */
if (register_fs_subsys(fs)) register_fs_subsys(fs);
printk(KERN_WARNING "Failed to register '%s' in sysfs\n",
fs->name);
} }
return res; return res;
} }
......
...@@ -33,7 +33,7 @@ static int sysfs_fill_super(struct super_block *sb, void *data, int silent) ...@@ -33,7 +33,7 @@ static int sysfs_fill_super(struct super_block *sb, void *data, int silent)
sb->s_op = &sysfs_ops; sb->s_op = &sysfs_ops;
sysfs_sb = sb; sysfs_sb = sb;
inode = sysfs_new_inode(S_IFDIR | S_IRUGO | S_IWUSR); inode = sysfs_new_inode(S_IFDIR | S_IRWXU | S_IRUGO | S_IXUGO);
if (inode) { if (inode) {
inode->i_op = &simple_dir_inode_operations; inode->i_op = &simple_dir_inode_operations;
inode->i_fop = &simple_dir_operations; inode->i_fop = &simple_dir_operations;
......
...@@ -2,15 +2,15 @@ ...@@ -2,15 +2,15 @@
#define _LINUX_PROFILE_H #define _LINUX_PROFILE_H
#ifdef __KERNEL__ #ifdef __KERNEL__
#include <linux/kernel.h> #include <linux/kernel.h>
#include <linux/config.h> #include <linux/config.h>
#include <linux/init.h> #include <linux/init.h>
#include <asm/errno.h> #include <asm/errno.h>
/* parse command line */ /* parse command line */
int __init profile_setup(char * str); int __init profile_setup(char * str);
/* init basic kernel profiler */ /* init basic kernel profiler */
void __init profile_init(void); void __init profile_init(void);
...@@ -27,14 +27,14 @@ enum profile_type { ...@@ -27,14 +27,14 @@ enum profile_type {
}; };
#ifdef CONFIG_PROFILING #ifdef CONFIG_PROFILING
struct notifier_block; struct notifier_block;
struct task_struct; struct task_struct;
struct mm_struct; struct mm_struct;
/* task is in do_exit() */ /* task is in do_exit() */
void profile_exit_task(struct task_struct * task); void profile_exit_task(struct task_struct * task);
/* change of vma mappings */ /* change of vma mappings */
void profile_exec_unmap(struct mm_struct * mm); void profile_exec_unmap(struct mm_struct * mm);
...@@ -44,10 +44,10 @@ void profile_exit_mmap(struct mm_struct * mm); ...@@ -44,10 +44,10 @@ void profile_exit_mmap(struct mm_struct * mm);
int profile_event_register(enum profile_type, struct notifier_block * n); int profile_event_register(enum profile_type, struct notifier_block * n);
int profile_event_unregister(enum profile_type, struct notifier_block * n); int profile_event_unregister(enum profile_type, struct notifier_block * n);
int register_profile_notifier(struct notifier_block * nb); int register_profile_notifier(struct notifier_block * nb);
int unregister_profile_notifier(struct notifier_block * nb); int unregister_profile_notifier(struct notifier_block * nb);
/* profiling hook activated on each timer interrupt */ /* profiling hook activated on each timer interrupt */
void profile_hook(struct pt_regs * regs); void profile_hook(struct pt_regs * regs);
...@@ -57,12 +57,12 @@ static inline int profile_event_register(enum profile_type t, struct notifier_bl ...@@ -57,12 +57,12 @@ static inline int profile_event_register(enum profile_type t, struct notifier_bl
{ {
return -ENOSYS; return -ENOSYS;
} }
static inline int profile_event_unregister(enum profile_type t, struct notifier_block * n) static inline int profile_event_unregister(enum profile_type t, struct notifier_block * n)
{ {
return -ENOSYS; return -ENOSYS;
} }
#define profile_exit_task(a) do { } while (0) #define profile_exit_task(a) do { } while (0)
#define profile_exec_unmap(a) do { } while (0) #define profile_exec_unmap(a) do { } while (0)
#define profile_exit_mmap(a) do { } while (0) #define profile_exit_mmap(a) do { } while (0)
...@@ -80,7 +80,7 @@ static inline int unregister_profile_notifier(struct notifier_block * nb) ...@@ -80,7 +80,7 @@ static inline int unregister_profile_notifier(struct notifier_block * nb)
#define profile_hook(regs) do { } while (0) #define profile_hook(regs) do { } while (0)
#endif /* CONFIG_PROFILING */ #endif /* CONFIG_PROFILING */
#endif /* __KERNEL__ */ #endif /* __KERNEL__ */
#endif /* _LINUX_PROFILE_H */ #endif /* _LINUX_PROFILE_H */
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment