Commit 18c16b33 authored by Linus Torvalds's avatar Linus Torvalds

Merge http://linux-isdn.bkbits.net/linux-2.5.make-drivers

into home.transmeta.com:/home/torvalds/v2.5/linux
parents bd7c5358 f1c83c72
...@@ -395,6 +395,11 @@ BUILD_INTERRUPT(call_function_interrupt,CALL_FUNCTION_VECTOR) ...@@ -395,6 +395,11 @@ BUILD_INTERRUPT(call_function_interrupt,CALL_FUNCTION_VECTOR)
BUILD_INTERRUPT(apic_timer_interrupt,LOCAL_TIMER_VECTOR) BUILD_INTERRUPT(apic_timer_interrupt,LOCAL_TIMER_VECTOR)
BUILD_INTERRUPT(error_interrupt,ERROR_APIC_VECTOR) BUILD_INTERRUPT(error_interrupt,ERROR_APIC_VECTOR)
BUILD_INTERRUPT(spurious_interrupt,SPURIOUS_APIC_VECTOR) BUILD_INTERRUPT(spurious_interrupt,SPURIOUS_APIC_VECTOR)
#ifdef CONFIG_X86_MCE_P4THERMAL
BUILD_INTERRUPT(thermal_interrupt,THERMAL_APIC_VECTOR)
#endif
#endif #endif
ENTRY(divide_error) ENTRY(divide_error)
......
...@@ -407,7 +407,7 @@ void __init init_IRQ(void) ...@@ -407,7 +407,7 @@ void __init init_IRQ(void)
/* thermal monitor LVT interrupt */ /* thermal monitor LVT interrupt */
#ifdef CONFIG_X86_MCE_P4THERMAL #ifdef CONFIG_X86_MCE_P4THERMAL
set_intr_gate(THERMAL_APIC_VECTOR, smp_thermal_interrupt); set_intr_gate(THERMAL_APIC_VECTOR, thermal_interrupt);
#endif #endif
#endif #endif
......
...@@ -50,13 +50,7 @@ extern int mac_floppy_init(void); ...@@ -50,13 +50,7 @@ extern int mac_floppy_init(void);
static kmem_cache_t *request_cachep; static kmem_cache_t *request_cachep;
static struct list_head blk_plug_list; static struct list_head blk_plug_list;
static spinlock_t blk_plug_lock = SPIN_LOCK_UNLOCKED; static spinlock_t blk_plug_lock __cacheline_aligned_in_smp = SPIN_LOCK_UNLOCKED;
/*
* The "disk" task queue is used to start the actual requests
* after a plug
*/
DECLARE_TASK_QUEUE(tq_disk);
/* blk_dev_struct is: /* blk_dev_struct is:
* request_queue * request_queue
...@@ -794,12 +788,11 @@ void blk_plug_device(request_queue_t *q) ...@@ -794,12 +788,11 @@ void blk_plug_device(request_queue_t *q)
if (!elv_queue_empty(q)) if (!elv_queue_empty(q))
return; return;
if (test_and_set_bit(QUEUE_FLAG_PLUGGED, &q->queue_flags)) if (!test_and_set_bit(QUEUE_FLAG_PLUGGED, &q->queue_flags)) {
return; spin_lock(&blk_plug_lock);
list_add_tail(&q->plug_list, &blk_plug_list);
spin_lock(&blk_plug_lock); spin_unlock(&blk_plug_lock);
list_add_tail(&q->plug.list, &blk_plug_list); }
spin_unlock(&blk_plug_lock);
} }
/* /*
...@@ -813,10 +806,8 @@ static inline void __generic_unplug_device(request_queue_t *q) ...@@ -813,10 +806,8 @@ static inline void __generic_unplug_device(request_queue_t *q)
if (!__test_and_clear_bit(QUEUE_FLAG_PLUGGED, &q->queue_flags)) if (!__test_and_clear_bit(QUEUE_FLAG_PLUGGED, &q->queue_flags))
return; return;
if (test_bit(QUEUE_FLAG_STOPPED, &q->queue_flags)) { if (test_bit(QUEUE_FLAG_STOPPED, &q->queue_flags))
printk("queue was stopped\n");
return; return;
}
/* /*
* was plugged, fire request_fn if queue has stuff to do * was plugged, fire request_fn if queue has stuff to do
...@@ -834,22 +825,12 @@ static inline void __generic_unplug_device(request_queue_t *q) ...@@ -834,22 +825,12 @@ static inline void __generic_unplug_device(request_queue_t *q)
* the device have at them. If a queue is plugged, the I/O scheduler * the device have at them. If a queue is plugged, the I/O scheduler
* is still adding and merging requests on the queue. Once the queue * is still adding and merging requests on the queue. Once the queue
* gets unplugged (either by manually calling this function, or by * gets unplugged (either by manually calling this function, or by
* running the tq_disk task queue), the request_fn defined for the * calling blk_run_queues()), the request_fn defined for the
* queue is invoked and transfers started. * queue is invoked and transfers started.
**/ **/
void generic_unplug_device(void *data) void generic_unplug_device(void *data)
{ {
request_queue_t *q = data; request_queue_t *q = data;
tasklet_schedule(&q->plug.task);
}
/*
* the plug tasklet
*/
static void blk_task_run(unsigned long data)
{
request_queue_t *q = (request_queue_t *) data;
unsigned long flags; unsigned long flags;
spin_lock_irqsave(q->queue_lock, flags); spin_lock_irqsave(q->queue_lock, flags);
...@@ -858,23 +839,26 @@ static void blk_task_run(unsigned long data) ...@@ -858,23 +839,26 @@ static void blk_task_run(unsigned long data)
} }
/* /*
* clear top flag and schedule tasklet for execution * clear stop flag and run queue
*/ */
void blk_start_queue(request_queue_t *q) void blk_start_queue(request_queue_t *q)
{ {
if (test_and_clear_bit(QUEUE_FLAG_STOPPED, &q->queue_flags)) if (test_and_clear_bit(QUEUE_FLAG_STOPPED, &q->queue_flags)) {
tasklet_enable(&q->plug.task); unsigned long flags;
tasklet_schedule(&q->plug.task); spin_lock_irqsave(q->queue_lock, flags);
if (!elv_queue_empty(q))
q->request_fn(q);
spin_unlock_irqrestore(q->queue_lock, flags);
}
} }
/* /*
* set stop bit and disable any pending tasklet * set stop bit, queue won't be run until blk_start_queue() is called
*/ */
void blk_stop_queue(request_queue_t *q) void blk_stop_queue(request_queue_t *q)
{ {
if (!test_and_set_bit(QUEUE_FLAG_STOPPED, &q->queue_flags)) set_bit(QUEUE_FLAG_STOPPED, &q->queue_flags);
tasklet_disable(&q->plug.task);
} }
/* /*
...@@ -882,22 +866,44 @@ void blk_stop_queue(request_queue_t *q) ...@@ -882,22 +866,44 @@ void blk_stop_queue(request_queue_t *q)
*/ */
void blk_run_queues(void) void blk_run_queues(void)
{ {
struct list_head *tmp, *n; struct list_head *n, *tmp, local_plug_list;
unsigned long flags; unsigned long flags;
INIT_LIST_HEAD(&local_plug_list);
/* /*
* we could splice to the stack prior to running * this will happen fairly often
*/ */
spin_lock_irqsave(&blk_plug_lock, flags); spin_lock_irqsave(&blk_plug_lock, flags);
list_for_each_safe(tmp, n, &blk_plug_list) { if (list_empty(&blk_plug_list)) {
request_queue_t *q = list_entry(tmp, request_queue_t,plug.list); spin_unlock_irqrestore(&blk_plug_lock, flags);
return;
}
list_splice(&blk_plug_list, &local_plug_list);
INIT_LIST_HEAD(&blk_plug_list);
spin_unlock_irqrestore(&blk_plug_lock, flags);
/*
* local_plug_list is now a private copy we can traverse lockless
*/
list_for_each_safe(n, tmp, &local_plug_list) {
request_queue_t *q = list_entry(n, request_queue_t, plug_list);
if (!test_bit(QUEUE_FLAG_STOPPED, &q->queue_flags)) { if (!test_bit(QUEUE_FLAG_STOPPED, &q->queue_flags)) {
list_del(&q->plug.list); list_del(&q->plug_list);
tasklet_schedule(&q->plug.task); generic_unplug_device(q);
} }
} }
spin_unlock_irqrestore(&blk_plug_lock, flags);
/*
* add any remaining queue back to plug list
*/
if (!list_empty(&local_plug_list)) {
spin_lock_irqsave(&blk_plug_lock, flags);
list_splice(&local_plug_list, &blk_plug_list);
spin_unlock_irqrestore(&blk_plug_lock, flags);
}
} }
static int __blk_cleanup_queue(struct request_list *list) static int __blk_cleanup_queue(struct request_list *list)
...@@ -1050,8 +1056,7 @@ int blk_init_queue(request_queue_t *q, request_fn_proc *rfn, spinlock_t *lock) ...@@ -1050,8 +1056,7 @@ int blk_init_queue(request_queue_t *q, request_fn_proc *rfn, spinlock_t *lock)
blk_queue_max_hw_segments(q, MAX_HW_SEGMENTS); blk_queue_max_hw_segments(q, MAX_HW_SEGMENTS);
blk_queue_max_phys_segments(q, MAX_PHYS_SEGMENTS); blk_queue_max_phys_segments(q, MAX_PHYS_SEGMENTS);
INIT_LIST_HEAD(&q->plug.list); INIT_LIST_HEAD(&q->plug_list);
tasklet_init(&q->plug.task, blk_task_run, (unsigned long) q);
return 0; return 0;
} }
......
...@@ -872,7 +872,8 @@ int fat_fill_super(struct super_block *sb, void *data, int silent, ...@@ -872,7 +872,8 @@ int fat_fill_super(struct super_block *sb, void *data, int silent,
if (!silent) if (!silent)
printk("FAT: Using codepage %s\n", sbi->nls_disk->charset); printk("FAT: Using codepage %s\n", sbi->nls_disk->charset);
if (sbi->options.isvfat && !sbi->options.utf8) { /* FIXME: utf8 is using iocharset for upper/lower conversion */
if (sbi->options.isvfat) {
if (sbi->options.iocharset != NULL) { if (sbi->options.iocharset != NULL) {
sbi->nls_io = load_nls(sbi->options.iocharset); sbi->nls_io = load_nls(sbi->options.iocharset);
if (!sbi->nls_io) { if (!sbi->nls_io) {
......
...@@ -28,6 +28,7 @@ extern int irq_vector[NR_IRQS]; ...@@ -28,6 +28,7 @@ extern int irq_vector[NR_IRQS];
extern void (*interrupt[NR_IRQS])(void); extern void (*interrupt[NR_IRQS])(void);
extern asmlinkage void thermal_interrupt(void);
extern asmlinkage void smp_thermal_interrupt(struct pt_regs); extern asmlinkage void smp_thermal_interrupt(struct pt_regs);
#ifdef CONFIG_SMP #ifdef CONFIG_SMP
......
...@@ -8,7 +8,6 @@ ...@@ -8,7 +8,6 @@
#include <linux/list.h> #include <linux/list.h>
#include <linux/pagemap.h> #include <linux/pagemap.h>
#include <linux/backing-dev.h> #include <linux/backing-dev.h>
#include <linux/interrupt.h>
#include <asm/scatterlist.h> #include <asm/scatterlist.h>
...@@ -137,11 +136,6 @@ struct blk_queue_tag { ...@@ -137,11 +136,6 @@ struct blk_queue_tag {
int max_depth; int max_depth;
}; };
struct blk_plug {
struct list_head list;
struct tasklet_struct task;
};
/* /*
* Default nr free requests per queue, ll_rw_blk will scale it down * Default nr free requests per queue, ll_rw_blk will scale it down
* according to available RAM at init time * according to available RAM at init time
...@@ -183,7 +177,7 @@ struct request_queue ...@@ -183,7 +177,7 @@ struct request_queue
unsigned long bounce_pfn; unsigned long bounce_pfn;
int bounce_gfp; int bounce_gfp;
struct blk_plug plug; struct list_head plug_list;
/* /*
* various queue flags, see QUEUE_* below * various queue flags, see QUEUE_* below
......
...@@ -13,7 +13,7 @@ ...@@ -13,7 +13,7 @@
* hybrid priority-list and round-robin design with * hybrid priority-list and round-robin design with
* an array-switch method of distributing timeslices * an array-switch method of distributing timeslices
* and per-CPU runqueues. Additional code by Davide * and per-CPU runqueues. Additional code by Davide
* Libenzi, Robert Love, and Rusty Russel. * Libenzi, Robert Love, and Rusty Russell.
*/ */
#include <linux/mm.h> #include <linux/mm.h>
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment