Commit f0054bb1 authored by Tejun Heo's avatar Tejun Heo Committed by Jens Axboe

writeback: move backing_dev_info->wb_lock and ->worklist into bdi_writeback

Currently, a bdi (backing_dev_info) embeds single wb (bdi_writeback)
and the role of the separation is unclear.  For cgroup support for
writeback IOs, a bdi will be updated to host multiple wb's where each
wb serves writeback IOs of a different cgroup on the bdi.  To achieve
that, a wb should carry all states necessary for servicing writeback
IOs for a cgroup independently.

This patch moves bdi->wb_lock and ->worklist into wb.

* The lock protects bdi->worklist and bdi->wb.dwork scheduling.  While
  moving, rename it to wb->work_lock as wb->wb_lock is confusing.
  Also, move wb->dwork downwards so that it's colocated with the new
  ->work_lock and ->work_list fields.

* bdi_writeback_workfn()		-> wb_workfn()
  bdi_wakeup_thread_delayed(bdi)	-> wb_wakeup_delayed(wb)
  bdi_wakeup_thread(bdi)		-> wb_wakeup(wb)
  bdi_queue_work(bdi, ...)		-> wb_queue_work(wb, ...)
  __bdi_start_writeback(bdi, ...)	-> __wb_start_writeback(wb, ...)
  get_next_work_item(bdi)		-> get_next_work_item(wb)

* bdi_wb_shutdown() is renamed to wb_shutdown() and now takes @wb.
  The function contained parts which belong to the containing bdi
  rather than the wb itself - testing cap_writeback_dirty and
  bdi_remove_from_list() invocation.  Those are moved to
  bdi_unregister().

* bdi_wb_{init|exit}() are renamed to wb_{init|exit}().
  Initializations of the moved bdi->wb_lock and ->work_list are
  relocated from bdi_init() to wb_init().

* As there's still only one bdi_writeback per backing_dev_info, all
  uses of bdi->state are mechanically replaced with bdi->wb.state
  introducing no behavior changes.
Signed-off-by: default avatarTejun Heo <tj@kernel.org>
Reviewed-by: default avatarJan Kara <jack@suse.cz>
Cc: Jens Axboe <axboe@kernel.dk>
Cc: Wu Fengguang <fengguang.wu@intel.com>
Signed-off-by: default avatarJens Axboe <axboe@fb.com>
parent de1fff37
...@@ -109,34 +109,33 @@ static inline struct inode *wb_inode(struct list_head *head) ...@@ -109,34 +109,33 @@ static inline struct inode *wb_inode(struct list_head *head)
EXPORT_TRACEPOINT_SYMBOL_GPL(wbc_writepage); EXPORT_TRACEPOINT_SYMBOL_GPL(wbc_writepage);
static void bdi_wakeup_thread(struct backing_dev_info *bdi) static void wb_wakeup(struct bdi_writeback *wb)
{ {
spin_lock_bh(&bdi->wb_lock); spin_lock_bh(&wb->work_lock);
if (test_bit(WB_registered, &bdi->wb.state)) if (test_bit(WB_registered, &wb->state))
mod_delayed_work(bdi_wq, &bdi->wb.dwork, 0); mod_delayed_work(bdi_wq, &wb->dwork, 0);
spin_unlock_bh(&bdi->wb_lock); spin_unlock_bh(&wb->work_lock);
} }
static void bdi_queue_work(struct backing_dev_info *bdi, static void wb_queue_work(struct bdi_writeback *wb,
struct wb_writeback_work *work) struct wb_writeback_work *work)
{ {
trace_writeback_queue(bdi, work); trace_writeback_queue(wb->bdi, work);
spin_lock_bh(&bdi->wb_lock); spin_lock_bh(&wb->work_lock);
if (!test_bit(WB_registered, &bdi->wb.state)) { if (!test_bit(WB_registered, &wb->state)) {
if (work->done) if (work->done)
complete(work->done); complete(work->done);
goto out_unlock; goto out_unlock;
} }
list_add_tail(&work->list, &bdi->work_list); list_add_tail(&work->list, &wb->work_list);
mod_delayed_work(bdi_wq, &bdi->wb.dwork, 0); mod_delayed_work(bdi_wq, &wb->dwork, 0);
out_unlock: out_unlock:
spin_unlock_bh(&bdi->wb_lock); spin_unlock_bh(&wb->work_lock);
} }
static void static void __wb_start_writeback(struct bdi_writeback *wb, long nr_pages,
__bdi_start_writeback(struct backing_dev_info *bdi, long nr_pages, bool range_cyclic, enum wb_reason reason)
bool range_cyclic, enum wb_reason reason)
{ {
struct wb_writeback_work *work; struct wb_writeback_work *work;
...@@ -146,8 +145,8 @@ __bdi_start_writeback(struct backing_dev_info *bdi, long nr_pages, ...@@ -146,8 +145,8 @@ __bdi_start_writeback(struct backing_dev_info *bdi, long nr_pages,
*/ */
work = kzalloc(sizeof(*work), GFP_ATOMIC); work = kzalloc(sizeof(*work), GFP_ATOMIC);
if (!work) { if (!work) {
trace_writeback_nowork(bdi); trace_writeback_nowork(wb->bdi);
bdi_wakeup_thread(bdi); wb_wakeup(wb);
return; return;
} }
...@@ -156,7 +155,7 @@ __bdi_start_writeback(struct backing_dev_info *bdi, long nr_pages, ...@@ -156,7 +155,7 @@ __bdi_start_writeback(struct backing_dev_info *bdi, long nr_pages,
work->range_cyclic = range_cyclic; work->range_cyclic = range_cyclic;
work->reason = reason; work->reason = reason;
bdi_queue_work(bdi, work); wb_queue_work(wb, work);
} }
/** /**
...@@ -174,7 +173,7 @@ __bdi_start_writeback(struct backing_dev_info *bdi, long nr_pages, ...@@ -174,7 +173,7 @@ __bdi_start_writeback(struct backing_dev_info *bdi, long nr_pages,
void bdi_start_writeback(struct backing_dev_info *bdi, long nr_pages, void bdi_start_writeback(struct backing_dev_info *bdi, long nr_pages,
enum wb_reason reason) enum wb_reason reason)
{ {
__bdi_start_writeback(bdi, nr_pages, true, reason); __wb_start_writeback(&bdi->wb, nr_pages, true, reason);
} }
/** /**
...@@ -194,7 +193,7 @@ void bdi_start_background_writeback(struct backing_dev_info *bdi) ...@@ -194,7 +193,7 @@ void bdi_start_background_writeback(struct backing_dev_info *bdi)
* writeback as soon as there is no other work to do. * writeback as soon as there is no other work to do.
*/ */
trace_writeback_wake_background(bdi); trace_writeback_wake_background(bdi);
bdi_wakeup_thread(bdi); wb_wakeup(&bdi->wb);
} }
/* /*
...@@ -898,7 +897,7 @@ static long wb_writeback(struct bdi_writeback *wb, ...@@ -898,7 +897,7 @@ static long wb_writeback(struct bdi_writeback *wb,
* after the other works are all done. * after the other works are all done.
*/ */
if ((work->for_background || work->for_kupdate) && if ((work->for_background || work->for_kupdate) &&
!list_empty(&wb->bdi->work_list)) !list_empty(&wb->work_list))
break; break;
/* /*
...@@ -969,18 +968,17 @@ static long wb_writeback(struct bdi_writeback *wb, ...@@ -969,18 +968,17 @@ static long wb_writeback(struct bdi_writeback *wb,
/* /*
* Return the next wb_writeback_work struct that hasn't been processed yet. * Return the next wb_writeback_work struct that hasn't been processed yet.
*/ */
static struct wb_writeback_work * static struct wb_writeback_work *get_next_work_item(struct bdi_writeback *wb)
get_next_work_item(struct backing_dev_info *bdi)
{ {
struct wb_writeback_work *work = NULL; struct wb_writeback_work *work = NULL;
spin_lock_bh(&bdi->wb_lock); spin_lock_bh(&wb->work_lock);
if (!list_empty(&bdi->work_list)) { if (!list_empty(&wb->work_list)) {
work = list_entry(bdi->work_list.next, work = list_entry(wb->work_list.next,
struct wb_writeback_work, list); struct wb_writeback_work, list);
list_del_init(&work->list); list_del_init(&work->list);
} }
spin_unlock_bh(&bdi->wb_lock); spin_unlock_bh(&wb->work_lock);
return work; return work;
} }
...@@ -1052,14 +1050,13 @@ static long wb_check_old_data_flush(struct bdi_writeback *wb) ...@@ -1052,14 +1050,13 @@ static long wb_check_old_data_flush(struct bdi_writeback *wb)
*/ */
static long wb_do_writeback(struct bdi_writeback *wb) static long wb_do_writeback(struct bdi_writeback *wb)
{ {
struct backing_dev_info *bdi = wb->bdi;
struct wb_writeback_work *work; struct wb_writeback_work *work;
long wrote = 0; long wrote = 0;
set_bit(WB_writeback_running, &wb->state); set_bit(WB_writeback_running, &wb->state);
while ((work = get_next_work_item(bdi)) != NULL) { while ((work = get_next_work_item(wb)) != NULL) {
trace_writeback_exec(bdi, work); trace_writeback_exec(wb->bdi, work);
wrote += wb_writeback(wb, work); wrote += wb_writeback(wb, work);
...@@ -1087,43 +1084,42 @@ static long wb_do_writeback(struct bdi_writeback *wb) ...@@ -1087,43 +1084,42 @@ static long wb_do_writeback(struct bdi_writeback *wb)
* Handle writeback of dirty data for the device backed by this bdi. Also * Handle writeback of dirty data for the device backed by this bdi. Also
* reschedules periodically and does kupdated style flushing. * reschedules periodically and does kupdated style flushing.
*/ */
void bdi_writeback_workfn(struct work_struct *work) void wb_workfn(struct work_struct *work)
{ {
struct bdi_writeback *wb = container_of(to_delayed_work(work), struct bdi_writeback *wb = container_of(to_delayed_work(work),
struct bdi_writeback, dwork); struct bdi_writeback, dwork);
struct backing_dev_info *bdi = wb->bdi;
long pages_written; long pages_written;
set_worker_desc("flush-%s", dev_name(bdi->dev)); set_worker_desc("flush-%s", dev_name(wb->bdi->dev));
current->flags |= PF_SWAPWRITE; current->flags |= PF_SWAPWRITE;
if (likely(!current_is_workqueue_rescuer() || if (likely(!current_is_workqueue_rescuer() ||
!test_bit(WB_registered, &wb->state))) { !test_bit(WB_registered, &wb->state))) {
/* /*
* The normal path. Keep writing back @bdi until its * The normal path. Keep writing back @wb until its
* work_list is empty. Note that this path is also taken * work_list is empty. Note that this path is also taken
* if @bdi is shutting down even when we're running off the * if @wb is shutting down even when we're running off the
* rescuer as work_list needs to be drained. * rescuer as work_list needs to be drained.
*/ */
do { do {
pages_written = wb_do_writeback(wb); pages_written = wb_do_writeback(wb);
trace_writeback_pages_written(pages_written); trace_writeback_pages_written(pages_written);
} while (!list_empty(&bdi->work_list)); } while (!list_empty(&wb->work_list));
} else { } else {
/* /*
* bdi_wq can't get enough workers and we're running off * bdi_wq can't get enough workers and we're running off
* the emergency worker. Don't hog it. Hopefully, 1024 is * the emergency worker. Don't hog it. Hopefully, 1024 is
* enough for efficient IO. * enough for efficient IO.
*/ */
pages_written = writeback_inodes_wb(&bdi->wb, 1024, pages_written = writeback_inodes_wb(wb, 1024,
WB_REASON_FORKER_THREAD); WB_REASON_FORKER_THREAD);
trace_writeback_pages_written(pages_written); trace_writeback_pages_written(pages_written);
} }
if (!list_empty(&bdi->work_list)) if (!list_empty(&wb->work_list))
mod_delayed_work(bdi_wq, &wb->dwork, 0); mod_delayed_work(bdi_wq, &wb->dwork, 0);
else if (wb_has_dirty_io(wb) && dirty_writeback_interval) else if (wb_has_dirty_io(wb) && dirty_writeback_interval)
bdi_wakeup_thread_delayed(bdi); wb_wakeup_delayed(wb);
current->flags &= ~PF_SWAPWRITE; current->flags &= ~PF_SWAPWRITE;
} }
...@@ -1143,7 +1139,7 @@ void wakeup_flusher_threads(long nr_pages, enum wb_reason reason) ...@@ -1143,7 +1139,7 @@ void wakeup_flusher_threads(long nr_pages, enum wb_reason reason)
list_for_each_entry_rcu(bdi, &bdi_list, bdi_list) { list_for_each_entry_rcu(bdi, &bdi_list, bdi_list) {
if (!bdi_has_dirty_io(bdi)) if (!bdi_has_dirty_io(bdi))
continue; continue;
__bdi_start_writeback(bdi, nr_pages, false, reason); __wb_start_writeback(&bdi->wb, nr_pages, false, reason);
} }
rcu_read_unlock(); rcu_read_unlock();
} }
...@@ -1174,7 +1170,7 @@ static void wakeup_dirtytime_writeback(struct work_struct *w) ...@@ -1174,7 +1170,7 @@ static void wakeup_dirtytime_writeback(struct work_struct *w)
list_for_each_entry_rcu(bdi, &bdi_list, bdi_list) { list_for_each_entry_rcu(bdi, &bdi_list, bdi_list) {
if (list_empty(&bdi->wb.b_dirty_time)) if (list_empty(&bdi->wb.b_dirty_time))
continue; continue;
bdi_wakeup_thread(bdi); wb_wakeup(&bdi->wb);
} }
rcu_read_unlock(); rcu_read_unlock();
schedule_delayed_work(&dirtytime_work, dirtytime_expire_interval * HZ); schedule_delayed_work(&dirtytime_work, dirtytime_expire_interval * HZ);
...@@ -1347,7 +1343,7 @@ void __mark_inode_dirty(struct inode *inode, int flags) ...@@ -1347,7 +1343,7 @@ void __mark_inode_dirty(struct inode *inode, int flags)
trace_writeback_dirty_inode_enqueue(inode); trace_writeback_dirty_inode_enqueue(inode);
if (wakeup_bdi) if (wakeup_bdi)
bdi_wakeup_thread_delayed(bdi); wb_wakeup_delayed(&bdi->wb);
return; return;
} }
} }
...@@ -1437,7 +1433,7 @@ void writeback_inodes_sb_nr(struct super_block *sb, ...@@ -1437,7 +1433,7 @@ void writeback_inodes_sb_nr(struct super_block *sb,
if (sb->s_bdi == &noop_backing_dev_info) if (sb->s_bdi == &noop_backing_dev_info)
return; return;
WARN_ON(!rwsem_is_locked(&sb->s_umount)); WARN_ON(!rwsem_is_locked(&sb->s_umount));
bdi_queue_work(sb->s_bdi, &work); wb_queue_work(&sb->s_bdi->wb, &work);
wait_for_completion(&done); wait_for_completion(&done);
} }
EXPORT_SYMBOL(writeback_inodes_sb_nr); EXPORT_SYMBOL(writeback_inodes_sb_nr);
...@@ -1521,7 +1517,7 @@ void sync_inodes_sb(struct super_block *sb) ...@@ -1521,7 +1517,7 @@ void sync_inodes_sb(struct super_block *sb)
return; return;
WARN_ON(!rwsem_is_locked(&sb->s_umount)); WARN_ON(!rwsem_is_locked(&sb->s_umount));
bdi_queue_work(sb->s_bdi, &work); wb_queue_work(&sb->s_bdi->wb, &work);
wait_for_completion(&done); wait_for_completion(&done);
wait_sb_inodes(sb); wait_sb_inodes(sb);
......
...@@ -52,7 +52,6 @@ struct bdi_writeback { ...@@ -52,7 +52,6 @@ struct bdi_writeback {
unsigned long state; /* Always use atomic bitops on this */ unsigned long state; /* Always use atomic bitops on this */
unsigned long last_old_flush; /* last old data flush */ unsigned long last_old_flush; /* last old data flush */
struct delayed_work dwork; /* work item used for writeback */
struct list_head b_dirty; /* dirty inodes */ struct list_head b_dirty; /* dirty inodes */
struct list_head b_io; /* parked for writeback */ struct list_head b_io; /* parked for writeback */
struct list_head b_more_io; /* parked for more writeback */ struct list_head b_more_io; /* parked for more writeback */
...@@ -78,6 +77,10 @@ struct bdi_writeback { ...@@ -78,6 +77,10 @@ struct bdi_writeback {
struct fprop_local_percpu completions; struct fprop_local_percpu completions;
int dirty_exceeded; int dirty_exceeded;
spinlock_t work_lock; /* protects work_list & dwork scheduling */
struct list_head work_list;
struct delayed_work dwork; /* work item used for writeback */
}; };
struct backing_dev_info { struct backing_dev_info {
...@@ -93,9 +96,6 @@ struct backing_dev_info { ...@@ -93,9 +96,6 @@ struct backing_dev_info {
unsigned int max_ratio, max_prop_frac; unsigned int max_ratio, max_prop_frac;
struct bdi_writeback wb; /* default writeback info for this bdi */ struct bdi_writeback wb; /* default writeback info for this bdi */
spinlock_t wb_lock; /* protects work_list & wb.dwork scheduling */
struct list_head work_list;
struct device *dev; struct device *dev;
...@@ -121,9 +121,9 @@ int __must_check bdi_setup_and_register(struct backing_dev_info *, char *); ...@@ -121,9 +121,9 @@ int __must_check bdi_setup_and_register(struct backing_dev_info *, char *);
void bdi_start_writeback(struct backing_dev_info *bdi, long nr_pages, void bdi_start_writeback(struct backing_dev_info *bdi, long nr_pages,
enum wb_reason reason); enum wb_reason reason);
void bdi_start_background_writeback(struct backing_dev_info *bdi); void bdi_start_background_writeback(struct backing_dev_info *bdi);
void bdi_writeback_workfn(struct work_struct *work); void wb_workfn(struct work_struct *work);
int bdi_has_dirty_io(struct backing_dev_info *bdi); int bdi_has_dirty_io(struct backing_dev_info *bdi);
void bdi_wakeup_thread_delayed(struct backing_dev_info *bdi); void wb_wakeup_delayed(struct bdi_writeback *wb);
extern spinlock_t bdi_lock; extern spinlock_t bdi_lock;
extern struct list_head bdi_list; extern struct list_head bdi_list;
......
...@@ -261,7 +261,7 @@ int bdi_has_dirty_io(struct backing_dev_info *bdi) ...@@ -261,7 +261,7 @@ int bdi_has_dirty_io(struct backing_dev_info *bdi)
} }
/* /*
* This function is used when the first inode for this bdi is marked dirty. It * This function is used when the first inode for this wb is marked dirty. It
* wakes-up the corresponding bdi thread which should then take care of the * wakes-up the corresponding bdi thread which should then take care of the
* periodic background write-out of dirty inodes. Since the write-out would * periodic background write-out of dirty inodes. Since the write-out would
* starts only 'dirty_writeback_interval' centisecs from now anyway, we just * starts only 'dirty_writeback_interval' centisecs from now anyway, we just
...@@ -274,15 +274,15 @@ int bdi_has_dirty_io(struct backing_dev_info *bdi) ...@@ -274,15 +274,15 @@ int bdi_has_dirty_io(struct backing_dev_info *bdi)
* We have to be careful not to postpone flush work if it is scheduled for * We have to be careful not to postpone flush work if it is scheduled for
* earlier. Thus we use queue_delayed_work(). * earlier. Thus we use queue_delayed_work().
*/ */
void bdi_wakeup_thread_delayed(struct backing_dev_info *bdi) void wb_wakeup_delayed(struct bdi_writeback *wb)
{ {
unsigned long timeout; unsigned long timeout;
timeout = msecs_to_jiffies(dirty_writeback_interval * 10); timeout = msecs_to_jiffies(dirty_writeback_interval * 10);
spin_lock_bh(&bdi->wb_lock); spin_lock_bh(&wb->work_lock);
if (test_bit(WB_registered, &bdi->wb.state)) if (test_bit(WB_registered, &wb->state))
queue_delayed_work(bdi_wq, &bdi->wb.dwork, timeout); queue_delayed_work(bdi_wq, &wb->dwork, timeout);
spin_unlock_bh(&bdi->wb_lock); spin_unlock_bh(&wb->work_lock);
} }
/* /*
...@@ -335,28 +335,24 @@ EXPORT_SYMBOL(bdi_register_dev); ...@@ -335,28 +335,24 @@ EXPORT_SYMBOL(bdi_register_dev);
/* /*
* Remove bdi from the global list and shutdown any threads we have running * Remove bdi from the global list and shutdown any threads we have running
*/ */
static void bdi_wb_shutdown(struct backing_dev_info *bdi) static void wb_shutdown(struct bdi_writeback *wb)
{ {
/* Make sure nobody queues further work */ /* Make sure nobody queues further work */
spin_lock_bh(&bdi->wb_lock); spin_lock_bh(&wb->work_lock);
if (!test_and_clear_bit(WB_registered, &bdi->wb.state)) { if (!test_and_clear_bit(WB_registered, &wb->state)) {
spin_unlock_bh(&bdi->wb_lock); spin_unlock_bh(&wb->work_lock);
return; return;
} }
spin_unlock_bh(&bdi->wb_lock); spin_unlock_bh(&wb->work_lock);
/* /*
* Make sure nobody finds us on the bdi_list anymore * Drain work list and shutdown the delayed_work. !WB_registered
* tells wb_workfn() that @wb is dying and its work_list needs to
* be drained no matter what.
*/ */
bdi_remove_from_list(bdi); mod_delayed_work(bdi_wq, &wb->dwork, 0);
flush_delayed_work(&wb->dwork);
/* WARN_ON(!list_empty(&wb->work_list));
* Drain work list and shutdown the delayed_work. At this point,
* @bdi->bdi_list is empty telling bdi_Writeback_workfn() that @bdi
* is dying and its work_list needs to be drained no matter what.
*/
mod_delayed_work(bdi_wq, &bdi->wb.dwork, 0);
flush_delayed_work(&bdi->wb.dwork);
} }
/* /*
...@@ -381,7 +377,7 @@ EXPORT_SYMBOL(bdi_unregister); ...@@ -381,7 +377,7 @@ EXPORT_SYMBOL(bdi_unregister);
*/ */
#define INIT_BW (100 << (20 - PAGE_SHIFT)) #define INIT_BW (100 << (20 - PAGE_SHIFT))
static int bdi_wb_init(struct bdi_writeback *wb, struct backing_dev_info *bdi) static int wb_init(struct bdi_writeback *wb, struct backing_dev_info *bdi)
{ {
int i, err; int i, err;
...@@ -394,7 +390,6 @@ static int bdi_wb_init(struct bdi_writeback *wb, struct backing_dev_info *bdi) ...@@ -394,7 +390,6 @@ static int bdi_wb_init(struct bdi_writeback *wb, struct backing_dev_info *bdi)
INIT_LIST_HEAD(&wb->b_more_io); INIT_LIST_HEAD(&wb->b_more_io);
INIT_LIST_HEAD(&wb->b_dirty_time); INIT_LIST_HEAD(&wb->b_dirty_time);
spin_lock_init(&wb->list_lock); spin_lock_init(&wb->list_lock);
INIT_DELAYED_WORK(&wb->dwork, bdi_writeback_workfn);
wb->bw_time_stamp = jiffies; wb->bw_time_stamp = jiffies;
wb->balanced_dirty_ratelimit = INIT_BW; wb->balanced_dirty_ratelimit = INIT_BW;
...@@ -402,6 +397,10 @@ static int bdi_wb_init(struct bdi_writeback *wb, struct backing_dev_info *bdi) ...@@ -402,6 +397,10 @@ static int bdi_wb_init(struct bdi_writeback *wb, struct backing_dev_info *bdi)
wb->write_bandwidth = INIT_BW; wb->write_bandwidth = INIT_BW;
wb->avg_write_bandwidth = INIT_BW; wb->avg_write_bandwidth = INIT_BW;
spin_lock_init(&wb->work_lock);
INIT_LIST_HEAD(&wb->work_list);
INIT_DELAYED_WORK(&wb->dwork, wb_workfn);
err = fprop_local_init_percpu(&wb->completions, GFP_KERNEL); err = fprop_local_init_percpu(&wb->completions, GFP_KERNEL);
if (err) if (err)
return err; return err;
...@@ -419,7 +418,7 @@ static int bdi_wb_init(struct bdi_writeback *wb, struct backing_dev_info *bdi) ...@@ -419,7 +418,7 @@ static int bdi_wb_init(struct bdi_writeback *wb, struct backing_dev_info *bdi)
return 0; return 0;
} }
static void bdi_wb_exit(struct bdi_writeback *wb) static void wb_exit(struct bdi_writeback *wb)
{ {
int i; int i;
...@@ -440,11 +439,9 @@ int bdi_init(struct backing_dev_info *bdi) ...@@ -440,11 +439,9 @@ int bdi_init(struct backing_dev_info *bdi)
bdi->min_ratio = 0; bdi->min_ratio = 0;
bdi->max_ratio = 100; bdi->max_ratio = 100;
bdi->max_prop_frac = FPROP_FRAC_BASE; bdi->max_prop_frac = FPROP_FRAC_BASE;
spin_lock_init(&bdi->wb_lock);
INIT_LIST_HEAD(&bdi->bdi_list); INIT_LIST_HEAD(&bdi->bdi_list);
INIT_LIST_HEAD(&bdi->work_list);
err = bdi_wb_init(&bdi->wb, bdi); err = wb_init(&bdi->wb, bdi);
if (err) if (err)
return err; return err;
...@@ -454,9 +451,9 @@ EXPORT_SYMBOL(bdi_init); ...@@ -454,9 +451,9 @@ EXPORT_SYMBOL(bdi_init);
void bdi_destroy(struct backing_dev_info *bdi) void bdi_destroy(struct backing_dev_info *bdi)
{ {
bdi_wb_shutdown(bdi); /* make sure nobody finds us on the bdi_list anymore */
bdi_remove_from_list(bdi);
WARN_ON(!list_empty(&bdi->work_list)); wb_shutdown(&bdi->wb);
if (bdi->dev) { if (bdi->dev) {
bdi_debug_unregister(bdi); bdi_debug_unregister(bdi);
...@@ -464,7 +461,7 @@ void bdi_destroy(struct backing_dev_info *bdi) ...@@ -464,7 +461,7 @@ void bdi_destroy(struct backing_dev_info *bdi)
bdi->dev = NULL; bdi->dev = NULL;
} }
bdi_wb_exit(&bdi->wb); wb_exit(&bdi->wb);
} }
EXPORT_SYMBOL(bdi_destroy); EXPORT_SYMBOL(bdi_destroy);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment