Commit b6e51316 authored by Jens Axboe's avatar Jens Axboe

writeback: separate starting of sync vs opportunistic writeback

bdi_start_writeback() is currently split into two paths, one for
WB_SYNC_NONE and one for WB_SYNC_ALL. Add bdi_sync_writeback()
for WB_SYNC_ALL writeback and let bdi_start_writeback() handle
only WB_SYNC_NONE.

Push down the writeback_control allocation and only accept the
parameters that make sense for each function. This cleans up
the API considerably.
Signed-off-by: default avatarJens Axboe <jens.axboe@oracle.com>
parent bcddc3f0
...@@ -74,14 +74,10 @@ static inline bool bdi_work_on_stack(struct bdi_work *work) ...@@ -74,14 +74,10 @@ static inline bool bdi_work_on_stack(struct bdi_work *work)
} }
static inline void bdi_work_init(struct bdi_work *work, static inline void bdi_work_init(struct bdi_work *work,
struct writeback_control *wbc) struct wb_writeback_args *args)
{ {
INIT_RCU_HEAD(&work->rcu_head); INIT_RCU_HEAD(&work->rcu_head);
work->args.sb = wbc->sb; work->args = *args;
work->args.nr_pages = wbc->nr_to_write;
work->args.sync_mode = wbc->sync_mode;
work->args.range_cyclic = wbc->range_cyclic;
work->args.for_kupdate = 0;
work->state = WS_USED; work->state = WS_USED;
} }
...@@ -194,7 +190,7 @@ static void bdi_wait_on_work_clear(struct bdi_work *work) ...@@ -194,7 +190,7 @@ static void bdi_wait_on_work_clear(struct bdi_work *work)
} }
static void bdi_alloc_queue_work(struct backing_dev_info *bdi, static void bdi_alloc_queue_work(struct backing_dev_info *bdi,
struct writeback_control *wbc) struct wb_writeback_args *args)
{ {
struct bdi_work *work; struct bdi_work *work;
...@@ -204,7 +200,7 @@ static void bdi_alloc_queue_work(struct backing_dev_info *bdi, ...@@ -204,7 +200,7 @@ static void bdi_alloc_queue_work(struct backing_dev_info *bdi,
*/ */
work = kmalloc(sizeof(*work), GFP_ATOMIC); work = kmalloc(sizeof(*work), GFP_ATOMIC);
if (work) { if (work) {
bdi_work_init(work, wbc); bdi_work_init(work, args);
bdi_queue_work(bdi, work); bdi_queue_work(bdi, work);
} else { } else {
struct bdi_writeback *wb = &bdi->wb; struct bdi_writeback *wb = &bdi->wb;
...@@ -214,24 +210,54 @@ static void bdi_alloc_queue_work(struct backing_dev_info *bdi, ...@@ -214,24 +210,54 @@ static void bdi_alloc_queue_work(struct backing_dev_info *bdi,
} }
} }
void bdi_start_writeback(struct writeback_control *wbc) /**
* bdi_sync_writeback - start and wait for writeback
* @bdi: the backing device to write from
* @sb: write inodes from this super_block
*
* Description:
* This does WB_SYNC_ALL data integrity writeback and waits for the
* IO to complete. Callers must hold the sb s_umount semaphore for
* reading, to avoid having the super disappear before we are done.
*/
static void bdi_sync_writeback(struct backing_dev_info *bdi,
struct super_block *sb)
{ {
/* struct wb_writeback_args args = {
* WB_SYNC_NONE is opportunistic writeback. If this allocation fails, .sb = sb,
* bdi_queue_work() will wake up the thread and flush old data. This .sync_mode = WB_SYNC_ALL,
* should ensure some amount of progress in freeing memory. .nr_pages = LONG_MAX,
*/ .range_cyclic = 0,
if (wbc->sync_mode != WB_SYNC_ALL) };
bdi_alloc_queue_work(wbc->bdi, wbc); struct bdi_work work;
else {
struct bdi_work work;
bdi_work_init(&work, wbc); bdi_work_init(&work, &args);
work.state |= WS_ONSTACK; work.state |= WS_ONSTACK;
bdi_queue_work(wbc->bdi, &work); bdi_queue_work(bdi, &work);
bdi_wait_on_work_clear(&work); bdi_wait_on_work_clear(&work);
} }
/**
* bdi_start_writeback - start writeback
* @bdi: the backing device to write from
* @nr_pages: the number of pages to write
*
* Description:
* This does WB_SYNC_NONE opportunistic writeback. The IO is only
* started when this function returns, we make no guarentees on
* completion. Caller need not hold sb s_umount semaphore.
*
*/
void bdi_start_writeback(struct backing_dev_info *bdi, long nr_pages)
{
struct wb_writeback_args args = {
.sync_mode = WB_SYNC_NONE,
.nr_pages = nr_pages,
.range_cyclic = 1,
};
bdi_alloc_queue_work(bdi, &args);
} }
/* /*
...@@ -863,23 +889,25 @@ int bdi_writeback_task(struct bdi_writeback *wb) ...@@ -863,23 +889,25 @@ int bdi_writeback_task(struct bdi_writeback *wb)
} }
/* /*
* Schedule writeback for all backing devices. Can only be used for * Schedule writeback for all backing devices. This does WB_SYNC_NONE
* WB_SYNC_NONE writeback, WB_SYNC_ALL should use bdi_start_writeback() * writeback, for integrity writeback see bdi_sync_writeback().
* and pass in the superblock.
*/ */
static void bdi_writeback_all(struct writeback_control *wbc) static void bdi_writeback_all(struct super_block *sb, long nr_pages)
{ {
struct wb_writeback_args args = {
.sb = sb,
.nr_pages = nr_pages,
.sync_mode = WB_SYNC_NONE,
};
struct backing_dev_info *bdi; struct backing_dev_info *bdi;
WARN_ON(wbc->sync_mode == WB_SYNC_ALL);
rcu_read_lock(); rcu_read_lock();
list_for_each_entry_rcu(bdi, &bdi_list, bdi_list) { list_for_each_entry_rcu(bdi, &bdi_list, bdi_list) {
if (!bdi_has_dirty_io(bdi)) if (!bdi_has_dirty_io(bdi))
continue; continue;
bdi_alloc_queue_work(bdi, wbc); bdi_alloc_queue_work(bdi, &args);
} }
rcu_read_unlock(); rcu_read_unlock();
...@@ -891,17 +919,10 @@ static void bdi_writeback_all(struct writeback_control *wbc) ...@@ -891,17 +919,10 @@ static void bdi_writeback_all(struct writeback_control *wbc)
*/ */
void wakeup_flusher_threads(long nr_pages) void wakeup_flusher_threads(long nr_pages)
{ {
struct writeback_control wbc = {
.sync_mode = WB_SYNC_NONE,
.older_than_this = NULL,
.range_cyclic = 1,
};
if (nr_pages == 0) if (nr_pages == 0)
nr_pages = global_page_state(NR_FILE_DIRTY) + nr_pages = global_page_state(NR_FILE_DIRTY) +
global_page_state(NR_UNSTABLE_NFS); global_page_state(NR_UNSTABLE_NFS);
wbc.nr_to_write = nr_pages; bdi_writeback_all(NULL, nr_pages);
bdi_writeback_all(&wbc);
} }
static noinline void block_dump___mark_inode_dirty(struct inode *inode) static noinline void block_dump___mark_inode_dirty(struct inode *inode)
...@@ -1048,7 +1069,7 @@ EXPORT_SYMBOL(__mark_inode_dirty); ...@@ -1048,7 +1069,7 @@ EXPORT_SYMBOL(__mark_inode_dirty);
* on the writer throttling path, and we get decent balancing between many * on the writer throttling path, and we get decent balancing between many
* throttled threads: we don't want them all piling up on inode_sync_wait. * throttled threads: we don't want them all piling up on inode_sync_wait.
*/ */
static void wait_sb_inodes(struct writeback_control *wbc) static void wait_sb_inodes(struct super_block *sb)
{ {
struct inode *inode, *old_inode = NULL; struct inode *inode, *old_inode = NULL;
...@@ -1056,7 +1077,7 @@ static void wait_sb_inodes(struct writeback_control *wbc) ...@@ -1056,7 +1077,7 @@ static void wait_sb_inodes(struct writeback_control *wbc)
* We need to be protected against the filesystem going from * We need to be protected against the filesystem going from
* r/o to r/w or vice versa. * r/o to r/w or vice versa.
*/ */
WARN_ON(!rwsem_is_locked(&wbc->sb->s_umount)); WARN_ON(!rwsem_is_locked(&sb->s_umount));
spin_lock(&inode_lock); spin_lock(&inode_lock);
...@@ -1067,7 +1088,7 @@ static void wait_sb_inodes(struct writeback_control *wbc) ...@@ -1067,7 +1088,7 @@ static void wait_sb_inodes(struct writeback_control *wbc)
* In which case, the inode may not be on the dirty list, but * In which case, the inode may not be on the dirty list, but
* we still have to wait for that writeout. * we still have to wait for that writeout.
*/ */
list_for_each_entry(inode, &wbc->sb->s_inodes, i_sb_list) { list_for_each_entry(inode, &sb->s_inodes, i_sb_list) {
struct address_space *mapping; struct address_space *mapping;
if (inode->i_state & (I_FREEING|I_CLEAR|I_WILL_FREE|I_NEW)) if (inode->i_state & (I_FREEING|I_CLEAR|I_WILL_FREE|I_NEW))
...@@ -1107,14 +1128,8 @@ static void wait_sb_inodes(struct writeback_control *wbc) ...@@ -1107,14 +1128,8 @@ static void wait_sb_inodes(struct writeback_control *wbc)
* for IO completion of submitted IO. The number of pages submitted is * for IO completion of submitted IO. The number of pages submitted is
* returned. * returned.
*/ */
long writeback_inodes_sb(struct super_block *sb) void writeback_inodes_sb(struct super_block *sb)
{ {
struct writeback_control wbc = {
.sb = sb,
.sync_mode = WB_SYNC_NONE,
.range_start = 0,
.range_end = LLONG_MAX,
};
unsigned long nr_dirty = global_page_state(NR_FILE_DIRTY); unsigned long nr_dirty = global_page_state(NR_FILE_DIRTY);
unsigned long nr_unstable = global_page_state(NR_UNSTABLE_NFS); unsigned long nr_unstable = global_page_state(NR_UNSTABLE_NFS);
long nr_to_write; long nr_to_write;
...@@ -1122,9 +1137,7 @@ long writeback_inodes_sb(struct super_block *sb) ...@@ -1122,9 +1137,7 @@ long writeback_inodes_sb(struct super_block *sb)
nr_to_write = nr_dirty + nr_unstable + nr_to_write = nr_dirty + nr_unstable +
(inodes_stat.nr_inodes - inodes_stat.nr_unused); (inodes_stat.nr_inodes - inodes_stat.nr_unused);
wbc.nr_to_write = nr_to_write; bdi_writeback_all(sb, nr_to_write);
bdi_writeback_all(&wbc);
return nr_to_write - wbc.nr_to_write;
} }
EXPORT_SYMBOL(writeback_inodes_sb); EXPORT_SYMBOL(writeback_inodes_sb);
...@@ -1135,21 +1148,10 @@ EXPORT_SYMBOL(writeback_inodes_sb); ...@@ -1135,21 +1148,10 @@ EXPORT_SYMBOL(writeback_inodes_sb);
* This function writes and waits on any dirty inode belonging to this * This function writes and waits on any dirty inode belonging to this
* super_block. The number of pages synced is returned. * super_block. The number of pages synced is returned.
*/ */
long sync_inodes_sb(struct super_block *sb) void sync_inodes_sb(struct super_block *sb)
{ {
struct writeback_control wbc = { bdi_sync_writeback(sb->s_bdi, sb);
.sb = sb, wait_sb_inodes(sb);
.bdi = sb->s_bdi,
.sync_mode = WB_SYNC_ALL,
.range_start = 0,
.range_end = LLONG_MAX,
};
long nr_to_write = LONG_MAX; /* doesn't actually matter */
wbc.nr_to_write = nr_to_write;
bdi_start_writeback(&wbc);
wait_sb_inodes(&wbc);
return nr_to_write - wbc.nr_to_write;
} }
EXPORT_SYMBOL(sync_inodes_sb); EXPORT_SYMBOL(sync_inodes_sb);
......
...@@ -54,29 +54,15 @@ ...@@ -54,29 +54,15 @@
* @nr_to_write: how many dirty pages to write-back * @nr_to_write: how many dirty pages to write-back
* *
* This function shrinks UBIFS liability by means of writing back some amount * This function shrinks UBIFS liability by means of writing back some amount
* of dirty inodes and their pages. Returns the amount of pages which were * of dirty inodes and their pages.
* written back. The returned value does not include dirty inodes which were
* synchronized.
* *
* Note, this function synchronizes even VFS inodes which are locked * Note, this function synchronizes even VFS inodes which are locked
* (@i_mutex) by the caller of the budgeting function, because write-back does * (@i_mutex) by the caller of the budgeting function, because write-back does
* not touch @i_mutex. * not touch @i_mutex.
*/ */
static int shrink_liability(struct ubifs_info *c, int nr_to_write) static void shrink_liability(struct ubifs_info *c, int nr_to_write)
{ {
int nr_written; writeback_inodes_sb(c->vfs_sb);
nr_written = writeback_inodes_sb(c->vfs_sb);
if (!nr_written) {
/*
* Re-try again but wait on pages/inodes which are being
* written-back concurrently (e.g., by pdflush).
*/
nr_written = sync_inodes_sb(c->vfs_sb);
}
dbg_budg("%d pages were written back", nr_written);
return nr_written;
} }
/** /**
......
...@@ -101,7 +101,7 @@ int bdi_register(struct backing_dev_info *bdi, struct device *parent, ...@@ -101,7 +101,7 @@ int bdi_register(struct backing_dev_info *bdi, struct device *parent,
const char *fmt, ...); const char *fmt, ...);
int bdi_register_dev(struct backing_dev_info *bdi, dev_t dev); int bdi_register_dev(struct backing_dev_info *bdi, dev_t dev);
void bdi_unregister(struct backing_dev_info *bdi); void bdi_unregister(struct backing_dev_info *bdi);
void bdi_start_writeback(struct writeback_control *wbc); void bdi_start_writeback(struct backing_dev_info *bdi, long nr_pages);
int bdi_writeback_task(struct bdi_writeback *wb); int bdi_writeback_task(struct bdi_writeback *wb);
int bdi_has_dirty_io(struct backing_dev_info *bdi); int bdi_has_dirty_io(struct backing_dev_info *bdi);
......
...@@ -68,8 +68,8 @@ struct writeback_control { ...@@ -68,8 +68,8 @@ struct writeback_control {
*/ */
struct bdi_writeback; struct bdi_writeback;
int inode_wait(void *); int inode_wait(void *);
long writeback_inodes_sb(struct super_block *); void writeback_inodes_sb(struct super_block *);
long sync_inodes_sb(struct super_block *); void sync_inodes_sb(struct super_block *);
void writeback_inodes_wbc(struct writeback_control *wbc); void writeback_inodes_wbc(struct writeback_control *wbc);
long wb_do_writeback(struct bdi_writeback *wb, int force_wait); long wb_do_writeback(struct bdi_writeback *wb, int force_wait);
void wakeup_flusher_threads(long nr_pages); void wakeup_flusher_threads(long nr_pages);
......
...@@ -582,16 +582,8 @@ static void balance_dirty_pages(struct address_space *mapping) ...@@ -582,16 +582,8 @@ static void balance_dirty_pages(struct address_space *mapping)
if ((laptop_mode && pages_written) || if ((laptop_mode && pages_written) ||
(!laptop_mode && ((nr_writeback = global_page_state(NR_FILE_DIRTY) (!laptop_mode && ((nr_writeback = global_page_state(NR_FILE_DIRTY)
+ global_page_state(NR_UNSTABLE_NFS)) + global_page_state(NR_UNSTABLE_NFS))
> background_thresh))) { > background_thresh)))
struct writeback_control wbc = { bdi_start_writeback(bdi, nr_writeback);
.bdi = bdi,
.sync_mode = WB_SYNC_NONE,
.nr_to_write = nr_writeback,
};
bdi_start_writeback(&wbc);
}
} }
void set_page_dirty_balance(struct page *page, int page_mkwrite) void set_page_dirty_balance(struct page *page, int page_mkwrite)
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment