提交 080dcec4 编写于 作者: A Artem Bityutskiy 提交者: Jens Axboe

writeback: simplify bdi code a little

This patch simplifies bdi code a little by removing the 'pending_list' which is
redundant. Indeed, currently the forker thread ('bdi_forker_thread()') is
working like this:

1. In a loop, fetch all bdi's which have works but have no writeback thread and
   move them to the 'pending_list'.
2. If the list is empty, sleep for 5 sec.
3. Otherwise, take one bdi from the list, fork the writeback thread for this
   bdi, and repeat the loop.

IOW, it first moves everything to the 'pending_list', then process only one
element, and so on. This patch simplifies the algorithm, which is now as
follows.

1. Find the first bdi which has a work and remove it from the global list of
   bdi's (bdi_list).
2. If there was not such bdi, sleep 5 sec.
3. Fork the writeback thread for this bdi and repeat the loop.

IOW, now we find the first bdi to process, process it, and so on. This is
simpler and involves less lists.

The bonus now is that we can get rid of a couple of functions, as well as
remove complications which involve 'rcu_call()' and 'bdi->rcu_head'.

This patch also makes sure we use 'list_add_tail_rcu()', instead of plain
'list_add_tail()', but this piece of code is going to be removed in the next
patch anyway.
Signed-off-by: NArtem Bityutskiy <Artem.Bityutskiy@nokia.com>
Reviewed-by: NChristoph Hellwig <hch@lst.de>
Signed-off-by: NJens Axboe <jaxboe@fusionio.com>
上级 297252c8
...@@ -58,7 +58,6 @@ struct bdi_writeback { ...@@ -58,7 +58,6 @@ struct bdi_writeback {
struct backing_dev_info { struct backing_dev_info {
struct list_head bdi_list; struct list_head bdi_list;
struct rcu_head rcu_head;
unsigned long ra_pages; /* max readahead in PAGE_CACHE_SIZE units */ unsigned long ra_pages; /* max readahead in PAGE_CACHE_SIZE units */
unsigned long state; /* Always use atomic bitops on this */ unsigned long state; /* Always use atomic bitops on this */
unsigned int capabilities; /* Device capabilities */ unsigned int capabilities; /* Device capabilities */
......
...@@ -50,8 +50,6 @@ static struct timer_list sync_supers_timer; ...@@ -50,8 +50,6 @@ static struct timer_list sync_supers_timer;
static int bdi_sync_supers(void *); static int bdi_sync_supers(void *);
static void sync_supers_timer_fn(unsigned long); static void sync_supers_timer_fn(unsigned long);
static void bdi_add_default_flusher_thread(struct backing_dev_info *bdi);
#ifdef CONFIG_DEBUG_FS #ifdef CONFIG_DEBUG_FS
#include <linux/debugfs.h> #include <linux/debugfs.h>
#include <linux/seq_file.h> #include <linux/seq_file.h>
...@@ -331,6 +329,7 @@ static int bdi_forker_thread(void *ptr) ...@@ -331,6 +329,7 @@ static int bdi_forker_thread(void *ptr)
set_user_nice(current, 0); set_user_nice(current, 0);
for (;;) { for (;;) {
bool fork = false;
struct task_struct *task; struct task_struct *task;
struct backing_dev_info *bdi, *tmp; struct backing_dev_info *bdi, *tmp;
...@@ -349,23 +348,30 @@ static int bdi_forker_thread(void *ptr) ...@@ -349,23 +348,30 @@ static int bdi_forker_thread(void *ptr)
* a thread registered. If so, set that up. * a thread registered. If so, set that up.
*/ */
list_for_each_entry_safe(bdi, tmp, &bdi_list, bdi_list) { list_for_each_entry_safe(bdi, tmp, &bdi_list, bdi_list) {
if (!bdi_cap_writeback_dirty(bdi))
continue;
if (bdi->wb.task) if (bdi->wb.task)
continue; continue;
if (list_empty(&bdi->work_list) && if (list_empty(&bdi->work_list) &&
!bdi_has_dirty_io(bdi)) !bdi_has_dirty_io(bdi))
continue; continue;
bdi_add_default_flusher_thread(bdi); WARN(!test_bit(BDI_registered, &bdi->state),
"bdi %p/%s is not registered!\n", bdi, bdi->name);
list_del_rcu(&bdi->bdi_list);
fork = true;
break;
} }
spin_unlock_bh(&bdi_lock);
/* Keep working if default bdi still has things to do */ /* Keep working if default bdi still has things to do */
if (!list_empty(&me->bdi->work_list)) if (!list_empty(&me->bdi->work_list))
__set_current_state(TASK_RUNNING); __set_current_state(TASK_RUNNING);
if (list_empty(&bdi_pending_list)) { if (!fork) {
unsigned long wait; unsigned long wait;
spin_unlock_bh(&bdi_lock);
wait = msecs_to_jiffies(dirty_writeback_interval * 10); wait = msecs_to_jiffies(dirty_writeback_interval * 10);
if (wait) if (wait)
schedule_timeout(wait); schedule_timeout(wait);
...@@ -378,13 +384,13 @@ static int bdi_forker_thread(void *ptr) ...@@ -378,13 +384,13 @@ static int bdi_forker_thread(void *ptr)
__set_current_state(TASK_RUNNING); __set_current_state(TASK_RUNNING);
/* /*
* This is our real job - check for pending entries in * Set the pending bit - if someone will try to unregister this
* bdi_pending_list, and create the threads that got added * bdi - it'll wait on this bit.
*/ */
bdi = list_entry(bdi_pending_list.next, struct backing_dev_info, set_bit(BDI_pending, &bdi->state);
bdi_list);
list_del_init(&bdi->bdi_list); /* Make sure no one uses the picked bdi */
spin_unlock_bh(&bdi_lock); synchronize_rcu();
task = kthread_run(bdi_writeback_thread, &bdi->wb, "flush-%s", task = kthread_run(bdi_writeback_thread, &bdi->wb, "flush-%s",
dev_name(bdi->dev)); dev_name(bdi->dev));
...@@ -397,7 +403,7 @@ static int bdi_forker_thread(void *ptr) ...@@ -397,7 +403,7 @@ static int bdi_forker_thread(void *ptr)
* flush other bdi's to free memory. * flush other bdi's to free memory.
*/ */
spin_lock_bh(&bdi_lock); spin_lock_bh(&bdi_lock);
list_add_tail(&bdi->bdi_list, &bdi_pending_list); list_add_tail_rcu(&bdi->bdi_list, &bdi_list);
spin_unlock_bh(&bdi_lock); spin_unlock_bh(&bdi_lock);
bdi_flush_io(bdi); bdi_flush_io(bdi);
...@@ -408,57 +414,6 @@ static int bdi_forker_thread(void *ptr) ...@@ -408,57 +414,6 @@ static int bdi_forker_thread(void *ptr)
return 0; return 0;
} }
static void bdi_add_to_pending(struct rcu_head *head)
{
struct backing_dev_info *bdi;
bdi = container_of(head, struct backing_dev_info, rcu_head);
INIT_LIST_HEAD(&bdi->bdi_list);
spin_lock(&bdi_lock);
list_add_tail(&bdi->bdi_list, &bdi_pending_list);
spin_unlock(&bdi_lock);
/*
* We are now on the pending list, wake up bdi_forker_task()
* to finish the job and add us back to the active bdi_list
*/
wake_up_process(default_backing_dev_info.wb.task);
}
/*
* Add the default flusher thread that gets created for any bdi
* that has dirty data pending writeout
*/
static void bdi_add_default_flusher_thread(struct backing_dev_info *bdi)
{
if (!bdi_cap_writeback_dirty(bdi))
return;
if (WARN_ON(!test_bit(BDI_registered, &bdi->state))) {
printk(KERN_ERR "bdi %p/%s is not registered!\n",
bdi, bdi->name);
return;
}
/*
* Check with the helper whether to proceed adding a thread. Will only
* abort if we two or more simultanous calls to
* bdi_add_default_flusher_thread() occured, further additions will
* block waiting for previous additions to finish.
*/
if (!test_and_set_bit(BDI_pending, &bdi->state)) {
list_del_rcu(&bdi->bdi_list);
/*
* We must wait for the current RCU period to end before
* moving to the pending list. So schedule that operation
* from an RCU callback.
*/
call_rcu(&bdi->rcu_head, bdi_add_to_pending);
}
}
/* /*
* Remove bdi from bdi_list, and ensure that it is no longer visible * Remove bdi from bdi_list, and ensure that it is no longer visible
*/ */
...@@ -599,7 +554,6 @@ int bdi_init(struct backing_dev_info *bdi) ...@@ -599,7 +554,6 @@ int bdi_init(struct backing_dev_info *bdi)
bdi->max_ratio = 100; bdi->max_ratio = 100;
bdi->max_prop_frac = PROP_FRAC_BASE; bdi->max_prop_frac = PROP_FRAC_BASE;
spin_lock_init(&bdi->wb_lock); spin_lock_init(&bdi->wb_lock);
INIT_RCU_HEAD(&bdi->rcu_head);
INIT_LIST_HEAD(&bdi->bdi_list); INIT_LIST_HEAD(&bdi->bdi_list);
INIT_LIST_HEAD(&bdi->work_list); INIT_LIST_HEAD(&bdi->work_list);
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册