提交 d38ecf93 编写于 作者: J Jens Axboe

io context sharing: preliminary support

Detach task state from ioc, instead keep track of how many processes
are accessing the ioc.
Signed-off-by: NJens Axboe <jens.axboe@oracle.com>
上级 fd0928df
...@@ -3854,12 +3854,13 @@ int __init blk_dev_init(void) ...@@ -3854,12 +3854,13 @@ int __init blk_dev_init(void)
} }
/* /*
* IO Context helper functions * IO Context helper functions. put_io_context() returns 1 if there are no
* more users of this io context, 0 otherwise.
*/ */
void put_io_context(struct io_context *ioc) int put_io_context(struct io_context *ioc)
{ {
if (ioc == NULL) if (ioc == NULL)
return; return 1;
BUG_ON(atomic_read(&ioc->refcount) == 0); BUG_ON(atomic_read(&ioc->refcount) == 0);
...@@ -3878,7 +3879,9 @@ void put_io_context(struct io_context *ioc) ...@@ -3878,7 +3879,9 @@ void put_io_context(struct io_context *ioc)
rcu_read_unlock(); rcu_read_unlock();
kmem_cache_free(iocontext_cachep, ioc); kmem_cache_free(iocontext_cachep, ioc);
return 1;
} }
return 0;
} }
EXPORT_SYMBOL(put_io_context); EXPORT_SYMBOL(put_io_context);
...@@ -3893,15 +3896,17 @@ void exit_io_context(void) ...@@ -3893,15 +3896,17 @@ void exit_io_context(void)
current->io_context = NULL; current->io_context = NULL;
task_unlock(current); task_unlock(current);
ioc->task = NULL; if (atomic_dec_and_test(&ioc->nr_tasks)) {
if (ioc->aic && ioc->aic->exit) if (ioc->aic && ioc->aic->exit)
ioc->aic->exit(ioc->aic); ioc->aic->exit(ioc->aic);
if (ioc->cic_root.rb_node != NULL) { if (ioc->cic_root.rb_node != NULL) {
cic = rb_entry(rb_first(&ioc->cic_root), struct cfq_io_context, rb_node); cic = rb_entry(rb_first(&ioc->cic_root),
cic->exit(ioc); struct cfq_io_context, rb_node);
} cic->exit(ioc);
}
put_io_context(ioc); put_io_context(ioc);
}
} }
struct io_context *alloc_io_context(gfp_t gfp_flags, int node) struct io_context *alloc_io_context(gfp_t gfp_flags, int node)
...@@ -3911,7 +3916,8 @@ struct io_context *alloc_io_context(gfp_t gfp_flags, int node) ...@@ -3911,7 +3916,8 @@ struct io_context *alloc_io_context(gfp_t gfp_flags, int node)
ret = kmem_cache_alloc_node(iocontext_cachep, gfp_flags, node); ret = kmem_cache_alloc_node(iocontext_cachep, gfp_flags, node);
if (ret) { if (ret) {
atomic_set(&ret->refcount, 1); atomic_set(&ret->refcount, 1);
ret->task = current; atomic_set(&ret->nr_tasks, 1);
spin_lock_init(&ret->lock);
ret->ioprio_changed = 0; ret->ioprio_changed = 0;
ret->ioprio = 0; ret->ioprio = 0;
ret->last_waited = jiffies; /* doesn't matter... */ ret->last_waited = jiffies; /* doesn't matter... */
...@@ -3959,10 +3965,18 @@ static struct io_context *current_io_context(gfp_t gfp_flags, int node) ...@@ -3959,10 +3965,18 @@ static struct io_context *current_io_context(gfp_t gfp_flags, int node)
*/ */
struct io_context *get_io_context(gfp_t gfp_flags, int node) struct io_context *get_io_context(gfp_t gfp_flags, int node)
{ {
struct io_context *ret; struct io_context *ret = NULL;
ret = current_io_context(gfp_flags, node);
if (likely(ret)) /*
atomic_inc(&ret->refcount); * Check for unlikely race with exiting task. ioc ref count is
* zero when ioc is being detached.
*/
do {
ret = current_io_context(gfp_flags, node);
if (unlikely(!ret))
break;
} while (!atomic_inc_not_zero(&ret->refcount));
return ret; return ret;
} }
EXPORT_SYMBOL(get_io_context); EXPORT_SYMBOL(get_io_context);
......
...@@ -54,7 +54,6 @@ static int set_task_ioprio(struct task_struct *task, int ioprio) ...@@ -54,7 +54,6 @@ static int set_task_ioprio(struct task_struct *task, int ioprio)
break; break;
} }
task->io_context = ioc; task->io_context = ioc;
ioc->task = task;
} while (1); } while (1);
if (!err) { if (!err) {
......
...@@ -34,7 +34,7 @@ struct sg_io_hdr; ...@@ -34,7 +34,7 @@ struct sg_io_hdr;
#define BLKDEV_MIN_RQ 4 #define BLKDEV_MIN_RQ 4
#define BLKDEV_MAX_RQ 128 /* Default maximum */ #define BLKDEV_MAX_RQ 128 /* Default maximum */
void put_io_context(struct io_context *ioc); int put_io_context(struct io_context *ioc);
void exit_io_context(void); void exit_io_context(void);
struct io_context *get_io_context(gfp_t gfp_flags, int node); struct io_context *get_io_context(gfp_t gfp_flags, int node);
struct io_context *alloc_io_context(gfp_t gfp_flags, int node); struct io_context *alloc_io_context(gfp_t gfp_flags, int node);
......
...@@ -54,13 +54,15 @@ struct cfq_io_context { ...@@ -54,13 +54,15 @@ struct cfq_io_context {
}; };
/* /*
* This is the per-process I/O subsystem state. It is refcounted and * I/O subsystem state of the associated processes. It is refcounted
* kmalloc'ed. Currently all fields are modified in process io context * and kmalloc'ed. These could be shared between processes.
* (apart from the atomic refcount), so require no locking.
*/ */
struct io_context { struct io_context {
atomic_t refcount; atomic_t refcount;
struct task_struct *task; atomic_t nr_tasks;
/* all the fields below are protected by this lock */
spinlock_t lock;
unsigned short ioprio; unsigned short ioprio;
unsigned short ioprio_changed; unsigned short ioprio_changed;
...@@ -76,4 +78,16 @@ struct io_context { ...@@ -76,4 +78,16 @@ struct io_context {
void *ioc_data; void *ioc_data;
}; };
static inline struct io_context *ioc_task_link(struct io_context *ioc)
{
/*
* if ref count is zero, don't allow sharing (ioc is going away, it's
* a race).
*/
if (ioc && atomic_inc_not_zero(&ioc->refcount))
return ioc;
return NULL;
}
#endif #endif
...@@ -805,7 +805,6 @@ static int copy_io(struct task_struct *tsk) ...@@ -805,7 +805,6 @@ static int copy_io(struct task_struct *tsk)
if (unlikely(!tsk->io_context)) if (unlikely(!tsk->io_context))
return -ENOMEM; return -ENOMEM;
tsk->io_context->task = tsk;
tsk->io_context->ioprio = ioc->ioprio; tsk->io_context->ioprio = ioc->ioprio;
} }
#endif #endif
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册