blk-ioc.c 3.7 KB
Newer Older
J
Jens Axboe 已提交
1 2 3 4 5 6 7 8 9
/*
 * Functions related to io context handling
 */
#include <linux/kernel.h>
#include <linux/module.h>
#include <linux/init.h>
#include <linux/bio.h>
#include <linux/blkdev.h>
#include <linux/bootmem.h>	/* for max_pfn/max_low_pfn */
10
#include <linux/slab.h>
J
Jens Axboe 已提交
11 12 13 14 15 16 17 18 19 20

#include "blk.h"

/*
 * For io context allocations
 */
static struct kmem_cache *iocontext_cachep;

static void cfq_dtor(struct io_context *ioc)
{
21 22 23
	if (!hlist_empty(&ioc->cic_list)) {
		struct cfq_io_context *cic;

24
		cic = hlist_entry(ioc->cic_list.first, struct cfq_io_context,
25 26 27
								cic_list);
		cic->dtor(ioc);
	}
J
Jens Axboe 已提交
28 29 30 31 32 33 34 35 36 37 38
}

/*
 * IO Context helper functions. put_io_context() returns 1 if there are no
 * more users of this io context, 0 otherwise.
 */
int put_io_context(struct io_context *ioc)
{
	if (ioc == NULL)
		return 1;

39
	BUG_ON(atomic_long_read(&ioc->refcount) == 0);
J
Jens Axboe 已提交
40

41
	if (atomic_long_dec_and_test(&ioc->refcount)) {
J
Jens Axboe 已提交
42 43
		rcu_read_lock();
		cfq_dtor(ioc);
44
		rcu_read_unlock();
J
Jens Axboe 已提交
45 46 47 48 49 50 51 52 53 54 55 56

		kmem_cache_free(iocontext_cachep, ioc);
		return 1;
	}
	return 0;
}
EXPORT_SYMBOL(put_io_context);

static void cfq_exit(struct io_context *ioc)
{
	rcu_read_lock();

57 58 59
	if (!hlist_empty(&ioc->cic_list)) {
		struct cfq_io_context *cic;

60
		cic = hlist_entry(ioc->cic_list.first, struct cfq_io_context,
61 62 63 64
								cic_list);
		cic->exit(ioc);
	}
	rcu_read_unlock();
J
Jens Axboe 已提交
65 66
}

67
/* Called by the exiting task */
68
void exit_io_context(struct task_struct *task)
J
Jens Axboe 已提交
69 70 71
{
	struct io_context *ioc;

72 73 74 75
	task_lock(task);
	ioc = task->io_context;
	task->io_context = NULL;
	task_unlock(task);
J
Jens Axboe 已提交
76

77
	if (atomic_dec_and_test(&ioc->nr_tasks))
J
Jens Axboe 已提交
78 79
		cfq_exit(ioc);

80
	put_io_context(ioc);
J
Jens Axboe 已提交
81 82 83 84
}

struct io_context *alloc_io_context(gfp_t gfp_flags, int node)
{
85
	struct io_context *ioc;
J
Jens Axboe 已提交
86

87 88 89 90 91 92 93 94 95 96 97 98
	ioc = kmem_cache_alloc_node(iocontext_cachep, gfp_flags, node);
	if (ioc) {
		atomic_long_set(&ioc->refcount, 1);
		atomic_set(&ioc->nr_tasks, 1);
		spin_lock_init(&ioc->lock);
		ioc->ioprio_changed = 0;
		ioc->ioprio = 0;
		ioc->last_waited = 0; /* doesn't matter... */
		ioc->nr_batch_requests = 0; /* because this is 0 */
		INIT_RADIX_TREE(&ioc->radix_root, GFP_ATOMIC | __GFP_HIGH);
		INIT_HLIST_HEAD(&ioc->cic_list);
		ioc->ioc_data = NULL;
99
#if defined(CONFIG_BLK_CGROUP) || defined(CONFIG_BLK_CGROUP_MODULE)
100
		ioc->cgroup_changed = 0;
101
#endif
J
Jens Axboe 已提交
102 103
	}

104
	return ioc;
J
Jens Axboe 已提交
105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141
}

/*
 * If the current task has no IO context then create one and initialise it.
 * Otherwise, return its existing IO context.
 *
 * This returned IO context doesn't have a specifically elevated refcount,
 * but since the current task itself holds a reference, the context can be
 * used in general code, so long as it stays within `current` context.
 */
struct io_context *current_io_context(gfp_t gfp_flags, int node)
{
	struct task_struct *tsk = current;
	struct io_context *ret;

	ret = tsk->io_context;
	if (likely(ret))
		return ret;

	ret = alloc_io_context(gfp_flags, node);
	if (ret) {
		/* make sure set_task_ioprio() sees the settings above */
		smp_wmb();
		tsk->io_context = ret;
	}

	return ret;
}

/*
 * If the current task has no IO context then create one and initialise it.
 * If it does have a context, take a ref on it.
 *
 * This is always called in the context of the task which submitted the I/O.
 */
struct io_context *get_io_context(gfp_t gfp_flags, int node)
{
142
	struct io_context *ioc = NULL;
J
Jens Axboe 已提交
143 144 145 146 147 148

	/*
	 * Check for unlikely race with exiting task. ioc ref count is
	 * zero when ioc is being detached.
	 */
	do {
149 150
		ioc = current_io_context(gfp_flags, node);
		if (unlikely(!ioc))
J
Jens Axboe 已提交
151
			break;
152
	} while (!atomic_long_inc_not_zero(&ioc->refcount));
J
Jens Axboe 已提交
153

154
	return ioc;
J
Jens Axboe 已提交
155 156 157
}
EXPORT_SYMBOL(get_io_context);

A
Adrian Bunk 已提交
158
static int __init blk_ioc_init(void)
J
Jens Axboe 已提交
159 160 161 162 163 164
{
	iocontext_cachep = kmem_cache_create("blkdev_ioc",
			sizeof(struct io_context), 0, SLAB_PANIC, NULL);
	return 0;
}
subsys_initcall(blk_ioc_init);