tree-defrag.c 6.5 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25
/*
 * Copyright (C) 2007 Oracle.  All rights reserved.
 *
 * This program is free software; you can redistribute it and/or
 * modify it under the terms of the GNU General Public
 * License v2 as published by the Free Software Foundation.
 *
 * This program is distributed in the hope that it will be useful,
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 * General Public License for more details.
 *
 * You should have received a copy of the GNU General Public
 * License along with this program; if not, write to the
 * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
 * Boston, MA 021110-1307, USA.
 */

#include <linux/sched.h>
#include "ctree.h"
#include "disk-io.h"
#include "print-tree.h"
#include "transaction.h"

static void reada_defrag(struct btrfs_root *root,
26
			 struct extent_buffer *node)
27 28 29
{
	int i;
	u32 nritems;
30
	u64 bytenr;
31
	u64 gen;
32
	u32 blocksize;
33 34
	int ret;

35
	blocksize = btrfs_level_size(root, btrfs_header_level(node) - 1);
36
	nritems = btrfs_header_nritems(node);
37
	for (i = 0; i < nritems; i++) {
38
		bytenr = btrfs_node_blockptr(node, i);
39 40
		gen = btrfs_node_ptr_generation(node, i);
		ret = readahead_tree_block(root, bytenr, blocksize, gen);
41 42 43 44 45 46 47 48
		if (ret)
			break;
	}
}

static int defrag_walk_down(struct btrfs_trans_handle *trans,
			    struct btrfs_root *root,
			    struct btrfs_path *path, int *level,
49
			    int cache_only, u64 *last_ret)
50
{
51 52
	struct extent_buffer *next;
	struct extent_buffer *cur;
53
	u64 bytenr;
54
	u64 ptr_gen;
55
	int ret = 0;
56
	int is_extent = 0;
57 58 59 60

	WARN_ON(*level < 0);
	WARN_ON(*level >= BTRFS_MAX_LEVEL);

61 62 63
	if (root->fs_info->extent_root == root)
		is_extent = 1;

64 65 66 67
	if (*level == 1 && cache_only && path->nodes[1] &&
	    !btrfs_buffer_defrag(path->nodes[1])) {
		goto out;
	}
68 69 70 71 72 73
	while(*level > 0) {
		WARN_ON(*level < 0);
		WARN_ON(*level >= BTRFS_MAX_LEVEL);
		cur = path->nodes[*level];

		if (!cache_only && *level > 1 && path->slots[*level] == 0)
74
			reada_defrag(root, cur);
75

76
		if (btrfs_header_level(cur) != *level)
77 78 79
			WARN_ON(1);

		if (path->slots[*level] >=
80
		    btrfs_header_nritems(cur))
81 82 83
			break;

		if (*level == 1) {
84 85
			WARN_ON(btrfs_header_generation(path->nodes[*level]) !=
							trans->transid);
86 87
			ret = btrfs_realloc_node(trans, root,
						 path->nodes[*level],
88 89 90
						 path->slots[*level],
						 cache_only, last_ret,
						 &root->defrag_progress);
91 92 93
			if (is_extent)
				btrfs_extent_post_op(trans, root);

94 95
			break;
		}
96
		bytenr = btrfs_node_blockptr(cur, path->slots[*level]);
97
		ptr_gen = btrfs_node_ptr_generation(cur, path->slots[*level]);
98 99

		if (cache_only) {
100 101
			next = btrfs_find_tree_block(root, bytenr,
					   btrfs_level_size(root, *level - 1));
102
			if (!next || !btrfs_buffer_uptodate(next, ptr_gen) ||
103
			    !btrfs_buffer_defrag(next)) {
104
				free_extent_buffer(next);
105 106 107 108
				path->slots[*level]++;
				continue;
			}
		} else {
109
			next = read_tree_block(root, bytenr,
110
				       btrfs_level_size(root, *level - 1),
111
				       ptr_gen);
112 113 114 115
		}
		ret = btrfs_cow_block(trans, root, next, path->nodes[*level],
				      path->slots[*level], &next);
		BUG_ON(ret);
116 117 118
		if (is_extent)
			btrfs_extent_post_op(trans, root);

119 120
		WARN_ON(*level <= 0);
		if (path->nodes[*level-1])
121
			free_extent_buffer(path->nodes[*level-1]);
122
		path->nodes[*level-1] = next;
123
		*level = btrfs_header_level(next);
124 125 126 127
		path->slots[*level] = 0;
	}
	WARN_ON(*level < 0);
	WARN_ON(*level >= BTRFS_MAX_LEVEL);
128 129

	btrfs_clear_buffer_defrag(path->nodes[*level]);
130
out:
131
	free_extent_buffer(path->nodes[*level]);
132 133
	path->nodes[*level] = NULL;
	*level += 1;
134 135
	WARN_ON(ret && ret != -EAGAIN);
	return ret;
136 137 138 139 140 141 142 143 144
}

static int defrag_walk_up(struct btrfs_trans_handle *trans,
			  struct btrfs_root *root,
			  struct btrfs_path *path, int *level,
			  int cache_only)
{
	int i;
	int slot;
145
	struct extent_buffer *node;
146 147 148

	for(i = *level; i < BTRFS_MAX_LEVEL - 1 && path->nodes[i]; i++) {
		slot = path->slots[i];
149
		if (slot < btrfs_header_nritems(path->nodes[i]) - 1) {
150 151
			path->slots[i]++;
			*level = i;
152
			node = path->nodes[i];
153
			WARN_ON(i == 0);
154 155
			btrfs_node_key_to_cpu(node, &root->defrag_progress,
					      path->slots[i]);
156 157 158
			root->defrag_level = i;
			return 0;
		} else {
159
			btrfs_clear_buffer_defrag(path->nodes[*level]);
160
			free_extent_buffer(path->nodes[*level]);
161 162 163 164 165 166 167 168 169 170 171
			path->nodes[*level] = NULL;
			*level = i + 1;
		}
	}
	return 1;
}

int btrfs_defrag_leaves(struct btrfs_trans_handle *trans,
			struct btrfs_root *root, int cache_only)
{
	struct btrfs_path *path = NULL;
172
	struct extent_buffer *tmp;
173 174 175 176 177
	int ret = 0;
	int wret;
	int level;
	int orig_level;
	int i;
178 179 180 181 182
	int is_extent = 0;
	u64 last_ret = 0;

	if (root->fs_info->extent_root == root)
		is_extent = 1;
183

184
	if (root->ref_cows == 0 && !is_extent)
185
		goto out;
186

187 188 189
	if (btrfs_test_opt(root, SSD))
		goto out;

190 191 192 193
	path = btrfs_alloc_path();
	if (!path)
		return -ENOMEM;

194
	level = btrfs_header_level(root->node);
195
	orig_level = level;
196

197 198 199 200
	if (level == 0) {
		goto out;
	}
	if (root->defrag_progress.objectid == 0) {
C
Chris Mason 已提交
201 202 203 204 205 206 207
		u32 nritems;

		nritems = btrfs_header_nritems(root->node);
		root->defrag_max.objectid = 0;
		/* from above we know this is not a leaf */
		btrfs_node_key_to_cpu(root->node, &root->defrag_max,
				      nritems - 1);
208
		extent_buffer_get(root->node);
209 210 211 212
		ret = btrfs_cow_block(trans, root, root->node, NULL, 0, &tmp);
		BUG_ON(ret);
		path->nodes[level] = root->node;
		path->slots[level] = 0;
213 214
		if (is_extent)
			btrfs_extent_post_op(trans, root);
215 216 217 218 219 220
	} else {
		level = root->defrag_level;
		path->lowest_level = level;
		wret = btrfs_search_slot(trans, root, &root->defrag_progress,
					 path, 0, 1);

221 222
		if (is_extent)
			btrfs_extent_post_op(trans, root);
223

224 225 226 227
		if (wret < 0) {
			ret = wret;
			goto out;
		}
228

229 230
		while(level > 0 && !path->nodes[level])
			level--;
231

232 233 234 235 236 237 238
		if (!path->nodes[level]) {
			ret = 0;
			goto out;
		}
	}

	while(1) {
239 240
		wret = defrag_walk_down(trans, root, path, &level, cache_only,
					&last_ret);
241 242 243 244 245 246 247 248 249 250
		if (wret > 0)
			break;
		if (wret < 0)
			ret = wret;

		wret = defrag_walk_up(trans, root, path, &level, cache_only);
		if (wret > 0)
			break;
		if (wret < 0)
			ret = wret;
251 252
		else
			ret = -EAGAIN;
253
		break;
254 255 256
	}
	for (i = 0; i <= orig_level; i++) {
		if (path->nodes[i]) {
257
			free_extent_buffer(path->nodes[i]);
258
			path->nodes[i] = NULL;
259 260 261 262 263
		}
	}
out:
	if (path)
		btrfs_free_path(path);
C
Chris Mason 已提交
264 265 266 267 268 269 270 271 272 273
	if (ret == -EAGAIN) {
		if (root->defrag_max.objectid > root->defrag_progress.objectid)
			goto done;
		if (root->defrag_max.type > root->defrag_progress.type)
			goto done;
		if (root->defrag_max.offset > root->defrag_progress.offset)
			goto done;
		ret = 0;
	}
done:
274 275 276 277 278 279
	if (ret != -EAGAIN) {
		memset(&root->defrag_progress, 0,
		       sizeof(root->defrag_progress));
	}
	return ret;
}