recovery.c 16.5 KB
Newer Older
J
Jaegeuk Kim 已提交
1
/*
2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
 * fs/f2fs/recovery.c
 *
 * Copyright (c) 2012 Samsung Electronics Co., Ltd.
 *             http://www.samsung.com/
 *
 * This program is free software; you can redistribute it and/or modify
 * it under the terms of the GNU General Public License version 2 as
 * published by the Free Software Foundation.
 */
#include <linux/fs.h>
#include <linux/f2fs_fs.h>
#include "f2fs.h"
#include "node.h"
#include "segment.h"

17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47
/*
 * Roll forward recovery scenarios.
 *
 * [Term] F: fsync_mark, D: dentry_mark
 *
 * 1. inode(x) | CP | inode(x) | dnode(F)
 * -> Update the latest inode(x).
 *
 * 2. inode(x) | CP | inode(F) | dnode(F)
 * -> No problem.
 *
 * 3. inode(x) | CP | dnode(F) | inode(x)
 * -> Recover to the latest dnode(F), and drop the last inode(x)
 *
 * 4. inode(x) | CP | dnode(F) | inode(F)
 * -> No problem.
 *
 * 5. CP | inode(x) | dnode(F)
 * -> The inode(DF) was missing. Should drop this dnode(F).
 *
 * 6. CP | inode(DF) | dnode(F)
 * -> No problem.
 *
 * 7. CP | dnode(F) | inode(DF)
 * -> If f2fs_iget fails, then goto next to find inode(DF).
 *
 * 8. CP | dnode(F) | inode(x)
 * -> If f2fs_iget fails, then goto next to find inode(DF).
 *    But it will fail due to no inode(DF).
 */

48 49 50 51
static struct kmem_cache *fsync_entry_slab;

bool space_for_roll_forward(struct f2fs_sb_info *sbi)
{
52 53 54
	s64 nalloc = percpu_counter_sum_positive(&sbi->alloc_valid_block_count);

	if (sbi->last_valid_block_count + nalloc > sbi->user_block_count)
55 56 57 58 59 60 61 62 63
		return false;
	return true;
}

static struct fsync_inode_entry *get_fsync_inode(struct list_head *head,
								nid_t ino)
{
	struct fsync_inode_entry *entry;

64
	list_for_each_entry(entry, head, list)
65 66
		if (entry->inode->i_ino == ino)
			return entry;
67

68 69 70
	return NULL;
}

71
static struct fsync_inode_entry *add_fsync_inode(struct f2fs_sb_info *sbi,
C
Chao Yu 已提交
72
			struct list_head *head, nid_t ino, bool quota_inode)
73
{
74
	struct inode *inode;
75
	struct fsync_inode_entry *entry;
C
Chao Yu 已提交
76
	int err;
77

78
	inode = f2fs_iget_retry(sbi->sb, ino);
79 80 81
	if (IS_ERR(inode))
		return ERR_CAST(inode);

C
Chao Yu 已提交
82 83 84 85 86 87 88 89 90 91
	err = dquot_initialize(inode);
	if (err)
		goto err_out;

	if (quota_inode) {
		err = dquot_alloc_inode(inode);
		if (err)
			goto err_out;
	}

92
	entry = f2fs_kmem_cache_alloc(fsync_entry_slab, GFP_F2FS_ZERO);
93 94 95 96
	entry->inode = inode;
	list_add_tail(&entry->list, head);

	return entry;
C
Chao Yu 已提交
97 98 99
err_out:
	iput(inode);
	return ERR_PTR(err);
100 101 102 103 104 105 106 107 108
}

static void del_fsync_inode(struct fsync_inode_entry *entry)
{
	iput(entry->inode);
	list_del(&entry->list);
	kmem_cache_free(fsync_entry_slab, entry);
}

C
Chao Yu 已提交
109 110
static int recover_dentry(struct inode *inode, struct page *ipage,
						struct list_head *dir_list)
111
{
112
	struct f2fs_inode *raw_inode = F2FS_INODE(ipage);
113
	nid_t pino = le32_to_cpu(raw_inode->i_pino);
J
Jaegeuk Kim 已提交
114
	struct f2fs_dir_entry *de;
115
	struct fscrypt_name fname;
116
	struct page *page;
J
Jaegeuk Kim 已提交
117
	struct inode *dir, *einode;
C
Chao Yu 已提交
118
	struct fsync_inode_entry *entry;
119
	int err = 0;
120
	char *name;
121

C
Chao Yu 已提交
122 123
	entry = get_fsync_inode(dir_list, pino);
	if (!entry) {
C
Chao Yu 已提交
124 125
		entry = add_fsync_inode(F2FS_I_SB(inode), dir_list,
							pino, false);
126 127 128
		if (IS_ERR(entry)) {
			dir = ERR_CAST(entry);
			err = PTR_ERR(entry);
C
Chao Yu 已提交
129 130
			goto out;
		}
131 132
	}

C
Chao Yu 已提交
133 134
	dir = entry->inode;

135 136 137
	memset(&fname, 0, sizeof(struct fscrypt_name));
	fname.disk_name.len = le32_to_cpu(raw_inode->i_namelen);
	fname.disk_name.name = raw_inode->i_name;
138

139
	if (unlikely(fname.disk_name.len > F2FS_NAME_LEN)) {
140 141
		WARN_ON(1);
		err = -ENAMETOOLONG;
C
Chao Yu 已提交
142
		goto out;
143
	}
J
Jaegeuk Kim 已提交
144
retry:
145
	de = __f2fs_find_entry(dir, &fname, &page);
146
	if (de && inode->i_ino == le32_to_cpu(de->ino))
147
		goto out_put;
148

J
Jaegeuk Kim 已提交
149
	if (de) {
150
		einode = f2fs_iget_retry(inode->i_sb, le32_to_cpu(de->ino));
J
Jaegeuk Kim 已提交
151 152
		if (IS_ERR(einode)) {
			WARN_ON(1);
153 154
			err = PTR_ERR(einode);
			if (err == -ENOENT)
J
Jaegeuk Kim 已提交
155
				err = -EEXIST;
156
			goto out_put;
157
		}
C
Chao Yu 已提交
158 159 160 161

		err = dquot_initialize(einode);
		if (err) {
			iput(einode);
162
			goto out_put;
C
Chao Yu 已提交
163 164
		}

165
		err = acquire_orphan_inode(F2FS_I_SB(inode));
166 167
		if (err) {
			iput(einode);
168
			goto out_put;
J
Jaegeuk Kim 已提交
169
		}
170
		f2fs_delete_entry(de, page, dir, einode);
J
Jaegeuk Kim 已提交
171 172
		iput(einode);
		goto retry;
173 174 175
	} else if (IS_ERR(page)) {
		err = PTR_ERR(page);
	} else {
176
		err = __f2fs_do_add_link(dir, &fname, inode,
177
					inode->i_ino, inode->i_mode);
178
	}
179 180
	if (err == -ENOMEM)
		goto retry;
181 182
	goto out;

183
out_put:
184
	f2fs_put_page(page, 0);
185
out:
186 187 188 189
	if (file_enc_name(inode))
		name = "<encrypted>";
	else
		name = raw_inode->i_name;
C
Chris Fries 已提交
190 191
	f2fs_msg(inode->i_sb, KERN_NOTICE,
			"%s: ino = %x, name = %s, dir = %lx, err = %d",
192
			__func__, ino_of_node(ipage), name,
D
Dan Carpenter 已提交
193
			IS_ERR(dir) ? 0 : dir->i_ino, err);
194 195 196
	return err;
}

J
Jaegeuk Kim 已提交
197 198 199 200 201 202 203 204 205 206 207 208
static void recover_inline_flags(struct inode *inode, struct f2fs_inode *ri)
{
	if (ri->i_inline & F2FS_PIN_FILE)
		set_inode_flag(inode, FI_PIN_FILE);
	else
		clear_inode_flag(inode, FI_PIN_FILE);
	if (ri->i_inline & F2FS_DATA_EXIST)
		set_inode_flag(inode, FI_DATA_EXIST);
	else
		clear_inode_flag(inode, FI_DATA_EXIST);
}

209
static void recover_inode(struct inode *inode, struct page *page)
210
{
211
	struct f2fs_inode *raw = F2FS_INODE(page);
212
	char *name;
213 214

	inode->i_mode = le16_to_cpu(raw->i_mode);
215
	f2fs_i_size_write(inode, le64_to_cpu(raw->i_size));
C
Chao Yu 已提交
216
	inode->i_atime.tv_sec = le64_to_cpu(raw->i_atime);
217 218
	inode->i_ctime.tv_sec = le64_to_cpu(raw->i_ctime);
	inode->i_mtime.tv_sec = le64_to_cpu(raw->i_mtime);
C
Chao Yu 已提交
219
	inode->i_atime.tv_nsec = le32_to_cpu(raw->i_atime_nsec);
220 221
	inode->i_ctime.tv_nsec = le32_to_cpu(raw->i_ctime_nsec);
	inode->i_mtime.tv_nsec = le32_to_cpu(raw->i_mtime_nsec);
222

223 224
	F2FS_I(inode)->i_advise = raw->i_advise;

J
Jaegeuk Kim 已提交
225 226
	recover_inline_flags(inode, raw);

227 228 229 230 231
	if (file_enc_name(inode))
		name = "<encrypted>";
	else
		name = F2FS_INODE(page)->i_name;

J
Jaegeuk Kim 已提交
232 233 234
	f2fs_msg(inode->i_sb, KERN_NOTICE,
		"recover_inode: ino = %x, name = %s, inline = %x",
			ino_of_node(page), name, raw->i_inline);
235 236
}

237 238
static int find_fsync_dnodes(struct f2fs_sb_info *sbi, struct list_head *head,
				bool check_only)
239 240
{
	struct curseg_info *curseg;
241
	struct page *page = NULL;
242
	block_t blkaddr;
243 244 245
	unsigned int loop_cnt = 0;
	unsigned int free_blocks = sbi->user_block_count -
					valid_user_blocks(sbi);
246 247 248 249
	int err = 0;

	/* get node pages in the current segment */
	curseg = CURSEG_I(sbi, CURSEG_WARM_NODE);
250
	blkaddr = NEXT_FREE_BLKADDR(sbi, curseg);
251 252 253 254

	while (1) {
		struct fsync_inode_entry *entry;

C
Chao Yu 已提交
255
		if (!is_valid_meta_blkaddr(sbi, blkaddr, META_POR))
256
			return 0;
257

258
		page = get_tmp_page(sbi, blkaddr);
259

260
		if (!is_recoverable_dnode(page))
261
			break;
262 263 264 265 266

		if (!is_fsync_dnode(page))
			goto next;

		entry = get_fsync_inode(head, ino_of_node(page));
267
		if (!entry) {
C
Chao Yu 已提交
268 269
			bool quota_inode = false;

270 271
			if (!check_only &&
					IS_INODE(page) && is_dent_dnode(page)) {
272 273
				err = recover_inode_page(sbi, page);
				if (err)
274
					break;
C
Chao Yu 已提交
275
				quota_inode = true;
276 277
			}

278 279 280 281
			/*
			 * CP | dnode(F) | inode(DF)
			 * For this case, we should not give up now.
			 */
C
Chao Yu 已提交
282 283
			entry = add_fsync_inode(sbi, head, ino_of_node(page),
								quota_inode);
284 285
			if (IS_ERR(entry)) {
				err = PTR_ERR(entry);
286 287
				if (err == -ENOENT) {
					err = 0;
288
					goto next;
289
				}
290
				break;
291 292
			}
		}
J
Jaegeuk Kim 已提交
293 294
		entry->blkaddr = blkaddr;

295 296
		if (IS_INODE(page) && is_dent_dnode(page))
			entry->last_dentry = blkaddr;
297
next:
298 299 300 301 302 303 304 305 306 307 308
		/* sanity check in order to detect looped node chain */
		if (++loop_cnt >= free_blocks ||
			blkaddr == next_blkaddr_of_node(page)) {
			f2fs_msg(sbi->sb, KERN_NOTICE,
				"%s: detect looped node chain, "
				"blkaddr:%u, next:%u",
				__func__, blkaddr, next_blkaddr_of_node(page));
			err = -EINVAL;
			break;
		}

309 310
		/* check next segment */
		blkaddr = next_blkaddr_of_node(page);
311
		f2fs_put_page(page, 1);
312 313

		ra_meta_pages_cond(sbi, blkaddr);
314
	}
315
	f2fs_put_page(page, 1);
316 317 318
	return err;
}

319
static void destroy_fsync_dnodes(struct list_head *head)
320
{
321 322
	struct fsync_inode_entry *entry, *tmp;

323 324
	list_for_each_entry_safe(entry, tmp, head, list)
		del_fsync_inode(entry);
325 326
}

327
static int check_index_in_prev_nodes(struct f2fs_sb_info *sbi,
328
			block_t blkaddr, struct dnode_of_data *dn)
329 330 331
{
	struct seg_entry *sentry;
	unsigned int segno = GET_SEGNO(sbi, blkaddr);
J
Jaegeuk Kim 已提交
332
	unsigned short blkoff = GET_BLKOFF_FROM_SEG0(sbi, blkaddr);
J
Jaegeuk Kim 已提交
333
	struct f2fs_summary_block *sum_node;
334
	struct f2fs_summary sum;
J
Jaegeuk Kim 已提交
335
	struct page *sum_page, *node_page;
336
	struct dnode_of_data tdn = *dn;
337
	nid_t ino, nid;
338
	struct inode *inode;
339
	unsigned int offset;
340 341 342 343 344
	block_t bidx;
	int i;

	sentry = get_seg_entry(sbi, segno);
	if (!f2fs_test_bit(blkoff, sentry->cur_valid_map))
345
		return 0;
346 347

	/* Get the previous summary */
348
	for (i = CURSEG_HOT_DATA; i <= CURSEG_COLD_DATA; i++) {
349 350 351
		struct curseg_info *curseg = CURSEG_I(sbi, i);
		if (curseg->segno == segno) {
			sum = curseg->sum_blk->entries[blkoff];
J
Jaegeuk Kim 已提交
352
			goto got_it;
353 354 355
		}
	}

J
Jaegeuk Kim 已提交
356 357 358 359 360
	sum_page = get_sum_page(sbi, segno);
	sum_node = (struct f2fs_summary_block *)page_address(sum_page);
	sum = sum_node->entries[blkoff];
	f2fs_put_page(sum_page, 1);
got_it:
361 362 363 364
	/* Use the locked dnode page and inode */
	nid = le32_to_cpu(sum.nid);
	if (dn->inode->i_ino == nid) {
		tdn.nid = nid;
365 366
		if (!dn->inode_page_locked)
			lock_page(dn->inode_page);
367
		tdn.node_page = dn->inode_page;
368
		tdn.ofs_in_node = le16_to_cpu(sum.ofs_in_node);
369
		goto truncate_out;
370
	} else if (dn->nid == nid) {
371
		tdn.ofs_in_node = le16_to_cpu(sum.ofs_in_node);
372
		goto truncate_out;
373 374
	}

375
	/* Get the node page */
376
	node_page = get_node_page(sbi, nid);
377 378
	if (IS_ERR(node_page))
		return PTR_ERR(node_page);
379 380

	offset = ofs_of_node(node_page);
381 382 383
	ino = ino_of_node(node_page);
	f2fs_put_page(node_page, 1);

384
	if (ino != dn->inode->i_ino) {
C
Chao Yu 已提交
385 386
		int ret;

387
		/* Deallocate previous index in the node page */
388
		inode = f2fs_iget_retry(sbi->sb, ino);
389 390
		if (IS_ERR(inode))
			return PTR_ERR(inode);
C
Chao Yu 已提交
391 392 393 394 395 396

		ret = dquot_initialize(inode);
		if (ret) {
			iput(inode);
			return ret;
		}
397 398 399
	} else {
		inode = dn->inode;
	}
400

401
	bidx = start_bidx_of_node(offset, inode) + le16_to_cpu(sum.ofs_in_node);
402

403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419
	/*
	 * if inode page is locked, unlock temporarily, but its reference
	 * count keeps alive.
	 */
	if (ino == dn->inode->i_ino && dn->inode_page_locked)
		unlock_page(dn->inode_page);

	set_new_dnode(&tdn, inode, NULL, NULL, 0);
	if (get_dnode_of_data(&tdn, bidx, LOOKUP_NODE))
		goto out;

	if (tdn.data_blkaddr == blkaddr)
		truncate_data_blocks_range(&tdn, 1);

	f2fs_put_dnode(&tdn);
out:
	if (ino != dn->inode->i_ino)
420
		iput(inode);
421 422 423 424 425
	else if (dn->inode_page_locked)
		lock_page(dn->inode_page);
	return 0;

truncate_out:
426 427
	if (datablock_addr(tdn.inode, tdn.node_page,
					tdn.ofs_in_node) == blkaddr)
428 429 430
		truncate_data_blocks_range(&tdn, 1);
	if (dn->inode->i_ino == nid && !dn->inode_page_locked)
		unlock_page(dn->inode_page);
431
	return 0;
432 433
}

434
static int do_recover_data(struct f2fs_sb_info *sbi, struct inode *inode,
S
Sheng Yong 已提交
435
					struct page *page)
436 437 438
{
	struct dnode_of_data dn;
	struct node_info ni;
439
	unsigned int start, end;
440
	int err = 0, recovered = 0;
441

442 443 444 445
	/* step 1: recover xattr */
	if (IS_INODE(page)) {
		recover_inline_xattr(inode, page);
	} else if (f2fs_has_xattr_block(ofs_of_node(page))) {
S
Sheng Yong 已提交
446
		err = recover_xattr_data(inode, page);
447 448
		if (!err)
			recovered++;
449
		goto out;
450
	}
451

452 453
	/* step 2: recover inline data */
	if (recover_inline_data(inode, page))
454 455
		goto out;

456
	/* step 3: recover data indices */
457 458
	start = start_bidx_of_node(ofs_of_node(page), inode);
	end = start + ADDRS_PER_PAGE(page, inode);
459 460

	set_new_dnode(&dn, inode, NULL, NULL, 0);
461
retry_dn:
462
	err = get_dnode_of_data(&dn, start, ALLOC_NODE);
463 464 465 466 467
	if (err) {
		if (err == -ENOMEM) {
			congestion_wait(BLK_RW_ASYNC, HZ/50);
			goto retry_dn;
		}
468
		goto out;
469
	}
470

471
	f2fs_wait_on_page_writeback(dn.node_page, NODE, true);
472 473

	get_node_info(sbi, dn.nid, &ni);
474 475
	f2fs_bug_on(sbi, ni.ino != ino_of_node(page));
	f2fs_bug_on(sbi, ofs_of_node(dn.node_page) != ofs_of_node(page));
476

477
	for (; start < end; start++, dn.ofs_in_node++) {
478 479
		block_t src, dest;

480 481
		src = datablock_addr(dn.inode, dn.node_page, dn.ofs_in_node);
		dest = datablock_addr(dn.inode, page, dn.ofs_in_node);
482

483 484 485 486 487 488 489 490 491 492
		/* skip recovering if dest is the same as src */
		if (src == dest)
			continue;

		/* dest is invalid, just invalidate src block */
		if (dest == NULL_ADDR) {
			truncate_data_blocks_range(&dn, 1);
			continue;
		}

493
		if (!file_keep_isize(inode) &&
494 495 496
			(i_size_read(inode) <= ((loff_t)start << PAGE_SHIFT)))
			f2fs_i_size_write(inode,
				(loff_t)(start + 1) << PAGE_SHIFT);
497

498 499 500 501 502 503
		/*
		 * dest is reserved block, invalidate src block
		 * and then reserve one new block in dnode page.
		 */
		if (dest == NEW_ADDR) {
			truncate_data_blocks_range(&dn, 1);
504
			reserve_new_block(&dn);
505 506 507 508
			continue;
		}

		/* dest is valid block, try to recover from src to dest */
C
Chao Yu 已提交
509
		if (is_valid_meta_blkaddr(sbi, dest, META_POR)) {
510

511
			if (src == NULL_ADDR) {
512
				err = reserve_new_block(&dn);
513 514 515 516
#ifdef CONFIG_F2FS_FAULT_INJECTION
				while (err)
					err = reserve_new_block(&dn);
#endif
517
				/* We should not get -ENOSPC */
518
				f2fs_bug_on(sbi, err);
519 520
				if (err)
					goto err;
521
			}
522
retry_prev:
523
			/* Check the previous node page having this index */
524
			err = check_index_in_prev_nodes(sbi, dest, &dn);
525 526 527 528 529
			if (err) {
				if (err == -ENOMEM) {
					congestion_wait(BLK_RW_ASYNC, HZ/50);
					goto retry_prev;
				}
530
				goto err;
531
			}
532 533

			/* write dummy data page */
534
			f2fs_replace_block(sbi, &dn, src, dest,
535
						ni.version, false, false);
536
			recovered++;
537 538 539 540 541 542 543
		}
	}

	copy_node_footer(dn.node_page, page);
	fill_node_footer(dn.node_page, dn.nid, ni.ino,
					ofs_of_node(page), false);
	set_page_dirty(dn.node_page);
544
err:
545
	f2fs_put_dnode(&dn);
546
out:
C
Chris Fries 已提交
547
	f2fs_msg(sbi->sb, KERN_NOTICE,
548 549 550 551
		"recover_data: ino = %lx (i_size: %s) recovered = %d, err = %d",
		inode->i_ino,
		file_keep_isize(inode) ? "keep" : "recover",
		recovered, err);
552
	return err;
553 554
}

C
Chao Yu 已提交
555 556
static int recover_data(struct f2fs_sb_info *sbi, struct list_head *inode_list,
						struct list_head *dir_list)
557 558
{
	struct curseg_info *curseg;
559
	struct page *page = NULL;
560
	int err = 0;
561 562 563
	block_t blkaddr;

	/* get node pages in the current segment */
C
Chao Yu 已提交
564
	curseg = CURSEG_I(sbi, CURSEG_WARM_NODE);
565 566 567 568 569
	blkaddr = NEXT_FREE_BLKADDR(sbi, curseg);

	while (1) {
		struct fsync_inode_entry *entry;

C
Chao Yu 已提交
570
		if (!is_valid_meta_blkaddr(sbi, blkaddr, META_POR))
571
			break;
572

573 574
		ra_meta_pages_cond(sbi, blkaddr);

575
		page = get_tmp_page(sbi, blkaddr);
576

577
		if (!is_recoverable_dnode(page)) {
578
			f2fs_put_page(page, 1);
579
			break;
580
		}
581

C
Chao Yu 已提交
582
		entry = get_fsync_inode(inode_list, ino_of_node(page));
583 584
		if (!entry)
			goto next;
585 586 587
		/*
		 * inode(x) | CP | inode(x) | dnode(F)
		 * In this case, we can lose the latest inode(x).
588
		 * So, call recover_inode for the inode update.
589
		 */
590
		if (IS_INODE(page))
591 592
			recover_inode(entry->inode, page);
		if (entry->last_dentry == blkaddr) {
C
Chao Yu 已提交
593
			err = recover_dentry(entry->inode, page, dir_list);
594 595 596 597 598
			if (err) {
				f2fs_put_page(page, 1);
				break;
			}
		}
S
Sheng Yong 已提交
599
		err = do_recover_data(sbi, entry->inode, page);
600 601
		if (err) {
			f2fs_put_page(page, 1);
602
			break;
603
		}
604

605 606
		if (entry->blkaddr == blkaddr)
			del_fsync_inode(entry);
607 608 609
next:
		/* check next segment */
		blkaddr = next_blkaddr_of_node(page);
610
		f2fs_put_page(page, 1);
611
	}
612 613 614
	if (!err)
		allocate_new_segments(sbi);
	return err;
615 616
}

617
int recover_fsync_data(struct f2fs_sb_info *sbi, bool check_only)
618 619
{
	struct list_head inode_list;
C
Chao Yu 已提交
620
	struct list_head dir_list;
621
	int err;
622
	int ret = 0;
C
Chao Yu 已提交
623
	unsigned long s_flags = sbi->sb->s_flags;
H
Haicheng Li 已提交
624
	bool need_writecp = false;
J
Jaegeuk Kim 已提交
625 626 627
#ifdef CONFIG_QUOTA
	int quota_enabled;
#endif
628

629
	if (s_flags & SB_RDONLY) {
C
Chao Yu 已提交
630
		f2fs_msg(sbi->sb, KERN_INFO, "orphan cleanup on readonly fs");
631
		sbi->sb->s_flags &= ~SB_RDONLY;
C
Chao Yu 已提交
632 633 634 635
	}

#ifdef CONFIG_QUOTA
	/* Needed for iput() to work correctly and not trash data */
636
	sbi->sb->s_flags |= SB_ACTIVE;
C
Chao Yu 已提交
637
	/* Turn on quotas so that they are updated correctly */
638
	quota_enabled = f2fs_enable_quota_files(sbi, s_flags & SB_RDONLY);
C
Chao Yu 已提交
639 640
#endif

641
	fsync_entry_slab = f2fs_kmem_cache_create("f2fs_fsync_inode_entry",
642
			sizeof(struct fsync_inode_entry));
C
Chao Yu 已提交
643 644 645 646
	if (!fsync_entry_slab) {
		err = -ENOMEM;
		goto out;
	}
647 648

	INIT_LIST_HEAD(&inode_list);
C
Chao Yu 已提交
649
	INIT_LIST_HEAD(&dir_list);
650

651 652 653
	/* prevent checkpoint */
	mutex_lock(&sbi->cp_mutex);

654
	/* step #1: find fsynced inode numbers */
655
	err = find_fsync_dnodes(sbi, &inode_list, check_only);
656
	if (err || list_empty(&inode_list))
C
Chao Yu 已提交
657
		goto skip;
658

659 660
	if (check_only) {
		ret = 1;
C
Chao Yu 已提交
661
		goto skip;
662
	}
663

H
Haicheng Li 已提交
664
	need_writecp = true;
665

666
	/* step #2: recover data */
C
Chao Yu 已提交
667
	err = recover_data(sbi, &inode_list, &dir_list);
668
	if (!err)
669
		f2fs_bug_on(sbi, !list_empty(&inode_list));
C
Chao Yu 已提交
670
skip:
671
	destroy_fsync_dnodes(&inode_list);
672

673 674
	/* truncate meta pages to be used by the recovery */
	truncate_inode_pages_range(META_MAPPING(sbi),
675
			(loff_t)MAIN_BLKADDR(sbi) << PAGE_SHIFT, -1);
676

677 678 679 680 681
	if (err) {
		truncate_inode_pages_final(NODE_MAPPING(sbi));
		truncate_inode_pages_final(META_MAPPING(sbi));
	}

682
	clear_sbi_flag(sbi, SBI_POR_DOING);
683 684
	mutex_unlock(&sbi->cp_mutex);

685 686 687
	/* let's drop all the directory inodes for clean checkpoint */
	destroy_fsync_dnodes(&dir_list);

688
	if (!err && need_writecp) {
689
		struct cp_control cpc = {
690
			.reason = CP_RECOVERY,
691
		};
C
Chao Yu 已提交
692
		err = write_checkpoint(sbi, &cpc);
693
	}
C
Chao Yu 已提交
694 695

	kmem_cache_destroy(fsync_entry_slab);
C
Chao Yu 已提交
696 697 698
out:
#ifdef CONFIG_QUOTA
	/* Turn quotas off */
J
Jaegeuk Kim 已提交
699 700
	if (quota_enabled)
		f2fs_quota_off_umount(sbi->sb);
C
Chao Yu 已提交
701
#endif
702
	sbi->sb->s_flags = s_flags; /* Restore SB_RDONLY status */
C
Chao Yu 已提交
703

704
	return ret ? ret: err;
705
}