data.c 99.9 KB
Newer Older
C
Chao Yu 已提交
1
// SPDX-License-Identifier: GPL-2.0
J
Jaegeuk Kim 已提交
2
/*
3 4 5 6 7 8 9 10 11 12 13
 * fs/f2fs/data.c
 *
 * Copyright (c) 2012 Samsung Electronics Co., Ltd.
 *             http://www.samsung.com/
 */
#include <linux/fs.h>
#include <linux/f2fs_fs.h>
#include <linux/buffer_head.h>
#include <linux/mpage.h>
#include <linux/writeback.h>
#include <linux/backing-dev.h>
C
Chao Yu 已提交
14
#include <linux/pagevec.h>
15 16
#include <linux/blkdev.h>
#include <linux/bio.h>
17
#include <linux/blk-crypto.h>
J
Jaegeuk Kim 已提交
18
#include <linux/swap.h>
19
#include <linux/prefetch.h>
20
#include <linux/uio.h>
J
Jaegeuk Kim 已提交
21
#include <linux/cleancache.h>
22
#include <linux/sched/signal.h>
23
#include <linux/fiemap.h>
24 25 26 27

#include "f2fs.h"
#include "node.h"
#include "segment.h"
J
Jaegeuk Kim 已提交
28
#include "trace.h"
29
#include <trace/events/f2fs.h>
30

31 32 33
#define NUM_PREALLOC_POST_READ_CTXS	128

static struct kmem_cache *bio_post_read_ctx_cache;
C
Chao Yu 已提交
34
static struct kmem_cache *bio_entry_slab;
35
static mempool_t *bio_post_read_ctx_pool;
C
Chao Yu 已提交
36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52
static struct bio_set f2fs_bioset;

#define	F2FS_BIO_POOL_SIZE	NR_CURSEG_TYPE

int __init f2fs_init_bioset(void)
{
	if (bioset_init(&f2fs_bioset, F2FS_BIO_POOL_SIZE,
					0, BIOSET_NEED_BVECS))
		return -ENOMEM;
	return 0;
}

void f2fs_destroy_bioset(void)
{
	bioset_exit(&f2fs_bioset);
}

53 54 55 56 57 58 59 60 61
static bool __is_cp_guaranteed(struct page *page)
{
	struct address_space *mapping = page->mapping;
	struct inode *inode;
	struct f2fs_sb_info *sbi;

	if (!mapping)
		return false;

C
Chao Yu 已提交
62 63 64
	if (f2fs_is_compressed_page(page))
		return false;

65 66 67 68
	inode = mapping->host;
	sbi = F2FS_I_SB(inode);

	if (inode->i_ino == F2FS_META_INO(sbi) ||
J
Jack Qiu 已提交
69
			inode->i_ino == F2FS_NODE_INO(sbi) ||
70
			S_ISDIR(inode->i_mode) ||
71
			(S_ISREG(inode->i_mode) &&
72
			(f2fs_is_atomic_file(inode) || IS_NOQUOTA(inode))) ||
73 74 75 76 77
			is_cold_data(page))
		return true;
	return false;
}

78 79
static enum count_type __read_io_type(struct page *page)
{
J
Jaegeuk Kim 已提交
80
	struct address_space *mapping = page_file_mapping(page);
81 82 83 84 85 86 87 88 89 90 91 92 93 94

	if (mapping) {
		struct inode *inode = mapping->host;
		struct f2fs_sb_info *sbi = F2FS_I_SB(inode);

		if (inode->i_ino == F2FS_META_INO(sbi))
			return F2FS_RD_META;

		if (inode->i_ino == F2FS_NODE_INO(sbi))
			return F2FS_RD_NODE;
	}
	return F2FS_RD_DATA;
}

95 96 97
/* postprocessing steps for read bios */
enum bio_post_read_step {
	STEP_DECRYPT,
98 99
	STEP_DECOMPRESS_NOWQ,		/* handle normal cluster data inplace */
	STEP_DECOMPRESS,		/* handle compressed cluster data in workqueue */
E
Eric Biggers 已提交
100
	STEP_VERITY,
101 102 103 104
};

struct bio_post_read_ctx {
	struct bio *bio;
C
Chao Yu 已提交
105
	struct f2fs_sb_info *sbi;
106 107 108 109
	struct work_struct work;
	unsigned int enabled_steps;
};

C
Chao Yu 已提交
110
static void __read_end_io(struct bio *bio, bool compr, bool verity)
111
{
112 113
	struct page *page;
	struct bio_vec *bv;
114
	struct bvec_iter_all iter_all;
115

116
	bio_for_each_segment_all(bv, bio, iter_all) {
117 118
		page = bv->bv_page;

C
Chao Yu 已提交
119 120 121 122 123
#ifdef CONFIG_F2FS_FS_COMPRESSION
		if (compr && f2fs_is_compressed_page(page)) {
			f2fs_decompress_pages(bio, page, verity);
			continue;
		}
C
Chao Yu 已提交
124 125
		if (verity)
			continue;
C
Chao Yu 已提交
126 127
#endif

128 129 130
		/* PG_error was set if any post_read step failed */
		if (bio->bi_status || PageError(page)) {
			ClearPageUptodate(page);
131 132
			/* will re-read again later */
			ClearPageError(page);
133 134 135
		} else {
			SetPageUptodate(page);
		}
136
		dec_page_count(F2FS_P_SB(page), __read_io_type(page));
137 138
		unlock_page(page);
	}
C
Chao Yu 已提交
139 140 141 142 143 144 145 146 147 148 149 150 151
}

static void f2fs_release_read_bio(struct bio *bio);
static void __f2fs_read_end_io(struct bio *bio, bool compr, bool verity)
{
	if (!compr)
		__read_end_io(bio, false, verity);
	f2fs_release_read_bio(bio);
}

static void f2fs_decompress_bio(struct bio *bio, bool verity)
{
	__read_end_io(bio, true, verity);
152 153 154 155
}

static void bio_post_read_processing(struct bio_post_read_ctx *ctx);

C
Chao Yu 已提交
156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173
static void f2fs_decrypt_work(struct bio_post_read_ctx *ctx)
{
	fscrypt_decrypt_bio(ctx->bio);
}

static void f2fs_decompress_work(struct bio_post_read_ctx *ctx)
{
	f2fs_decompress_bio(ctx->bio, ctx->enabled_steps & (1 << STEP_VERITY));
}

#ifdef CONFIG_F2FS_FS_COMPRESSION
static void f2fs_verify_pages(struct page **rpages, unsigned int cluster_size)
{
	f2fs_decompress_end_io(rpages, cluster_size, false, true);
}

static void f2fs_verify_bio(struct bio *bio)
{
174 175 176 177 178 179
	struct bio_vec *bv;
	struct bvec_iter_all iter_all;

	bio_for_each_segment_all(bv, bio, iter_all) {
		struct page *page = bv->bv_page;
		struct decompress_io_ctx *dic;
C
Chao Yu 已提交
180

181 182 183
		dic = (struct decompress_io_ctx *)page_private(page);

		if (dic) {
184
			if (atomic_dec_return(&dic->verity_pages))
185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202
				continue;
			f2fs_verify_pages(dic->rpages,
						dic->cluster_size);
			f2fs_free_dic(dic);
			continue;
		}

		if (bio->bi_status || PageError(page))
			goto clear_uptodate;

		if (fsverity_verify_page(page)) {
			SetPageUptodate(page);
			goto unlock;
		}
clear_uptodate:
		ClearPageUptodate(page);
		ClearPageError(page);
unlock:
C
Chao Yu 已提交
203
		dec_page_count(F2FS_P_SB(page), __read_io_type(page));
204 205
		unlock_page(page);
	}
C
Chao Yu 已提交
206 207 208 209
}
#endif

static void f2fs_verity_work(struct work_struct *work)
210 211 212
{
	struct bio_post_read_ctx *ctx =
		container_of(work, struct bio_post_read_ctx, work);
213 214 215 216 217 218 219 220 221 222 223 224 225 226
	struct bio *bio = ctx->bio;
#ifdef CONFIG_F2FS_FS_COMPRESSION
	unsigned int enabled_steps = ctx->enabled_steps;
#endif

	/*
	 * fsverity_verify_bio() may call readpages() again, and while verity
	 * will be disabled for this, decryption may still be needed, resulting
	 * in another bio_post_read_ctx being allocated.  So to prevent
	 * deadlocks we need to release the current ctx to the mempool first.
	 * This assumes that verity is the last post-read step.
	 */
	mempool_free(ctx, bio_post_read_ctx_pool);
	bio->bi_private = NULL;
227

C
Chao Yu 已提交
228 229
#ifdef CONFIG_F2FS_FS_COMPRESSION
	/* previous step is decompression */
230 231 232
	if (enabled_steps & (1 << STEP_DECOMPRESS)) {
		f2fs_verify_bio(bio);
		f2fs_release_read_bio(bio);
C
Chao Yu 已提交
233 234 235
		return;
	}
#endif
236

237 238
	fsverity_verify_bio(bio);
	__f2fs_read_end_io(bio, false, false);
239 240
}

C
Chao Yu 已提交
241
static void f2fs_post_read_work(struct work_struct *work)
E
Eric Biggers 已提交
242 243 244 245
{
	struct bio_post_read_ctx *ctx =
		container_of(work, struct bio_post_read_ctx, work);

C
Chao Yu 已提交
246 247 248 249 250 251 252 253 254 255 256 257 258 259 260
	if (ctx->enabled_steps & (1 << STEP_DECRYPT))
		f2fs_decrypt_work(ctx);

	if (ctx->enabled_steps & (1 << STEP_DECOMPRESS))
		f2fs_decompress_work(ctx);

	if (ctx->enabled_steps & (1 << STEP_VERITY)) {
		INIT_WORK(&ctx->work, f2fs_verity_work);
		fsverity_enqueue_verify_work(&ctx->work);
		return;
	}

	__f2fs_read_end_io(ctx->bio,
		ctx->enabled_steps & (1 << STEP_DECOMPRESS), false);
}
E
Eric Biggers 已提交
261

C
Chao Yu 已提交
262 263 264 265
static void f2fs_enqueue_post_read_work(struct f2fs_sb_info *sbi,
						struct work_struct *work)
{
	queue_work(sbi->post_read_wq, work);
E
Eric Biggers 已提交
266 267
}

268 269
static void bio_post_read_processing(struct bio_post_read_ctx *ctx)
{
E
Eric Biggers 已提交
270 271 272 273 274
	/*
	 * We use different work queues for decryption and for verity because
	 * verity may require reading metadata pages that need decryption, and
	 * we shouldn't recurse to the same workqueue.
	 */
C
Chao Yu 已提交
275 276 277 278 279 280 281 282 283 284 285 286

	if (ctx->enabled_steps & (1 << STEP_DECRYPT) ||
		ctx->enabled_steps & (1 << STEP_DECOMPRESS)) {
		INIT_WORK(&ctx->work, f2fs_post_read_work);
		f2fs_enqueue_post_read_work(ctx->sbi, &ctx->work);
		return;
	}

	if (ctx->enabled_steps & (1 << STEP_VERITY)) {
		INIT_WORK(&ctx->work, f2fs_verity_work);
		fsverity_enqueue_verify_work(&ctx->work);
		return;
287
	}
C
Chao Yu 已提交
288 289

	__f2fs_read_end_io(ctx->bio, false, false);
290 291 292 293
}

static bool f2fs_bio_post_read_required(struct bio *bio)
{
C
Chao Yu 已提交
294
	return bio->bi_private;
295 296 297 298
}

static void f2fs_read_end_io(struct bio *bio)
{
299 300 301 302
	struct f2fs_sb_info *sbi = F2FS_P_SB(bio_first_page_all(bio));

	if (time_to_inject(sbi, FAULT_READ_IO)) {
		f2fs_show_injection_info(sbi, FAULT_READ_IO);
303
		bio->bi_status = BLK_STS_IOERR;
304
	}
C
Chao Yu 已提交
305

306 307
	if (f2fs_bio_post_read_required(bio)) {
		struct bio_post_read_ctx *ctx = bio->bi_private;
J
Jaegeuk Kim 已提交
308

309 310
		bio_post_read_processing(ctx);
		return;
J
Jaegeuk Kim 已提交
311
	}
312

C
Chao Yu 已提交
313
	__f2fs_read_end_io(bio, false, false);
J
Jaegeuk Kim 已提交
314 315
}

316
static void f2fs_write_end_io(struct bio *bio)
317
{
318
	struct f2fs_sb_info *sbi = bio->bi_private;
319
	struct bio_vec *bvec;
320
	struct bvec_iter_all iter_all;
321

322
	if (time_to_inject(sbi, FAULT_WRITE_IO)) {
323
		f2fs_show_injection_info(sbi, FAULT_WRITE_IO);
324 325 326
		bio->bi_status = BLK_STS_IOERR;
	}

327
	bio_for_each_segment_all(bvec, bio, iter_all) {
328
		struct page *page = bvec->bv_page;
329
		enum count_type type = WB_DATA_TYPE(page);
330

331 332 333 334 335 336
		if (IS_DUMMY_WRITTEN_PAGE(page)) {
			set_page_private(page, (unsigned long)NULL);
			ClearPagePrivate(page);
			unlock_page(page);
			mempool_free(page, sbi->write_io_dummy);

337
			if (unlikely(bio->bi_status))
338 339 340 341
				f2fs_stop_checkpoint(sbi, true);
			continue;
		}

342
		fscrypt_finalize_bounce_page(&page);
343

C
Chao Yu 已提交
344 345 346 347 348 349 350
#ifdef CONFIG_F2FS_FS_COMPRESSION
		if (f2fs_is_compressed_page(page)) {
			f2fs_compress_write_end_io(bio, page);
			continue;
		}
#endif

351
		if (unlikely(bio->bi_status)) {
352
			mapping_set_error(page->mapping, -EIO);
353 354
			if (type == F2FS_WB_CP_DATA)
				f2fs_stop_checkpoint(sbi, true);
355
		}
356 357 358 359

		f2fs_bug_on(sbi, page->mapping == NODE_MAPPING(sbi) &&
					page->index != nid_of_node(page));

360
		dec_page_count(sbi, type);
361 362
		if (f2fs_in_warm_node_list(sbi, page))
			f2fs_del_fsync_node_entry(sbi, page);
363
		clear_cold_data(page);
364
		end_page_writeback(page);
365
	}
366
	if (!get_pages(sbi, F2FS_WB_CP_DATA) &&
367
				wq_has_sleeper(&sbi->cp_wait))
368 369 370 371 372
		wake_up(&sbi->cp_wait);

	bio_put(bio);
}

J
Jaegeuk Kim 已提交
373 374 375 376 377 378
struct block_device *f2fs_target_device(struct f2fs_sb_info *sbi,
				block_t blk_addr, struct bio *bio)
{
	struct block_device *bdev = sbi->sb->s_bdev;
	int i;

379 380 381 382 383 384 385 386
	if (f2fs_is_multi_device(sbi)) {
		for (i = 0; i < sbi->s_ndevs; i++) {
			if (FDEV(i).start_blk <= blk_addr &&
			    FDEV(i).end_blk >= blk_addr) {
				blk_addr -= FDEV(i).start_blk;
				bdev = FDEV(i).bdev;
				break;
			}
J
Jaegeuk Kim 已提交
387 388 389
		}
	}
	if (bio) {
390
		bio_set_dev(bio, bdev);
J
Jaegeuk Kim 已提交
391 392 393 394 395 396 397 398 399
		bio->bi_iter.bi_sector = SECTOR_FROM_BLOCK(blk_addr);
	}
	return bdev;
}

int f2fs_target_device_index(struct f2fs_sb_info *sbi, block_t blkaddr)
{
	int i;

400 401 402
	if (!f2fs_is_multi_device(sbi))
		return 0;

J
Jaegeuk Kim 已提交
403 404 405 406 407 408
	for (i = 0; i < sbi->s_ndevs; i++)
		if (FDEV(i).start_blk <= blkaddr && FDEV(i).end_blk >= blkaddr)
			return i;
	return 0;
}

409
static struct bio *__bio_alloc(struct f2fs_io_info *fio, int npages)
410
{
411
	struct f2fs_sb_info *sbi = fio->sbi;
412 413
	struct bio *bio;

C
Christoph Hellwig 已提交
414
	bio = bio_alloc_bioset(GFP_NOIO, npages, &f2fs_bioset);
415

416 417
	f2fs_target_device(sbi, fio->new_blkaddr, bio);
	if (is_read_io(fio->op)) {
418 419 420 421 422
		bio->bi_end_io = f2fs_read_end_io;
		bio->bi_private = NULL;
	} else {
		bio->bi_end_io = f2fs_write_end_io;
		bio->bi_private = sbi;
423 424
		bio->bi_write_hint = f2fs_io_type_to_rw_hint(sbi,
						fio->type, fio->temp);
425
	}
426 427
	if (fio->io_wbc)
		wbc_init_bio(fio->io_wbc, bio);
428 429 430 431

	return bio;
}

432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458
static void f2fs_set_bio_crypt_ctx(struct bio *bio, const struct inode *inode,
				  pgoff_t first_idx,
				  const struct f2fs_io_info *fio,
				  gfp_t gfp_mask)
{
	/*
	 * The f2fs garbage collector sets ->encrypted_page when it wants to
	 * read/write raw data without encryption.
	 */
	if (!fio || !fio->encrypted_page)
		fscrypt_set_bio_crypt_ctx(bio, inode, first_idx, gfp_mask);
}

static bool f2fs_crypt_mergeable_bio(struct bio *bio, const struct inode *inode,
				     pgoff_t next_idx,
				     const struct f2fs_io_info *fio)
{
	/*
	 * The f2fs garbage collector sets ->encrypted_page when it wants to
	 * read/write raw data without encryption.
	 */
	if (fio && fio->encrypted_page)
		return !bio_has_crypt_ctx(bio);

	return fscrypt_mergeable_bio(bio, inode, next_idx);
}

459 460
static inline void __submit_bio(struct f2fs_sb_info *sbi,
				struct bio *bio, enum page_type type)
461
{
462
	if (!is_read_io(bio_op(bio))) {
463 464 465 466 467
		unsigned int start;

		if (type != DATA && type != NODE)
			goto submit_io;

468
		if (f2fs_lfs_mode(sbi) && current->plug)
469 470
			blk_finish_plug(current->plug);

471 472 473
		if (F2FS_IO_ALIGNED(sbi))
			goto submit_io;

474 475 476 477 478 479 480 481 482 483
		start = bio->bi_iter.bi_size >> F2FS_BLKSIZE_BITS;
		start %= F2FS_IO_SIZE(sbi);

		if (start == 0)
			goto submit_io;

		/* fill dummy pages */
		for (; start < F2FS_IO_SIZE(sbi); start++) {
			struct page *page =
				mempool_alloc(sbi->write_io_dummy,
484
					      GFP_NOIO | __GFP_NOFAIL);
485 486
			f2fs_bug_on(sbi, !page);

487
			zero_user_segment(page, 0, PAGE_SIZE);
488
			SetPagePrivate(page);
489
			set_page_private(page, DUMMY_WRITTEN_PAGE);
490 491 492 493 494 495 496 497 498 499
			lock_page(page);
			if (bio_add_page(bio, page, PAGE_SIZE, 0) < PAGE_SIZE)
				f2fs_bug_on(sbi, 1);
		}
		/*
		 * In the NODE case, we lose next block address chain. So, we
		 * need to do checkpoint in f2fs_sync_file.
		 */
		if (type == NODE)
			set_sbi_flag(sbi, SBI_NEED_CP);
J
Jaegeuk Kim 已提交
500
	}
501
submit_io:
J
Jaegeuk Kim 已提交
502 503 504 505
	if (is_read_io(bio_op(bio)))
		trace_f2fs_submit_read_bio(sbi->sb, type, bio);
	else
		trace_f2fs_submit_write_bio(sbi->sb, type, bio);
506
	submit_bio(bio);
507 508
}

C
Chao Yu 已提交
509 510 511 512 513 514
void f2fs_submit_bio(struct f2fs_sb_info *sbi,
				struct bio *bio, enum page_type type)
{
	__submit_bio(sbi, bio, type);
}

515
static void __attach_io_flag(struct f2fs_io_info *fio)
516 517 518
{
	struct f2fs_sb_info *sbi = fio->sbi;
	unsigned int temp_mask = (1 << NR_TEMP_TYPE) - 1;
519 520 521 522 523 524 525 526 527 528 529 530
	unsigned int io_flag, fua_flag, meta_flag;

	if (fio->type == DATA)
		io_flag = sbi->data_io_flag;
	else if (fio->type == NODE)
		io_flag = sbi->node_io_flag;
	else
		return;

	fua_flag = io_flag & temp_mask;
	meta_flag = (io_flag >> NR_TEMP_TYPE) & temp_mask;

531
	/*
532
	 * data/node io flag bits per temp:
533 534 535 536 537 538 539 540 541 542
	 *      REQ_META     |      REQ_FUA      |
	 *    5 |    4 |   3 |    2 |    1 |   0 |
	 * Cold | Warm | Hot | Cold | Warm | Hot |
	 */
	if ((1 << fio->temp) & meta_flag)
		fio->op_flags |= REQ_META;
	if ((1 << fio->temp) & fua_flag)
		fio->op_flags |= REQ_FUA;
}

J
Jaegeuk Kim 已提交
543
static void __submit_merged_bio(struct f2fs_bio_info *io)
544
{
J
Jaegeuk Kim 已提交
545
	struct f2fs_io_info *fio = &io->fio;
546 547 548 549

	if (!io->bio)
		return;

550
	__attach_io_flag(fio);
J
Jaegeuk Kim 已提交
551 552
	bio_set_op_attrs(io->bio, fio->op, fio->op_flags);

M
Mike Christie 已提交
553
	if (is_read_io(fio->op))
J
Jaegeuk Kim 已提交
554
		trace_f2fs_prepare_read_bio(io->sbi->sb, fio->type, io->bio);
555
	else
J
Jaegeuk Kim 已提交
556
		trace_f2fs_prepare_write_bio(io->sbi->sb, fio->type, io->bio);
M
Mike Christie 已提交
557

558
	__submit_bio(io->sbi, io->bio, fio->type);
559 560 561
	io->bio = NULL;
}

C
Chao Yu 已提交
562
static bool __has_merged_page(struct bio *bio, struct inode *inode,
563
						struct page *page, nid_t ino)
C
Chao Yu 已提交
564 565
{
	struct bio_vec *bvec;
566
	struct bvec_iter_all iter_all;
C
Chao Yu 已提交
567

C
Chao Yu 已提交
568
	if (!bio)
C
Chao Yu 已提交
569
		return false;
570

571
	if (!inode && !page && !ino)
572
		return true;
C
Chao Yu 已提交
573

C
Chao Yu 已提交
574
	bio_for_each_segment_all(bvec, bio, iter_all) {
C
Chao Yu 已提交
575
		struct page *target = bvec->bv_page;
C
Chao Yu 已提交
576

C
Chao Yu 已提交
577
		if (fscrypt_is_bounce_page(target)) {
578
			target = fscrypt_pagecache_page(target);
C
Chao Yu 已提交
579 580 581 582 583 584 585 586
			if (IS_ERR(target))
				continue;
		}
		if (f2fs_is_compressed_page(target)) {
			target = f2fs_compress_control_page(target);
			if (IS_ERR(target))
				continue;
		}
C
Chao Yu 已提交
587

588 589
		if (inode && inode == target->mapping->host)
			return true;
590 591
		if (page && page == target)
			return true;
592
		if (ino && ino == ino_of_node(target))
C
Chao Yu 已提交
593 594 595 596 597 598
			return true;
	}

	return false;
}

599
static void __f2fs_submit_merged_write(struct f2fs_sb_info *sbi,
J
Jaegeuk Kim 已提交
600
				enum page_type type, enum temp_type temp)
601 602
{
	enum page_type btype = PAGE_TYPE_OF_BIO(type);
J
Jaegeuk Kim 已提交
603
	struct f2fs_bio_info *io = sbi->write_io[btype] + temp;
604

605
	down_write(&io->io_rwsem);
J
Jaegeuk Kim 已提交
606 607 608 609

	/* change META to META_FLUSH in the checkpoint procedure */
	if (type >= META_FLUSH) {
		io->fio.type = META_FLUSH;
M
Mike Christie 已提交
610
		io->fio.op = REQ_OP_WRITE;
611
		io->fio.op_flags = REQ_META | REQ_PRIO | REQ_SYNC;
612
		if (!test_opt(sbi, NOBARRIER))
613
			io->fio.op_flags |= REQ_PREFLUSH | REQ_FUA;
J
Jaegeuk Kim 已提交
614 615
	}
	__submit_merged_bio(io);
616
	up_write(&io->io_rwsem);
617 618
}

J
Jaegeuk Kim 已提交
619
static void __submit_merged_write_cond(struct f2fs_sb_info *sbi,
620 621
				struct inode *inode, struct page *page,
				nid_t ino, enum page_type type, bool force)
622
{
J
Jaegeuk Kim 已提交
623
	enum temp_type temp;
624
	bool ret = true;
J
Jaegeuk Kim 已提交
625 626

	for (temp = HOT; temp < NR_TEMP_TYPE; temp++) {
627 628 629
		if (!force)	{
			enum page_type btype = PAGE_TYPE_OF_BIO(type);
			struct f2fs_bio_info *io = sbi->write_io[btype] + temp;
J
Jaegeuk Kim 已提交
630

631
			down_read(&io->io_rwsem);
C
Chao Yu 已提交
632
			ret = __has_merged_page(io->bio, inode, page, ino);
633 634 635 636
			up_read(&io->io_rwsem);
		}
		if (ret)
			__f2fs_submit_merged_write(sbi, type, temp);
J
Jaegeuk Kim 已提交
637 638 639 640 641

		/* TODO: use HOT temp only for meta pages now. */
		if (type >= META)
			break;
	}
642 643
}

644
void f2fs_submit_merged_write(struct f2fs_sb_info *sbi, enum page_type type)
645
{
646
	__submit_merged_write_cond(sbi, NULL, NULL, 0, type, true);
647 648
}

649
void f2fs_submit_merged_write_cond(struct f2fs_sb_info *sbi,
650 651
				struct inode *inode, struct page *page,
				nid_t ino, enum page_type type)
652
{
653
	__submit_merged_write_cond(sbi, inode, page, ino, type, false);
654 655
}

656
void f2fs_flush_merged_writes(struct f2fs_sb_info *sbi)
657
{
658 659 660
	f2fs_submit_merged_write(sbi, DATA);
	f2fs_submit_merged_write(sbi, NODE);
	f2fs_submit_merged_write(sbi, META);
661 662
}

663 664
/*
 * Fill the locked page with data located in the block address.
665
 * A caller needs to unlock the page on failure.
666
 */
667
int f2fs_submit_page_bio(struct f2fs_io_info *fio)
668 669
{
	struct bio *bio;
670 671
	struct page *page = fio->encrypted_page ?
			fio->encrypted_page : fio->page;
672

673
	if (!f2fs_is_valid_blkaddr(fio->sbi, fio->new_blkaddr,
C
Chao Yu 已提交
674 675
			fio->is_por ? META_POR : (__is_meta_io(fio) ?
			META_GENERIC : DATA_GENERIC_ENHANCE)))
676
		return -EFSCORRUPTED;
677

678
	trace_f2fs_submit_page_bio(page, fio);
679
	f2fs_trace_ios(fio, 0);
680 681

	/* Allocate a new bio */
682
	bio = __bio_alloc(fio, 1);
683

684 685 686
	f2fs_set_bio_crypt_ctx(bio, fio->page->mapping->host,
			       fio->page->index, fio, GFP_NOIO);

687
	if (bio_add_page(bio, page, PAGE_SIZE, 0) < PAGE_SIZE) {
688 689 690
		bio_put(bio);
		return -EFAULT;
	}
691 692

	if (fio->io_wbc && !is_read_io(fio->op))
693
		wbc_account_cgroup_owner(fio->io_wbc, page, PAGE_SIZE);
694

695
	__attach_io_flag(fio);
M
Mike Christie 已提交
696
	bio_set_op_attrs(bio, fio->op, fio->op_flags);
697

698 699
	inc_page_count(fio->sbi, is_read_io(fio->op) ?
			__read_io_type(page): WB_DATA_TYPE(fio->page));
C
Chao Yu 已提交
700 701

	__submit_bio(fio->sbi, bio, fio->type);
702 703 704
	return 0;
}

705 706 707
static bool page_is_mergeable(struct f2fs_sb_info *sbi, struct bio *bio,
				block_t last_blkaddr, block_t cur_blkaddr)
{
708 709 710
	if (unlikely(sbi->max_io_bytes &&
			bio->bi_iter.bi_size >= sbi->max_io_bytes))
		return false;
711 712
	if (last_blkaddr + 1 != cur_blkaddr)
		return false;
713
	return bio->bi_bdev == f2fs_target_device(sbi, cur_blkaddr, NULL);
714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729
}

static bool io_type_is_mergeable(struct f2fs_bio_info *io,
						struct f2fs_io_info *fio)
{
	if (io->fio.op != fio->op)
		return false;
	return io->fio.op_flags == fio->op_flags;
}

static bool io_is_mergeable(struct f2fs_sb_info *sbi, struct bio *bio,
					struct f2fs_bio_info *io,
					struct f2fs_io_info *fio,
					block_t last_blkaddr,
					block_t cur_blkaddr)
{
730 731 732 733 734 735 736 737 738 739
	if (F2FS_IO_ALIGNED(sbi) && (fio->type == DATA || fio->type == NODE)) {
		unsigned int filled_blocks =
				F2FS_BYTES_TO_BLK(bio->bi_iter.bi_size);
		unsigned int io_size = F2FS_IO_SIZE(sbi);
		unsigned int left_vecs = bio->bi_max_vecs - bio->bi_vcnt;

		/* IOs in bio is aligned and left space of vectors is not enough */
		if (!(filled_blocks % io_size) && left_vecs < io_size)
			return false;
	}
740 741 742 743 744
	if (!page_is_mergeable(sbi, bio, last_blkaddr, cur_blkaddr))
		return false;
	return io_type_is_mergeable(io, fio);
}

C
Chao Yu 已提交
745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768
static void add_bio_entry(struct f2fs_sb_info *sbi, struct bio *bio,
				struct page *page, enum temp_type temp)
{
	struct f2fs_bio_info *io = sbi->write_io[DATA] + temp;
	struct bio_entry *be;

	be = f2fs_kmem_cache_alloc(bio_entry_slab, GFP_NOFS);
	be->bio = bio;
	bio_get(bio);

	if (bio_add_page(bio, page, PAGE_SIZE, 0) != PAGE_SIZE)
		f2fs_bug_on(sbi, 1);

	down_write(&io->bio_list_lock);
	list_add_tail(&be->list, &io->bio_list);
	up_write(&io->bio_list_lock);
}

static void del_bio_entry(struct bio_entry *be)
{
	list_del(&be->list);
	kmem_cache_free(bio_entry_slab, be);
}

769
static int add_ipu_page(struct f2fs_io_info *fio, struct bio **bio,
C
Chao Yu 已提交
770 771
							struct page *page)
{
772
	struct f2fs_sb_info *sbi = fio->sbi;
C
Chao Yu 已提交
773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788
	enum temp_type temp;
	bool found = false;
	int ret = -EAGAIN;

	for (temp = HOT; temp < NR_TEMP_TYPE && !found; temp++) {
		struct f2fs_bio_info *io = sbi->write_io[DATA] + temp;
		struct list_head *head = &io->bio_list;
		struct bio_entry *be;

		down_write(&io->bio_list_lock);
		list_for_each_entry(be, head, list) {
			if (be->bio != *bio)
				continue;

			found = true;

789 790 791 792 793 794 795 796
			f2fs_bug_on(sbi, !page_is_mergeable(sbi, *bio,
							    *fio->last_block,
							    fio->new_blkaddr));
			if (f2fs_crypt_mergeable_bio(*bio,
					fio->page->mapping->host,
					fio->page->index, fio) &&
			    bio_add_page(*bio, page, PAGE_SIZE, 0) ==
					PAGE_SIZE) {
C
Chao Yu 已提交
797 798 799 800
				ret = 0;
				break;
			}

801
			/* page can't be merged into bio; submit the bio */
C
Chao Yu 已提交
802 803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872
			del_bio_entry(be);
			__submit_bio(sbi, *bio, DATA);
			break;
		}
		up_write(&io->bio_list_lock);
	}

	if (ret) {
		bio_put(*bio);
		*bio = NULL;
	}

	return ret;
}

void f2fs_submit_merged_ipu_write(struct f2fs_sb_info *sbi,
					struct bio **bio, struct page *page)
{
	enum temp_type temp;
	bool found = false;
	struct bio *target = bio ? *bio : NULL;

	for (temp = HOT; temp < NR_TEMP_TYPE && !found; temp++) {
		struct f2fs_bio_info *io = sbi->write_io[DATA] + temp;
		struct list_head *head = &io->bio_list;
		struct bio_entry *be;

		if (list_empty(head))
			continue;

		down_read(&io->bio_list_lock);
		list_for_each_entry(be, head, list) {
			if (target)
				found = (target == be->bio);
			else
				found = __has_merged_page(be->bio, NULL,
								page, 0);
			if (found)
				break;
		}
		up_read(&io->bio_list_lock);

		if (!found)
			continue;

		found = false;

		down_write(&io->bio_list_lock);
		list_for_each_entry(be, head, list) {
			if (target)
				found = (target == be->bio);
			else
				found = __has_merged_page(be->bio, NULL,
								page, 0);
			if (found) {
				target = be->bio;
				del_bio_entry(be);
				break;
			}
		}
		up_write(&io->bio_list_lock);
	}

	if (found)
		__submit_bio(sbi, target, DATA);
	if (bio && *bio) {
		bio_put(*bio);
		*bio = NULL;
	}
}

C
Chao Yu 已提交
873 874 875 876 877 878 879 880
int f2fs_merge_page_bio(struct f2fs_io_info *fio)
{
	struct bio *bio = *fio->bio;
	struct page *page = fio->encrypted_page ?
			fio->encrypted_page : fio->page;

	if (!f2fs_is_valid_blkaddr(fio->sbi, fio->new_blkaddr,
			__is_meta_io(fio) ? META_GENERIC : DATA_GENERIC))
881
		return -EFSCORRUPTED;
C
Chao Yu 已提交
882 883 884 885

	trace_f2fs_submit_page_bio(page, fio);
	f2fs_trace_ios(fio, 0);

886
	if (bio && !page_is_mergeable(fio->sbi, bio, *fio->last_block,
C
Chao Yu 已提交
887 888
						fio->new_blkaddr))
		f2fs_submit_merged_ipu_write(fio->sbi, &bio, NULL);
C
Chao Yu 已提交
889 890
alloc_new:
	if (!bio) {
891
		bio = __bio_alloc(fio, BIO_MAX_PAGES);
892
		__attach_io_flag(fio);
893 894
		f2fs_set_bio_crypt_ctx(bio, fio->page->mapping->host,
				       fio->page->index, fio, GFP_NOIO);
C
Chao Yu 已提交
895 896
		bio_set_op_attrs(bio, fio->op, fio->op_flags);

C
Chao Yu 已提交
897 898
		add_bio_entry(fio->sbi, bio, page, fio->temp);
	} else {
899
		if (add_ipu_page(fio, &bio, page))
C
Chao Yu 已提交
900
			goto alloc_new;
C
Chao Yu 已提交
901 902 903
	}

	if (fio->io_wbc)
904
		wbc_account_cgroup_owner(fio->io_wbc, page, PAGE_SIZE);
C
Chao Yu 已提交
905 906 907 908 909 910 911 912 913

	inc_page_count(fio->sbi, WB_DATA_TYPE(page));

	*fio->last_block = fio->new_blkaddr;
	*fio->bio = bio;

	return 0;
}

914
void f2fs_submit_page_write(struct f2fs_io_info *fio)
915
{
916
	struct f2fs_sb_info *sbi = fio->sbi;
J
Jaegeuk Kim 已提交
917
	enum page_type btype = PAGE_TYPE_OF_BIO(fio->type);
J
Jaegeuk Kim 已提交
918
	struct f2fs_bio_info *io = sbi->write_io[btype] + fio->temp;
919
	struct page *bio_page;
920

921
	f2fs_bug_on(sbi, is_read_io(fio->op));
922

923 924 925 926 927 928
	down_write(&io->io_rwsem);
next:
	if (fio->in_list) {
		spin_lock(&io->io_lock);
		if (list_empty(&io->io_list)) {
			spin_unlock(&io->io_lock);
929
			goto out;
930 931 932 933 934 935
		}
		fio = list_first_entry(&io->io_list,
						struct f2fs_io_info, list);
		list_del(&fio->list);
		spin_unlock(&io->io_lock);
	}
936

C
Chao Yu 已提交
937
	verify_fio_blkaddr(fio);
938

C
Chao Yu 已提交
939 940 941 942 943 944
	if (fio->encrypted_page)
		bio_page = fio->encrypted_page;
	else if (fio->compressed_page)
		bio_page = fio->compressed_page;
	else
		bio_page = fio->page;
945

946 947
	/* set submitted = true as a return value */
	fio->submitted = true;
948

949
	inc_page_count(sbi, WB_DATA_TYPE(bio_page));
950

951 952 953 954 955
	if (io->bio &&
	    (!io_is_mergeable(sbi, io->bio, io, fio, io->last_block_in_bio,
			      fio->new_blkaddr) ||
	     !f2fs_crypt_mergeable_bio(io->bio, fio->page->mapping->host,
				       bio_page->index, fio)))
J
Jaegeuk Kim 已提交
956
		__submit_merged_bio(io);
957 958
alloc_new:
	if (io->bio == NULL) {
959 960
		if (F2FS_IO_ALIGNED(sbi) &&
				(fio->type == DATA || fio->type == NODE) &&
961
				fio->new_blkaddr & F2FS_IO_SIZE_MASK(sbi)) {
962
			dec_page_count(sbi, WB_DATA_TYPE(bio_page));
963 964
			fio->retry = true;
			goto skip;
965
		}
966
		io->bio = __bio_alloc(fio, BIO_MAX_PAGES);
967 968
		f2fs_set_bio_crypt_ctx(io->bio, fio->page->mapping->host,
				       bio_page->index, fio, GFP_NOIO);
J
Jaegeuk Kim 已提交
969
		io->fio = *fio;
970 971
	}

J
Jaegeuk Kim 已提交
972
	if (bio_add_page(io->bio, bio_page, PAGE_SIZE, 0) < PAGE_SIZE) {
J
Jaegeuk Kim 已提交
973
		__submit_merged_bio(io);
974 975 976
		goto alloc_new;
	}

977
	if (fio->io_wbc)
978
		wbc_account_cgroup_owner(fio->io_wbc, bio_page, PAGE_SIZE);
979

980
	io->last_block_in_bio = fio->new_blkaddr;
981
	f2fs_trace_ios(fio, 0);
982 983

	trace_f2fs_submit_page_write(fio->page, fio);
984
skip:
985 986
	if (fio->in_list)
		goto next;
987
out:
D
Daniel Rosenberg 已提交
988
	if (is_sbi_flag_set(sbi, SBI_IS_SHUTDOWN) ||
989
				!f2fs_is_checkpoint_ready(sbi))
J
Jaegeuk Kim 已提交
990
		__submit_merged_bio(io);
991
	up_write(&io->io_rwsem);
992 993
}

E
Eric Biggers 已提交
994 995 996 997 998 999
static inline bool f2fs_need_verity(const struct inode *inode, pgoff_t idx)
{
	return fsverity_active(inode) &&
	       idx < DIV_ROUND_UP(inode->i_size, PAGE_SIZE);
}

1000
static struct bio *f2fs_grab_read_bio(struct inode *inode, block_t blkaddr,
E
Eric Biggers 已提交
1001
				      unsigned nr_pages, unsigned op_flag,
1002 1003
				      pgoff_t first_idx, bool for_write,
				      bool for_verity)
1004 1005 1006
{
	struct f2fs_sb_info *sbi = F2FS_I_SB(inode);
	struct bio *bio;
1007 1008
	struct bio_post_read_ctx *ctx;
	unsigned int post_read_steps = 0;
1009

C
Christoph Hellwig 已提交
1010 1011 1012
	bio = bio_alloc_bioset(for_write ? GFP_NOIO : GFP_KERNEL,
			       min_t(int, nr_pages, BIO_MAX_PAGES),
			       &f2fs_bioset);
1013
	if (!bio)
1014
		return ERR_PTR(-ENOMEM);
1015 1016 1017

	f2fs_set_bio_crypt_ctx(bio, inode, first_idx, NULL, GFP_NOFS);

1018 1019
	f2fs_target_device(sbi, blkaddr, bio);
	bio->bi_end_io = f2fs_read_end_io;
1020
	bio_set_op_attrs(bio, REQ_OP_READ, op_flag);
1021

1022
	if (fscrypt_inode_uses_fs_layer_crypto(inode))
1023
		post_read_steps |= 1 << STEP_DECRYPT;
C
Chao Yu 已提交
1024
	if (f2fs_compressed_file(inode))
1025
		post_read_steps |= 1 << STEP_DECOMPRESS_NOWQ;
1026
	if (for_verity && f2fs_need_verity(inode, first_idx))
E
Eric Biggers 已提交
1027 1028
		post_read_steps |= 1 << STEP_VERITY;

1029
	if (post_read_steps) {
1030
		/* Due to the mempool, this never fails. */
1031 1032
		ctx = mempool_alloc(bio_post_read_ctx_pool, GFP_NOFS);
		ctx->bio = bio;
C
Chao Yu 已提交
1033
		ctx->sbi = sbi;
1034 1035 1036 1037
		ctx->enabled_steps = post_read_steps;
		bio->bi_private = ctx;
	}

1038 1039 1040
	return bio;
}

C
Chao Yu 已提交
1041 1042 1043 1044 1045 1046 1047
static void f2fs_release_read_bio(struct bio *bio)
{
	if (bio->bi_private)
		mempool_free(bio->bi_private, bio_post_read_ctx_pool);
	bio_put(bio);
}

1048 1049
/* This can handle encryption stuffs */
static int f2fs_submit_page_read(struct inode *inode, struct page *page,
1050
				 block_t blkaddr, int op_flags, bool for_write)
1051
{
C
Chao Yu 已提交
1052 1053
	struct f2fs_sb_info *sbi = F2FS_I_SB(inode);
	struct bio *bio;
1054

1055
	bio = f2fs_grab_read_bio(inode, blkaddr, 1, op_flags,
1056
					page->index, for_write, true);
1057 1058 1059
	if (IS_ERR(bio))
		return PTR_ERR(bio);

1060 1061 1062
	/* wait for GCed page writeback via META_MAPPING */
	f2fs_wait_on_block_writeback(inode, blkaddr);

1063 1064 1065 1066
	if (bio_add_page(bio, page, PAGE_SIZE, 0) < PAGE_SIZE) {
		bio_put(bio);
		return -EFAULT;
	}
1067
	ClearPageError(page);
C
Chao Yu 已提交
1068
	inc_page_count(sbi, F2FS_RD_DATA);
C
Chao Yu 已提交
1069
	f2fs_update_iostat(sbi, FS_DATA_READ_IO, F2FS_BLKSIZE);
C
Chao Yu 已提交
1070
	__submit_bio(sbi, bio, DATA);
1071 1072 1073
	return 0;
}

1074 1075 1076 1077
static void __set_data_blkaddr(struct dnode_of_data *dn)
{
	struct f2fs_node *rn = F2FS_NODE(dn->node_page);
	__le32 *addr_array;
1078 1079 1080 1081
	int base = 0;

	if (IS_INODE(dn->node_page) && f2fs_has_extra_attr(dn->inode))
		base = get_extra_isize(dn->inode);
1082 1083 1084

	/* Get physical address of data block */
	addr_array = blkaddr_in_node(rn);
1085
	addr_array[base + dn->ofs_in_node] = cpu_to_le32(dn->data_blkaddr);
1086 1087
}

J
Jaegeuk Kim 已提交
1088
/*
1089 1090 1091 1092 1093
 * Lock ordering for the change of data block address:
 * ->data_page
 *  ->node_page
 *    update block addresses in the node page
 */
C
Chao Yu 已提交
1094
void f2fs_set_data_blkaddr(struct dnode_of_data *dn)
1095
{
1096
	f2fs_wait_on_page_writeback(dn->node_page, NODE, true, true);
1097 1098
	__set_data_blkaddr(dn);
	if (set_page_dirty(dn->node_page))
1099
		dn->node_changed = true;
1100 1101
}

1102 1103 1104
void f2fs_update_data_blkaddr(struct dnode_of_data *dn, block_t blkaddr)
{
	dn->data_blkaddr = blkaddr;
C
Chao Yu 已提交
1105
	f2fs_set_data_blkaddr(dn);
1106 1107 1108
	f2fs_update_extent_cache(dn);
}

1109
/* dn->ofs_in_node will be returned with up-to-date last block pointer */
C
Chao Yu 已提交
1110
int f2fs_reserve_new_blocks(struct dnode_of_data *dn, blkcnt_t count)
1111
{
1112
	struct f2fs_sb_info *sbi = F2FS_I_SB(dn->inode);
C
Chao Yu 已提交
1113
	int err;
1114

1115 1116 1117
	if (!count)
		return 0;

1118
	if (unlikely(is_inode_flag_set(dn->inode, FI_NO_ALLOC)))
1119
		return -EPERM;
C
Chao Yu 已提交
1120 1121
	if (unlikely((err = inc_valid_block_count(sbi, dn->inode, &count))))
		return err;
1122

1123 1124 1125
	trace_f2fs_reserve_new_blocks(dn->inode, dn->nid,
						dn->ofs_in_node, count);

1126
	f2fs_wait_on_page_writeback(dn->node_page, NODE, true, true);
1127 1128

	for (; count > 0; dn->ofs_in_node++) {
1129
		block_t blkaddr = f2fs_data_blkaddr(dn);
1130 1131 1132 1133 1134 1135 1136 1137 1138
		if (blkaddr == NULL_ADDR) {
			dn->data_blkaddr = NEW_ADDR;
			__set_data_blkaddr(dn);
			count--;
		}
	}

	if (set_page_dirty(dn->node_page))
		dn->node_changed = true;
1139 1140 1141
	return 0;
}

1142
/* Should keep dn->ofs_in_node unchanged */
C
Chao Yu 已提交
1143
int f2fs_reserve_new_block(struct dnode_of_data *dn)
1144 1145 1146 1147
{
	unsigned int ofs_in_node = dn->ofs_in_node;
	int ret;

C
Chao Yu 已提交
1148
	ret = f2fs_reserve_new_blocks(dn, 1);
1149 1150 1151 1152
	dn->ofs_in_node = ofs_in_node;
	return ret;
}

1153 1154 1155 1156 1157
int f2fs_reserve_block(struct dnode_of_data *dn, pgoff_t index)
{
	bool need_put = dn->inode_page ? false : true;
	int err;

C
Chao Yu 已提交
1158
	err = f2fs_get_dnode_of_data(dn, index, ALLOC_NODE);
1159 1160
	if (err)
		return err;
1161

1162
	if (dn->data_blkaddr == NULL_ADDR)
C
Chao Yu 已提交
1163
		err = f2fs_reserve_new_block(dn);
1164
	if (err || need_put)
1165 1166 1167 1168
		f2fs_put_dnode(dn);
	return err;
}

1169
int f2fs_get_block(struct dnode_of_data *dn, pgoff_t index)
1170
{
J
Jack Qiu 已提交
1171
	struct extent_info ei = {0, 0, 0};
1172
	struct inode *inode = dn->inode;
1173

1174 1175 1176
	if (f2fs_lookup_extent_cache(inode, index, &ei)) {
		dn->data_blkaddr = ei.blk + index - ei.fofs;
		return 0;
1177
	}
1178

1179
	return f2fs_reserve_block(dn, index);
1180 1181
}

C
Chao Yu 已提交
1182
struct page *f2fs_get_read_data_page(struct inode *inode, pgoff_t index,
M
Mike Christie 已提交
1183
						int op_flags, bool for_write)
1184 1185 1186 1187
{
	struct address_space *mapping = inode->i_mapping;
	struct dnode_of_data dn;
	struct page *page;
1188
	struct extent_info ei = {0,0,0};
1189
	int err;
1190

1191
	page = f2fs_grab_cache_page(mapping, index, for_write);
1192 1193 1194
	if (!page)
		return ERR_PTR(-ENOMEM);

C
Chao Yu 已提交
1195 1196
	if (f2fs_lookup_extent_cache(inode, index, &ei)) {
		dn.data_blkaddr = ei.blk + index - ei.fofs;
C
Chao Yu 已提交
1197 1198
		if (!f2fs_is_valid_blkaddr(F2FS_I_SB(inode), dn.data_blkaddr,
						DATA_GENERIC_ENHANCE_READ)) {
1199
			err = -EFSCORRUPTED;
C
Chao Yu 已提交
1200 1201
			goto put_err;
		}
C
Chao Yu 已提交
1202 1203 1204
		goto got_it;
	}

1205
	set_new_dnode(&dn, inode, NULL, NULL, 0);
C
Chao Yu 已提交
1206
	err = f2fs_get_dnode_of_data(&dn, index, LOOKUP_NODE);
1207 1208
	if (err)
		goto put_err;
1209 1210
	f2fs_put_dnode(&dn);

1211
	if (unlikely(dn.data_blkaddr == NULL_ADDR)) {
1212 1213
		err = -ENOENT;
		goto put_err;
1214
	}
C
Chao Yu 已提交
1215 1216 1217 1218
	if (dn.data_blkaddr != NEW_ADDR &&
			!f2fs_is_valid_blkaddr(F2FS_I_SB(inode),
						dn.data_blkaddr,
						DATA_GENERIC_ENHANCE)) {
1219
		err = -EFSCORRUPTED;
C
Chao Yu 已提交
1220 1221
		goto put_err;
	}
C
Chao Yu 已提交
1222
got_it:
1223 1224
	if (PageUptodate(page)) {
		unlock_page(page);
1225
		return page;
1226
	}
1227

J
Jaegeuk Kim 已提交
1228 1229 1230 1231
	/*
	 * A new dentry page is allocated but not able to be written, since its
	 * new inode page couldn't be allocated due to -ENOSPC.
	 * In such the case, its blkaddr can be remained as NEW_ADDR.
C
Chao Yu 已提交
1232 1233
	 * see, f2fs_add_link -> f2fs_get_new_data_page ->
	 * f2fs_init_inode_metadata.
J
Jaegeuk Kim 已提交
1234 1235
	 */
	if (dn.data_blkaddr == NEW_ADDR) {
1236
		zero_user_segment(page, 0, PAGE_SIZE);
1237 1238
		if (!PageUptodate(page))
			SetPageUptodate(page);
1239
		unlock_page(page);
J
Jaegeuk Kim 已提交
1240 1241
		return page;
	}
1242

1243 1244
	err = f2fs_submit_page_read(inode, page, dn.data_blkaddr,
						op_flags, for_write);
1245
	if (err)
1246
		goto put_err;
1247
	return page;
1248 1249 1250 1251

put_err:
	f2fs_put_page(page, 1);
	return ERR_PTR(err);
1252 1253
}

C
Chao Yu 已提交
1254
struct page *f2fs_find_data_page(struct inode *inode, pgoff_t index)
1255 1256 1257 1258 1259 1260 1261 1262 1263
{
	struct address_space *mapping = inode->i_mapping;
	struct page *page;

	page = find_get_page(mapping, index);
	if (page && PageUptodate(page))
		return page;
	f2fs_put_page(page, 0);

C
Chao Yu 已提交
1264
	page = f2fs_get_read_data_page(inode, index, 0, false);
1265 1266 1267 1268 1269 1270 1271 1272 1273 1274 1275 1276 1277 1278 1279 1280 1281 1282 1283
	if (IS_ERR(page))
		return page;

	if (PageUptodate(page))
		return page;

	wait_on_page_locked(page);
	if (unlikely(!PageUptodate(page))) {
		f2fs_put_page(page, 0);
		return ERR_PTR(-EIO);
	}
	return page;
}

/*
 * If it tries to access a hole, return an error.
 * Because, the callers, functions in dir.c and GC, should be able to know
 * whether this page exists or not.
 */
C
Chao Yu 已提交
1284
struct page *f2fs_get_lock_data_page(struct inode *inode, pgoff_t index,
1285
							bool for_write)
1286 1287 1288 1289
{
	struct address_space *mapping = inode->i_mapping;
	struct page *page;
repeat:
C
Chao Yu 已提交
1290
	page = f2fs_get_read_data_page(inode, index, 0, for_write);
1291 1292
	if (IS_ERR(page))
		return page;
1293

1294
	/* wait for read completion */
1295
	lock_page(page);
1296
	if (unlikely(page->mapping != mapping)) {
1297 1298
		f2fs_put_page(page, 1);
		goto repeat;
1299
	}
1300 1301 1302 1303
	if (unlikely(!PageUptodate(page))) {
		f2fs_put_page(page, 1);
		return ERR_PTR(-EIO);
	}
1304 1305 1306
	return page;
}

J
Jaegeuk Kim 已提交
1307
/*
1308 1309
 * Caller ensures that this data page is never allocated.
 * A new zero-filled data page is allocated in the page cache.
1310
 *
C
Chao Yu 已提交
1311 1312
 * Also, caller should grab and release a rwsem by calling f2fs_lock_op() and
 * f2fs_unlock_op().
1313 1314
 * Note that, ipage is set only by make_empty_dir, and if any error occur,
 * ipage should be released by this function.
1315
 */
C
Chao Yu 已提交
1316
struct page *f2fs_get_new_data_page(struct inode *inode,
1317
		struct page *ipage, pgoff_t index, bool new_i_size)
1318 1319 1320 1321 1322
{
	struct address_space *mapping = inode->i_mapping;
	struct page *page;
	struct dnode_of_data dn;
	int err;
1323

1324
	page = f2fs_grab_cache_page(mapping, index, true);
1325 1326 1327 1328 1329 1330
	if (!page) {
		/*
		 * before exiting, we should make sure ipage will be released
		 * if any error occur.
		 */
		f2fs_put_page(ipage, 1);
1331
		return ERR_PTR(-ENOMEM);
1332
	}
1333

1334
	set_new_dnode(&dn, inode, ipage, NULL, 0);
1335
	err = f2fs_reserve_block(&dn, index);
1336 1337
	if (err) {
		f2fs_put_page(page, 1);
1338
		return ERR_PTR(err);
1339
	}
1340 1341
	if (!ipage)
		f2fs_put_dnode(&dn);
1342 1343

	if (PageUptodate(page))
1344
		goto got_it;
1345 1346

	if (dn.data_blkaddr == NEW_ADDR) {
1347
		zero_user_segment(page, 0, PAGE_SIZE);
1348 1349
		if (!PageUptodate(page))
			SetPageUptodate(page);
1350
	} else {
1351
		f2fs_put_page(page, 1);
1352

1353 1354
		/* if ipage exists, blkaddr should be NEW_ADDR */
		f2fs_bug_on(F2FS_I_SB(inode), ipage);
C
Chao Yu 已提交
1355
		page = f2fs_get_lock_data_page(inode, index, true);
1356
		if (IS_ERR(page))
1357
			return page;
1358
	}
1359
got_it:
C
Chao Yu 已提交
1360
	if (new_i_size && i_size_read(inode) <
1361
				((loff_t)(index + 1) << PAGE_SHIFT))
1362
		f2fs_i_size_write(inode, ((loff_t)(index + 1) << PAGE_SHIFT));
1363 1364 1365
	return page;
}

1366
static int __allocate_data_block(struct dnode_of_data *dn, int seg_type)
1367
{
1368
	struct f2fs_sb_info *sbi = F2FS_I_SB(dn->inode);
1369 1370
	struct f2fs_summary sum;
	struct node_info ni;
1371
	block_t old_blkaddr;
1372
	blkcnt_t count = 1;
C
Chao Yu 已提交
1373
	int err;
1374

1375
	if (unlikely(is_inode_flag_set(dn->inode, FI_NO_ALLOC)))
1376
		return -EPERM;
1377

1378 1379 1380 1381
	err = f2fs_get_node_info(sbi, dn->nid, &ni);
	if (err)
		return err;

1382
	dn->data_blkaddr = f2fs_data_blkaddr(dn);
1383
	if (dn->data_blkaddr != NULL_ADDR)
1384 1385
		goto alloc;

C
Chao Yu 已提交
1386 1387
	if (unlikely((err = inc_valid_block_count(sbi, dn->inode, &count))))
		return err;
1388

1389
alloc:
1390
	set_summary(&sum, dn->nid, dn->ofs_in_node, ni.version);
1391 1392
	old_blkaddr = dn->data_blkaddr;
	f2fs_allocate_data_block(sbi, NULL, old_blkaddr, &dn->data_blkaddr,
1393
				&sum, seg_type, NULL);
1394 1395 1396
	if (GET_SEGNO(sbi, old_blkaddr) != NULL_SEGNO)
		invalidate_mapping_pages(META_MAPPING(sbi),
					old_blkaddr, old_blkaddr);
1397
	f2fs_update_data_blkaddr(dn, dn->data_blkaddr);
1398

1399 1400 1401 1402
	/*
	 * i_size will be updated by direct_IO. Otherwise, we'll get stale
	 * data from unwritten block via dio_read.
	 */
1403 1404 1405
	return 0;
}

1406
int f2fs_preallocate_blocks(struct kiocb *iocb, struct iov_iter *from)
1407
{
1408
	struct inode *inode = file_inode(iocb->ki_filp);
C
Chao Yu 已提交
1409
	struct f2fs_map_blocks map;
1410
	int flag;
1411
	int err = 0;
1412
	bool direct_io = iocb->ki_flags & IOCB_DIRECT;
1413

1414
	map.m_lblk = F2FS_BLK_ALIGN(iocb->ki_pos);
1415 1416 1417 1418 1419 1420
	map.m_len = F2FS_BYTES_TO_BLK(iocb->ki_pos + iov_iter_count(from));
	if (map.m_len > map.m_lblk)
		map.m_len -= map.m_lblk;
	else
		map.m_len = 0;

1421
	map.m_next_pgofs = NULL;
1422
	map.m_next_extent = NULL;
1423
	map.m_seg_type = NO_CHECK_TYPE;
C
Chao Yu 已提交
1424
	map.m_may_create = true;
1425

1426
	if (direct_io) {
C
Chao Yu 已提交
1427
		map.m_seg_type = f2fs_rw_hint_to_seg_type(iocb->ki_hint);
1428
		flag = f2fs_force_buffered_io(inode, iocb, from) ?
1429 1430 1431
					F2FS_GET_BLOCK_PRE_AIO :
					F2FS_GET_BLOCK_PRE_DIO;
		goto map_blocks;
1432
	}
C
Chao Yu 已提交
1433
	if (iocb->ki_pos + iov_iter_count(from) > MAX_INLINE_DATA(inode)) {
1434 1435 1436
		err = f2fs_convert_inline_inode(inode);
		if (err)
			return err;
1437
	}
1438
	if (f2fs_has_inline_data(inode))
1439
		return err;
1440 1441 1442 1443 1444 1445 1446 1447 1448

	flag = F2FS_GET_BLOCK_PRE_AIO;

map_blocks:
	err = f2fs_map_blocks(inode, &map, 1, flag);
	if (map.m_len > 0 && err == -ENOSPC) {
		if (!direct_io)
			set_inode_flag(inode, FI_NO_PREALLOC);
		err = 0;
1449
	}
1450
	return err;
1451 1452
}

C
Chao Yu 已提交
1453
void f2fs_do_map_lock(struct f2fs_sb_info *sbi, int flag, bool lock)
1454 1455 1456 1457 1458 1459 1460 1461 1462 1463 1464 1465 1466 1467
{
	if (flag == F2FS_GET_BLOCK_PRE_AIO) {
		if (lock)
			down_read(&sbi->node_change);
		else
			up_read(&sbi->node_change);
	} else {
		if (lock)
			f2fs_lock_op(sbi);
		else
			f2fs_unlock_op(sbi);
	}
}

J
Jaegeuk Kim 已提交
1468
/*
C
Chao Yu 已提交
1469 1470 1471
 * f2fs_map_blocks() tries to find or build mapping relationship which
 * maps continuous logical blocks to physical blocks, and return such
 * info via f2fs_map_blocks structure.
1472
 */
C
Chao Yu 已提交
1473
int f2fs_map_blocks(struct inode *inode, struct f2fs_map_blocks *map,
C
Chao Yu 已提交
1474
						int create, int flag)
1475
{
J
Jaegeuk Kim 已提交
1476
	unsigned int maxblocks = map->m_len;
1477
	struct dnode_of_data dn;
1478
	struct f2fs_sb_info *sbi = F2FS_I_SB(inode);
C
Chao Yu 已提交
1479
	int mode = map->m_may_create ? ALLOC_NODE : LOOKUP_NODE;
1480
	pgoff_t pgofs, end_offset, end;
1481
	int err = 0, ofs = 1;
1482 1483
	unsigned int ofs_in_node, last_ofs_in_node;
	blkcnt_t prealloc;
1484
	struct extent_info ei = {0,0,0};
1485
	block_t blkaddr;
1486
	unsigned int start_pgofs;
1487

1488 1489 1490
	if (!maxblocks)
		return 0;

J
Jaegeuk Kim 已提交
1491 1492 1493 1494 1495
	map->m_len = 0;
	map->m_flags = 0;

	/* it only supports block size == page size */
	pgofs =	(pgoff_t)map->m_lblk;
1496
	end = pgofs + maxblocks;
1497

1498
	if (!create && f2fs_lookup_extent_cache(inode, pgofs, &ei)) {
1499
		if (f2fs_lfs_mode(sbi) && flag == F2FS_GET_BLOCK_DIO &&
1500 1501 1502
							map->m_may_create)
			goto next_dnode;

J
Jaegeuk Kim 已提交
1503 1504 1505
		map->m_pblk = ei.blk + pgofs - ei.fofs;
		map->m_len = min((pgoff_t)maxblocks, ei.fofs + ei.len - pgofs);
		map->m_flags = F2FS_MAP_MAPPED;
1506 1507
		if (map->m_next_extent)
			*map->m_next_extent = pgofs + map->m_len;
1508 1509 1510 1511 1512

		/* for hardware encryption, but to avoid potential issue in future */
		if (flag == F2FS_GET_BLOCK_DIO)
			f2fs_wait_on_block_writeback_range(inode,
						map->m_pblk, map->m_len);
1513
		goto out;
1514
	}
1515

C
Chao Yu 已提交
1516
next_dnode:
C
Chao Yu 已提交
1517
	if (map->m_may_create)
C
Chao Yu 已提交
1518
		f2fs_do_map_lock(sbi, flag, true);
1519 1520 1521

	/* When reading holes, we need its node page */
	set_new_dnode(&dn, inode, NULL, NULL, 0);
C
Chao Yu 已提交
1522
	err = f2fs_get_dnode_of_data(&dn, pgofs, mode);
1523
	if (err) {
C
Chao Yu 已提交
1524 1525
		if (flag == F2FS_GET_BLOCK_BMAP)
			map->m_pblk = 0;
1526
		if (err == -ENOENT) {
1527
			err = 0;
1528 1529
			if (map->m_next_pgofs)
				*map->m_next_pgofs =
C
Chao Yu 已提交
1530
					f2fs_get_next_page_offset(&dn, pgofs);
1531 1532
			if (map->m_next_extent)
				*map->m_next_extent =
C
Chao Yu 已提交
1533
					f2fs_get_next_page_offset(&dn, pgofs);
1534
		}
1535
		goto unlock_out;
1536
	}
C
Chao Yu 已提交
1537

1538
	start_pgofs = pgofs;
1539
	prealloc = 0;
1540
	last_ofs_in_node = ofs_in_node = dn.ofs_in_node;
1541
	end_offset = ADDRS_PER_PAGE(dn.node_page, inode);
C
Chao Yu 已提交
1542 1543

next_block:
1544
	blkaddr = f2fs_data_blkaddr(&dn);
C
Chao Yu 已提交
1545

1546
	if (__is_valid_data_blkaddr(blkaddr) &&
C
Chao Yu 已提交
1547
		!f2fs_is_valid_blkaddr(sbi, blkaddr, DATA_GENERIC_ENHANCE)) {
1548
		err = -EFSCORRUPTED;
1549 1550 1551
		goto sync_out;
	}

C
Chao Yu 已提交
1552
	if (__is_valid_data_blkaddr(blkaddr)) {
1553
		/* use out-place-update for driect IO under LFS mode */
1554
		if (f2fs_lfs_mode(sbi) && flag == F2FS_GET_BLOCK_DIO &&
C
Chao Yu 已提交
1555
							map->m_may_create) {
1556
			err = __allocate_data_block(&dn, map->m_seg_type);
1557 1558 1559 1560
			if (err)
				goto sync_out;
			blkaddr = dn.data_blkaddr;
			set_inode_flag(inode, FI_APPEND_WRITE);
1561 1562
		}
	} else {
C
Chao Yu 已提交
1563
		if (create) {
1564 1565
			if (unlikely(f2fs_cp_error(sbi))) {
				err = -EIO;
C
Chao Yu 已提交
1566
				goto sync_out;
1567
			}
1568
			if (flag == F2FS_GET_BLOCK_PRE_AIO) {
1569 1570 1571 1572
				if (blkaddr == NULL_ADDR) {
					prealloc++;
					last_ofs_in_node = dn.ofs_in_node;
				}
1573
			} else {
1574 1575
				WARN_ON(flag != F2FS_GET_BLOCK_PRE_DIO &&
					flag != F2FS_GET_BLOCK_DIO);
1576 1577
				err = __allocate_data_block(&dn,
							map->m_seg_type);
1578
				if (!err)
1579
					set_inode_flag(inode, FI_APPEND_WRITE);
1580
			}
C
Chao Yu 已提交
1581
			if (err)
C
Chao Yu 已提交
1582
				goto sync_out;
1583
			map->m_flags |= F2FS_MAP_NEW;
C
Chao Yu 已提交
1584
			blkaddr = dn.data_blkaddr;
C
Chao Yu 已提交
1585
		} else {
C
Chao Yu 已提交
1586 1587 1588 1589
			if (flag == F2FS_GET_BLOCK_BMAP) {
				map->m_pblk = 0;
				goto sync_out;
			}
1590 1591
			if (flag == F2FS_GET_BLOCK_PRECACHE)
				goto sync_out;
1592 1593 1594 1595
			if (flag == F2FS_GET_BLOCK_FIEMAP &&
						blkaddr == NULL_ADDR) {
				if (map->m_next_pgofs)
					*map->m_next_pgofs = pgofs + 1;
C
Chao Yu 已提交
1596
				goto sync_out;
1597
			}
1598 1599 1600 1601
			if (flag != F2FS_GET_BLOCK_FIEMAP) {
				/* for defragment case */
				if (map->m_next_pgofs)
					*map->m_next_pgofs = pgofs + 1;
C
Chao Yu 已提交
1602
				goto sync_out;
1603
			}
C
Chao Yu 已提交
1604 1605
		}
	}
1606

1607 1608 1609
	if (flag == F2FS_GET_BLOCK_PRE_AIO)
		goto skip;

C
Chao Yu 已提交
1610 1611 1612 1613 1614 1615 1616 1617 1618 1619
	if (map->m_len == 0) {
		/* preallocated unwritten block should be mapped for fiemap. */
		if (blkaddr == NEW_ADDR)
			map->m_flags |= F2FS_MAP_UNWRITTEN;
		map->m_flags |= F2FS_MAP_MAPPED;

		map->m_pblk = blkaddr;
		map->m_len = 1;
	} else if ((map->m_pblk != NEW_ADDR &&
			blkaddr == (map->m_pblk + ofs)) ||
1620
			(map->m_pblk == NEW_ADDR && blkaddr == NEW_ADDR) ||
1621
			flag == F2FS_GET_BLOCK_PRE_DIO) {
C
Chao Yu 已提交
1622 1623 1624 1625 1626
		ofs++;
		map->m_len++;
	} else {
		goto sync_out;
	}
1627

1628
skip:
1629 1630 1631
	dn.ofs_in_node++;
	pgofs++;

1632 1633 1634
	/* preallocate blocks in batch for one dnode page */
	if (flag == F2FS_GET_BLOCK_PRE_AIO &&
			(pgofs == end || dn.ofs_in_node == end_offset)) {
1635

1636
		dn.ofs_in_node = ofs_in_node;
C
Chao Yu 已提交
1637
		err = f2fs_reserve_new_blocks(&dn, prealloc);
1638 1639
		if (err)
			goto sync_out;
1640

1641 1642 1643 1644
		map->m_len += dn.ofs_in_node - ofs_in_node;
		if (prealloc && dn.ofs_in_node != last_ofs_in_node + 1) {
			err = -ENOSPC;
			goto sync_out;
1645
		}
1646 1647 1648 1649 1650 1651 1652 1653
		dn.ofs_in_node = end_offset;
	}

	if (pgofs >= end)
		goto sync_out;
	else if (dn.ofs_in_node < end_offset)
		goto next_block;

1654 1655 1656 1657 1658 1659 1660 1661 1662 1663
	if (flag == F2FS_GET_BLOCK_PRECACHE) {
		if (map->m_flags & F2FS_MAP_MAPPED) {
			unsigned int ofs = start_pgofs - map->m_lblk;

			f2fs_update_extent_cache_range(&dn,
				start_pgofs, map->m_pblk + ofs,
				map->m_len - ofs);
		}
	}

1664 1665
	f2fs_put_dnode(&dn);

C
Chao Yu 已提交
1666
	if (map->m_may_create) {
C
Chao Yu 已提交
1667
		f2fs_do_map_lock(sbi, flag, false);
1668
		f2fs_balance_fs(sbi, dn.node_changed);
1669
	}
1670
	goto next_dnode;
1671

1672
sync_out:
1673 1674 1675 1676 1677 1678

	/* for hardware encryption, but to avoid potential issue in future */
	if (flag == F2FS_GET_BLOCK_DIO && map->m_flags & F2FS_MAP_MAPPED)
		f2fs_wait_on_block_writeback_range(inode,
						map->m_pblk, map->m_len);

1679 1680 1681 1682 1683 1684 1685 1686 1687 1688 1689
	if (flag == F2FS_GET_BLOCK_PRECACHE) {
		if (map->m_flags & F2FS_MAP_MAPPED) {
			unsigned int ofs = start_pgofs - map->m_lblk;

			f2fs_update_extent_cache_range(&dn,
				start_pgofs, map->m_pblk + ofs,
				map->m_len - ofs);
		}
		if (map->m_next_extent)
			*map->m_next_extent = pgofs + 1;
	}
1690
	f2fs_put_dnode(&dn);
1691
unlock_out:
C
Chao Yu 已提交
1692
	if (map->m_may_create) {
C
Chao Yu 已提交
1693
		f2fs_do_map_lock(sbi, flag, false);
1694
		f2fs_balance_fs(sbi, dn.node_changed);
1695
	}
1696
out:
J
Jaegeuk Kim 已提交
1697
	trace_f2fs_map_blocks(inode, map, err);
1698
	return err;
1699 1700
}

H
Hyunchul Lee 已提交
1701 1702 1703 1704 1705 1706 1707 1708 1709 1710 1711 1712 1713
bool f2fs_overwrite_io(struct inode *inode, loff_t pos, size_t len)
{
	struct f2fs_map_blocks map;
	block_t last_lblk;
	int err;

	if (pos + len > i_size_read(inode))
		return false;

	map.m_lblk = F2FS_BYTES_TO_BLK(pos);
	map.m_next_pgofs = NULL;
	map.m_next_extent = NULL;
	map.m_seg_type = NO_CHECK_TYPE;
1714
	map.m_may_create = false;
H
Hyunchul Lee 已提交
1715 1716 1717 1718 1719 1720 1721 1722 1723 1724 1725 1726
	last_lblk = F2FS_BLK_ALIGN(pos + len);

	while (map.m_lblk < last_lblk) {
		map.m_len = last_lblk - map.m_lblk;
		err = f2fs_map_blocks(inode, &map, 0, F2FS_GET_BLOCK_DEFAULT);
		if (err || map.m_len == 0)
			return false;
		map.m_lblk += map.m_len;
	}
	return true;
}

1727 1728 1729 1730 1731 1732 1733 1734 1735 1736
static inline u64 bytes_to_blks(struct inode *inode, u64 bytes)
{
	return (bytes >> inode->i_blkbits);
}

static inline u64 blks_to_bytes(struct inode *inode, u64 blks)
{
	return (blks << inode->i_blkbits);
}

J
Jaegeuk Kim 已提交
1737
static int __get_data_block(struct inode *inode, sector_t iblock,
1738
			struct buffer_head *bh, int create, int flag,
C
Chao Yu 已提交
1739
			pgoff_t *next_pgofs, int seg_type, bool may_write)
J
Jaegeuk Kim 已提交
1740 1741
{
	struct f2fs_map_blocks map;
1742
	int err;
J
Jaegeuk Kim 已提交
1743 1744

	map.m_lblk = iblock;
1745
	map.m_len = bytes_to_blks(inode, bh->b_size);
1746
	map.m_next_pgofs = next_pgofs;
1747
	map.m_next_extent = NULL;
1748
	map.m_seg_type = seg_type;
C
Chao Yu 已提交
1749
	map.m_may_create = may_write;
J
Jaegeuk Kim 已提交
1750

1751 1752
	err = f2fs_map_blocks(inode, &map, create, flag);
	if (!err) {
J
Jaegeuk Kim 已提交
1753 1754
		map_bh(bh, inode->i_sb, map.m_pblk);
		bh->b_state = (bh->b_state & ~F2FS_MAP_FLAGS) | map.m_flags;
1755
		bh->b_size = blks_to_bytes(inode, map.m_len);
J
Jaegeuk Kim 已提交
1756
	}
1757
	return err;
J
Jaegeuk Kim 已提交
1758 1759
}

C
Chao Yu 已提交
1760 1761 1762 1763 1764 1765
static int get_data_block_dio_write(struct inode *inode, sector_t iblock,
			struct buffer_head *bh_result, int create)
{
	return __get_data_block(inode, iblock, bh_result, create,
				F2FS_GET_BLOCK_DIO, NULL,
				f2fs_rw_hint_to_seg_type(inode->i_write_hint),
1766
				IS_SWAPFILE(inode) ? false : true);
C
Chao Yu 已提交
1767 1768 1769
}

static int get_data_block_dio(struct inode *inode, sector_t iblock,
1770 1771
			struct buffer_head *bh_result, int create)
{
C
Chao Yu 已提交
1772
	return __get_data_block(inode, iblock, bh_result, create,
C
Chao Yu 已提交
1773 1774 1775
				F2FS_GET_BLOCK_DIO, NULL,
				f2fs_rw_hint_to_seg_type(inode->i_write_hint),
				false);
1776 1777
}

C
Chao Yu 已提交
1778 1779 1780 1781 1782 1783 1784 1785 1786 1787 1788 1789 1790 1791 1792 1793 1794 1795 1796
static int f2fs_xattr_fiemap(struct inode *inode,
				struct fiemap_extent_info *fieinfo)
{
	struct f2fs_sb_info *sbi = F2FS_I_SB(inode);
	struct page *page;
	struct node_info ni;
	__u64 phys = 0, len;
	__u32 flags;
	nid_t xnid = F2FS_I(inode)->i_xattr_nid;
	int err = 0;

	if (f2fs_has_inline_xattr(inode)) {
		int offset;

		page = f2fs_grab_cache_page(NODE_MAPPING(sbi),
						inode->i_ino, false);
		if (!page)
			return -ENOMEM;

1797 1798 1799 1800 1801
		err = f2fs_get_node_info(sbi, inode->i_ino, &ni);
		if (err) {
			f2fs_put_page(page, 1);
			return err;
		}
C
Chao Yu 已提交
1802

1803
		phys = blks_to_bytes(inode, ni.blk_addr);
C
Chao Yu 已提交
1804 1805
		offset = offsetof(struct f2fs_inode, i_addr) +
					sizeof(__le32) * (DEF_ADDRS_PER_INODE -
1806
					get_inline_xattr_addrs(inode));
C
Chao Yu 已提交
1807 1808 1809 1810 1811 1812 1813 1814 1815 1816 1817 1818

		phys += offset;
		len = inline_xattr_size(inode);

		f2fs_put_page(page, 1);

		flags = FIEMAP_EXTENT_DATA_INLINE | FIEMAP_EXTENT_NOT_ALIGNED;

		if (!xnid)
			flags |= FIEMAP_EXTENT_LAST;

		err = fiemap_fill_next_extent(fieinfo, 0, phys, len, flags);
C
Chao Yu 已提交
1819
		trace_f2fs_fiemap(inode, 0, phys, len, flags, err);
C
Chao Yu 已提交
1820 1821 1822 1823 1824 1825 1826 1827 1828
		if (err || err == 1)
			return err;
	}

	if (xnid) {
		page = f2fs_grab_cache_page(NODE_MAPPING(sbi), xnid, false);
		if (!page)
			return -ENOMEM;

1829 1830 1831 1832 1833
		err = f2fs_get_node_info(sbi, xnid, &ni);
		if (err) {
			f2fs_put_page(page, 1);
			return err;
		}
C
Chao Yu 已提交
1834

1835
		phys = blks_to_bytes(inode, ni.blk_addr);
C
Chao Yu 已提交
1836 1837 1838 1839 1840 1841 1842
		len = inode->i_sb->s_blocksize;

		f2fs_put_page(page, 1);

		flags = FIEMAP_EXTENT_LAST;
	}

C
Chao Yu 已提交
1843
	if (phys) {
C
Chao Yu 已提交
1844
		err = fiemap_fill_next_extent(fieinfo, 0, phys, len, flags);
C
Chao Yu 已提交
1845 1846
		trace_f2fs_fiemap(inode, 0, phys, len, flags, err);
	}
C
Chao Yu 已提交
1847 1848 1849 1850

	return (err < 0 ? err : 0);
}

1851 1852 1853 1854 1855 1856 1857 1858 1859 1860 1861 1862 1863 1864 1865 1866 1867 1868 1869
static loff_t max_inode_blocks(struct inode *inode)
{
	loff_t result = ADDRS_PER_INODE(inode);
	loff_t leaf_count = ADDRS_PER_BLOCK(inode);

	/* two direct node blocks */
	result += (leaf_count * 2);

	/* two indirect node blocks */
	leaf_count *= NIDS_PER_BLOCK;
	result += (leaf_count * 2);

	/* one double indirect node block */
	leaf_count *= NIDS_PER_BLOCK;
	result += leaf_count;

	return result;
}

J
Jaegeuk Kim 已提交
1870 1871 1872
int f2fs_fiemap(struct inode *inode, struct fiemap_extent_info *fieinfo,
		u64 start, u64 len)
{
1873
	struct f2fs_map_blocks map;
1874
	sector_t start_blk, last_blk;
1875
	pgoff_t next_pgofs;
1876 1877 1878
	u64 logical = 0, phys = 0, size = 0;
	u32 flags = 0;
	int ret = 0;
1879 1880
	bool compr_cluster = false;
	unsigned int cluster_size = F2FS_I(inode)->i_cluster_size;
1881

1882 1883 1884 1885 1886 1887
	if (fieinfo->fi_flags & FIEMAP_FLAG_CACHE) {
		ret = f2fs_precache_extents(inode);
		if (ret)
			return ret;
	}

1888
	ret = fiemap_prep(inode, fieinfo, start, &len, FIEMAP_FLAG_XATTR);
1889 1890 1891
	if (ret)
		return ret;

1892 1893
	inode_lock(inode);

C
Chao Yu 已提交
1894 1895 1896 1897 1898
	if (fieinfo->fi_flags & FIEMAP_FLAG_XATTR) {
		ret = f2fs_xattr_fiemap(inode, fieinfo);
		goto out;
	}

1899
	if (f2fs_has_inline_data(inode) || f2fs_has_inline_dentry(inode)) {
J
Jaegeuk Kim 已提交
1900 1901
		ret = f2fs_inline_data_fiemap(inode, fieinfo, start, len);
		if (ret != -EAGAIN)
1902
			goto out;
J
Jaegeuk Kim 已提交
1903 1904
	}

1905 1906
	if (bytes_to_blks(inode, len) == 0)
		len = blks_to_bytes(inode, 1);
1907

1908 1909
	start_blk = bytes_to_blks(inode, start);
	last_blk = bytes_to_blks(inode, start + len - 1);
1910

1911
next:
1912 1913 1914 1915 1916
	memset(&map, 0, sizeof(map));
	map.m_lblk = start_blk;
	map.m_len = bytes_to_blks(inode, len);
	map.m_next_pgofs = &next_pgofs;
	map.m_seg_type = NO_CHECK_TYPE;
1917

1918
	if (compr_cluster)
1919
		map.m_len = cluster_size - 1;
1920

1921
	ret = f2fs_map_blocks(inode, &map, 0, F2FS_GET_BLOCK_FIEMAP);
1922 1923 1924 1925
	if (ret)
		goto out;

	/* HOLE */
1926
	if (!(map.m_flags & F2FS_MAP_FLAGS)) {
1927
		start_blk = next_pgofs;
1928

1929
		if (blks_to_bytes(inode, start_blk) < blks_to_bytes(inode,
1930
						max_inode_blocks(inode)))
1931
			goto prep_next;
1932

1933 1934
		flags |= FIEMAP_EXTENT_LAST;
	}
1935

1936
	if (size) {
1937
		if (IS_ENCRYPTED(inode))
1938 1939
			flags |= FIEMAP_EXTENT_DATA_ENCRYPTED;

1940 1941
		ret = fiemap_fill_next_extent(fieinfo, logical,
				phys, size, flags);
C
Chao Yu 已提交
1942
		trace_f2fs_fiemap(inode, logical, phys, size, flags, ret);
1943 1944 1945
		if (ret)
			goto out;
		size = 0;
1946
	}
1947

1948
	if (start_blk > last_blk)
1949
		goto out;
1950

1951 1952 1953 1954
	if (compr_cluster) {
		compr_cluster = false;


1955
		logical = blks_to_bytes(inode, start_blk - 1);
1956
		phys = blks_to_bytes(inode, map.m_pblk);
1957
		size = blks_to_bytes(inode, cluster_size);
1958 1959 1960 1961 1962 1963 1964 1965 1966 1967 1968

		flags |= FIEMAP_EXTENT_ENCODED;

		start_blk += cluster_size - 1;

		if (start_blk > last_blk)
			goto out;

		goto prep_next;
	}

1969
	if (map.m_pblk == COMPRESS_ADDR) {
1970 1971 1972 1973 1974
		compr_cluster = true;
		start_blk++;
		goto prep_next;
	}

1975
	logical = blks_to_bytes(inode, start_blk);
1976 1977
	phys = blks_to_bytes(inode, map.m_pblk);
	size = blks_to_bytes(inode, map.m_len);
1978
	flags = 0;
1979
	if (map.m_flags & F2FS_MAP_UNWRITTEN)
1980
		flags = FIEMAP_EXTENT_UNWRITTEN;
1981

1982
	start_blk += bytes_to_blks(inode, size);
1983

1984
prep_next:
1985 1986 1987 1988 1989 1990 1991 1992 1993
	cond_resched();
	if (fatal_signal_pending(current))
		ret = -EINTR;
	else
		goto next;
out:
	if (ret == 1)
		ret = 0;

A
Al Viro 已提交
1994
	inode_unlock(inode);
1995
	return ret;
J
Jaegeuk Kim 已提交
1996 1997
}

E
Eric Biggers 已提交
1998 1999 2000 2001 2002 2003 2004 2005 2006
static inline loff_t f2fs_readpage_limit(struct inode *inode)
{
	if (IS_ENABLED(CONFIG_FS_VERITY) &&
	    (IS_VERITY(inode) || f2fs_verity_in_progress(inode)))
		return inode->i_sb->s_maxbytes;

	return i_size_read(inode);
}

2007 2008 2009 2010 2011 2012 2013 2014
static int f2fs_read_single_page(struct inode *inode, struct page *page,
					unsigned nr_pages,
					struct f2fs_map_blocks *map,
					struct bio **bio_ret,
					sector_t *last_block_in_bio,
					bool is_readahead)
{
	struct bio *bio = *bio_ret;
2015
	const unsigned blocksize = blks_to_bytes(inode, 1);
2016 2017 2018 2019 2020 2021
	sector_t block_in_file;
	sector_t last_block;
	sector_t last_block_in_file;
	sector_t block_nr;
	int ret = 0;

J
Jaegeuk Kim 已提交
2022
	block_in_file = (sector_t)page_index(page);
2023
	last_block = block_in_file + nr_pages;
2024 2025
	last_block_in_file = bytes_to_blks(inode,
			f2fs_readpage_limit(inode) + blocksize - 1);
2026 2027 2028 2029 2030 2031 2032 2033 2034 2035 2036 2037 2038 2039 2040 2041 2042 2043 2044 2045 2046 2047 2048 2049 2050 2051 2052 2053 2054
	if (last_block > last_block_in_file)
		last_block = last_block_in_file;

	/* just zeroing out page which is beyond EOF */
	if (block_in_file >= last_block)
		goto zero_out;
	/*
	 * Map blocks using the previous result first.
	 */
	if ((map->m_flags & F2FS_MAP_MAPPED) &&
			block_in_file > map->m_lblk &&
			block_in_file < (map->m_lblk + map->m_len))
		goto got_it;

	/*
	 * Then do more f2fs_map_blocks() calls until we are
	 * done with this page.
	 */
	map->m_lblk = block_in_file;
	map->m_len = last_block - block_in_file;

	ret = f2fs_map_blocks(inode, map, 0, F2FS_GET_BLOCK_DEFAULT);
	if (ret)
		goto out;
got_it:
	if ((map->m_flags & F2FS_MAP_MAPPED)) {
		block_nr = map->m_pblk + block_in_file - map->m_lblk;
		SetPageMappedToDisk(page);

J
Jaegeuk Kim 已提交
2055 2056
		if (!PageUptodate(page) && (!PageSwapCache(page) &&
					!cleancache_get_page(page))) {
2057 2058 2059 2060 2061
			SetPageUptodate(page);
			goto confused;
		}

		if (!f2fs_is_valid_blkaddr(F2FS_I_SB(inode), block_nr,
C
Chao Yu 已提交
2062
						DATA_GENERIC_ENHANCE_READ)) {
2063
			ret = -EFSCORRUPTED;
2064 2065 2066 2067 2068
			goto out;
		}
	} else {
zero_out:
		zero_user_segment(page, 0, PAGE_SIZE);
E
Eric Biggers 已提交
2069 2070 2071 2072 2073
		if (f2fs_need_verity(inode, page->index) &&
		    !fsverity_verify_page(page)) {
			ret = -EIO;
			goto out;
		}
2074 2075 2076 2077 2078 2079 2080 2081 2082 2083
		if (!PageUptodate(page))
			SetPageUptodate(page);
		unlock_page(page);
		goto out;
	}

	/*
	 * This page will go to BIO.  Do we need to send this
	 * BIO off first?
	 */
2084 2085 2086
	if (bio && (!page_is_mergeable(F2FS_I_SB(inode), bio,
				       *last_block_in_bio, block_nr) ||
		    !f2fs_crypt_mergeable_bio(bio, inode, page->index, NULL))) {
2087 2088 2089 2090 2091 2092
submit_and_realloc:
		__submit_bio(F2FS_I_SB(inode), bio, DATA);
		bio = NULL;
	}
	if (bio == NULL) {
		bio = f2fs_grab_read_bio(inode, block_nr, nr_pages,
2093
				is_readahead ? REQ_RAHEAD : 0, page->index,
2094
				false, true);
2095 2096 2097 2098 2099 2100 2101 2102 2103 2104 2105 2106 2107 2108 2109 2110 2111
		if (IS_ERR(bio)) {
			ret = PTR_ERR(bio);
			bio = NULL;
			goto out;
		}
	}

	/*
	 * If the page is under writeback, we need to wait for
	 * its completion to see the correct decrypted data.
	 */
	f2fs_wait_on_block_writeback(inode, block_nr);

	if (bio_add_page(bio, page, blocksize, 0) < blocksize)
		goto submit_and_realloc;

	inc_page_count(F2FS_I_SB(inode), F2FS_RD_DATA);
C
Chao Yu 已提交
2112
	f2fs_update_iostat(F2FS_I_SB(inode), FS_DATA_READ_IO, F2FS_BLKSIZE);
2113 2114 2115 2116 2117 2118 2119 2120 2121 2122 2123 2124 2125 2126
	ClearPageError(page);
	*last_block_in_bio = block_nr;
	goto out;
confused:
	if (bio) {
		__submit_bio(F2FS_I_SB(inode), bio, DATA);
		bio = NULL;
	}
	unlock_page(page);
out:
	*bio_ret = bio;
	return ret;
}

C
Chao Yu 已提交
2127 2128 2129
#ifdef CONFIG_F2FS_FS_COMPRESSION
int f2fs_read_multi_pages(struct compress_ctx *cc, struct bio **bio_ret,
				unsigned nr_pages, sector_t *last_block_in_bio,
2130
				bool is_readahead, bool for_write)
C
Chao Yu 已提交
2131 2132 2133 2134 2135 2136 2137
{
	struct dnode_of_data dn;
	struct inode *inode = cc->inode;
	struct f2fs_sb_info *sbi = F2FS_I_SB(inode);
	struct bio *bio = *bio_ret;
	unsigned int start_idx = cc->cluster_idx << cc->log_cluster_size;
	sector_t last_block_in_file;
2138
	const unsigned blocksize = blks_to_bytes(inode, 1);
C
Chao Yu 已提交
2139
	struct decompress_io_ctx *dic = NULL;
2140 2141
	struct bio_post_read_ctx *ctx;
	bool for_verity = false;
C
Chao Yu 已提交
2142 2143 2144 2145 2146
	int i;
	int ret = 0;

	f2fs_bug_on(sbi, f2fs_cluster_is_empty(cc));

2147 2148
	last_block_in_file = bytes_to_blks(inode,
			f2fs_readpage_limit(inode) + blocksize - 1);
C
Chao Yu 已提交
2149 2150 2151 2152 2153 2154 2155 2156 2157 2158 2159 2160 2161 2162 2163 2164 2165 2166 2167 2168 2169 2170 2171 2172 2173 2174 2175 2176

	/* get rid of pages beyond EOF */
	for (i = 0; i < cc->cluster_size; i++) {
		struct page *page = cc->rpages[i];

		if (!page)
			continue;
		if ((sector_t)page->index >= last_block_in_file) {
			zero_user_segment(page, 0, PAGE_SIZE);
			if (!PageUptodate(page))
				SetPageUptodate(page);
		} else if (!PageUptodate(page)) {
			continue;
		}
		unlock_page(page);
		cc->rpages[i] = NULL;
		cc->nr_rpages--;
	}

	/* we are done since all pages are beyond EOF */
	if (f2fs_cluster_is_empty(cc))
		goto out;

	set_new_dnode(&dn, inode, NULL, NULL, 0);
	ret = f2fs_get_dnode_of_data(&dn, start_idx, LOOKUP_NODE);
	if (ret)
		goto out;

2177
	f2fs_bug_on(sbi, dn.data_blkaddr != COMPRESS_ADDR);
C
Chao Yu 已提交
2178 2179 2180 2181

	for (i = 1; i < cc->cluster_size; i++) {
		block_t blkaddr;

2182
		blkaddr = data_blkaddr(dn.inode, dn.node_page,
C
Chao Yu 已提交
2183 2184 2185 2186 2187 2188 2189 2190 2191 2192 2193 2194 2195 2196 2197 2198 2199 2200 2201 2202 2203 2204 2205 2206
						dn.ofs_in_node + i);

		if (!__is_valid_data_blkaddr(blkaddr))
			break;

		if (!f2fs_is_valid_blkaddr(sbi, blkaddr, DATA_GENERIC)) {
			ret = -EFAULT;
			goto out_put_dnode;
		}
		cc->nr_cpages++;
	}

	/* nothing to decompress */
	if (cc->nr_cpages == 0) {
		ret = 0;
		goto out_put_dnode;
	}

	dic = f2fs_alloc_dic(cc);
	if (IS_ERR(dic)) {
		ret = PTR_ERR(dic);
		goto out_put_dnode;
	}

2207 2208 2209 2210 2211 2212 2213 2214 2215 2216 2217 2218 2219 2220 2221 2222 2223 2224 2225 2226
	/*
	 * It's possible to enable fsverity on the fly when handling a cluster,
	 * which requires complicated error handling. Instead of adding more
	 * complexity, let's give a rule where end_io post-processes fsverity
	 * per cluster. In order to do that, we need to submit bio, if previous
	 * bio sets a different post-process policy.
	 */
	if (fsverity_active(cc->inode)) {
		atomic_set(&dic->verity_pages, cc->nr_cpages);
		for_verity = true;

		if (bio) {
			ctx = bio->bi_private;
			if (!(ctx->enabled_steps & (1 << STEP_VERITY))) {
				__submit_bio(sbi, bio, DATA);
				bio = NULL;
			}
		}
	}

C
Chao Yu 已提交
2227 2228 2229 2230
	for (i = 0; i < dic->nr_cpages; i++) {
		struct page *page = dic->cpages[i];
		block_t blkaddr;

2231
		blkaddr = data_blkaddr(dn.inode, dn.node_page,
C
Chao Yu 已提交
2232 2233
						dn.ofs_in_node + i + 1);

2234 2235 2236
		if (bio && (!page_is_mergeable(sbi, bio,
					*last_block_in_bio, blkaddr) ||
		    !f2fs_crypt_mergeable_bio(bio, inode, page->index, NULL))) {
C
Chao Yu 已提交
2237 2238 2239 2240 2241 2242 2243 2244
submit_and_realloc:
			__submit_bio(sbi, bio, DATA);
			bio = NULL;
		}

		if (!bio) {
			bio = f2fs_grab_read_bio(inode, blkaddr, nr_pages,
					is_readahead ? REQ_RAHEAD : 0,
2245
					page->index, for_write, for_verity);
C
Chao Yu 已提交
2246
			if (IS_ERR(bio)) {
2247 2248 2249
				unsigned int remained = dic->nr_cpages - i;
				bool release = false;

C
Chao Yu 已提交
2250 2251
				ret = PTR_ERR(bio);
				dic->failed = true;
2252 2253 2254 2255 2256 2257 2258 2259 2260 2261 2262 2263

				if (for_verity) {
					if (!atomic_sub_return(remained,
						&dic->verity_pages))
						release = true;
				} else {
					if (!atomic_sub_return(remained,
						&dic->pending_pages))
						release = true;
				}

				if (release) {
C
Chao Yu 已提交
2264
					f2fs_decompress_end_io(dic->rpages,
2265 2266
						cc->cluster_size, true,
						false);
2267 2268
					f2fs_free_dic(dic);
				}
2269

C
Chao Yu 已提交
2270
				f2fs_put_dnode(&dn);
2271
				*bio_ret = NULL;
C
Chao Yu 已提交
2272 2273 2274 2275 2276 2277 2278 2279 2280
				return ret;
			}
		}

		f2fs_wait_on_block_writeback(inode, blkaddr);

		if (bio_add_page(bio, page, blocksize, 0) < blocksize)
			goto submit_and_realloc;

2281 2282 2283 2284 2285
		/* tag STEP_DECOMPRESS to handle IO in wq */
		ctx = bio->bi_private;
		if (!(ctx->enabled_steps & (1 << STEP_DECOMPRESS)))
			ctx->enabled_steps |= 1 << STEP_DECOMPRESS;

C
Chao Yu 已提交
2286
		inc_page_count(sbi, F2FS_RD_DATA);
C
Chao Yu 已提交
2287
		f2fs_update_iostat(sbi, FS_DATA_READ_IO, F2FS_BLKSIZE);
2288
		f2fs_update_iostat(sbi, FS_CDATA_READ_IO, F2FS_BLKSIZE);
C
Chao Yu 已提交
2289 2290 2291 2292 2293 2294 2295 2296 2297 2298 2299 2300 2301 2302 2303 2304 2305 2306
		ClearPageError(page);
		*last_block_in_bio = blkaddr;
	}

	f2fs_put_dnode(&dn);

	*bio_ret = bio;
	return 0;

out_put_dnode:
	f2fs_put_dnode(&dn);
out:
	f2fs_decompress_end_io(cc->rpages, cc->cluster_size, true, false);
	*bio_ret = bio;
	return ret;
}
#endif

J
Jaegeuk Kim 已提交
2307 2308 2309
/*
 * This function was originally taken from fs/mpage.c, and customized for f2fs.
 * Major change was from block_size == page_size in f2fs by default.
2310 2311 2312 2313 2314
 *
 * Note that the aops->readpages() function is ONLY used for read-ahead. If
 * this function ever deviates from doing just read-ahead, it should either
 * use ->readpage() or do the necessary surgery to decouple ->readpages()
 * from read-ahead.
J
Jaegeuk Kim 已提交
2315
 */
2316
static int f2fs_mpage_readpages(struct inode *inode,
2317
		struct readahead_control *rac, struct page *page)
J
Jaegeuk Kim 已提交
2318 2319 2320 2321
{
	struct bio *bio = NULL;
	sector_t last_block_in_bio = 0;
	struct f2fs_map_blocks map;
C
Chao Yu 已提交
2322 2323 2324 2325 2326 2327 2328 2329 2330 2331 2332 2333
#ifdef CONFIG_F2FS_FS_COMPRESSION
	struct compress_ctx cc = {
		.inode = inode,
		.log_cluster_size = F2FS_I(inode)->i_log_cluster_size,
		.cluster_size = F2FS_I(inode)->i_cluster_size,
		.cluster_idx = NULL_CLUSTER,
		.rpages = NULL,
		.cpages = NULL,
		.nr_rpages = 0,
		.nr_cpages = 0,
	};
#endif
2334
	unsigned nr_pages = rac ? readahead_count(rac) : 1;
C
Chao Yu 已提交
2335
	unsigned max_nr_pages = nr_pages;
2336
	int ret = 0;
2337
	bool drop_ra = false;
J
Jaegeuk Kim 已提交
2338 2339 2340 2341 2342

	map.m_pblk = 0;
	map.m_lblk = 0;
	map.m_len = 0;
	map.m_flags = 0;
2343
	map.m_next_pgofs = NULL;
2344
	map.m_next_extent = NULL;
2345
	map.m_seg_type = NO_CHECK_TYPE;
C
Chao Yu 已提交
2346
	map.m_may_create = false;
J
Jaegeuk Kim 已提交
2347

2348 2349 2350 2351 2352 2353 2354 2355 2356 2357 2358 2359
	/*
	 * Two readahead threads for same address range can cause race condition
	 * which fragments sequential read IOs. So let's avoid each other.
	 */
	if (rac && readahead_count(rac)) {
		if (READ_ONCE(F2FS_I(inode)->ra_offset) == readahead_index(rac))
			drop_ra = true;
		else
			WRITE_ONCE(F2FS_I(inode)->ra_offset,
						readahead_index(rac));
	}

L
LiFan 已提交
2360
	for (; nr_pages; nr_pages--) {
2361 2362
		if (rac) {
			page = readahead_page(rac);
2363
			prefetchw(&page->flags);
2364 2365 2366 2367
			if (drop_ra) {
				f2fs_put_page(page, 1);
				continue;
			}
J
Jaegeuk Kim 已提交
2368 2369
		}

C
Chao Yu 已提交
2370 2371 2372 2373 2374 2375 2376
#ifdef CONFIG_F2FS_FS_COMPRESSION
		if (f2fs_compressed_file(inode)) {
			/* there are remained comressed pages, submit them */
			if (!f2fs_cluster_can_merge_page(&cc, page->index)) {
				ret = f2fs_read_multi_pages(&cc, &bio,
							max_nr_pages,
							&last_block_in_bio,
2377
							rac != NULL, false);
C
Chao Yu 已提交
2378 2379 2380 2381 2382 2383 2384 2385 2386 2387 2388 2389 2390 2391 2392 2393 2394 2395 2396 2397 2398 2399
				f2fs_destroy_compress_ctx(&cc);
				if (ret)
					goto set_error_page;
			}
			ret = f2fs_is_compressed_cluster(inode, page->index);
			if (ret < 0)
				goto set_error_page;
			else if (!ret)
				goto read_single_page;

			ret = f2fs_init_compress_ctx(&cc);
			if (ret)
				goto set_error_page;

			f2fs_compress_ctx_add_page(&cc, page);

			goto next_page;
		}
read_single_page:
#endif

		ret = f2fs_read_single_page(inode, page, max_nr_pages, &map,
2400
					&bio, &last_block_in_bio, rac);
2401
		if (ret) {
C
Chao Yu 已提交
2402 2403 2404
#ifdef CONFIG_F2FS_FS_COMPRESSION
set_error_page:
#endif
2405
			SetPageError(page);
2406
			zero_user_segment(page, 0, PAGE_SIZE);
J
Jaegeuk Kim 已提交
2407 2408
			unlock_page(page);
		}
2409
#ifdef CONFIG_F2FS_FS_COMPRESSION
J
Jaegeuk Kim 已提交
2410
next_page:
2411 2412
#endif
		if (rac)
2413
			put_page(page);
C
Chao Yu 已提交
2414 2415 2416 2417 2418 2419 2420 2421

#ifdef CONFIG_F2FS_FS_COMPRESSION
		if (f2fs_compressed_file(inode)) {
			/* last page */
			if (nr_pages == 1 && !f2fs_cluster_is_empty(&cc)) {
				ret = f2fs_read_multi_pages(&cc, &bio,
							max_nr_pages,
							&last_block_in_bio,
2422
							rac != NULL, false);
C
Chao Yu 已提交
2423 2424 2425 2426
				f2fs_destroy_compress_ctx(&cc);
			}
		}
#endif
J
Jaegeuk Kim 已提交
2427 2428
	}
	if (bio)
2429
		__submit_bio(F2FS_I_SB(inode), bio, DATA);
2430 2431 2432

	if (rac && readahead_count(rac) && !drop_ra)
		WRITE_ONCE(F2FS_I(inode)->ra_offset, -1);
2433
	return ret;
J
Jaegeuk Kim 已提交
2434 2435
}

2436 2437
static int f2fs_read_data_page(struct file *file, struct page *page)
{
J
Jaegeuk Kim 已提交
2438
	struct inode *inode = page_file_mapping(page)->host;
2439
	int ret = -EAGAIN;
H
Huajun Li 已提交
2440

2441 2442
	trace_f2fs_readpage(page, DATA);

C
Chao Yu 已提交
2443 2444 2445 2446 2447
	if (!f2fs_is_compress_backend_ready(inode)) {
		unlock_page(page);
		return -EOPNOTSUPP;
	}

A
arter97 已提交
2448
	/* If the file has inline data, try to read it directly */
H
Huajun Li 已提交
2449 2450
	if (f2fs_has_inline_data(inode))
		ret = f2fs_read_inline_data(inode, page);
2451
	if (ret == -EAGAIN)
2452
		ret = f2fs_mpage_readpages(inode, NULL, page);
H
Huajun Li 已提交
2453
	return ret;
2454 2455
}

2456
static void f2fs_readahead(struct readahead_control *rac)
2457
{
2458
	struct inode *inode = rac->mapping->host;
2459

2460
	trace_f2fs_readpages(inode, readahead_index(rac), readahead_count(rac));
H
Huajun Li 已提交
2461

C
Chao Yu 已提交
2462
	if (!f2fs_is_compress_backend_ready(inode))
2463
		return;
C
Chao Yu 已提交
2464

H
Huajun Li 已提交
2465 2466
	/* If the file has inline data, skip readpages */
	if (f2fs_has_inline_data(inode))
2467
		return;
H
Huajun Li 已提交
2468

2469
	f2fs_mpage_readpages(inode, rac, NULL);
2470 2471
}

C
Chao Yu 已提交
2472
int f2fs_encrypt_one_page(struct f2fs_io_info *fio)
2473 2474
{
	struct inode *inode = fio->page->mapping->host;
C
Chao Yu 已提交
2475
	struct page *mpage, *page;
2476 2477
	gfp_t gfp_flags = GFP_NOFS;

2478
	if (!f2fs_encrypted_file(inode))
2479 2480
		return 0;

C
Chao Yu 已提交
2481 2482
	page = fio->compressed_page ? fio->compressed_page : fio->page;

2483
	/* wait for GCed page writeback via META_MAPPING */
2484
	f2fs_wait_on_block_writeback(inode, fio->old_blkaddr);
2485

2486 2487 2488
	if (fscrypt_inode_uses_inline_crypto(inode))
		return 0;

2489
retry_encrypt:
C
Chao Yu 已提交
2490 2491
	fio->encrypted_page = fscrypt_encrypt_pagecache_blocks(page,
					PAGE_SIZE, 0, gfp_flags);
2492 2493 2494 2495
	if (IS_ERR(fio->encrypted_page)) {
		/* flush pending IOs and wait for a while in the ENOMEM case */
		if (PTR_ERR(fio->encrypted_page) == -ENOMEM) {
			f2fs_flush_merged_writes(fio->sbi);
C
Chao Yu 已提交
2496
			congestion_wait(BLK_RW_ASYNC, DEFAULT_IO_TIMEOUT);
2497 2498 2499 2500 2501
			gfp_flags |= __GFP_NOFAIL;
			goto retry_encrypt;
		}
		return PTR_ERR(fio->encrypted_page);
	}
2502

2503 2504 2505 2506 2507 2508
	mpage = find_lock_page(META_MAPPING(fio->sbi), fio->old_blkaddr);
	if (mpage) {
		if (PageUptodate(mpage))
			memcpy(page_address(mpage),
				page_address(fio->encrypted_page), PAGE_SIZE);
		f2fs_put_page(mpage, 1);
2509
	}
2510
	return 0;
2511 2512
}

C
Chao Yu 已提交
2513 2514
static inline bool check_inplace_update_policy(struct inode *inode,
				struct f2fs_io_info *fio)
2515
{
C
Chao Yu 已提交
2516 2517
	struct f2fs_sb_info *sbi = F2FS_I_SB(inode);
	unsigned int policy = SM_I(sbi)->ipu_policy;
2518

C
Chao Yu 已提交
2519 2520
	if (policy & (0x1 << F2FS_IPU_FORCE))
		return true;
C
Chao Yu 已提交
2521
	if (policy & (0x1 << F2FS_IPU_SSR) && f2fs_need_SSR(sbi))
C
Chao Yu 已提交
2522 2523 2524 2525
		return true;
	if (policy & (0x1 << F2FS_IPU_UTIL) &&
			utilization(sbi) > SM_I(sbi)->min_ipu_util)
		return true;
C
Chao Yu 已提交
2526
	if (policy & (0x1 << F2FS_IPU_SSR_UTIL) && f2fs_need_SSR(sbi) &&
C
Chao Yu 已提交
2527 2528 2529 2530 2531 2532 2533 2534 2535
			utilization(sbi) > SM_I(sbi)->min_ipu_util)
		return true;

	/*
	 * IPU for rewrite async pages
	 */
	if (policy & (0x1 << F2FS_IPU_ASYNC) &&
			fio && fio->op == REQ_OP_WRITE &&
			!(fio->op_flags & REQ_SYNC) &&
2536
			!IS_ENCRYPTED(inode))
C
Chao Yu 已提交
2537 2538 2539 2540 2541 2542 2543
		return true;

	/* this is only set during fdatasync */
	if (policy & (0x1 << F2FS_IPU_FSYNC) &&
			is_inode_flag_set(inode, FI_NEED_IPU))
		return true;

D
Daniel Rosenberg 已提交
2544 2545 2546 2547
	if (unlikely(fio && is_sbi_flag_set(sbi, SBI_CP_DISABLED) &&
			!f2fs_is_checkpointed_data(sbi, fio->old_blkaddr)))
		return true;

C
Chao Yu 已提交
2548 2549 2550
	return false;
}

C
Chao Yu 已提交
2551
bool f2fs_should_update_inplace(struct inode *inode, struct f2fs_io_info *fio)
C
Chao Yu 已提交
2552
{
2553 2554
	if (f2fs_is_pinned_file(inode))
		return true;
C
Chao Yu 已提交
2555 2556 2557 2558 2559 2560 2561 2562

	/* if this is cold file, we should overwrite to avoid fragmentation */
	if (file_is_cold(inode))
		return true;

	return check_inplace_update_policy(inode, fio);
}

C
Chao Yu 已提交
2563
bool f2fs_should_update_outplace(struct inode *inode, struct f2fs_io_info *fio)
C
Chao Yu 已提交
2564 2565 2566
{
	struct f2fs_sb_info *sbi = F2FS_I_SB(inode);

2567
	if (f2fs_lfs_mode(sbi))
C
Chao Yu 已提交
2568 2569 2570
		return true;
	if (S_ISDIR(inode->i_mode))
		return true;
2571 2572
	if (IS_NOQUOTA(inode))
		return true;
C
Chao Yu 已提交
2573 2574 2575 2576 2577 2578 2579
	if (f2fs_is_atomic_file(inode))
		return true;
	if (fio) {
		if (is_cold_data(fio->page))
			return true;
		if (IS_ATOMIC_WRITTEN_PAGE(fio->page))
			return true;
D
Daniel Rosenberg 已提交
2580 2581 2582
		if (unlikely(is_sbi_flag_set(sbi, SBI_CP_DISABLED) &&
			f2fs_is_checkpointed_data(sbi, fio->old_blkaddr)))
			return true;
C
Chao Yu 已提交
2583 2584 2585 2586
	}
	return false;
}

2587 2588 2589 2590
static inline bool need_inplace_update(struct f2fs_io_info *fio)
{
	struct inode *inode = fio->page->mapping->host;

C
Chao Yu 已提交
2591
	if (f2fs_should_update_outplace(inode, fio))
2592 2593
		return false;

C
Chao Yu 已提交
2594
	return f2fs_should_update_inplace(inode, fio);
2595 2596
}

C
Chao Yu 已提交
2597
int f2fs_do_write_data_page(struct f2fs_io_info *fio)
2598
{
2599
	struct page *page = fio->page;
2600 2601
	struct inode *inode = page->mapping->host;
	struct dnode_of_data dn;
2602
	struct extent_info ei = {0,0,0};
2603
	struct node_info ni;
2604
	bool ipu_force = false;
2605 2606 2607
	int err = 0;

	set_new_dnode(&dn, inode, NULL, NULL, 0);
2608 2609 2610
	if (need_inplace_update(fio) &&
			f2fs_lookup_extent_cache(inode, page->index, &ei)) {
		fio->old_blkaddr = ei.blk + page->index - ei.fofs;
2611

2612
		if (!f2fs_is_valid_blkaddr(fio->sbi, fio->old_blkaddr,
C
Chao Yu 已提交
2613
						DATA_GENERIC_ENHANCE))
2614
			return -EFSCORRUPTED;
2615 2616 2617 2618

		ipu_force = true;
		fio->need_lock = LOCK_DONE;
		goto got_it;
2619
	}
2620

2621 2622 2623
	/* Deadlock due to between page->lock and f2fs_lock_op */
	if (fio->need_lock == LOCK_REQ && !f2fs_trylock_op(fio->sbi))
		return -EAGAIN;
2624

C
Chao Yu 已提交
2625
	err = f2fs_get_dnode_of_data(&dn, page->index, LOOKUP_NODE);
2626
	if (err)
2627
		goto out;
2628

2629
	fio->old_blkaddr = dn.data_blkaddr;
2630 2631

	/* This page is already truncated */
2632
	if (fio->old_blkaddr == NULL_ADDR) {
2633
		ClearPageUptodate(page);
C
Chao Yu 已提交
2634
		clear_cold_data(page);
2635
		goto out_writepage;
2636
	}
2637
got_it:
2638 2639
	if (__is_valid_data_blkaddr(fio->old_blkaddr) &&
		!f2fs_is_valid_blkaddr(fio->sbi, fio->old_blkaddr,
C
Chao Yu 已提交
2640
						DATA_GENERIC_ENHANCE)) {
2641
		err = -EFSCORRUPTED;
2642 2643
		goto out_writepage;
	}
2644 2645 2646 2647
	/*
	 * If current allocation needs SSR,
	 * it had better in-place writes for updated data.
	 */
C
Chao Yu 已提交
2648 2649
	if (ipu_force ||
		(__is_valid_data_blkaddr(fio->old_blkaddr) &&
C
Chao Yu 已提交
2650
					need_inplace_update(fio))) {
C
Chao Yu 已提交
2651
		err = f2fs_encrypt_one_page(fio);
2652 2653 2654 2655
		if (err)
			goto out_writepage;

		set_page_writeback(page);
J
Jaegeuk Kim 已提交
2656
		ClearPageError(page);
2657
		f2fs_put_dnode(&dn);
2658
		if (fio->need_lock == LOCK_REQ)
2659
			f2fs_unlock_op(fio->sbi);
C
Chao Yu 已提交
2660
		err = f2fs_inplace_write_data(fio);
C
Chao Yu 已提交
2661
		if (err) {
2662
			if (fscrypt_inode_uses_fs_layer_crypto(inode))
2663
				fscrypt_finalize_bounce_page(&fio->encrypted_page);
C
Chao Yu 已提交
2664 2665
			if (PageWriteback(page))
				end_page_writeback(page);
2666 2667
		} else {
			set_inode_flag(inode, FI_UPDATE_WRITE);
C
Chao Yu 已提交
2668
		}
2669
		trace_f2fs_do_write_data_page(fio->page, IPU);
2670
		return err;
2671
	}
2672

2673 2674 2675 2676 2677 2678 2679 2680
	if (fio->need_lock == LOCK_RETRY) {
		if (!f2fs_trylock_op(fio->sbi)) {
			err = -EAGAIN;
			goto out_writepage;
		}
		fio->need_lock = LOCK_REQ;
	}

2681 2682 2683 2684 2685 2686
	err = f2fs_get_node_info(fio->sbi, dn.nid, &ni);
	if (err)
		goto out_writepage;

	fio->version = ni.version;

C
Chao Yu 已提交
2687
	err = f2fs_encrypt_one_page(fio);
2688 2689 2690 2691
	if (err)
		goto out_writepage;

	set_page_writeback(page);
J
Jaegeuk Kim 已提交
2692
	ClearPageError(page);
2693

C
Chao Yu 已提交
2694 2695 2696
	if (fio->compr_blocks && fio->old_blkaddr == COMPRESS_ADDR)
		f2fs_i_compr_blocks_update(inode, fio->compr_blocks - 1, false);

2697
	/* LFS mode write path */
C
Chao Yu 已提交
2698
	f2fs_outplace_write_data(&dn, fio);
2699 2700 2701 2702
	trace_f2fs_do_write_data_page(page, OPU);
	set_inode_flag(inode, FI_APPEND_WRITE);
	if (page->index == 0)
		set_inode_flag(inode, FI_FIRST_BLOCK_WRITTEN);
2703 2704
out_writepage:
	f2fs_put_dnode(&dn);
2705
out:
2706
	if (fio->need_lock == LOCK_REQ)
2707
		f2fs_unlock_op(fio->sbi);
2708 2709 2710
	return err;
}

C
Chao Yu 已提交
2711
int f2fs_write_single_data_page(struct page *page, int *submitted,
C
Chao Yu 已提交
2712 2713
				struct bio **bio,
				sector_t *last_block,
C
Chao Yu 已提交
2714
				struct writeback_control *wbc,
C
Chao Yu 已提交
2715 2716
				enum iostat_type io_type,
				int compr_blocks)
2717 2718
{
	struct inode *inode = page->mapping->host;
2719
	struct f2fs_sb_info *sbi = F2FS_I_SB(inode);
2720
	loff_t i_size = i_size_read(inode);
C
Chao Yu 已提交
2721
	const pgoff_t end_index = ((unsigned long long)i_size)
2722
							>> PAGE_SHIFT;
C
Chao Yu 已提交
2723
	loff_t psize = (loff_t)(page->index + 1) << PAGE_SHIFT;
H
Huajun Li 已提交
2724
	unsigned offset = 0;
2725
	bool need_balance_fs = false;
2726
	int err = 0;
J
Jaegeuk Kim 已提交
2727
	struct f2fs_io_info fio = {
2728
		.sbi = sbi,
C
Chao Yu 已提交
2729
		.ino = inode->i_ino,
J
Jaegeuk Kim 已提交
2730
		.type = DATA,
M
Mike Christie 已提交
2731
		.op = REQ_OP_WRITE,
J
Jens Axboe 已提交
2732
		.op_flags = wbc_to_write_flags(wbc),
2733
		.old_blkaddr = NULL_ADDR,
2734
		.page = page,
2735
		.encrypted_page = NULL,
2736
		.submitted = false,
C
Chao Yu 已提交
2737
		.compr_blocks = compr_blocks,
2738
		.need_lock = LOCK_RETRY,
C
Chao Yu 已提交
2739
		.io_type = io_type,
2740
		.io_wbc = wbc,
C
Chao Yu 已提交
2741 2742
		.bio = bio,
		.last_block = last_block,
J
Jaegeuk Kim 已提交
2743
	};
2744

2745 2746
	trace_f2fs_writepage(page, DATA);

2747 2748 2749
	/* we should bypass data pages to proceed the kworkder jobs */
	if (unlikely(f2fs_cp_error(sbi))) {
		mapping_set_error(page->mapping, -EIO);
2750 2751 2752 2753 2754 2755
		/*
		 * don't drop any dirty dentry pages for keeping lastest
		 * directory structure.
		 */
		if (S_ISDIR(inode->i_mode))
			goto redirty_out;
2756 2757 2758
		goto out;
	}

2759 2760 2761
	if (unlikely(is_sbi_flag_set(sbi, SBI_POR_DOING)))
		goto redirty_out;

C
Chao Yu 已提交
2762 2763 2764
	if (page->index < end_index ||
			f2fs_verity_in_progress(inode) ||
			compr_blocks)
2765
		goto write;
2766 2767 2768 2769 2770

	/*
	 * If the offset is out-of-range of file size,
	 * this page does not have to be written to disk.
	 */
2771
	offset = i_size & (PAGE_SIZE - 1);
2772
	if ((page->index >= end_index + 1) || !offset)
2773
		goto out;
2774

2775
	zero_user_segment(page, offset, PAGE_SIZE);
2776
write:
2777 2778
	if (f2fs_is_drop_cache(inode))
		goto out;
2779 2780 2781
	/* we should not write 0'th page having journal header */
	if (f2fs_is_volatile_file(inode) && (!page->index ||
			(!wbc->for_reclaim &&
C
Chao Yu 已提交
2782
			f2fs_available_free_memory(sbi, BASE_CHECK))))
2783
		goto redirty_out;
2784

2785 2786
	/* Dentry/quota blocks are controlled by checkpoint */
	if (S_ISDIR(inode->i_mode) || IS_NOQUOTA(inode)) {
C
Chao Yu 已提交
2787 2788 2789 2790 2791 2792 2793 2794
		/*
		 * We need to wait for node_write to avoid block allocation during
		 * checkpoint. This can only happen to quota writes which can cause
		 * the below discard race condition.
		 */
		if (IS_NOQUOTA(inode))
			down_read(&sbi->node_write);

2795
		fio.need_lock = LOCK_DONE;
C
Chao Yu 已提交
2796
		err = f2fs_do_write_data_page(&fio);
C
Chao Yu 已提交
2797 2798 2799 2800

		if (IS_NOQUOTA(inode))
			up_read(&sbi->node_write);

2801 2802
		goto done;
	}
H
Huajun Li 已提交
2803

2804
	if (!wbc->for_reclaim)
2805
		need_balance_fs = true;
2806
	else if (has_not_enough_free_secs(sbi, 0, 0))
2807
		goto redirty_out;
2808 2809
	else
		set_inode_flag(inode, FI_HOT_DATA);
2810

2811
	err = -EAGAIN;
2812
	if (f2fs_has_inline_data(inode)) {
2813
		err = f2fs_write_inline_data(inode, page);
2814 2815 2816
		if (!err)
			goto out;
	}
2817

2818
	if (err == -EAGAIN) {
C
Chao Yu 已提交
2819
		err = f2fs_do_write_data_page(&fio);
2820 2821
		if (err == -EAGAIN) {
			fio.need_lock = LOCK_REQ;
C
Chao Yu 已提交
2822
			err = f2fs_do_write_data_page(&fio);
2823 2824
		}
	}
2825

2826 2827 2828
	if (err) {
		file_set_keep_isize(inode);
	} else {
2829
		spin_lock(&F2FS_I(inode)->i_size_lock);
2830 2831
		if (F2FS_I(inode)->last_disk_size < psize)
			F2FS_I(inode)->last_disk_size = psize;
2832
		spin_unlock(&F2FS_I(inode)->i_size_lock);
2833
	}
2834

2835 2836 2837
done:
	if (err && err != -ENOENT)
		goto redirty_out;
2838

2839
out:
2840
	inode_dec_dirty_pages(inode);
C
Chao Yu 已提交
2841
	if (err) {
2842
		ClearPageUptodate(page);
C
Chao Yu 已提交
2843 2844
		clear_cold_data(page);
	}
2845 2846

	if (wbc->for_reclaim) {
2847
		f2fs_submit_merged_write_cond(sbi, NULL, page, 0, DATA);
2848
		clear_inode_flag(inode, FI_HOT_DATA);
C
Chao Yu 已提交
2849
		f2fs_remove_dirty_inode(inode);
2850
		submitted = NULL;
2851
	}
2852
	unlock_page(page);
2853
	if (!S_ISDIR(inode->i_mode) && !IS_NOQUOTA(inode) &&
C
Chao Yu 已提交
2854
					!F2FS_I(inode)->cp_task)
J
Jaegeuk Kim 已提交
2855
		f2fs_balance_fs(sbi, need_balance_fs);
2856

2857
	if (unlikely(f2fs_cp_error(sbi))) {
2858
		f2fs_submit_merged_write(sbi, DATA);
C
Chao Yu 已提交
2859
		f2fs_submit_merged_ipu_write(sbi, bio, NULL);
2860 2861 2862 2863
		submitted = NULL;
	}

	if (submitted)
C
Chao Yu 已提交
2864
		*submitted = fio.submitted ? 1 : 0;
2865

2866 2867 2868
	return 0;

redirty_out:
2869
	redirty_page_for_writepage(wbc, page);
2870 2871 2872 2873 2874 2875 2876
	/*
	 * pageout() in MM traslates EAGAIN, so calls handle_write_error()
	 * -> mapping_set_error() -> set_bit(AS_EIO, ...).
	 * file_write_and_wait_range() will see EIO error, which is critical
	 * to return value of fsync() followed by atomic_write failure to user.
	 */
	if (!err || wbc->for_reclaim)
2877
		return AOP_WRITEPAGE_ACTIVATE;
J
Jaegeuk Kim 已提交
2878 2879
	unlock_page(page);
	return err;
2880 2881
}

2882 2883 2884
static int f2fs_write_data_page(struct page *page,
					struct writeback_control *wbc)
{
C
Chao Yu 已提交
2885 2886 2887 2888 2889 2890 2891 2892 2893 2894 2895 2896 2897 2898 2899 2900 2901
#ifdef CONFIG_F2FS_FS_COMPRESSION
	struct inode *inode = page->mapping->host;

	if (unlikely(f2fs_cp_error(F2FS_I_SB(inode))))
		goto out;

	if (f2fs_compressed_file(inode)) {
		if (f2fs_is_compressed_cluster(inode, page->index)) {
			redirty_page_for_writepage(wbc, page);
			return AOP_WRITEPAGE_ACTIVATE;
		}
	}
out:
#endif

	return f2fs_write_single_data_page(page, NULL, NULL, NULL,
						wbc, FS_DATA_IO, 0);
2902 2903
}

C
Chao Yu 已提交
2904 2905 2906 2907 2908 2909
/*
 * This function was copied from write_cche_pages from mm/page-writeback.c.
 * The major change is making write step of cold data page separately from
 * warm/hot data page.
 */
static int f2fs_write_cache_pages(struct address_space *mapping,
C
Chao Yu 已提交
2910 2911
					struct writeback_control *wbc,
					enum iostat_type io_type)
C
Chao Yu 已提交
2912 2913
{
	int ret = 0;
C
Chao Yu 已提交
2914
	int done = 0, retry = 0;
C
Chao Yu 已提交
2915
	struct pagevec pvec;
2916
	struct f2fs_sb_info *sbi = F2FS_M_SB(mapping);
C
Chao Yu 已提交
2917 2918
	struct bio *bio = NULL;
	sector_t last_block;
C
Chao Yu 已提交
2919 2920 2921 2922 2923 2924 2925 2926 2927 2928 2929 2930 2931 2932 2933 2934
#ifdef CONFIG_F2FS_FS_COMPRESSION
	struct inode *inode = mapping->host;
	struct compress_ctx cc = {
		.inode = inode,
		.log_cluster_size = F2FS_I(inode)->i_log_cluster_size,
		.cluster_size = F2FS_I(inode)->i_cluster_size,
		.cluster_idx = NULL_CLUSTER,
		.rpages = NULL,
		.nr_rpages = 0,
		.cpages = NULL,
		.rbuf = NULL,
		.cbuf = NULL,
		.rlen = PAGE_SIZE * F2FS_I(inode)->i_cluster_size,
		.private = NULL,
	};
#endif
C
Chao Yu 已提交
2935 2936 2937 2938 2939
	int nr_pages;
	pgoff_t index;
	pgoff_t end;		/* Inclusive */
	pgoff_t done_index;
	int range_whole = 0;
M
Matthew Wilcox 已提交
2940
	xa_mark_t tag;
2941
	int nwritten = 0;
C
Chao Yu 已提交
2942 2943
	int submitted = 0;
	int i;
C
Chao Yu 已提交
2944

2945
	pagevec_init(&pvec);
2946

2947 2948 2949 2950 2951 2952
	if (get_dirty_pages(mapping->host) <=
				SM_I(F2FS_M_SB(mapping))->min_hot_blocks)
		set_inode_flag(mapping->host, FI_HOT_DATA);
	else
		clear_inode_flag(mapping->host, FI_HOT_DATA);

C
Chao Yu 已提交
2953
	if (wbc->range_cyclic) {
2954
		index = mapping->writeback_index; /* prev offset */
C
Chao Yu 已提交
2955 2956
		end = -1;
	} else {
2957 2958
		index = wbc->range_start >> PAGE_SHIFT;
		end = wbc->range_end >> PAGE_SHIFT;
C
Chao Yu 已提交
2959 2960 2961 2962 2963 2964 2965 2966
		if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX)
			range_whole = 1;
	}
	if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages)
		tag = PAGECACHE_TAG_TOWRITE;
	else
		tag = PAGECACHE_TAG_DIRTY;
retry:
C
Chao Yu 已提交
2967
	retry = 0;
C
Chao Yu 已提交
2968 2969 2970
	if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages)
		tag_pages_for_writeback(mapping, index, end);
	done_index = index;
C
Chao Yu 已提交
2971
	while (!done && !retry && (index <= end)) {
J
Jan Kara 已提交
2972
		nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, end,
2973
				tag);
C
Chao Yu 已提交
2974 2975 2976 2977 2978
		if (nr_pages == 0)
			break;

		for (i = 0; i < nr_pages; i++) {
			struct page *page = pvec.pages[i];
C
Chao Yu 已提交
2979 2980 2981 2982 2983 2984 2985 2986 2987 2988 2989 2990 2991 2992 2993 2994 2995 2996 2997
			bool need_readd;
readd:
			need_readd = false;
#ifdef CONFIG_F2FS_FS_COMPRESSION
			if (f2fs_compressed_file(inode)) {
				ret = f2fs_init_compress_ctx(&cc);
				if (ret) {
					done = 1;
					break;
				}

				if (!f2fs_cluster_can_merge_page(&cc,
								page->index)) {
					ret = f2fs_write_multi_pages(&cc,
						&submitted, wbc, io_type);
					if (!ret)
						need_readd = true;
					goto result;
				}
C
Chao Yu 已提交
2998

C
Chao Yu 已提交
2999 3000 3001 3002 3003 3004 3005 3006 3007 3008 3009 3010 3011 3012 3013 3014 3015 3016 3017 3018 3019 3020 3021 3022 3023 3024 3025
				if (unlikely(f2fs_cp_error(sbi)))
					goto lock_page;

				if (f2fs_cluster_is_empty(&cc)) {
					void *fsdata = NULL;
					struct page *pagep;
					int ret2;

					ret2 = f2fs_prepare_compress_overwrite(
							inode, &pagep,
							page->index, &fsdata);
					if (ret2 < 0) {
						ret = ret2;
						done = 1;
						break;
					} else if (ret2 &&
						!f2fs_compress_write_end(inode,
								fsdata, page->index,
								1)) {
						retry = 1;
						break;
					}
				} else {
					goto lock_page;
				}
			}
#endif
3026
			/* give a priority to WB_SYNC threads */
3027
			if (atomic_read(&sbi->wb_sync_req[DATA]) &&
3028 3029 3030 3031
					wbc->sync_mode == WB_SYNC_NONE) {
				done = 1;
				break;
			}
C
Chao Yu 已提交
3032 3033 3034
#ifdef CONFIG_F2FS_FS_COMPRESSION
lock_page:
#endif
C
Chao Yu 已提交
3035
			done_index = page->index;
3036
retry_write:
C
Chao Yu 已提交
3037 3038 3039 3040 3041 3042 3043 3044 3045 3046 3047 3048 3049 3050
			lock_page(page);

			if (unlikely(page->mapping != mapping)) {
continue_unlock:
				unlock_page(page);
				continue;
			}

			if (!PageDirty(page)) {
				/* someone wrote it for us */
				goto continue_unlock;
			}

			if (PageWriteback(page)) {
C
Chao Yu 已提交
3051
				if (wbc->sync_mode != WB_SYNC_NONE)
3052
					f2fs_wait_on_page_writeback(page,
3053
							DATA, true, true);
C
Chao Yu 已提交
3054
				else
C
Chao Yu 已提交
3055 3056 3057 3058 3059 3060
					goto continue_unlock;
			}

			if (!clear_page_dirty_for_io(page))
				goto continue_unlock;

C
Chao Yu 已提交
3061 3062 3063 3064 3065 3066 3067 3068 3069 3070 3071 3072 3073 3074 3075 3076 3077
#ifdef CONFIG_F2FS_FS_COMPRESSION
			if (f2fs_compressed_file(inode)) {
				get_page(page);
				f2fs_compress_ctx_add_page(&cc, page);
				continue;
			}
#endif
			ret = f2fs_write_single_data_page(page, &submitted,
					&bio, &last_block, wbc, io_type, 0);
			if (ret == AOP_WRITEPAGE_ACTIVATE)
				unlock_page(page);
#ifdef CONFIG_F2FS_FS_COMPRESSION
result:
#endif
			nwritten += submitted;
			wbc->nr_to_write -= submitted;

C
Chao Yu 已提交
3078
			if (unlikely(ret)) {
3079 3080 3081 3082 3083 3084
				/*
				 * keep nr_to_write, since vfs uses this to
				 * get # of written pages.
				 */
				if (ret == AOP_WRITEPAGE_ACTIVATE) {
					ret = 0;
C
Chao Yu 已提交
3085
					goto next;
3086 3087 3088 3089 3090
				} else if (ret == -EAGAIN) {
					ret = 0;
					if (wbc->sync_mode == WB_SYNC_ALL) {
						cond_resched();
						congestion_wait(BLK_RW_ASYNC,
C
Chao Yu 已提交
3091
							DEFAULT_IO_TIMEOUT);
3092 3093
						goto retry_write;
					}
C
Chao Yu 已提交
3094
					goto next;
3095
				}
J
Jaegeuk Kim 已提交
3096 3097 3098
				done_index = page->index + 1;
				done = 1;
				break;
C
Chao Yu 已提交
3099 3100
			}

C
Chao Yu 已提交
3101
			if (wbc->nr_to_write <= 0 &&
3102
					wbc->sync_mode == WB_SYNC_NONE) {
C
Chao Yu 已提交
3103 3104 3105
				done = 1;
				break;
			}
C
Chao Yu 已提交
3106 3107 3108
next:
			if (need_readd)
				goto readd;
C
Chao Yu 已提交
3109 3110 3111 3112
		}
		pagevec_release(&pvec);
		cond_resched();
	}
C
Chao Yu 已提交
3113 3114 3115 3116 3117 3118 3119 3120 3121 3122 3123
#ifdef CONFIG_F2FS_FS_COMPRESSION
	/* flush remained pages in compress cluster */
	if (f2fs_compressed_file(inode) && !f2fs_cluster_is_empty(&cc)) {
		ret = f2fs_write_multi_pages(&cc, &submitted, wbc, io_type);
		nwritten += submitted;
		wbc->nr_to_write -= submitted;
		if (ret) {
			done = 1;
			retry = 0;
		}
	}
3124 3125
	if (f2fs_compressed_file(inode))
		f2fs_destroy_compress_ctx(&cc);
C
Chao Yu 已提交
3126
#endif
3127
	if (retry) {
C
Chao Yu 已提交
3128
		index = 0;
3129
		end = -1;
C
Chao Yu 已提交
3130 3131
		goto retry;
	}
3132 3133
	if (wbc->range_cyclic && !done)
		done_index = 0;
C
Chao Yu 已提交
3134 3135 3136
	if (wbc->range_cyclic || (range_whole && wbc->nr_to_write > 0))
		mapping->writeback_index = done_index;

3137
	if (nwritten)
3138
		f2fs_submit_merged_write_cond(F2FS_M_SB(mapping), mapping->host,
3139
								NULL, 0, DATA);
C
Chao Yu 已提交
3140 3141
	/* submit cached bio of IPU write */
	if (bio)
C
Chao Yu 已提交
3142
		f2fs_submit_merged_ipu_write(sbi, &bio, NULL);
C
Chao Yu 已提交
3143

C
Chao Yu 已提交
3144 3145 3146
	return ret;
}

3147 3148 3149
static inline bool __should_serialize_io(struct inode *inode,
					struct writeback_control *wbc)
{
C
Chao Yu 已提交
3150 3151 3152 3153
	/* to avoid deadlock in path of data flush */
	if (F2FS_I(inode)->cp_task)
		return false;

3154 3155
	if (!S_ISREG(inode->i_mode))
		return false;
3156 3157
	if (IS_NOQUOTA(inode))
		return false;
C
Chao Yu 已提交
3158

D
Daeho Jeong 已提交
3159
	if (f2fs_need_compress_data(inode))
C
Chao Yu 已提交
3160
		return true;
3161 3162 3163 3164 3165 3166 3167
	if (wbc->sync_mode != WB_SYNC_ALL)
		return true;
	if (get_dirty_pages(inode) >= SM_I(F2FS_I_SB(inode))->min_seq_blocks)
		return true;
	return false;
}

3168
static int __f2fs_write_data_pages(struct address_space *mapping,
C
Chao Yu 已提交
3169 3170
						struct writeback_control *wbc,
						enum iostat_type io_type)
3171 3172
{
	struct inode *inode = mapping->host;
3173
	struct f2fs_sb_info *sbi = F2FS_I_SB(inode);
3174
	struct blk_plug plug;
3175
	int ret;
3176
	bool locked = false;
3177

P
P J P 已提交
3178 3179 3180 3181
	/* deal with chardevs and other special file */
	if (!mapping->a_ops->writepage)
		return 0;

3182 3183 3184 3185
	/* skip writing if there is no dirty page in this inode */
	if (!get_dirty_pages(inode) && wbc->sync_mode == WB_SYNC_NONE)
		return 0;

3186 3187 3188 3189
	/* during POR, we don't need to trigger writepage at all. */
	if (unlikely(is_sbi_flag_set(sbi, SBI_POR_DOING)))
		goto skip_write;

3190 3191
	if ((S_ISDIR(inode->i_mode) || IS_NOQUOTA(inode)) &&
			wbc->sync_mode == WB_SYNC_NONE &&
3192
			get_dirty_pages(inode) < nr_pages_to_skip(sbi, DATA) &&
C
Chao Yu 已提交
3193
			f2fs_available_free_memory(sbi, DIRTY_DENTS))
3194 3195
		goto skip_write;

C
Chao Yu 已提交
3196
	/* skip writing during file defragment */
3197
	if (is_inode_flag_set(inode, FI_DO_DEFRAG))
C
Chao Yu 已提交
3198 3199
		goto skip_write;

Y
Yunlei He 已提交
3200 3201
	trace_f2fs_writepages(mapping->host, wbc, DATA);

3202 3203
	/* to avoid spliting IOs due to mixed WB_SYNC_ALL and WB_SYNC_NONE */
	if (wbc->sync_mode == WB_SYNC_ALL)
3204 3205
		atomic_inc(&sbi->wb_sync_req[DATA]);
	else if (atomic_read(&sbi->wb_sync_req[DATA]))
3206 3207
		goto skip_write;

3208 3209 3210 3211 3212
	if (__should_serialize_io(inode, wbc)) {
		mutex_lock(&sbi->writepages);
		locked = true;
	}

3213
	blk_start_plug(&plug);
C
Chao Yu 已提交
3214
	ret = f2fs_write_cache_pages(mapping, wbc, io_type);
3215
	blk_finish_plug(&plug);
3216

3217 3218 3219
	if (locked)
		mutex_unlock(&sbi->writepages);

3220
	if (wbc->sync_mode == WB_SYNC_ALL)
3221
		atomic_dec(&sbi->wb_sync_req[DATA]);
3222 3223 3224 3225
	/*
	 * if some pages were truncated, we cannot guarantee its mapping->host
	 * to detect pending bios.
	 */
J
Jaegeuk Kim 已提交
3226

C
Chao Yu 已提交
3227
	f2fs_remove_dirty_inode(inode);
3228
	return ret;
3229 3230

skip_write:
3231
	wbc->pages_skipped += get_dirty_pages(inode);
Y
Yunlei He 已提交
3232
	trace_f2fs_writepages(mapping->host, wbc, DATA);
3233
	return 0;
3234 3235
}

C
Chao Yu 已提交
3236 3237 3238 3239 3240 3241 3242 3243 3244 3245
static int f2fs_write_data_pages(struct address_space *mapping,
			    struct writeback_control *wbc)
{
	struct inode *inode = mapping->host;

	return __f2fs_write_data_pages(mapping, wbc,
			F2FS_I(inode)->cp_task == current ?
			FS_CP_DATA_IO : FS_DATA_IO);
}

3246 3247 3248
static void f2fs_write_failed(struct address_space *mapping, loff_t to)
{
	struct inode *inode = mapping->host;
J
Jaegeuk Kim 已提交
3249
	loff_t i_size = i_size_read(inode);
3250

3251 3252 3253
	if (IS_NOQUOTA(inode))
		return;

E
Eric Biggers 已提交
3254 3255
	/* In the fs-verity case, f2fs_end_enable_verity() does the truncate */
	if (to > i_size && !f2fs_verity_in_progress(inode)) {
3256
		down_write(&F2FS_I(inode)->i_gc_rwsem[WRITE]);
3257
		down_write(&F2FS_I(inode)->i_mmap_sem);
3258

J
Jaegeuk Kim 已提交
3259
		truncate_pagecache(inode, i_size);
3260
		f2fs_truncate_blocks(inode, i_size, true);
3261

3262
		up_write(&F2FS_I(inode)->i_mmap_sem);
3263
		up_write(&F2FS_I(inode)->i_gc_rwsem[WRITE]);
3264 3265 3266
	}
}

3267 3268 3269 3270 3271 3272 3273 3274
static int prepare_write_begin(struct f2fs_sb_info *sbi,
			struct page *page, loff_t pos, unsigned len,
			block_t *blk_addr, bool *node_changed)
{
	struct inode *inode = page->mapping->host;
	pgoff_t index = page->index;
	struct dnode_of_data dn;
	struct page *ipage;
3275
	bool locked = false;
3276
	struct extent_info ei = {0,0,0};
3277
	int err = 0;
3278
	int flag;
3279

3280 3281 3282 3283
	/*
	 * we already allocated all the blocks, so we don't need to get
	 * the block addresses when there is no need to fill the page.
	 */
3284
	if (!f2fs_has_inline_data(inode) && len == PAGE_SIZE &&
E
Eric Biggers 已提交
3285 3286
	    !is_inode_flag_set(inode, FI_NO_PREALLOC) &&
	    !f2fs_verity_in_progress(inode))
3287 3288
		return 0;

3289 3290 3291 3292 3293 3294
	/* f2fs_lock_op avoids race between write CP and convert_inline_page */
	if (f2fs_has_inline_data(inode) && pos + len > MAX_INLINE_DATA(inode))
		flag = F2FS_GET_BLOCK_DEFAULT;
	else
		flag = F2FS_GET_BLOCK_PRE_AIO;

3295
	if (f2fs_has_inline_data(inode) ||
3296
			(pos & PAGE_MASK) >= i_size_read(inode)) {
C
Chao Yu 已提交
3297
		f2fs_do_map_lock(sbi, flag, true);
3298 3299
		locked = true;
	}
C
Chao Yu 已提交
3300

3301
restart:
3302
	/* check inline_data */
C
Chao Yu 已提交
3303
	ipage = f2fs_get_node_page(sbi, inode->i_ino);
3304 3305 3306 3307 3308 3309 3310 3311
	if (IS_ERR(ipage)) {
		err = PTR_ERR(ipage);
		goto unlock_out;
	}

	set_new_dnode(&dn, inode, ipage, ipage, 0);

	if (f2fs_has_inline_data(inode)) {
C
Chao Yu 已提交
3312
		if (pos + len <= MAX_INLINE_DATA(inode)) {
C
Chao Yu 已提交
3313
			f2fs_do_read_inline_data(page, ipage);
3314
			set_inode_flag(inode, FI_DATA_EXIST);
3315 3316
			if (inode->i_nlink)
				set_inline_node(ipage);
3317 3318 3319
		} else {
			err = f2fs_convert_inline_page(&dn, page);
			if (err)
3320 3321 3322 3323 3324 3325 3326 3327 3328 3329 3330
				goto out;
			if (dn.data_blkaddr == NULL_ADDR)
				err = f2fs_get_block(&dn, index);
		}
	} else if (locked) {
		err = f2fs_get_block(&dn, index);
	} else {
		if (f2fs_lookup_extent_cache(inode, index, &ei)) {
			dn.data_blkaddr = ei.blk + index - ei.fofs;
		} else {
			/* hole case */
C
Chao Yu 已提交
3331
			err = f2fs_get_dnode_of_data(&dn, index, LOOKUP_NODE);
3332
			if (err || dn.data_blkaddr == NULL_ADDR) {
3333
				f2fs_put_dnode(&dn);
C
Chao Yu 已提交
3334
				f2fs_do_map_lock(sbi, F2FS_GET_BLOCK_PRE_AIO,
3335
								true);
3336
				WARN_ON(flag != F2FS_GET_BLOCK_PRE_AIO);
3337 3338 3339
				locked = true;
				goto restart;
			}
3340 3341
		}
	}
3342

3343 3344 3345
	/* convert_inline_page can make node_changed */
	*blk_addr = dn.data_blkaddr;
	*node_changed = dn.node_changed;
3346
out:
3347 3348
	f2fs_put_dnode(&dn);
unlock_out:
3349
	if (locked)
C
Chao Yu 已提交
3350
		f2fs_do_map_lock(sbi, flag, false);
3351 3352 3353
	return err;
}

3354 3355 3356 3357 3358
static int f2fs_write_begin(struct file *file, struct address_space *mapping,
		loff_t pos, unsigned len, unsigned flags,
		struct page **pagep, void **fsdata)
{
	struct inode *inode = mapping->host;
3359
	struct f2fs_sb_info *sbi = F2FS_I_SB(inode);
3360
	struct page *page = NULL;
3361
	pgoff_t index = ((unsigned long long) pos) >> PAGE_SHIFT;
3362
	bool need_balance = false, drop_atomic = false;
3363
	block_t blkaddr = NULL_ADDR;
3364 3365
	int err = 0;

3366 3367
	trace_f2fs_write_begin(inode, pos, len, flags);

3368 3369
	if (!f2fs_is_checkpoint_ready(sbi)) {
		err = -ENOSPC;
D
Daniel Rosenberg 已提交
3370
		goto fail;
3371
	}
D
Daniel Rosenberg 已提交
3372

3373 3374 3375
	if ((f2fs_is_atomic_file(inode) &&
			!f2fs_available_free_memory(sbi, INMEM_PAGES)) ||
			is_inode_flag_set(inode, FI_ATOMIC_REVOKE_REQUEST)) {
J
Jaegeuk Kim 已提交
3376
		err = -ENOMEM;
3377
		drop_atomic = true;
J
Jaegeuk Kim 已提交
3378 3379 3380
		goto fail;
	}

3381 3382 3383 3384 3385 3386 3387 3388 3389 3390
	/*
	 * We should check this at this moment to avoid deadlock on inode page
	 * and #0 page. The locking rule for inline_data conversion should be:
	 * lock_page(page #0) -> lock_page(inode_page)
	 */
	if (index != 0) {
		err = f2fs_convert_inline_inode(inode);
		if (err)
			goto fail;
	}
C
Chao Yu 已提交
3391 3392 3393 3394 3395 3396 3397 3398 3399 3400 3401 3402 3403 3404 3405 3406 3407 3408

#ifdef CONFIG_F2FS_FS_COMPRESSION
	if (f2fs_compressed_file(inode)) {
		int ret;

		*fsdata = NULL;

		ret = f2fs_prepare_compress_overwrite(inode, pagep,
							index, fsdata);
		if (ret < 0) {
			err = ret;
			goto fail;
		} else if (ret) {
			return 0;
		}
	}
#endif

3409
repeat:
3410 3411 3412 3413
	/*
	 * Do not use grab_cache_page_write_begin() to avoid deadlock due to
	 * wait_for_stable_page. Will wait that below with our IO control.
	 */
C
Chao Yu 已提交
3414
	page = f2fs_pagecache_get_page(mapping, index,
3415
				FGP_LOCK | FGP_WRITE | FGP_CREAT, GFP_NOFS);
3416 3417 3418 3419
	if (!page) {
		err = -ENOMEM;
		goto fail;
	}
3420

C
Chao Yu 已提交
3421 3422
	/* TODO: cluster can be compressed due to race with .writepage */

3423 3424
	*pagep = page;

3425 3426
	err = prepare_write_begin(sbi, page, pos, len,
					&blkaddr, &need_balance);
3427
	if (err)
3428
		goto fail;
3429

3430 3431
	if (need_balance && !IS_NOQUOTA(inode) &&
			has_not_enough_free_secs(sbi, 0, 0)) {
3432
		unlock_page(page);
J
Jaegeuk Kim 已提交
3433
		f2fs_balance_fs(sbi, true);
3434 3435 3436 3437 3438 3439 3440 3441
		lock_page(page);
		if (page->mapping != mapping) {
			/* The page got truncated from under us */
			f2fs_put_page(page, 1);
			goto repeat;
		}
	}

3442
	f2fs_wait_on_page_writeback(page, DATA, false, true);
3443

3444 3445
	if (len == PAGE_SIZE || PageUptodate(page))
		return 0;
3446

E
Eric Biggers 已提交
3447 3448
	if (!(pos & (PAGE_SIZE - 1)) && (pos + len) >= i_size_read(inode) &&
	    !f2fs_verity_in_progress(inode)) {
3449 3450 3451 3452
		zero_user_segment(page, len, PAGE_SIZE);
		return 0;
	}

3453
	if (blkaddr == NEW_ADDR) {
3454
		zero_user_segment(page, 0, PAGE_SIZE);
3455
		SetPageUptodate(page);
3456
	} else {
C
Chao Yu 已提交
3457 3458
		if (!f2fs_is_valid_blkaddr(sbi, blkaddr,
				DATA_GENERIC_ENHANCE_READ)) {
3459
			err = -EFSCORRUPTED;
C
Chao Yu 已提交
3460 3461
			goto fail;
		}
3462
		err = f2fs_submit_page_read(inode, page, blkaddr, 0, true);
3463
		if (err)
3464
			goto fail;
3465

3466
		lock_page(page);
3467
		if (unlikely(page->mapping != mapping)) {
3468 3469
			f2fs_put_page(page, 1);
			goto repeat;
3470
		}
3471 3472 3473
		if (unlikely(!PageUptodate(page))) {
			err = -EIO;
			goto fail;
3474
		}
3475 3476
	}
	return 0;
3477

3478
fail:
3479
	f2fs_put_page(page, 1);
3480
	f2fs_write_failed(mapping, pos + len);
3481
	if (drop_atomic)
C
Chao Yu 已提交
3482
		f2fs_drop_inmem_pages_all(sbi, false);
3483
	return err;
3484 3485
}

3486 3487 3488 3489 3490 3491 3492
static int f2fs_write_end(struct file *file,
			struct address_space *mapping,
			loff_t pos, unsigned len, unsigned copied,
			struct page *page, void *fsdata)
{
	struct inode *inode = page->mapping->host;

3493 3494
	trace_f2fs_write_end(inode, pos, len, copied);

3495 3496 3497 3498 3499 3500
	/*
	 * This should be come from len == PAGE_SIZE, and we expect copied
	 * should be PAGE_SIZE. Otherwise, we treat it with zero copied and
	 * let generic_perform_write() try to copy data again through copied=0.
	 */
	if (!PageUptodate(page)) {
3501
		if (unlikely(copied != len))
3502 3503 3504 3505
			copied = 0;
		else
			SetPageUptodate(page);
	}
C
Chao Yu 已提交
3506 3507 3508 3509 3510 3511

#ifdef CONFIG_F2FS_FS_COMPRESSION
	/* overwrite compressed file */
	if (f2fs_compressed_file(inode) && fsdata) {
		f2fs_compress_write_end(inode, fsdata, page->index, copied);
		f2fs_update_time(F2FS_I_SB(inode), REQ_TIME);
3512 3513 3514 3515

		if (pos + copied > i_size_read(inode) &&
				!f2fs_verity_in_progress(inode))
			f2fs_i_size_write(inode, pos + copied);
C
Chao Yu 已提交
3516 3517 3518 3519
		return copied;
	}
#endif

3520 3521 3522
	if (!copied)
		goto unlock_out;

3523
	set_page_dirty(page);
3524

E
Eric Biggers 已提交
3525 3526
	if (pos + copied > i_size_read(inode) &&
	    !f2fs_verity_in_progress(inode))
3527
		f2fs_i_size_write(inode, pos + copied);
3528
unlock_out:
3529
	f2fs_put_page(page, 1);
3530
	f2fs_update_time(F2FS_I_SB(inode), REQ_TIME);
3531 3532 3533
	return copied;
}

3534 3535
static int check_direct_IO(struct inode *inode, struct iov_iter *iter,
			   loff_t offset)
3536
{
3537 3538 3539 3540 3541 3542
	unsigned i_blkbits = READ_ONCE(inode->i_blkbits);
	unsigned blkbits = i_blkbits;
	unsigned blocksize_mask = (1 << blkbits) - 1;
	unsigned long align = offset | iov_iter_alignment(iter);
	struct block_device *bdev = inode->i_sb->s_bdev;

3543 3544 3545
	if (iov_iter_rw(iter) == READ && offset >= i_size_read(inode))
		return 1;

3546 3547 3548 3549 3550 3551 3552 3553
	if (align & blocksize_mask) {
		if (bdev)
			blkbits = blksize_bits(bdev_logical_block_size(bdev));
		blocksize_mask = (1 << blkbits) - 1;
		if (align & blocksize_mask)
			return -EINVAL;
		return 1;
	}
3554 3555 3556
	return 0;
}

C
Chao Yu 已提交
3557 3558 3559 3560 3561 3562 3563 3564 3565 3566
static void f2fs_dio_end_io(struct bio *bio)
{
	struct f2fs_private_dio *dio = bio->bi_private;

	dec_page_count(F2FS_I_SB(dio->inode),
			dio->write ? F2FS_DIO_WRITE : F2FS_DIO_READ);

	bio->bi_private = dio->orig_private;
	bio->bi_end_io = dio->orig_end_io;

C
Chao Yu 已提交
3567
	kfree(dio);
C
Chao Yu 已提交
3568 3569 3570 3571 3572 3573 3574 3575 3576 3577 3578 3579

	bio_endio(bio);
}

static void f2fs_dio_submit_bio(struct bio *bio, struct inode *inode,
							loff_t file_offset)
{
	struct f2fs_private_dio *dio;
	bool write = (bio_op(bio) == REQ_OP_WRITE);

	dio = f2fs_kzalloc(F2FS_I_SB(inode),
			sizeof(struct f2fs_private_dio), GFP_NOFS);
3580
	if (!dio)
C
Chao Yu 已提交
3581 3582 3583 3584 3585 3586 3587 3588 3589 3590 3591 3592 3593 3594 3595 3596 3597 3598 3599 3600
		goto out;

	dio->inode = inode;
	dio->orig_end_io = bio->bi_end_io;
	dio->orig_private = bio->bi_private;
	dio->write = write;

	bio->bi_end_io = f2fs_dio_end_io;
	bio->bi_private = dio;

	inc_page_count(F2FS_I_SB(inode),
			write ? F2FS_DIO_WRITE : F2FS_DIO_READ);

	submit_bio(bio);
	return;
out:
	bio->bi_status = BLK_STS_IOERR;
	bio_endio(bio);
}

3601
static ssize_t f2fs_direct_IO(struct kiocb *iocb, struct iov_iter *iter)
3602
{
3603
	struct address_space *mapping = iocb->ki_filp->f_mapping;
3604
	struct inode *inode = mapping->host;
3605
	struct f2fs_sb_info *sbi = F2FS_I_SB(inode);
3606
	struct f2fs_inode_info *fi = F2FS_I(inode);
3607
	size_t count = iov_iter_count(iter);
3608
	loff_t offset = iocb->ki_pos;
3609
	int rw = iov_iter_rw(iter);
3610
	int err;
3611
	enum rw_hint hint = iocb->ki_hint;
3612
	int whint_mode = F2FS_OPTION(sbi).whint_mode;
3613
	bool do_opu;
3614

3615
	err = check_direct_IO(inode, iter, offset);
3616
	if (err)
3617
		return err < 0 ? err : 0;
H
Huajun Li 已提交
3618

3619
	if (f2fs_force_buffered_io(inode, iocb, iter))
3620
		return 0;
3621

3622 3623
	do_opu = allow_outplace_dio(inode, iocb, iter);

3624
	trace_f2fs_direct_IO_enter(inode, offset, count, rw);
3625

3626 3627 3628
	if (rw == WRITE && whint_mode == WHINT_MODE_OFF)
		iocb->ki_hint = WRITE_LIFE_NOT_SET;

3629 3630 3631 3632 3633 3634 3635 3636
	if (iocb->ki_flags & IOCB_NOWAIT) {
		if (!down_read_trylock(&fi->i_gc_rwsem[rw])) {
			iocb->ki_hint = hint;
			err = -EAGAIN;
			goto out;
		}
		if (do_opu && !down_read_trylock(&fi->i_gc_rwsem[READ])) {
			up_read(&fi->i_gc_rwsem[rw]);
H
Hyunchul Lee 已提交
3637 3638 3639 3640
			iocb->ki_hint = hint;
			err = -EAGAIN;
			goto out;
		}
3641 3642 3643 3644
	} else {
		down_read(&fi->i_gc_rwsem[rw]);
		if (do_opu)
			down_read(&fi->i_gc_rwsem[READ]);
H
Hyunchul Lee 已提交
3645 3646
	}

C
Chao Yu 已提交
3647
	err = __blockdev_direct_IO(iocb, inode, inode->i_sb->s_bdev,
C
Chao Yu 已提交
3648 3649
			iter, rw == WRITE ? get_data_block_dio_write :
			get_data_block_dio, NULL, f2fs_dio_submit_bio,
D
DongDongJu 已提交
3650 3651
			rw == WRITE ? DIO_LOCKING | DIO_SKIP_HOLES :
			DIO_SKIP_HOLES);
3652 3653 3654 3655 3656

	if (do_opu)
		up_read(&fi->i_gc_rwsem[READ]);

	up_read(&fi->i_gc_rwsem[rw]);
3657 3658

	if (rw == WRITE) {
3659 3660
		if (whint_mode == WHINT_MODE_OFF)
			iocb->ki_hint = hint;
C
Chao Yu 已提交
3661 3662 3663
		if (err > 0) {
			f2fs_update_iostat(F2FS_I_SB(inode), APP_DIRECT_IO,
									err);
3664 3665
			if (!do_opu)
				set_inode_flag(inode, FI_UPDATE_WRITE);
3666 3667 3668
		} else if (err == -EIOCBQUEUED) {
			f2fs_update_iostat(F2FS_I_SB(inode), APP_DIRECT_IO,
						count - iov_iter_count(iter));
C
Chao Yu 已提交
3669
		} else if (err < 0) {
3670
			f2fs_write_failed(mapping, offset + count);
C
Chao Yu 已提交
3671
		}
C
Chao Yu 已提交
3672 3673 3674
	} else {
		if (err > 0)
			f2fs_update_iostat(sbi, APP_DIRECT_READ_IO, err);
3675 3676 3677
		else if (err == -EIOCBQUEUED)
			f2fs_update_iostat(F2FS_I_SB(inode), APP_DIRECT_READ_IO,
						count - iov_iter_count(iter));
3678
	}
3679

H
Hyunchul Lee 已提交
3680
out:
3681
	trace_f2fs_direct_IO_exit(inode, offset, count, rw, err);
3682

3683
	return err;
3684 3685
}

3686 3687
void f2fs_invalidate_page(struct page *page, unsigned int offset,
							unsigned int length)
3688 3689
{
	struct inode *inode = page->mapping->host;
3690
	struct f2fs_sb_info *sbi = F2FS_I_SB(inode);
3691

3692
	if (inode->i_ino >= F2FS_ROOT_INO(sbi) &&
3693
		(offset % PAGE_SIZE || length != PAGE_SIZE))
3694 3695
		return;

3696
	if (PageDirty(page)) {
3697
		if (inode->i_ino == F2FS_META_INO(sbi)) {
3698
			dec_page_count(sbi, F2FS_DIRTY_META);
3699
		} else if (inode->i_ino == F2FS_NODE_INO(sbi)) {
3700
			dec_page_count(sbi, F2FS_DIRTY_NODES);
3701
		} else {
3702
			inode_dec_dirty_pages(inode);
C
Chao Yu 已提交
3703
			f2fs_remove_dirty_inode(inode);
3704
		}
3705
	}
C
Chao Yu 已提交
3706

C
Chao Yu 已提交
3707 3708
	clear_cold_data(page);

C
Chao Yu 已提交
3709
	if (IS_ATOMIC_WRITTEN_PAGE(page))
C
Chao Yu 已提交
3710
		return f2fs_drop_inmem_page(inode, page);
C
Chao Yu 已提交
3711

3712
	f2fs_clear_page_private(page);
3713 3714
}

3715
int f2fs_release_page(struct page *page, gfp_t wait)
3716
{
3717 3718 3719 3720
	/* If this is dirty page, keep PagePrivate */
	if (PageDirty(page))
		return 0;

C
Chao Yu 已提交
3721 3722 3723 3724
	/* This is atomic written page, keep Private */
	if (IS_ATOMIC_WRITTEN_PAGE(page))
		return 0;

C
Chao Yu 已提交
3725
	clear_cold_data(page);
3726
	f2fs_clear_page_private(page);
3727
	return 1;
3728 3729 3730 3731
}

static int f2fs_set_data_page_dirty(struct page *page)
{
J
Jaegeuk Kim 已提交
3732
	struct inode *inode = page_file_mapping(page)->host;
3733

3734 3735
	trace_f2fs_set_page_dirty(page, DATA);

3736 3737
	if (!PageUptodate(page))
		SetPageUptodate(page);
J
Jaegeuk Kim 已提交
3738 3739
	if (PageSwapCache(page))
		return __set_page_dirty_nobuffers(page);
3740

C
Chao Yu 已提交
3741
	if (f2fs_is_atomic_file(inode) && !f2fs_is_commit_atomic_write(inode)) {
C
Chao Yu 已提交
3742
		if (!IS_ATOMIC_WRITTEN_PAGE(page)) {
C
Chao Yu 已提交
3743
			f2fs_register_inmem_page(inode, page);
C
Chao Yu 已提交
3744 3745 3746 3747 3748 3749 3750
			return 1;
		}
		/*
		 * Previously, this page has been registered, we just
		 * return here.
		 */
		return 0;
3751 3752
	}

3753
	if (!PageDirty(page)) {
3754
		__set_page_dirty_nobuffers(page);
C
Chao Yu 已提交
3755
		f2fs_update_dirty_page(inode, page);
3756 3757 3758 3759 3760
		return 1;
	}
	return 0;
}

C
Chao Yu 已提交
3761 3762 3763 3764 3765 3766 3767 3768 3769 3770 3771 3772 3773 3774 3775 3776 3777 3778 3779 3780 3781 3782 3783 3784 3785

static sector_t f2fs_bmap_compress(struct inode *inode, sector_t block)
{
#ifdef CONFIG_F2FS_FS_COMPRESSION
	struct dnode_of_data dn;
	sector_t start_idx, blknr = 0;
	int ret;

	start_idx = round_down(block, F2FS_I(inode)->i_cluster_size);

	set_new_dnode(&dn, inode, NULL, NULL, 0);
	ret = f2fs_get_dnode_of_data(&dn, start_idx, LOOKUP_NODE);
	if (ret)
		return 0;

	if (dn.data_blkaddr != COMPRESS_ADDR) {
		dn.ofs_in_node += block - start_idx;
		blknr = f2fs_data_blkaddr(&dn);
		if (!__is_valid_data_blkaddr(blknr))
			blknr = 0;
	}

	f2fs_put_dnode(&dn);
	return blknr;
#else
3786
	return 0;
C
Chao Yu 已提交
3787 3788 3789 3790
#endif
}


J
Jaegeuk Kim 已提交
3791 3792
static sector_t f2fs_bmap(struct address_space *mapping, sector_t block)
{
3793
	struct inode *inode = mapping->host;
C
Chao Yu 已提交
3794
	sector_t blknr = 0;
3795

J
Jaegeuk Kim 已提交
3796
	if (f2fs_has_inline_data(inode))
C
Chao Yu 已提交
3797
		goto out;
J
Jaegeuk Kim 已提交
3798 3799 3800 3801 3802

	/* make sure allocating whole blocks */
	if (mapping_tagged(mapping, PAGECACHE_TAG_DIRTY))
		filemap_write_and_wait(mapping);

3803 3804 3805
	/* Block number less than F2FS MAX BLOCKS */
	if (unlikely(block >= F2FS_I_SB(inode)->max_file_blocks))
		goto out;
C
Chao Yu 已提交
3806

3807 3808 3809
	if (f2fs_compressed_file(inode)) {
		blknr = f2fs_bmap_compress(inode, block);
	} else {
3810 3811 3812 3813 3814 3815 3816 3817 3818 3819
		struct f2fs_map_blocks map;

		memset(&map, 0, sizeof(map));
		map.m_lblk = block;
		map.m_len = 1;
		map.m_next_pgofs = NULL;
		map.m_seg_type = NO_CHECK_TYPE;

		if (!f2fs_map_blocks(inode, &map, 0, F2FS_GET_BLOCK_BMAP))
			blknr = map.m_pblk;
3820
	}
C
Chao Yu 已提交
3821 3822 3823
out:
	trace_f2fs_bmap(inode, block, blknr);
	return blknr;
3824 3825
}

3826 3827 3828 3829 3830 3831 3832 3833 3834 3835 3836 3837 3838
#ifdef CONFIG_MIGRATION
#include <linux/migrate.h>

int f2fs_migrate_page(struct address_space *mapping,
		struct page *newpage, struct page *page, enum migrate_mode mode)
{
	int rc, extra_count;
	struct f2fs_inode_info *fi = F2FS_I(mapping->host);
	bool atomic_written = IS_ATOMIC_WRITTEN_PAGE(page);

	BUG_ON(PageWriteback(page));

	/* migrating an atomic written page is safe with the inmem_lock hold */
3839 3840 3841 3842 3843 3844
	if (atomic_written) {
		if (mode != MIGRATE_SYNC)
			return -EBUSY;
		if (!mutex_trylock(&fi->inmem_lock))
			return -EAGAIN;
	}
3845

3846 3847
	/* one extra reference was held for atomic_write page */
	extra_count = atomic_written ? 1 : 0;
3848
	rc = migrate_page_move_mapping(mapping, newpage,
3849
				page, extra_count);
3850 3851 3852 3853 3854 3855 3856 3857 3858 3859 3860 3861 3862 3863 3864 3865 3866 3867
	if (rc != MIGRATEPAGE_SUCCESS) {
		if (atomic_written)
			mutex_unlock(&fi->inmem_lock);
		return rc;
	}

	if (atomic_written) {
		struct inmem_pages *cur;
		list_for_each_entry(cur, &fi->inmem_pages, list)
			if (cur->page == page) {
				cur->page = newpage;
				break;
			}
		mutex_unlock(&fi->inmem_lock);
		put_page(page);
		get_page(newpage);
	}

3868 3869 3870 3871
	if (PagePrivate(page)) {
		f2fs_set_page_private(newpage, page_private(page));
		f2fs_clear_page_private(page);
	}
3872

3873 3874 3875 3876
	if (mode != MIGRATE_SYNC_NO_COPY)
		migrate_page_copy(newpage, page);
	else
		migrate_page_states(newpage, page);
3877 3878 3879 3880 3881

	return MIGRATEPAGE_SUCCESS;
}
#endif

J
Jaegeuk Kim 已提交
3882
#ifdef CONFIG_SWAP
3883 3884 3885 3886 3887 3888 3889 3890 3891 3892 3893 3894
static int check_swap_activate_fast(struct swap_info_struct *sis,
				struct file *swap_file, sector_t *span)
{
	struct address_space *mapping = swap_file->f_mapping;
	struct inode *inode = mapping->host;
	sector_t cur_lblock;
	sector_t last_lblock;
	sector_t pblock;
	sector_t lowest_pblock = -1;
	sector_t highest_pblock = 0;
	int nr_extents = 0;
	unsigned long nr_pblocks;
3895
	u64 len;
3896 3897 3898 3899 3900 3901 3902
	int ret;

	/*
	 * Map all the blocks into the extent list.  This code doesn't try
	 * to be very smart.
	 */
	cur_lblock = 0;
3903
	last_lblock = bytes_to_blks(inode, i_size_read(inode));
3904 3905 3906
	len = i_size_read(inode);

	while (cur_lblock <= last_lblock && cur_lblock < sis->max) {
3907
		struct f2fs_map_blocks map;
3908 3909 3910 3911
		pgoff_t next_pgofs;

		cond_resched();

3912 3913 3914 3915 3916
		memset(&map, 0, sizeof(map));
		map.m_lblk = cur_lblock;
		map.m_len = bytes_to_blks(inode, len) - cur_lblock;
		map.m_next_pgofs = &next_pgofs;
		map.m_seg_type = NO_CHECK_TYPE;
3917

3918
		ret = f2fs_map_blocks(inode, &map, 0, F2FS_GET_BLOCK_FIEMAP);
3919 3920 3921 3922
		if (ret)
			goto err_out;

		/* hole */
3923
		if (!(map.m_flags & F2FS_MAP_FLAGS))
3924 3925
			goto err_out;

3926 3927
		pblock = map.m_pblk;
		nr_pblocks = map.m_len;
3928 3929 3930 3931 3932 3933 3934 3935 3936 3937 3938 3939 3940 3941 3942 3943 3944 3945 3946 3947 3948 3949 3950 3951 3952 3953 3954 3955 3956 3957 3958 3959 3960 3961

		if (cur_lblock + nr_pblocks >= sis->max)
			nr_pblocks = sis->max - cur_lblock;

		if (cur_lblock) {	/* exclude the header page */
			if (pblock < lowest_pblock)
				lowest_pblock = pblock;
			if (pblock + nr_pblocks - 1 > highest_pblock)
				highest_pblock = pblock + nr_pblocks - 1;
		}

		/*
		 * We found a PAGE_SIZE-length, PAGE_SIZE-aligned run of blocks
		 */
		ret = add_swap_extent(sis, cur_lblock, nr_pblocks, pblock);
		if (ret < 0)
			goto out;
		nr_extents += ret;
		cur_lblock += nr_pblocks;
	}
	ret = nr_extents;
	*span = 1 + highest_pblock - lowest_pblock;
	if (cur_lblock == 0)
		cur_lblock = 1;	/* force Empty message */
	sis->max = cur_lblock;
	sis->pages = cur_lblock - 1;
	sis->highest_bit = cur_lblock - 1;
out:
	return ret;
err_out:
	pr_err("swapon: swapfile has holes\n");
	return -EINVAL;
}

J
Jaegeuk Kim 已提交
3962
/* Copied from generic_swapfile_activate() to check any holes */
C
Chao Yu 已提交
3963 3964
static int check_swap_activate(struct swap_info_struct *sis,
				struct file *swap_file, sector_t *span)
J
Jaegeuk Kim 已提交
3965 3966 3967 3968 3969 3970 3971 3972 3973
{
	struct address_space *mapping = swap_file->f_mapping;
	struct inode *inode = mapping->host;
	unsigned blocks_per_page;
	unsigned long page_no;
	sector_t probe_block;
	sector_t last_block;
	sector_t lowest_block = -1;
	sector_t highest_block = 0;
C
Chao Yu 已提交
3974 3975
	int nr_extents = 0;
	int ret;
J
Jaegeuk Kim 已提交
3976

3977 3978 3979
	if (PAGE_SIZE == F2FS_BLKSIZE)
		return check_swap_activate_fast(sis, swap_file, span);

3980
	blocks_per_page = bytes_to_blks(inode, PAGE_SIZE);
J
Jaegeuk Kim 已提交
3981 3982 3983 3984 3985 3986 3987

	/*
	 * Map all the blocks into the extent list.  This code doesn't try
	 * to be very smart.
	 */
	probe_block = 0;
	page_no = 0;
3988
	last_block = bytes_to_blks(inode, i_size_read(inode));
C
Chao Yu 已提交
3989 3990
	while ((probe_block + blocks_per_page) <= last_block &&
			page_no < sis->max) {
J
Jaegeuk Kim 已提交
3991 3992
		unsigned block_in_page;
		sector_t first_block;
3993 3994
		sector_t block = 0;
		int	 err = 0;
J
Jaegeuk Kim 已提交
3995 3996 3997

		cond_resched();

3998 3999 4000
		block = probe_block;
		err = bmap(inode, &block);
		if (err || !block)
J
Jaegeuk Kim 已提交
4001
			goto bad_bmap;
4002
		first_block = block;
J
Jaegeuk Kim 已提交
4003 4004 4005 4006 4007 4008 4009 4010 4011 4012 4013 4014

		/*
		 * It must be PAGE_SIZE aligned on-disk
		 */
		if (first_block & (blocks_per_page - 1)) {
			probe_block++;
			goto reprobe;
		}

		for (block_in_page = 1; block_in_page < blocks_per_page;
					block_in_page++) {

4015 4016 4017 4018
			block = probe_block + block_in_page;
			err = bmap(inode, &block);

			if (err || !block)
J
Jaegeuk Kim 已提交
4019
				goto bad_bmap;
4020

J
Jaegeuk Kim 已提交
4021 4022 4023 4024 4025 4026 4027
			if (block != first_block + block_in_page) {
				/* Discontiguity */
				probe_block++;
				goto reprobe;
			}
		}

4028
		first_block >>= (PAGE_SHIFT - inode->i_blkbits);
J
Jaegeuk Kim 已提交
4029 4030 4031 4032 4033 4034 4035
		if (page_no) {	/* exclude the header page */
			if (first_block < lowest_block)
				lowest_block = first_block;
			if (first_block > highest_block)
				highest_block = first_block;
		}

C
Chao Yu 已提交
4036 4037 4038 4039 4040 4041 4042
		/*
		 * We found a PAGE_SIZE-length, PAGE_SIZE-aligned run of blocks
		 */
		ret = add_swap_extent(sis, page_no, 1, first_block);
		if (ret < 0)
			goto out;
		nr_extents += ret;
J
Jaegeuk Kim 已提交
4043 4044 4045 4046 4047
		page_no++;
		probe_block += blocks_per_page;
reprobe:
		continue;
	}
C
Chao Yu 已提交
4048 4049 4050 4051 4052 4053 4054 4055 4056
	ret = nr_extents;
	*span = 1 + highest_block - lowest_block;
	if (page_no == 0)
		page_no = 1;	/* force Empty message */
	sis->max = page_no;
	sis->pages = page_no - 1;
	sis->highest_bit = page_no - 1;
out:
	return ret;
J
Jaegeuk Kim 已提交
4057 4058 4059 4060 4061 4062 4063 4064 4065 4066 4067 4068 4069 4070 4071 4072 4073 4074 4075 4076 4077
bad_bmap:
	pr_err("swapon: swapfile has holes\n");
	return -EINVAL;
}

static int f2fs_swap_activate(struct swap_info_struct *sis, struct file *file,
				sector_t *span)
{
	struct inode *inode = file_inode(file);
	int ret;

	if (!S_ISREG(inode->i_mode))
		return -EINVAL;

	if (f2fs_readonly(F2FS_I_SB(inode)->sb))
		return -EROFS;

	ret = f2fs_convert_inline_inode(inode);
	if (ret)
		return ret;

4078
	if (!f2fs_disable_compressed_file(inode))
C
Chao Yu 已提交
4079 4080
		return -EINVAL;

C
Chao Yu 已提交
4081 4082
	ret = check_swap_activate(sis, file, span);
	if (ret < 0)
J
Jaegeuk Kim 已提交
4083 4084 4085 4086 4087
		return ret;

	set_inode_flag(inode, FI_PIN_FILE);
	f2fs_precache_extents(inode);
	f2fs_update_time(F2FS_I_SB(inode), REQ_TIME);
C
Chao Yu 已提交
4088
	return ret;
J
Jaegeuk Kim 已提交
4089 4090 4091 4092 4093 4094 4095 4096 4097 4098 4099 4100 4101 4102 4103 4104 4105 4106 4107 4108
}

static void f2fs_swap_deactivate(struct file *file)
{
	struct inode *inode = file_inode(file);

	clear_inode_flag(inode, FI_PIN_FILE);
}
#else
static int f2fs_swap_activate(struct swap_info_struct *sis, struct file *file,
				sector_t *span)
{
	return -EOPNOTSUPP;
}

static void f2fs_swap_deactivate(struct file *file)
{
}
#endif

4109 4110
const struct address_space_operations f2fs_dblock_aops = {
	.readpage	= f2fs_read_data_page,
4111
	.readahead	= f2fs_readahead,
4112 4113 4114
	.writepage	= f2fs_write_data_page,
	.writepages	= f2fs_write_data_pages,
	.write_begin	= f2fs_write_begin,
4115
	.write_end	= f2fs_write_end,
4116
	.set_page_dirty	= f2fs_set_data_page_dirty,
4117 4118
	.invalidatepage	= f2fs_invalidate_page,
	.releasepage	= f2fs_release_page,
4119
	.direct_IO	= f2fs_direct_IO,
J
Jaegeuk Kim 已提交
4120
	.bmap		= f2fs_bmap,
J
Jaegeuk Kim 已提交
4121 4122
	.swap_activate  = f2fs_swap_activate,
	.swap_deactivate = f2fs_swap_deactivate,
4123 4124 4125
#ifdef CONFIG_MIGRATION
	.migratepage    = f2fs_migrate_page,
#endif
4126
};
4127

M
Matthew Wilcox 已提交
4128
void f2fs_clear_page_cache_dirty_tag(struct page *page)
4129 4130 4131 4132 4133
{
	struct address_space *mapping = page_mapping(page);
	unsigned long flags;

	xa_lock_irqsave(&mapping->i_pages, flags);
M
Matthew Wilcox 已提交
4134
	__xa_clear_mark(&mapping->i_pages, page_index(page),
4135 4136 4137 4138
						PAGECACHE_TAG_DIRTY);
	xa_unlock_irqrestore(&mapping->i_pages, flags);
}

4139 4140
int __init f2fs_init_post_read_processing(void)
{
E
Eric Biggers 已提交
4141 4142 4143
	bio_post_read_ctx_cache =
		kmem_cache_create("f2fs_bio_post_read_ctx",
				  sizeof(struct bio_post_read_ctx), 0, 0, NULL);
4144 4145 4146 4147 4148 4149 4150 4151 4152 4153 4154 4155 4156 4157 4158
	if (!bio_post_read_ctx_cache)
		goto fail;
	bio_post_read_ctx_pool =
		mempool_create_slab_pool(NUM_PREALLOC_POST_READ_CTXS,
					 bio_post_read_ctx_cache);
	if (!bio_post_read_ctx_pool)
		goto fail_free_cache;
	return 0;

fail_free_cache:
	kmem_cache_destroy(bio_post_read_ctx_cache);
fail:
	return -ENOMEM;
}

C
Chao Yu 已提交
4159
void f2fs_destroy_post_read_processing(void)
4160 4161 4162 4163
{
	mempool_destroy(bio_post_read_ctx_pool);
	kmem_cache_destroy(bio_post_read_ctx_cache);
}
C
Chao Yu 已提交
4164

C
Chao Yu 已提交
4165 4166 4167 4168 4169 4170 4171 4172 4173 4174 4175 4176 4177 4178 4179 4180 4181 4182 4183 4184 4185
int f2fs_init_post_read_wq(struct f2fs_sb_info *sbi)
{
	if (!f2fs_sb_has_encrypt(sbi) &&
		!f2fs_sb_has_verity(sbi) &&
		!f2fs_sb_has_compression(sbi))
		return 0;

	sbi->post_read_wq = alloc_workqueue("f2fs_post_read_wq",
						 WQ_UNBOUND | WQ_HIGHPRI,
						 num_online_cpus());
	if (!sbi->post_read_wq)
		return -ENOMEM;
	return 0;
}

void f2fs_destroy_post_read_wq(struct f2fs_sb_info *sbi)
{
	if (sbi->post_read_wq)
		destroy_workqueue(sbi->post_read_wq);
}

C
Chao Yu 已提交
4186 4187
int __init f2fs_init_bio_entry_cache(void)
{
4188
	bio_entry_slab = f2fs_kmem_cache_create("f2fs_bio_entry_slab",
C
Chao Yu 已提交
4189 4190 4191 4192 4193 4194
			sizeof(struct bio_entry));
	if (!bio_entry_slab)
		return -ENOMEM;
	return 0;
}

C
Chao Yu 已提交
4195
void f2fs_destroy_bio_entry_cache(void)
C
Chao Yu 已提交
4196 4197 4198
{
	kmem_cache_destroy(bio_entry_slab);
}