提交 058e09e6 编写于 作者: Q Qu Wenruo 提交者: David Sterba

btrfs: scrub: introduce a writeback helper for scrub_stripe

Add a new helper, scrub_write_sectors(), to submit write bios for
specified sectors to the target disk.

There are several differences compared to read path:

- Utilize btrfs_submit_scrub_write()
  Now we still rely on the @mirror_num based writeback, but the
  requirement is also a little different than regular writeback or read,
  thus we have to call btrfs_submit_scrub_write().

- We cannot write the full stripe back
  We can only write the sectors we have.  There will be two call sites
  later, one for repaired sectors, one for all utilized sectors of
  dev-replace.

  Thus the callers should specify their own write_bitmap.

This function only submit the bios, will not wait for them unless for
zoned case.

Caller must explicitly wait for the IO to finish.
Signed-off-by: NQu Wenruo <wqu@suse.com>
Reviewed-by: NDavid Sterba <dsterba@suse.com>
Signed-off-by: NDavid Sterba <dsterba@suse.com>
上级 9ecb5ef5
......@@ -152,6 +152,12 @@ struct scrub_stripe {
unsigned long csum_error_bitmap;
unsigned long meta_error_bitmap;
/* For writeback (repair or replace) error reporting. */
unsigned long write_error_bitmap;
/* Writeback can be concurrent, thus we need to protect the bitmap. */
spinlock_t write_error_lock;
/*
* Checksum for the whole stripe if this stripe is inside a data block
* group.
......@@ -386,6 +392,7 @@ int init_scrub_stripe(struct btrfs_fs_info *fs_info, struct scrub_stripe *stripe
init_waitqueue_head(&stripe->io_wait);
init_waitqueue_head(&stripe->repair_wait);
atomic_set(&stripe->pending_io, 0);
spin_lock_init(&stripe->write_error_lock);
ret = btrfs_alloc_page_array(SCRUB_STRIPE_PAGES, stripe->pages);
if (ret < 0)
......@@ -2541,6 +2548,92 @@ void scrub_read_endio(struct btrfs_bio *bbio)
}
}
static void scrub_write_endio(struct btrfs_bio *bbio)
{
struct scrub_stripe *stripe = bbio->private;
struct btrfs_fs_info *fs_info = stripe->bg->fs_info;
struct bio_vec *bvec;
int sector_nr = calc_sector_number(stripe, bio_first_bvec_all(&bbio->bio));
u32 bio_size = 0;
int i;
bio_for_each_bvec_all(bvec, &bbio->bio, i)
bio_size += bvec->bv_len;
if (bbio->bio.bi_status) {
unsigned long flags;
spin_lock_irqsave(&stripe->write_error_lock, flags);
bitmap_set(&stripe->write_error_bitmap, sector_nr,
bio_size >> fs_info->sectorsize_bits);
spin_unlock_irqrestore(&stripe->write_error_lock, flags);
}
bio_put(&bbio->bio);
if (atomic_dec_and_test(&stripe->pending_io))
wake_up(&stripe->io_wait);
}
/*
* Submit the write bio(s) for the sectors specified by @write_bitmap.
*
* Here we utilize btrfs_submit_repair_write(), which has some extra benefits:
*
* - Only needs logical bytenr and mirror_num
* Just like the scrub read path
*
* - Would only result in writes to the specified mirror
* Unlike the regular writeback path, which would write back to all stripes
*
* - Handle dev-replace and read-repair writeback differently
*/
void scrub_write_sectors(struct scrub_ctx *sctx, struct scrub_stripe *stripe,
unsigned long write_bitmap, bool dev_replace)
{
struct btrfs_fs_info *fs_info = stripe->bg->fs_info;
struct btrfs_bio *bbio = NULL;
const bool zoned = btrfs_is_zoned(fs_info);
int sector_nr;
for_each_set_bit(sector_nr, &write_bitmap, stripe->nr_sectors) {
struct page *page = scrub_stripe_get_page(stripe, sector_nr);
unsigned int pgoff = scrub_stripe_get_page_offset(stripe, sector_nr);
int ret;
/* We should only writeback sectors covered by an extent. */
ASSERT(test_bit(sector_nr, &stripe->extent_sector_bitmap));
/* Cannot merge with previous sector, submit the current one. */
if (bbio && sector_nr && !test_bit(sector_nr - 1, &write_bitmap)) {
fill_writer_pointer_gap(sctx, stripe->physical +
(sector_nr << fs_info->sectorsize_bits));
atomic_inc(&stripe->pending_io);
btrfs_submit_repair_write(bbio, stripe->mirror_num, dev_replace);
/* For zoned writeback, queue depth must be 1. */
if (zoned)
wait_scrub_stripe_io(stripe);
bbio = NULL;
}
if (!bbio) {
bbio = btrfs_bio_alloc(stripe->nr_sectors, REQ_OP_WRITE,
fs_info, scrub_write_endio, stripe);
bbio->bio.bi_iter.bi_sector = (stripe->logical +
(sector_nr << fs_info->sectorsize_bits)) >>
SECTOR_SHIFT;
}
ret = bio_add_page(&bbio->bio, page, fs_info->sectorsize, pgoff);
ASSERT(ret == fs_info->sectorsize);
}
if (bbio) {
fill_writer_pointer_gap(sctx, bbio->bio.bi_iter.bi_sector <<
SECTOR_SHIFT);
atomic_inc(&stripe->pending_io);
btrfs_submit_repair_write(bbio, stripe->mirror_num, dev_replace);
if (zoned)
wait_scrub_stripe_io(stripe);
}
}
static int scrub_checksum_tree_block(struct scrub_block *sblock)
{
struct scrub_ctx *sctx = sblock->sctx;
......
......@@ -24,5 +24,8 @@ int scrub_find_fill_first_stripe(struct btrfs_block_group *bg,
int mirror_num, u64 logical_start,
u32 logical_len, struct scrub_stripe *stripe);
void scrub_read_endio(struct btrfs_bio *bbio);
void scrub_write_sectors(struct scrub_ctx *sctx,
struct scrub_stripe *stripe,
unsigned long write_bitmap, bool dev_replace);
#endif
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册