提交 4845e44f 编写于 作者: C Chris Mason

Btrfs: rework O_DIRECT enospc handling

This changes O_DIRECT write code to mark extents as delalloc
while it is processing them.  Yan Zheng has reworked the
enospc accounting based on tracking delalloc extents and
this makes it much easier to track enospc in the O_DIRECT code.

There are a few space cases with the O_DIRECT code though,
it only sets the EXTENT_DELALLOC bits, instead of doing
EXTENT_DELALLOC | EXTENT_DIRTY | EXTENT_UPTODATE, because
we don't want to mess with clearing the dirty and uptodate
bits when things go wrong.  This is important because there
are no pages in the page cache, so any extent state structs
that we put in the tree won't get freed by releasepage.  We have
to clear them ourselves as the DIO ends.

With this commit, we reserve space at in btrfs_file_aio_write,
and then as each btrfs_direct_IO call progresses it sets
EXTENT_DELALLOC on the range.

btrfs_get_blocks_direct is responsible for clearing the delalloc
at the same time it drops the extent lock.
Signed-off-by: NChris Mason <chris.mason@oracle.com>
上级 eaf25d93
...@@ -135,7 +135,7 @@ static struct extent_state *alloc_extent_state(gfp_t mask) ...@@ -135,7 +135,7 @@ static struct extent_state *alloc_extent_state(gfp_t mask)
return state; return state;
} }
static void free_extent_state(struct extent_state *state) void free_extent_state(struct extent_state *state)
{ {
if (!state) if (!state)
return; return;
...@@ -745,10 +745,9 @@ static void cache_state(struct extent_state *state, ...@@ -745,10 +745,9 @@ static void cache_state(struct extent_state *state,
* [start, end] is inclusive This takes the tree lock. * [start, end] is inclusive This takes the tree lock.
*/ */
static int set_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, int set_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
int bits, int exclusive_bits, u64 *failed_start, int bits, int exclusive_bits, u64 *failed_start,
struct extent_state **cached_state, struct extent_state **cached_state, gfp_t mask)
gfp_t mask)
{ {
struct extent_state *state; struct extent_state *state;
struct extent_state *prealloc = NULL; struct extent_state *prealloc = NULL;
......
...@@ -178,6 +178,7 @@ u64 count_range_bits(struct extent_io_tree *tree, ...@@ -178,6 +178,7 @@ u64 count_range_bits(struct extent_io_tree *tree,
u64 *start, u64 search_end, u64 *start, u64 search_end,
u64 max_bytes, unsigned long bits); u64 max_bytes, unsigned long bits);
void free_extent_state(struct extent_state *state);
int test_range_bit(struct extent_io_tree *tree, u64 start, u64 end, int test_range_bit(struct extent_io_tree *tree, u64 start, u64 end,
int bits, int filled, struct extent_state *cached_state); int bits, int filled, struct extent_state *cached_state);
int clear_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, int clear_extent_bits(struct extent_io_tree *tree, u64 start, u64 end,
...@@ -187,6 +188,9 @@ int clear_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, ...@@ -187,6 +188,9 @@ int clear_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
gfp_t mask); gfp_t mask);
int set_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, int set_extent_bits(struct extent_io_tree *tree, u64 start, u64 end,
int bits, gfp_t mask); int bits, gfp_t mask);
int set_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
int bits, int exclusive_bits, u64 *failed_start,
struct extent_state **cached_state, gfp_t mask);
int set_extent_uptodate(struct extent_io_tree *tree, u64 start, u64 end, int set_extent_uptodate(struct extent_io_tree *tree, u64 start, u64 end,
gfp_t mask); gfp_t mask);
int set_extent_new(struct extent_io_tree *tree, u64 start, u64 end, int set_extent_new(struct extent_io_tree *tree, u64 start, u64 end,
......
...@@ -909,13 +909,6 @@ static ssize_t btrfs_file_aio_write(struct kiocb *iocb, ...@@ -909,13 +909,6 @@ static ssize_t btrfs_file_aio_write(struct kiocb *iocb,
} }
if (num_written < 0) { if (num_written < 0) {
if (num_written != -EIOCBQUEUED) {
/*
* aio land will take care of releasing the
* delalloc
*/
btrfs_delalloc_release_space(inode, count);
}
ret = num_written; ret = num_written;
num_written = 0; num_written = 0;
goto out; goto out;
...@@ -924,13 +917,6 @@ static ssize_t btrfs_file_aio_write(struct kiocb *iocb, ...@@ -924,13 +917,6 @@ static ssize_t btrfs_file_aio_write(struct kiocb *iocb,
pos = *ppos; pos = *ppos;
goto out; goto out;
} }
/*
* the buffered IO will reserve bytes for the rest of the
* range, don't double count them here
*/
btrfs_delalloc_release_space(inode, count - num_written);
/* /*
* We are going to do buffered for the rest of the range, so we * We are going to do buffered for the rest of the range, so we
* need to make sure to invalidate the buffered pages when we're * need to make sure to invalidate the buffered pages when we're
......
...@@ -5327,8 +5327,9 @@ static int btrfs_get_blocks_direct(struct inode *inode, sector_t iblock, ...@@ -5327,8 +5327,9 @@ static int btrfs_get_blocks_direct(struct inode *inode, sector_t iblock,
return PTR_ERR(em); return PTR_ERR(em);
len = min(len, em->block_len); len = min(len, em->block_len);
} }
unlock_extent(&BTRFS_I(inode)->io_tree, start, start + len - 1, clear_extent_bit(&BTRFS_I(inode)->io_tree, start, start + len - 1,
GFP_NOFS); EXTENT_LOCKED | EXTENT_DELALLOC | EXTENT_DIRTY, 1,
0, NULL, GFP_NOFS);
map: map:
bh_result->b_blocknr = (em->block_start + (start - em->start)) >> bh_result->b_blocknr = (em->block_start + (start - em->start)) >>
inode->i_blkbits; inode->i_blkbits;
...@@ -5596,14 +5597,18 @@ static ssize_t btrfs_direct_IO(int rw, struct kiocb *iocb, ...@@ -5596,14 +5597,18 @@ static ssize_t btrfs_direct_IO(int rw, struct kiocb *iocb,
struct file *file = iocb->ki_filp; struct file *file = iocb->ki_filp;
struct inode *inode = file->f_mapping->host; struct inode *inode = file->f_mapping->host;
struct btrfs_ordered_extent *ordered; struct btrfs_ordered_extent *ordered;
struct extent_state *cached_state = NULL;
u64 lockstart, lockend; u64 lockstart, lockend;
ssize_t ret; ssize_t ret;
int writing = rw & WRITE;
int write_bits = 0;
lockstart = offset; lockstart = offset;
lockend = offset + iov_length(iov, nr_segs) - 1; lockend = offset + iov_length(iov, nr_segs) - 1;
while (1) { while (1) {
lock_extent(&BTRFS_I(inode)->io_tree, lockstart, lockend, lock_extent_bits(&BTRFS_I(inode)->io_tree, lockstart, lockend,
GFP_NOFS); 0, &cached_state, GFP_NOFS);
/* /*
* We're concerned with the entire range that we're going to be * We're concerned with the entire range that we're going to be
* doing DIO to, so we need to make sure theres no ordered * doing DIO to, so we need to make sure theres no ordered
...@@ -5613,29 +5618,54 @@ static ssize_t btrfs_direct_IO(int rw, struct kiocb *iocb, ...@@ -5613,29 +5618,54 @@ static ssize_t btrfs_direct_IO(int rw, struct kiocb *iocb,
lockend - lockstart + 1); lockend - lockstart + 1);
if (!ordered) if (!ordered)
break; break;
unlock_extent(&BTRFS_I(inode)->io_tree, lockstart, lockend, unlock_extent_cached(&BTRFS_I(inode)->io_tree, lockstart, lockend,
GFP_NOFS); &cached_state, GFP_NOFS);
btrfs_start_ordered_extent(inode, ordered, 1); btrfs_start_ordered_extent(inode, ordered, 1);
btrfs_put_ordered_extent(ordered); btrfs_put_ordered_extent(ordered);
cond_resched(); cond_resched();
} }
/*
* we don't use btrfs_set_extent_delalloc because we don't want
* the dirty or uptodate bits
*/
if (writing) {
write_bits = EXTENT_DELALLOC | EXTENT_DO_ACCOUNTING;
ret = set_extent_bit(&BTRFS_I(inode)->io_tree, lockstart, lockend,
EXTENT_DELALLOC, 0, NULL, &cached_state,
GFP_NOFS);
if (ret) {
clear_extent_bit(&BTRFS_I(inode)->io_tree, lockstart,
lockend, EXTENT_LOCKED | write_bits,
1, 0, &cached_state, GFP_NOFS);
goto out;
}
}
free_extent_state(cached_state);
cached_state = NULL;
ret = __blockdev_direct_IO(rw, iocb, inode, NULL, iov, offset, nr_segs, ret = __blockdev_direct_IO(rw, iocb, inode, NULL, iov, offset, nr_segs,
btrfs_get_blocks_direct, NULL, btrfs_get_blocks_direct, NULL,
btrfs_submit_direct, 0); btrfs_submit_direct, 0);
if (ret < 0 && ret != -EIOCBQUEUED) { if (ret < 0 && ret != -EIOCBQUEUED) {
unlock_extent(&BTRFS_I(inode)->io_tree, offset, clear_extent_bit(&BTRFS_I(inode)->io_tree, offset,
offset + iov_length(iov, nr_segs) - 1, GFP_NOFS); offset + iov_length(iov, nr_segs) - 1,
EXTENT_LOCKED | write_bits, 1, 0,
&cached_state, GFP_NOFS);
} else if (ret >= 0 && ret < iov_length(iov, nr_segs)) { } else if (ret >= 0 && ret < iov_length(iov, nr_segs)) {
/* /*
* We're falling back to buffered, unlock the section we didn't * We're falling back to buffered, unlock the section we didn't
* do IO on. * do IO on.
*/ */
unlock_extent(&BTRFS_I(inode)->io_tree, offset + ret, clear_extent_bit(&BTRFS_I(inode)->io_tree, offset + ret,
offset + iov_length(iov, nr_segs) - 1, GFP_NOFS); offset + iov_length(iov, nr_segs) - 1,
EXTENT_LOCKED | write_bits, 1, 0,
&cached_state, GFP_NOFS);
} }
out:
free_extent_state(cached_state);
return ret; return ret;
} }
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册