提交 21b40200 编写于 作者: K Kent Overstreet 提交者: Linus Torvalds

aio: use flush_dcache_page()

This wasn't causing problems before because it's not needed on x86, but
it is needed on other architectures.
Signed-off-by: NKent Overstreet <koverstreet@google.com>
Cc: Zach Brown <zab@redhat.com>
Cc: Felipe Balbi <balbi@ti.com>
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: Mark Fasheh <mfasheh@suse.com>
Cc: Joel Becker <jlbec@evilplan.org>
Cc: Rusty Russell <rusty@rustcorp.com.au>
Cc: Jens Axboe <axboe@kernel.dk>
Cc: Asai Thambi S P <asamymuthupa@micron.com>
Cc: Selvan Mani <smani@micron.com>
Cc: Sam Bradshaw <sbradshaw@micron.com>
Cc: Jeff Moyer <jmoyer@redhat.com>
Cc: Al Viro <viro@zeniv.linux.org.uk>
Cc: Benjamin LaHaise <bcrl@kvack.org>
Cc: Theodore Ts'o <tytso@mit.edu>
Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
上级 a31ad380
...@@ -211,33 +211,15 @@ static int aio_setup_ring(struct kioctx *ctx) ...@@ -211,33 +211,15 @@ static int aio_setup_ring(struct kioctx *ctx)
ring->incompat_features = AIO_RING_INCOMPAT_FEATURES; ring->incompat_features = AIO_RING_INCOMPAT_FEATURES;
ring->header_length = sizeof(struct aio_ring); ring->header_length = sizeof(struct aio_ring);
kunmap_atomic(ring); kunmap_atomic(ring);
flush_dcache_page(info->ring_pages[0]);
return 0; return 0;
} }
/* aio_ring_event: returns a pointer to the event at the given index from
* kmap_atomic(). Release the pointer with put_aio_ring_event();
*/
#define AIO_EVENTS_PER_PAGE (PAGE_SIZE / sizeof(struct io_event)) #define AIO_EVENTS_PER_PAGE (PAGE_SIZE / sizeof(struct io_event))
#define AIO_EVENTS_FIRST_PAGE ((PAGE_SIZE - sizeof(struct aio_ring)) / sizeof(struct io_event)) #define AIO_EVENTS_FIRST_PAGE ((PAGE_SIZE - sizeof(struct aio_ring)) / sizeof(struct io_event))
#define AIO_EVENTS_OFFSET (AIO_EVENTS_PER_PAGE - AIO_EVENTS_FIRST_PAGE) #define AIO_EVENTS_OFFSET (AIO_EVENTS_PER_PAGE - AIO_EVENTS_FIRST_PAGE)
#define aio_ring_event(info, nr) ({ \
unsigned pos = (nr) + AIO_EVENTS_OFFSET; \
struct io_event *__event; \
__event = kmap_atomic( \
(info)->ring_pages[pos / AIO_EVENTS_PER_PAGE]); \
__event += pos % AIO_EVENTS_PER_PAGE; \
__event; \
})
#define put_aio_ring_event(event) do { \
struct io_event *__event = (event); \
(void)__event; \
kunmap_atomic((void *)((unsigned long)__event & PAGE_MASK)); \
} while(0)
static int kiocb_cancel(struct kioctx *ctx, struct kiocb *kiocb, static int kiocb_cancel(struct kioctx *ctx, struct kiocb *kiocb,
struct io_event *res) struct io_event *res)
{ {
...@@ -649,9 +631,9 @@ void aio_complete(struct kiocb *iocb, long res, long res2) ...@@ -649,9 +631,9 @@ void aio_complete(struct kiocb *iocb, long res, long res2)
struct kioctx *ctx = iocb->ki_ctx; struct kioctx *ctx = iocb->ki_ctx;
struct aio_ring_info *info; struct aio_ring_info *info;
struct aio_ring *ring; struct aio_ring *ring;
struct io_event *event; struct io_event *ev_page, *event;
unsigned long flags; unsigned long flags;
unsigned long tail; unsigned tail, pos;
/* /*
* Special case handling for sync iocbs: * Special case handling for sync iocbs:
...@@ -690,19 +672,24 @@ void aio_complete(struct kiocb *iocb, long res, long res2) ...@@ -690,19 +672,24 @@ void aio_complete(struct kiocb *iocb, long res, long res2)
if (kiocbIsCancelled(iocb)) if (kiocbIsCancelled(iocb))
goto put_rq; goto put_rq;
ring = kmap_atomic(info->ring_pages[0]);
tail = info->tail; tail = info->tail;
event = aio_ring_event(info, tail); pos = tail + AIO_EVENTS_OFFSET;
if (++tail >= info->nr) if (++tail >= info->nr)
tail = 0; tail = 0;
ev_page = kmap_atomic(info->ring_pages[pos / AIO_EVENTS_PER_PAGE]);
event = ev_page + pos % AIO_EVENTS_PER_PAGE;
event->obj = (u64)(unsigned long)iocb->ki_obj.user; event->obj = (u64)(unsigned long)iocb->ki_obj.user;
event->data = iocb->ki_user_data; event->data = iocb->ki_user_data;
event->res = res; event->res = res;
event->res2 = res2; event->res2 = res2;
pr_debug("%p[%lu]: %p: %p %Lx %lx %lx\n", kunmap_atomic(ev_page);
flush_dcache_page(info->ring_pages[pos / AIO_EVENTS_PER_PAGE]);
pr_debug("%p[%u]: %p: %p %Lx %lx %lx\n",
ctx, tail, iocb, iocb->ki_obj.user, iocb->ki_user_data, ctx, tail, iocb, iocb->ki_obj.user, iocb->ki_user_data,
res, res2); res, res2);
...@@ -712,12 +699,13 @@ void aio_complete(struct kiocb *iocb, long res, long res2) ...@@ -712,12 +699,13 @@ void aio_complete(struct kiocb *iocb, long res, long res2)
smp_wmb(); /* make event visible before updating tail */ smp_wmb(); /* make event visible before updating tail */
info->tail = tail; info->tail = tail;
ring->tail = tail;
put_aio_ring_event(event); ring = kmap_atomic(info->ring_pages[0]);
ring->tail = tail;
kunmap_atomic(ring); kunmap_atomic(ring);
flush_dcache_page(info->ring_pages[0]);
pr_debug("added to ring %p at [%lu]\n", iocb, tail); pr_debug("added to ring %p at [%u]\n", iocb, tail);
/* /*
* Check if the user asked us to deliver the result through an * Check if the user asked us to deliver the result through an
...@@ -807,6 +795,7 @@ static long aio_read_events_ring(struct kioctx *ctx, ...@@ -807,6 +795,7 @@ static long aio_read_events_ring(struct kioctx *ctx,
ring = kmap_atomic(info->ring_pages[0]); ring = kmap_atomic(info->ring_pages[0]);
ring->head = head; ring->head = head;
kunmap_atomic(ring); kunmap_atomic(ring);
flush_dcache_page(info->ring_pages[0]);
pr_debug("%li h%u t%u\n", ret, head, info->tail); pr_debug("%li h%u t%u\n", ret, head, info->tail);
out: out:
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册