Commit 8459d86a authored by Zach Brown's avatar Zach Brown Committed by Linus Torvalds

[PATCH] dio: only call aio_complete() after returning -EIOCBQUEUED

The only time it is safe to call aio_complete() is when the ->ki_retry
function returns -EIOCBQUEUED to the AIO core.  direct_io_worker() has
historically done this by relying on its caller to translate positive return
codes into -EIOCBQUEUED for the aio case.  It did this by trying to keep
conditionals in sync.  direct_io_worker() knew when finished_one_bio() was
going to call aio_complete().  It would reverse the test and wait and free the
dio in the cases it thought that finished_one_bio() wasn't going to.

Not surprisingly, it ended up getting it wrong.  'ret' could be a negative
errno from the submission path but it failed to communicate this to
finished_one_bio().  direct_io_worker() would return < 0, it's callers
wouldn't raise -EIOCBQUEUED, and aio_complete() would be called.  In the
future finished_one_bio()'s tests wouldn't reflect this and aio_complete()
would be called for a second time which can manifest as an oops.

The previous cleanups have whittled the sync and async completion paths down
to the point where we can collapse them and clearly reassert the invariant
that we must only call aio_complete() after returning -EIOCBQUEUED.
direct_io_worker() will only return -EIOCBQUEUED when it is not the last to
drop the dio refcount and the aio bio completion path will only call
aio_complete() when it is the last to drop the dio refcount.
direct_io_worker() can ensure that it is the last to drop the reference count
by waiting for bios to drain.  It does this for sync ops, of course, and for
partial dio writes that must fall back to buffered and for aio ops that saw
errors during submission.

This means that operations that end up waiting, even if they were issued as
aio ops, will not call aio_complete() from dio.  Instead we return the return
code of the operation and let the aio core call aio_complete().  This is
purposely done to fix a bug where AIO DIO file extensions would call
aio_complete() before their callers have a chance to update i_size.

Now that direct_io_worker() is explicitly returning -EIOCBQUEUED its callers
no longer have to translate for it.  XFS needs to be careful not to free
resources that will be used during AIO completion if -EIOCBQUEUED is returned.
 We maintain the previous behaviour of trying to write fs metadata for O_SYNC
aio+dio writes.
Signed-off-by: default avatarZach Brown <zach.brown@oracle.com>
Cc: Badari Pulavarty <pbadari@us.ibm.com>
Cc: Suparna Bhattacharya <suparna@in.ibm.com>
Acked-by: default avatarJeff Moyer <jmoyer@redhat.com>
Cc: <xfs-masters@oss.sgi.com>
Signed-off-by: default avatarAndrew Morton <akpm@osdl.org>
Signed-off-by: default avatarLinus Torvalds <torvalds@osdl.org>
parent 20258b2b
...@@ -226,6 +226,15 @@ static int dio_complete(struct dio *dio, loff_t offset, int ret) ...@@ -226,6 +226,15 @@ static int dio_complete(struct dio *dio, loff_t offset, int ret)
{ {
ssize_t transferred = 0; ssize_t transferred = 0;
/*
* AIO submission can race with bio completion to get here while
* expecting to have the last io completed by bio completion.
* In that case -EIOCBQUEUED is in fact not an error we want
* to preserve through this call.
*/
if (ret == -EIOCBQUEUED)
ret = 0;
if (dio->result) { if (dio->result) {
transferred = dio->result; transferred = dio->result;
...@@ -251,24 +260,6 @@ static int dio_complete(struct dio *dio, loff_t offset, int ret) ...@@ -251,24 +260,6 @@ static int dio_complete(struct dio *dio, loff_t offset, int ret)
return ret; return ret;
} }
/*
* Called when a BIO has been processed. If the count goes to zero then IO is
* complete and we can signal this to the AIO layer.
*/
static void dio_complete_aio(struct dio *dio)
{
int ret;
ret = dio_complete(dio, dio->iocb->ki_pos, 0);
/* Complete AIO later if falling back to buffered i/o */
if (dio->result == dio->size ||
((dio->rw == READ) && dio->result)) {
aio_complete(dio->iocb, ret, 0);
kfree(dio);
}
}
static int dio_bio_complete(struct dio *dio, struct bio *bio); static int dio_bio_complete(struct dio *dio, struct bio *bio);
/* /*
* Asynchronous IO callback. * Asynchronous IO callback.
...@@ -290,8 +281,11 @@ static int dio_bio_end_aio(struct bio *bio, unsigned int bytes_done, int error) ...@@ -290,8 +281,11 @@ static int dio_bio_end_aio(struct bio *bio, unsigned int bytes_done, int error)
if (remaining == 1 && waiter_holds_ref) if (remaining == 1 && waiter_holds_ref)
wake_up_process(dio->waiter); wake_up_process(dio->waiter);
if (remaining == 0) if (remaining == 0) {
dio_complete_aio(dio); int ret = dio_complete(dio, dio->iocb->ki_pos, 0);
aio_complete(dio->iocb, ret, 0);
kfree(dio);
}
return 0; return 0;
} }
...@@ -1082,47 +1076,33 @@ direct_io_worker(int rw, struct kiocb *iocb, struct inode *inode, ...@@ -1082,47 +1076,33 @@ direct_io_worker(int rw, struct kiocb *iocb, struct inode *inode,
mutex_unlock(&dio->inode->i_mutex); mutex_unlock(&dio->inode->i_mutex);
/* /*
* OK, all BIOs are submitted, so we can decrement bio_count to truly * The only time we want to leave bios in flight is when a successful
* reflect the number of to-be-processed BIOs. * partial aio read or full aio write have been setup. In that case
* bio completion will call aio_complete. The only time it's safe to
* call aio_complete is when we return -EIOCBQUEUED, so we key on that.
* This had *better* be the only place that raises -EIOCBQUEUED.
*/ */
if (dio->is_async) { BUG_ON(ret == -EIOCBQUEUED);
int should_wait = 0; if (dio->is_async && ret == 0 && dio->result &&
((rw & READ) || (dio->result == dio->size)))
if (dio->result < dio->size && (rw & WRITE)) { ret = -EIOCBQUEUED;
dio->waiter = current;
should_wait = 1;
}
if (ret == 0)
ret = dio->result;
if (should_wait) if (ret != -EIOCBQUEUED)
dio_await_completion(dio); dio_await_completion(dio);
/* this can free the dio */
if (atomic_dec_and_test(&dio->refcount))
dio_complete_aio(dio);
if (should_wait)
kfree(dio);
} else {
dio_await_completion(dio);
ret = dio_complete(dio, offset, ret);
/* We could have also come here on an AIO file extend */
if (!is_sync_kiocb(iocb) && (rw & WRITE) &&
ret >= 0 && dio->result == dio->size)
/* /*
* For AIO writes where we have completed the * Sync will always be dropping the final ref and completing the
* i/o, we have to mark the the aio complete. * operation. AIO can if it was a broken operation described above
* or in fact if all the bios race to complete before we get here.
* In that case dio_complete() translates the EIOCBQUEUED into
* the proper return code that the caller will hand to aio_complete().
*/ */
aio_complete(iocb, ret, 0); if (atomic_dec_and_test(&dio->refcount)) {
ret = dio_complete(dio, offset, ret);
if (atomic_dec_and_test(&dio->refcount))
kfree(dio); kfree(dio);
else } else
BUG(); BUG_ON(ret != -EIOCBQUEUED);
}
return ret; return ret;
} }
......
...@@ -1406,7 +1406,7 @@ xfs_vm_direct_IO( ...@@ -1406,7 +1406,7 @@ xfs_vm_direct_IO(
xfs_end_io_direct); xfs_end_io_direct);
} }
if (unlikely(ret <= 0 && iocb->private)) if (unlikely(ret != -EIOCBQUEUED && iocb->private))
xfs_destroy_ioend(iocb->private); xfs_destroy_ioend(iocb->private);
return ret; return ret;
} }
......
...@@ -1181,8 +1181,6 @@ generic_file_aio_read(struct kiocb *iocb, const struct iovec *iov, ...@@ -1181,8 +1181,6 @@ generic_file_aio_read(struct kiocb *iocb, const struct iovec *iov,
if (pos < size) { if (pos < size) {
retval = generic_file_direct_IO(READ, iocb, retval = generic_file_direct_IO(READ, iocb,
iov, pos, nr_segs); iov, pos, nr_segs);
if (retval > 0 && !is_sync_kiocb(iocb))
retval = -EIOCBQUEUED;
if (retval > 0) if (retval > 0)
*ppos = pos + retval; *ppos = pos + retval;
} }
...@@ -2047,15 +2045,14 @@ generic_file_direct_write(struct kiocb *iocb, const struct iovec *iov, ...@@ -2047,15 +2045,14 @@ generic_file_direct_write(struct kiocb *iocb, const struct iovec *iov,
* Sync the fs metadata but not the minor inode changes and * Sync the fs metadata but not the minor inode changes and
* of course not the data as we did direct DMA for the IO. * of course not the data as we did direct DMA for the IO.
* i_mutex is held, which protects generic_osync_inode() from * i_mutex is held, which protects generic_osync_inode() from
* livelocking. * livelocking. AIO O_DIRECT ops attempt to sync metadata here.
*/ */
if (written >= 0 && ((file->f_flags & O_SYNC) || IS_SYNC(inode))) { if ((written >= 0 || written == -EIOCBQUEUED) &&
((file->f_flags & O_SYNC) || IS_SYNC(inode))) {
int err = generic_osync_inode(inode, mapping, OSYNC_METADATA); int err = generic_osync_inode(inode, mapping, OSYNC_METADATA);
if (err < 0) if (err < 0)
written = err; written = err;
} }
if (written == count && !is_sync_kiocb(iocb))
written = -EIOCBQUEUED;
return written; return written;
} }
EXPORT_SYMBOL(generic_file_direct_write); EXPORT_SYMBOL(generic_file_direct_write);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment