 b49f4755c7
			
		
	
	
		b49f4755c7
		
	
	
	
	
		
			
			This is the big patch that removes aio_context_acquire()/aio_context_release() from the block layer and affected block layer users. There isn't a clean way to split this patch and the reviewers are likely the same group of people, so I decided to do it in one patch. Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> Reviewed-by: Eric Blake <eblake@redhat.com> Reviewed-by: Kevin Wolf <kwolf@redhat.com> Reviewed-by: Paul Durrant <paul@xen.org> Message-ID: <20231205182011.1976568-7-stefanha@redhat.com> Signed-off-by: Kevin Wolf <kwolf@redhat.com>
		
			
				
	
	
		
			899 lines
		
	
	
		
			26 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			899 lines
		
	
	
		
			26 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * Block tests for iothreads
 | |
|  *
 | |
|  * Copyright (c) 2018 Kevin Wolf <kwolf@redhat.com>
 | |
|  *
 | |
|  * Permission is hereby granted, free of charge, to any person obtaining a copy
 | |
|  * of this software and associated documentation files (the "Software"), to deal
 | |
|  * in the Software without restriction, including without limitation the rights
 | |
|  * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
 | |
|  * copies of the Software, and to permit persons to whom the Software is
 | |
|  * furnished to do so, subject to the following conditions:
 | |
|  *
 | |
|  * The above copyright notice and this permission notice shall be included in
 | |
|  * all copies or substantial portions of the Software.
 | |
|  *
 | |
|  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
 | |
|  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
 | |
|  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
 | |
|  * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
 | |
|  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
 | |
|  * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
 | |
|  * THE SOFTWARE.
 | |
|  */
 | |
| 
 | |
| #include "qemu/osdep.h"
 | |
| #include "block/block.h"
 | |
| #include "block/block_int-global-state.h"
 | |
| #include "block/blockjob_int.h"
 | |
| #include "sysemu/block-backend.h"
 | |
| #include "qapi/error.h"
 | |
| #include "qapi/qmp/qdict.h"
 | |
| #include "qemu/main-loop.h"
 | |
| #include "iothread.h"
 | |
| 
 | |
| static int coroutine_fn bdrv_test_co_preadv(BlockDriverState *bs,
 | |
|                                             int64_t offset, int64_t bytes,
 | |
|                                             QEMUIOVector *qiov,
 | |
|                                             BdrvRequestFlags flags)
 | |
| {
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static int coroutine_fn bdrv_test_co_pwritev(BlockDriverState *bs,
 | |
|                                              int64_t offset, int64_t bytes,
 | |
|                                              QEMUIOVector *qiov,
 | |
|                                              BdrvRequestFlags flags)
 | |
| {
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static int coroutine_fn bdrv_test_co_pdiscard(BlockDriverState *bs,
 | |
|                                               int64_t offset, int64_t bytes)
 | |
| {
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static int coroutine_fn
 | |
| bdrv_test_co_truncate(BlockDriverState *bs, int64_t offset, bool exact,
 | |
|                       PreallocMode prealloc, BdrvRequestFlags flags,
 | |
|                       Error **errp)
 | |
| {
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static int coroutine_fn bdrv_test_co_block_status(BlockDriverState *bs,
 | |
|                                                   bool want_zero,
 | |
|                                                   int64_t offset, int64_t count,
 | |
|                                                   int64_t *pnum, int64_t *map,
 | |
|                                                   BlockDriverState **file)
 | |
| {
 | |
|     *pnum = count;
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static BlockDriver bdrv_test = {
 | |
|     .format_name            = "test",
 | |
|     .instance_size          = 1,
 | |
| 
 | |
|     .bdrv_co_preadv         = bdrv_test_co_preadv,
 | |
|     .bdrv_co_pwritev        = bdrv_test_co_pwritev,
 | |
|     .bdrv_co_pdiscard       = bdrv_test_co_pdiscard,
 | |
|     .bdrv_co_truncate       = bdrv_test_co_truncate,
 | |
|     .bdrv_co_block_status   = bdrv_test_co_block_status,
 | |
| };
 | |
| 
 | |
| static void test_sync_op_pread(BdrvChild *c)
 | |
| {
 | |
|     uint8_t buf[512];
 | |
|     int ret;
 | |
| 
 | |
|     /* Success */
 | |
|     ret = bdrv_pread(c, 0, sizeof(buf), buf, 0);
 | |
|     g_assert_cmpint(ret, ==, 0);
 | |
| 
 | |
|     /* Early error: Negative offset */
 | |
|     ret = bdrv_pread(c, -2, sizeof(buf), buf, 0);
 | |
|     g_assert_cmpint(ret, ==, -EIO);
 | |
| }
 | |
| 
 | |
| static void test_sync_op_pwrite(BdrvChild *c)
 | |
| {
 | |
|     uint8_t buf[512] = { 0 };
 | |
|     int ret;
 | |
| 
 | |
|     /* Success */
 | |
|     ret = bdrv_pwrite(c, 0, sizeof(buf), buf, 0);
 | |
|     g_assert_cmpint(ret, ==, 0);
 | |
| 
 | |
|     /* Early error: Negative offset */
 | |
|     ret = bdrv_pwrite(c, -2, sizeof(buf), buf, 0);
 | |
|     g_assert_cmpint(ret, ==, -EIO);
 | |
| }
 | |
| 
 | |
| static void test_sync_op_blk_pread(BlockBackend *blk)
 | |
| {
 | |
|     uint8_t buf[512];
 | |
|     int ret;
 | |
| 
 | |
|     /* Success */
 | |
|     ret = blk_pread(blk, 0, sizeof(buf), buf, 0);
 | |
|     g_assert_cmpint(ret, ==, 0);
 | |
| 
 | |
|     /* Early error: Negative offset */
 | |
|     ret = blk_pread(blk, -2, sizeof(buf), buf, 0);
 | |
|     g_assert_cmpint(ret, ==, -EIO);
 | |
| }
 | |
| 
 | |
| static void test_sync_op_blk_pwrite(BlockBackend *blk)
 | |
| {
 | |
|     uint8_t buf[512] = { 0 };
 | |
|     int ret;
 | |
| 
 | |
|     /* Success */
 | |
|     ret = blk_pwrite(blk, 0, sizeof(buf), buf, 0);
 | |
|     g_assert_cmpint(ret, ==, 0);
 | |
| 
 | |
|     /* Early error: Negative offset */
 | |
|     ret = blk_pwrite(blk, -2, sizeof(buf), buf, 0);
 | |
|     g_assert_cmpint(ret, ==, -EIO);
 | |
| }
 | |
| 
 | |
| static void test_sync_op_blk_preadv(BlockBackend *blk)
 | |
| {
 | |
|     uint8_t buf[512];
 | |
|     QEMUIOVector qiov = QEMU_IOVEC_INIT_BUF(qiov, buf, sizeof(buf));
 | |
|     int ret;
 | |
| 
 | |
|     /* Success */
 | |
|     ret = blk_preadv(blk, 0, sizeof(buf), &qiov, 0);
 | |
|     g_assert_cmpint(ret, ==, 0);
 | |
| 
 | |
|     /* Early error: Negative offset */
 | |
|     ret = blk_preadv(blk, -2, sizeof(buf), &qiov, 0);
 | |
|     g_assert_cmpint(ret, ==, -EIO);
 | |
| }
 | |
| 
 | |
| static void test_sync_op_blk_pwritev(BlockBackend *blk)
 | |
| {
 | |
|     uint8_t buf[512] = { 0 };
 | |
|     QEMUIOVector qiov = QEMU_IOVEC_INIT_BUF(qiov, buf, sizeof(buf));
 | |
|     int ret;
 | |
| 
 | |
|     /* Success */
 | |
|     ret = blk_pwritev(blk, 0, sizeof(buf), &qiov, 0);
 | |
|     g_assert_cmpint(ret, ==, 0);
 | |
| 
 | |
|     /* Early error: Negative offset */
 | |
|     ret = blk_pwritev(blk, -2, sizeof(buf), &qiov, 0);
 | |
|     g_assert_cmpint(ret, ==, -EIO);
 | |
| }
 | |
| 
 | |
| static void test_sync_op_blk_preadv_part(BlockBackend *blk)
 | |
| {
 | |
|     uint8_t buf[512];
 | |
|     QEMUIOVector qiov = QEMU_IOVEC_INIT_BUF(qiov, buf, sizeof(buf));
 | |
|     int ret;
 | |
| 
 | |
|     /* Success */
 | |
|     ret = blk_preadv_part(blk, 0, sizeof(buf), &qiov, 0, 0);
 | |
|     g_assert_cmpint(ret, ==, 0);
 | |
| 
 | |
|     /* Early error: Negative offset */
 | |
|     ret = blk_preadv_part(blk, -2, sizeof(buf), &qiov, 0, 0);
 | |
|     g_assert_cmpint(ret, ==, -EIO);
 | |
| }
 | |
| 
 | |
| static void test_sync_op_blk_pwritev_part(BlockBackend *blk)
 | |
| {
 | |
|     uint8_t buf[512] = { 0 };
 | |
|     QEMUIOVector qiov = QEMU_IOVEC_INIT_BUF(qiov, buf, sizeof(buf));
 | |
|     int ret;
 | |
| 
 | |
|     /* Success */
 | |
|     ret = blk_pwritev_part(blk, 0, sizeof(buf), &qiov, 0, 0);
 | |
|     g_assert_cmpint(ret, ==, 0);
 | |
| 
 | |
|     /* Early error: Negative offset */
 | |
|     ret = blk_pwritev_part(blk, -2, sizeof(buf), &qiov, 0, 0);
 | |
|     g_assert_cmpint(ret, ==, -EIO);
 | |
| }
 | |
| 
 | |
| static void test_sync_op_blk_pwrite_compressed(BlockBackend *blk)
 | |
| {
 | |
|     uint8_t buf[512] = { 0 };
 | |
|     int ret;
 | |
| 
 | |
|     /* Late error: Not supported */
 | |
|     ret = blk_pwrite_compressed(blk, 0, sizeof(buf), buf);
 | |
|     g_assert_cmpint(ret, ==, -ENOTSUP);
 | |
| 
 | |
|     /* Early error: Negative offset */
 | |
|     ret = blk_pwrite_compressed(blk, -2, sizeof(buf), buf);
 | |
|     g_assert_cmpint(ret, ==, -EIO);
 | |
| }
 | |
| 
 | |
| static void test_sync_op_blk_pwrite_zeroes(BlockBackend *blk)
 | |
| {
 | |
|     int ret;
 | |
| 
 | |
|     /* Success */
 | |
|     ret = blk_pwrite_zeroes(blk, 0, 512, 0);
 | |
|     g_assert_cmpint(ret, ==, 0);
 | |
| 
 | |
|     /* Early error: Negative offset */
 | |
|     ret = blk_pwrite_zeroes(blk, -2, 512, 0);
 | |
|     g_assert_cmpint(ret, ==, -EIO);
 | |
| }
 | |
| 
 | |
| static void test_sync_op_load_vmstate(BdrvChild *c)
 | |
| {
 | |
|     uint8_t buf[512];
 | |
|     int ret;
 | |
| 
 | |
|     /* Error: Driver does not support snapshots */
 | |
|     ret = bdrv_load_vmstate(c->bs, buf, 0, sizeof(buf));
 | |
|     g_assert_cmpint(ret, ==, -ENOTSUP);
 | |
| }
 | |
| 
 | |
| static void test_sync_op_save_vmstate(BdrvChild *c)
 | |
| {
 | |
|     uint8_t buf[512] = { 0 };
 | |
|     int ret;
 | |
| 
 | |
|     /* Error: Driver does not support snapshots */
 | |
|     ret = bdrv_save_vmstate(c->bs, buf, 0, sizeof(buf));
 | |
|     g_assert_cmpint(ret, ==, -ENOTSUP);
 | |
| }
 | |
| 
 | |
| static void test_sync_op_pdiscard(BdrvChild *c)
 | |
| {
 | |
|     int ret;
 | |
| 
 | |
|     /* Normal success path */
 | |
|     c->bs->open_flags |= BDRV_O_UNMAP;
 | |
|     ret = bdrv_pdiscard(c, 0, 512);
 | |
|     g_assert_cmpint(ret, ==, 0);
 | |
| 
 | |
|     /* Early success: UNMAP not supported */
 | |
|     c->bs->open_flags &= ~BDRV_O_UNMAP;
 | |
|     ret = bdrv_pdiscard(c, 0, 512);
 | |
|     g_assert_cmpint(ret, ==, 0);
 | |
| 
 | |
|     /* Early error: Negative offset */
 | |
|     ret = bdrv_pdiscard(c, -2, 512);
 | |
|     g_assert_cmpint(ret, ==, -EIO);
 | |
| }
 | |
| 
 | |
| static void test_sync_op_blk_pdiscard(BlockBackend *blk)
 | |
| {
 | |
|     int ret;
 | |
| 
 | |
|     /* Early success: UNMAP not supported */
 | |
|     ret = blk_pdiscard(blk, 0, 512);
 | |
|     g_assert_cmpint(ret, ==, 0);
 | |
| 
 | |
|     /* Early error: Negative offset */
 | |
|     ret = blk_pdiscard(blk, -2, 512);
 | |
|     g_assert_cmpint(ret, ==, -EIO);
 | |
| }
 | |
| 
 | |
| static void test_sync_op_truncate(BdrvChild *c)
 | |
| {
 | |
|     int ret;
 | |
| 
 | |
|     /* Normal success path */
 | |
|     ret = bdrv_truncate(c, 65536, false, PREALLOC_MODE_OFF, 0, NULL);
 | |
|     g_assert_cmpint(ret, ==, 0);
 | |
| 
 | |
|     /* Early error: Negative offset */
 | |
|     ret = bdrv_truncate(c, -2, false, PREALLOC_MODE_OFF, 0, NULL);
 | |
|     g_assert_cmpint(ret, ==, -EINVAL);
 | |
| 
 | |
|     /* Error: Read-only image */
 | |
|     c->bs->open_flags &= ~BDRV_O_RDWR;
 | |
| 
 | |
|     ret = bdrv_truncate(c, 65536, false, PREALLOC_MODE_OFF, 0, NULL);
 | |
|     g_assert_cmpint(ret, ==, -EACCES);
 | |
| 
 | |
|     c->bs->open_flags |= BDRV_O_RDWR;
 | |
| }
 | |
| 
 | |
| static void test_sync_op_blk_truncate(BlockBackend *blk)
 | |
| {
 | |
|     int ret;
 | |
| 
 | |
|     /* Normal success path */
 | |
|     ret = blk_truncate(blk, 65536, false, PREALLOC_MODE_OFF, 0, NULL);
 | |
|     g_assert_cmpint(ret, ==, 0);
 | |
| 
 | |
|     /* Early error: Negative offset */
 | |
|     ret = blk_truncate(blk, -2, false, PREALLOC_MODE_OFF, 0, NULL);
 | |
|     g_assert_cmpint(ret, ==, -EINVAL);
 | |
| }
 | |
| 
 | |
| /* Disable TSA to make bdrv_test.bdrv_co_block_status writable */
 | |
| static void TSA_NO_TSA test_sync_op_block_status(BdrvChild *c)
 | |
| {
 | |
|     int ret;
 | |
|     int64_t n;
 | |
| 
 | |
|     /* Normal success path */
 | |
|     ret = bdrv_is_allocated(c->bs, 0, 65536, &n);
 | |
|     g_assert_cmpint(ret, ==, 0);
 | |
| 
 | |
|     /* Early success: No driver support */
 | |
|     bdrv_test.bdrv_co_block_status = NULL;
 | |
|     ret = bdrv_is_allocated(c->bs, 0, 65536, &n);
 | |
|     g_assert_cmpint(ret, ==, 1);
 | |
| 
 | |
|     /* Early success: bytes = 0 */
 | |
|     ret = bdrv_is_allocated(c->bs, 0, 0, &n);
 | |
|     g_assert_cmpint(ret, ==, 0);
 | |
| 
 | |
|     /* Early success: Offset > image size*/
 | |
|     ret = bdrv_is_allocated(c->bs, 0x1000000, 0x1000000, &n);
 | |
|     g_assert_cmpint(ret, ==, 0);
 | |
| }
 | |
| 
 | |
| static void test_sync_op_flush(BdrvChild *c)
 | |
| {
 | |
|     int ret;
 | |
| 
 | |
|     /* Normal success path */
 | |
|     ret = bdrv_flush(c->bs);
 | |
|     g_assert_cmpint(ret, ==, 0);
 | |
| 
 | |
|     /* Early success: Read-only image */
 | |
|     c->bs->open_flags &= ~BDRV_O_RDWR;
 | |
| 
 | |
|     ret = bdrv_flush(c->bs);
 | |
|     g_assert_cmpint(ret, ==, 0);
 | |
| 
 | |
|     c->bs->open_flags |= BDRV_O_RDWR;
 | |
| }
 | |
| 
 | |
| static void test_sync_op_blk_flush(BlockBackend *blk)
 | |
| {
 | |
|     BlockDriverState *bs = blk_bs(blk);
 | |
|     int ret;
 | |
| 
 | |
|     /* Normal success path */
 | |
|     ret = blk_flush(blk);
 | |
|     g_assert_cmpint(ret, ==, 0);
 | |
| 
 | |
|     /* Early success: Read-only image */
 | |
|     bs->open_flags &= ~BDRV_O_RDWR;
 | |
| 
 | |
|     ret = blk_flush(blk);
 | |
|     g_assert_cmpint(ret, ==, 0);
 | |
| 
 | |
|     bs->open_flags |= BDRV_O_RDWR;
 | |
| }
 | |
| 
 | |
| static void test_sync_op_check(BdrvChild *c)
 | |
| {
 | |
|     BdrvCheckResult result;
 | |
|     int ret;
 | |
| 
 | |
|     /* Error: Driver does not implement check */
 | |
|     ret = bdrv_check(c->bs, &result, 0);
 | |
|     g_assert_cmpint(ret, ==, -ENOTSUP);
 | |
| }
 | |
| 
 | |
| static void test_sync_op_activate(BdrvChild *c)
 | |
| {
 | |
|     GLOBAL_STATE_CODE();
 | |
|     GRAPH_RDLOCK_GUARD_MAINLOOP();
 | |
| 
 | |
|     /* Early success: Image is not inactive */
 | |
|     bdrv_activate(c->bs, NULL);
 | |
| }
 | |
| 
 | |
| 
 | |
| typedef struct SyncOpTest {
 | |
|     const char *name;
 | |
|     void (*fn)(BdrvChild *c);
 | |
|     void (*blkfn)(BlockBackend *blk);
 | |
| } SyncOpTest;
 | |
| 
 | |
| const SyncOpTest sync_op_tests[] = {
 | |
|     {
 | |
|         .name   = "/sync-op/pread",
 | |
|         .fn     = test_sync_op_pread,
 | |
|         .blkfn  = test_sync_op_blk_pread,
 | |
|     }, {
 | |
|         .name   = "/sync-op/pwrite",
 | |
|         .fn     = test_sync_op_pwrite,
 | |
|         .blkfn  = test_sync_op_blk_pwrite,
 | |
|     }, {
 | |
|         .name   = "/sync-op/preadv",
 | |
|         .fn     = NULL,
 | |
|         .blkfn  = test_sync_op_blk_preadv,
 | |
|     }, {
 | |
|         .name   = "/sync-op/pwritev",
 | |
|         .fn     = NULL,
 | |
|         .blkfn  = test_sync_op_blk_pwritev,
 | |
|     }, {
 | |
|         .name   = "/sync-op/preadv_part",
 | |
|         .fn     = NULL,
 | |
|         .blkfn  = test_sync_op_blk_preadv_part,
 | |
|     }, {
 | |
|         .name   = "/sync-op/pwritev_part",
 | |
|         .fn     = NULL,
 | |
|         .blkfn  = test_sync_op_blk_pwritev_part,
 | |
|     }, {
 | |
|         .name   = "/sync-op/pwrite_compressed",
 | |
|         .fn     = NULL,
 | |
|         .blkfn  = test_sync_op_blk_pwrite_compressed,
 | |
|     }, {
 | |
|         .name   = "/sync-op/pwrite_zeroes",
 | |
|         .fn     = NULL,
 | |
|         .blkfn  = test_sync_op_blk_pwrite_zeroes,
 | |
|     }, {
 | |
|         .name   = "/sync-op/load_vmstate",
 | |
|         .fn     = test_sync_op_load_vmstate,
 | |
|     }, {
 | |
|         .name   = "/sync-op/save_vmstate",
 | |
|         .fn     = test_sync_op_save_vmstate,
 | |
|     }, {
 | |
|         .name   = "/sync-op/pdiscard",
 | |
|         .fn     = test_sync_op_pdiscard,
 | |
|         .blkfn  = test_sync_op_blk_pdiscard,
 | |
|     }, {
 | |
|         .name   = "/sync-op/truncate",
 | |
|         .fn     = test_sync_op_truncate,
 | |
|         .blkfn  = test_sync_op_blk_truncate,
 | |
|     }, {
 | |
|         .name   = "/sync-op/block_status",
 | |
|         .fn     = test_sync_op_block_status,
 | |
|     }, {
 | |
|         .name   = "/sync-op/flush",
 | |
|         .fn     = test_sync_op_flush,
 | |
|         .blkfn  = test_sync_op_blk_flush,
 | |
|     }, {
 | |
|         .name   = "/sync-op/check",
 | |
|         .fn     = test_sync_op_check,
 | |
|     }, {
 | |
|         .name   = "/sync-op/activate",
 | |
|         .fn     = test_sync_op_activate,
 | |
|     },
 | |
| };
 | |
| 
 | |
| /* Test synchronous operations that run in a different iothread, so we have to
 | |
|  * poll for the coroutine there to return. */
 | |
| static void test_sync_op(const void *opaque)
 | |
| {
 | |
|     const SyncOpTest *t = opaque;
 | |
|     IOThread *iothread = iothread_new();
 | |
|     AioContext *ctx = iothread_get_aio_context(iothread);
 | |
|     BlockBackend *blk;
 | |
|     BlockDriverState *bs;
 | |
|     BdrvChild *c;
 | |
| 
 | |
|     GLOBAL_STATE_CODE();
 | |
| 
 | |
|     blk = blk_new(qemu_get_aio_context(), BLK_PERM_ALL, BLK_PERM_ALL);
 | |
|     bs = bdrv_new_open_driver(&bdrv_test, "base", BDRV_O_RDWR, &error_abort);
 | |
|     bs->total_sectors = 65536 / BDRV_SECTOR_SIZE;
 | |
|     blk_insert_bs(blk, bs, &error_abort);
 | |
| 
 | |
|     bdrv_graph_rdlock_main_loop();
 | |
|     c = QLIST_FIRST(&bs->parents);
 | |
|     bdrv_graph_rdunlock_main_loop();
 | |
| 
 | |
|     blk_set_aio_context(blk, ctx, &error_abort);
 | |
|     if (t->fn) {
 | |
|         t->fn(c);
 | |
|     }
 | |
|     if (t->blkfn) {
 | |
|         t->blkfn(blk);
 | |
|     }
 | |
|     blk_set_aio_context(blk, qemu_get_aio_context(), &error_abort);
 | |
| 
 | |
|     bdrv_unref(bs);
 | |
|     blk_unref(blk);
 | |
| }
 | |
| 
 | |
| typedef struct TestBlockJob {
 | |
|     BlockJob common;
 | |
|     bool should_complete;
 | |
|     int n;
 | |
| } TestBlockJob;
 | |
| 
 | |
| static int test_job_prepare(Job *job)
 | |
| {
 | |
|     g_assert(qemu_get_current_aio_context() == qemu_get_aio_context());
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static int coroutine_fn test_job_run(Job *job, Error **errp)
 | |
| {
 | |
|     TestBlockJob *s = container_of(job, TestBlockJob, common.job);
 | |
| 
 | |
|     job_transition_to_ready(&s->common.job);
 | |
|     while (!s->should_complete) {
 | |
|         s->n++;
 | |
|         g_assert(qemu_get_current_aio_context() == job->aio_context);
 | |
| 
 | |
|         /* Avoid job_sleep_ns() because it marks the job as !busy. We want to
 | |
|          * emulate some actual activity (probably some I/O) here so that the
 | |
|          * drain involved in AioContext switches has to wait for this activity
 | |
|          * to stop. */
 | |
|         qemu_co_sleep_ns(QEMU_CLOCK_REALTIME, 1000000);
 | |
| 
 | |
|         job_pause_point(&s->common.job);
 | |
|     }
 | |
| 
 | |
|     g_assert(qemu_get_current_aio_context() == job->aio_context);
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static void test_job_complete(Job *job, Error **errp)
 | |
| {
 | |
|     TestBlockJob *s = container_of(job, TestBlockJob, common.job);
 | |
|     s->should_complete = true;
 | |
| }
 | |
| 
 | |
| BlockJobDriver test_job_driver = {
 | |
|     .job_driver = {
 | |
|         .instance_size  = sizeof(TestBlockJob),
 | |
|         .free           = block_job_free,
 | |
|         .user_resume    = block_job_user_resume,
 | |
|         .run            = test_job_run,
 | |
|         .complete       = test_job_complete,
 | |
|         .prepare        = test_job_prepare,
 | |
|     },
 | |
| };
 | |
| 
 | |
| static void test_attach_blockjob(void)
 | |
| {
 | |
|     IOThread *iothread = iothread_new();
 | |
|     AioContext *ctx = iothread_get_aio_context(iothread);
 | |
|     BlockBackend *blk;
 | |
|     BlockDriverState *bs;
 | |
|     TestBlockJob *tjob;
 | |
| 
 | |
|     blk = blk_new(qemu_get_aio_context(), BLK_PERM_ALL, BLK_PERM_ALL);
 | |
|     bs = bdrv_new_open_driver(&bdrv_test, "base", BDRV_O_RDWR, &error_abort);
 | |
|     blk_insert_bs(blk, bs, &error_abort);
 | |
| 
 | |
|     tjob = block_job_create("job0", &test_job_driver, NULL, bs,
 | |
|                             0, BLK_PERM_ALL,
 | |
|                             0, 0, NULL, NULL, &error_abort);
 | |
|     job_start(&tjob->common.job);
 | |
| 
 | |
|     while (tjob->n == 0) {
 | |
|         aio_poll(qemu_get_aio_context(), false);
 | |
|     }
 | |
| 
 | |
|     blk_set_aio_context(blk, ctx, &error_abort);
 | |
| 
 | |
|     tjob->n = 0;
 | |
|     while (tjob->n == 0) {
 | |
|         aio_poll(qemu_get_aio_context(), false);
 | |
|     }
 | |
| 
 | |
|     blk_set_aio_context(blk, qemu_get_aio_context(), &error_abort);
 | |
| 
 | |
|     tjob->n = 0;
 | |
|     while (tjob->n == 0) {
 | |
|         aio_poll(qemu_get_aio_context(), false);
 | |
|     }
 | |
| 
 | |
|     blk_set_aio_context(blk, ctx, &error_abort);
 | |
| 
 | |
|     tjob->n = 0;
 | |
|     while (tjob->n == 0) {
 | |
|         aio_poll(qemu_get_aio_context(), false);
 | |
|     }
 | |
| 
 | |
|     WITH_JOB_LOCK_GUARD() {
 | |
|         job_complete_sync_locked(&tjob->common.job, &error_abort);
 | |
|     }
 | |
|     blk_set_aio_context(blk, qemu_get_aio_context(), &error_abort);
 | |
| 
 | |
|     bdrv_unref(bs);
 | |
|     blk_unref(blk);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Test that changing the AioContext for one node in a tree (here through blk)
 | |
|  * changes all other nodes as well:
 | |
|  *
 | |
|  *  blk
 | |
|  *   |
 | |
|  *   |  bs_verify [blkverify]
 | |
|  *   |   /               \
 | |
|  *   |  /                 \
 | |
|  *  bs_a [bdrv_test]    bs_b [bdrv_test]
 | |
|  *
 | |
|  */
 | |
| static void test_propagate_basic(void)
 | |
| {
 | |
|     IOThread *iothread = iothread_new();
 | |
|     AioContext *ctx = iothread_get_aio_context(iothread);
 | |
|     AioContext *main_ctx;
 | |
|     BlockBackend *blk;
 | |
|     BlockDriverState *bs_a, *bs_b, *bs_verify;
 | |
|     QDict *options;
 | |
| 
 | |
|     /*
 | |
|      * Create bs_a and its BlockBackend.  We cannot take the RESIZE
 | |
|      * permission because blkverify will not share it on the test
 | |
|      * image.
 | |
|      */
 | |
|     blk = blk_new(qemu_get_aio_context(), BLK_PERM_ALL & ~BLK_PERM_RESIZE,
 | |
|                   BLK_PERM_ALL);
 | |
|     bs_a = bdrv_new_open_driver(&bdrv_test, "bs_a", BDRV_O_RDWR, &error_abort);
 | |
|     blk_insert_bs(blk, bs_a, &error_abort);
 | |
| 
 | |
|     /* Create bs_b */
 | |
|     bs_b = bdrv_new_open_driver(&bdrv_test, "bs_b", BDRV_O_RDWR, &error_abort);
 | |
| 
 | |
|     /* Create blkverify filter that references both bs_a and bs_b */
 | |
|     options = qdict_new();
 | |
|     qdict_put_str(options, "driver", "blkverify");
 | |
|     qdict_put_str(options, "test", "bs_a");
 | |
|     qdict_put_str(options, "raw", "bs_b");
 | |
| 
 | |
|     bs_verify = bdrv_open(NULL, NULL, options, BDRV_O_RDWR, &error_abort);
 | |
| 
 | |
|     /* Switch the AioContext */
 | |
|     blk_set_aio_context(blk, ctx, &error_abort);
 | |
|     g_assert(blk_get_aio_context(blk) == ctx);
 | |
|     g_assert(bdrv_get_aio_context(bs_a) == ctx);
 | |
|     g_assert(bdrv_get_aio_context(bs_verify) == ctx);
 | |
|     g_assert(bdrv_get_aio_context(bs_b) == ctx);
 | |
| 
 | |
|     /* Switch the AioContext back */
 | |
|     main_ctx = qemu_get_aio_context();
 | |
|     blk_set_aio_context(blk, main_ctx, &error_abort);
 | |
|     g_assert(blk_get_aio_context(blk) == main_ctx);
 | |
|     g_assert(bdrv_get_aio_context(bs_a) == main_ctx);
 | |
|     g_assert(bdrv_get_aio_context(bs_verify) == main_ctx);
 | |
|     g_assert(bdrv_get_aio_context(bs_b) == main_ctx);
 | |
| 
 | |
|     bdrv_unref(bs_verify);
 | |
|     bdrv_unref(bs_b);
 | |
|     bdrv_unref(bs_a);
 | |
|     blk_unref(blk);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Test that diamonds in the graph don't lead to endless recursion:
 | |
|  *
 | |
|  *              blk
 | |
|  *               |
 | |
|  *      bs_verify [blkverify]
 | |
|  *       /              \
 | |
|  *      /                \
 | |
|  *   bs_b [raw]         bs_c[raw]
 | |
|  *      \                /
 | |
|  *       \              /
 | |
|  *       bs_a [bdrv_test]
 | |
|  */
 | |
| static void test_propagate_diamond(void)
 | |
| {
 | |
|     IOThread *iothread = iothread_new();
 | |
|     AioContext *ctx = iothread_get_aio_context(iothread);
 | |
|     AioContext *main_ctx;
 | |
|     BlockBackend *blk;
 | |
|     BlockDriverState *bs_a, *bs_b, *bs_c, *bs_verify;
 | |
|     QDict *options;
 | |
| 
 | |
|     /* Create bs_a */
 | |
|     bs_a = bdrv_new_open_driver(&bdrv_test, "bs_a", BDRV_O_RDWR, &error_abort);
 | |
| 
 | |
|     /* Create bs_b and bc_c */
 | |
|     options = qdict_new();
 | |
|     qdict_put_str(options, "driver", "raw");
 | |
|     qdict_put_str(options, "file", "bs_a");
 | |
|     qdict_put_str(options, "node-name", "bs_b");
 | |
|     bs_b = bdrv_open(NULL, NULL, options, BDRV_O_RDWR, &error_abort);
 | |
| 
 | |
|     options = qdict_new();
 | |
|     qdict_put_str(options, "driver", "raw");
 | |
|     qdict_put_str(options, "file", "bs_a");
 | |
|     qdict_put_str(options, "node-name", "bs_c");
 | |
|     bs_c = bdrv_open(NULL, NULL, options, BDRV_O_RDWR, &error_abort);
 | |
| 
 | |
|     /* Create blkverify filter that references both bs_b and bs_c */
 | |
|     options = qdict_new();
 | |
|     qdict_put_str(options, "driver", "blkverify");
 | |
|     qdict_put_str(options, "test", "bs_b");
 | |
|     qdict_put_str(options, "raw", "bs_c");
 | |
| 
 | |
|     bs_verify = bdrv_open(NULL, NULL, options, BDRV_O_RDWR, &error_abort);
 | |
|     /*
 | |
|      * Do not take the RESIZE permission: This would require the same
 | |
|      * from bs_c and thus from bs_a; however, blkverify will not share
 | |
|      * it on bs_b, and thus it will not be available for bs_a.
 | |
|      */
 | |
|     blk = blk_new(qemu_get_aio_context(), BLK_PERM_ALL & ~BLK_PERM_RESIZE,
 | |
|                   BLK_PERM_ALL);
 | |
|     blk_insert_bs(blk, bs_verify, &error_abort);
 | |
| 
 | |
|     /* Switch the AioContext */
 | |
|     blk_set_aio_context(blk, ctx, &error_abort);
 | |
|     g_assert(blk_get_aio_context(blk) == ctx);
 | |
|     g_assert(bdrv_get_aio_context(bs_verify) == ctx);
 | |
|     g_assert(bdrv_get_aio_context(bs_a) == ctx);
 | |
|     g_assert(bdrv_get_aio_context(bs_b) == ctx);
 | |
|     g_assert(bdrv_get_aio_context(bs_c) == ctx);
 | |
| 
 | |
|     /* Switch the AioContext back */
 | |
|     main_ctx = qemu_get_aio_context();
 | |
|     blk_set_aio_context(blk, main_ctx, &error_abort);
 | |
|     g_assert(blk_get_aio_context(blk) == main_ctx);
 | |
|     g_assert(bdrv_get_aio_context(bs_verify) == main_ctx);
 | |
|     g_assert(bdrv_get_aio_context(bs_a) == main_ctx);
 | |
|     g_assert(bdrv_get_aio_context(bs_b) == main_ctx);
 | |
|     g_assert(bdrv_get_aio_context(bs_c) == main_ctx);
 | |
| 
 | |
|     blk_unref(blk);
 | |
|     bdrv_unref(bs_verify);
 | |
|     bdrv_unref(bs_c);
 | |
|     bdrv_unref(bs_b);
 | |
|     bdrv_unref(bs_a);
 | |
| }
 | |
| 
 | |
| static void test_propagate_mirror(void)
 | |
| {
 | |
|     IOThread *iothread = iothread_new();
 | |
|     AioContext *ctx = iothread_get_aio_context(iothread);
 | |
|     AioContext *main_ctx = qemu_get_aio_context();
 | |
|     BlockDriverState *src, *target, *filter;
 | |
|     BlockBackend *blk;
 | |
|     Job *job;
 | |
|     Error *local_err = NULL;
 | |
| 
 | |
|     /* Create src and target*/
 | |
|     src = bdrv_new_open_driver(&bdrv_test, "src", BDRV_O_RDWR, &error_abort);
 | |
|     target = bdrv_new_open_driver(&bdrv_test, "target", BDRV_O_RDWR,
 | |
|                                   &error_abort);
 | |
| 
 | |
|     /* Start a mirror job */
 | |
|     mirror_start("job0", src, target, NULL, JOB_DEFAULT, 0, 0, 0,
 | |
|                  MIRROR_SYNC_MODE_NONE, MIRROR_OPEN_BACKING_CHAIN, false,
 | |
|                  BLOCKDEV_ON_ERROR_REPORT, BLOCKDEV_ON_ERROR_REPORT,
 | |
|                  false, "filter_node", MIRROR_COPY_MODE_BACKGROUND,
 | |
|                  &error_abort);
 | |
| 
 | |
|     WITH_JOB_LOCK_GUARD() {
 | |
|         job = job_get_locked("job0");
 | |
|     }
 | |
|     filter = bdrv_find_node("filter_node");
 | |
| 
 | |
|     /* Change the AioContext of src */
 | |
|     bdrv_try_change_aio_context(src, ctx, NULL, &error_abort);
 | |
|     g_assert(bdrv_get_aio_context(src) == ctx);
 | |
|     g_assert(bdrv_get_aio_context(target) == ctx);
 | |
|     g_assert(bdrv_get_aio_context(filter) == ctx);
 | |
|     g_assert(job->aio_context == ctx);
 | |
| 
 | |
|     /* Change the AioContext of target */
 | |
|     bdrv_try_change_aio_context(target, main_ctx, NULL, &error_abort);
 | |
|     g_assert(bdrv_get_aio_context(src) == main_ctx);
 | |
|     g_assert(bdrv_get_aio_context(target) == main_ctx);
 | |
|     g_assert(bdrv_get_aio_context(filter) == main_ctx);
 | |
| 
 | |
|     /* With a BlockBackend on src, changing target must fail */
 | |
|     blk = blk_new(qemu_get_aio_context(), 0, BLK_PERM_ALL);
 | |
|     blk_insert_bs(blk, src, &error_abort);
 | |
| 
 | |
|     bdrv_try_change_aio_context(target, ctx, NULL, &local_err);
 | |
|     error_free_or_abort(&local_err);
 | |
| 
 | |
|     g_assert(blk_get_aio_context(blk) == main_ctx);
 | |
|     g_assert(bdrv_get_aio_context(src) == main_ctx);
 | |
|     g_assert(bdrv_get_aio_context(target) == main_ctx);
 | |
|     g_assert(bdrv_get_aio_context(filter) == main_ctx);
 | |
| 
 | |
|     /* ...unless we explicitly allow it */
 | |
|     blk_set_allow_aio_context_change(blk, true);
 | |
|     bdrv_try_change_aio_context(target, ctx, NULL, &error_abort);
 | |
| 
 | |
|     g_assert(blk_get_aio_context(blk) == ctx);
 | |
|     g_assert(bdrv_get_aio_context(src) == ctx);
 | |
|     g_assert(bdrv_get_aio_context(target) == ctx);
 | |
|     g_assert(bdrv_get_aio_context(filter) == ctx);
 | |
| 
 | |
|     job_cancel_sync_all();
 | |
| 
 | |
|     blk_set_aio_context(blk, main_ctx, &error_abort);
 | |
|     bdrv_try_change_aio_context(target, main_ctx, NULL, &error_abort);
 | |
| 
 | |
|     blk_unref(blk);
 | |
|     bdrv_unref(src);
 | |
|     bdrv_unref(target);
 | |
| }
 | |
| 
 | |
| static void test_attach_second_node(void)
 | |
| {
 | |
|     IOThread *iothread = iothread_new();
 | |
|     AioContext *ctx = iothread_get_aio_context(iothread);
 | |
|     AioContext *main_ctx = qemu_get_aio_context();
 | |
|     BlockBackend *blk;
 | |
|     BlockDriverState *bs, *filter;
 | |
|     QDict *options;
 | |
| 
 | |
|     blk = blk_new(ctx, BLK_PERM_ALL, BLK_PERM_ALL);
 | |
|     bs = bdrv_new_open_driver(&bdrv_test, "base", BDRV_O_RDWR, &error_abort);
 | |
|     blk_insert_bs(blk, bs, &error_abort);
 | |
| 
 | |
|     options = qdict_new();
 | |
|     qdict_put_str(options, "driver", "raw");
 | |
|     qdict_put_str(options, "file", "base");
 | |
| 
 | |
|     filter = bdrv_open(NULL, NULL, options, BDRV_O_RDWR, &error_abort);
 | |
| 
 | |
|     g_assert(blk_get_aio_context(blk) == ctx);
 | |
|     g_assert(bdrv_get_aio_context(bs) == ctx);
 | |
|     g_assert(bdrv_get_aio_context(filter) == ctx);
 | |
| 
 | |
|     blk_set_aio_context(blk, main_ctx, &error_abort);
 | |
|     g_assert(blk_get_aio_context(blk) == main_ctx);
 | |
|     g_assert(bdrv_get_aio_context(bs) == main_ctx);
 | |
|     g_assert(bdrv_get_aio_context(filter) == main_ctx);
 | |
| 
 | |
|     bdrv_unref(filter);
 | |
|     bdrv_unref(bs);
 | |
|     blk_unref(blk);
 | |
| }
 | |
| 
 | |
| static void test_attach_preserve_blk_ctx(void)
 | |
| {
 | |
|     IOThread *iothread = iothread_new();
 | |
|     AioContext *ctx = iothread_get_aio_context(iothread);
 | |
|     BlockBackend *blk;
 | |
|     BlockDriverState *bs;
 | |
| 
 | |
|     blk = blk_new(ctx, BLK_PERM_ALL, BLK_PERM_ALL);
 | |
|     bs = bdrv_new_open_driver(&bdrv_test, "base", BDRV_O_RDWR, &error_abort);
 | |
|     bs->total_sectors = 65536 / BDRV_SECTOR_SIZE;
 | |
| 
 | |
|     /* Add node to BlockBackend that has an iothread context assigned */
 | |
|     blk_insert_bs(blk, bs, &error_abort);
 | |
|     g_assert(blk_get_aio_context(blk) == ctx);
 | |
|     g_assert(bdrv_get_aio_context(bs) == ctx);
 | |
| 
 | |
|     /* Remove the node again */
 | |
|     blk_remove_bs(blk);
 | |
|     g_assert(blk_get_aio_context(blk) == ctx);
 | |
|     g_assert(bdrv_get_aio_context(bs) == qemu_get_aio_context());
 | |
| 
 | |
|     /* Re-attach the node */
 | |
|     blk_insert_bs(blk, bs, &error_abort);
 | |
|     g_assert(blk_get_aio_context(blk) == ctx);
 | |
|     g_assert(bdrv_get_aio_context(bs) == ctx);
 | |
| 
 | |
|     blk_set_aio_context(blk, qemu_get_aio_context(), &error_abort);
 | |
|     bdrv_unref(bs);
 | |
|     blk_unref(blk);
 | |
| }
 | |
| 
 | |
| int main(int argc, char **argv)
 | |
| {
 | |
|     int i;
 | |
| 
 | |
|     bdrv_init();
 | |
|     qemu_init_main_loop(&error_abort);
 | |
| 
 | |
|     g_test_init(&argc, &argv, NULL);
 | |
| 
 | |
|     for (i = 0; i < ARRAY_SIZE(sync_op_tests); i++) {
 | |
|         const SyncOpTest *t = &sync_op_tests[i];
 | |
|         g_test_add_data_func(t->name, t, test_sync_op);
 | |
|     }
 | |
| 
 | |
|     g_test_add_func("/attach/blockjob", test_attach_blockjob);
 | |
|     g_test_add_func("/attach/second_node", test_attach_second_node);
 | |
|     g_test_add_func("/attach/preserve_blk_ctx", test_attach_preserve_blk_ctx);
 | |
|     g_test_add_func("/propagate/basic", test_propagate_basic);
 | |
|     g_test_add_func("/propagate/diamond", test_propagate_diamond);
 | |
|     g_test_add_func("/propagate/mirror", test_propagate_mirror);
 | |
| 
 | |
|     return g_test_run();
 | |
| }
 |