new file mode 100755
@@ -0,0 +1,89 @@
+#!/usr/bin/env python3
+# group: rw
+#
+# Tests for block-job-complete immediately after a drain
+#
+# Copyright (c) 2021 Red Hat, Inc.
+#
+# This program is free software; you can redistribute it and/or modify
+# it under the terms of the GNU General Public License as published by
+# the Free Software Foundation; either version 2 of the License, or
+# (at your option) any later version.
+#
+# This program is distributed in the hope that it will be useful,
+# but WITHOUT ANY WARRANTY; without even the implied warranty of
+# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+# GNU General Public License for more details.
+#
+# You should have received a copy of the GNU General Public License
+# along with this program. If not, see <http://www.gnu.org/licenses/>.
+#
+
+import iotests
+
+iotests.script_initialize(supported_fmts=['raw'])
+
+DISK_JOBS = 4
+NULL_JOBS = 1
+
+
+# We cannot auto-generate these in a loop because the files are
+# deleted when their scope ends
+src_imgs = iotests.file_path('src0', 'src1', 'src2', 'src3')
+dst_imgs = iotests.file_path('dst0', 'dst1', 'dst2', 'dst3')
+
+assert len(src_imgs) == DISK_JOBS
+assert len(dst_imgs) == DISK_JOBS
+
+
+for i in range(DISK_JOBS):
+ ret = iotests.qemu_img('create', '-f', iotests.imgfmt, src_imgs[i], '128M')
+ assert ret == 0
+
+ ret = iotests.qemu_img('create', '-f', iotests.imgfmt, dst_imgs[i], '128M')
+ assert ret == 0
+
+with iotests.VM() as vm:
+ vm.add_object('iothread,id=iothr0')
+ vm.add_device('virtio-scsi,iothread=iothr0')
+
+ for i in range(DISK_JOBS):
+ vm.add_blockdev(f'file,node-name=source-disk-{i},'
+ f'filename={src_imgs[i]}')
+
+ vm.add_blockdev(f'file,node-name=target-disk-{i},'
+ f'filename={dst_imgs[i]}')
+
+ vm.add_device(f'scsi-hd,id=device-disk-{i},drive=source-disk-{i}')
+
+ for i in range(NULL_JOBS):
+ vm.add_blockdev(f'null-co,node-name=source-null-{i},read-zeroes=on')
+ vm.add_blockdev(f'null-co,node-name=target-null-{i},read-zeroes=on')
+ vm.add_device(f'scsi-hd,id=device-null-{i},drive=source-null-{i}')
+
+ vm.launch()
+
+ for i in range(DISK_JOBS):
+ vm.qmp_log('blockdev-mirror',
+ job_id=f'mirror-disk-{i}',
+ device=f'source-disk-{i}',
+ target=f'target-disk-{i}',
+ sync='full',
+ granularity=1048576,
+ buf_size=(16 * 1048576))
+
+ for i in range(NULL_JOBS):
+ vm.qmp_log('blockdev-mirror',
+ job_id=f'mirror-null-{i}',
+ device=f'source-null-{i}',
+ target=f'target-null-{i}',
+ sync='full')
+
+ for i in range(DISK_JOBS + NULL_JOBS):
+ vm.event_wait('BLOCK_JOB_READY')
+
+ for i in range(DISK_JOBS):
+ vm.hmp(f'qemu-io -d device-disk-{i} "write 0 128M"')
+
+ vm.qmp_log('transaction', actions=[])
+ vm.qmp_log('block-job-complete', device='mirror-null-0')
new file mode 100644
@@ -0,0 +1,14 @@
+{"execute": "blockdev-mirror", "arguments": {"buf-size": 16777216, "device": "source-disk-0", "granularity": 1048576, "job-id": "mirror-disk-0", "sync": "full", "target": "target-disk-0"}}
+{"return": {}}
+{"execute": "blockdev-mirror", "arguments": {"buf-size": 16777216, "device": "source-disk-1", "granularity": 1048576, "job-id": "mirror-disk-1", "sync": "full", "target": "target-disk-1"}}
+{"return": {}}
+{"execute": "blockdev-mirror", "arguments": {"buf-size": 16777216, "device": "source-disk-2", "granularity": 1048576, "job-id": "mirror-disk-2", "sync": "full", "target": "target-disk-2"}}
+{"return": {}}
+{"execute": "blockdev-mirror", "arguments": {"buf-size": 16777216, "device": "source-disk-3", "granularity": 1048576, "job-id": "mirror-disk-3", "sync": "full", "target": "target-disk-3"}}
+{"return": {}}
+{"execute": "blockdev-mirror", "arguments": {"device": "source-null-0", "job-id": "mirror-null-0", "sync": "full", "target": "target-null-0"}}
+{"return": {}}
+{"execute": "transaction", "arguments": {"actions": []}}
+{"return": {}}
+{"execute": "block-job-complete", "arguments": {"device": "mirror-null-0"}}
+{"return": {}}
Test what happens when you have multiple busy block jobs, drain all (via an empty transaction), and immediately issue a block-job-complete on one of the jobs. Sometimes it will still be in STANDBY, in which case block-job-complete used to fail. It should not. Signed-off-by: Max Reitz <mreitz@redhat.com> --- .../tests/mirror-complete-after-drain | 89 +++++++++++++++++++ .../tests/mirror-complete-after-drain.out | 14 +++ 2 files changed, 103 insertions(+) create mode 100755 tests/qemu-iotests/tests/mirror-complete-after-drain create mode 100644 tests/qemu-iotests/tests/mirror-complete-after-drain.out