8cfbe929e8
Put an NBD block device into an I/O thread, and then read data from it,
hoping that the NBD connection will yield during that read. When it
does, the coroutine must be reentered in the block device's I/O thread,
which will only happen if the NBD block driver attaches the connection's
QIOChannel to the new AioContext. It did not do that after 4ddb5d2fde
("block/nbd: drop connection_co") and prior to "block/nbd: Move s->ioc
on AioContext change", which would cause an assertion failure.
To improve our chances of yielding, the NBD server is throttled to
reading 64 kB/s, and the NBD client reads 128 kB, so it should yield at
some point.
Reviewed-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
Signed-off-by: Hanna Reitz <hreitz@redhat.com>
Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
347 lines
13 KiB
Python
Executable File
347 lines
13 KiB
Python
Executable File
#!/usr/bin/env python3
|
|
# group: rw
|
|
#
|
|
# Test cases for blockdev + IOThread interactions
|
|
#
|
|
# Copyright (C) 2019 Red Hat, Inc.
|
|
#
|
|
# This program is free software; you can redistribute it and/or modify
|
|
# it under the terms of the GNU General Public License as published by
|
|
# the Free Software Foundation; either version 2 of the License, or
|
|
# (at your option) any later version.
|
|
#
|
|
# This program is distributed in the hope that it will be useful,
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
# GNU General Public License for more details.
|
|
#
|
|
# You should have received a copy of the GNU General Public License
|
|
# along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
#
|
|
|
|
import os
|
|
import time
|
|
import iotests
|
|
from iotests import qemu_img, QemuStorageDaemon
|
|
|
|
image_len = 64 * 1024 * 1024
|
|
|
|
# Test for RHBZ#1782175
|
|
class TestDirtyBitmapIOThread(iotests.QMPTestCase):
|
|
drive0_img = os.path.join(iotests.test_dir, 'drive0.img')
|
|
images = { 'drive0': drive0_img }
|
|
|
|
def setUp(self):
|
|
for name in self.images:
|
|
qemu_img('create', '-f', iotests.imgfmt,
|
|
self.images[name], str(image_len))
|
|
|
|
self.vm = iotests.VM()
|
|
self.vm.add_object('iothread,id=iothread0')
|
|
|
|
for name in self.images:
|
|
self.vm.add_blockdev('driver=file,filename=%s,node-name=file_%s'
|
|
% (self.images[name], name))
|
|
self.vm.add_blockdev('driver=qcow2,file=file_%s,node-name=%s'
|
|
% (name, name))
|
|
|
|
self.vm.launch()
|
|
self.vm.qmp('x-blockdev-set-iothread',
|
|
node_name='drive0', iothread='iothread0',
|
|
force=True)
|
|
|
|
def tearDown(self):
|
|
self.vm.shutdown()
|
|
for name in self.images:
|
|
os.remove(self.images[name])
|
|
|
|
def test_add_dirty_bitmap(self):
|
|
result = self.vm.qmp(
|
|
'block-dirty-bitmap-add',
|
|
node='drive0',
|
|
name='bitmap1',
|
|
persistent=True,
|
|
)
|
|
|
|
self.assert_qmp(result, 'return', {})
|
|
|
|
|
|
# Test for RHBZ#1746217 & RHBZ#1773517
|
|
class TestNBDMirrorIOThread(iotests.QMPTestCase):
|
|
nbd_sock = os.path.join(iotests.sock_dir, 'nbd.sock')
|
|
drive0_img = os.path.join(iotests.test_dir, 'drive0.img')
|
|
mirror_img = os.path.join(iotests.test_dir, 'mirror.img')
|
|
images = { 'drive0': drive0_img, 'mirror': mirror_img }
|
|
|
|
def setUp(self):
|
|
for name in self.images:
|
|
qemu_img('create', '-f', iotests.imgfmt,
|
|
self.images[name], str(image_len))
|
|
|
|
self.vm_src = iotests.VM(path_suffix='src')
|
|
self.vm_src.add_object('iothread,id=iothread0')
|
|
self.vm_src.add_blockdev('driver=file,filename=%s,node-name=file0'
|
|
% (self.drive0_img))
|
|
self.vm_src.add_blockdev('driver=qcow2,file=file0,node-name=drive0')
|
|
self.vm_src.launch()
|
|
self.vm_src.qmp('x-blockdev-set-iothread',
|
|
node_name='drive0', iothread='iothread0',
|
|
force=True)
|
|
|
|
self.vm_tgt = iotests.VM(path_suffix='tgt')
|
|
self.vm_tgt.add_object('iothread,id=iothread0')
|
|
self.vm_tgt.add_blockdev('driver=file,filename=%s,node-name=file0'
|
|
% (self.mirror_img))
|
|
self.vm_tgt.add_blockdev('driver=qcow2,file=file0,node-name=drive0')
|
|
self.vm_tgt.launch()
|
|
self.vm_tgt.qmp('x-blockdev-set-iothread',
|
|
node_name='drive0', iothread='iothread0',
|
|
force=True)
|
|
|
|
def tearDown(self):
|
|
self.vm_src.shutdown()
|
|
self.vm_tgt.shutdown()
|
|
for name in self.images:
|
|
os.remove(self.images[name])
|
|
|
|
def test_nbd_mirror(self):
|
|
result = self.vm_tgt.qmp(
|
|
'nbd-server-start',
|
|
addr={
|
|
'type': 'unix',
|
|
'data': { 'path': self.nbd_sock }
|
|
}
|
|
)
|
|
self.assert_qmp(result, 'return', {})
|
|
|
|
result = self.vm_tgt.qmp(
|
|
'nbd-server-add',
|
|
device='drive0',
|
|
writable=True
|
|
)
|
|
self.assert_qmp(result, 'return', {})
|
|
|
|
result = self.vm_src.qmp(
|
|
'drive-mirror',
|
|
device='drive0',
|
|
target='nbd+unix:///drive0?socket=' + self.nbd_sock,
|
|
sync='full',
|
|
mode='existing',
|
|
speed=64*1024*1024,
|
|
job_id='j1'
|
|
)
|
|
self.assert_qmp(result, 'return', {})
|
|
|
|
self.vm_src.event_wait(name="BLOCK_JOB_READY")
|
|
|
|
|
|
# Test for RHBZ#1779036
|
|
class TestExternalSnapshotAbort(iotests.QMPTestCase):
|
|
drive0_img = os.path.join(iotests.test_dir, 'drive0.img')
|
|
snapshot_img = os.path.join(iotests.test_dir, 'snapshot.img')
|
|
images = { 'drive0': drive0_img, 'snapshot': snapshot_img }
|
|
|
|
def setUp(self):
|
|
for name in self.images:
|
|
qemu_img('create', '-f', iotests.imgfmt,
|
|
self.images[name], str(image_len))
|
|
|
|
self.vm = iotests.VM()
|
|
self.vm.add_object('iothread,id=iothread0')
|
|
self.vm.add_blockdev('driver=file,filename=%s,node-name=file0'
|
|
% (self.drive0_img))
|
|
self.vm.add_blockdev('driver=qcow2,file=file0,node-name=drive0')
|
|
self.vm.launch()
|
|
self.vm.qmp('x-blockdev-set-iothread',
|
|
node_name='drive0', iothread='iothread0',
|
|
force=True)
|
|
|
|
def tearDown(self):
|
|
self.vm.shutdown()
|
|
for name in self.images:
|
|
os.remove(self.images[name])
|
|
|
|
def test_external_snapshot_abort(self):
|
|
# Use a two actions transaction with a bogus values on the second
|
|
# one to trigger an abort of the transaction.
|
|
result = self.vm.qmp('transaction', actions=[
|
|
{
|
|
'type': 'blockdev-snapshot-sync',
|
|
'data': { 'node-name': 'drive0',
|
|
'snapshot-file': self.snapshot_img,
|
|
'snapshot-node-name': 'snap1',
|
|
'mode': 'absolute-paths',
|
|
'format': 'qcow2' }
|
|
},
|
|
{
|
|
'type': 'blockdev-snapshot-sync',
|
|
'data': { 'node-name': 'drive0',
|
|
'snapshot-file': '/fakesnapshot',
|
|
'snapshot-node-name': 'snap2',
|
|
'mode': 'absolute-paths',
|
|
'format': 'qcow2' }
|
|
},
|
|
])
|
|
|
|
# Crashes on failure, we expect this error.
|
|
self.assert_qmp(result, 'error/class', 'GenericError')
|
|
|
|
|
|
# Test for RHBZ#1782111
|
|
class TestBlockdevBackupAbort(iotests.QMPTestCase):
|
|
drive0_img = os.path.join(iotests.test_dir, 'drive0.img')
|
|
drive1_img = os.path.join(iotests.test_dir, 'drive1.img')
|
|
snap0_img = os.path.join(iotests.test_dir, 'snap0.img')
|
|
snap1_img = os.path.join(iotests.test_dir, 'snap1.img')
|
|
images = { 'drive0': drive0_img,
|
|
'drive1': drive1_img,
|
|
'snap0': snap0_img,
|
|
'snap1': snap1_img }
|
|
|
|
def setUp(self):
|
|
for name in self.images:
|
|
qemu_img('create', '-f', iotests.imgfmt,
|
|
self.images[name], str(image_len))
|
|
|
|
self.vm = iotests.VM()
|
|
self.vm.add_object('iothread,id=iothread0')
|
|
self.vm.add_device('virtio-scsi,iothread=iothread0')
|
|
|
|
for name in self.images:
|
|
self.vm.add_blockdev('driver=file,filename=%s,node-name=file_%s'
|
|
% (self.images[name], name))
|
|
self.vm.add_blockdev('driver=qcow2,file=file_%s,node-name=%s'
|
|
% (name, name))
|
|
|
|
self.vm.add_device('scsi-hd,drive=drive0')
|
|
self.vm.add_device('scsi-hd,drive=drive1')
|
|
self.vm.launch()
|
|
|
|
def tearDown(self):
|
|
self.vm.shutdown()
|
|
for name in self.images:
|
|
os.remove(self.images[name])
|
|
|
|
def test_blockdev_backup_abort(self):
|
|
# Use a two actions transaction with a bogus values on the second
|
|
# one to trigger an abort of the transaction.
|
|
result = self.vm.qmp('transaction', actions=[
|
|
{
|
|
'type': 'blockdev-backup',
|
|
'data': { 'device': 'drive0',
|
|
'target': 'snap0',
|
|
'sync': 'full',
|
|
'job-id': 'j1' }
|
|
},
|
|
{
|
|
'type': 'blockdev-backup',
|
|
'data': { 'device': 'drive1',
|
|
'target': 'snap1',
|
|
'sync': 'full' }
|
|
},
|
|
])
|
|
|
|
# Hangs on failure, we expect this error.
|
|
self.assert_qmp(result, 'error/class', 'GenericError')
|
|
|
|
# Test for RHBZ#2033626
|
|
class TestYieldingAndTimers(iotests.QMPTestCase):
|
|
sock = os.path.join(iotests.sock_dir, 'nbd.sock')
|
|
qsd = None
|
|
|
|
def setUp(self):
|
|
self.create_nbd_export()
|
|
|
|
# Simple VM with an NBD block device connected to the NBD export
|
|
# provided by the QSD, and an (initially unused) iothread
|
|
self.vm = iotests.VM()
|
|
self.vm.add_object('iothread,id=iothr')
|
|
self.vm.add_blockdev('nbd,node-name=nbd,server.type=unix,' +
|
|
f'server.path={self.sock},export=exp,' +
|
|
'reconnect-delay=1,open-timeout=1')
|
|
|
|
self.vm.launch()
|
|
|
|
def tearDown(self):
|
|
self.stop_nbd_export()
|
|
self.vm.shutdown()
|
|
|
|
def test_timers_with_blockdev_del(self):
|
|
# The NBD BDS will have had an active open timer, because setUp() gave
|
|
# a positive value for @open-timeout. It should be gone once the BDS
|
|
# has been opened.
|
|
# (But there used to be a bug where it remained active, which will
|
|
# become important below.)
|
|
|
|
# Stop and restart the NBD server, and do some I/O on the client to
|
|
# trigger a reconnect and start the reconnect delay timer
|
|
self.stop_nbd_export()
|
|
self.create_nbd_export()
|
|
|
|
result = self.vm.qmp('human-monitor-command',
|
|
command_line='qemu-io nbd "write 0 512"')
|
|
self.assert_qmp(result, 'return', '')
|
|
|
|
# Reconnect is done, so the reconnect delay timer should be gone.
|
|
# (This is similar to how the open timer should be gone after open,
|
|
# and similarly there used to be a bug where it was not gone.)
|
|
|
|
# Delete the BDS to see whether both timers are gone. If they are not,
|
|
# they will remain active, fire later, and then access freed data.
|
|
# (Or, with "block/nbd: Assert there are no timers when closed"
|
|
# applied, the assertions added in that patch will fail.)
|
|
result = self.vm.qmp('blockdev-del', node_name='nbd')
|
|
self.assert_qmp(result, 'return', {})
|
|
|
|
# Give the timers some time to fire (both have a timeout of 1 s).
|
|
# (Sleeping in an iotest may ring some alarm bells, but note that if
|
|
# the timing is off here, the test will just always pass. If we kill
|
|
# the VM too early, then we just kill the timers before they can fire,
|
|
# thus not see the error, and so the test will pass.)
|
|
time.sleep(2)
|
|
|
|
def test_yield_in_iothread(self):
|
|
# Move the NBD node to the I/O thread; the NBD block driver should
|
|
# attach the connection's QIOChannel to that thread's AioContext, too
|
|
result = self.vm.qmp('x-blockdev-set-iothread',
|
|
node_name='nbd', iothread='iothr')
|
|
self.assert_qmp(result, 'return', {})
|
|
|
|
# Do some I/O that will be throttled by the QSD, so that the network
|
|
# connection hopefully will yield here. When it is resumed, it must
|
|
# then be resumed in the I/O thread's AioContext.
|
|
result = self.vm.qmp('human-monitor-command',
|
|
command_line='qemu-io nbd "read 0 128K"')
|
|
self.assert_qmp(result, 'return', '')
|
|
|
|
def create_nbd_export(self):
|
|
assert self.qsd is None
|
|
|
|
# Export a throttled null-co BDS: Reads are throttled (max 64 kB/s),
|
|
# writes are not.
|
|
self.qsd = QemuStorageDaemon(
|
|
'--object',
|
|
'throttle-group,id=thrgr,x-bps-read=65536,x-bps-read-max=65536',
|
|
|
|
'--blockdev',
|
|
'null-co,node-name=null,read-zeroes=true',
|
|
|
|
'--blockdev',
|
|
'throttle,node-name=thr,file=null,throttle-group=thrgr',
|
|
|
|
'--nbd-server',
|
|
f'addr.type=unix,addr.path={self.sock}',
|
|
|
|
'--export',
|
|
'nbd,id=exp,node-name=thr,name=exp,writable=true'
|
|
)
|
|
|
|
def stop_nbd_export(self):
|
|
self.qsd.stop()
|
|
self.qsd = None
|
|
|
|
if __name__ == '__main__':
|
|
iotests.main(supported_fmts=['qcow2'],
|
|
supported_protocols=['file'],
|
|
unsupported_imgopts=['compat'])
|