-----BEGIN PGP SIGNATURE-----
iQEcBAABAgAGBQJZrqaUAAoJEJykq7OBq3PI1o0H/RbjT8TJrMdMdaZRyVLDp9ER kmzTnG0v7sz1mWSpOVcjcMfwjnHHw9kW4hyt2OQPhxSmeoHdlcKCHQ2CEW1ODFv4 m183XFIGrSoE62akcPlkfg2jUiLF6GrdZP3XEeWDcCre5mMONzN8R2gQZfGk3QLV vNkR1TBJUcA9ZBIhNObvmY8sGDag63pLkSufQzMD67b1SZAhFHFLrkYmmt+ftbRi GCNntQs8hGAlcHjVpZ/aRJQrlugDxYDikkKK/bpt26D7PjyFTb6O+RLfexK5+FHE 81PDWjOY87yFaSo6o5zvRbsxoMrWa6oeW8LCOeVf/vMtEDxJWRm91iUvXSRJkUs= =BoAP -----END PGP SIGNATURE----- Merge remote-tracking branch 'remotes/stefanha/tags/block-pull-request' into staging # gpg: Signature made Tue 05 Sep 2017 14:28:52 BST # gpg: using RSA key 0x9CA4ABB381AB73C8 # gpg: Good signature from "Stefan Hajnoczi <stefanha@redhat.com>" # gpg: aka "Stefan Hajnoczi <stefanha@gmail.com>" # Primary key fingerprint: 8695 A8BF D3F9 7CDA AC35 775A 9CA4 ABB3 81AB 73C8 * remotes/stefanha/tags/block-pull-request: block: document semantics of bdrv_co_preadv|pwritev qemu-iotests: use context managers for resource cleanup in 194 iotests.py: add FilePath context manager qemu.py: make VM() a context manager Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
This commit is contained in:
commit
6bbd7e2728
|
@ -146,12 +146,43 @@ struct BlockDriver {
|
||||||
|
|
||||||
int coroutine_fn (*bdrv_co_readv)(BlockDriverState *bs,
|
int coroutine_fn (*bdrv_co_readv)(BlockDriverState *bs,
|
||||||
int64_t sector_num, int nb_sectors, QEMUIOVector *qiov);
|
int64_t sector_num, int nb_sectors, QEMUIOVector *qiov);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @offset: position in bytes to read at
|
||||||
|
* @bytes: number of bytes to read
|
||||||
|
* @qiov: the buffers to fill with read data
|
||||||
|
* @flags: currently unused, always 0
|
||||||
|
*
|
||||||
|
* @offset and @bytes will be a multiple of 'request_alignment',
|
||||||
|
* but the length of individual @qiov elements does not have to
|
||||||
|
* be a multiple.
|
||||||
|
*
|
||||||
|
* @bytes will always equal the total size of @qiov, and will be
|
||||||
|
* no larger than 'max_transfer'.
|
||||||
|
*
|
||||||
|
* The buffer in @qiov may point directly to guest memory.
|
||||||
|
*/
|
||||||
int coroutine_fn (*bdrv_co_preadv)(BlockDriverState *bs,
|
int coroutine_fn (*bdrv_co_preadv)(BlockDriverState *bs,
|
||||||
uint64_t offset, uint64_t bytes, QEMUIOVector *qiov, int flags);
|
uint64_t offset, uint64_t bytes, QEMUIOVector *qiov, int flags);
|
||||||
int coroutine_fn (*bdrv_co_writev)(BlockDriverState *bs,
|
int coroutine_fn (*bdrv_co_writev)(BlockDriverState *bs,
|
||||||
int64_t sector_num, int nb_sectors, QEMUIOVector *qiov);
|
int64_t sector_num, int nb_sectors, QEMUIOVector *qiov);
|
||||||
int coroutine_fn (*bdrv_co_writev_flags)(BlockDriverState *bs,
|
int coroutine_fn (*bdrv_co_writev_flags)(BlockDriverState *bs,
|
||||||
int64_t sector_num, int nb_sectors, QEMUIOVector *qiov, int flags);
|
int64_t sector_num, int nb_sectors, QEMUIOVector *qiov, int flags);
|
||||||
|
/**
|
||||||
|
* @offset: position in bytes to write at
|
||||||
|
* @bytes: number of bytes to write
|
||||||
|
* @qiov: the buffers containing data to write
|
||||||
|
* @flags: zero or more bits allowed by 'supported_write_flags'
|
||||||
|
*
|
||||||
|
* @offset and @bytes will be a multiple of 'request_alignment',
|
||||||
|
* but the length of individual @qiov elements does not have to
|
||||||
|
* be a multiple.
|
||||||
|
*
|
||||||
|
* @bytes will always equal the total size of @qiov, and will be
|
||||||
|
* no larger than 'max_transfer'.
|
||||||
|
*
|
||||||
|
* The buffer in @qiov may point directly to guest memory.
|
||||||
|
*/
|
||||||
int coroutine_fn (*bdrv_co_pwritev)(BlockDriverState *bs,
|
int coroutine_fn (*bdrv_co_pwritev)(BlockDriverState *bs,
|
||||||
uint64_t offset, uint64_t bytes, QEMUIOVector *qiov, int flags);
|
uint64_t offset, uint64_t bytes, QEMUIOVector *qiov, int flags);
|
||||||
|
|
||||||
|
|
|
@ -21,7 +21,14 @@ import qmp.qmp
|
||||||
|
|
||||||
|
|
||||||
class QEMUMachine(object):
|
class QEMUMachine(object):
|
||||||
'''A QEMU VM'''
|
'''A QEMU VM
|
||||||
|
|
||||||
|
Use this object as a context manager to ensure the QEMU process terminates::
|
||||||
|
|
||||||
|
with VM(binary) as vm:
|
||||||
|
...
|
||||||
|
# vm is guaranteed to be shut down here
|
||||||
|
'''
|
||||||
|
|
||||||
def __init__(self, binary, args=[], wrapper=[], name=None, test_dir="/var/tmp",
|
def __init__(self, binary, args=[], wrapper=[], name=None, test_dir="/var/tmp",
|
||||||
monitor_address=None, socket_scm_helper=None, debug=False):
|
monitor_address=None, socket_scm_helper=None, debug=False):
|
||||||
|
@ -40,6 +47,13 @@ class QEMUMachine(object):
|
||||||
self._socket_scm_helper = socket_scm_helper
|
self._socket_scm_helper = socket_scm_helper
|
||||||
self._debug = debug
|
self._debug = debug
|
||||||
|
|
||||||
|
def __enter__(self):
|
||||||
|
return self
|
||||||
|
|
||||||
|
def __exit__(self, exc_type, exc_val, exc_tb):
|
||||||
|
self.shutdown()
|
||||||
|
return False
|
||||||
|
|
||||||
# This can be used to add an unused monitor instance.
|
# This can be used to add an unused monitor instance.
|
||||||
def add_monitor_telnet(self, ip, port):
|
def add_monitor_telnet(self, ip, port):
|
||||||
args = 'tcp:%s:%d,server,nowait,telnet' % (ip, port)
|
args = 'tcp:%s:%d,server,nowait,telnet' % (ip, port)
|
||||||
|
|
|
@ -19,66 +19,65 @@
|
||||||
#
|
#
|
||||||
# Non-shared storage migration test using NBD server and drive-mirror
|
# Non-shared storage migration test using NBD server and drive-mirror
|
||||||
|
|
||||||
import os
|
|
||||||
import atexit
|
|
||||||
import iotests
|
import iotests
|
||||||
|
|
||||||
iotests.verify_platform(['linux'])
|
iotests.verify_platform(['linux'])
|
||||||
|
|
||||||
img_size = '1G'
|
with iotests.FilePath('source.img') as source_img_path, \
|
||||||
source_img_path = os.path.join(iotests.test_dir, 'source.img')
|
iotests.FilePath('dest.img') as dest_img_path, \
|
||||||
dest_img_path = os.path.join(iotests.test_dir, 'dest.img')
|
iotests.FilePath('migration.sock') as migration_sock_path, \
|
||||||
iotests.qemu_img_pipe('create', '-f', iotests.imgfmt, source_img_path, img_size)
|
iotests.FilePath('nbd.sock') as nbd_sock_path, \
|
||||||
iotests.qemu_img_pipe('create', '-f', iotests.imgfmt, dest_img_path, img_size)
|
iotests.VM('source') as source_vm, \
|
||||||
|
iotests.VM('dest') as dest_vm:
|
||||||
|
|
||||||
iotests.log('Launching VMs...')
|
img_size = '1G'
|
||||||
migration_sock_path = os.path.join(iotests.test_dir, 'migration.sock')
|
iotests.qemu_img_pipe('create', '-f', iotests.imgfmt, source_img_path, img_size)
|
||||||
nbd_sock_path = os.path.join(iotests.test_dir, 'nbd.sock')
|
iotests.qemu_img_pipe('create', '-f', iotests.imgfmt, dest_img_path, img_size)
|
||||||
source_vm = iotests.VM('source').add_drive(source_img_path)
|
|
||||||
dest_vm = (iotests.VM('dest').add_drive(dest_img_path)
|
|
||||||
.add_incoming('unix:{0}'.format(migration_sock_path)))
|
|
||||||
source_vm.launch()
|
|
||||||
atexit.register(source_vm.shutdown)
|
|
||||||
dest_vm.launch()
|
|
||||||
atexit.register(dest_vm.shutdown)
|
|
||||||
|
|
||||||
iotests.log('Launching NBD server on destination...')
|
iotests.log('Launching VMs...')
|
||||||
iotests.log(dest_vm.qmp('nbd-server-start', addr={'type': 'unix', 'data': {'path': nbd_sock_path}}))
|
(source_vm.add_drive(source_img_path)
|
||||||
iotests.log(dest_vm.qmp('nbd-server-add', device='drive0', writable=True))
|
.launch())
|
||||||
|
(dest_vm.add_drive(dest_img_path)
|
||||||
|
.add_incoming('unix:{0}'.format(migration_sock_path))
|
||||||
|
.launch())
|
||||||
|
|
||||||
iotests.log('Starting `drive-mirror` on source...')
|
iotests.log('Launching NBD server on destination...')
|
||||||
iotests.log(source_vm.qmp(
|
iotests.log(dest_vm.qmp('nbd-server-start', addr={'type': 'unix', 'data': {'path': nbd_sock_path}}))
|
||||||
'drive-mirror',
|
iotests.log(dest_vm.qmp('nbd-server-add', device='drive0', writable=True))
|
||||||
device='drive0',
|
|
||||||
target='nbd+unix:///drive0?socket={0}'.format(nbd_sock_path),
|
|
||||||
sync='full',
|
|
||||||
format='raw', # always raw, the server handles the format
|
|
||||||
mode='existing',
|
|
||||||
job_id='mirror-job0'))
|
|
||||||
|
|
||||||
iotests.log('Waiting for `drive-mirror` to complete...')
|
iotests.log('Starting `drive-mirror` on source...')
|
||||||
iotests.log(source_vm.event_wait('BLOCK_JOB_READY'),
|
iotests.log(source_vm.qmp(
|
||||||
filters=[iotests.filter_qmp_event])
|
'drive-mirror',
|
||||||
|
device='drive0',
|
||||||
|
target='nbd+unix:///drive0?socket={0}'.format(nbd_sock_path),
|
||||||
|
sync='full',
|
||||||
|
format='raw', # always raw, the server handles the format
|
||||||
|
mode='existing',
|
||||||
|
job_id='mirror-job0'))
|
||||||
|
|
||||||
iotests.log('Starting migration...')
|
iotests.log('Waiting for `drive-mirror` to complete...')
|
||||||
source_vm.qmp('migrate-set-capabilities',
|
iotests.log(source_vm.event_wait('BLOCK_JOB_READY'),
|
||||||
capabilities=[{'capability': 'events', 'state': True}])
|
filters=[iotests.filter_qmp_event])
|
||||||
dest_vm.qmp('migrate-set-capabilities',
|
|
||||||
capabilities=[{'capability': 'events', 'state': True}])
|
|
||||||
iotests.log(source_vm.qmp('migrate', uri='unix:{0}'.format(migration_sock_path)))
|
|
||||||
|
|
||||||
while True:
|
iotests.log('Starting migration...')
|
||||||
event1 = source_vm.event_wait('MIGRATION')
|
source_vm.qmp('migrate-set-capabilities',
|
||||||
iotests.log(event1, filters=[iotests.filter_qmp_event])
|
capabilities=[{'capability': 'events', 'state': True}])
|
||||||
if event1['data']['status'] in ('completed', 'failed'):
|
dest_vm.qmp('migrate-set-capabilities',
|
||||||
iotests.log('Gracefully ending the `drive-mirror` job on source...')
|
capabilities=[{'capability': 'events', 'state': True}])
|
||||||
iotests.log(source_vm.qmp('block-job-cancel', device='mirror-job0'))
|
iotests.log(source_vm.qmp('migrate', uri='unix:{0}'.format(migration_sock_path)))
|
||||||
break
|
|
||||||
|
|
||||||
while True:
|
while True:
|
||||||
event2 = source_vm.event_wait('BLOCK_JOB_COMPLETED')
|
event1 = source_vm.event_wait('MIGRATION')
|
||||||
iotests.log(event2, filters=[iotests.filter_qmp_event])
|
iotests.log(event1, filters=[iotests.filter_qmp_event])
|
||||||
if event2['event'] == 'BLOCK_JOB_COMPLETED':
|
if event1['data']['status'] in ('completed', 'failed'):
|
||||||
iotests.log('Stopping the NBD server on destination...')
|
iotests.log('Gracefully ending the `drive-mirror` job on source...')
|
||||||
iotests.log(dest_vm.qmp('nbd-server-stop'))
|
iotests.log(source_vm.qmp('block-job-cancel', device='mirror-job0'))
|
||||||
break
|
break
|
||||||
|
|
||||||
|
while True:
|
||||||
|
event2 = source_vm.event_wait('BLOCK_JOB_COMPLETED')
|
||||||
|
iotests.log(event2, filters=[iotests.filter_qmp_event])
|
||||||
|
if event2['event'] == 'BLOCK_JOB_COMPLETED':
|
||||||
|
iotests.log('Stopping the NBD server on destination...')
|
||||||
|
iotests.log(dest_vm.qmp('nbd-server-stop'))
|
||||||
|
break
|
||||||
|
|
|
@ -160,6 +160,32 @@ class Timeout:
|
||||||
def timeout(self, signum, frame):
|
def timeout(self, signum, frame):
|
||||||
raise Exception(self.errmsg)
|
raise Exception(self.errmsg)
|
||||||
|
|
||||||
|
|
||||||
|
class FilePath(object):
|
||||||
|
'''An auto-generated filename that cleans itself up.
|
||||||
|
|
||||||
|
Use this context manager to generate filenames and ensure that the file
|
||||||
|
gets deleted::
|
||||||
|
|
||||||
|
with TestFilePath('test.img') as img_path:
|
||||||
|
qemu_img('create', img_path, '1G')
|
||||||
|
# migration_sock_path is automatically deleted
|
||||||
|
'''
|
||||||
|
def __init__(self, name):
|
||||||
|
filename = '{0}-{1}'.format(os.getpid(), name)
|
||||||
|
self.path = os.path.join(test_dir, filename)
|
||||||
|
|
||||||
|
def __enter__(self):
|
||||||
|
return self.path
|
||||||
|
|
||||||
|
def __exit__(self, exc_type, exc_val, exc_tb):
|
||||||
|
try:
|
||||||
|
os.remove(self.path)
|
||||||
|
except OSError:
|
||||||
|
pass
|
||||||
|
return False
|
||||||
|
|
||||||
|
|
||||||
class VM(qtest.QEMUQtestMachine):
|
class VM(qtest.QEMUQtestMachine):
|
||||||
'''A QEMU VM'''
|
'''A QEMU VM'''
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue