mirror of
https://github.com/xemu-project/xemu.git
synced 2024-11-24 20:19:44 +00:00
34a5de525a
The further change of moving backup to be a one block-copy call will make copying chunk-size and cluster-size two separate things. So, even with 64k cluster sized qcow2 image, default chunk would be 1M. Test 219 depends on specified chunk-size. Update it for explicit chunk-size for backup as for mirror. Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com> Reviewed-by: Max Reitz <mreitz@redhat.com> Message-Id: <20210116214705.822267-14-vsementsov@virtuozzo.com> Signed-off-by: Max Reitz <mreitz@redhat.com>
236 lines
9.0 KiB
Python
Executable File
236 lines
9.0 KiB
Python
Executable File
#!/usr/bin/env python3
|
|
# group: rw
|
|
#
|
|
# Copyright (C) 2018 Red Hat, Inc.
|
|
#
|
|
# This program is free software; you can redistribute it and/or modify
|
|
# it under the terms of the GNU General Public License as published by
|
|
# the Free Software Foundation; either version 2 of the License, or
|
|
# (at your option) any later version.
|
|
#
|
|
# This program is distributed in the hope that it will be useful,
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
# GNU General Public License for more details.
|
|
#
|
|
# You should have received a copy of the GNU General Public License
|
|
# along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
#
|
|
# Creator/Owner: Kevin Wolf <kwolf@redhat.com>
|
|
#
|
|
# Check using the job-* QMP commands with block jobs
|
|
|
|
import iotests
|
|
|
|
iotests.script_initialize(supported_fmts=['qcow2'])
|
|
|
|
img_size = 4 * 1024 * 1024
|
|
|
|
def pause_wait(vm, job_id):
|
|
with iotests.Timeout(3, "Timeout waiting for job to pause"):
|
|
while True:
|
|
result = vm.qmp('query-jobs')
|
|
for job in result['return']:
|
|
if job['id'] == job_id and job['status'] in ['paused', 'standby']:
|
|
return job
|
|
|
|
# Test that block-job-pause/resume and job-pause/resume can be mixed
|
|
def test_pause_resume(vm):
|
|
for pause_cmd, pause_arg in [('block-job-pause', 'device'),
|
|
('job-pause', 'id')]:
|
|
for resume_cmd, resume_arg in [('block-job-resume', 'device'),
|
|
('job-resume', 'id')]:
|
|
iotests.log('=== Testing %s/%s ===' % (pause_cmd, resume_cmd))
|
|
|
|
iotests.log(vm.qmp(pause_cmd, **{pause_arg: 'job0'}))
|
|
pause_wait(vm, 'job0')
|
|
iotests.log(iotests.filter_qmp_event(vm.event_wait('JOB_STATUS_CHANGE')))
|
|
result = vm.qmp('query-jobs')
|
|
iotests.log(result)
|
|
|
|
old_progress = result['return'][0]['current-progress']
|
|
total_progress = result['return'][0]['total-progress']
|
|
|
|
iotests.log(vm.qmp(resume_cmd, **{resume_arg: 'job0'}))
|
|
iotests.log(iotests.filter_qmp_event(vm.event_wait('JOB_STATUS_CHANGE')))
|
|
if old_progress < total_progress:
|
|
# Wait for the job to advance
|
|
while result['return'][0]['current-progress'] == old_progress:
|
|
result = vm.qmp('query-jobs')
|
|
iotests.log(result)
|
|
else:
|
|
# Already reached the end, so the job cannot advance
|
|
# any further; therefore, the query-jobs result can be
|
|
# logged immediately
|
|
iotests.log(vm.qmp('query-jobs'))
|
|
|
|
def test_job_lifecycle(vm, job, job_args, has_ready=False, is_mirror=False):
|
|
global img_size
|
|
|
|
iotests.log('')
|
|
iotests.log('')
|
|
iotests.log('Starting block job: %s (auto-finalize: %s; auto-dismiss: %s)' %
|
|
(job,
|
|
job_args.get('auto-finalize', True),
|
|
job_args.get('auto-dismiss', True)))
|
|
iotests.log(vm.qmp(job, job_id='job0', **job_args))
|
|
|
|
# Depending on the storage, the first request may or may not have completed
|
|
# yet (and the total progress may not have been fully determined yet), so
|
|
# filter out the progress. Later query-job calls don't need the filtering
|
|
# because the progress is made deterministic by the block job speed
|
|
result = vm.qmp('query-jobs')
|
|
for j in result['return']:
|
|
j['current-progress'] = 'FILTERED'
|
|
j['total-progress'] = 'FILTERED'
|
|
iotests.log(result)
|
|
|
|
# undefined -> created -> running
|
|
iotests.log(iotests.filter_qmp_event(vm.event_wait('JOB_STATUS_CHANGE')))
|
|
iotests.log(iotests.filter_qmp_event(vm.event_wait('JOB_STATUS_CHANGE')))
|
|
|
|
# Wait for total-progress to stabilize
|
|
while vm.qmp('query-jobs')['return'][0]['total-progress'] < img_size:
|
|
pass
|
|
|
|
# RUNNING state:
|
|
# pause/resume should work, complete/finalize/dismiss should error out
|
|
iotests.log('')
|
|
iotests.log('Pause/resume in RUNNING')
|
|
test_pause_resume(vm)
|
|
|
|
iotests.log(vm.qmp('job-complete', id='job0'))
|
|
iotests.log(vm.qmp('job-finalize', id='job0'))
|
|
iotests.log(vm.qmp('job-dismiss', id='job0'))
|
|
|
|
iotests.log(vm.qmp('block-job-complete', device='job0'))
|
|
iotests.log(vm.qmp('block-job-finalize', id='job0'))
|
|
iotests.log(vm.qmp('block-job-dismiss', id='job0'))
|
|
|
|
# Let the job complete (or transition to READY if it supports that)
|
|
iotests.log(vm.qmp('block-job-set-speed', device='job0', speed=0))
|
|
if has_ready:
|
|
iotests.log('')
|
|
iotests.log('Waiting for READY state...')
|
|
vm.event_wait('BLOCK_JOB_READY')
|
|
iotests.log(iotests.filter_qmp_event(vm.event_wait('JOB_STATUS_CHANGE')))
|
|
iotests.log(vm.qmp('query-jobs'))
|
|
|
|
# READY state:
|
|
# pause/resume/complete should work, finalize/dismiss should error out
|
|
iotests.log('')
|
|
iotests.log('Pause/resume in READY')
|
|
test_pause_resume(vm)
|
|
|
|
iotests.log(vm.qmp('job-finalize', id='job0'))
|
|
iotests.log(vm.qmp('job-dismiss', id='job0'))
|
|
|
|
iotests.log(vm.qmp('block-job-finalize', id='job0'))
|
|
iotests.log(vm.qmp('block-job-dismiss', id='job0'))
|
|
|
|
# Transition to WAITING
|
|
iotests.log(vm.qmp('job-complete', id='job0'))
|
|
|
|
# Move to WAITING and PENDING state
|
|
iotests.log('')
|
|
iotests.log('Waiting for PENDING state...')
|
|
iotests.log(iotests.filter_qmp_event(vm.event_wait('JOB_STATUS_CHANGE')))
|
|
iotests.log(iotests.filter_qmp_event(vm.event_wait('JOB_STATUS_CHANGE')))
|
|
if is_mirror:
|
|
iotests.log(iotests.filter_qmp_event(vm.event_wait('JOB_STATUS_CHANGE')))
|
|
iotests.log(iotests.filter_qmp_event(vm.event_wait('JOB_STATUS_CHANGE')))
|
|
|
|
if not job_args.get('auto-finalize', True):
|
|
# PENDING state:
|
|
# finalize should work, pause/complete/dismiss should error out
|
|
iotests.log(vm.qmp('query-jobs'))
|
|
|
|
iotests.log(vm.qmp('job-pause', id='job0'))
|
|
iotests.log(vm.qmp('job-complete', id='job0'))
|
|
iotests.log(vm.qmp('job-dismiss', id='job0'))
|
|
|
|
iotests.log(vm.qmp('block-job-pause', device='job0'))
|
|
iotests.log(vm.qmp('block-job-complete', device='job0'))
|
|
iotests.log(vm.qmp('block-job-dismiss', id='job0'))
|
|
|
|
# Transition to CONCLUDED
|
|
iotests.log(vm.qmp('job-finalize', id='job0'))
|
|
|
|
|
|
# Move to CONCLUDED state
|
|
iotests.log(iotests.filter_qmp_event(vm.event_wait('JOB_STATUS_CHANGE')))
|
|
|
|
if not job_args.get('auto-dismiss', True):
|
|
# CONCLUDED state:
|
|
# dismiss should work, pause/complete/finalize should error out
|
|
iotests.log(vm.qmp('query-jobs'))
|
|
|
|
iotests.log(vm.qmp('job-pause', id='job0'))
|
|
iotests.log(vm.qmp('job-complete', id='job0'))
|
|
iotests.log(vm.qmp('job-finalize', id='job0'))
|
|
|
|
iotests.log(vm.qmp('block-job-pause', device='job0'))
|
|
iotests.log(vm.qmp('block-job-complete', device='job0'))
|
|
iotests.log(vm.qmp('block-job-finalize', id='job0'))
|
|
|
|
# Transition to NULL
|
|
iotests.log(vm.qmp('job-dismiss', id='job0'))
|
|
|
|
# Move to NULL state
|
|
iotests.log(iotests.filter_qmp_event(vm.event_wait('JOB_STATUS_CHANGE')))
|
|
iotests.log(vm.qmp('query-jobs'))
|
|
|
|
|
|
with iotests.FilePath('disk.img') as disk_path, \
|
|
iotests.FilePath('copy.img') as copy_path, \
|
|
iotests.VM() as vm:
|
|
|
|
iotests.qemu_img_create('-f', iotests.imgfmt, disk_path, str(img_size))
|
|
iotests.qemu_io('-c', 'write 0 %i' % (img_size),
|
|
'-f', iotests.imgfmt, disk_path)
|
|
|
|
iotests.log('Launching VM...')
|
|
vm.add_blockdev(vm.qmp_to_opts({
|
|
'driver': iotests.imgfmt,
|
|
'node-name': 'drive0-node',
|
|
'file': {
|
|
'driver': 'file',
|
|
'filename': disk_path,
|
|
},
|
|
}))
|
|
vm.launch()
|
|
|
|
# In order to keep things deterministic (especially progress in query-job,
|
|
# but related to this also automatic state transitions like job
|
|
# completion), but still get pause points often enough to avoid making this
|
|
# test very slow, it's important to have the right ratio between speed and
|
|
# copy-chunk-size.
|
|
#
|
|
# Chose 64k copy-chunk-size both for mirror (by buf_size) and backup (by
|
|
# x-max-chunk). The slice time, i.e. the granularity of the rate limiting
|
|
# is 100ms. With a speed of 256k per second, we can get four pause points
|
|
# per second. This gives us 250ms per iteration, which should be enough to
|
|
# stay deterministic.
|
|
|
|
test_job_lifecycle(vm, 'drive-mirror', has_ready=True, job_args={
|
|
'device': 'drive0-node',
|
|
'target': copy_path,
|
|
'sync': 'full',
|
|
'speed': 262144,
|
|
'buf_size': 65536,
|
|
})
|
|
|
|
for auto_finalize in [True, False]:
|
|
for auto_dismiss in [True, False]:
|
|
test_job_lifecycle(vm, 'drive-backup', is_mirror=True, job_args={
|
|
'device': 'drive0-node',
|
|
'target': copy_path,
|
|
'sync': 'full',
|
|
'speed': 262144,
|
|
'x-perf': {'max-chunk': 65536},
|
|
'auto-finalize': auto_finalize,
|
|
'auto-dismiss': auto_dismiss,
|
|
})
|
|
|
|
vm.shutdown()
|