64ed6f92ff
The binaries move to the root directory, e.g. qemu-system-i386 or qemu-arm. This requires changes to qtests, CI, etc. Signed-off-by: Marc-André Lureau <marcandre.lureau@redhat.com> Signed-off-by: Paolo Bonzini <pbonzini@redhat.com>
1103 lines
37 KiB
ReStructuredText
1103 lines
37 KiB
ReStructuredText
..
|
|
Copyright (C) 2017 Red Hat Inc.
|
|
|
|
This work is licensed under the terms of the GNU GPL, version 2 or
|
|
later. See the COPYING file in the top-level directory.
|
|
|
|
============================
|
|
Live Block Device Operations
|
|
============================
|
|
|
|
QEMU Block Layer currently (as of QEMU 2.9) supports four major kinds of
|
|
live block device jobs -- stream, commit, mirror, and backup. These can
|
|
be used to manipulate disk image chains to accomplish certain tasks,
|
|
namely: live copy data from backing files into overlays; shorten long
|
|
disk image chains by merging data from overlays into backing files; live
|
|
synchronize data from a disk image chain (including current active disk)
|
|
to another target image; and point-in-time (and incremental) backups of
|
|
a block device. Below is a description of the said block (QMP)
|
|
primitives, and some (non-exhaustive list of) examples to illustrate
|
|
their use.
|
|
|
|
.. note::
|
|
The file ``qapi/block-core.json`` in the QEMU source tree has the
|
|
canonical QEMU API (QAPI) schema documentation for the QMP
|
|
primitives discussed here.
|
|
|
|
.. todo (kashyapc):: Remove the ".. contents::" directive when Sphinx is
|
|
integrated.
|
|
|
|
.. contents::
|
|
|
|
Disk image backing chain notation
|
|
---------------------------------
|
|
|
|
A simple disk image chain. (This can be created live using QMP
|
|
``blockdev-snapshot-sync``, or offline via ``qemu-img``)::
|
|
|
|
(Live QEMU)
|
|
|
|
|
.
|
|
V
|
|
|
|
[A] <----- [B]
|
|
|
|
(backing file) (overlay)
|
|
|
|
The arrow can be read as: Image [A] is the backing file of disk image
|
|
[B]. And live QEMU is currently writing to image [B], consequently, it
|
|
is also referred to as the "active layer".
|
|
|
|
There are two kinds of terminology that are common when referring to
|
|
files in a disk image backing chain:
|
|
|
|
(1) Directional: 'base' and 'top'. Given the simple disk image chain
|
|
above, image [A] can be referred to as 'base', and image [B] as
|
|
'top'. (This terminology can be seen in in QAPI schema file,
|
|
block-core.json.)
|
|
|
|
(2) Relational: 'backing file' and 'overlay'. Again, taking the same
|
|
simple disk image chain from the above, disk image [A] is referred
|
|
to as the backing file, and image [B] as overlay.
|
|
|
|
Throughout this document, we will use the relational terminology.
|
|
|
|
.. important::
|
|
The overlay files can generally be any format that supports a
|
|
backing file, although QCOW2 is the preferred format and the one
|
|
used in this document.
|
|
|
|
|
|
Brief overview of live block QMP primitives
|
|
-------------------------------------------
|
|
|
|
The following are the four different kinds of live block operations that
|
|
QEMU block layer supports.
|
|
|
|
(1) ``block-stream``: Live copy of data from backing files into overlay
|
|
files.
|
|
|
|
.. note:: Once the 'stream' operation has finished, three things to
|
|
note:
|
|
|
|
(a) QEMU rewrites the backing chain to remove
|
|
reference to the now-streamed and redundant backing
|
|
file;
|
|
|
|
(b) the streamed file *itself* won't be removed by QEMU,
|
|
and must be explicitly discarded by the user;
|
|
|
|
(c) the streamed file remains valid -- i.e. further
|
|
overlays can be created based on it. Refer the
|
|
``block-stream`` section further below for more
|
|
details.
|
|
|
|
(2) ``block-commit``: Live merge of data from overlay files into backing
|
|
files (with the optional goal of removing the overlay file from the
|
|
chain). Since QEMU 2.0, this includes "active ``block-commit``"
|
|
(i.e. merge the current active layer into the base image).
|
|
|
|
.. note:: Once the 'commit' operation has finished, there are three
|
|
things to note here as well:
|
|
|
|
(a) QEMU rewrites the backing chain to remove reference
|
|
to now-redundant overlay images that have been
|
|
committed into a backing file;
|
|
|
|
(b) the committed file *itself* won't be removed by QEMU
|
|
-- it ought to be manually removed;
|
|
|
|
(c) however, unlike in the case of ``block-stream``, the
|
|
intermediate images will be rendered invalid -- i.e.
|
|
no more further overlays can be created based on
|
|
them. Refer the ``block-commit`` section further
|
|
below for more details.
|
|
|
|
(3) ``drive-mirror`` (and ``blockdev-mirror``): Synchronize a running
|
|
disk to another image.
|
|
|
|
(4) ``drive-backup`` (and ``blockdev-backup``): Point-in-time (live) copy
|
|
of a block device to a destination.
|
|
|
|
|
|
.. _`Interacting with a QEMU instance`:
|
|
|
|
Interacting with a QEMU instance
|
|
--------------------------------
|
|
|
|
To show some example invocations of command-line, we will use the
|
|
following invocation of QEMU, with a QMP server running over UNIX
|
|
socket::
|
|
|
|
$ ./qemu-system-x86_64 -display none -no-user-config \
|
|
-M q35 -nodefaults -m 512 \
|
|
-blockdev node-name=node-A,driver=qcow2,file.driver=file,file.node-name=file,file.filename=./a.qcow2 \
|
|
-device virtio-blk,drive=node-A,id=virtio0 \
|
|
-monitor stdio -qmp unix:/tmp/qmp-sock,server,nowait
|
|
|
|
The ``-blockdev`` command-line option, used above, is available from
|
|
QEMU 2.9 onwards. In the above invocation, notice the ``node-name``
|
|
parameter that is used to refer to the disk image a.qcow2 ('node-A') --
|
|
this is a cleaner way to refer to a disk image (as opposed to referring
|
|
to it by spelling out file paths). So, we will continue to designate a
|
|
``node-name`` to each further disk image created (either via
|
|
``blockdev-snapshot-sync``, or ``blockdev-add``) as part of the disk
|
|
image chain, and continue to refer to the disks using their
|
|
``node-name`` (where possible, because ``block-commit`` does not yet, as
|
|
of QEMU 2.9, accept ``node-name`` parameter) when performing various
|
|
block operations.
|
|
|
|
To interact with the QEMU instance launched above, we will use the
|
|
``qmp-shell`` utility (located at: ``qemu/scripts/qmp``, as part of the
|
|
QEMU source directory), which takes key-value pairs for QMP commands.
|
|
Invoke it as below (which will also print out the complete raw JSON
|
|
syntax for reference -- examples in the following sections)::
|
|
|
|
$ ./qmp-shell -v -p /tmp/qmp-sock
|
|
(QEMU)
|
|
|
|
.. note::
|
|
In the event we have to repeat a certain QMP command, we will: for
|
|
the first occurrence of it, show the ``qmp-shell`` invocation, *and*
|
|
the corresponding raw JSON QMP syntax; but for subsequent
|
|
invocations, present just the ``qmp-shell`` syntax, and omit the
|
|
equivalent JSON output.
|
|
|
|
|
|
Example disk image chain
|
|
------------------------
|
|
|
|
We will use the below disk image chain (and occasionally spelling it
|
|
out where appropriate) when discussing various primitives::
|
|
|
|
[A] <-- [B] <-- [C] <-- [D]
|
|
|
|
Where [A] is the original base image; [B] and [C] are intermediate
|
|
overlay images; image [D] is the active layer -- i.e. live QEMU is
|
|
writing to it. (The rule of thumb is: live QEMU will always be pointing
|
|
to the rightmost image in a disk image chain.)
|
|
|
|
The above image chain can be created by invoking
|
|
``blockdev-snapshot-sync`` commands as following (which shows the
|
|
creation of overlay image [B]) using the ``qmp-shell`` (our invocation
|
|
also prints the raw JSON invocation of it)::
|
|
|
|
(QEMU) blockdev-snapshot-sync node-name=node-A snapshot-file=b.qcow2 snapshot-node-name=node-B format=qcow2
|
|
{
|
|
"execute": "blockdev-snapshot-sync",
|
|
"arguments": {
|
|
"node-name": "node-A",
|
|
"snapshot-file": "b.qcow2",
|
|
"format": "qcow2",
|
|
"snapshot-node-name": "node-B"
|
|
}
|
|
}
|
|
|
|
Here, "node-A" is the name QEMU internally uses to refer to the base
|
|
image [A] -- it is the backing file, based on which the overlay image,
|
|
[B], is created.
|
|
|
|
To create the rest of the overlay images, [C], and [D] (omitting the raw
|
|
JSON output for brevity)::
|
|
|
|
(QEMU) blockdev-snapshot-sync node-name=node-B snapshot-file=c.qcow2 snapshot-node-name=node-C format=qcow2
|
|
(QEMU) blockdev-snapshot-sync node-name=node-C snapshot-file=d.qcow2 snapshot-node-name=node-D format=qcow2
|
|
|
|
|
|
A note on points-in-time vs file names
|
|
--------------------------------------
|
|
|
|
In our disk image chain::
|
|
|
|
[A] <-- [B] <-- [C] <-- [D]
|
|
|
|
We have *three* points in time and an active layer:
|
|
|
|
- Point 1: Guest state when [B] was created is contained in file [A]
|
|
- Point 2: Guest state when [C] was created is contained in [A] + [B]
|
|
- Point 3: Guest state when [D] was created is contained in
|
|
[A] + [B] + [C]
|
|
- Active layer: Current guest state is contained in [A] + [B] + [C] +
|
|
[D]
|
|
|
|
Therefore, be aware with naming choices:
|
|
|
|
- Naming a file after the time it is created is misleading -- the
|
|
guest data for that point in time is *not* contained in that file
|
|
(as explained earlier)
|
|
- Rather, think of files as a *delta* from the backing file
|
|
|
|
|
|
Live block streaming --- ``block-stream``
|
|
-----------------------------------------
|
|
|
|
The ``block-stream`` command allows you to do live copy data from backing
|
|
files into overlay images.
|
|
|
|
Given our original example disk image chain from earlier::
|
|
|
|
[A] <-- [B] <-- [C] <-- [D]
|
|
|
|
The disk image chain can be shortened in one of the following different
|
|
ways (not an exhaustive list).
|
|
|
|
.. _`Case-1`:
|
|
|
|
(1) Merge everything into the active layer: I.e. copy all contents from
|
|
the base image, [A], and overlay images, [B] and [C], into [D],
|
|
*while* the guest is running. The resulting chain will be a
|
|
standalone image, [D] -- with contents from [A], [B] and [C] merged
|
|
into it (where live QEMU writes go to)::
|
|
|
|
[D]
|
|
|
|
.. _`Case-2`:
|
|
|
|
(2) Taking the same example disk image chain mentioned earlier, merge
|
|
only images [B] and [C] into [D], the active layer. The result will
|
|
be contents of images [B] and [C] will be copied into [D], and the
|
|
backing file pointer of image [D] will be adjusted to point to image
|
|
[A]. The resulting chain will be::
|
|
|
|
[A] <-- [D]
|
|
|
|
.. _`Case-3`:
|
|
|
|
(3) Intermediate streaming (available since QEMU 2.8): Starting afresh
|
|
with the original example disk image chain, with a total of four
|
|
images, it is possible to copy contents from image [B] into image
|
|
[C]. Once the copy is finished, image [B] can now be (optionally)
|
|
discarded; and the backing file pointer of image [C] will be
|
|
adjusted to point to [A]. I.e. after performing "intermediate
|
|
streaming" of [B] into [C], the resulting image chain will be (where
|
|
live QEMU is writing to [D])::
|
|
|
|
[A] <-- [C] <-- [D]
|
|
|
|
|
|
QMP invocation for ``block-stream``
|
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|
|
|
For `Case-1`_, to merge contents of all the backing files into the
|
|
active layer, where 'node-D' is the current active image (by default
|
|
``block-stream`` will flatten the entire chain); ``qmp-shell`` (and its
|
|
corresponding JSON output)::
|
|
|
|
(QEMU) block-stream device=node-D job-id=job0
|
|
{
|
|
"execute": "block-stream",
|
|
"arguments": {
|
|
"device": "node-D",
|
|
"job-id": "job0"
|
|
}
|
|
}
|
|
|
|
For `Case-2`_, merge contents of the images [B] and [C] into [D], where
|
|
image [D] ends up referring to image [A] as its backing file::
|
|
|
|
(QEMU) block-stream device=node-D base-node=node-A job-id=job0
|
|
|
|
And for `Case-3`_, of "intermediate" streaming", merge contents of
|
|
images [B] into [C], where [C] ends up referring to [A] as its backing
|
|
image::
|
|
|
|
(QEMU) block-stream device=node-C base-node=node-A job-id=job0
|
|
|
|
Progress of a ``block-stream`` operation can be monitored via the QMP
|
|
command::
|
|
|
|
(QEMU) query-block-jobs
|
|
{
|
|
"execute": "query-block-jobs",
|
|
"arguments": {}
|
|
}
|
|
|
|
|
|
Once the ``block-stream`` operation has completed, QEMU will emit an
|
|
event, ``BLOCK_JOB_COMPLETED``. The intermediate overlays remain valid,
|
|
and can now be (optionally) discarded, or retained to create further
|
|
overlays based on them. Finally, the ``block-stream`` jobs can be
|
|
restarted at anytime.
|
|
|
|
|
|
Live block commit --- ``block-commit``
|
|
--------------------------------------
|
|
|
|
The ``block-commit`` command lets you merge live data from overlay
|
|
images into backing file(s). Since QEMU 2.0, this includes "live active
|
|
commit" (i.e. it is possible to merge the "active layer", the right-most
|
|
image in a disk image chain where live QEMU will be writing to, into the
|
|
base image). This is analogous to ``block-stream``, but in the opposite
|
|
direction.
|
|
|
|
Again, starting afresh with our example disk image chain, where live
|
|
QEMU is writing to the right-most image in the chain, [D]::
|
|
|
|
[A] <-- [B] <-- [C] <-- [D]
|
|
|
|
The disk image chain can be shortened in one of the following ways:
|
|
|
|
.. _`block-commit_Case-1`:
|
|
|
|
(1) Commit content from only image [B] into image [A]. The resulting
|
|
chain is the following, where image [C] is adjusted to point at [A]
|
|
as its new backing file::
|
|
|
|
[A] <-- [C] <-- [D]
|
|
|
|
(2) Commit content from images [B] and [C] into image [A]. The
|
|
resulting chain, where image [D] is adjusted to point to image [A]
|
|
as its new backing file::
|
|
|
|
[A] <-- [D]
|
|
|
|
.. _`block-commit_Case-3`:
|
|
|
|
(3) Commit content from images [B], [C], and the active layer [D] into
|
|
image [A]. The resulting chain (in this case, a consolidated single
|
|
image)::
|
|
|
|
[A]
|
|
|
|
(4) Commit content from image only image [C] into image [B]. The
|
|
resulting chain::
|
|
|
|
[A] <-- [B] <-- [D]
|
|
|
|
(5) Commit content from image [C] and the active layer [D] into image
|
|
[B]. The resulting chain::
|
|
|
|
[A] <-- [B]
|
|
|
|
|
|
QMP invocation for ``block-commit``
|
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|
|
|
For :ref:`Case-1 <block-commit_Case-1>`, to merge contents only from
|
|
image [B] into image [A], the invocation is as follows::
|
|
|
|
(QEMU) block-commit device=node-D base=a.qcow2 top=b.qcow2 job-id=job0
|
|
{
|
|
"execute": "block-commit",
|
|
"arguments": {
|
|
"device": "node-D",
|
|
"job-id": "job0",
|
|
"top": "b.qcow2",
|
|
"base": "a.qcow2"
|
|
}
|
|
}
|
|
|
|
Once the above ``block-commit`` operation has completed, a
|
|
``BLOCK_JOB_COMPLETED`` event will be issued, and no further action is
|
|
required. As the end result, the backing file of image [C] is adjusted
|
|
to point to image [A], and the original 4-image chain will end up being
|
|
transformed to::
|
|
|
|
[A] <-- [C] <-- [D]
|
|
|
|
.. note::
|
|
The intermediate image [B] is invalid (as in: no more further
|
|
overlays based on it can be created).
|
|
|
|
Reasoning: An intermediate image after a 'stream' operation still
|
|
represents that old point-in-time, and may be valid in that context.
|
|
However, an intermediate image after a 'commit' operation no longer
|
|
represents any point-in-time, and is invalid in any context.
|
|
|
|
|
|
However, :ref:`Case-3 <block-commit_Case-3>` (also called: "active
|
|
``block-commit``") is a *two-phase* operation: In the first phase, the
|
|
content from the active overlay, along with the intermediate overlays,
|
|
is copied into the backing file (also called the base image). In the
|
|
second phase, adjust the said backing file as the current active image
|
|
-- possible via issuing the command ``block-job-complete``. Optionally,
|
|
the ``block-commit`` operation can be cancelled by issuing the command
|
|
``block-job-cancel``, but be careful when doing this.
|
|
|
|
Once the ``block-commit`` operation has completed, the event
|
|
``BLOCK_JOB_READY`` will be emitted, signalling that the synchronization
|
|
has finished. Now the job can be gracefully completed by issuing the
|
|
command ``block-job-complete`` -- until such a command is issued, the
|
|
'commit' operation remains active.
|
|
|
|
The following is the flow for :ref:`Case-3 <block-commit_Case-3>` to
|
|
convert a disk image chain such as this::
|
|
|
|
[A] <-- [B] <-- [C] <-- [D]
|
|
|
|
Into::
|
|
|
|
[A]
|
|
|
|
Where content from all the subsequent overlays, [B], and [C], including
|
|
the active layer, [D], is committed back to [A] -- which is where live
|
|
QEMU is performing all its current writes).
|
|
|
|
Start the "active ``block-commit``" operation::
|
|
|
|
(QEMU) block-commit device=node-D base=a.qcow2 top=d.qcow2 job-id=job0
|
|
{
|
|
"execute": "block-commit",
|
|
"arguments": {
|
|
"device": "node-D",
|
|
"job-id": "job0",
|
|
"top": "d.qcow2",
|
|
"base": "a.qcow2"
|
|
}
|
|
}
|
|
|
|
|
|
Once the synchronization has completed, the event ``BLOCK_JOB_READY`` will
|
|
be emitted.
|
|
|
|
Then, optionally query for the status of the active block operations.
|
|
We can see the 'commit' job is now ready to be completed, as indicated
|
|
by the line *"ready": true*::
|
|
|
|
(QEMU) query-block-jobs
|
|
{
|
|
"execute": "query-block-jobs",
|
|
"arguments": {}
|
|
}
|
|
{
|
|
"return": [
|
|
{
|
|
"busy": false,
|
|
"type": "commit",
|
|
"len": 1376256,
|
|
"paused": false,
|
|
"ready": true,
|
|
"io-status": "ok",
|
|
"offset": 1376256,
|
|
"device": "job0",
|
|
"speed": 0
|
|
}
|
|
]
|
|
}
|
|
|
|
Gracefully complete the 'commit' block device job::
|
|
|
|
(QEMU) block-job-complete device=job0
|
|
{
|
|
"execute": "block-job-complete",
|
|
"arguments": {
|
|
"device": "job0"
|
|
}
|
|
}
|
|
{
|
|
"return": {}
|
|
}
|
|
|
|
Finally, once the above job is completed, an event
|
|
``BLOCK_JOB_COMPLETED`` will be emitted.
|
|
|
|
.. note::
|
|
The invocation for rest of the cases (2, 4, and 5), discussed in the
|
|
previous section, is omitted for brevity.
|
|
|
|
|
|
Live disk synchronization --- ``drive-mirror`` and ``blockdev-mirror``
|
|
----------------------------------------------------------------------
|
|
|
|
Synchronize a running disk image chain (all or part of it) to a target
|
|
image.
|
|
|
|
Again, given our familiar disk image chain::
|
|
|
|
[A] <-- [B] <-- [C] <-- [D]
|
|
|
|
The ``drive-mirror`` (and its newer equivalent ``blockdev-mirror``)
|
|
allows you to copy data from the entire chain into a single target image
|
|
(which can be located on a different host), [E].
|
|
|
|
.. note::
|
|
|
|
When you cancel an in-progress 'mirror' job *before* the source and
|
|
target are synchronized, ``block-job-cancel`` will emit the event
|
|
``BLOCK_JOB_CANCELLED``. However, note that if you cancel a
|
|
'mirror' job *after* it has indicated (via the event
|
|
``BLOCK_JOB_READY``) that the source and target have reached
|
|
synchronization, then the event emitted by ``block-job-cancel``
|
|
changes to ``BLOCK_JOB_COMPLETED``.
|
|
|
|
Besides the 'mirror' job, the "active ``block-commit``" is the only
|
|
other block device job that emits the event ``BLOCK_JOB_READY``.
|
|
The rest of the block device jobs ('stream', "non-active
|
|
``block-commit``", and 'backup') end automatically.
|
|
|
|
So there are two possible actions to take, after a 'mirror' job has
|
|
emitted the event ``BLOCK_JOB_READY``, indicating that the source and
|
|
target have reached synchronization:
|
|
|
|
(1) Issuing the command ``block-job-cancel`` (after it emits the event
|
|
``BLOCK_JOB_COMPLETED``) will create a point-in-time (which is at
|
|
the time of *triggering* the cancel command) copy of the entire disk
|
|
image chain (or only the top-most image, depending on the ``sync``
|
|
mode), contained in the target image [E]. One use case for this is
|
|
live VM migration with non-shared storage.
|
|
|
|
(2) Issuing the command ``block-job-complete`` (after it emits the event
|
|
``BLOCK_JOB_COMPLETED``) will adjust the guest device (i.e. live
|
|
QEMU) to point to the target image, [E], causing all the new writes
|
|
from this point on to happen there.
|
|
|
|
About synchronization modes: The synchronization mode determines
|
|
*which* part of the disk image chain will be copied to the target.
|
|
Currently, there are four different kinds:
|
|
|
|
(1) ``full`` -- Synchronize the content of entire disk image chain to
|
|
the target
|
|
|
|
(2) ``top`` -- Synchronize only the contents of the top-most disk image
|
|
in the chain to the target
|
|
|
|
(3) ``none`` -- Synchronize only the new writes from this point on.
|
|
|
|
.. note:: In the case of ``drive-backup`` (or ``blockdev-backup``),
|
|
the behavior of ``none`` synchronization mode is different.
|
|
Normally, a ``backup`` job consists of two parts: Anything
|
|
that is overwritten by the guest is first copied out to
|
|
the backup, and in the background the whole image is
|
|
copied from start to end. With ``sync=none``, it's only
|
|
the first part.
|
|
|
|
(4) ``incremental`` -- Synchronize content that is described by the
|
|
dirty bitmap
|
|
|
|
.. note::
|
|
Refer to the :doc:`bitmaps` document in the QEMU source
|
|
tree to learn about the detailed workings of the ``incremental``
|
|
synchronization mode.
|
|
|
|
|
|
QMP invocation for ``drive-mirror``
|
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|
|
|
To copy the contents of the entire disk image chain, from [A] all the
|
|
way to [D], to a new target (``drive-mirror`` will create the destination
|
|
file, if it doesn't already exist), call it [E]::
|
|
|
|
(QEMU) drive-mirror device=node-D target=e.qcow2 sync=full job-id=job0
|
|
{
|
|
"execute": "drive-mirror",
|
|
"arguments": {
|
|
"device": "node-D",
|
|
"job-id": "job0",
|
|
"target": "e.qcow2",
|
|
"sync": "full"
|
|
}
|
|
}
|
|
|
|
The ``"sync": "full"``, from the above, means: copy the *entire* chain
|
|
to the destination.
|
|
|
|
Following the above, querying for active block jobs will show that a
|
|
'mirror' job is "ready" to be completed (and QEMU will also emit an
|
|
event, ``BLOCK_JOB_READY``)::
|
|
|
|
(QEMU) query-block-jobs
|
|
{
|
|
"execute": "query-block-jobs",
|
|
"arguments": {}
|
|
}
|
|
{
|
|
"return": [
|
|
{
|
|
"busy": false,
|
|
"type": "mirror",
|
|
"len": 21757952,
|
|
"paused": false,
|
|
"ready": true,
|
|
"io-status": "ok",
|
|
"offset": 21757952,
|
|
"device": "job0",
|
|
"speed": 0
|
|
}
|
|
]
|
|
}
|
|
|
|
And, as noted in the previous section, there are two possible actions
|
|
at this point:
|
|
|
|
(a) Create a point-in-time snapshot by ending the synchronization. The
|
|
point-in-time is at the time of *ending* the sync. (The result of
|
|
the following being: the target image, [E], will be populated with
|
|
content from the entire chain, [A] to [D])::
|
|
|
|
(QEMU) block-job-cancel device=job0
|
|
{
|
|
"execute": "block-job-cancel",
|
|
"arguments": {
|
|
"device": "job0"
|
|
}
|
|
}
|
|
|
|
(b) Or, complete the operation and pivot the live QEMU to the target
|
|
copy::
|
|
|
|
(QEMU) block-job-complete device=job0
|
|
|
|
In either of the above cases, if you once again run the
|
|
`query-block-jobs` command, there should not be any active block
|
|
operation.
|
|
|
|
Comparing 'commit' and 'mirror': In both then cases, the overlay images
|
|
can be discarded. However, with 'commit', the *existing* base image
|
|
will be modified (by updating it with contents from overlays); while in
|
|
the case of 'mirror', a *new* target image is populated with the data
|
|
from the disk image chain.
|
|
|
|
|
|
QMP invocation for live storage migration with ``drive-mirror`` + NBD
|
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|
|
|
Live storage migration (without shared storage setup) is one of the most
|
|
common use-cases that takes advantage of the ``drive-mirror`` primitive
|
|
and QEMU's built-in Network Block Device (NBD) server. Here's a quick
|
|
walk-through of this setup.
|
|
|
|
Given the disk image chain::
|
|
|
|
[A] <-- [B] <-- [C] <-- [D]
|
|
|
|
Instead of copying content from the entire chain, synchronize *only* the
|
|
contents of the *top*-most disk image (i.e. the active layer), [D], to a
|
|
target, say, [TargetDisk].
|
|
|
|
.. important::
|
|
The destination host must already have the contents of the backing
|
|
chain, involving images [A], [B], and [C], visible via other means
|
|
-- whether by ``cp``, ``rsync``, or by some storage array-specific
|
|
command.)
|
|
|
|
Sometimes, this is also referred to as "shallow copy" -- because only
|
|
the "active layer", and not the rest of the image chain, is copied to
|
|
the destination.
|
|
|
|
.. note::
|
|
In this example, for the sake of simplicity, we'll be using the same
|
|
``localhost`` as both source and destination.
|
|
|
|
As noted earlier, on the destination host the contents of the backing
|
|
chain -- from images [A] to [C] -- are already expected to exist in some
|
|
form (e.g. in a file called, ``Contents-of-A-B-C.qcow2``). Now, on the
|
|
destination host, let's create a target overlay image (with the image
|
|
``Contents-of-A-B-C.qcow2`` as its backing file), to which the contents
|
|
of image [D] (from the source QEMU) will be mirrored to::
|
|
|
|
$ qemu-img create -f qcow2 -b ./Contents-of-A-B-C.qcow2 \
|
|
-F qcow2 ./target-disk.qcow2
|
|
|
|
And start the destination QEMU (we already have the source QEMU running
|
|
-- discussed in the section: `Interacting with a QEMU instance`_)
|
|
instance, with the following invocation. (As noted earlier, for
|
|
simplicity's sake, the destination QEMU is started on the same host, but
|
|
it could be located elsewhere)::
|
|
|
|
$ ./qemu-system-x86_64 -display none -no-user-config \
|
|
-M q35 -nodefaults -m 512 \
|
|
-blockdev node-name=node-TargetDisk,driver=qcow2,file.driver=file,file.node-name=file,file.filename=./target-disk.qcow2 \
|
|
-device virtio-blk,drive=node-TargetDisk,id=virtio0 \
|
|
-S -monitor stdio -qmp unix:./qmp-sock2,server,nowait \
|
|
-incoming tcp:localhost:6666
|
|
|
|
Given the disk image chain on source QEMU::
|
|
|
|
[A] <-- [B] <-- [C] <-- [D]
|
|
|
|
On the destination host, it is expected that the contents of the chain
|
|
``[A] <-- [B] <-- [C]`` are *already* present, and therefore copy *only*
|
|
the content of image [D].
|
|
|
|
(1) [On *destination* QEMU] As part of the first step, start the
|
|
built-in NBD server on a given host (local host, represented by
|
|
``::``)and port::
|
|
|
|
(QEMU) nbd-server-start addr={"type":"inet","data":{"host":"::","port":"49153"}}
|
|
{
|
|
"execute": "nbd-server-start",
|
|
"arguments": {
|
|
"addr": {
|
|
"data": {
|
|
"host": "::",
|
|
"port": "49153"
|
|
},
|
|
"type": "inet"
|
|
}
|
|
}
|
|
}
|
|
|
|
(2) [On *destination* QEMU] And export the destination disk image using
|
|
QEMU's built-in NBD server::
|
|
|
|
(QEMU) nbd-server-add device=node-TargetDisk writable=true
|
|
{
|
|
"execute": "nbd-server-add",
|
|
"arguments": {
|
|
"device": "node-TargetDisk"
|
|
}
|
|
}
|
|
|
|
(3) [On *source* QEMU] Then, invoke ``drive-mirror`` (NB: since we're
|
|
running ``drive-mirror`` with ``mode=existing`` (meaning:
|
|
synchronize to a pre-created file, therefore 'existing', file on the
|
|
target host), with the synchronization mode as 'top' (``"sync:
|
|
"top"``)::
|
|
|
|
(QEMU) drive-mirror device=node-D target=nbd:localhost:49153:exportname=node-TargetDisk sync=top mode=existing job-id=job0
|
|
{
|
|
"execute": "drive-mirror",
|
|
"arguments": {
|
|
"device": "node-D",
|
|
"mode": "existing",
|
|
"job-id": "job0",
|
|
"target": "nbd:localhost:49153:exportname=node-TargetDisk",
|
|
"sync": "top"
|
|
}
|
|
}
|
|
|
|
(4) [On *source* QEMU] Once ``drive-mirror`` copies the entire data, and the
|
|
event ``BLOCK_JOB_READY`` is emitted, issue ``block-job-cancel`` to
|
|
gracefully end the synchronization, from source QEMU::
|
|
|
|
(QEMU) block-job-cancel device=job0
|
|
{
|
|
"execute": "block-job-cancel",
|
|
"arguments": {
|
|
"device": "job0"
|
|
}
|
|
}
|
|
|
|
(5) [On *destination* QEMU] Then, stop the NBD server::
|
|
|
|
(QEMU) nbd-server-stop
|
|
{
|
|
"execute": "nbd-server-stop",
|
|
"arguments": {}
|
|
}
|
|
|
|
(6) [On *destination* QEMU] Finally, resume the guest vCPUs by issuing the
|
|
QMP command `cont`::
|
|
|
|
(QEMU) cont
|
|
{
|
|
"execute": "cont",
|
|
"arguments": {}
|
|
}
|
|
|
|
.. note::
|
|
Higher-level libraries (e.g. libvirt) automate the entire above
|
|
process (although note that libvirt does not allow same-host
|
|
migrations to localhost for other reasons).
|
|
|
|
|
|
Notes on ``blockdev-mirror``
|
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|
|
|
The ``blockdev-mirror`` command is equivalent in core functionality to
|
|
``drive-mirror``, except that it operates at node-level in a BDS graph.
|
|
|
|
Also: for ``blockdev-mirror``, the 'target' image needs to be explicitly
|
|
created (using ``qemu-img``) and attach it to live QEMU via
|
|
``blockdev-add``, which assigns a name to the to-be created target node.
|
|
|
|
E.g. the sequence of actions to create a point-in-time backup of an
|
|
entire disk image chain, to a target, using ``blockdev-mirror`` would be:
|
|
|
|
(0) Create the QCOW2 overlays, to arrive at a backing chain of desired
|
|
depth
|
|
|
|
(1) Create the target image (using ``qemu-img``), say, ``e.qcow2``
|
|
|
|
(2) Attach the above created file (``e.qcow2``), run-time, using
|
|
``blockdev-add`` to QEMU
|
|
|
|
(3) Perform ``blockdev-mirror`` (use ``"sync": "full"`` to copy the
|
|
entire chain to the target). And notice the event
|
|
``BLOCK_JOB_READY``
|
|
|
|
(4) Optionally, query for active block jobs, there should be a 'mirror'
|
|
job ready to be completed
|
|
|
|
(5) Gracefully complete the 'mirror' block device job, and notice the
|
|
the event ``BLOCK_JOB_COMPLETED``
|
|
|
|
(6) Shutdown the guest by issuing the QMP ``quit`` command so that
|
|
caches are flushed
|
|
|
|
(7) Then, finally, compare the contents of the disk image chain, and
|
|
the target copy with ``qemu-img compare``. You should notice:
|
|
"Images are identical"
|
|
|
|
|
|
QMP invocation for ``blockdev-mirror``
|
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|
|
|
Given the disk image chain::
|
|
|
|
[A] <-- [B] <-- [C] <-- [D]
|
|
|
|
To copy the contents of the entire disk image chain, from [A] all the
|
|
way to [D], to a new target, call it [E]. The following is the flow.
|
|
|
|
Create the overlay images, [B], [C], and [D]::
|
|
|
|
(QEMU) blockdev-snapshot-sync node-name=node-A snapshot-file=b.qcow2 snapshot-node-name=node-B format=qcow2
|
|
(QEMU) blockdev-snapshot-sync node-name=node-B snapshot-file=c.qcow2 snapshot-node-name=node-C format=qcow2
|
|
(QEMU) blockdev-snapshot-sync node-name=node-C snapshot-file=d.qcow2 snapshot-node-name=node-D format=qcow2
|
|
|
|
Create the target image, [E]::
|
|
|
|
$ qemu-img create -f qcow2 e.qcow2 39M
|
|
|
|
Add the above created target image to QEMU, via ``blockdev-add``::
|
|
|
|
(QEMU) blockdev-add driver=qcow2 node-name=node-E file={"driver":"file","filename":"e.qcow2"}
|
|
{
|
|
"execute": "blockdev-add",
|
|
"arguments": {
|
|
"node-name": "node-E",
|
|
"driver": "qcow2",
|
|
"file": {
|
|
"driver": "file",
|
|
"filename": "e.qcow2"
|
|
}
|
|
}
|
|
}
|
|
|
|
Perform ``blockdev-mirror``, and notice the event ``BLOCK_JOB_READY``::
|
|
|
|
(QEMU) blockdev-mirror device=node-B target=node-E sync=full job-id=job0
|
|
{
|
|
"execute": "blockdev-mirror",
|
|
"arguments": {
|
|
"device": "node-D",
|
|
"job-id": "job0",
|
|
"target": "node-E",
|
|
"sync": "full"
|
|
}
|
|
}
|
|
|
|
Query for active block jobs, there should be a 'mirror' job ready::
|
|
|
|
(QEMU) query-block-jobs
|
|
{
|
|
"execute": "query-block-jobs",
|
|
"arguments": {}
|
|
}
|
|
{
|
|
"return": [
|
|
{
|
|
"busy": false,
|
|
"type": "mirror",
|
|
"len": 21561344,
|
|
"paused": false,
|
|
"ready": true,
|
|
"io-status": "ok",
|
|
"offset": 21561344,
|
|
"device": "job0",
|
|
"speed": 0
|
|
}
|
|
]
|
|
}
|
|
|
|
Gracefully complete the block device job operation, and notice the
|
|
event ``BLOCK_JOB_COMPLETED``::
|
|
|
|
(QEMU) block-job-complete device=job0
|
|
{
|
|
"execute": "block-job-complete",
|
|
"arguments": {
|
|
"device": "job0"
|
|
}
|
|
}
|
|
{
|
|
"return": {}
|
|
}
|
|
|
|
Shutdown the guest, by issuing the ``quit`` QMP command::
|
|
|
|
(QEMU) quit
|
|
{
|
|
"execute": "quit",
|
|
"arguments": {}
|
|
}
|
|
|
|
|
|
Live disk backup --- ``drive-backup`` and ``blockdev-backup``
|
|
-------------------------------------------------------------
|
|
|
|
The ``drive-backup`` (and its newer equivalent ``blockdev-backup``) allows
|
|
you to create a point-in-time snapshot.
|
|
|
|
In this case, the point-in-time is when you *start* the ``drive-backup``
|
|
(or its newer equivalent ``blockdev-backup``) command.
|
|
|
|
|
|
QMP invocation for ``drive-backup``
|
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|
|
|
Yet again, starting afresh with our example disk image chain::
|
|
|
|
[A] <-- [B] <-- [C] <-- [D]
|
|
|
|
To create a target image [E], with content populated from image [A] to
|
|
[D], from the above chain, the following is the syntax. (If the target
|
|
image does not exist, ``drive-backup`` will create it)::
|
|
|
|
(QEMU) drive-backup device=node-D sync=full target=e.qcow2 job-id=job0
|
|
{
|
|
"execute": "drive-backup",
|
|
"arguments": {
|
|
"device": "node-D",
|
|
"job-id": "job0",
|
|
"sync": "full",
|
|
"target": "e.qcow2"
|
|
}
|
|
}
|
|
|
|
Once the above ``drive-backup`` has completed, a ``BLOCK_JOB_COMPLETED`` event
|
|
will be issued, indicating the live block device job operation has
|
|
completed, and no further action is required.
|
|
|
|
|
|
Notes on ``blockdev-backup``
|
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|
|
|
The ``blockdev-backup`` command is equivalent in functionality to
|
|
``drive-backup``, except that it operates at node-level in a Block Driver
|
|
State (BDS) graph.
|
|
|
|
E.g. the sequence of actions to create a point-in-time backup
|
|
of an entire disk image chain, to a target, using ``blockdev-backup``
|
|
would be:
|
|
|
|
(0) Create the QCOW2 overlays, to arrive at a backing chain of desired
|
|
depth
|
|
|
|
(1) Create the target image (using ``qemu-img``), say, ``e.qcow2``
|
|
|
|
(2) Attach the above created file (``e.qcow2``), run-time, using
|
|
``blockdev-add`` to QEMU
|
|
|
|
(3) Perform ``blockdev-backup`` (use ``"sync": "full"`` to copy the
|
|
entire chain to the target). And notice the event
|
|
``BLOCK_JOB_COMPLETED``
|
|
|
|
(4) Shutdown the guest, by issuing the QMP ``quit`` command, so that
|
|
caches are flushed
|
|
|
|
(5) Then, finally, compare the contents of the disk image chain, and
|
|
the target copy with ``qemu-img compare``. You should notice:
|
|
"Images are identical"
|
|
|
|
The following section shows an example QMP invocation for
|
|
``blockdev-backup``.
|
|
|
|
QMP invocation for ``blockdev-backup``
|
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|
|
|
Given a disk image chain of depth 1 where image [B] is the active
|
|
overlay (live QEMU is writing to it)::
|
|
|
|
[A] <-- [B]
|
|
|
|
The following is the procedure to copy the content from the entire chain
|
|
to a target image (say, [E]), which has the full content from [A] and
|
|
[B].
|
|
|
|
Create the overlay [B]::
|
|
|
|
(QEMU) blockdev-snapshot-sync node-name=node-A snapshot-file=b.qcow2 snapshot-node-name=node-B format=qcow2
|
|
{
|
|
"execute": "blockdev-snapshot-sync",
|
|
"arguments": {
|
|
"node-name": "node-A",
|
|
"snapshot-file": "b.qcow2",
|
|
"format": "qcow2",
|
|
"snapshot-node-name": "node-B"
|
|
}
|
|
}
|
|
|
|
|
|
Create a target image that will contain the copy::
|
|
|
|
$ qemu-img create -f qcow2 e.qcow2 39M
|
|
|
|
Then add it to QEMU via ``blockdev-add``::
|
|
|
|
(QEMU) blockdev-add driver=qcow2 node-name=node-E file={"driver":"file","filename":"e.qcow2"}
|
|
{
|
|
"execute": "blockdev-add",
|
|
"arguments": {
|
|
"node-name": "node-E",
|
|
"driver": "qcow2",
|
|
"file": {
|
|
"driver": "file",
|
|
"filename": "e.qcow2"
|
|
}
|
|
}
|
|
}
|
|
|
|
Then invoke ``blockdev-backup`` to copy the contents from the entire
|
|
image chain, consisting of images [A] and [B] to the target image
|
|
'e.qcow2'::
|
|
|
|
(QEMU) blockdev-backup device=node-B target=node-E sync=full job-id=job0
|
|
{
|
|
"execute": "blockdev-backup",
|
|
"arguments": {
|
|
"device": "node-B",
|
|
"job-id": "job0",
|
|
"target": "node-E",
|
|
"sync": "full"
|
|
}
|
|
}
|
|
|
|
Once the above 'backup' operation has completed, the event,
|
|
``BLOCK_JOB_COMPLETED`` will be emitted, signalling successful
|
|
completion.
|
|
|
|
Next, query for any active block device jobs (there should be none)::
|
|
|
|
(QEMU) query-block-jobs
|
|
{
|
|
"execute": "query-block-jobs",
|
|
"arguments": {}
|
|
}
|
|
|
|
Shutdown the guest::
|
|
|
|
(QEMU) quit
|
|
{
|
|
"execute": "quit",
|
|
"arguments": {}
|
|
}
|
|
"return": {}
|
|
}
|
|
|
|
.. note::
|
|
The above step is really important; if forgotten, an error, "Failed
|
|
to get shared "write" lock on e.qcow2", will be thrown when you do
|
|
``qemu-img compare`` to verify the integrity of the disk image
|
|
with the backup content.
|
|
|
|
|
|
The end result will be the image 'e.qcow2' containing a
|
|
point-in-time backup of the disk image chain -- i.e. contents from
|
|
images [A] and [B] at the time the ``blockdev-backup`` command was
|
|
initiated.
|
|
|
|
One way to confirm the backup disk image contains the identical content
|
|
with the disk image chain is to compare the backup and the contents of
|
|
the chain, you should see "Images are identical". (NB: this is assuming
|
|
QEMU was launched with ``-S`` option, which will not start the CPUs at
|
|
guest boot up)::
|
|
|
|
$ qemu-img compare b.qcow2 e.qcow2
|
|
Warning: Image size mismatch!
|
|
Images are identical.
|
|
|
|
NOTE: The "Warning: Image size mismatch!" is expected, as we created the
|
|
target image (e.qcow2) with 39M size.
|