2016-07-27 09:01:50 +02:00
|
|
|
/*
|
|
|
|
* Replication Block filter
|
|
|
|
*
|
|
|
|
* Copyright (c) 2016 HUAWEI TECHNOLOGIES CO., LTD.
|
|
|
|
* Copyright (c) 2016 Intel Corporation
|
|
|
|
* Copyright (c) 2016 FUJITSU LIMITED
|
|
|
|
*
|
|
|
|
* Author:
|
|
|
|
* Wen Congyang <wency@cn.fujitsu.com>
|
|
|
|
*
|
|
|
|
* This work is licensed under the terms of the GNU GPL, version 2 or later.
|
|
|
|
* See the COPYING file in the top-level directory.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include "qemu/osdep.h"
|
|
|
|
#include "qemu-common.h"
|
|
|
|
#include "block/nbd.h"
|
|
|
|
#include "block/blockjob.h"
|
|
|
|
#include "block/block_int.h"
|
|
|
|
#include "block/block_backup.h"
|
|
|
|
#include "sysemu/block-backend.h"
|
|
|
|
#include "qapi/error.h"
|
|
|
|
#include "replication.h"
|
|
|
|
|
2017-03-17 03:17:39 +01:00
|
|
|
typedef enum {
|
|
|
|
BLOCK_REPLICATION_NONE, /* block replication is not started */
|
|
|
|
BLOCK_REPLICATION_RUNNING, /* block replication is running */
|
|
|
|
BLOCK_REPLICATION_FAILOVER, /* failover is running in background */
|
|
|
|
BLOCK_REPLICATION_FAILOVER_FAILED, /* failover failed */
|
|
|
|
BLOCK_REPLICATION_DONE, /* block replication is done */
|
|
|
|
} ReplicationStage;
|
|
|
|
|
2016-07-27 09:01:50 +02:00
|
|
|
typedef struct BDRVReplicationState {
|
|
|
|
ReplicationMode mode;
|
2017-03-17 03:17:39 +01:00
|
|
|
ReplicationStage stage;
|
2016-07-27 09:01:50 +02:00
|
|
|
BdrvChild *active_disk;
|
|
|
|
BdrvChild *hidden_disk;
|
|
|
|
BdrvChild *secondary_disk;
|
|
|
|
char *top_id;
|
|
|
|
ReplicationState *rs;
|
|
|
|
Error *blocker;
|
|
|
|
int orig_hidden_flags;
|
|
|
|
int orig_secondary_flags;
|
|
|
|
int error;
|
|
|
|
} BDRVReplicationState;
|
|
|
|
|
|
|
|
static void replication_start(ReplicationState *rs, ReplicationMode mode,
|
|
|
|
Error **errp);
|
|
|
|
static void replication_do_checkpoint(ReplicationState *rs, Error **errp);
|
|
|
|
static void replication_get_error(ReplicationState *rs, Error **errp);
|
|
|
|
static void replication_stop(ReplicationState *rs, bool failover,
|
|
|
|
Error **errp);
|
|
|
|
|
|
|
|
#define REPLICATION_MODE "mode"
|
|
|
|
#define REPLICATION_TOP_ID "top-id"
|
|
|
|
static QemuOptsList replication_runtime_opts = {
|
|
|
|
.name = "replication",
|
|
|
|
.head = QTAILQ_HEAD_INITIALIZER(replication_runtime_opts.head),
|
|
|
|
.desc = {
|
|
|
|
{
|
|
|
|
.name = REPLICATION_MODE,
|
|
|
|
.type = QEMU_OPT_STRING,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.name = REPLICATION_TOP_ID,
|
|
|
|
.type = QEMU_OPT_STRING,
|
|
|
|
},
|
|
|
|
{ /* end of list */ }
|
|
|
|
},
|
|
|
|
};
|
|
|
|
|
|
|
|
static ReplicationOps replication_ops = {
|
|
|
|
.start = replication_start,
|
|
|
|
.checkpoint = replication_do_checkpoint,
|
|
|
|
.get_error = replication_get_error,
|
|
|
|
.stop = replication_stop,
|
|
|
|
};
|
|
|
|
|
|
|
|
static int replication_open(BlockDriverState *bs, QDict *options,
|
|
|
|
int flags, Error **errp)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
BDRVReplicationState *s = bs->opaque;
|
|
|
|
Error *local_err = NULL;
|
|
|
|
QemuOpts *opts = NULL;
|
|
|
|
const char *mode;
|
|
|
|
const char *top_id;
|
|
|
|
|
2016-12-16 18:52:37 +01:00
|
|
|
bs->file = bdrv_open_child(NULL, options, "file", bs, &child_file,
|
|
|
|
false, errp);
|
|
|
|
if (!bs->file) {
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
2016-07-27 09:01:50 +02:00
|
|
|
ret = -EINVAL;
|
|
|
|
opts = qemu_opts_create(&replication_runtime_opts, NULL, 0, &error_abort);
|
|
|
|
qemu_opts_absorb_qdict(opts, options, &local_err);
|
|
|
|
if (local_err) {
|
|
|
|
goto fail;
|
|
|
|
}
|
|
|
|
|
|
|
|
mode = qemu_opt_get(opts, REPLICATION_MODE);
|
|
|
|
if (!mode) {
|
|
|
|
error_setg(&local_err, "Missing the option mode");
|
|
|
|
goto fail;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!strcmp(mode, "primary")) {
|
|
|
|
s->mode = REPLICATION_MODE_PRIMARY;
|
2016-10-12 06:50:08 +02:00
|
|
|
top_id = qemu_opt_get(opts, REPLICATION_TOP_ID);
|
|
|
|
if (top_id) {
|
|
|
|
error_setg(&local_err, "The primary side does not support option top-id");
|
|
|
|
goto fail;
|
|
|
|
}
|
2016-07-27 09:01:50 +02:00
|
|
|
} else if (!strcmp(mode, "secondary")) {
|
|
|
|
s->mode = REPLICATION_MODE_SECONDARY;
|
|
|
|
top_id = qemu_opt_get(opts, REPLICATION_TOP_ID);
|
|
|
|
s->top_id = g_strdup(top_id);
|
|
|
|
if (!s->top_id) {
|
|
|
|
error_setg(&local_err, "Missing the option top-id");
|
|
|
|
goto fail;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
error_setg(&local_err,
|
|
|
|
"The option mode's value should be primary or secondary");
|
|
|
|
goto fail;
|
|
|
|
}
|
|
|
|
|
|
|
|
s->rs = replication_new(bs, &replication_ops);
|
|
|
|
|
|
|
|
ret = 0;
|
|
|
|
|
|
|
|
fail:
|
|
|
|
qemu_opts_del(opts);
|
|
|
|
error_propagate(errp, local_err);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void replication_close(BlockDriverState *bs)
|
|
|
|
{
|
|
|
|
BDRVReplicationState *s = bs->opaque;
|
|
|
|
|
2017-03-17 03:17:39 +01:00
|
|
|
if (s->stage == BLOCK_REPLICATION_RUNNING) {
|
2016-07-27 09:01:50 +02:00
|
|
|
replication_stop(s->rs, false, NULL);
|
|
|
|
}
|
2017-03-17 03:17:39 +01:00
|
|
|
if (s->stage == BLOCK_REPLICATION_FAILOVER) {
|
2016-10-27 12:48:49 +02:00
|
|
|
block_job_cancel_sync(s->active_disk->bs->job);
|
|
|
|
}
|
2016-07-27 09:01:50 +02:00
|
|
|
|
|
|
|
if (s->mode == REPLICATION_MODE_SECONDARY) {
|
|
|
|
g_free(s->top_id);
|
|
|
|
}
|
|
|
|
|
|
|
|
replication_remove(s->rs);
|
|
|
|
}
|
|
|
|
|
2017-03-14 12:46:52 +01:00
|
|
|
static void replication_child_perm(BlockDriverState *bs, BdrvChild *c,
|
|
|
|
const BdrvChildRole *role,
|
|
|
|
uint64_t perm, uint64_t shared,
|
|
|
|
uint64_t *nperm, uint64_t *nshared)
|
|
|
|
{
|
|
|
|
*nperm = *nshared = BLK_PERM_CONSISTENT_READ \
|
|
|
|
| BLK_PERM_WRITE \
|
|
|
|
| BLK_PERM_WRITE_UNCHANGED;
|
|
|
|
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2016-07-27 09:01:50 +02:00
|
|
|
static int64_t replication_getlength(BlockDriverState *bs)
|
|
|
|
{
|
|
|
|
return bdrv_getlength(bs->file->bs);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int replication_get_io_status(BDRVReplicationState *s)
|
|
|
|
{
|
2017-03-17 03:17:39 +01:00
|
|
|
switch (s->stage) {
|
2016-07-27 09:01:50 +02:00
|
|
|
case BLOCK_REPLICATION_NONE:
|
|
|
|
return -EIO;
|
|
|
|
case BLOCK_REPLICATION_RUNNING:
|
|
|
|
return 0;
|
|
|
|
case BLOCK_REPLICATION_FAILOVER:
|
|
|
|
return s->mode == REPLICATION_MODE_PRIMARY ? -EIO : 0;
|
|
|
|
case BLOCK_REPLICATION_FAILOVER_FAILED:
|
|
|
|
return s->mode == REPLICATION_MODE_PRIMARY ? -EIO : 1;
|
|
|
|
case BLOCK_REPLICATION_DONE:
|
|
|
|
/*
|
|
|
|
* active commit job completes, and active disk and secondary_disk
|
|
|
|
* is swapped, so we can operate bs->file directly
|
|
|
|
*/
|
|
|
|
return s->mode == REPLICATION_MODE_PRIMARY ? -EIO : 0;
|
|
|
|
default:
|
|
|
|
abort();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static int replication_return_value(BDRVReplicationState *s, int ret)
|
|
|
|
{
|
|
|
|
if (s->mode == REPLICATION_MODE_SECONDARY) {
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (ret < 0) {
|
|
|
|
s->error = ret;
|
|
|
|
ret = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static coroutine_fn int replication_co_readv(BlockDriverState *bs,
|
|
|
|
int64_t sector_num,
|
|
|
|
int remaining_sectors,
|
|
|
|
QEMUIOVector *qiov)
|
|
|
|
{
|
|
|
|
BDRVReplicationState *s = bs->opaque;
|
|
|
|
BdrvChild *child = s->secondary_disk;
|
|
|
|
BlockJob *job = NULL;
|
|
|
|
CowRequest req;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
if (s->mode == REPLICATION_MODE_PRIMARY) {
|
|
|
|
/* We only use it to forward primary write requests */
|
|
|
|
return -EIO;
|
|
|
|
}
|
|
|
|
|
|
|
|
ret = replication_get_io_status(s);
|
|
|
|
if (ret < 0) {
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (child && child->bs) {
|
|
|
|
job = child->bs->job;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (job) {
|
2017-07-07 14:44:54 +02:00
|
|
|
uint64_t remaining_bytes = remaining_sectors * BDRV_SECTOR_SIZE;
|
|
|
|
|
|
|
|
backup_wait_for_overlapping_requests(child->bs->job,
|
|
|
|
sector_num * BDRV_SECTOR_SIZE,
|
|
|
|
remaining_bytes);
|
|
|
|
backup_cow_request_begin(&req, child->bs->job,
|
|
|
|
sector_num * BDRV_SECTOR_SIZE,
|
|
|
|
remaining_bytes);
|
2016-07-27 09:01:50 +02:00
|
|
|
ret = bdrv_co_readv(bs->file, sector_num, remaining_sectors,
|
|
|
|
qiov);
|
|
|
|
backup_cow_request_end(&req);
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
ret = bdrv_co_readv(bs->file, sector_num, remaining_sectors, qiov);
|
|
|
|
out:
|
|
|
|
return replication_return_value(s, ret);
|
|
|
|
}
|
|
|
|
|
|
|
|
static coroutine_fn int replication_co_writev(BlockDriverState *bs,
|
|
|
|
int64_t sector_num,
|
|
|
|
int remaining_sectors,
|
|
|
|
QEMUIOVector *qiov)
|
|
|
|
{
|
|
|
|
BDRVReplicationState *s = bs->opaque;
|
|
|
|
QEMUIOVector hd_qiov;
|
|
|
|
uint64_t bytes_done = 0;
|
|
|
|
BdrvChild *top = bs->file;
|
|
|
|
BdrvChild *base = s->secondary_disk;
|
|
|
|
BdrvChild *target;
|
block: Make bdrv_is_allocated_above() byte-based
We are gradually moving away from sector-based interfaces, towards
byte-based. In the common case, allocation is unlikely to ever use
values that are not naturally sector-aligned, but it is possible
that byte-based values will let us be more precise about allocation
at the end of an unaligned file that can do byte-based access.
Changing the signature of the function to use int64_t *pnum ensures
that the compiler enforces that all callers are updated. For now,
the io.c layer still assert()s that all callers are sector-aligned,
but that can be relaxed when a later patch implements byte-based
block status. Therefore, for the most part this patch is just the
addition of scaling at the callers followed by inverse scaling at
bdrv_is_allocated(). But some code, particularly stream_run(),
gets a lot simpler because it no longer has to mess with sectors.
Leave comments where we can further simplify by switching to
byte-based iterations, once later patches eliminate the need for
sector-aligned operations.
For ease of review, bdrv_is_allocated() was tackled separately.
Signed-off-by: Eric Blake <eblake@redhat.com>
Signed-off-by: Kevin Wolf <kwolf@redhat.com>
2017-07-07 14:44:59 +02:00
|
|
|
int ret;
|
|
|
|
int64_t n;
|
2016-07-27 09:01:50 +02:00
|
|
|
|
|
|
|
ret = replication_get_io_status(s);
|
|
|
|
if (ret < 0) {
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (ret == 0) {
|
|
|
|
ret = bdrv_co_writev(top, sector_num,
|
|
|
|
remaining_sectors, qiov);
|
|
|
|
return replication_return_value(s, ret);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Failover failed, only write to active disk if the sectors
|
|
|
|
* have already been allocated in active disk/hidden disk.
|
|
|
|
*/
|
|
|
|
qemu_iovec_init(&hd_qiov, qiov->niov);
|
|
|
|
while (remaining_sectors > 0) {
|
block: Make bdrv_is_allocated_above() byte-based
We are gradually moving away from sector-based interfaces, towards
byte-based. In the common case, allocation is unlikely to ever use
values that are not naturally sector-aligned, but it is possible
that byte-based values will let us be more precise about allocation
at the end of an unaligned file that can do byte-based access.
Changing the signature of the function to use int64_t *pnum ensures
that the compiler enforces that all callers are updated. For now,
the io.c layer still assert()s that all callers are sector-aligned,
but that can be relaxed when a later patch implements byte-based
block status. Therefore, for the most part this patch is just the
addition of scaling at the callers followed by inverse scaling at
bdrv_is_allocated(). But some code, particularly stream_run(),
gets a lot simpler because it no longer has to mess with sectors.
Leave comments where we can further simplify by switching to
byte-based iterations, once later patches eliminate the need for
sector-aligned operations.
For ease of review, bdrv_is_allocated() was tackled separately.
Signed-off-by: Eric Blake <eblake@redhat.com>
Signed-off-by: Kevin Wolf <kwolf@redhat.com>
2017-07-07 14:44:59 +02:00
|
|
|
int64_t count;
|
|
|
|
|
|
|
|
ret = bdrv_is_allocated_above(top->bs, base->bs,
|
|
|
|
sector_num * BDRV_SECTOR_SIZE,
|
|
|
|
remaining_sectors * BDRV_SECTOR_SIZE,
|
|
|
|
&count);
|
2016-07-27 09:01:50 +02:00
|
|
|
if (ret < 0) {
|
|
|
|
goto out1;
|
|
|
|
}
|
|
|
|
|
block: Make bdrv_is_allocated_above() byte-based
We are gradually moving away from sector-based interfaces, towards
byte-based. In the common case, allocation is unlikely to ever use
values that are not naturally sector-aligned, but it is possible
that byte-based values will let us be more precise about allocation
at the end of an unaligned file that can do byte-based access.
Changing the signature of the function to use int64_t *pnum ensures
that the compiler enforces that all callers are updated. For now,
the io.c layer still assert()s that all callers are sector-aligned,
but that can be relaxed when a later patch implements byte-based
block status. Therefore, for the most part this patch is just the
addition of scaling at the callers followed by inverse scaling at
bdrv_is_allocated(). But some code, particularly stream_run(),
gets a lot simpler because it no longer has to mess with sectors.
Leave comments where we can further simplify by switching to
byte-based iterations, once later patches eliminate the need for
sector-aligned operations.
For ease of review, bdrv_is_allocated() was tackled separately.
Signed-off-by: Eric Blake <eblake@redhat.com>
Signed-off-by: Kevin Wolf <kwolf@redhat.com>
2017-07-07 14:44:59 +02:00
|
|
|
assert(QEMU_IS_ALIGNED(count, BDRV_SECTOR_SIZE));
|
|
|
|
n = count >> BDRV_SECTOR_BITS;
|
2016-07-27 09:01:50 +02:00
|
|
|
qemu_iovec_reset(&hd_qiov);
|
block: Make bdrv_is_allocated_above() byte-based
We are gradually moving away from sector-based interfaces, towards
byte-based. In the common case, allocation is unlikely to ever use
values that are not naturally sector-aligned, but it is possible
that byte-based values will let us be more precise about allocation
at the end of an unaligned file that can do byte-based access.
Changing the signature of the function to use int64_t *pnum ensures
that the compiler enforces that all callers are updated. For now,
the io.c layer still assert()s that all callers are sector-aligned,
but that can be relaxed when a later patch implements byte-based
block status. Therefore, for the most part this patch is just the
addition of scaling at the callers followed by inverse scaling at
bdrv_is_allocated(). But some code, particularly stream_run(),
gets a lot simpler because it no longer has to mess with sectors.
Leave comments where we can further simplify by switching to
byte-based iterations, once later patches eliminate the need for
sector-aligned operations.
For ease of review, bdrv_is_allocated() was tackled separately.
Signed-off-by: Eric Blake <eblake@redhat.com>
Signed-off-by: Kevin Wolf <kwolf@redhat.com>
2017-07-07 14:44:59 +02:00
|
|
|
qemu_iovec_concat(&hd_qiov, qiov, bytes_done, count);
|
2016-07-27 09:01:50 +02:00
|
|
|
|
|
|
|
target = ret ? top : base;
|
|
|
|
ret = bdrv_co_writev(target, sector_num, n, &hd_qiov);
|
|
|
|
if (ret < 0) {
|
|
|
|
goto out1;
|
|
|
|
}
|
|
|
|
|
|
|
|
remaining_sectors -= n;
|
|
|
|
sector_num += n;
|
block: Make bdrv_is_allocated_above() byte-based
We are gradually moving away from sector-based interfaces, towards
byte-based. In the common case, allocation is unlikely to ever use
values that are not naturally sector-aligned, but it is possible
that byte-based values will let us be more precise about allocation
at the end of an unaligned file that can do byte-based access.
Changing the signature of the function to use int64_t *pnum ensures
that the compiler enforces that all callers are updated. For now,
the io.c layer still assert()s that all callers are sector-aligned,
but that can be relaxed when a later patch implements byte-based
block status. Therefore, for the most part this patch is just the
addition of scaling at the callers followed by inverse scaling at
bdrv_is_allocated(). But some code, particularly stream_run(),
gets a lot simpler because it no longer has to mess with sectors.
Leave comments where we can further simplify by switching to
byte-based iterations, once later patches eliminate the need for
sector-aligned operations.
For ease of review, bdrv_is_allocated() was tackled separately.
Signed-off-by: Eric Blake <eblake@redhat.com>
Signed-off-by: Kevin Wolf <kwolf@redhat.com>
2017-07-07 14:44:59 +02:00
|
|
|
bytes_done += count;
|
2016-07-27 09:01:50 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
out1:
|
|
|
|
qemu_iovec_destroy(&hd_qiov);
|
|
|
|
out:
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static bool replication_recurse_is_first_non_filter(BlockDriverState *bs,
|
|
|
|
BlockDriverState *candidate)
|
|
|
|
{
|
|
|
|
return bdrv_recurse_is_first_non_filter(bs->file->bs, candidate);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void secondary_do_checkpoint(BDRVReplicationState *s, Error **errp)
|
|
|
|
{
|
|
|
|
Error *local_err = NULL;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
if (!s->secondary_disk->bs->job) {
|
|
|
|
error_setg(errp, "Backup job was cancelled unexpectedly");
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
backup_do_checkpoint(s->secondary_disk->bs->job, &local_err);
|
|
|
|
if (local_err) {
|
|
|
|
error_propagate(errp, local_err);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
ret = s->active_disk->bs->drv->bdrv_make_empty(s->active_disk->bs);
|
|
|
|
if (ret < 0) {
|
|
|
|
error_setg(errp, "Cannot make active disk empty");
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
ret = s->hidden_disk->bs->drv->bdrv_make_empty(s->hidden_disk->bs);
|
|
|
|
if (ret < 0) {
|
|
|
|
error_setg(errp, "Cannot make hidden disk empty");
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2016-10-27 12:49:01 +02:00
|
|
|
static void reopen_backing_file(BlockDriverState *bs, bool writable,
|
2016-07-27 09:01:50 +02:00
|
|
|
Error **errp)
|
|
|
|
{
|
2016-10-27 12:49:01 +02:00
|
|
|
BDRVReplicationState *s = bs->opaque;
|
2016-07-27 09:01:50 +02:00
|
|
|
BlockReopenQueue *reopen_queue = NULL;
|
|
|
|
int orig_hidden_flags, orig_secondary_flags;
|
|
|
|
int new_hidden_flags, new_secondary_flags;
|
|
|
|
Error *local_err = NULL;
|
|
|
|
|
|
|
|
if (writable) {
|
|
|
|
orig_hidden_flags = s->orig_hidden_flags =
|
|
|
|
bdrv_get_flags(s->hidden_disk->bs);
|
|
|
|
new_hidden_flags = (orig_hidden_flags | BDRV_O_RDWR) &
|
|
|
|
~BDRV_O_INACTIVE;
|
|
|
|
orig_secondary_flags = s->orig_secondary_flags =
|
|
|
|
bdrv_get_flags(s->secondary_disk->bs);
|
|
|
|
new_secondary_flags = (orig_secondary_flags | BDRV_O_RDWR) &
|
|
|
|
~BDRV_O_INACTIVE;
|
|
|
|
} else {
|
|
|
|
orig_hidden_flags = (s->orig_hidden_flags | BDRV_O_RDWR) &
|
|
|
|
~BDRV_O_INACTIVE;
|
|
|
|
new_hidden_flags = s->orig_hidden_flags;
|
|
|
|
orig_secondary_flags = (s->orig_secondary_flags | BDRV_O_RDWR) &
|
|
|
|
~BDRV_O_INACTIVE;
|
|
|
|
new_secondary_flags = s->orig_secondary_flags;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (orig_hidden_flags != new_hidden_flags) {
|
|
|
|
reopen_queue = bdrv_reopen_queue(reopen_queue, s->hidden_disk->bs, NULL,
|
|
|
|
new_hidden_flags);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!(orig_secondary_flags & BDRV_O_RDWR)) {
|
|
|
|
reopen_queue = bdrv_reopen_queue(reopen_queue, s->secondary_disk->bs,
|
|
|
|
NULL, new_secondary_flags);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (reopen_queue) {
|
2016-10-27 12:49:02 +02:00
|
|
|
bdrv_reopen_multiple(bdrv_get_aio_context(bs),
|
|
|
|
reopen_queue, &local_err);
|
2016-07-27 09:01:50 +02:00
|
|
|
error_propagate(errp, local_err);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2016-10-27 12:49:01 +02:00
|
|
|
static void backup_job_cleanup(BlockDriverState *bs)
|
2016-07-27 09:01:50 +02:00
|
|
|
{
|
2016-10-27 12:49:01 +02:00
|
|
|
BDRVReplicationState *s = bs->opaque;
|
2016-07-27 09:01:50 +02:00
|
|
|
BlockDriverState *top_bs;
|
|
|
|
|
|
|
|
top_bs = bdrv_lookup_bs(s->top_id, s->top_id, NULL);
|
|
|
|
if (!top_bs) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
bdrv_op_unblock_all(top_bs, s->blocker);
|
|
|
|
error_free(s->blocker);
|
2016-10-27 12:49:01 +02:00
|
|
|
reopen_backing_file(bs, false, NULL);
|
2016-07-27 09:01:50 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
static void backup_job_completed(void *opaque, int ret)
|
|
|
|
{
|
2016-10-27 12:49:01 +02:00
|
|
|
BlockDriverState *bs = opaque;
|
|
|
|
BDRVReplicationState *s = bs->opaque;
|
2016-07-27 09:01:50 +02:00
|
|
|
|
2017-03-17 03:17:39 +01:00
|
|
|
if (s->stage != BLOCK_REPLICATION_FAILOVER) {
|
2016-07-27 09:01:50 +02:00
|
|
|
/* The backup job is cancelled unexpectedly */
|
|
|
|
s->error = -EIO;
|
|
|
|
}
|
|
|
|
|
2016-10-27 12:49:01 +02:00
|
|
|
backup_job_cleanup(bs);
|
2016-07-27 09:01:50 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
static bool check_top_bs(BlockDriverState *top_bs, BlockDriverState *bs)
|
|
|
|
{
|
|
|
|
BdrvChild *child;
|
|
|
|
|
|
|
|
/* The bs itself is the top_bs */
|
|
|
|
if (top_bs == bs) {
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Iterate over top_bs's children */
|
|
|
|
QLIST_FOREACH(child, &top_bs->children, next) {
|
|
|
|
if (child->bs == bs || check_top_bs(child->bs, bs)) {
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void replication_start(ReplicationState *rs, ReplicationMode mode,
|
|
|
|
Error **errp)
|
|
|
|
{
|
|
|
|
BlockDriverState *bs = rs->opaque;
|
|
|
|
BDRVReplicationState *s;
|
|
|
|
BlockDriverState *top_bs;
|
|
|
|
int64_t active_length, hidden_length, disk_length;
|
|
|
|
AioContext *aio_context;
|
|
|
|
Error *local_err = NULL;
|
2016-11-08 07:50:38 +01:00
|
|
|
BlockJob *job;
|
2016-07-27 09:01:50 +02:00
|
|
|
|
|
|
|
aio_context = bdrv_get_aio_context(bs);
|
|
|
|
aio_context_acquire(aio_context);
|
|
|
|
s = bs->opaque;
|
|
|
|
|
2017-03-17 03:17:39 +01:00
|
|
|
if (s->stage != BLOCK_REPLICATION_NONE) {
|
2016-07-27 09:01:50 +02:00
|
|
|
error_setg(errp, "Block replication is running or done");
|
|
|
|
aio_context_release(aio_context);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (s->mode != mode) {
|
|
|
|
error_setg(errp, "The parameter mode's value is invalid, needs %d,"
|
|
|
|
" but got %d", s->mode, mode);
|
|
|
|
aio_context_release(aio_context);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
switch (s->mode) {
|
|
|
|
case REPLICATION_MODE_PRIMARY:
|
|
|
|
break;
|
|
|
|
case REPLICATION_MODE_SECONDARY:
|
|
|
|
s->active_disk = bs->file;
|
|
|
|
if (!s->active_disk || !s->active_disk->bs ||
|
|
|
|
!s->active_disk->bs->backing) {
|
|
|
|
error_setg(errp, "Active disk doesn't have backing file");
|
|
|
|
aio_context_release(aio_context);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
s->hidden_disk = s->active_disk->bs->backing;
|
|
|
|
if (!s->hidden_disk->bs || !s->hidden_disk->bs->backing) {
|
|
|
|
error_setg(errp, "Hidden disk doesn't have backing file");
|
|
|
|
aio_context_release(aio_context);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
s->secondary_disk = s->hidden_disk->bs->backing;
|
|
|
|
if (!s->secondary_disk->bs || !bdrv_has_blk(s->secondary_disk->bs)) {
|
|
|
|
error_setg(errp, "The secondary disk doesn't have block backend");
|
|
|
|
aio_context_release(aio_context);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* verify the length */
|
|
|
|
active_length = bdrv_getlength(s->active_disk->bs);
|
|
|
|
hidden_length = bdrv_getlength(s->hidden_disk->bs);
|
|
|
|
disk_length = bdrv_getlength(s->secondary_disk->bs);
|
|
|
|
if (active_length < 0 || hidden_length < 0 || disk_length < 0 ||
|
|
|
|
active_length != hidden_length || hidden_length != disk_length) {
|
|
|
|
error_setg(errp, "Active disk, hidden disk, secondary disk's length"
|
|
|
|
" are not the same");
|
|
|
|
aio_context_release(aio_context);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!s->active_disk->bs->drv->bdrv_make_empty ||
|
|
|
|
!s->hidden_disk->bs->drv->bdrv_make_empty) {
|
|
|
|
error_setg(errp,
|
|
|
|
"Active disk or hidden disk doesn't support make_empty");
|
|
|
|
aio_context_release(aio_context);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* reopen the backing file in r/w mode */
|
2016-10-27 12:49:01 +02:00
|
|
|
reopen_backing_file(bs, true, &local_err);
|
2016-07-27 09:01:50 +02:00
|
|
|
if (local_err) {
|
|
|
|
error_propagate(errp, local_err);
|
|
|
|
aio_context_release(aio_context);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* start backup job now */
|
|
|
|
error_setg(&s->blocker,
|
|
|
|
"Block device is in use by internal backup job");
|
|
|
|
|
|
|
|
top_bs = bdrv_lookup_bs(s->top_id, s->top_id, NULL);
|
|
|
|
if (!top_bs || !bdrv_is_root_node(top_bs) ||
|
|
|
|
!check_top_bs(top_bs, bs)) {
|
|
|
|
error_setg(errp, "No top_bs or it is invalid");
|
2016-10-27 12:49:01 +02:00
|
|
|
reopen_backing_file(bs, false, NULL);
|
2016-07-27 09:01:50 +02:00
|
|
|
aio_context_release(aio_context);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
bdrv_op_block_all(top_bs, s->blocker);
|
|
|
|
bdrv_op_unblock(top_bs, BLOCK_OP_TYPE_DATAPLANE, s->blocker);
|
|
|
|
|
2016-11-08 07:50:38 +01:00
|
|
|
job = backup_job_create(NULL, s->secondary_disk->bs, s->hidden_disk->bs,
|
|
|
|
0, MIRROR_SYNC_MODE_NONE, NULL, false,
|
|
|
|
BLOCKDEV_ON_ERROR_REPORT,
|
|
|
|
BLOCKDEV_ON_ERROR_REPORT, BLOCK_JOB_INTERNAL,
|
|
|
|
backup_job_completed, bs, NULL, &local_err);
|
2016-07-27 09:01:50 +02:00
|
|
|
if (local_err) {
|
|
|
|
error_propagate(errp, local_err);
|
2016-10-27 12:49:01 +02:00
|
|
|
backup_job_cleanup(bs);
|
2016-07-27 09:01:50 +02:00
|
|
|
aio_context_release(aio_context);
|
|
|
|
return;
|
|
|
|
}
|
2016-11-08 07:50:38 +01:00
|
|
|
block_job_start(job);
|
2016-07-27 09:01:50 +02:00
|
|
|
break;
|
|
|
|
default:
|
|
|
|
aio_context_release(aio_context);
|
|
|
|
abort();
|
|
|
|
}
|
|
|
|
|
2017-03-17 03:17:39 +01:00
|
|
|
s->stage = BLOCK_REPLICATION_RUNNING;
|
2016-07-27 09:01:50 +02:00
|
|
|
|
|
|
|
if (s->mode == REPLICATION_MODE_SECONDARY) {
|
|
|
|
secondary_do_checkpoint(s, errp);
|
|
|
|
}
|
|
|
|
|
|
|
|
s->error = 0;
|
|
|
|
aio_context_release(aio_context);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void replication_do_checkpoint(ReplicationState *rs, Error **errp)
|
|
|
|
{
|
|
|
|
BlockDriverState *bs = rs->opaque;
|
|
|
|
BDRVReplicationState *s;
|
|
|
|
AioContext *aio_context;
|
|
|
|
|
|
|
|
aio_context = bdrv_get_aio_context(bs);
|
|
|
|
aio_context_acquire(aio_context);
|
|
|
|
s = bs->opaque;
|
|
|
|
|
|
|
|
if (s->mode == REPLICATION_MODE_SECONDARY) {
|
|
|
|
secondary_do_checkpoint(s, errp);
|
|
|
|
}
|
|
|
|
aio_context_release(aio_context);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void replication_get_error(ReplicationState *rs, Error **errp)
|
|
|
|
{
|
|
|
|
BlockDriverState *bs = rs->opaque;
|
|
|
|
BDRVReplicationState *s;
|
|
|
|
AioContext *aio_context;
|
|
|
|
|
|
|
|
aio_context = bdrv_get_aio_context(bs);
|
|
|
|
aio_context_acquire(aio_context);
|
|
|
|
s = bs->opaque;
|
|
|
|
|
2017-03-17 03:17:39 +01:00
|
|
|
if (s->stage != BLOCK_REPLICATION_RUNNING) {
|
2016-07-27 09:01:50 +02:00
|
|
|
error_setg(errp, "Block replication is not running");
|
|
|
|
aio_context_release(aio_context);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (s->error) {
|
|
|
|
error_setg(errp, "I/O error occurred");
|
|
|
|
aio_context_release(aio_context);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
aio_context_release(aio_context);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void replication_done(void *opaque, int ret)
|
|
|
|
{
|
|
|
|
BlockDriverState *bs = opaque;
|
|
|
|
BDRVReplicationState *s = bs->opaque;
|
|
|
|
|
|
|
|
if (ret == 0) {
|
2017-03-17 03:17:39 +01:00
|
|
|
s->stage = BLOCK_REPLICATION_DONE;
|
2016-07-27 09:01:50 +02:00
|
|
|
|
|
|
|
/* refresh top bs's filename */
|
|
|
|
bdrv_refresh_filename(bs);
|
|
|
|
s->active_disk = NULL;
|
|
|
|
s->secondary_disk = NULL;
|
|
|
|
s->hidden_disk = NULL;
|
|
|
|
s->error = 0;
|
|
|
|
} else {
|
2017-03-17 03:17:39 +01:00
|
|
|
s->stage = BLOCK_REPLICATION_FAILOVER_FAILED;
|
2016-07-27 09:01:50 +02:00
|
|
|
s->error = -EIO;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void replication_stop(ReplicationState *rs, bool failover, Error **errp)
|
|
|
|
{
|
|
|
|
BlockDriverState *bs = rs->opaque;
|
|
|
|
BDRVReplicationState *s;
|
|
|
|
AioContext *aio_context;
|
|
|
|
|
|
|
|
aio_context = bdrv_get_aio_context(bs);
|
|
|
|
aio_context_acquire(aio_context);
|
|
|
|
s = bs->opaque;
|
|
|
|
|
2017-03-17 03:17:39 +01:00
|
|
|
if (s->stage != BLOCK_REPLICATION_RUNNING) {
|
2016-07-27 09:01:50 +02:00
|
|
|
error_setg(errp, "Block replication is not running");
|
|
|
|
aio_context_release(aio_context);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
switch (s->mode) {
|
|
|
|
case REPLICATION_MODE_PRIMARY:
|
2017-03-17 03:17:39 +01:00
|
|
|
s->stage = BLOCK_REPLICATION_DONE;
|
2016-07-27 09:01:50 +02:00
|
|
|
s->error = 0;
|
|
|
|
break;
|
|
|
|
case REPLICATION_MODE_SECONDARY:
|
|
|
|
/*
|
|
|
|
* This BDS will be closed, and the job should be completed
|
|
|
|
* before the BDS is closed, because we will access hidden
|
|
|
|
* disk, secondary disk in backup_job_completed().
|
|
|
|
*/
|
|
|
|
if (s->secondary_disk->bs->job) {
|
|
|
|
block_job_cancel_sync(s->secondary_disk->bs->job);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!failover) {
|
|
|
|
secondary_do_checkpoint(s, errp);
|
2017-03-17 03:17:39 +01:00
|
|
|
s->stage = BLOCK_REPLICATION_DONE;
|
2016-07-27 09:01:50 +02:00
|
|
|
aio_context_release(aio_context);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2017-03-17 03:17:39 +01:00
|
|
|
s->stage = BLOCK_REPLICATION_FAILOVER;
|
2016-10-27 18:06:57 +02:00
|
|
|
commit_active_start(NULL, s->active_disk->bs, s->secondary_disk->bs,
|
|
|
|
BLOCK_JOB_INTERNAL, 0, BLOCKDEV_ON_ERROR_REPORT,
|
2017-04-21 14:27:04 +02:00
|
|
|
NULL, replication_done, bs, true, errp);
|
2016-07-27 09:01:50 +02:00
|
|
|
break;
|
|
|
|
default:
|
|
|
|
aio_context_release(aio_context);
|
|
|
|
abort();
|
|
|
|
}
|
|
|
|
aio_context_release(aio_context);
|
|
|
|
}
|
|
|
|
|
|
|
|
BlockDriver bdrv_replication = {
|
|
|
|
.format_name = "replication",
|
|
|
|
.protocol_name = "replication",
|
|
|
|
.instance_size = sizeof(BDRVReplicationState),
|
|
|
|
|
|
|
|
.bdrv_open = replication_open,
|
|
|
|
.bdrv_close = replication_close,
|
2017-03-14 12:46:52 +01:00
|
|
|
.bdrv_child_perm = replication_child_perm,
|
2016-07-27 09:01:50 +02:00
|
|
|
|
|
|
|
.bdrv_getlength = replication_getlength,
|
|
|
|
.bdrv_co_readv = replication_co_readv,
|
|
|
|
.bdrv_co_writev = replication_co_writev,
|
|
|
|
|
|
|
|
.is_filter = true,
|
|
|
|
.bdrv_recurse_is_first_non_filter = replication_recurse_is_first_non_filter,
|
|
|
|
|
|
|
|
.has_variable_length = true,
|
|
|
|
};
|
|
|
|
|
|
|
|
static void bdrv_replication_init(void)
|
|
|
|
{
|
|
|
|
bdrv_register(&bdrv_replication);
|
|
|
|
}
|
|
|
|
|
|
|
|
block_init(bdrv_replication_init);
|