qed: Add return value to qed_aio_write_inplace/alloc()
Don't recurse into qed_aio_next_io() and qed_aio_complete() here, but just return an error code and let the caller handle it. Signed-off-by: Kevin Wolf <kwolf@redhat.com> Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com>
This commit is contained in:
parent
a101341aa0
commit
d6daddcdeb
43
block/qed.c
43
block/qed.c
@ -1118,7 +1118,7 @@ static bool qed_should_set_need_check(BDRVQEDState *s)
|
|||||||
*
|
*
|
||||||
* This path is taken when writing to previously unallocated clusters.
|
* This path is taken when writing to previously unallocated clusters.
|
||||||
*/
|
*/
|
||||||
static void qed_aio_write_alloc(QEDAIOCB *acb, size_t len)
|
static int qed_aio_write_alloc(QEDAIOCB *acb, size_t len)
|
||||||
{
|
{
|
||||||
BDRVQEDState *s = acb_to_s(acb);
|
BDRVQEDState *s = acb_to_s(acb);
|
||||||
int ret;
|
int ret;
|
||||||
@ -1134,7 +1134,7 @@ static void qed_aio_write_alloc(QEDAIOCB *acb, size_t len)
|
|||||||
}
|
}
|
||||||
if (acb != QSIMPLEQ_FIRST(&s->allocating_write_reqs) ||
|
if (acb != QSIMPLEQ_FIRST(&s->allocating_write_reqs) ||
|
||||||
s->allocating_write_reqs_plugged) {
|
s->allocating_write_reqs_plugged) {
|
||||||
return; /* wait for existing request to finish */
|
return -EINPROGRESS; /* wait for existing request to finish */
|
||||||
}
|
}
|
||||||
|
|
||||||
acb->cur_nclusters = qed_bytes_to_clusters(s,
|
acb->cur_nclusters = qed_bytes_to_clusters(s,
|
||||||
@ -1144,8 +1144,7 @@ static void qed_aio_write_alloc(QEDAIOCB *acb, size_t len)
|
|||||||
if (acb->flags & QED_AIOCB_ZERO) {
|
if (acb->flags & QED_AIOCB_ZERO) {
|
||||||
/* Skip ahead if the clusters are already zero */
|
/* Skip ahead if the clusters are already zero */
|
||||||
if (acb->find_cluster_ret == QED_CLUSTER_ZERO) {
|
if (acb->find_cluster_ret == QED_CLUSTER_ZERO) {
|
||||||
qed_aio_start_io(acb);
|
return 0;
|
||||||
return;
|
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
acb->cur_cluster = qed_alloc_clusters(s, acb->cur_nclusters);
|
acb->cur_cluster = qed_alloc_clusters(s, acb->cur_nclusters);
|
||||||
@ -1155,8 +1154,7 @@ static void qed_aio_write_alloc(QEDAIOCB *acb, size_t len)
|
|||||||
s->header.features |= QED_F_NEED_CHECK;
|
s->header.features |= QED_F_NEED_CHECK;
|
||||||
ret = qed_write_header(s);
|
ret = qed_write_header(s);
|
||||||
if (ret < 0) {
|
if (ret < 0) {
|
||||||
qed_aio_complete(acb, ret);
|
return ret;
|
||||||
return;
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -1166,10 +1164,9 @@ static void qed_aio_write_alloc(QEDAIOCB *acb, size_t len)
|
|||||||
ret = qed_aio_write_cow(acb);
|
ret = qed_aio_write_cow(acb);
|
||||||
}
|
}
|
||||||
if (ret < 0) {
|
if (ret < 0) {
|
||||||
qed_aio_complete(acb, ret);
|
return ret;
|
||||||
return;
|
|
||||||
}
|
}
|
||||||
qed_aio_next_io(acb, 0);
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@ -1181,10 +1178,8 @@ static void qed_aio_write_alloc(QEDAIOCB *acb, size_t len)
|
|||||||
*
|
*
|
||||||
* This path is taken when writing to already allocated clusters.
|
* This path is taken when writing to already allocated clusters.
|
||||||
*/
|
*/
|
||||||
static void qed_aio_write_inplace(QEDAIOCB *acb, uint64_t offset, size_t len)
|
static int qed_aio_write_inplace(QEDAIOCB *acb, uint64_t offset, size_t len)
|
||||||
{
|
{
|
||||||
int ret;
|
|
||||||
|
|
||||||
/* Allocate buffer for zero writes */
|
/* Allocate buffer for zero writes */
|
||||||
if (acb->flags & QED_AIOCB_ZERO) {
|
if (acb->flags & QED_AIOCB_ZERO) {
|
||||||
struct iovec *iov = acb->qiov->iov;
|
struct iovec *iov = acb->qiov->iov;
|
||||||
@ -1192,8 +1187,7 @@ static void qed_aio_write_inplace(QEDAIOCB *acb, uint64_t offset, size_t len)
|
|||||||
if (!iov->iov_base) {
|
if (!iov->iov_base) {
|
||||||
iov->iov_base = qemu_try_blockalign(acb->common.bs, iov->iov_len);
|
iov->iov_base = qemu_try_blockalign(acb->common.bs, iov->iov_len);
|
||||||
if (iov->iov_base == NULL) {
|
if (iov->iov_base == NULL) {
|
||||||
qed_aio_complete(acb, -ENOMEM);
|
return -ENOMEM;
|
||||||
return;
|
|
||||||
}
|
}
|
||||||
memset(iov->iov_base, 0, iov->iov_len);
|
memset(iov->iov_base, 0, iov->iov_len);
|
||||||
}
|
}
|
||||||
@ -1204,12 +1198,7 @@ static void qed_aio_write_inplace(QEDAIOCB *acb, uint64_t offset, size_t len)
|
|||||||
qemu_iovec_concat(&acb->cur_qiov, acb->qiov, acb->qiov_offset, len);
|
qemu_iovec_concat(&acb->cur_qiov, acb->qiov, acb->qiov_offset, len);
|
||||||
|
|
||||||
/* Do the actual write */
|
/* Do the actual write */
|
||||||
ret = qed_aio_write_main(acb);
|
return qed_aio_write_main(acb);
|
||||||
if (ret < 0) {
|
|
||||||
qed_aio_complete(acb, ret);
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
qed_aio_next_io(acb, 0);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@ -1232,19 +1221,27 @@ static void qed_aio_write_data(void *opaque, int ret,
|
|||||||
|
|
||||||
switch (ret) {
|
switch (ret) {
|
||||||
case QED_CLUSTER_FOUND:
|
case QED_CLUSTER_FOUND:
|
||||||
qed_aio_write_inplace(acb, offset, len);
|
ret = qed_aio_write_inplace(acb, offset, len);
|
||||||
break;
|
break;
|
||||||
|
|
||||||
case QED_CLUSTER_L2:
|
case QED_CLUSTER_L2:
|
||||||
case QED_CLUSTER_L1:
|
case QED_CLUSTER_L1:
|
||||||
case QED_CLUSTER_ZERO:
|
case QED_CLUSTER_ZERO:
|
||||||
qed_aio_write_alloc(acb, len);
|
ret = qed_aio_write_alloc(acb, len);
|
||||||
break;
|
break;
|
||||||
|
|
||||||
default:
|
default:
|
||||||
qed_aio_complete(acb, ret);
|
assert(ret < 0);
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (ret < 0) {
|
||||||
|
if (ret != -EINPROGRESS) {
|
||||||
|
qed_aio_complete(acb, ret);
|
||||||
|
}
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
qed_aio_next_io(acb, 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
Loading…
Reference in New Issue
Block a user