-----BEGIN PGP SIGNATURE-----
Version: GnuPG v1 iQEcBAABAgAGBQJVevNrAAoJEJykq7OBq3PIn3wH/iTkoxK+8c6vHY4gAyW/3Vel JAWyDULliLg/mJiyDyz+9hbyb3FmDsQ7Gve5laFCqSES8IChe/fAoR2dXAapHbJm VroBii+C4MIakAjB1eSGVTMeJVgyq0SxCbEgo6MSgCNRg8pH/ne4kGMn0elsyMQm VrxqsoW1qda0gLbybu7uRKmdZlb5Eklk0zQmBlzd25j3xgMXmo7ZcXW0d7dCi/xd ot01aPNHWKy2FyYsI+N1SMagCedjn0FFCHMMc08W34olUOAschV+Croxi6gupbTl RdlYNnB6BBYLgT7mxYuKKd20r4Q8Si3E5YjIsYaXIskbCOxDuryKboyheMXWqdA= =mx5x -----END PGP SIGNATURE----- Merge remote-tracking branch 'remotes/stefanha/tags/block-pull-request' into staging # gpg: Signature made Fri Jun 12 15:57:47 2015 BST using RSA key ID 81AB73C8 # gpg: Good signature from "Stefan Hajnoczi <stefanha@redhat.com>" # gpg: aka "Stefan Hajnoczi <stefanha@gmail.com>" * remotes/stefanha/tags/block-pull-request: qemu-iotests: expand test 093 to support group throttling throttle: Update throttle infrastructure copyright throttle: add the name of the ThrottleGroup to BlockDeviceInfo throttle: acquire the ThrottleGroup lock in bdrv_swap() throttle: Add throttle group support throttle: Add throttle group infrastructure tests throttle: Add throttle group infrastructure throttle: Extract timers from ThrottleState into a separate structure raw-posix: Fix .bdrv_co_get_block_status() for unaligned image size Revert "iothread: release iothread around aio_poll" Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
This commit is contained in:
commit
8aeaa055f5
8
async.c
8
async.c
|
@ -280,6 +280,12 @@ static void aio_timerlist_notify(void *opaque)
|
||||||
aio_notify(opaque);
|
aio_notify(opaque);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void aio_rfifolock_cb(void *opaque)
|
||||||
|
{
|
||||||
|
/* Kick owner thread in case they are blocked in aio_poll() */
|
||||||
|
aio_notify(opaque);
|
||||||
|
}
|
||||||
|
|
||||||
AioContext *aio_context_new(Error **errp)
|
AioContext *aio_context_new(Error **errp)
|
||||||
{
|
{
|
||||||
int ret;
|
int ret;
|
||||||
|
@ -297,7 +303,7 @@ AioContext *aio_context_new(Error **errp)
|
||||||
event_notifier_test_and_clear);
|
event_notifier_test_and_clear);
|
||||||
ctx->thread_pool = NULL;
|
ctx->thread_pool = NULL;
|
||||||
qemu_mutex_init(&ctx->bh_lock);
|
qemu_mutex_init(&ctx->bh_lock);
|
||||||
rfifolock_init(&ctx->lock, NULL, NULL);
|
rfifolock_init(&ctx->lock, aio_rfifolock_cb, ctx);
|
||||||
timerlistgroup_init(&ctx->tlg, aio_timerlist_notify, ctx);
|
timerlistgroup_init(&ctx->tlg, aio_timerlist_notify, ctx);
|
||||||
|
|
||||||
return ctx;
|
return ctx;
|
||||||
|
|
38
block.c
38
block.c
|
@ -36,6 +36,7 @@
|
||||||
#include "qmp-commands.h"
|
#include "qmp-commands.h"
|
||||||
#include "qemu/timer.h"
|
#include "qemu/timer.h"
|
||||||
#include "qapi-event.h"
|
#include "qapi-event.h"
|
||||||
|
#include "block/throttle-groups.h"
|
||||||
|
|
||||||
#ifdef CONFIG_BSD
|
#ifdef CONFIG_BSD
|
||||||
#include <sys/types.h>
|
#include <sys/types.h>
|
||||||
|
@ -1822,12 +1823,18 @@ static void bdrv_move_feature_fields(BlockDriverState *bs_dest,
|
||||||
bs_dest->enable_write_cache = bs_src->enable_write_cache;
|
bs_dest->enable_write_cache = bs_src->enable_write_cache;
|
||||||
|
|
||||||
/* i/o throttled req */
|
/* i/o throttled req */
|
||||||
memcpy(&bs_dest->throttle_state,
|
bs_dest->throttle_state = bs_src->throttle_state,
|
||||||
&bs_src->throttle_state,
|
bs_dest->io_limits_enabled = bs_src->io_limits_enabled;
|
||||||
sizeof(ThrottleState));
|
bs_dest->pending_reqs[0] = bs_src->pending_reqs[0];
|
||||||
|
bs_dest->pending_reqs[1] = bs_src->pending_reqs[1];
|
||||||
bs_dest->throttled_reqs[0] = bs_src->throttled_reqs[0];
|
bs_dest->throttled_reqs[0] = bs_src->throttled_reqs[0];
|
||||||
bs_dest->throttled_reqs[1] = bs_src->throttled_reqs[1];
|
bs_dest->throttled_reqs[1] = bs_src->throttled_reqs[1];
|
||||||
bs_dest->io_limits_enabled = bs_src->io_limits_enabled;
|
memcpy(&bs_dest->round_robin,
|
||||||
|
&bs_src->round_robin,
|
||||||
|
sizeof(bs_dest->round_robin));
|
||||||
|
memcpy(&bs_dest->throttle_timers,
|
||||||
|
&bs_src->throttle_timers,
|
||||||
|
sizeof(ThrottleTimers));
|
||||||
|
|
||||||
/* r/w error */
|
/* r/w error */
|
||||||
bs_dest->on_read_error = bs_src->on_read_error;
|
bs_dest->on_read_error = bs_src->on_read_error;
|
||||||
|
@ -1881,12 +1888,21 @@ void bdrv_swap(BlockDriverState *bs_new, BlockDriverState *bs_old)
|
||||||
QTAILQ_REMOVE(&graph_bdrv_states, bs_old, node_list);
|
QTAILQ_REMOVE(&graph_bdrv_states, bs_old, node_list);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* If the BlockDriverState is part of a throttling group acquire
|
||||||
|
* its lock since we're going to mess with the protected fields.
|
||||||
|
* Otherwise there's no need to worry since no one else can touch
|
||||||
|
* them. */
|
||||||
|
if (bs_old->throttle_state) {
|
||||||
|
throttle_group_lock(bs_old);
|
||||||
|
}
|
||||||
|
|
||||||
/* bs_new must be unattached and shouldn't have anything fancy enabled */
|
/* bs_new must be unattached and shouldn't have anything fancy enabled */
|
||||||
assert(!bs_new->blk);
|
assert(!bs_new->blk);
|
||||||
assert(QLIST_EMPTY(&bs_new->dirty_bitmaps));
|
assert(QLIST_EMPTY(&bs_new->dirty_bitmaps));
|
||||||
assert(bs_new->job == NULL);
|
assert(bs_new->job == NULL);
|
||||||
assert(bs_new->io_limits_enabled == false);
|
assert(bs_new->io_limits_enabled == false);
|
||||||
assert(!throttle_have_timer(&bs_new->throttle_state));
|
assert(bs_new->throttle_state == NULL);
|
||||||
|
assert(!throttle_timers_are_initialized(&bs_new->throttle_timers));
|
||||||
|
|
||||||
tmp = *bs_new;
|
tmp = *bs_new;
|
||||||
*bs_new = *bs_old;
|
*bs_new = *bs_old;
|
||||||
|
@ -1903,7 +1919,13 @@ void bdrv_swap(BlockDriverState *bs_new, BlockDriverState *bs_old)
|
||||||
/* Check a few fields that should remain attached to the device */
|
/* Check a few fields that should remain attached to the device */
|
||||||
assert(bs_new->job == NULL);
|
assert(bs_new->job == NULL);
|
||||||
assert(bs_new->io_limits_enabled == false);
|
assert(bs_new->io_limits_enabled == false);
|
||||||
assert(!throttle_have_timer(&bs_new->throttle_state));
|
assert(bs_new->throttle_state == NULL);
|
||||||
|
assert(!throttle_timers_are_initialized(&bs_new->throttle_timers));
|
||||||
|
|
||||||
|
/* Release the ThrottleGroup lock */
|
||||||
|
if (bs_old->throttle_state) {
|
||||||
|
throttle_group_unlock(bs_old);
|
||||||
|
}
|
||||||
|
|
||||||
/* insert the nodes back into the graph node list if needed */
|
/* insert the nodes back into the graph node list if needed */
|
||||||
if (bs_new->node_name[0] != '\0') {
|
if (bs_new->node_name[0] != '\0') {
|
||||||
|
@ -3691,7 +3713,7 @@ void bdrv_detach_aio_context(BlockDriverState *bs)
|
||||||
}
|
}
|
||||||
|
|
||||||
if (bs->io_limits_enabled) {
|
if (bs->io_limits_enabled) {
|
||||||
throttle_detach_aio_context(&bs->throttle_state);
|
throttle_timers_detach_aio_context(&bs->throttle_timers);
|
||||||
}
|
}
|
||||||
if (bs->drv->bdrv_detach_aio_context) {
|
if (bs->drv->bdrv_detach_aio_context) {
|
||||||
bs->drv->bdrv_detach_aio_context(bs);
|
bs->drv->bdrv_detach_aio_context(bs);
|
||||||
|
@ -3727,7 +3749,7 @@ void bdrv_attach_aio_context(BlockDriverState *bs,
|
||||||
bs->drv->bdrv_attach_aio_context(bs, new_context);
|
bs->drv->bdrv_attach_aio_context(bs, new_context);
|
||||||
}
|
}
|
||||||
if (bs->io_limits_enabled) {
|
if (bs->io_limits_enabled) {
|
||||||
throttle_attach_aio_context(&bs->throttle_state, new_context);
|
throttle_timers_attach_aio_context(&bs->throttle_timers, new_context);
|
||||||
}
|
}
|
||||||
|
|
||||||
QLIST_FOREACH(ban, &bs->aio_notifiers, list) {
|
QLIST_FOREACH(ban, &bs->aio_notifiers, list) {
|
||||||
|
|
|
@ -10,6 +10,7 @@ block-obj-$(CONFIG_WIN32) += raw-win32.o win32-aio.o
|
||||||
block-obj-$(CONFIG_POSIX) += raw-posix.o
|
block-obj-$(CONFIG_POSIX) += raw-posix.o
|
||||||
block-obj-$(CONFIG_LINUX_AIO) += linux-aio.o
|
block-obj-$(CONFIG_LINUX_AIO) += linux-aio.o
|
||||||
block-obj-y += null.o mirror.o io.o
|
block-obj-y += null.o mirror.o io.o
|
||||||
|
block-obj-y += throttle-groups.o
|
||||||
|
|
||||||
block-obj-y += nbd.o nbd-client.o sheepdog.o
|
block-obj-y += nbd.o nbd-client.o sheepdog.o
|
||||||
block-obj-$(CONFIG_LIBISCSI) += iscsi.o
|
block-obj-$(CONFIG_LIBISCSI) += iscsi.o
|
||||||
|
|
75
block/io.c
75
block/io.c
|
@ -23,9 +23,9 @@
|
||||||
*/
|
*/
|
||||||
|
|
||||||
#include "trace.h"
|
#include "trace.h"
|
||||||
#include "sysemu/qtest.h"
|
|
||||||
#include "block/blockjob.h"
|
#include "block/blockjob.h"
|
||||||
#include "block/block_int.h"
|
#include "block/block_int.h"
|
||||||
|
#include "block/throttle-groups.h"
|
||||||
|
|
||||||
#define NOT_DONE 0x7fffffff /* used while emulated sync operation in progress */
|
#define NOT_DONE 0x7fffffff /* used while emulated sync operation in progress */
|
||||||
|
|
||||||
|
@ -65,7 +65,7 @@ void bdrv_set_io_limits(BlockDriverState *bs,
|
||||||
{
|
{
|
||||||
int i;
|
int i;
|
||||||
|
|
||||||
throttle_config(&bs->throttle_state, cfg);
|
throttle_group_config(bs, cfg);
|
||||||
|
|
||||||
for (i = 0; i < 2; i++) {
|
for (i = 0; i < 2; i++) {
|
||||||
qemu_co_enter_next(&bs->throttled_reqs[i]);
|
qemu_co_enter_next(&bs->throttled_reqs[i]);
|
||||||
|
@ -95,72 +95,33 @@ static bool bdrv_start_throttled_reqs(BlockDriverState *bs)
|
||||||
void bdrv_io_limits_disable(BlockDriverState *bs)
|
void bdrv_io_limits_disable(BlockDriverState *bs)
|
||||||
{
|
{
|
||||||
bs->io_limits_enabled = false;
|
bs->io_limits_enabled = false;
|
||||||
|
|
||||||
bdrv_start_throttled_reqs(bs);
|
bdrv_start_throttled_reqs(bs);
|
||||||
|
throttle_group_unregister_bs(bs);
|
||||||
throttle_destroy(&bs->throttle_state);
|
|
||||||
}
|
|
||||||
|
|
||||||
static void bdrv_throttle_read_timer_cb(void *opaque)
|
|
||||||
{
|
|
||||||
BlockDriverState *bs = opaque;
|
|
||||||
qemu_co_enter_next(&bs->throttled_reqs[0]);
|
|
||||||
}
|
|
||||||
|
|
||||||
static void bdrv_throttle_write_timer_cb(void *opaque)
|
|
||||||
{
|
|
||||||
BlockDriverState *bs = opaque;
|
|
||||||
qemu_co_enter_next(&bs->throttled_reqs[1]);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/* should be called before bdrv_set_io_limits if a limit is set */
|
/* should be called before bdrv_set_io_limits if a limit is set */
|
||||||
void bdrv_io_limits_enable(BlockDriverState *bs)
|
void bdrv_io_limits_enable(BlockDriverState *bs, const char *group)
|
||||||
{
|
{
|
||||||
int clock_type = QEMU_CLOCK_REALTIME;
|
|
||||||
|
|
||||||
if (qtest_enabled()) {
|
|
||||||
/* For testing block IO throttling only */
|
|
||||||
clock_type = QEMU_CLOCK_VIRTUAL;
|
|
||||||
}
|
|
||||||
assert(!bs->io_limits_enabled);
|
assert(!bs->io_limits_enabled);
|
||||||
throttle_init(&bs->throttle_state,
|
throttle_group_register_bs(bs, group);
|
||||||
bdrv_get_aio_context(bs),
|
|
||||||
clock_type,
|
|
||||||
bdrv_throttle_read_timer_cb,
|
|
||||||
bdrv_throttle_write_timer_cb,
|
|
||||||
bs);
|
|
||||||
bs->io_limits_enabled = true;
|
bs->io_limits_enabled = true;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* This function makes an IO wait if needed
|
void bdrv_io_limits_update_group(BlockDriverState *bs, const char *group)
|
||||||
*
|
|
||||||
* @nb_sectors: the number of sectors of the IO
|
|
||||||
* @is_write: is the IO a write
|
|
||||||
*/
|
|
||||||
static void bdrv_io_limits_intercept(BlockDriverState *bs,
|
|
||||||
unsigned int bytes,
|
|
||||||
bool is_write)
|
|
||||||
{
|
{
|
||||||
/* does this io must wait */
|
/* this bs is not part of any group */
|
||||||
bool must_wait = throttle_schedule_timer(&bs->throttle_state, is_write);
|
if (!bs->throttle_state) {
|
||||||
|
|
||||||
/* if must wait or any request of this type throttled queue the IO */
|
|
||||||
if (must_wait ||
|
|
||||||
!qemu_co_queue_empty(&bs->throttled_reqs[is_write])) {
|
|
||||||
qemu_co_queue_wait(&bs->throttled_reqs[is_write]);
|
|
||||||
}
|
|
||||||
|
|
||||||
/* the IO will be executed, do the accounting */
|
|
||||||
throttle_account(&bs->throttle_state, is_write, bytes);
|
|
||||||
|
|
||||||
|
|
||||||
/* if the next request must wait -> do nothing */
|
|
||||||
if (throttle_schedule_timer(&bs->throttle_state, is_write)) {
|
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* else queue next request for execution */
|
/* this bs is a part of the same group than the one we want */
|
||||||
qemu_co_queue_next(&bs->throttled_reqs[is_write]);
|
if (!g_strcmp0(throttle_group_get_name(bs), group)) {
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* need to change the group this bs belong to */
|
||||||
|
bdrv_io_limits_disable(bs);
|
||||||
|
bdrv_io_limits_enable(bs, group);
|
||||||
}
|
}
|
||||||
|
|
||||||
void bdrv_setup_io_funcs(BlockDriver *bdrv)
|
void bdrv_setup_io_funcs(BlockDriver *bdrv)
|
||||||
|
@ -967,7 +928,7 @@ static int coroutine_fn bdrv_co_do_preadv(BlockDriverState *bs,
|
||||||
|
|
||||||
/* throttling disk I/O */
|
/* throttling disk I/O */
|
||||||
if (bs->io_limits_enabled) {
|
if (bs->io_limits_enabled) {
|
||||||
bdrv_io_limits_intercept(bs, bytes, false);
|
throttle_group_co_io_limits_intercept(bs, bytes, false);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Align read if necessary by padding qiov */
|
/* Align read if necessary by padding qiov */
|
||||||
|
@ -1297,7 +1258,7 @@ static int coroutine_fn bdrv_co_do_pwritev(BlockDriverState *bs,
|
||||||
|
|
||||||
/* throttling disk I/O */
|
/* throttling disk I/O */
|
||||||
if (bs->io_limits_enabled) {
|
if (bs->io_limits_enabled) {
|
||||||
bdrv_io_limits_intercept(bs, bytes, true);
|
throttle_group_co_io_limits_intercept(bs, bytes, true);
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
|
|
@ -24,6 +24,7 @@
|
||||||
|
|
||||||
#include "block/qapi.h"
|
#include "block/qapi.h"
|
||||||
#include "block/block_int.h"
|
#include "block/block_int.h"
|
||||||
|
#include "block/throttle-groups.h"
|
||||||
#include "block/write-threshold.h"
|
#include "block/write-threshold.h"
|
||||||
#include "qmp-commands.h"
|
#include "qmp-commands.h"
|
||||||
#include "qapi-visit.h"
|
#include "qapi-visit.h"
|
||||||
|
@ -65,7 +66,9 @@ BlockDeviceInfo *bdrv_block_device_info(BlockDriverState *bs, Error **errp)
|
||||||
|
|
||||||
if (bs->io_limits_enabled) {
|
if (bs->io_limits_enabled) {
|
||||||
ThrottleConfig cfg;
|
ThrottleConfig cfg;
|
||||||
throttle_get_config(&bs->throttle_state, &cfg);
|
|
||||||
|
throttle_group_get_config(bs, &cfg);
|
||||||
|
|
||||||
info->bps = cfg.buckets[THROTTLE_BPS_TOTAL].avg;
|
info->bps = cfg.buckets[THROTTLE_BPS_TOTAL].avg;
|
||||||
info->bps_rd = cfg.buckets[THROTTLE_BPS_READ].avg;
|
info->bps_rd = cfg.buckets[THROTTLE_BPS_READ].avg;
|
||||||
info->bps_wr = cfg.buckets[THROTTLE_BPS_WRITE].avg;
|
info->bps_wr = cfg.buckets[THROTTLE_BPS_WRITE].avg;
|
||||||
|
@ -90,6 +93,9 @@ BlockDeviceInfo *bdrv_block_device_info(BlockDriverState *bs, Error **errp)
|
||||||
|
|
||||||
info->has_iops_size = cfg.op_size;
|
info->has_iops_size = cfg.op_size;
|
||||||
info->iops_size = cfg.op_size;
|
info->iops_size = cfg.op_size;
|
||||||
|
|
||||||
|
info->has_group = true;
|
||||||
|
info->group = g_strdup(throttle_group_get_name(bs));
|
||||||
}
|
}
|
||||||
|
|
||||||
info->write_threshold = bdrv_write_threshold_get(bs);
|
info->write_threshold = bdrv_write_threshold_get(bs);
|
||||||
|
|
|
@ -1848,8 +1848,9 @@ static int64_t coroutine_fn raw_co_get_block_status(BlockDriverState *bs,
|
||||||
*pnum = nb_sectors;
|
*pnum = nb_sectors;
|
||||||
ret = BDRV_BLOCK_DATA;
|
ret = BDRV_BLOCK_DATA;
|
||||||
} else if (data == start) {
|
} else if (data == start) {
|
||||||
/* On a data extent, compute sectors to the end of the extent. */
|
/* On a data extent, compute sectors to the end of the extent,
|
||||||
*pnum = MIN(nb_sectors, (hole - start) / BDRV_SECTOR_SIZE);
|
* possibly including a partial sector at EOF. */
|
||||||
|
*pnum = MIN(nb_sectors, DIV_ROUND_UP(hole - start, BDRV_SECTOR_SIZE));
|
||||||
ret = BDRV_BLOCK_DATA;
|
ret = BDRV_BLOCK_DATA;
|
||||||
} else {
|
} else {
|
||||||
/* On a hole, compute sectors to the beginning of the next extent. */
|
/* On a hole, compute sectors to the beginning of the next extent. */
|
||||||
|
|
|
@ -0,0 +1,496 @@
|
||||||
|
/*
|
||||||
|
* QEMU block throttling group infrastructure
|
||||||
|
*
|
||||||
|
* Copyright (C) Nodalink, EURL. 2014
|
||||||
|
* Copyright (C) Igalia, S.L. 2015
|
||||||
|
*
|
||||||
|
* Authors:
|
||||||
|
* Benoît Canet <benoit.canet@nodalink.com>
|
||||||
|
* Alberto Garcia <berto@igalia.com>
|
||||||
|
*
|
||||||
|
* This program is free software; you can redistribute it and/or
|
||||||
|
* modify it under the terms of the GNU General Public License as
|
||||||
|
* published by the Free Software Foundation; either version 2 or
|
||||||
|
* (at your option) version 3 of the License.
|
||||||
|
*
|
||||||
|
* This program is distributed in the hope that it will be useful,
|
||||||
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
* GNU General Public License for more details.
|
||||||
|
*
|
||||||
|
* You should have received a copy of the GNU General Public License
|
||||||
|
* along with this program; if not, see <http://www.gnu.org/licenses/>.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#include "block/throttle-groups.h"
|
||||||
|
#include "qemu/queue.h"
|
||||||
|
#include "qemu/thread.h"
|
||||||
|
#include "sysemu/qtest.h"
|
||||||
|
|
||||||
|
/* The ThrottleGroup structure (with its ThrottleState) is shared
|
||||||
|
* among different BlockDriverState and it's independent from
|
||||||
|
* AioContext, so in order to use it from different threads it needs
|
||||||
|
* its own locking.
|
||||||
|
*
|
||||||
|
* This locking is however handled internally in this file, so it's
|
||||||
|
* mostly transparent to outside users (but see the documentation in
|
||||||
|
* throttle_groups_lock()).
|
||||||
|
*
|
||||||
|
* The whole ThrottleGroup structure is private and invisible to
|
||||||
|
* outside users, that only use it through its ThrottleState.
|
||||||
|
*
|
||||||
|
* In addition to the ThrottleGroup structure, BlockDriverState has
|
||||||
|
* fields that need to be accessed by other members of the group and
|
||||||
|
* therefore also need to be protected by this lock. Once a BDS is
|
||||||
|
* registered in a group those fields can be accessed by other threads
|
||||||
|
* any time.
|
||||||
|
*
|
||||||
|
* Again, all this is handled internally and is mostly transparent to
|
||||||
|
* the outside. The 'throttle_timers' field however has an additional
|
||||||
|
* constraint because it may be temporarily invalid (see for example
|
||||||
|
* bdrv_set_aio_context()). Therefore in this file a thread will
|
||||||
|
* access some other BDS's timers only after verifying that that BDS
|
||||||
|
* has throttled requests in the queue.
|
||||||
|
*/
|
||||||
|
typedef struct ThrottleGroup {
|
||||||
|
char *name; /* This is constant during the lifetime of the group */
|
||||||
|
|
||||||
|
QemuMutex lock; /* This lock protects the following four fields */
|
||||||
|
ThrottleState ts;
|
||||||
|
QLIST_HEAD(, BlockDriverState) head;
|
||||||
|
BlockDriverState *tokens[2];
|
||||||
|
bool any_timer_armed[2];
|
||||||
|
|
||||||
|
/* These two are protected by the global throttle_groups_lock */
|
||||||
|
unsigned refcount;
|
||||||
|
QTAILQ_ENTRY(ThrottleGroup) list;
|
||||||
|
} ThrottleGroup;
|
||||||
|
|
||||||
|
static QemuMutex throttle_groups_lock;
|
||||||
|
static QTAILQ_HEAD(, ThrottleGroup) throttle_groups =
|
||||||
|
QTAILQ_HEAD_INITIALIZER(throttle_groups);
|
||||||
|
|
||||||
|
/* Increments the reference count of a ThrottleGroup given its name.
|
||||||
|
*
|
||||||
|
* If no ThrottleGroup is found with the given name a new one is
|
||||||
|
* created.
|
||||||
|
*
|
||||||
|
* @name: the name of the ThrottleGroup
|
||||||
|
* @ret: the ThrottleGroup
|
||||||
|
*/
|
||||||
|
static ThrottleGroup *throttle_group_incref(const char *name)
|
||||||
|
{
|
||||||
|
ThrottleGroup *tg = NULL;
|
||||||
|
ThrottleGroup *iter;
|
||||||
|
|
||||||
|
qemu_mutex_lock(&throttle_groups_lock);
|
||||||
|
|
||||||
|
/* Look for an existing group with that name */
|
||||||
|
QTAILQ_FOREACH(iter, &throttle_groups, list) {
|
||||||
|
if (!strcmp(name, iter->name)) {
|
||||||
|
tg = iter;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Create a new one if not found */
|
||||||
|
if (!tg) {
|
||||||
|
tg = g_new0(ThrottleGroup, 1);
|
||||||
|
tg->name = g_strdup(name);
|
||||||
|
qemu_mutex_init(&tg->lock);
|
||||||
|
throttle_init(&tg->ts);
|
||||||
|
QLIST_INIT(&tg->head);
|
||||||
|
|
||||||
|
QTAILQ_INSERT_TAIL(&throttle_groups, tg, list);
|
||||||
|
}
|
||||||
|
|
||||||
|
tg->refcount++;
|
||||||
|
|
||||||
|
qemu_mutex_unlock(&throttle_groups_lock);
|
||||||
|
|
||||||
|
return tg;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Decrease the reference count of a ThrottleGroup.
|
||||||
|
*
|
||||||
|
* When the reference count reaches zero the ThrottleGroup is
|
||||||
|
* destroyed.
|
||||||
|
*
|
||||||
|
* @tg: The ThrottleGroup to unref
|
||||||
|
*/
|
||||||
|
static void throttle_group_unref(ThrottleGroup *tg)
|
||||||
|
{
|
||||||
|
qemu_mutex_lock(&throttle_groups_lock);
|
||||||
|
if (--tg->refcount == 0) {
|
||||||
|
QTAILQ_REMOVE(&throttle_groups, tg, list);
|
||||||
|
qemu_mutex_destroy(&tg->lock);
|
||||||
|
g_free(tg->name);
|
||||||
|
g_free(tg);
|
||||||
|
}
|
||||||
|
qemu_mutex_unlock(&throttle_groups_lock);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Get the name from a BlockDriverState's ThrottleGroup. The name (and
|
||||||
|
* the pointer) is guaranteed to remain constant during the lifetime
|
||||||
|
* of the group.
|
||||||
|
*
|
||||||
|
* @bs: a BlockDriverState that is member of a throttling group
|
||||||
|
* @ret: the name of the group.
|
||||||
|
*/
|
||||||
|
const char *throttle_group_get_name(BlockDriverState *bs)
|
||||||
|
{
|
||||||
|
ThrottleGroup *tg = container_of(bs->throttle_state, ThrottleGroup, ts);
|
||||||
|
return tg->name;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Return the next BlockDriverState in the round-robin sequence,
|
||||||
|
* simulating a circular list.
|
||||||
|
*
|
||||||
|
* This assumes that tg->lock is held.
|
||||||
|
*
|
||||||
|
* @bs: the current BlockDriverState
|
||||||
|
* @ret: the next BlockDriverState in the sequence
|
||||||
|
*/
|
||||||
|
static BlockDriverState *throttle_group_next_bs(BlockDriverState *bs)
|
||||||
|
{
|
||||||
|
ThrottleState *ts = bs->throttle_state;
|
||||||
|
ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts);
|
||||||
|
BlockDriverState *next = QLIST_NEXT(bs, round_robin);
|
||||||
|
|
||||||
|
if (!next) {
|
||||||
|
return QLIST_FIRST(&tg->head);
|
||||||
|
}
|
||||||
|
|
||||||
|
return next;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Return the next BlockDriverState in the round-robin sequence with
|
||||||
|
* pending I/O requests.
|
||||||
|
*
|
||||||
|
* This assumes that tg->lock is held.
|
||||||
|
*
|
||||||
|
* @bs: the current BlockDriverState
|
||||||
|
* @is_write: the type of operation (read/write)
|
||||||
|
* @ret: the next BlockDriverState with pending requests, or bs
|
||||||
|
* if there is none.
|
||||||
|
*/
|
||||||
|
static BlockDriverState *next_throttle_token(BlockDriverState *bs,
|
||||||
|
bool is_write)
|
||||||
|
{
|
||||||
|
ThrottleGroup *tg = container_of(bs->throttle_state, ThrottleGroup, ts);
|
||||||
|
BlockDriverState *token, *start;
|
||||||
|
|
||||||
|
start = token = tg->tokens[is_write];
|
||||||
|
|
||||||
|
/* get next bs round in round robin style */
|
||||||
|
token = throttle_group_next_bs(token);
|
||||||
|
while (token != start && !token->pending_reqs[is_write]) {
|
||||||
|
token = throttle_group_next_bs(token);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* If no IO are queued for scheduling on the next round robin token
|
||||||
|
* then decide the token is the current bs because chances are
|
||||||
|
* the current bs get the current request queued.
|
||||||
|
*/
|
||||||
|
if (token == start && !token->pending_reqs[is_write]) {
|
||||||
|
token = bs;
|
||||||
|
}
|
||||||
|
|
||||||
|
return token;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Check if the next I/O request for a BlockDriverState needs to be
|
||||||
|
* throttled or not. If there's no timer set in this group, set one
|
||||||
|
* and update the token accordingly.
|
||||||
|
*
|
||||||
|
* This assumes that tg->lock is held.
|
||||||
|
*
|
||||||
|
* @bs: the current BlockDriverState
|
||||||
|
* @is_write: the type of operation (read/write)
|
||||||
|
* @ret: whether the I/O request needs to be throttled or not
|
||||||
|
*/
|
||||||
|
static bool throttle_group_schedule_timer(BlockDriverState *bs,
|
||||||
|
bool is_write)
|
||||||
|
{
|
||||||
|
ThrottleState *ts = bs->throttle_state;
|
||||||
|
ThrottleTimers *tt = &bs->throttle_timers;
|
||||||
|
ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts);
|
||||||
|
bool must_wait;
|
||||||
|
|
||||||
|
/* Check if any of the timers in this group is already armed */
|
||||||
|
if (tg->any_timer_armed[is_write]) {
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
must_wait = throttle_schedule_timer(ts, tt, is_write);
|
||||||
|
|
||||||
|
/* If a timer just got armed, set bs as the current token */
|
||||||
|
if (must_wait) {
|
||||||
|
tg->tokens[is_write] = bs;
|
||||||
|
tg->any_timer_armed[is_write] = true;
|
||||||
|
}
|
||||||
|
|
||||||
|
return must_wait;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Look for the next pending I/O request and schedule it.
|
||||||
|
*
|
||||||
|
* This assumes that tg->lock is held.
|
||||||
|
*
|
||||||
|
* @bs: the current BlockDriverState
|
||||||
|
* @is_write: the type of operation (read/write)
|
||||||
|
*/
|
||||||
|
static void schedule_next_request(BlockDriverState *bs, bool is_write)
|
||||||
|
{
|
||||||
|
ThrottleGroup *tg = container_of(bs->throttle_state, ThrottleGroup, ts);
|
||||||
|
bool must_wait;
|
||||||
|
BlockDriverState *token;
|
||||||
|
|
||||||
|
/* Check if there's any pending request to schedule next */
|
||||||
|
token = next_throttle_token(bs, is_write);
|
||||||
|
if (!token->pending_reqs[is_write]) {
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Set a timer for the request if it needs to be throttled */
|
||||||
|
must_wait = throttle_group_schedule_timer(token, is_write);
|
||||||
|
|
||||||
|
/* If it doesn't have to wait, queue it for immediate execution */
|
||||||
|
if (!must_wait) {
|
||||||
|
/* Give preference to requests from the current bs */
|
||||||
|
if (qemu_in_coroutine() &&
|
||||||
|
qemu_co_queue_next(&bs->throttled_reqs[is_write])) {
|
||||||
|
token = bs;
|
||||||
|
} else {
|
||||||
|
ThrottleTimers *tt = &token->throttle_timers;
|
||||||
|
int64_t now = qemu_clock_get_ns(tt->clock_type);
|
||||||
|
timer_mod(tt->timers[is_write], now + 1);
|
||||||
|
tg->any_timer_armed[is_write] = true;
|
||||||
|
}
|
||||||
|
tg->tokens[is_write] = token;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Check if an I/O request needs to be throttled, wait and set a timer
|
||||||
|
* if necessary, and schedule the next request using a round robin
|
||||||
|
* algorithm.
|
||||||
|
*
|
||||||
|
* @bs: the current BlockDriverState
|
||||||
|
* @bytes: the number of bytes for this I/O
|
||||||
|
* @is_write: the type of operation (read/write)
|
||||||
|
*/
|
||||||
|
void coroutine_fn throttle_group_co_io_limits_intercept(BlockDriverState *bs,
|
||||||
|
unsigned int bytes,
|
||||||
|
bool is_write)
|
||||||
|
{
|
||||||
|
bool must_wait;
|
||||||
|
BlockDriverState *token;
|
||||||
|
|
||||||
|
ThrottleGroup *tg = container_of(bs->throttle_state, ThrottleGroup, ts);
|
||||||
|
qemu_mutex_lock(&tg->lock);
|
||||||
|
|
||||||
|
/* First we check if this I/O has to be throttled. */
|
||||||
|
token = next_throttle_token(bs, is_write);
|
||||||
|
must_wait = throttle_group_schedule_timer(token, is_write);
|
||||||
|
|
||||||
|
/* Wait if there's a timer set or queued requests of this type */
|
||||||
|
if (must_wait || bs->pending_reqs[is_write]) {
|
||||||
|
bs->pending_reqs[is_write]++;
|
||||||
|
qemu_mutex_unlock(&tg->lock);
|
||||||
|
qemu_co_queue_wait(&bs->throttled_reqs[is_write]);
|
||||||
|
qemu_mutex_lock(&tg->lock);
|
||||||
|
bs->pending_reqs[is_write]--;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* The I/O will be executed, so do the accounting */
|
||||||
|
throttle_account(bs->throttle_state, is_write, bytes);
|
||||||
|
|
||||||
|
/* Schedule the next request */
|
||||||
|
schedule_next_request(bs, is_write);
|
||||||
|
|
||||||
|
qemu_mutex_unlock(&tg->lock);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Update the throttle configuration for a particular group. Similar
|
||||||
|
* to throttle_config(), but guarantees atomicity within the
|
||||||
|
* throttling group.
|
||||||
|
*
|
||||||
|
* @bs: a BlockDriverState that is member of the group
|
||||||
|
* @cfg: the configuration to set
|
||||||
|
*/
|
||||||
|
void throttle_group_config(BlockDriverState *bs, ThrottleConfig *cfg)
|
||||||
|
{
|
||||||
|
ThrottleTimers *tt = &bs->throttle_timers;
|
||||||
|
ThrottleState *ts = bs->throttle_state;
|
||||||
|
ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts);
|
||||||
|
qemu_mutex_lock(&tg->lock);
|
||||||
|
throttle_config(ts, tt, cfg);
|
||||||
|
/* throttle_config() cancels the timers */
|
||||||
|
tg->any_timer_armed[0] = tg->any_timer_armed[1] = false;
|
||||||
|
qemu_mutex_unlock(&tg->lock);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Get the throttle configuration from a particular group. Similar to
|
||||||
|
* throttle_get_config(), but guarantees atomicity within the
|
||||||
|
* throttling group.
|
||||||
|
*
|
||||||
|
* @bs: a BlockDriverState that is member of the group
|
||||||
|
* @cfg: the configuration will be written here
|
||||||
|
*/
|
||||||
|
void throttle_group_get_config(BlockDriverState *bs, ThrottleConfig *cfg)
|
||||||
|
{
|
||||||
|
ThrottleState *ts = bs->throttle_state;
|
||||||
|
ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts);
|
||||||
|
qemu_mutex_lock(&tg->lock);
|
||||||
|
throttle_get_config(ts, cfg);
|
||||||
|
qemu_mutex_unlock(&tg->lock);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* ThrottleTimers callback. This wakes up a request that was waiting
|
||||||
|
* because it had been throttled.
|
||||||
|
*
|
||||||
|
* @bs: the BlockDriverState whose request had been throttled
|
||||||
|
* @is_write: the type of operation (read/write)
|
||||||
|
*/
|
||||||
|
static void timer_cb(BlockDriverState *bs, bool is_write)
|
||||||
|
{
|
||||||
|
ThrottleState *ts = bs->throttle_state;
|
||||||
|
ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts);
|
||||||
|
bool empty_queue;
|
||||||
|
|
||||||
|
/* The timer has just been fired, so we can update the flag */
|
||||||
|
qemu_mutex_lock(&tg->lock);
|
||||||
|
tg->any_timer_armed[is_write] = false;
|
||||||
|
qemu_mutex_unlock(&tg->lock);
|
||||||
|
|
||||||
|
/* Run the request that was waiting for this timer */
|
||||||
|
empty_queue = !qemu_co_enter_next(&bs->throttled_reqs[is_write]);
|
||||||
|
|
||||||
|
/* If the request queue was empty then we have to take care of
|
||||||
|
* scheduling the next one */
|
||||||
|
if (empty_queue) {
|
||||||
|
qemu_mutex_lock(&tg->lock);
|
||||||
|
schedule_next_request(bs, is_write);
|
||||||
|
qemu_mutex_unlock(&tg->lock);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
static void read_timer_cb(void *opaque)
|
||||||
|
{
|
||||||
|
timer_cb(opaque, false);
|
||||||
|
}
|
||||||
|
|
||||||
|
static void write_timer_cb(void *opaque)
|
||||||
|
{
|
||||||
|
timer_cb(opaque, true);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Register a BlockDriverState in the throttling group, also
|
||||||
|
* initializing its timers and updating its throttle_state pointer to
|
||||||
|
* point to it. If a throttling group with that name does not exist
|
||||||
|
* yet, it will be created.
|
||||||
|
*
|
||||||
|
* @bs: the BlockDriverState to insert
|
||||||
|
* @groupname: the name of the group
|
||||||
|
*/
|
||||||
|
void throttle_group_register_bs(BlockDriverState *bs, const char *groupname)
|
||||||
|
{
|
||||||
|
int i;
|
||||||
|
ThrottleGroup *tg = throttle_group_incref(groupname);
|
||||||
|
int clock_type = QEMU_CLOCK_REALTIME;
|
||||||
|
|
||||||
|
if (qtest_enabled()) {
|
||||||
|
/* For testing block IO throttling only */
|
||||||
|
clock_type = QEMU_CLOCK_VIRTUAL;
|
||||||
|
}
|
||||||
|
|
||||||
|
bs->throttle_state = &tg->ts;
|
||||||
|
|
||||||
|
qemu_mutex_lock(&tg->lock);
|
||||||
|
/* If the ThrottleGroup is new set this BlockDriverState as the token */
|
||||||
|
for (i = 0; i < 2; i++) {
|
||||||
|
if (!tg->tokens[i]) {
|
||||||
|
tg->tokens[i] = bs;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
QLIST_INSERT_HEAD(&tg->head, bs, round_robin);
|
||||||
|
|
||||||
|
throttle_timers_init(&bs->throttle_timers,
|
||||||
|
bdrv_get_aio_context(bs),
|
||||||
|
clock_type,
|
||||||
|
read_timer_cb,
|
||||||
|
write_timer_cb,
|
||||||
|
bs);
|
||||||
|
|
||||||
|
qemu_mutex_unlock(&tg->lock);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Unregister a BlockDriverState from its group, removing it from the
|
||||||
|
* list, destroying the timers and setting the throttle_state pointer
|
||||||
|
* to NULL.
|
||||||
|
*
|
||||||
|
* The group will be destroyed if it's empty after this operation.
|
||||||
|
*
|
||||||
|
* @bs: the BlockDriverState to remove
|
||||||
|
*/
|
||||||
|
void throttle_group_unregister_bs(BlockDriverState *bs)
|
||||||
|
{
|
||||||
|
ThrottleGroup *tg = container_of(bs->throttle_state, ThrottleGroup, ts);
|
||||||
|
int i;
|
||||||
|
|
||||||
|
qemu_mutex_lock(&tg->lock);
|
||||||
|
for (i = 0; i < 2; i++) {
|
||||||
|
if (tg->tokens[i] == bs) {
|
||||||
|
BlockDriverState *token = throttle_group_next_bs(bs);
|
||||||
|
/* Take care of the case where this is the last bs in the group */
|
||||||
|
if (token == bs) {
|
||||||
|
token = NULL;
|
||||||
|
}
|
||||||
|
tg->tokens[i] = token;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/* remove the current bs from the list */
|
||||||
|
QLIST_REMOVE(bs, round_robin);
|
||||||
|
throttle_timers_destroy(&bs->throttle_timers);
|
||||||
|
qemu_mutex_unlock(&tg->lock);
|
||||||
|
|
||||||
|
throttle_group_unref(tg);
|
||||||
|
bs->throttle_state = NULL;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Acquire the lock of this throttling group.
|
||||||
|
*
|
||||||
|
* You won't normally need to use this. None of the functions from the
|
||||||
|
* ThrottleGroup API require you to acquire the lock since all of them
|
||||||
|
* deal with it internally.
|
||||||
|
*
|
||||||
|
* This should only be used in exceptional cases when you want to
|
||||||
|
* access the protected fields of a BlockDriverState directly
|
||||||
|
* (e.g. bdrv_swap()).
|
||||||
|
*
|
||||||
|
* @bs: a BlockDriverState that is member of the group
|
||||||
|
*/
|
||||||
|
void throttle_group_lock(BlockDriverState *bs)
|
||||||
|
{
|
||||||
|
ThrottleGroup *tg = container_of(bs->throttle_state, ThrottleGroup, ts);
|
||||||
|
qemu_mutex_lock(&tg->lock);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Release the lock of this throttling group.
|
||||||
|
*
|
||||||
|
* See the comments in throttle_group_lock().
|
||||||
|
*/
|
||||||
|
void throttle_group_unlock(BlockDriverState *bs)
|
||||||
|
{
|
||||||
|
ThrottleGroup *tg = container_of(bs->throttle_state, ThrottleGroup, ts);
|
||||||
|
qemu_mutex_unlock(&tg->lock);
|
||||||
|
}
|
||||||
|
|
||||||
|
static void throttle_groups_init(void)
|
||||||
|
{
|
||||||
|
qemu_mutex_init(&throttle_groups_lock);
|
||||||
|
}
|
||||||
|
|
||||||
|
block_init(throttle_groups_init);
|
38
blockdev.c
38
blockdev.c
|
@ -34,6 +34,7 @@
|
||||||
#include "sysemu/blockdev.h"
|
#include "sysemu/blockdev.h"
|
||||||
#include "hw/block/block.h"
|
#include "hw/block/block.h"
|
||||||
#include "block/blockjob.h"
|
#include "block/blockjob.h"
|
||||||
|
#include "block/throttle-groups.h"
|
||||||
#include "monitor/monitor.h"
|
#include "monitor/monitor.h"
|
||||||
#include "qemu/option.h"
|
#include "qemu/option.h"
|
||||||
#include "qemu/config-file.h"
|
#include "qemu/config-file.h"
|
||||||
|
@ -357,6 +358,7 @@ static BlockBackend *blockdev_init(const char *file, QDict *bs_opts,
|
||||||
const char *id;
|
const char *id;
|
||||||
bool has_driver_specific_opts;
|
bool has_driver_specific_opts;
|
||||||
BlockdevDetectZeroesOptions detect_zeroes;
|
BlockdevDetectZeroesOptions detect_zeroes;
|
||||||
|
const char *throttling_group;
|
||||||
|
|
||||||
/* Check common options by copying from bs_opts to opts, all other options
|
/* Check common options by copying from bs_opts to opts, all other options
|
||||||
* stay in bs_opts for processing by bdrv_open(). */
|
* stay in bs_opts for processing by bdrv_open(). */
|
||||||
|
@ -459,6 +461,8 @@ static BlockBackend *blockdev_init(const char *file, QDict *bs_opts,
|
||||||
|
|
||||||
cfg.op_size = qemu_opt_get_number(opts, "throttling.iops-size", 0);
|
cfg.op_size = qemu_opt_get_number(opts, "throttling.iops-size", 0);
|
||||||
|
|
||||||
|
throttling_group = qemu_opt_get(opts, "throttling.group");
|
||||||
|
|
||||||
if (!check_throttle_config(&cfg, &error)) {
|
if (!check_throttle_config(&cfg, &error)) {
|
||||||
error_propagate(errp, error);
|
error_propagate(errp, error);
|
||||||
goto early_err;
|
goto early_err;
|
||||||
|
@ -547,7 +551,10 @@ static BlockBackend *blockdev_init(const char *file, QDict *bs_opts,
|
||||||
|
|
||||||
/* disk I/O throttling */
|
/* disk I/O throttling */
|
||||||
if (throttle_enabled(&cfg)) {
|
if (throttle_enabled(&cfg)) {
|
||||||
bdrv_io_limits_enable(bs);
|
if (!throttling_group) {
|
||||||
|
throttling_group = blk_name(blk);
|
||||||
|
}
|
||||||
|
bdrv_io_limits_enable(bs, throttling_group);
|
||||||
bdrv_set_io_limits(bs, &cfg);
|
bdrv_set_io_limits(bs, &cfg);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -711,6 +718,8 @@ DriveInfo *drive_new(QemuOpts *all_opts, BlockInterfaceType block_default_type)
|
||||||
|
|
||||||
{ "iops_size", "throttling.iops-size" },
|
{ "iops_size", "throttling.iops-size" },
|
||||||
|
|
||||||
|
{ "group", "throttling.group" },
|
||||||
|
|
||||||
{ "readonly", "read-only" },
|
{ "readonly", "read-only" },
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -1951,7 +1960,9 @@ void qmp_block_set_io_throttle(const char *device, int64_t bps, int64_t bps_rd,
|
||||||
bool has_iops_wr_max,
|
bool has_iops_wr_max,
|
||||||
int64_t iops_wr_max,
|
int64_t iops_wr_max,
|
||||||
bool has_iops_size,
|
bool has_iops_size,
|
||||||
int64_t iops_size, Error **errp)
|
int64_t iops_size,
|
||||||
|
bool has_group,
|
||||||
|
const char *group, Error **errp)
|
||||||
{
|
{
|
||||||
ThrottleConfig cfg;
|
ThrottleConfig cfg;
|
||||||
BlockDriverState *bs;
|
BlockDriverState *bs;
|
||||||
|
@ -2004,14 +2015,19 @@ void qmp_block_set_io_throttle(const char *device, int64_t bps, int64_t bps_rd,
|
||||||
aio_context = bdrv_get_aio_context(bs);
|
aio_context = bdrv_get_aio_context(bs);
|
||||||
aio_context_acquire(aio_context);
|
aio_context_acquire(aio_context);
|
||||||
|
|
||||||
if (!bs->io_limits_enabled && throttle_enabled(&cfg)) {
|
if (throttle_enabled(&cfg)) {
|
||||||
bdrv_io_limits_enable(bs);
|
/* Enable I/O limits if they're not enabled yet, otherwise
|
||||||
} else if (bs->io_limits_enabled && !throttle_enabled(&cfg)) {
|
* just update the throttling group. */
|
||||||
bdrv_io_limits_disable(bs);
|
if (!bs->io_limits_enabled) {
|
||||||
}
|
bdrv_io_limits_enable(bs, has_group ? group : device);
|
||||||
|
} else if (has_group) {
|
||||||
if (bs->io_limits_enabled) {
|
bdrv_io_limits_update_group(bs, group);
|
||||||
|
}
|
||||||
|
/* Set the new throttling configuration */
|
||||||
bdrv_set_io_limits(bs, &cfg);
|
bdrv_set_io_limits(bs, &cfg);
|
||||||
|
} else if (bs->io_limits_enabled) {
|
||||||
|
/* If all throttling settings are set to 0, disable I/O limits */
|
||||||
|
bdrv_io_limits_disable(bs);
|
||||||
}
|
}
|
||||||
|
|
||||||
aio_context_release(aio_context);
|
aio_context_release(aio_context);
|
||||||
|
@ -3188,6 +3204,10 @@ QemuOptsList qemu_common_drive_opts = {
|
||||||
.name = "throttling.iops-size",
|
.name = "throttling.iops-size",
|
||||||
.type = QEMU_OPT_NUMBER,
|
.type = QEMU_OPT_NUMBER,
|
||||||
.help = "when limiting by iops max size of an I/O in bytes",
|
.help = "when limiting by iops max size of an I/O in bytes",
|
||||||
|
},{
|
||||||
|
.name = "throttling.group",
|
||||||
|
.type = QEMU_OPT_STRING,
|
||||||
|
.help = "name of the block throttling group",
|
||||||
},{
|
},{
|
||||||
.name = "copy-on-read",
|
.name = "copy-on-read",
|
||||||
.type = QEMU_OPT_BOOL,
|
.type = QEMU_OPT_BOOL,
|
||||||
|
|
10
hmp.c
10
hmp.c
|
@ -400,7 +400,8 @@ static void print_block_info(Monitor *mon, BlockInfo *info,
|
||||||
" iops_max=%" PRId64
|
" iops_max=%" PRId64
|
||||||
" iops_rd_max=%" PRId64
|
" iops_rd_max=%" PRId64
|
||||||
" iops_wr_max=%" PRId64
|
" iops_wr_max=%" PRId64
|
||||||
" iops_size=%" PRId64 "\n",
|
" iops_size=%" PRId64
|
||||||
|
" group=%s\n",
|
||||||
inserted->bps,
|
inserted->bps,
|
||||||
inserted->bps_rd,
|
inserted->bps_rd,
|
||||||
inserted->bps_wr,
|
inserted->bps_wr,
|
||||||
|
@ -413,7 +414,8 @@ static void print_block_info(Monitor *mon, BlockInfo *info,
|
||||||
inserted->iops_max,
|
inserted->iops_max,
|
||||||
inserted->iops_rd_max,
|
inserted->iops_rd_max,
|
||||||
inserted->iops_wr_max,
|
inserted->iops_wr_max,
|
||||||
inserted->iops_size);
|
inserted->iops_size,
|
||||||
|
inserted->group);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (verbose) {
|
if (verbose) {
|
||||||
|
@ -1357,7 +1359,9 @@ void hmp_block_set_io_throttle(Monitor *mon, const QDict *qdict)
|
||||||
false,
|
false,
|
||||||
0,
|
0,
|
||||||
false, /* No default I/O size */
|
false, /* No default I/O size */
|
||||||
0, &err);
|
0,
|
||||||
|
false,
|
||||||
|
NULL, &err);
|
||||||
hmp_handle_error(mon, &err);
|
hmp_handle_error(mon, &err);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -173,8 +173,9 @@ void bdrv_stats_print(Monitor *mon, const QObject *data);
|
||||||
void bdrv_info_stats(Monitor *mon, QObject **ret_data);
|
void bdrv_info_stats(Monitor *mon, QObject **ret_data);
|
||||||
|
|
||||||
/* disk I/O throttling */
|
/* disk I/O throttling */
|
||||||
void bdrv_io_limits_enable(BlockDriverState *bs);
|
void bdrv_io_limits_enable(BlockDriverState *bs, const char *group);
|
||||||
void bdrv_io_limits_disable(BlockDriverState *bs);
|
void bdrv_io_limits_disable(BlockDriverState *bs);
|
||||||
|
void bdrv_io_limits_update_group(BlockDriverState *bs, const char *group);
|
||||||
|
|
||||||
void bdrv_init(void);
|
void bdrv_init(void);
|
||||||
void bdrv_init_with_whitelist(void);
|
void bdrv_init_with_whitelist(void);
|
||||||
|
|
|
@ -379,9 +379,14 @@ struct BlockDriverState {
|
||||||
unsigned int serialising_in_flight;
|
unsigned int serialising_in_flight;
|
||||||
|
|
||||||
/* I/O throttling */
|
/* I/O throttling */
|
||||||
ThrottleState throttle_state;
|
|
||||||
CoQueue throttled_reqs[2];
|
CoQueue throttled_reqs[2];
|
||||||
bool io_limits_enabled;
|
bool io_limits_enabled;
|
||||||
|
/* The following fields are protected by the ThrottleGroup lock.
|
||||||
|
* See the ThrottleGroup documentation for details. */
|
||||||
|
ThrottleState *throttle_state;
|
||||||
|
ThrottleTimers throttle_timers;
|
||||||
|
unsigned pending_reqs[2];
|
||||||
|
QLIST_ENTRY(BlockDriverState) round_robin;
|
||||||
|
|
||||||
/* I/O stats (display with "info blockstats"). */
|
/* I/O stats (display with "info blockstats"). */
|
||||||
BlockAcctStats stats;
|
BlockAcctStats stats;
|
||||||
|
|
|
@ -0,0 +1,46 @@
|
||||||
|
/*
|
||||||
|
* QEMU block throttling group infrastructure
|
||||||
|
*
|
||||||
|
* Copyright (C) Nodalink, EURL. 2014
|
||||||
|
* Copyright (C) Igalia, S.L. 2015
|
||||||
|
*
|
||||||
|
* Authors:
|
||||||
|
* Benoît Canet <benoit.canet@nodalink.com>
|
||||||
|
* Alberto Garcia <berto@igalia.com>
|
||||||
|
*
|
||||||
|
* This program is free software; you can redistribute it and/or
|
||||||
|
* modify it under the terms of the GNU General Public License as
|
||||||
|
* published by the Free Software Foundation; either version 2 or
|
||||||
|
* (at your option) version 3 of the License.
|
||||||
|
*
|
||||||
|
* This program is distributed in the hope that it will be useful,
|
||||||
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
* GNU General Public License for more details.
|
||||||
|
*
|
||||||
|
* You should have received a copy of the GNU General Public License
|
||||||
|
* along with this program; if not, see <http://www.gnu.org/licenses/>.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef THROTTLE_GROUPS_H
|
||||||
|
#define THROTTLE_GROUPS_H
|
||||||
|
|
||||||
|
#include "qemu/throttle.h"
|
||||||
|
#include "block/block_int.h"
|
||||||
|
|
||||||
|
const char *throttle_group_get_name(BlockDriverState *bs);
|
||||||
|
|
||||||
|
void throttle_group_config(BlockDriverState *bs, ThrottleConfig *cfg);
|
||||||
|
void throttle_group_get_config(BlockDriverState *bs, ThrottleConfig *cfg);
|
||||||
|
|
||||||
|
void throttle_group_register_bs(BlockDriverState *bs, const char *groupname);
|
||||||
|
void throttle_group_unregister_bs(BlockDriverState *bs);
|
||||||
|
|
||||||
|
void coroutine_fn throttle_group_co_io_limits_intercept(BlockDriverState *bs,
|
||||||
|
unsigned int bytes,
|
||||||
|
bool is_write);
|
||||||
|
|
||||||
|
void throttle_group_lock(BlockDriverState *bs);
|
||||||
|
void throttle_group_unlock(BlockDriverState *bs);
|
||||||
|
|
||||||
|
#endif
|
|
@ -1,10 +1,12 @@
|
||||||
/*
|
/*
|
||||||
* QEMU throttling infrastructure
|
* QEMU throttling infrastructure
|
||||||
*
|
*
|
||||||
* Copyright (C) Nodalink, SARL. 2013
|
* Copyright (C) Nodalink, EURL. 2013-2014
|
||||||
|
* Copyright (C) Igalia, S.L. 2015
|
||||||
*
|
*
|
||||||
* Author:
|
* Authors:
|
||||||
* Benoît Canet <benoit.canet@irqsave.net>
|
* Benoît Canet <benoit.canet@nodalink.com>
|
||||||
|
* Alberto Garcia <berto@igalia.com>
|
||||||
*
|
*
|
||||||
* This program is free software; you can redistribute it and/or
|
* This program is free software; you can redistribute it and/or
|
||||||
* modify it under the terms of the GNU General Public License as
|
* modify it under the terms of the GNU General Public License as
|
||||||
|
@ -65,14 +67,17 @@ typedef struct ThrottleConfig {
|
||||||
typedef struct ThrottleState {
|
typedef struct ThrottleState {
|
||||||
ThrottleConfig cfg; /* configuration */
|
ThrottleConfig cfg; /* configuration */
|
||||||
int64_t previous_leak; /* timestamp of the last leak done */
|
int64_t previous_leak; /* timestamp of the last leak done */
|
||||||
QEMUTimer * timers[2]; /* timers used to do the throttling */
|
} ThrottleState;
|
||||||
|
|
||||||
|
typedef struct ThrottleTimers {
|
||||||
|
QEMUTimer *timers[2]; /* timers used to do the throttling */
|
||||||
QEMUClockType clock_type; /* the clock used */
|
QEMUClockType clock_type; /* the clock used */
|
||||||
|
|
||||||
/* Callbacks */
|
/* Callbacks */
|
||||||
QEMUTimerCB *read_timer_cb;
|
QEMUTimerCB *read_timer_cb;
|
||||||
QEMUTimerCB *write_timer_cb;
|
QEMUTimerCB *write_timer_cb;
|
||||||
void *timer_opaque;
|
void *timer_opaque;
|
||||||
} ThrottleState;
|
} ThrottleTimers;
|
||||||
|
|
||||||
/* operations on single leaky buckets */
|
/* operations on single leaky buckets */
|
||||||
void throttle_leak_bucket(LeakyBucket *bkt, int64_t delta);
|
void throttle_leak_bucket(LeakyBucket *bkt, int64_t delta);
|
||||||
|
@ -86,20 +91,23 @@ bool throttle_compute_timer(ThrottleState *ts,
|
||||||
int64_t *next_timestamp);
|
int64_t *next_timestamp);
|
||||||
|
|
||||||
/* init/destroy cycle */
|
/* init/destroy cycle */
|
||||||
void throttle_init(ThrottleState *ts,
|
void throttle_init(ThrottleState *ts);
|
||||||
AioContext *aio_context,
|
|
||||||
QEMUClockType clock_type,
|
|
||||||
void (read_timer)(void *),
|
|
||||||
void (write_timer)(void *),
|
|
||||||
void *timer_opaque);
|
|
||||||
|
|
||||||
void throttle_destroy(ThrottleState *ts);
|
void throttle_timers_init(ThrottleTimers *tt,
|
||||||
|
AioContext *aio_context,
|
||||||
|
QEMUClockType clock_type,
|
||||||
|
QEMUTimerCB *read_timer_cb,
|
||||||
|
QEMUTimerCB *write_timer_cb,
|
||||||
|
void *timer_opaque);
|
||||||
|
|
||||||
void throttle_detach_aio_context(ThrottleState *ts);
|
void throttle_timers_destroy(ThrottleTimers *tt);
|
||||||
|
|
||||||
void throttle_attach_aio_context(ThrottleState *ts, AioContext *new_context);
|
void throttle_timers_detach_aio_context(ThrottleTimers *tt);
|
||||||
|
|
||||||
bool throttle_have_timer(ThrottleState *ts);
|
void throttle_timers_attach_aio_context(ThrottleTimers *tt,
|
||||||
|
AioContext *new_context);
|
||||||
|
|
||||||
|
bool throttle_timers_are_initialized(ThrottleTimers *tt);
|
||||||
|
|
||||||
/* configuration */
|
/* configuration */
|
||||||
bool throttle_enabled(ThrottleConfig *cfg);
|
bool throttle_enabled(ThrottleConfig *cfg);
|
||||||
|
@ -108,12 +116,16 @@ bool throttle_conflicting(ThrottleConfig *cfg);
|
||||||
|
|
||||||
bool throttle_is_valid(ThrottleConfig *cfg);
|
bool throttle_is_valid(ThrottleConfig *cfg);
|
||||||
|
|
||||||
void throttle_config(ThrottleState *ts, ThrottleConfig *cfg);
|
void throttle_config(ThrottleState *ts,
|
||||||
|
ThrottleTimers *tt,
|
||||||
|
ThrottleConfig *cfg);
|
||||||
|
|
||||||
void throttle_get_config(ThrottleState *ts, ThrottleConfig *cfg);
|
void throttle_get_config(ThrottleState *ts, ThrottleConfig *cfg);
|
||||||
|
|
||||||
/* usage */
|
/* usage */
|
||||||
bool throttle_schedule_timer(ThrottleState *ts, bool is_write);
|
bool throttle_schedule_timer(ThrottleState *ts,
|
||||||
|
ThrottleTimers *tt,
|
||||||
|
bool is_write);
|
||||||
|
|
||||||
void throttle_account(ThrottleState *ts, bool is_write, uint64_t size);
|
void throttle_account(ThrottleState *ts, bool is_write, uint64_t size);
|
||||||
|
|
||||||
|
|
11
iothread.c
11
iothread.c
|
@ -31,14 +31,21 @@ typedef ObjectClass IOThreadClass;
|
||||||
static void *iothread_run(void *opaque)
|
static void *iothread_run(void *opaque)
|
||||||
{
|
{
|
||||||
IOThread *iothread = opaque;
|
IOThread *iothread = opaque;
|
||||||
|
bool blocking;
|
||||||
|
|
||||||
qemu_mutex_lock(&iothread->init_done_lock);
|
qemu_mutex_lock(&iothread->init_done_lock);
|
||||||
iothread->thread_id = qemu_get_thread_id();
|
iothread->thread_id = qemu_get_thread_id();
|
||||||
qemu_cond_signal(&iothread->init_done_cond);
|
qemu_cond_signal(&iothread->init_done_cond);
|
||||||
qemu_mutex_unlock(&iothread->init_done_lock);
|
qemu_mutex_unlock(&iothread->init_done_lock);
|
||||||
|
|
||||||
while (!atomic_read(&iothread->stopping)) {
|
while (!iothread->stopping) {
|
||||||
aio_poll(iothread->ctx, true);
|
aio_context_acquire(iothread->ctx);
|
||||||
|
blocking = true;
|
||||||
|
while (!iothread->stopping && aio_poll(iothread->ctx, blocking)) {
|
||||||
|
/* Progress was made, keep going */
|
||||||
|
blocking = false;
|
||||||
|
}
|
||||||
|
aio_context_release(iothread->ctx);
|
||||||
}
|
}
|
||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
|
|
|
@ -259,6 +259,8 @@
|
||||||
#
|
#
|
||||||
# @iops_size: #optional an I/O size in bytes (Since 1.7)
|
# @iops_size: #optional an I/O size in bytes (Since 1.7)
|
||||||
#
|
#
|
||||||
|
# @group: #optional throttle group name (Since 2.4)
|
||||||
|
#
|
||||||
# @cache: the cache mode used for the block device (since: 2.3)
|
# @cache: the cache mode used for the block device (since: 2.3)
|
||||||
#
|
#
|
||||||
# @write_threshold: configured write threshold for the device.
|
# @write_threshold: configured write threshold for the device.
|
||||||
|
@ -278,7 +280,7 @@
|
||||||
'*bps_max': 'int', '*bps_rd_max': 'int',
|
'*bps_max': 'int', '*bps_rd_max': 'int',
|
||||||
'*bps_wr_max': 'int', '*iops_max': 'int',
|
'*bps_wr_max': 'int', '*iops_max': 'int',
|
||||||
'*iops_rd_max': 'int', '*iops_wr_max': 'int',
|
'*iops_rd_max': 'int', '*iops_wr_max': 'int',
|
||||||
'*iops_size': 'int', 'cache': 'BlockdevCacheInfo',
|
'*iops_size': 'int', '*group': 'str', 'cache': 'BlockdevCacheInfo',
|
||||||
'write_threshold': 'int' } }
|
'write_threshold': 'int' } }
|
||||||
|
|
||||||
##
|
##
|
||||||
|
@ -1062,6 +1064,27 @@
|
||||||
#
|
#
|
||||||
# Change I/O throttle limits for a block drive.
|
# Change I/O throttle limits for a block drive.
|
||||||
#
|
#
|
||||||
|
# Since QEMU 2.4, each device with I/O limits is member of a throttle
|
||||||
|
# group.
|
||||||
|
#
|
||||||
|
# If two or more devices are members of the same group, the limits
|
||||||
|
# will apply to the combined I/O of the whole group in a round-robin
|
||||||
|
# fashion. Therefore, setting new I/O limits to a device will affect
|
||||||
|
# the whole group.
|
||||||
|
#
|
||||||
|
# The name of the group can be specified using the 'group' parameter.
|
||||||
|
# If the parameter is unset, it is assumed to be the current group of
|
||||||
|
# that device. If it's not in any group yet, the name of the device
|
||||||
|
# will be used as the name for its group.
|
||||||
|
#
|
||||||
|
# The 'group' parameter can also be used to move a device to a
|
||||||
|
# different group. In this case the limits specified in the parameters
|
||||||
|
# will be applied to the new group only.
|
||||||
|
#
|
||||||
|
# I/O limits can be disabled by setting all of them to 0. In this case
|
||||||
|
# the device will be removed from its group and the rest of its
|
||||||
|
# members will no be affected. The 'group' parameter is ignored.
|
||||||
|
#
|
||||||
# @device: The name of the device
|
# @device: The name of the device
|
||||||
#
|
#
|
||||||
# @bps: total throughput limit in bytes per second
|
# @bps: total throughput limit in bytes per second
|
||||||
|
@ -1090,6 +1113,8 @@
|
||||||
#
|
#
|
||||||
# @iops_size: #optional an I/O size in bytes (Since 1.7)
|
# @iops_size: #optional an I/O size in bytes (Since 1.7)
|
||||||
#
|
#
|
||||||
|
# @group: #optional throttle group name (Since 2.4)
|
||||||
|
#
|
||||||
# Returns: Nothing on success
|
# Returns: Nothing on success
|
||||||
# If @device is not a valid block device, DeviceNotFound
|
# If @device is not a valid block device, DeviceNotFound
|
||||||
#
|
#
|
||||||
|
@ -1101,7 +1126,7 @@
|
||||||
'*bps_max': 'int', '*bps_rd_max': 'int',
|
'*bps_max': 'int', '*bps_rd_max': 'int',
|
||||||
'*bps_wr_max': 'int', '*iops_max': 'int',
|
'*bps_wr_max': 'int', '*iops_max': 'int',
|
||||||
'*iops_rd_max': 'int', '*iops_wr_max': 'int',
|
'*iops_rd_max': 'int', '*iops_wr_max': 'int',
|
||||||
'*iops_size': 'int' } }
|
'*iops_size': 'int', '*group': 'str' } }
|
||||||
|
|
||||||
##
|
##
|
||||||
# @block-stream:
|
# @block-stream:
|
||||||
|
|
|
@ -468,6 +468,7 @@ DEF("drive", HAS_ARG, QEMU_OPTION_drive,
|
||||||
" [[,bps_max=bm]|[[,bps_rd_max=rm][,bps_wr_max=wm]]]\n"
|
" [[,bps_max=bm]|[[,bps_rd_max=rm][,bps_wr_max=wm]]]\n"
|
||||||
" [[,iops_max=im]|[[,iops_rd_max=irm][,iops_wr_max=iwm]]]\n"
|
" [[,iops_max=im]|[[,iops_rd_max=irm][,iops_wr_max=iwm]]]\n"
|
||||||
" [[,iops_size=is]]\n"
|
" [[,iops_size=is]]\n"
|
||||||
|
" [[,group=g]]\n"
|
||||||
" use 'file' as a drive image\n", QEMU_ARCH_ALL)
|
" use 'file' as a drive image\n", QEMU_ARCH_ALL)
|
||||||
STEXI
|
STEXI
|
||||||
@item -drive @var{option}[,@var{option}[,@var{option}[,...]]]
|
@item -drive @var{option}[,@var{option}[,@var{option}[,...]]]
|
||||||
|
|
|
@ -1853,7 +1853,7 @@ EQMP
|
||||||
|
|
||||||
{
|
{
|
||||||
.name = "block_set_io_throttle",
|
.name = "block_set_io_throttle",
|
||||||
.args_type = "device:B,bps:l,bps_rd:l,bps_wr:l,iops:l,iops_rd:l,iops_wr:l,bps_max:l?,bps_rd_max:l?,bps_wr_max:l?,iops_max:l?,iops_rd_max:l?,iops_wr_max:l?,iops_size:l?",
|
.args_type = "device:B,bps:l,bps_rd:l,bps_wr:l,iops:l,iops_rd:l,iops_wr:l,bps_max:l?,bps_rd_max:l?,bps_wr_max:l?,iops_max:l?,iops_rd_max:l?,iops_wr_max:l?,iops_size:l?,group:s?",
|
||||||
.mhandler.cmd_new = qmp_marshal_input_block_set_io_throttle,
|
.mhandler.cmd_new = qmp_marshal_input_block_set_io_throttle,
|
||||||
},
|
},
|
||||||
|
|
||||||
|
@ -1879,6 +1879,7 @@ Arguments:
|
||||||
- "iops_rd_max": read I/O operations max (json-int)
|
- "iops_rd_max": read I/O operations max (json-int)
|
||||||
- "iops_wr_max": write I/O operations max (json-int)
|
- "iops_wr_max": write I/O operations max (json-int)
|
||||||
- "iops_size": I/O size in bytes when limiting (json-int)
|
- "iops_size": I/O size in bytes when limiting (json-int)
|
||||||
|
- "group": throttle group name (json-string)
|
||||||
|
|
||||||
Example:
|
Example:
|
||||||
|
|
||||||
|
|
|
@ -3,6 +3,7 @@
|
||||||
# Tests for IO throttling
|
# Tests for IO throttling
|
||||||
#
|
#
|
||||||
# Copyright (C) 2015 Red Hat, Inc.
|
# Copyright (C) 2015 Red Hat, Inc.
|
||||||
|
# Copyright (C) 2015 Igalia, S.L.
|
||||||
#
|
#
|
||||||
# This program is free software; you can redistribute it and/or modify
|
# This program is free software; you can redistribute it and/or modify
|
||||||
# it under the terms of the GNU General Public License as published by
|
# it under the terms of the GNU General Public License as published by
|
||||||
|
@ -22,6 +23,7 @@ import iotests
|
||||||
|
|
||||||
class ThrottleTestCase(iotests.QMPTestCase):
|
class ThrottleTestCase(iotests.QMPTestCase):
|
||||||
test_img = "null-aio://"
|
test_img = "null-aio://"
|
||||||
|
max_drives = 3
|
||||||
|
|
||||||
def blockstats(self, device):
|
def blockstats(self, device):
|
||||||
result = self.vm.qmp("query-blockstats")
|
result = self.vm.qmp("query-blockstats")
|
||||||
|
@ -32,26 +34,31 @@ class ThrottleTestCase(iotests.QMPTestCase):
|
||||||
raise Exception("Device not found for blockstats: %s" % device)
|
raise Exception("Device not found for blockstats: %s" % device)
|
||||||
|
|
||||||
def setUp(self):
|
def setUp(self):
|
||||||
self.vm = iotests.VM().add_drive(self.test_img)
|
self.vm = iotests.VM()
|
||||||
|
for i in range(0, self.max_drives):
|
||||||
|
self.vm.add_drive(self.test_img)
|
||||||
self.vm.launch()
|
self.vm.launch()
|
||||||
|
|
||||||
def tearDown(self):
|
def tearDown(self):
|
||||||
self.vm.shutdown()
|
self.vm.shutdown()
|
||||||
|
|
||||||
def do_test_throttle(self, seconds, params):
|
def do_test_throttle(self, ndrives, seconds, params):
|
||||||
def check_limit(limit, num):
|
def check_limit(limit, num):
|
||||||
# IO throttling algorithm is discrete, allow 10% error so the test
|
# IO throttling algorithm is discrete, allow 10% error so the test
|
||||||
# is more robust
|
# is more robust
|
||||||
return limit == 0 or \
|
return limit == 0 or \
|
||||||
(num < seconds * limit * 1.1
|
(num < seconds * limit * 1.1 / ndrives
|
||||||
and num > seconds * limit * 0.9)
|
and num > seconds * limit * 0.9 / ndrives)
|
||||||
|
|
||||||
nsec_per_sec = 1000000000
|
nsec_per_sec = 1000000000
|
||||||
|
|
||||||
params['device'] = 'drive0'
|
params['group'] = 'test'
|
||||||
|
|
||||||
result = self.vm.qmp("block_set_io_throttle", conv_keys=False, **params)
|
# Set the I/O throttling parameters to all drives
|
||||||
self.assert_qmp(result, 'return', {})
|
for i in range(0, ndrives):
|
||||||
|
params['device'] = 'drive%d' % i
|
||||||
|
result = self.vm.qmp("block_set_io_throttle", conv_keys=False, **params)
|
||||||
|
self.assert_qmp(result, 'return', {})
|
||||||
|
|
||||||
# Set vm clock to a known value
|
# Set vm clock to a known value
|
||||||
ns = seconds * nsec_per_sec
|
ns = seconds * nsec_per_sec
|
||||||
|
@ -66,32 +73,60 @@ class ThrottleTestCase(iotests.QMPTestCase):
|
||||||
params['iops'] / 2,
|
params['iops'] / 2,
|
||||||
params['iops_rd'])
|
params['iops_rd'])
|
||||||
rd_nr *= seconds * 2
|
rd_nr *= seconds * 2
|
||||||
|
rd_nr /= ndrives
|
||||||
wr_nr = max(params['bps'] / rq_size / 2,
|
wr_nr = max(params['bps'] / rq_size / 2,
|
||||||
params['bps_wr'] / rq_size,
|
params['bps_wr'] / rq_size,
|
||||||
params['iops'] / 2,
|
params['iops'] / 2,
|
||||||
params['iops_wr'])
|
params['iops_wr'])
|
||||||
wr_nr *= seconds * 2
|
wr_nr *= seconds * 2
|
||||||
for i in range(rd_nr):
|
wr_nr /= ndrives
|
||||||
self.vm.hmp_qemu_io("drive0", "aio_read %d %d" % (i * rq_size, rq_size))
|
|
||||||
for i in range(wr_nr):
|
|
||||||
self.vm.hmp_qemu_io("drive0", "aio_write %d %d" % (i * rq_size, rq_size))
|
|
||||||
|
|
||||||
start_rd_bytes, start_rd_iops, start_wr_bytes, start_wr_iops = self.blockstats('drive0')
|
# Send I/O requests to all drives
|
||||||
|
for i in range(rd_nr):
|
||||||
|
for drive in range(0, ndrives):
|
||||||
|
self.vm.hmp_qemu_io("drive%d" % drive, "aio_read %d %d" %
|
||||||
|
(i * rq_size, rq_size))
|
||||||
|
|
||||||
|
for i in range(wr_nr):
|
||||||
|
for drive in range(0, ndrives):
|
||||||
|
self.vm.hmp_qemu_io("drive%d" % drive, "aio_write %d %d" %
|
||||||
|
(i * rq_size, rq_size))
|
||||||
|
|
||||||
|
# We'll store the I/O stats for each drive in these arrays
|
||||||
|
start_rd_bytes = [0] * ndrives
|
||||||
|
start_rd_iops = [0] * ndrives
|
||||||
|
start_wr_bytes = [0] * ndrives
|
||||||
|
start_wr_iops = [0] * ndrives
|
||||||
|
end_rd_bytes = [0] * ndrives
|
||||||
|
end_rd_iops = [0] * ndrives
|
||||||
|
end_wr_bytes = [0] * ndrives
|
||||||
|
end_wr_iops = [0] * ndrives
|
||||||
|
|
||||||
|
# Read the stats before advancing the clock
|
||||||
|
for i in range(0, ndrives):
|
||||||
|
start_rd_bytes[i], start_rd_iops[i], start_wr_bytes[i], \
|
||||||
|
start_wr_iops[i] = self.blockstats('drive%d' % i)
|
||||||
|
|
||||||
self.vm.qtest("clock_step %d" % ns)
|
self.vm.qtest("clock_step %d" % ns)
|
||||||
end_rd_bytes, end_rd_iops, end_wr_bytes, end_wr_iops = self.blockstats('drive0')
|
|
||||||
|
|
||||||
rd_bytes = end_rd_bytes - start_rd_bytes
|
# Read the stats after advancing the clock
|
||||||
rd_iops = end_rd_iops - start_rd_iops
|
for i in range(0, ndrives):
|
||||||
wr_bytes = end_wr_bytes - start_wr_bytes
|
end_rd_bytes[i], end_rd_iops[i], end_wr_bytes[i], \
|
||||||
wr_iops = end_wr_iops - start_wr_iops
|
end_wr_iops[i] = self.blockstats('drive%d' % i)
|
||||||
|
|
||||||
self.assertTrue(check_limit(params['bps'], rd_bytes + wr_bytes))
|
# Check that the I/O is within the limits and evenly distributed
|
||||||
self.assertTrue(check_limit(params['bps_rd'], rd_bytes))
|
for i in range(0, ndrives):
|
||||||
self.assertTrue(check_limit(params['bps_wr'], wr_bytes))
|
rd_bytes = end_rd_bytes[i] - start_rd_bytes[i]
|
||||||
self.assertTrue(check_limit(params['iops'], rd_iops + wr_iops))
|
rd_iops = end_rd_iops[i] - start_rd_iops[i]
|
||||||
self.assertTrue(check_limit(params['iops_rd'], rd_iops))
|
wr_bytes = end_wr_bytes[i] - start_wr_bytes[i]
|
||||||
self.assertTrue(check_limit(params['iops_wr'], wr_iops))
|
wr_iops = end_wr_iops[i] - start_wr_iops[i]
|
||||||
|
|
||||||
|
self.assertTrue(check_limit(params['bps'], rd_bytes + wr_bytes))
|
||||||
|
self.assertTrue(check_limit(params['bps_rd'], rd_bytes))
|
||||||
|
self.assertTrue(check_limit(params['bps_wr'], wr_bytes))
|
||||||
|
self.assertTrue(check_limit(params['iops'], rd_iops + wr_iops))
|
||||||
|
self.assertTrue(check_limit(params['iops_rd'], rd_iops))
|
||||||
|
self.assertTrue(check_limit(params['iops_wr'], wr_iops))
|
||||||
|
|
||||||
def test_all(self):
|
def test_all(self):
|
||||||
params = {"bps": 4096,
|
params = {"bps": 4096,
|
||||||
|
@ -101,11 +136,13 @@ class ThrottleTestCase(iotests.QMPTestCase):
|
||||||
"iops_rd": 10,
|
"iops_rd": 10,
|
||||||
"iops_wr": 10,
|
"iops_wr": 10,
|
||||||
}
|
}
|
||||||
# Pick each out of all possible params and test
|
# Repeat the test with different numbers of drives
|
||||||
for tk in params:
|
for ndrives in range(1, self.max_drives + 1):
|
||||||
limits = dict([(k, 0) for k in params])
|
# Pick each out of all possible params and test
|
||||||
limits[tk] = params[tk]
|
for tk in params:
|
||||||
self.do_test_throttle(5, limits)
|
limits = dict([(k, 0) for k in params])
|
||||||
|
limits[tk] = params[tk] * ndrives
|
||||||
|
self.do_test_throttle(ndrives, 5, limits)
|
||||||
|
|
||||||
class ThrottleTestCoroutine(ThrottleTestCase):
|
class ThrottleTestCoroutine(ThrottleTestCase):
|
||||||
test_img = "null-co://"
|
test_img = "null-co://"
|
||||||
|
|
|
@ -107,7 +107,6 @@ static void test_notify(void)
|
||||||
|
|
||||||
typedef struct {
|
typedef struct {
|
||||||
QemuMutex start_lock;
|
QemuMutex start_lock;
|
||||||
EventNotifier notifier;
|
|
||||||
bool thread_acquired;
|
bool thread_acquired;
|
||||||
} AcquireTestData;
|
} AcquireTestData;
|
||||||
|
|
||||||
|
@ -119,8 +118,6 @@ static void *test_acquire_thread(void *opaque)
|
||||||
qemu_mutex_lock(&data->start_lock);
|
qemu_mutex_lock(&data->start_lock);
|
||||||
qemu_mutex_unlock(&data->start_lock);
|
qemu_mutex_unlock(&data->start_lock);
|
||||||
|
|
||||||
g_usleep(500000);
|
|
||||||
event_notifier_set(&data->notifier);
|
|
||||||
aio_context_acquire(ctx);
|
aio_context_acquire(ctx);
|
||||||
aio_context_release(ctx);
|
aio_context_release(ctx);
|
||||||
|
|
||||||
|
@ -129,19 +126,20 @@ static void *test_acquire_thread(void *opaque)
|
||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void dummy_notifier_read(EventNotifier *n)
|
static void dummy_notifier_read(EventNotifier *unused)
|
||||||
{
|
{
|
||||||
event_notifier_test_and_clear(n);
|
g_assert(false); /* should never be invoked */
|
||||||
}
|
}
|
||||||
|
|
||||||
static void test_acquire(void)
|
static void test_acquire(void)
|
||||||
{
|
{
|
||||||
QemuThread thread;
|
QemuThread thread;
|
||||||
|
EventNotifier notifier;
|
||||||
AcquireTestData data;
|
AcquireTestData data;
|
||||||
|
|
||||||
/* Dummy event notifier ensures aio_poll() will block */
|
/* Dummy event notifier ensures aio_poll() will block */
|
||||||
event_notifier_init(&data.notifier, false);
|
event_notifier_init(¬ifier, false);
|
||||||
aio_set_event_notifier(ctx, &data.notifier, dummy_notifier_read);
|
aio_set_event_notifier(ctx, ¬ifier, dummy_notifier_read);
|
||||||
g_assert(!aio_poll(ctx, false)); /* consume aio_notify() */
|
g_assert(!aio_poll(ctx, false)); /* consume aio_notify() */
|
||||||
|
|
||||||
qemu_mutex_init(&data.start_lock);
|
qemu_mutex_init(&data.start_lock);
|
||||||
|
@ -155,13 +153,12 @@ static void test_acquire(void)
|
||||||
/* Block in aio_poll(), let other thread kick us and acquire context */
|
/* Block in aio_poll(), let other thread kick us and acquire context */
|
||||||
aio_context_acquire(ctx);
|
aio_context_acquire(ctx);
|
||||||
qemu_mutex_unlock(&data.start_lock); /* let the thread run */
|
qemu_mutex_unlock(&data.start_lock); /* let the thread run */
|
||||||
g_assert(aio_poll(ctx, true));
|
g_assert(!aio_poll(ctx, true));
|
||||||
g_assert(!data.thread_acquired);
|
|
||||||
aio_context_release(ctx);
|
aio_context_release(ctx);
|
||||||
|
|
||||||
qemu_thread_join(&thread);
|
qemu_thread_join(&thread);
|
||||||
aio_set_event_notifier(ctx, &data.notifier, NULL);
|
aio_set_event_notifier(ctx, ¬ifier, NULL);
|
||||||
event_notifier_cleanup(&data.notifier);
|
event_notifier_cleanup(¬ifier);
|
||||||
|
|
||||||
g_assert(data.thread_acquired);
|
g_assert(data.thread_acquired);
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,10 +1,12 @@
|
||||||
/*
|
/*
|
||||||
* Throttle infrastructure tests
|
* Throttle infrastructure tests
|
||||||
*
|
*
|
||||||
* Copyright Nodalink, SARL. 2013
|
* Copyright Nodalink, EURL. 2013-2014
|
||||||
|
* Copyright Igalia, S.L. 2015
|
||||||
*
|
*
|
||||||
* Authors:
|
* Authors:
|
||||||
* Benoît Canet <benoit.canet@irqsave.net>
|
* Benoît Canet <benoit.canet@nodalink.com>
|
||||||
|
* Alberto Garcia <berto@igalia.com>
|
||||||
*
|
*
|
||||||
* This work is licensed under the terms of the GNU LGPL, version 2 or later.
|
* This work is licensed under the terms of the GNU LGPL, version 2 or later.
|
||||||
* See the COPYING.LIB file in the top-level directory.
|
* See the COPYING.LIB file in the top-level directory.
|
||||||
|
@ -15,11 +17,13 @@
|
||||||
#include "block/aio.h"
|
#include "block/aio.h"
|
||||||
#include "qemu/throttle.h"
|
#include "qemu/throttle.h"
|
||||||
#include "qemu/error-report.h"
|
#include "qemu/error-report.h"
|
||||||
|
#include "block/throttle-groups.h"
|
||||||
|
|
||||||
static AioContext *ctx;
|
static AioContext *ctx;
|
||||||
static LeakyBucket bkt;
|
static LeakyBucket bkt;
|
||||||
static ThrottleConfig cfg;
|
static ThrottleConfig cfg;
|
||||||
static ThrottleState ts;
|
static ThrottleState ts;
|
||||||
|
static ThrottleTimers tt;
|
||||||
|
|
||||||
/* useful function */
|
/* useful function */
|
||||||
static bool double_cmp(double x, double y)
|
static bool double_cmp(double x, double y)
|
||||||
|
@ -103,17 +107,19 @@ static void test_init(void)
|
||||||
{
|
{
|
||||||
int i;
|
int i;
|
||||||
|
|
||||||
/* fill the structure with crap */
|
/* fill the structures with crap */
|
||||||
memset(&ts, 1, sizeof(ts));
|
memset(&ts, 1, sizeof(ts));
|
||||||
|
memset(&tt, 1, sizeof(tt));
|
||||||
|
|
||||||
/* init the structure */
|
/* init structures */
|
||||||
throttle_init(&ts, ctx, QEMU_CLOCK_VIRTUAL,
|
throttle_init(&ts);
|
||||||
read_timer_cb, write_timer_cb, &ts);
|
throttle_timers_init(&tt, ctx, QEMU_CLOCK_VIRTUAL,
|
||||||
|
read_timer_cb, write_timer_cb, &ts);
|
||||||
|
|
||||||
/* check initialized fields */
|
/* check initialized fields */
|
||||||
g_assert(ts.clock_type == QEMU_CLOCK_VIRTUAL);
|
g_assert(tt.clock_type == QEMU_CLOCK_VIRTUAL);
|
||||||
g_assert(ts.timers[0]);
|
g_assert(tt.timers[0]);
|
||||||
g_assert(ts.timers[1]);
|
g_assert(tt.timers[1]);
|
||||||
|
|
||||||
/* check other fields where cleared */
|
/* check other fields where cleared */
|
||||||
g_assert(!ts.previous_leak);
|
g_assert(!ts.previous_leak);
|
||||||
|
@ -124,17 +130,18 @@ static void test_init(void)
|
||||||
g_assert(!ts.cfg.buckets[i].level);
|
g_assert(!ts.cfg.buckets[i].level);
|
||||||
}
|
}
|
||||||
|
|
||||||
throttle_destroy(&ts);
|
throttle_timers_destroy(&tt);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void test_destroy(void)
|
static void test_destroy(void)
|
||||||
{
|
{
|
||||||
int i;
|
int i;
|
||||||
throttle_init(&ts, ctx, QEMU_CLOCK_VIRTUAL,
|
throttle_init(&ts);
|
||||||
read_timer_cb, write_timer_cb, &ts);
|
throttle_timers_init(&tt, ctx, QEMU_CLOCK_VIRTUAL,
|
||||||
throttle_destroy(&ts);
|
read_timer_cb, write_timer_cb, &ts);
|
||||||
|
throttle_timers_destroy(&tt);
|
||||||
for (i = 0; i < 2; i++) {
|
for (i = 0; i < 2; i++) {
|
||||||
g_assert(!ts.timers[i]);
|
g_assert(!tt.timers[i]);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -170,11 +177,12 @@ static void test_config_functions(void)
|
||||||
|
|
||||||
orig_cfg.op_size = 1;
|
orig_cfg.op_size = 1;
|
||||||
|
|
||||||
throttle_init(&ts, ctx, QEMU_CLOCK_VIRTUAL,
|
throttle_init(&ts);
|
||||||
read_timer_cb, write_timer_cb, &ts);
|
throttle_timers_init(&tt, ctx, QEMU_CLOCK_VIRTUAL,
|
||||||
|
read_timer_cb, write_timer_cb, &ts);
|
||||||
/* structure reset by throttle_init previous_leak should be null */
|
/* structure reset by throttle_init previous_leak should be null */
|
||||||
g_assert(!ts.previous_leak);
|
g_assert(!ts.previous_leak);
|
||||||
throttle_config(&ts, &orig_cfg);
|
throttle_config(&ts, &tt, &orig_cfg);
|
||||||
|
|
||||||
/* has previous leak been initialized by throttle_config ? */
|
/* has previous leak been initialized by throttle_config ? */
|
||||||
g_assert(ts.previous_leak);
|
g_assert(ts.previous_leak);
|
||||||
|
@ -182,7 +190,7 @@ static void test_config_functions(void)
|
||||||
/* get back the fixed configuration */
|
/* get back the fixed configuration */
|
||||||
throttle_get_config(&ts, &final_cfg);
|
throttle_get_config(&ts, &final_cfg);
|
||||||
|
|
||||||
throttle_destroy(&ts);
|
throttle_timers_destroy(&tt);
|
||||||
|
|
||||||
g_assert(final_cfg.buckets[THROTTLE_BPS_TOTAL].avg == 153);
|
g_assert(final_cfg.buckets[THROTTLE_BPS_TOTAL].avg == 153);
|
||||||
g_assert(final_cfg.buckets[THROTTLE_BPS_READ].avg == 56);
|
g_assert(final_cfg.buckets[THROTTLE_BPS_READ].avg == 56);
|
||||||
|
@ -323,43 +331,47 @@ static void test_is_valid(void)
|
||||||
|
|
||||||
static void test_have_timer(void)
|
static void test_have_timer(void)
|
||||||
{
|
{
|
||||||
/* zero the structure */
|
/* zero structures */
|
||||||
memset(&ts, 0, sizeof(ts));
|
memset(&ts, 0, sizeof(ts));
|
||||||
|
memset(&tt, 0, sizeof(tt));
|
||||||
|
|
||||||
/* no timer set should return false */
|
/* no timer set should return false */
|
||||||
g_assert(!throttle_have_timer(&ts));
|
g_assert(!throttle_timers_are_initialized(&tt));
|
||||||
|
|
||||||
/* init the structure */
|
/* init structures */
|
||||||
throttle_init(&ts, ctx, QEMU_CLOCK_VIRTUAL,
|
throttle_init(&ts);
|
||||||
read_timer_cb, write_timer_cb, &ts);
|
throttle_timers_init(&tt, ctx, QEMU_CLOCK_VIRTUAL,
|
||||||
|
read_timer_cb, write_timer_cb, &ts);
|
||||||
|
|
||||||
/* timer set by init should return true */
|
/* timer set by init should return true */
|
||||||
g_assert(throttle_have_timer(&ts));
|
g_assert(throttle_timers_are_initialized(&tt));
|
||||||
|
|
||||||
throttle_destroy(&ts);
|
throttle_timers_destroy(&tt);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void test_detach_attach(void)
|
static void test_detach_attach(void)
|
||||||
{
|
{
|
||||||
/* zero the structure */
|
/* zero structures */
|
||||||
memset(&ts, 0, sizeof(ts));
|
memset(&ts, 0, sizeof(ts));
|
||||||
|
memset(&tt, 0, sizeof(tt));
|
||||||
|
|
||||||
/* init the structure */
|
/* init the structure */
|
||||||
throttle_init(&ts, ctx, QEMU_CLOCK_VIRTUAL,
|
throttle_init(&ts);
|
||||||
read_timer_cb, write_timer_cb, &ts);
|
throttle_timers_init(&tt, ctx, QEMU_CLOCK_VIRTUAL,
|
||||||
|
read_timer_cb, write_timer_cb, &ts);
|
||||||
|
|
||||||
/* timer set by init should return true */
|
/* timer set by init should return true */
|
||||||
g_assert(throttle_have_timer(&ts));
|
g_assert(throttle_timers_are_initialized(&tt));
|
||||||
|
|
||||||
/* timer should no longer exist after detaching */
|
/* timer should no longer exist after detaching */
|
||||||
throttle_detach_aio_context(&ts);
|
throttle_timers_detach_aio_context(&tt);
|
||||||
g_assert(!throttle_have_timer(&ts));
|
g_assert(!throttle_timers_are_initialized(&tt));
|
||||||
|
|
||||||
/* timer should exist again after attaching */
|
/* timer should exist again after attaching */
|
||||||
throttle_attach_aio_context(&ts, ctx);
|
throttle_timers_attach_aio_context(&tt, ctx);
|
||||||
g_assert(throttle_have_timer(&ts));
|
g_assert(throttle_timers_are_initialized(&tt));
|
||||||
|
|
||||||
throttle_destroy(&ts);
|
throttle_timers_destroy(&tt);
|
||||||
}
|
}
|
||||||
|
|
||||||
static bool do_test_accounting(bool is_ops, /* are we testing bps or ops */
|
static bool do_test_accounting(bool is_ops, /* are we testing bps or ops */
|
||||||
|
@ -387,9 +399,10 @@ static bool do_test_accounting(bool is_ops, /* are we testing bps or ops */
|
||||||
|
|
||||||
cfg.op_size = op_size;
|
cfg.op_size = op_size;
|
||||||
|
|
||||||
throttle_init(&ts, ctx, QEMU_CLOCK_VIRTUAL,
|
throttle_init(&ts);
|
||||||
read_timer_cb, write_timer_cb, &ts);
|
throttle_timers_init(&tt, ctx, QEMU_CLOCK_VIRTUAL,
|
||||||
throttle_config(&ts, &cfg);
|
read_timer_cb, write_timer_cb, &ts);
|
||||||
|
throttle_config(&ts, &tt, &cfg);
|
||||||
|
|
||||||
/* account a read */
|
/* account a read */
|
||||||
throttle_account(&ts, false, size);
|
throttle_account(&ts, false, size);
|
||||||
|
@ -414,7 +427,7 @@ static bool do_test_accounting(bool is_ops, /* are we testing bps or ops */
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
throttle_destroy(&ts);
|
throttle_timers_destroy(&tt);
|
||||||
|
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
@ -490,23 +503,80 @@ static void test_accounting(void)
|
||||||
(64.0 / 13)));
|
(64.0 / 13)));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void test_groups(void)
|
||||||
|
{
|
||||||
|
ThrottleConfig cfg1, cfg2;
|
||||||
|
BlockDriverState *bdrv1, *bdrv2, *bdrv3;
|
||||||
|
|
||||||
|
bdrv1 = bdrv_new();
|
||||||
|
bdrv2 = bdrv_new();
|
||||||
|
bdrv3 = bdrv_new();
|
||||||
|
|
||||||
|
g_assert(bdrv1->throttle_state == NULL);
|
||||||
|
g_assert(bdrv2->throttle_state == NULL);
|
||||||
|
g_assert(bdrv3->throttle_state == NULL);
|
||||||
|
|
||||||
|
throttle_group_register_bs(bdrv1, "bar");
|
||||||
|
throttle_group_register_bs(bdrv2, "foo");
|
||||||
|
throttle_group_register_bs(bdrv3, "bar");
|
||||||
|
|
||||||
|
g_assert(bdrv1->throttle_state != NULL);
|
||||||
|
g_assert(bdrv2->throttle_state != NULL);
|
||||||
|
g_assert(bdrv3->throttle_state != NULL);
|
||||||
|
|
||||||
|
g_assert(!strcmp(throttle_group_get_name(bdrv1), "bar"));
|
||||||
|
g_assert(!strcmp(throttle_group_get_name(bdrv2), "foo"));
|
||||||
|
g_assert(bdrv1->throttle_state == bdrv3->throttle_state);
|
||||||
|
|
||||||
|
/* Setting the config of a group member affects the whole group */
|
||||||
|
memset(&cfg1, 0, sizeof(cfg1));
|
||||||
|
cfg1.buckets[THROTTLE_BPS_READ].avg = 500000;
|
||||||
|
cfg1.buckets[THROTTLE_BPS_WRITE].avg = 285000;
|
||||||
|
cfg1.buckets[THROTTLE_OPS_READ].avg = 20000;
|
||||||
|
cfg1.buckets[THROTTLE_OPS_WRITE].avg = 12000;
|
||||||
|
throttle_group_config(bdrv1, &cfg1);
|
||||||
|
|
||||||
|
throttle_group_get_config(bdrv1, &cfg1);
|
||||||
|
throttle_group_get_config(bdrv3, &cfg2);
|
||||||
|
g_assert(!memcmp(&cfg1, &cfg2, sizeof(cfg1)));
|
||||||
|
|
||||||
|
cfg2.buckets[THROTTLE_BPS_READ].avg = 4547;
|
||||||
|
cfg2.buckets[THROTTLE_BPS_WRITE].avg = 1349;
|
||||||
|
cfg2.buckets[THROTTLE_OPS_READ].avg = 123;
|
||||||
|
cfg2.buckets[THROTTLE_OPS_WRITE].avg = 86;
|
||||||
|
throttle_group_config(bdrv3, &cfg1);
|
||||||
|
|
||||||
|
throttle_group_get_config(bdrv1, &cfg1);
|
||||||
|
throttle_group_get_config(bdrv3, &cfg2);
|
||||||
|
g_assert(!memcmp(&cfg1, &cfg2, sizeof(cfg1)));
|
||||||
|
|
||||||
|
throttle_group_unregister_bs(bdrv1);
|
||||||
|
throttle_group_unregister_bs(bdrv2);
|
||||||
|
throttle_group_unregister_bs(bdrv3);
|
||||||
|
|
||||||
|
g_assert(bdrv1->throttle_state == NULL);
|
||||||
|
g_assert(bdrv2->throttle_state == NULL);
|
||||||
|
g_assert(bdrv3->throttle_state == NULL);
|
||||||
|
}
|
||||||
|
|
||||||
int main(int argc, char **argv)
|
int main(int argc, char **argv)
|
||||||
{
|
{
|
||||||
GSource *src;
|
|
||||||
Error *local_error = NULL;
|
Error *local_error = NULL;
|
||||||
|
|
||||||
init_clocks();
|
qemu_init_main_loop(&local_error);
|
||||||
|
ctx = qemu_get_aio_context();
|
||||||
|
|
||||||
ctx = aio_context_new(&local_error);
|
|
||||||
if (!ctx) {
|
if (!ctx) {
|
||||||
error_report("Failed to create AIO Context: '%s'",
|
error_report("Failed to create AIO Context: '%s'",
|
||||||
error_get_pretty(local_error));
|
local_error ? error_get_pretty(local_error) :
|
||||||
error_free(local_error);
|
"Failed to initialize the QEMU main loop");
|
||||||
|
if (local_error) {
|
||||||
|
error_free(local_error);
|
||||||
|
}
|
||||||
exit(1);
|
exit(1);
|
||||||
}
|
}
|
||||||
src = aio_get_g_source(ctx);
|
|
||||||
g_source_attach(src, NULL);
|
bdrv_init();
|
||||||
g_source_unref(src);
|
|
||||||
|
|
||||||
do {} while (g_main_context_iteration(NULL, false));
|
do {} while (g_main_context_iteration(NULL, false));
|
||||||
|
|
||||||
|
@ -523,6 +593,7 @@ int main(int argc, char **argv)
|
||||||
g_test_add_func("/throttle/config/is_valid", test_is_valid);
|
g_test_add_func("/throttle/config/is_valid", test_is_valid);
|
||||||
g_test_add_func("/throttle/config_functions", test_config_functions);
|
g_test_add_func("/throttle/config_functions", test_config_functions);
|
||||||
g_test_add_func("/throttle/accounting", test_accounting);
|
g_test_add_func("/throttle/accounting", test_accounting);
|
||||||
|
g_test_add_func("/throttle/groups", test_groups);
|
||||||
return g_test_run();
|
return g_test_run();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -1,10 +1,12 @@
|
||||||
/*
|
/*
|
||||||
* QEMU throttling infrastructure
|
* QEMU throttling infrastructure
|
||||||
*
|
*
|
||||||
* Copyright (C) Nodalink, SARL. 2013
|
* Copyright (C) Nodalink, EURL. 2013-2014
|
||||||
|
* Copyright (C) Igalia, S.L. 2015
|
||||||
*
|
*
|
||||||
* Author:
|
* Authors:
|
||||||
* Benoît Canet <benoit.canet@irqsave.net>
|
* Benoît Canet <benoit.canet@nodalink.com>
|
||||||
|
* Alberto Garcia <berto@igalia.com>
|
||||||
*
|
*
|
||||||
* This program is free software; you can redistribute it and/or
|
* This program is free software; you can redistribute it and/or
|
||||||
* modify it under the terms of the GNU General Public License as
|
* modify it under the terms of the GNU General Public License as
|
||||||
|
@ -159,29 +161,36 @@ bool throttle_compute_timer(ThrottleState *ts,
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Add timers to event loop */
|
/* Add timers to event loop */
|
||||||
void throttle_attach_aio_context(ThrottleState *ts, AioContext *new_context)
|
void throttle_timers_attach_aio_context(ThrottleTimers *tt,
|
||||||
|
AioContext *new_context)
|
||||||
{
|
{
|
||||||
ts->timers[0] = aio_timer_new(new_context, ts->clock_type, SCALE_NS,
|
tt->timers[0] = aio_timer_new(new_context, tt->clock_type, SCALE_NS,
|
||||||
ts->read_timer_cb, ts->timer_opaque);
|
tt->read_timer_cb, tt->timer_opaque);
|
||||||
ts->timers[1] = aio_timer_new(new_context, ts->clock_type, SCALE_NS,
|
tt->timers[1] = aio_timer_new(new_context, tt->clock_type, SCALE_NS,
|
||||||
ts->write_timer_cb, ts->timer_opaque);
|
tt->write_timer_cb, tt->timer_opaque);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* To be called first on the ThrottleState */
|
/* To be called first on the ThrottleState */
|
||||||
void throttle_init(ThrottleState *ts,
|
void throttle_init(ThrottleState *ts)
|
||||||
AioContext *aio_context,
|
|
||||||
QEMUClockType clock_type,
|
|
||||||
QEMUTimerCB *read_timer_cb,
|
|
||||||
QEMUTimerCB *write_timer_cb,
|
|
||||||
void *timer_opaque)
|
|
||||||
{
|
{
|
||||||
memset(ts, 0, sizeof(ThrottleState));
|
memset(ts, 0, sizeof(ThrottleState));
|
||||||
|
}
|
||||||
|
|
||||||
ts->clock_type = clock_type;
|
/* To be called first on the ThrottleTimers */
|
||||||
ts->read_timer_cb = read_timer_cb;
|
void throttle_timers_init(ThrottleTimers *tt,
|
||||||
ts->write_timer_cb = write_timer_cb;
|
AioContext *aio_context,
|
||||||
ts->timer_opaque = timer_opaque;
|
QEMUClockType clock_type,
|
||||||
throttle_attach_aio_context(ts, aio_context);
|
QEMUTimerCB *read_timer_cb,
|
||||||
|
QEMUTimerCB *write_timer_cb,
|
||||||
|
void *timer_opaque)
|
||||||
|
{
|
||||||
|
memset(tt, 0, sizeof(ThrottleTimers));
|
||||||
|
|
||||||
|
tt->clock_type = clock_type;
|
||||||
|
tt->read_timer_cb = read_timer_cb;
|
||||||
|
tt->write_timer_cb = write_timer_cb;
|
||||||
|
tt->timer_opaque = timer_opaque;
|
||||||
|
throttle_timers_attach_aio_context(tt, aio_context);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* destroy a timer */
|
/* destroy a timer */
|
||||||
|
@ -195,25 +204,25 @@ static void throttle_timer_destroy(QEMUTimer **timer)
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Remove timers from event loop */
|
/* Remove timers from event loop */
|
||||||
void throttle_detach_aio_context(ThrottleState *ts)
|
void throttle_timers_detach_aio_context(ThrottleTimers *tt)
|
||||||
{
|
{
|
||||||
int i;
|
int i;
|
||||||
|
|
||||||
for (i = 0; i < 2; i++) {
|
for (i = 0; i < 2; i++) {
|
||||||
throttle_timer_destroy(&ts->timers[i]);
|
throttle_timer_destroy(&tt->timers[i]);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/* To be called last on the ThrottleState */
|
/* To be called last on the ThrottleTimers */
|
||||||
void throttle_destroy(ThrottleState *ts)
|
void throttle_timers_destroy(ThrottleTimers *tt)
|
||||||
{
|
{
|
||||||
throttle_detach_aio_context(ts);
|
throttle_timers_detach_aio_context(tt);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* is any throttling timer configured */
|
/* is any throttling timer configured */
|
||||||
bool throttle_have_timer(ThrottleState *ts)
|
bool throttle_timers_are_initialized(ThrottleTimers *tt)
|
||||||
{
|
{
|
||||||
if (ts->timers[0]) {
|
if (tt->timers[0]) {
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -324,9 +333,12 @@ static void throttle_cancel_timer(QEMUTimer *timer)
|
||||||
/* Used to configure the throttle
|
/* Used to configure the throttle
|
||||||
*
|
*
|
||||||
* @ts: the throttle state we are working on
|
* @ts: the throttle state we are working on
|
||||||
|
* @tt: the throttle timers we use in this aio context
|
||||||
* @cfg: the config to set
|
* @cfg: the config to set
|
||||||
*/
|
*/
|
||||||
void throttle_config(ThrottleState *ts, ThrottleConfig *cfg)
|
void throttle_config(ThrottleState *ts,
|
||||||
|
ThrottleTimers *tt,
|
||||||
|
ThrottleConfig *cfg)
|
||||||
{
|
{
|
||||||
int i;
|
int i;
|
||||||
|
|
||||||
|
@ -336,10 +348,10 @@ void throttle_config(ThrottleState *ts, ThrottleConfig *cfg)
|
||||||
throttle_fix_bucket(&ts->cfg.buckets[i]);
|
throttle_fix_bucket(&ts->cfg.buckets[i]);
|
||||||
}
|
}
|
||||||
|
|
||||||
ts->previous_leak = qemu_clock_get_ns(ts->clock_type);
|
ts->previous_leak = qemu_clock_get_ns(tt->clock_type);
|
||||||
|
|
||||||
for (i = 0; i < 2; i++) {
|
for (i = 0; i < 2; i++) {
|
||||||
throttle_cancel_timer(ts->timers[i]);
|
throttle_cancel_timer(tt->timers[i]);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -358,12 +370,15 @@ void throttle_get_config(ThrottleState *ts, ThrottleConfig *cfg)
|
||||||
*
|
*
|
||||||
* NOTE: this function is not unit tested due to it's usage of timer_mod
|
* NOTE: this function is not unit tested due to it's usage of timer_mod
|
||||||
*
|
*
|
||||||
|
* @tt: the timers structure
|
||||||
* @is_write: the type of operation (read/write)
|
* @is_write: the type of operation (read/write)
|
||||||
* @ret: true if the timer has been scheduled else false
|
* @ret: true if the timer has been scheduled else false
|
||||||
*/
|
*/
|
||||||
bool throttle_schedule_timer(ThrottleState *ts, bool is_write)
|
bool throttle_schedule_timer(ThrottleState *ts,
|
||||||
|
ThrottleTimers *tt,
|
||||||
|
bool is_write)
|
||||||
{
|
{
|
||||||
int64_t now = qemu_clock_get_ns(ts->clock_type);
|
int64_t now = qemu_clock_get_ns(tt->clock_type);
|
||||||
int64_t next_timestamp;
|
int64_t next_timestamp;
|
||||||
bool must_wait;
|
bool must_wait;
|
||||||
|
|
||||||
|
@ -378,12 +393,12 @@ bool throttle_schedule_timer(ThrottleState *ts, bool is_write)
|
||||||
}
|
}
|
||||||
|
|
||||||
/* request throttled and timer pending -> do nothing */
|
/* request throttled and timer pending -> do nothing */
|
||||||
if (timer_pending(ts->timers[is_write])) {
|
if (timer_pending(tt->timers[is_write])) {
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* request throttled and timer not pending -> arm timer */
|
/* request throttled and timer not pending -> arm timer */
|
||||||
timer_mod(ts->timers[is_write], next_timestamp);
|
timer_mod(tt->timers[is_write], next_timestamp);
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue