aio / timers: Add test harness for AioContext timers
Add a test harness for AioContext timers. The g_source equivalent is unsatisfactory as it suffers from false wakeups. Signed-off-by: Alex Bligh <alex@alex.org.uk> Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
This commit is contained in:
parent
7483d1e547
commit
b53edf971f
134
tests/test-aio.c
134
tests/test-aio.c
@ -47,6 +47,15 @@ typedef struct {
|
||||
int max;
|
||||
} BHTestData;
|
||||
|
||||
typedef struct {
|
||||
QEMUTimer timer;
|
||||
QEMUClockType clock_type;
|
||||
int n;
|
||||
int max;
|
||||
int64_t ns;
|
||||
AioContext *ctx;
|
||||
} TimerTestData;
|
||||
|
||||
static void bh_test_cb(void *opaque)
|
||||
{
|
||||
BHTestData *data = opaque;
|
||||
@ -55,6 +64,24 @@ static void bh_test_cb(void *opaque)
|
||||
}
|
||||
}
|
||||
|
||||
static void timer_test_cb(void *opaque)
|
||||
{
|
||||
TimerTestData *data = opaque;
|
||||
if (++data->n < data->max) {
|
||||
timer_mod(&data->timer,
|
||||
qemu_clock_get_ns(data->clock_type) + data->ns);
|
||||
}
|
||||
}
|
||||
|
||||
static void dummy_io_handler_read(void *opaque)
|
||||
{
|
||||
}
|
||||
|
||||
static int dummy_io_handler_flush(void *opaque)
|
||||
{
|
||||
return 1;
|
||||
}
|
||||
|
||||
static void bh_delete_cb(void *opaque)
|
||||
{
|
||||
BHTestData *data = opaque;
|
||||
@ -343,6 +370,64 @@ static void test_wait_event_notifier_noflush(void)
|
||||
event_notifier_cleanup(&data.e);
|
||||
}
|
||||
|
||||
static void test_timer_schedule(void)
|
||||
{
|
||||
TimerTestData data = { .n = 0, .ctx = ctx, .ns = SCALE_MS * 750LL,
|
||||
.max = 2,
|
||||
.clock_type = QEMU_CLOCK_VIRTUAL };
|
||||
int pipefd[2];
|
||||
|
||||
/* aio_poll will not block to wait for timers to complete unless it has
|
||||
* an fd to wait on. Fixing this breaks other tests. So create a dummy one.
|
||||
*/
|
||||
g_assert(!pipe2(pipefd, O_NONBLOCK));
|
||||
aio_set_fd_handler(ctx, pipefd[0],
|
||||
dummy_io_handler_read, NULL, dummy_io_handler_flush);
|
||||
aio_poll(ctx, false);
|
||||
|
||||
aio_timer_init(ctx, &data.timer, data.clock_type,
|
||||
SCALE_NS, timer_test_cb, &data);
|
||||
timer_mod(&data.timer,
|
||||
qemu_clock_get_ns(data.clock_type) +
|
||||
data.ns);
|
||||
|
||||
g_assert_cmpint(data.n, ==, 0);
|
||||
|
||||
/* timer_mod may well cause an event notifer to have gone off,
|
||||
* so clear that
|
||||
*/
|
||||
do {} while (aio_poll(ctx, false));
|
||||
|
||||
g_assert(!aio_poll(ctx, false));
|
||||
g_assert_cmpint(data.n, ==, 0);
|
||||
|
||||
sleep(1);
|
||||
g_assert_cmpint(data.n, ==, 0);
|
||||
|
||||
g_assert(aio_poll(ctx, false));
|
||||
g_assert_cmpint(data.n, ==, 1);
|
||||
|
||||
/* timer_mod called by our callback */
|
||||
do {} while (aio_poll(ctx, false));
|
||||
|
||||
g_assert(!aio_poll(ctx, false));
|
||||
g_assert_cmpint(data.n, ==, 1);
|
||||
|
||||
g_assert(aio_poll(ctx, true));
|
||||
g_assert_cmpint(data.n, ==, 2);
|
||||
|
||||
/* As max is now 2, an event notifier should not have gone off */
|
||||
|
||||
g_assert(!aio_poll(ctx, false));
|
||||
g_assert_cmpint(data.n, ==, 2);
|
||||
|
||||
aio_set_fd_handler(ctx, pipefd[0], NULL, NULL, NULL);
|
||||
close(pipefd[0]);
|
||||
close(pipefd[1]);
|
||||
|
||||
timer_del(&data.timer);
|
||||
}
|
||||
|
||||
/* Now the same tests, using the context as a GSource. They are
|
||||
* very similar to the ones above, with g_main_context_iteration
|
||||
* replacing aio_poll. However:
|
||||
@ -625,6 +710,53 @@ static void test_source_wait_event_notifier_noflush(void)
|
||||
event_notifier_cleanup(&data.e);
|
||||
}
|
||||
|
||||
static void test_source_timer_schedule(void)
|
||||
{
|
||||
TimerTestData data = { .n = 0, .ctx = ctx, .ns = SCALE_MS * 750LL,
|
||||
.max = 2,
|
||||
.clock_type = QEMU_CLOCK_VIRTUAL };
|
||||
int pipefd[2];
|
||||
int64_t expiry;
|
||||
|
||||
/* aio_poll will not block to wait for timers to complete unless it has
|
||||
* an fd to wait on. Fixing this breaks other tests. So create a dummy one.
|
||||
*/
|
||||
g_assert(!pipe2(pipefd, O_NONBLOCK));
|
||||
aio_set_fd_handler(ctx, pipefd[0],
|
||||
dummy_io_handler_read, NULL, dummy_io_handler_flush);
|
||||
do {} while (g_main_context_iteration(NULL, false));
|
||||
|
||||
aio_timer_init(ctx, &data.timer, data.clock_type,
|
||||
SCALE_NS, timer_test_cb, &data);
|
||||
expiry = qemu_clock_get_ns(data.clock_type) +
|
||||
data.ns;
|
||||
timer_mod(&data.timer, expiry);
|
||||
|
||||
g_assert_cmpint(data.n, ==, 0);
|
||||
|
||||
sleep(1);
|
||||
g_assert_cmpint(data.n, ==, 0);
|
||||
|
||||
g_assert(g_main_context_iteration(NULL, false));
|
||||
g_assert_cmpint(data.n, ==, 1);
|
||||
|
||||
/* The comment above was not kidding when it said this wakes up itself */
|
||||
do {
|
||||
g_assert(g_main_context_iteration(NULL, true));
|
||||
} while (qemu_clock_get_ns(data.clock_type) <= expiry);
|
||||
sleep(1);
|
||||
g_main_context_iteration(NULL, false);
|
||||
|
||||
g_assert_cmpint(data.n, ==, 2);
|
||||
|
||||
aio_set_fd_handler(ctx, pipefd[0], NULL, NULL, NULL);
|
||||
close(pipefd[0]);
|
||||
close(pipefd[1]);
|
||||
|
||||
timer_del(&data.timer);
|
||||
}
|
||||
|
||||
|
||||
/* End of tests. */
|
||||
|
||||
int main(int argc, char **argv)
|
||||
@ -653,6 +785,7 @@ int main(int argc, char **argv)
|
||||
g_test_add_func("/aio/event/wait", test_wait_event_notifier);
|
||||
g_test_add_func("/aio/event/wait/no-flush-cb", test_wait_event_notifier_noflush);
|
||||
g_test_add_func("/aio/event/flush", test_flush_event_notifier);
|
||||
g_test_add_func("/aio/timer/schedule", test_timer_schedule);
|
||||
|
||||
g_test_add_func("/aio-gsource/notify", test_source_notify);
|
||||
g_test_add_func("/aio-gsource/flush", test_source_flush);
|
||||
@ -667,5 +800,6 @@ int main(int argc, char **argv)
|
||||
g_test_add_func("/aio-gsource/event/wait", test_source_wait_event_notifier);
|
||||
g_test_add_func("/aio-gsource/event/wait/no-flush-cb", test_source_wait_event_notifier_noflush);
|
||||
g_test_add_func("/aio-gsource/event/flush", test_source_flush_event_notifier);
|
||||
g_test_add_func("/aio-gsource/timer/schedule", test_source_timer_schedule);
|
||||
return g_test_run();
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user