625a5befc2
When using ballooning to manage overcommitted memory on a host, a system for guests to communicate their memory usage to the host can provide information that will minimize the impact of ballooning on the guests. The current method employs a daemon running in each guest that communicates memory statistics to a host daemon at a specified time interval. The host daemon aggregates this information and inflates and/or deflates balloons according to the level of host memory pressure. This approach is effective but overly complex since a daemon must be installed inside each guest and coordinated to communicate with the host. A simpler approach is to collect memory statistics in the virtio balloon driver and communicate them directly to the hypervisor. Signed-off-by: Adam Litke <agl@us.ibm.com> Signed-off-by: Anthony Liguori <aliguori@us.ibm.com>
312 lines
9.1 KiB
C
312 lines
9.1 KiB
C
/*
|
|
* Virtio Block Device
|
|
*
|
|
* Copyright IBM, Corp. 2008
|
|
*
|
|
* Authors:
|
|
* Anthony Liguori <aliguori@us.ibm.com>
|
|
*
|
|
* This work is licensed under the terms of the GNU GPL, version 2. See
|
|
* the COPYING file in the top-level directory.
|
|
*
|
|
*/
|
|
|
|
#include "qemu-common.h"
|
|
#include "virtio.h"
|
|
#include "pc.h"
|
|
#include "sysemu.h"
|
|
#include "cpu.h"
|
|
#include "monitor.h"
|
|
#include "balloon.h"
|
|
#include "virtio-balloon.h"
|
|
#include "kvm.h"
|
|
#include "qlist.h"
|
|
#include "qint.h"
|
|
#include "qstring.h"
|
|
|
|
#if defined(__linux__)
|
|
#include <sys/mman.h>
|
|
#endif
|
|
|
|
typedef struct VirtIOBalloon
|
|
{
|
|
VirtIODevice vdev;
|
|
VirtQueue *ivq, *dvq, *svq;
|
|
uint32_t num_pages;
|
|
uint32_t actual;
|
|
uint64_t stats[VIRTIO_BALLOON_S_NR];
|
|
VirtQueueElement stats_vq_elem;
|
|
size_t stats_vq_offset;
|
|
MonitorCompletion *stats_callback;
|
|
void *stats_opaque_callback_data;
|
|
} VirtIOBalloon;
|
|
|
|
static VirtIOBalloon *to_virtio_balloon(VirtIODevice *vdev)
|
|
{
|
|
return (VirtIOBalloon *)vdev;
|
|
}
|
|
|
|
static void balloon_page(void *addr, int deflate)
|
|
{
|
|
#if defined(__linux__)
|
|
if (!kvm_enabled() || kvm_has_sync_mmu())
|
|
madvise(addr, TARGET_PAGE_SIZE,
|
|
deflate ? MADV_WILLNEED : MADV_DONTNEED);
|
|
#endif
|
|
}
|
|
|
|
/*
|
|
* reset_stats - Mark all items in the stats array as unset
|
|
*
|
|
* This function needs to be called at device intialization and before
|
|
* before updating to a set of newly-generated stats. This will ensure that no
|
|
* stale values stick around in case the guest reports a subset of the supported
|
|
* statistics.
|
|
*/
|
|
static inline void reset_stats(VirtIOBalloon *dev)
|
|
{
|
|
int i;
|
|
for (i = 0; i < VIRTIO_BALLOON_S_NR; dev->stats[i++] = -1);
|
|
}
|
|
|
|
static void stat_put(QDict *dict, const char *label, uint64_t val)
|
|
{
|
|
if (val != -1)
|
|
qdict_put(dict, label, qint_from_int(val));
|
|
}
|
|
|
|
static QObject *get_stats_qobject(VirtIOBalloon *dev)
|
|
{
|
|
QDict *dict = qdict_new();
|
|
uint32_t actual = ram_size - (dev->actual << VIRTIO_BALLOON_PFN_SHIFT);
|
|
|
|
stat_put(dict, "actual", actual);
|
|
stat_put(dict, "mem_swapped_in", dev->stats[VIRTIO_BALLOON_S_SWAP_IN]);
|
|
stat_put(dict, "mem_swapped_out", dev->stats[VIRTIO_BALLOON_S_SWAP_OUT]);
|
|
stat_put(dict, "major_page_faults", dev->stats[VIRTIO_BALLOON_S_MAJFLT]);
|
|
stat_put(dict, "minor_page_faults", dev->stats[VIRTIO_BALLOON_S_MINFLT]);
|
|
stat_put(dict, "free_mem", dev->stats[VIRTIO_BALLOON_S_MEMFREE]);
|
|
stat_put(dict, "total_mem", dev->stats[VIRTIO_BALLOON_S_MEMTOT]);
|
|
|
|
return QOBJECT(dict);
|
|
}
|
|
|
|
/* FIXME: once we do a virtio refactoring, this will get subsumed into common
|
|
* code */
|
|
static size_t memcpy_from_iovector(void *data, size_t offset, size_t size,
|
|
struct iovec *iov, int iovlen)
|
|
{
|
|
int i;
|
|
uint8_t *ptr = data;
|
|
size_t iov_off = 0;
|
|
size_t data_off = 0;
|
|
|
|
for (i = 0; i < iovlen && size; i++) {
|
|
if (offset < (iov_off + iov[i].iov_len)) {
|
|
size_t len = MIN((iov_off + iov[i].iov_len) - offset , size);
|
|
|
|
memcpy(ptr + data_off, iov[i].iov_base + (offset - iov_off), len);
|
|
|
|
data_off += len;
|
|
offset += len;
|
|
size -= len;
|
|
}
|
|
|
|
iov_off += iov[i].iov_len;
|
|
}
|
|
|
|
return data_off;
|
|
}
|
|
|
|
static void virtio_balloon_handle_output(VirtIODevice *vdev, VirtQueue *vq)
|
|
{
|
|
VirtIOBalloon *s = to_virtio_balloon(vdev);
|
|
VirtQueueElement elem;
|
|
|
|
while (virtqueue_pop(vq, &elem)) {
|
|
size_t offset = 0;
|
|
uint32_t pfn;
|
|
|
|
while (memcpy_from_iovector(&pfn, offset, 4,
|
|
elem.out_sg, elem.out_num) == 4) {
|
|
ram_addr_t pa;
|
|
ram_addr_t addr;
|
|
|
|
pa = (ram_addr_t)ldl_p(&pfn) << VIRTIO_BALLOON_PFN_SHIFT;
|
|
offset += 4;
|
|
|
|
addr = cpu_get_physical_page_desc(pa);
|
|
if ((addr & ~TARGET_PAGE_MASK) != IO_MEM_RAM)
|
|
continue;
|
|
|
|
/* Using qemu_get_ram_ptr is bending the rules a bit, but
|
|
should be OK because we only want a single page. */
|
|
balloon_page(qemu_get_ram_ptr(addr), !!(vq == s->dvq));
|
|
}
|
|
|
|
virtqueue_push(vq, &elem, offset);
|
|
virtio_notify(vdev, vq);
|
|
}
|
|
}
|
|
|
|
static void complete_stats_request(VirtIOBalloon *vb)
|
|
{
|
|
QObject *stats;
|
|
|
|
if (!vb->stats_opaque_callback_data)
|
|
return;
|
|
|
|
stats = get_stats_qobject(vb);
|
|
vb->stats_callback(vb->stats_opaque_callback_data, stats);
|
|
qobject_decref(stats);
|
|
vb->stats_opaque_callback_data = NULL;
|
|
vb->stats_callback = NULL;
|
|
}
|
|
|
|
static void virtio_balloon_receive_stats(VirtIODevice *vdev, VirtQueue *vq)
|
|
{
|
|
VirtIOBalloon *s = DO_UPCAST(VirtIOBalloon, vdev, vdev);
|
|
VirtQueueElement *elem = &s->stats_vq_elem;
|
|
VirtIOBalloonStat stat;
|
|
size_t offset = 0;
|
|
|
|
if (!virtqueue_pop(vq, elem)) {
|
|
return;
|
|
}
|
|
|
|
/* Initialize the stats to get rid of any stale values. This is only
|
|
* needed to handle the case where a guest supports fewer stats than it
|
|
* used to (ie. it has booted into an old kernel).
|
|
*/
|
|
reset_stats(s);
|
|
|
|
while (memcpy_from_iovector(&stat, offset, sizeof(stat), elem->out_sg,
|
|
elem->out_num) == sizeof(stat)) {
|
|
uint16_t tag = tswap16(stat.tag);
|
|
uint64_t val = tswap64(stat.val);
|
|
|
|
offset += sizeof(stat);
|
|
if (tag < VIRTIO_BALLOON_S_NR)
|
|
s->stats[tag] = val;
|
|
}
|
|
s->stats_vq_offset = offset;
|
|
|
|
complete_stats_request(s);
|
|
}
|
|
|
|
static void virtio_balloon_get_config(VirtIODevice *vdev, uint8_t *config_data)
|
|
{
|
|
VirtIOBalloon *dev = to_virtio_balloon(vdev);
|
|
struct virtio_balloon_config config;
|
|
|
|
config.num_pages = cpu_to_le32(dev->num_pages);
|
|
config.actual = cpu_to_le32(dev->actual);
|
|
|
|
memcpy(config_data, &config, 8);
|
|
}
|
|
|
|
static void virtio_balloon_set_config(VirtIODevice *vdev,
|
|
const uint8_t *config_data)
|
|
{
|
|
VirtIOBalloon *dev = to_virtio_balloon(vdev);
|
|
struct virtio_balloon_config config;
|
|
memcpy(&config, config_data, 8);
|
|
dev->actual = config.actual;
|
|
}
|
|
|
|
static uint32_t virtio_balloon_get_features(VirtIODevice *vdev, uint32_t f)
|
|
{
|
|
f |= (1 << VIRTIO_BALLOON_F_STATS_VQ);
|
|
return f;
|
|
}
|
|
|
|
static void virtio_balloon_to_target(void *opaque, ram_addr_t target,
|
|
MonitorCompletion cb, void *cb_data)
|
|
{
|
|
VirtIOBalloon *dev = opaque;
|
|
|
|
if (target > ram_size)
|
|
target = ram_size;
|
|
|
|
if (target) {
|
|
dev->num_pages = (ram_size - target) >> VIRTIO_BALLOON_PFN_SHIFT;
|
|
virtio_notify_config(&dev->vdev);
|
|
} else {
|
|
/* For now, only allow one request at a time. This restriction can be
|
|
* removed later by queueing callback and data pairs.
|
|
*/
|
|
if (dev->stats_callback != NULL) {
|
|
return;
|
|
}
|
|
dev->stats_callback = cb;
|
|
dev->stats_opaque_callback_data = cb_data;
|
|
if (dev->vdev.guest_features & (1 << VIRTIO_BALLOON_F_STATS_VQ)) {
|
|
virtqueue_push(dev->svq, &dev->stats_vq_elem, dev->stats_vq_offset);
|
|
virtio_notify(&dev->vdev, dev->svq);
|
|
} else {
|
|
/* Stats are not supported. Clear out any stale values that might
|
|
* have been set by a more featureful guest kernel.
|
|
*/
|
|
reset_stats(dev);
|
|
complete_stats_request(dev);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void virtio_balloon_save(QEMUFile *f, void *opaque)
|
|
{
|
|
VirtIOBalloon *s = opaque;
|
|
|
|
virtio_save(&s->vdev, f);
|
|
|
|
qemu_put_be32(f, s->num_pages);
|
|
qemu_put_be32(f, s->actual);
|
|
qemu_put_buffer(f, (uint8_t *)&s->stats_vq_elem, sizeof(VirtQueueElement));
|
|
qemu_put_buffer(f, (uint8_t *)&s->stats_vq_offset, sizeof(size_t));
|
|
qemu_put_buffer(f, (uint8_t *)&s->stats_callback, sizeof(MonitorCompletion));
|
|
qemu_put_buffer(f, (uint8_t *)&s->stats_opaque_callback_data, sizeof(void));
|
|
}
|
|
|
|
static int virtio_balloon_load(QEMUFile *f, void *opaque, int version_id)
|
|
{
|
|
VirtIOBalloon *s = opaque;
|
|
|
|
if (version_id != 1)
|
|
return -EINVAL;
|
|
|
|
virtio_load(&s->vdev, f);
|
|
|
|
s->num_pages = qemu_get_be32(f);
|
|
s->actual = qemu_get_be32(f);
|
|
qemu_get_buffer(f, (uint8_t *)&s->stats_vq_elem, sizeof(VirtQueueElement));
|
|
qemu_get_buffer(f, (uint8_t *)&s->stats_vq_offset, sizeof(size_t));
|
|
qemu_get_buffer(f, (uint8_t *)&s->stats_callback, sizeof(MonitorCompletion));
|
|
qemu_get_buffer(f, (uint8_t *)&s->stats_opaque_callback_data, sizeof(void));
|
|
|
|
return 0;
|
|
}
|
|
|
|
VirtIODevice *virtio_balloon_init(DeviceState *dev)
|
|
{
|
|
VirtIOBalloon *s;
|
|
|
|
s = (VirtIOBalloon *)virtio_common_init("virtio-balloon",
|
|
VIRTIO_ID_BALLOON,
|
|
8, sizeof(VirtIOBalloon));
|
|
|
|
s->vdev.get_config = virtio_balloon_get_config;
|
|
s->vdev.set_config = virtio_balloon_set_config;
|
|
s->vdev.get_features = virtio_balloon_get_features;
|
|
|
|
s->ivq = virtio_add_queue(&s->vdev, 128, virtio_balloon_handle_output);
|
|
s->dvq = virtio_add_queue(&s->vdev, 128, virtio_balloon_handle_output);
|
|
s->svq = virtio_add_queue(&s->vdev, 128, virtio_balloon_receive_stats);
|
|
|
|
reset_stats(s);
|
|
qemu_add_balloon_handler(virtio_balloon_to_target, s);
|
|
|
|
register_savevm("virtio-balloon", -1, 1, virtio_balloon_save, virtio_balloon_load, s);
|
|
|
|
return &s->vdev;
|
|
}
|