qemu-e2k/hw/hyperv/vmbus.c
Maciej S. Szmigiero 6093637b4d vmbus: Print a warning when enabled without the recommended set of features
Some Windows versions crash at boot or fail to enable the VMBus device if
they don't see the expected set of Hyper-V features (enlightenments).

Since this provides poor user experience let's warn user if the VMBus
device is enabled without the recommended set of Hyper-V features.

The recommended set is the minimum set of Hyper-V features required to make
the VMBus device work properly in Windows Server versions 2016, 2019 and
2022.

Acked-by: Paolo Bonzini <pbonzini@redhat.com>
Signed-off-by: Maciej S. Szmigiero <maciej.szmigiero@oracle.com>
2024-03-08 14:18:56 +01:00

2694 lines
74 KiB
C

/*
* QEMU Hyper-V VMBus
*
* Copyright (c) 2017-2018 Virtuozzo International GmbH.
*
* This work is licensed under the terms of the GNU GPL, version 2 or later.
* See the COPYING file in the top-level directory.
*/
#include "qemu/osdep.h"
#include "qemu/error-report.h"
#include "qemu/main-loop.h"
#include "qapi/error.h"
#include "migration/vmstate.h"
#include "hw/qdev-properties.h"
#include "hw/qdev-properties-system.h"
#include "hw/hyperv/hyperv.h"
#include "hw/hyperv/vmbus.h"
#include "hw/hyperv/vmbus-bridge.h"
#include "hw/sysbus.h"
#include "cpu.h"
#include "trace.h"
enum {
VMGPADL_INIT,
VMGPADL_ALIVE,
VMGPADL_TEARINGDOWN,
VMGPADL_TORNDOWN,
};
struct VMBusGpadl {
/* GPADL id */
uint32_t id;
/* associated channel id (rudimentary?) */
uint32_t child_relid;
/* number of pages in the GPADL as declared in GPADL_HEADER message */
uint32_t num_gfns;
/*
* Due to limited message size, GPADL may not fit fully in a single
* GPADL_HEADER message, and is further popluated using GPADL_BODY
* messages. @seen_gfns is the number of pages seen so far; once it
* reaches @num_gfns, the GPADL is ready to use.
*/
uint32_t seen_gfns;
/* array of GFNs (of size @num_gfns once allocated) */
uint64_t *gfns;
uint8_t state;
QTAILQ_ENTRY(VMBusGpadl) link;
VMBus *vmbus;
unsigned refcount;
};
/*
* Wrap sequential read from / write to GPADL.
*/
typedef struct GpadlIter {
VMBusGpadl *gpadl;
AddressSpace *as;
DMADirection dir;
/* offset into GPADL where the next i/o will be performed */
uint32_t off;
/*
* Cached mapping of the currently accessed page, up to page boundary.
* Updated lazily on i/o.
* Note: MemoryRegionCache can not be used here because pages in the GPADL
* are non-contiguous and may belong to different memory regions.
*/
void *map;
/* offset after last i/o (i.e. not affected by seek) */
uint32_t last_off;
/*
* Indicator that the iterator is active and may have a cached mapping.
* Allows to enforce bracketing of all i/o (which may create cached
* mappings) and thus exclude mapping leaks.
*/
bool active;
} GpadlIter;
/*
* Ring buffer. There are two of them, sitting in the same GPADL, for each
* channel.
* Each ring buffer consists of a set of pages, with the first page containing
* the ring buffer header, and the remaining pages being for data packets.
*/
typedef struct VMBusRingBufCommon {
AddressSpace *as;
/* GPA of the ring buffer header */
dma_addr_t rb_addr;
/* start and length of the ring buffer data area within GPADL */
uint32_t base;
uint32_t len;
GpadlIter iter;
} VMBusRingBufCommon;
typedef struct VMBusSendRingBuf {
VMBusRingBufCommon common;
/* current write index, to be committed at the end of send */
uint32_t wr_idx;
/* write index at the start of send */
uint32_t last_wr_idx;
/* space to be requested from the guest */
uint32_t wanted;
/* space reserved for planned sends */
uint32_t reserved;
/* last seen read index */
uint32_t last_seen_rd_idx;
} VMBusSendRingBuf;
typedef struct VMBusRecvRingBuf {
VMBusRingBufCommon common;
/* current read index, to be committed at the end of receive */
uint32_t rd_idx;
/* read index at the start of receive */
uint32_t last_rd_idx;
/* last seen write index */
uint32_t last_seen_wr_idx;
} VMBusRecvRingBuf;
enum {
VMOFFER_INIT,
VMOFFER_SENDING,
VMOFFER_SENT,
};
enum {
VMCHAN_INIT,
VMCHAN_OPENING,
VMCHAN_OPEN,
};
struct VMBusChannel {
VMBusDevice *dev;
/* channel id */
uint32_t id;
/*
* subchannel index within the device; subchannel #0 is "primary" and
* always exists
*/
uint16_t subchan_idx;
uint32_t open_id;
/* VP_INDEX of the vCPU to notify with (synthetic) interrupts */
uint32_t target_vp;
/* GPADL id to use for the ring buffers */
uint32_t ringbuf_gpadl;
/* start (in pages) of the send ring buffer within @ringbuf_gpadl */
uint32_t ringbuf_send_offset;
uint8_t offer_state;
uint8_t state;
bool is_open;
/* main device worker; copied from the device class */
VMBusChannelNotifyCb notify_cb;
/*
* guest->host notifications, either sent directly or dispatched via
* interrupt page (older VMBus)
*/
EventNotifier notifier;
VMBus *vmbus;
/*
* SINT route to signal with host->guest notifications; may be shared with
* the main VMBus SINT route
*/
HvSintRoute *notify_route;
VMBusGpadl *gpadl;
VMBusSendRingBuf send_ringbuf;
VMBusRecvRingBuf recv_ringbuf;
QTAILQ_ENTRY(VMBusChannel) link;
};
/*
* Hyper-V spec mandates that every message port has 16 buffers, which means
* that the guest can post up to this many messages without blocking.
* Therefore a queue for incoming messages has to be provided.
* For outgoing (i.e. host->guest) messages there's no queue; the VMBus just
* doesn't transition to a new state until the message is known to have been
* successfully delivered to the respective SynIC message slot.
*/
#define HV_MSG_QUEUE_LEN 16
/* Hyper-V devices never use channel #0. Must be something special. */
#define VMBUS_FIRST_CHANID 1
/* Each channel occupies one bit within a single event page sint slot. */
#define VMBUS_CHANID_COUNT (HV_EVENT_FLAGS_COUNT - VMBUS_FIRST_CHANID)
/* Leave a few connection numbers for other purposes. */
#define VMBUS_CHAN_CONNECTION_OFFSET 16
/*
* Since the success or failure of sending a message is reported
* asynchronously, the VMBus state machine has effectively two entry points:
* vmbus_run and vmbus_msg_cb (the latter is called when the host->guest
* message delivery status becomes known). Both are run as oneshot BHs on the
* main aio context, ensuring serialization.
*/
enum {
VMBUS_LISTEN,
VMBUS_HANDSHAKE,
VMBUS_OFFER,
VMBUS_CREATE_GPADL,
VMBUS_TEARDOWN_GPADL,
VMBUS_OPEN_CHANNEL,
VMBUS_UNLOAD,
VMBUS_STATE_MAX
};
struct VMBus {
BusState parent;
uint8_t state;
/* protection against recursive aio_poll (see vmbus_run) */
bool in_progress;
/* whether there's a message being delivered to the guest */
bool msg_in_progress;
uint32_t version;
/* VP_INDEX of the vCPU to send messages and interrupts to */
uint32_t target_vp;
HvSintRoute *sint_route;
/*
* interrupt page for older protocol versions; newer ones use SynIC event
* flags directly
*/
hwaddr int_page_gpa;
DECLARE_BITMAP(chanid_bitmap, VMBUS_CHANID_COUNT);
/* incoming message queue */
struct hyperv_post_message_input rx_queue[HV_MSG_QUEUE_LEN];
uint8_t rx_queue_head;
uint8_t rx_queue_size;
QemuMutex rx_queue_lock;
QTAILQ_HEAD(, VMBusGpadl) gpadl_list;
QTAILQ_HEAD(, VMBusChannel) channel_list;
/*
* guest->host notifications for older VMBus, to be dispatched via
* interrupt page
*/
EventNotifier notifier;
};
static bool gpadl_full(VMBusGpadl *gpadl)
{
return gpadl->seen_gfns == gpadl->num_gfns;
}
static VMBusGpadl *create_gpadl(VMBus *vmbus, uint32_t id,
uint32_t child_relid, uint32_t num_gfns)
{
VMBusGpadl *gpadl = g_new0(VMBusGpadl, 1);
gpadl->id = id;
gpadl->child_relid = child_relid;
gpadl->num_gfns = num_gfns;
gpadl->gfns = g_new(uint64_t, num_gfns);
QTAILQ_INSERT_HEAD(&vmbus->gpadl_list, gpadl, link);
gpadl->vmbus = vmbus;
gpadl->refcount = 1;
return gpadl;
}
static void free_gpadl(VMBusGpadl *gpadl)
{
QTAILQ_REMOVE(&gpadl->vmbus->gpadl_list, gpadl, link);
g_free(gpadl->gfns);
g_free(gpadl);
}
static VMBusGpadl *find_gpadl(VMBus *vmbus, uint32_t gpadl_id)
{
VMBusGpadl *gpadl;
QTAILQ_FOREACH(gpadl, &vmbus->gpadl_list, link) {
if (gpadl->id == gpadl_id) {
return gpadl;
}
}
return NULL;
}
VMBusGpadl *vmbus_get_gpadl(VMBusChannel *chan, uint32_t gpadl_id)
{
VMBusGpadl *gpadl = find_gpadl(chan->vmbus, gpadl_id);
if (!gpadl || !gpadl_full(gpadl)) {
return NULL;
}
gpadl->refcount++;
return gpadl;
}
void vmbus_put_gpadl(VMBusGpadl *gpadl)
{
if (!gpadl) {
return;
}
if (--gpadl->refcount) {
return;
}
free_gpadl(gpadl);
}
uint32_t vmbus_gpadl_len(VMBusGpadl *gpadl)
{
return gpadl->num_gfns * TARGET_PAGE_SIZE;
}
static void gpadl_iter_init(GpadlIter *iter, VMBusGpadl *gpadl,
AddressSpace *as, DMADirection dir)
{
iter->gpadl = gpadl;
iter->as = as;
iter->dir = dir;
iter->active = false;
}
static inline void gpadl_iter_cache_unmap(GpadlIter *iter)
{
uint32_t map_start_in_page = (uintptr_t)iter->map & ~TARGET_PAGE_MASK;
uint32_t io_end_in_page = ((iter->last_off - 1) & ~TARGET_PAGE_MASK) + 1;
/* mapping is only done to do non-zero amount of i/o */
assert(iter->last_off > 0);
assert(map_start_in_page < io_end_in_page);
dma_memory_unmap(iter->as, iter->map, TARGET_PAGE_SIZE - map_start_in_page,
iter->dir, io_end_in_page - map_start_in_page);
}
/*
* Copy exactly @len bytes between the GPADL pointed to by @iter and @buf.
* The direction of the copy is determined by @iter->dir.
* The caller must ensure the operation overflows neither @buf nor the GPADL
* (there's an assert for the latter).
* Reuse the currently mapped page in the GPADL if possible.
*/
static ssize_t gpadl_iter_io(GpadlIter *iter, void *buf, uint32_t len)
{
ssize_t ret = len;
assert(iter->active);
while (len) {
uint32_t off_in_page = iter->off & ~TARGET_PAGE_MASK;
uint32_t pgleft = TARGET_PAGE_SIZE - off_in_page;
uint32_t cplen = MIN(pgleft, len);
void *p;
/* try to reuse the cached mapping */
if (iter->map) {
uint32_t map_start_in_page =
(uintptr_t)iter->map & ~TARGET_PAGE_MASK;
uint32_t off_base = iter->off & ~TARGET_PAGE_MASK;
uint32_t mapped_base = (iter->last_off - 1) & ~TARGET_PAGE_MASK;
if (off_base != mapped_base || off_in_page < map_start_in_page) {
gpadl_iter_cache_unmap(iter);
iter->map = NULL;
}
}
if (!iter->map) {
dma_addr_t maddr;
dma_addr_t mlen = pgleft;
uint32_t idx = iter->off >> TARGET_PAGE_BITS;
assert(idx < iter->gpadl->num_gfns);
maddr = (iter->gpadl->gfns[idx] << TARGET_PAGE_BITS) | off_in_page;
iter->map = dma_memory_map(iter->as, maddr, &mlen, iter->dir,
MEMTXATTRS_UNSPECIFIED);
if (mlen != pgleft) {
dma_memory_unmap(iter->as, iter->map, mlen, iter->dir, 0);
iter->map = NULL;
return -EFAULT;
}
}
p = (void *)(uintptr_t)(((uintptr_t)iter->map & TARGET_PAGE_MASK) |
off_in_page);
if (iter->dir == DMA_DIRECTION_FROM_DEVICE) {
memcpy(p, buf, cplen);
} else {
memcpy(buf, p, cplen);
}
buf += cplen;
len -= cplen;
iter->off += cplen;
iter->last_off = iter->off;
}
return ret;
}
/*
* Position the iterator @iter at new offset @new_off.
* If this results in the cached mapping being unusable with the new offset,
* unmap it.
*/
static inline void gpadl_iter_seek(GpadlIter *iter, uint32_t new_off)
{
assert(iter->active);
iter->off = new_off;
}
/*
* Start a series of i/o on the GPADL.
* After this i/o and seek operations on @iter become legal.
*/
static inline void gpadl_iter_start_io(GpadlIter *iter)
{
assert(!iter->active);
/* mapping is cached lazily on i/o */
iter->map = NULL;
iter->active = true;
}
/*
* End the eariler started series of i/o on the GPADL and release the cached
* mapping if any.
*/
static inline void gpadl_iter_end_io(GpadlIter *iter)
{
assert(iter->active);
if (iter->map) {
gpadl_iter_cache_unmap(iter);
}
iter->active = false;
}
static void vmbus_resched(VMBus *vmbus);
static void vmbus_msg_cb(void *data, int status);
ssize_t vmbus_iov_to_gpadl(VMBusChannel *chan, VMBusGpadl *gpadl, uint32_t off,
const struct iovec *iov, size_t iov_cnt)
{
GpadlIter iter;
size_t i;
ssize_t ret = 0;
gpadl_iter_init(&iter, gpadl, chan->dev->dma_as,
DMA_DIRECTION_FROM_DEVICE);
gpadl_iter_start_io(&iter);
gpadl_iter_seek(&iter, off);
for (i = 0; i < iov_cnt; i++) {
ret = gpadl_iter_io(&iter, iov[i].iov_base, iov[i].iov_len);
if (ret < 0) {
goto out;
}
}
out:
gpadl_iter_end_io(&iter);
return ret;
}
int vmbus_map_sgl(VMBusChanReq *req, DMADirection dir, struct iovec *iov,
unsigned iov_cnt, size_t len, size_t off)
{
int ret_cnt = 0, ret;
unsigned i;
QEMUSGList *sgl = &req->sgl;
ScatterGatherEntry *sg = sgl->sg;
for (i = 0; i < sgl->nsg; i++) {
if (sg[i].len > off) {
break;
}
off -= sg[i].len;
}
for (; len && i < sgl->nsg; i++) {
dma_addr_t mlen = MIN(sg[i].len - off, len);
dma_addr_t addr = sg[i].base + off;
len -= mlen;
off = 0;
for (; mlen; ret_cnt++) {
dma_addr_t l = mlen;
dma_addr_t a = addr;
if (ret_cnt == iov_cnt) {
ret = -ENOBUFS;
goto err;
}
iov[ret_cnt].iov_base = dma_memory_map(sgl->as, a, &l, dir,
MEMTXATTRS_UNSPECIFIED);
if (!l) {
ret = -EFAULT;
goto err;
}
iov[ret_cnt].iov_len = l;
addr += l;
mlen -= l;
}
}
return ret_cnt;
err:
vmbus_unmap_sgl(req, dir, iov, ret_cnt, 0);
return ret;
}
void vmbus_unmap_sgl(VMBusChanReq *req, DMADirection dir, struct iovec *iov,
unsigned iov_cnt, size_t accessed)
{
QEMUSGList *sgl = &req->sgl;
unsigned i;
for (i = 0; i < iov_cnt; i++) {
size_t acsd = MIN(accessed, iov[i].iov_len);
dma_memory_unmap(sgl->as, iov[i].iov_base, iov[i].iov_len, dir, acsd);
accessed -= acsd;
}
}
static const VMStateDescription vmstate_gpadl = {
.name = "vmbus/gpadl",
.version_id = 0,
.minimum_version_id = 0,
.fields = (const VMStateField[]) {
VMSTATE_UINT32(id, VMBusGpadl),
VMSTATE_UINT32(child_relid, VMBusGpadl),
VMSTATE_UINT32(num_gfns, VMBusGpadl),
VMSTATE_UINT32(seen_gfns, VMBusGpadl),
VMSTATE_VARRAY_UINT32_ALLOC(gfns, VMBusGpadl, num_gfns, 0,
vmstate_info_uint64, uint64_t),
VMSTATE_UINT8(state, VMBusGpadl),
VMSTATE_END_OF_LIST()
}
};
/*
* Wrap the index into a ring buffer of @len bytes.
* @idx is assumed not to exceed twice the size of the ringbuffer, so only
* single wraparound is considered.
*/
static inline uint32_t rb_idx_wrap(uint32_t idx, uint32_t len)
{
if (idx >= len) {
idx -= len;
}
return idx;
}
/*
* Circular difference between two indices into a ring buffer of @len bytes.
* @allow_catchup - whether @idx1 may catch up @idx2; e.g. read index may catch
* up write index but not vice versa.
*/
static inline uint32_t rb_idx_delta(uint32_t idx1, uint32_t idx2, uint32_t len,
bool allow_catchup)
{
return rb_idx_wrap(idx2 + len - idx1 - !allow_catchup, len);
}
static vmbus_ring_buffer *ringbuf_map_hdr(VMBusRingBufCommon *ringbuf)
{
vmbus_ring_buffer *rb;
dma_addr_t mlen = sizeof(*rb);
rb = dma_memory_map(ringbuf->as, ringbuf->rb_addr, &mlen,
DMA_DIRECTION_FROM_DEVICE, MEMTXATTRS_UNSPECIFIED);
if (mlen != sizeof(*rb)) {
dma_memory_unmap(ringbuf->as, rb, mlen,
DMA_DIRECTION_FROM_DEVICE, 0);
return NULL;
}
return rb;
}
static void ringbuf_unmap_hdr(VMBusRingBufCommon *ringbuf,
vmbus_ring_buffer *rb, bool dirty)
{
assert(rb);
dma_memory_unmap(ringbuf->as, rb, sizeof(*rb), DMA_DIRECTION_FROM_DEVICE,
dirty ? sizeof(*rb) : 0);
}
static void ringbuf_init_common(VMBusRingBufCommon *ringbuf, VMBusGpadl *gpadl,
AddressSpace *as, DMADirection dir,
uint32_t begin, uint32_t end)
{
ringbuf->as = as;
ringbuf->rb_addr = gpadl->gfns[begin] << TARGET_PAGE_BITS;
ringbuf->base = (begin + 1) << TARGET_PAGE_BITS;
ringbuf->len = (end - begin - 1) << TARGET_PAGE_BITS;
gpadl_iter_init(&ringbuf->iter, gpadl, as, dir);
}
static int ringbufs_init(VMBusChannel *chan)
{
vmbus_ring_buffer *rb;
VMBusSendRingBuf *send_ringbuf = &chan->send_ringbuf;
VMBusRecvRingBuf *recv_ringbuf = &chan->recv_ringbuf;
if (chan->ringbuf_send_offset <= 1 ||
chan->gpadl->num_gfns <= chan->ringbuf_send_offset + 1) {
return -EINVAL;
}
ringbuf_init_common(&recv_ringbuf->common, chan->gpadl, chan->dev->dma_as,
DMA_DIRECTION_TO_DEVICE, 0, chan->ringbuf_send_offset);
ringbuf_init_common(&send_ringbuf->common, chan->gpadl, chan->dev->dma_as,
DMA_DIRECTION_FROM_DEVICE, chan->ringbuf_send_offset,
chan->gpadl->num_gfns);
send_ringbuf->wanted = 0;
send_ringbuf->reserved = 0;
rb = ringbuf_map_hdr(&recv_ringbuf->common);
if (!rb) {
return -EFAULT;
}
recv_ringbuf->rd_idx = recv_ringbuf->last_rd_idx = rb->read_index;
ringbuf_unmap_hdr(&recv_ringbuf->common, rb, false);
rb = ringbuf_map_hdr(&send_ringbuf->common);
if (!rb) {
return -EFAULT;
}
send_ringbuf->wr_idx = send_ringbuf->last_wr_idx = rb->write_index;
send_ringbuf->last_seen_rd_idx = rb->read_index;
rb->feature_bits |= VMBUS_RING_BUFFER_FEAT_PENDING_SZ;
ringbuf_unmap_hdr(&send_ringbuf->common, rb, true);
if (recv_ringbuf->rd_idx >= recv_ringbuf->common.len ||
send_ringbuf->wr_idx >= send_ringbuf->common.len) {
return -EOVERFLOW;
}
return 0;
}
/*
* Perform io between the GPADL-backed ringbuffer @ringbuf and @buf, wrapping
* around if needed.
* @len is assumed not to exceed the size of the ringbuffer, so only single
* wraparound is considered.
*/
static ssize_t ringbuf_io(VMBusRingBufCommon *ringbuf, void *buf, uint32_t len)
{
ssize_t ret1 = 0, ret2 = 0;
uint32_t remain = ringbuf->len + ringbuf->base - ringbuf->iter.off;
if (len >= remain) {
ret1 = gpadl_iter_io(&ringbuf->iter, buf, remain);
if (ret1 < 0) {
return ret1;
}
gpadl_iter_seek(&ringbuf->iter, ringbuf->base);
buf += remain;
len -= remain;
}
ret2 = gpadl_iter_io(&ringbuf->iter, buf, len);
if (ret2 < 0) {
return ret2;
}
return ret1 + ret2;
}
/*
* Position the circular iterator within @ringbuf to offset @new_off, wrapping
* around if needed.
* @new_off is assumed not to exceed twice the size of the ringbuffer, so only
* single wraparound is considered.
*/
static inline void ringbuf_seek(VMBusRingBufCommon *ringbuf, uint32_t new_off)
{
gpadl_iter_seek(&ringbuf->iter,
ringbuf->base + rb_idx_wrap(new_off, ringbuf->len));
}
static inline uint32_t ringbuf_tell(VMBusRingBufCommon *ringbuf)
{
return ringbuf->iter.off - ringbuf->base;
}
static inline void ringbuf_start_io(VMBusRingBufCommon *ringbuf)
{
gpadl_iter_start_io(&ringbuf->iter);
}
static inline void ringbuf_end_io(VMBusRingBufCommon *ringbuf)
{
gpadl_iter_end_io(&ringbuf->iter);
}
VMBusDevice *vmbus_channel_device(VMBusChannel *chan)
{
return chan->dev;
}
VMBusChannel *vmbus_device_channel(VMBusDevice *dev, uint32_t chan_idx)
{
if (chan_idx >= dev->num_channels) {
return NULL;
}
return &dev->channels[chan_idx];
}
uint32_t vmbus_channel_idx(VMBusChannel *chan)
{
return chan - chan->dev->channels;
}
void vmbus_channel_notify_host(VMBusChannel *chan)
{
event_notifier_set(&chan->notifier);
}
bool vmbus_channel_is_open(VMBusChannel *chan)
{
return chan->is_open;
}
/*
* Notify the guest side about the data to work on in the channel ring buffer.
* The notification is done by signaling a dedicated per-channel SynIC event
* flag (more recent guests) or setting a bit in the interrupt page and firing
* the VMBus SINT (older guests).
*/
static int vmbus_channel_notify_guest(VMBusChannel *chan)
{
int res = 0;
unsigned long *int_map, mask;
unsigned idx;
hwaddr addr = chan->vmbus->int_page_gpa;
hwaddr len = TARGET_PAGE_SIZE / 2, dirty = 0;
trace_vmbus_channel_notify_guest(chan->id);
if (!addr) {
return hyperv_set_event_flag(chan->notify_route, chan->id);
}
int_map = cpu_physical_memory_map(addr, &len, 1);
if (len != TARGET_PAGE_SIZE / 2) {
res = -ENXIO;
goto unmap;
}
idx = BIT_WORD(chan->id);
mask = BIT_MASK(chan->id);
if ((qatomic_fetch_or(&int_map[idx], mask) & mask) != mask) {
res = hyperv_sint_route_set_sint(chan->notify_route);
dirty = len;
}
unmap:
cpu_physical_memory_unmap(int_map, len, 1, dirty);
return res;
}
#define VMBUS_PKT_TRAILER sizeof(uint64_t)
static uint32_t vmbus_pkt_hdr_set_offsets(vmbus_packet_hdr *hdr,
uint32_t desclen, uint32_t msglen)
{
hdr->offset_qwords = sizeof(*hdr) / sizeof(uint64_t) +
DIV_ROUND_UP(desclen, sizeof(uint64_t));
hdr->len_qwords = hdr->offset_qwords +
DIV_ROUND_UP(msglen, sizeof(uint64_t));
return hdr->len_qwords * sizeof(uint64_t) + VMBUS_PKT_TRAILER;
}
/*
* Simplified ring buffer operation with paired barriers annotations in the
* producer and consumer loops:
*
* producer * consumer
* ~~~~~~~~ * ~~~~~~~~
* write pending_send_sz * read write_index
* smp_mb [A] * smp_mb [C]
* read read_index * read packet
* smp_mb [B] * read/write out-of-band data
* read/write out-of-band data * smp_mb [B]
* write packet * write read_index
* smp_mb [C] * smp_mb [A]
* write write_index * read pending_send_sz
* smp_wmb [D] * smp_rmb [D]
* write pending_send_sz * read write_index
* ... * ...
*/
static inline uint32_t ringbuf_send_avail(VMBusSendRingBuf *ringbuf)
{
/* don't trust guest data */
if (ringbuf->last_seen_rd_idx >= ringbuf->common.len) {
return 0;
}
return rb_idx_delta(ringbuf->wr_idx, ringbuf->last_seen_rd_idx,
ringbuf->common.len, false);
}
static ssize_t ringbuf_send_update_idx(VMBusChannel *chan)
{
VMBusSendRingBuf *ringbuf = &chan->send_ringbuf;
vmbus_ring_buffer *rb;
uint32_t written;
written = rb_idx_delta(ringbuf->last_wr_idx, ringbuf->wr_idx,
ringbuf->common.len, true);
if (!written) {
return 0;
}
rb = ringbuf_map_hdr(&ringbuf->common);
if (!rb) {
return -EFAULT;
}
ringbuf->reserved -= written;
/* prevent reorder with the data operation and packet write */
smp_mb(); /* barrier pair [C] */
rb->write_index = ringbuf->wr_idx;
/*
* If the producer earlier indicated that it wants to be notified when the
* consumer frees certain amount of space in the ring buffer, that amount
* is reduced by the size of the completed write.
*/
if (ringbuf->wanted) {
/* otherwise reservation would fail */
assert(ringbuf->wanted < written);
ringbuf->wanted -= written;
/* prevent reorder with write_index write */
smp_wmb(); /* barrier pair [D] */
rb->pending_send_sz = ringbuf->wanted;
}
/* prevent reorder with write_index or pending_send_sz write */
smp_mb(); /* barrier pair [A] */
ringbuf->last_seen_rd_idx = rb->read_index;
/*
* The consumer may have missed the reduction of pending_send_sz and skip
* notification, so re-check the blocking condition, and, if it's no longer
* true, ensure processing another iteration by simulating consumer's
* notification.
*/
if (ringbuf_send_avail(ringbuf) >= ringbuf->wanted) {
vmbus_channel_notify_host(chan);
}
/* skip notification by consumer's request */
if (rb->interrupt_mask) {
goto out;
}
/*
* The consumer hasn't caught up with the producer's previous state so it's
* not blocked.
* (last_seen_rd_idx comes from the guest but it's safe to use w/o
* validation here as it only affects notification.)
*/
if (rb_idx_delta(ringbuf->last_seen_rd_idx, ringbuf->wr_idx,
ringbuf->common.len, true) > written) {
goto out;
}
vmbus_channel_notify_guest(chan);
out:
ringbuf_unmap_hdr(&ringbuf->common, rb, true);
ringbuf->last_wr_idx = ringbuf->wr_idx;
return written;
}
int vmbus_channel_reserve(VMBusChannel *chan,
uint32_t desclen, uint32_t msglen)
{
VMBusSendRingBuf *ringbuf = &chan->send_ringbuf;
vmbus_ring_buffer *rb = NULL;
vmbus_packet_hdr hdr;
uint32_t needed = ringbuf->reserved +
vmbus_pkt_hdr_set_offsets(&hdr, desclen, msglen);
/* avoid touching the guest memory if possible */
if (likely(needed <= ringbuf_send_avail(ringbuf))) {
goto success;
}
rb = ringbuf_map_hdr(&ringbuf->common);
if (!rb) {
return -EFAULT;
}
/* fetch read index from guest memory and try again */
ringbuf->last_seen_rd_idx = rb->read_index;
if (likely(needed <= ringbuf_send_avail(ringbuf))) {
goto success;
}
rb->pending_send_sz = needed;
/*
* The consumer may have made progress and freed up some space before
* seeing updated pending_send_sz, so re-read read_index (preventing
* reorder with the pending_send_sz write) and try again.
*/
smp_mb(); /* barrier pair [A] */
ringbuf->last_seen_rd_idx = rb->read_index;
if (needed > ringbuf_send_avail(ringbuf)) {
goto out;
}
success:
ringbuf->reserved = needed;
needed = 0;
/* clear pending_send_sz if it was set */
if (ringbuf->wanted) {
if (!rb) {
rb = ringbuf_map_hdr(&ringbuf->common);
if (!rb) {
/* failure to clear pending_send_sz is non-fatal */
goto out;
}
}
rb->pending_send_sz = 0;
}
/* prevent reorder of the following data operation with read_index read */
smp_mb(); /* barrier pair [B] */
out:
if (rb) {
ringbuf_unmap_hdr(&ringbuf->common, rb, ringbuf->wanted == needed);
}
ringbuf->wanted = needed;
return needed ? -ENOSPC : 0;
}
ssize_t vmbus_channel_send(VMBusChannel *chan, uint16_t pkt_type,
void *desc, uint32_t desclen,
void *msg, uint32_t msglen,
bool need_comp, uint64_t transaction_id)
{
ssize_t ret = 0;
vmbus_packet_hdr hdr;
uint32_t totlen;
VMBusSendRingBuf *ringbuf = &chan->send_ringbuf;
if (!vmbus_channel_is_open(chan)) {
return -EINVAL;
}
totlen = vmbus_pkt_hdr_set_offsets(&hdr, desclen, msglen);
hdr.type = pkt_type;
hdr.flags = need_comp ? VMBUS_PACKET_FLAG_REQUEST_COMPLETION : 0;
hdr.transaction_id = transaction_id;
assert(totlen <= ringbuf->reserved);
ringbuf_start_io(&ringbuf->common);
ringbuf_seek(&ringbuf->common, ringbuf->wr_idx);
ret = ringbuf_io(&ringbuf->common, &hdr, sizeof(hdr));
if (ret < 0) {
goto out;
}
if (desclen) {
assert(desc);
ret = ringbuf_io(&ringbuf->common, desc, desclen);
if (ret < 0) {
goto out;
}
ringbuf_seek(&ringbuf->common,
ringbuf->wr_idx + hdr.offset_qwords * sizeof(uint64_t));
}
ret = ringbuf_io(&ringbuf->common, msg, msglen);
if (ret < 0) {
goto out;
}
ringbuf_seek(&ringbuf->common, ringbuf->wr_idx + totlen);
ringbuf->wr_idx = ringbuf_tell(&ringbuf->common);
ret = 0;
out:
ringbuf_end_io(&ringbuf->common);
if (ret) {
return ret;
}
return ringbuf_send_update_idx(chan);
}
ssize_t vmbus_channel_send_completion(VMBusChanReq *req,
void *msg, uint32_t msglen)
{
assert(req->need_comp);
return vmbus_channel_send(req->chan, VMBUS_PACKET_COMP, NULL, 0,
msg, msglen, false, req->transaction_id);
}
static int sgl_from_gpa_ranges(QEMUSGList *sgl, VMBusDevice *dev,
VMBusRingBufCommon *ringbuf, uint32_t len)
{
int ret;
vmbus_pkt_gpa_direct hdr;
hwaddr curaddr = 0;
hwaddr curlen = 0;
int num;
if (len < sizeof(hdr)) {
return -EIO;
}
ret = ringbuf_io(ringbuf, &hdr, sizeof(hdr));
if (ret < 0) {
return ret;
}
len -= sizeof(hdr);
num = (len - hdr.rangecount * sizeof(vmbus_gpa_range)) / sizeof(uint64_t);
if (num < 0) {
return -EIO;
}
qemu_sglist_init(sgl, DEVICE(dev), num, ringbuf->as);
for (; hdr.rangecount; hdr.rangecount--) {
vmbus_gpa_range range;
if (len < sizeof(range)) {
goto eio;
}
ret = ringbuf_io(ringbuf, &range, sizeof(range));
if (ret < 0) {
goto err;
}
len -= sizeof(range);
if (range.byte_offset & TARGET_PAGE_MASK) {
goto eio;
}
for (; range.byte_count; range.byte_offset = 0) {
uint64_t paddr;
uint32_t plen = MIN(range.byte_count,
TARGET_PAGE_SIZE - range.byte_offset);
if (len < sizeof(uint64_t)) {
goto eio;
}
ret = ringbuf_io(ringbuf, &paddr, sizeof(paddr));
if (ret < 0) {
goto err;
}
len -= sizeof(uint64_t);
paddr <<= TARGET_PAGE_BITS;
paddr |= range.byte_offset;
range.byte_count -= plen;
if (curaddr + curlen == paddr) {
/* consecutive fragments - join */
curlen += plen;
} else {
if (curlen) {
qemu_sglist_add(sgl, curaddr, curlen);
}
curaddr = paddr;
curlen = plen;
}
}
}
if (curlen) {
qemu_sglist_add(sgl, curaddr, curlen);
}
return 0;
eio:
ret = -EIO;
err:
qemu_sglist_destroy(sgl);
return ret;
}
static VMBusChanReq *vmbus_alloc_req(VMBusChannel *chan,
uint32_t size, uint16_t pkt_type,
uint32_t msglen, uint64_t transaction_id,
bool need_comp)
{
VMBusChanReq *req;
uint32_t msgoff = QEMU_ALIGN_UP(size, __alignof__(*req->msg));
uint32_t totlen = msgoff + msglen;
req = g_malloc0(totlen);
req->chan = chan;
req->pkt_type = pkt_type;
req->msg = (void *)req + msgoff;
req->msglen = msglen;
req->transaction_id = transaction_id;
req->need_comp = need_comp;
return req;
}
int vmbus_channel_recv_start(VMBusChannel *chan)
{
VMBusRecvRingBuf *ringbuf = &chan->recv_ringbuf;
vmbus_ring_buffer *rb;
rb = ringbuf_map_hdr(&ringbuf->common);
if (!rb) {
return -EFAULT;
}
ringbuf->last_seen_wr_idx = rb->write_index;
ringbuf_unmap_hdr(&ringbuf->common, rb, false);
if (ringbuf->last_seen_wr_idx >= ringbuf->common.len) {
return -EOVERFLOW;
}
/* prevent reorder of the following data operation with write_index read */
smp_mb(); /* barrier pair [C] */
return 0;
}
void *vmbus_channel_recv_peek(VMBusChannel *chan, uint32_t size)
{
VMBusRecvRingBuf *ringbuf = &chan->recv_ringbuf;
vmbus_packet_hdr hdr = {};
VMBusChanReq *req;
uint32_t avail;
uint32_t totlen, pktlen, msglen, msgoff, desclen;
assert(size >= sizeof(*req));
/* safe as last_seen_wr_idx is validated in vmbus_channel_recv_start */
avail = rb_idx_delta(ringbuf->rd_idx, ringbuf->last_seen_wr_idx,
ringbuf->common.len, true);
if (avail < sizeof(hdr)) {
return NULL;
}
ringbuf_seek(&ringbuf->common, ringbuf->rd_idx);
if (ringbuf_io(&ringbuf->common, &hdr, sizeof(hdr)) < 0) {
return NULL;
}
pktlen = hdr.len_qwords * sizeof(uint64_t);
totlen = pktlen + VMBUS_PKT_TRAILER;
if (totlen > avail) {
return NULL;
}
msgoff = hdr.offset_qwords * sizeof(uint64_t);
if (msgoff > pktlen || msgoff < sizeof(hdr)) {
error_report("%s: malformed packet: %u %u", __func__, msgoff, pktlen);
return NULL;
}
msglen = pktlen - msgoff;
req = vmbus_alloc_req(chan, size, hdr.type, msglen, hdr.transaction_id,
hdr.flags & VMBUS_PACKET_FLAG_REQUEST_COMPLETION);
switch (hdr.type) {
case VMBUS_PACKET_DATA_USING_GPA_DIRECT:
desclen = msgoff - sizeof(hdr);
if (sgl_from_gpa_ranges(&req->sgl, chan->dev, &ringbuf->common,
desclen) < 0) {
error_report("%s: failed to convert GPA ranges to SGL", __func__);
goto free_req;
}
break;
case VMBUS_PACKET_DATA_INBAND:
case VMBUS_PACKET_COMP:
break;
default:
error_report("%s: unexpected msg type: %x", __func__, hdr.type);
goto free_req;
}
ringbuf_seek(&ringbuf->common, ringbuf->rd_idx + msgoff);
if (ringbuf_io(&ringbuf->common, req->msg, msglen) < 0) {
goto free_req;
}
ringbuf_seek(&ringbuf->common, ringbuf->rd_idx + totlen);
return req;
free_req:
vmbus_free_req(req);
return NULL;
}
void vmbus_channel_recv_pop(VMBusChannel *chan)
{
VMBusRecvRingBuf *ringbuf = &chan->recv_ringbuf;
ringbuf->rd_idx = ringbuf_tell(&ringbuf->common);
}
ssize_t vmbus_channel_recv_done(VMBusChannel *chan)
{
VMBusRecvRingBuf *ringbuf = &chan->recv_ringbuf;
vmbus_ring_buffer *rb;
uint32_t read;
read = rb_idx_delta(ringbuf->last_rd_idx, ringbuf->rd_idx,
ringbuf->common.len, true);
if (!read) {
return 0;
}
rb = ringbuf_map_hdr(&ringbuf->common);
if (!rb) {
return -EFAULT;
}
/* prevent reorder with the data operation and packet read */
smp_mb(); /* barrier pair [B] */
rb->read_index = ringbuf->rd_idx;
/* prevent reorder of the following pending_send_sz read */
smp_mb(); /* barrier pair [A] */
if (rb->interrupt_mask) {
goto out;
}
if (rb->feature_bits & VMBUS_RING_BUFFER_FEAT_PENDING_SZ) {
uint32_t wr_idx, wr_avail;
uint32_t wanted = rb->pending_send_sz;
if (!wanted) {
goto out;
}
/* prevent reorder with pending_send_sz read */
smp_rmb(); /* barrier pair [D] */
wr_idx = rb->write_index;
wr_avail = rb_idx_delta(wr_idx, ringbuf->rd_idx, ringbuf->common.len,
true);
/* the producer wasn't blocked on the consumer state */
if (wr_avail >= read + wanted) {
goto out;
}
/* there's not enough space for the producer to make progress */
if (wr_avail < wanted) {
goto out;
}
}
vmbus_channel_notify_guest(chan);
out:
ringbuf_unmap_hdr(&ringbuf->common, rb, true);
ringbuf->last_rd_idx = ringbuf->rd_idx;
return read;
}
void vmbus_free_req(void *req)
{
VMBusChanReq *r = req;
if (!req) {
return;
}
if (r->sgl.dev) {
qemu_sglist_destroy(&r->sgl);
}
g_free(req);
}
static void channel_event_cb(EventNotifier *e)
{
VMBusChannel *chan = container_of(e, VMBusChannel, notifier);
if (event_notifier_test_and_clear(e)) {
/*
* All receives are supposed to happen within the device worker, so
* bracket it with ringbuf_start/end_io on the receive ringbuffer, and
* potentially reuse the cached mapping throughout the worker.
* Can't do this for sends as they may happen outside the device
* worker.
*/
VMBusRecvRingBuf *ringbuf = &chan->recv_ringbuf;
ringbuf_start_io(&ringbuf->common);
chan->notify_cb(chan);
ringbuf_end_io(&ringbuf->common);
}
}
static int alloc_chan_id(VMBus *vmbus)
{
int ret;
ret = find_next_zero_bit(vmbus->chanid_bitmap, VMBUS_CHANID_COUNT, 0);
if (ret == VMBUS_CHANID_COUNT) {
return -ENOMEM;
}
return ret + VMBUS_FIRST_CHANID;
}
static int register_chan_id(VMBusChannel *chan)
{
return test_and_set_bit(chan->id - VMBUS_FIRST_CHANID,
chan->vmbus->chanid_bitmap) ? -EEXIST : 0;
}
static void unregister_chan_id(VMBusChannel *chan)
{
clear_bit(chan->id - VMBUS_FIRST_CHANID, chan->vmbus->chanid_bitmap);
}
static uint32_t chan_connection_id(VMBusChannel *chan)
{
return VMBUS_CHAN_CONNECTION_OFFSET + chan->id;
}
static void init_channel(VMBus *vmbus, VMBusDevice *dev, VMBusDeviceClass *vdc,
VMBusChannel *chan, uint16_t idx, Error **errp)
{
int res;
chan->dev = dev;
chan->notify_cb = vdc->chan_notify_cb;
chan->subchan_idx = idx;
chan->vmbus = vmbus;
res = alloc_chan_id(vmbus);
if (res < 0) {
error_setg(errp, "no spare channel id");
return;
}
chan->id = res;
register_chan_id(chan);
/*
* The guest drivers depend on the device subchannels (idx #1+) to be
* offered after the primary channel (idx #0) of that device. To ensure
* that, record the channels on the channel list in the order they appear
* within the device.
*/
QTAILQ_INSERT_TAIL(&vmbus->channel_list, chan, link);
}
static void deinit_channel(VMBusChannel *chan)
{
assert(chan->state == VMCHAN_INIT);
QTAILQ_REMOVE(&chan->vmbus->channel_list, chan, link);
unregister_chan_id(chan);
}
static void create_channels(VMBus *vmbus, VMBusDevice *dev, Error **errp)
{
uint16_t i;
VMBusDeviceClass *vdc = VMBUS_DEVICE_GET_CLASS(dev);
Error *err = NULL;
dev->num_channels = vdc->num_channels ? vdc->num_channels(dev) : 1;
if (dev->num_channels < 1) {
error_setg(errp, "invalid #channels: %u", dev->num_channels);
return;
}
dev->channels = g_new0(VMBusChannel, dev->num_channels);
for (i = 0; i < dev->num_channels; i++) {
init_channel(vmbus, dev, vdc, &dev->channels[i], i, &err);
if (err) {
goto err_init;
}
}
return;
err_init:
while (i--) {
deinit_channel(&dev->channels[i]);
}
error_propagate(errp, err);
}
static void free_channels(VMBusDevice *dev)
{
uint16_t i;
for (i = 0; i < dev->num_channels; i++) {
deinit_channel(&dev->channels[i]);
}
g_free(dev->channels);
}
static HvSintRoute *make_sint_route(VMBus *vmbus, uint32_t vp_index)
{
VMBusChannel *chan;
if (vp_index == vmbus->target_vp) {
hyperv_sint_route_ref(vmbus->sint_route);
return vmbus->sint_route;
}
QTAILQ_FOREACH(chan, &vmbus->channel_list, link) {
if (chan->target_vp == vp_index && vmbus_channel_is_open(chan)) {
hyperv_sint_route_ref(chan->notify_route);
return chan->notify_route;
}
}
return hyperv_sint_route_new(vp_index, VMBUS_SINT, NULL, NULL);
}
static void open_channel(VMBusChannel *chan)
{
VMBusDeviceClass *vdc = VMBUS_DEVICE_GET_CLASS(chan->dev);
chan->gpadl = vmbus_get_gpadl(chan, chan->ringbuf_gpadl);
if (!chan->gpadl) {
return;
}
if (ringbufs_init(chan)) {
goto put_gpadl;
}
if (event_notifier_init(&chan->notifier, 0)) {
goto put_gpadl;
}
event_notifier_set_handler(&chan->notifier, channel_event_cb);
if (hyperv_set_event_flag_handler(chan_connection_id(chan),
&chan->notifier)) {
goto cleanup_notifier;
}
chan->notify_route = make_sint_route(chan->vmbus, chan->target_vp);
if (!chan->notify_route) {
goto clear_event_flag_handler;
}
if (vdc->open_channel && vdc->open_channel(chan)) {
goto unref_sint_route;
}
chan->is_open = true;
return;
unref_sint_route:
hyperv_sint_route_unref(chan->notify_route);
clear_event_flag_handler:
hyperv_set_event_flag_handler(chan_connection_id(chan), NULL);
cleanup_notifier:
event_notifier_set_handler(&chan->notifier, NULL);
event_notifier_cleanup(&chan->notifier);
put_gpadl:
vmbus_put_gpadl(chan->gpadl);
}
static void close_channel(VMBusChannel *chan)
{
VMBusDeviceClass *vdc = VMBUS_DEVICE_GET_CLASS(chan->dev);
if (!chan->is_open) {
return;
}
if (vdc->close_channel) {
vdc->close_channel(chan);
}
hyperv_sint_route_unref(chan->notify_route);
hyperv_set_event_flag_handler(chan_connection_id(chan), NULL);
event_notifier_set_handler(&chan->notifier, NULL);
event_notifier_cleanup(&chan->notifier);
vmbus_put_gpadl(chan->gpadl);
chan->is_open = false;
}
static int channel_post_load(void *opaque, int version_id)
{
VMBusChannel *chan = opaque;
return register_chan_id(chan);
}
static const VMStateDescription vmstate_channel = {
.name = "vmbus/channel",
.version_id = 0,
.minimum_version_id = 0,
.post_load = channel_post_load,
.fields = (const VMStateField[]) {
VMSTATE_UINT32(id, VMBusChannel),
VMSTATE_UINT16(subchan_idx, VMBusChannel),
VMSTATE_UINT32(open_id, VMBusChannel),
VMSTATE_UINT32(target_vp, VMBusChannel),
VMSTATE_UINT32(ringbuf_gpadl, VMBusChannel),
VMSTATE_UINT32(ringbuf_send_offset, VMBusChannel),
VMSTATE_UINT8(offer_state, VMBusChannel),
VMSTATE_UINT8(state, VMBusChannel),
VMSTATE_END_OF_LIST()
}
};
static VMBusChannel *find_channel(VMBus *vmbus, uint32_t id)
{
VMBusChannel *chan;
QTAILQ_FOREACH(chan, &vmbus->channel_list, link) {
if (chan->id == id) {
return chan;
}
}
return NULL;
}
static int enqueue_incoming_message(VMBus *vmbus,
const struct hyperv_post_message_input *msg)
{
int ret = 0;
uint8_t idx, prev_size;
qemu_mutex_lock(&vmbus->rx_queue_lock);
if (vmbus->rx_queue_size == HV_MSG_QUEUE_LEN) {
ret = -ENOBUFS;
goto out;
}
prev_size = vmbus->rx_queue_size;
idx = (vmbus->rx_queue_head + vmbus->rx_queue_size) % HV_MSG_QUEUE_LEN;
memcpy(&vmbus->rx_queue[idx], msg, sizeof(*msg));
vmbus->rx_queue_size++;
/* only need to resched if the queue was empty before */
if (!prev_size) {
vmbus_resched(vmbus);
}
out:
qemu_mutex_unlock(&vmbus->rx_queue_lock);
return ret;
}
static uint16_t vmbus_recv_message(const struct hyperv_post_message_input *msg,
void *data)
{
VMBus *vmbus = data;
struct vmbus_message_header *vmbus_msg;
if (msg->message_type != HV_MESSAGE_VMBUS) {
return HV_STATUS_INVALID_HYPERCALL_INPUT;
}
if (msg->payload_size < sizeof(struct vmbus_message_header)) {
return HV_STATUS_INVALID_HYPERCALL_INPUT;
}
vmbus_msg = (struct vmbus_message_header *)msg->payload;
trace_vmbus_recv_message(vmbus_msg->message_type, msg->payload_size);
if (vmbus_msg->message_type == VMBUS_MSG_INVALID ||
vmbus_msg->message_type >= VMBUS_MSG_COUNT) {
error_report("vmbus: unknown message type %#x",
vmbus_msg->message_type);
return HV_STATUS_INVALID_HYPERCALL_INPUT;
}
if (enqueue_incoming_message(vmbus, msg)) {
return HV_STATUS_INSUFFICIENT_BUFFERS;
}
return HV_STATUS_SUCCESS;
}
static bool vmbus_initialized(VMBus *vmbus)
{
return vmbus->version > 0 && vmbus->version <= VMBUS_VERSION_CURRENT;
}
static void vmbus_reset_all(VMBus *vmbus)
{
bus_cold_reset(BUS(vmbus));
}
static void post_msg(VMBus *vmbus, void *msgdata, uint32_t msglen)
{
int ret;
struct hyperv_message msg = {
.header.message_type = HV_MESSAGE_VMBUS,
};
assert(!vmbus->msg_in_progress);
assert(msglen <= sizeof(msg.payload));
assert(msglen >= sizeof(struct vmbus_message_header));
vmbus->msg_in_progress = true;
trace_vmbus_post_msg(((struct vmbus_message_header *)msgdata)->message_type,
msglen);
memcpy(msg.payload, msgdata, msglen);
msg.header.payload_size = ROUND_UP(msglen, VMBUS_MESSAGE_SIZE_ALIGN);
ret = hyperv_post_msg(vmbus->sint_route, &msg);
if (ret == 0 || ret == -EAGAIN) {
return;
}
error_report("message delivery fatal failure: %d; aborting vmbus", ret);
vmbus_reset_all(vmbus);
}
static int vmbus_init(VMBus *vmbus)
{
if (vmbus->target_vp != (uint32_t)-1) {
vmbus->sint_route = hyperv_sint_route_new(vmbus->target_vp, VMBUS_SINT,
vmbus_msg_cb, vmbus);
if (!vmbus->sint_route) {
error_report("failed to set up SINT route");
return -ENOMEM;
}
}
return 0;
}
static void vmbus_deinit(VMBus *vmbus)
{
VMBusGpadl *gpadl, *tmp_gpadl;
VMBusChannel *chan;
QTAILQ_FOREACH_SAFE(gpadl, &vmbus->gpadl_list, link, tmp_gpadl) {
if (gpadl->state == VMGPADL_TORNDOWN) {
continue;
}
vmbus_put_gpadl(gpadl);
}
QTAILQ_FOREACH(chan, &vmbus->channel_list, link) {
chan->offer_state = VMOFFER_INIT;
}
hyperv_sint_route_unref(vmbus->sint_route);
vmbus->sint_route = NULL;
vmbus->int_page_gpa = 0;
vmbus->target_vp = (uint32_t)-1;
vmbus->version = 0;
vmbus->state = VMBUS_LISTEN;
vmbus->msg_in_progress = false;
}
static void handle_initiate_contact(VMBus *vmbus,
vmbus_message_initiate_contact *msg,
uint32_t msglen)
{
if (msglen < sizeof(*msg)) {
return;
}
trace_vmbus_initiate_contact(msg->version_requested >> 16,
msg->version_requested & 0xffff,
msg->target_vcpu, msg->monitor_page1,
msg->monitor_page2, msg->interrupt_page);
/*
* Reset vmbus on INITIATE_CONTACT regardless of its previous state.
* Useful, in particular, with vmbus-aware BIOS which can't shut vmbus down
* before handing over to OS loader.
*/
vmbus_reset_all(vmbus);
vmbus->target_vp = msg->target_vcpu;
vmbus->version = msg->version_requested;
if (vmbus->version < VMBUS_VERSION_WIN8) {
/* linux passes interrupt page even when it doesn't need it */
vmbus->int_page_gpa = msg->interrupt_page;
}
vmbus->state = VMBUS_HANDSHAKE;
if (vmbus_init(vmbus)) {
error_report("failed to init vmbus; aborting");
vmbus_deinit(vmbus);
return;
}
}
static void send_handshake(VMBus *vmbus)
{
struct vmbus_message_version_response msg = {
.header.message_type = VMBUS_MSG_VERSION_RESPONSE,
.version_supported = vmbus_initialized(vmbus),
};
post_msg(vmbus, &msg, sizeof(msg));
}
static void handle_request_offers(VMBus *vmbus, void *msgdata, uint32_t msglen)
{
VMBusChannel *chan;
if (!vmbus_initialized(vmbus)) {
return;
}
QTAILQ_FOREACH(chan, &vmbus->channel_list, link) {
if (chan->offer_state == VMOFFER_INIT) {
chan->offer_state = VMOFFER_SENDING;
break;
}
}
vmbus->state = VMBUS_OFFER;
}
static void send_offer(VMBus *vmbus)
{
VMBusChannel *chan;
struct vmbus_message_header alloffers_msg = {
.message_type = VMBUS_MSG_ALLOFFERS_DELIVERED,
};
QTAILQ_FOREACH(chan, &vmbus->channel_list, link) {
if (chan->offer_state == VMOFFER_SENDING) {
VMBusDeviceClass *vdc = VMBUS_DEVICE_GET_CLASS(chan->dev);
/* Hyper-V wants LE GUIDs */
QemuUUID classid = qemu_uuid_bswap(vdc->classid);
QemuUUID instanceid = qemu_uuid_bswap(chan->dev->instanceid);
struct vmbus_message_offer_channel msg = {
.header.message_type = VMBUS_MSG_OFFERCHANNEL,
.child_relid = chan->id,
.connection_id = chan_connection_id(chan),
.channel_flags = vdc->channel_flags,
.mmio_size_mb = vdc->mmio_size_mb,
.sub_channel_index = vmbus_channel_idx(chan),
.interrupt_flags = VMBUS_OFFER_INTERRUPT_DEDICATED,
};
memcpy(msg.type_uuid, &classid, sizeof(classid));
memcpy(msg.instance_uuid, &instanceid, sizeof(instanceid));
trace_vmbus_send_offer(chan->id, chan->dev);
post_msg(vmbus, &msg, sizeof(msg));
return;
}
}
/* no more offers, send terminator message */
trace_vmbus_terminate_offers();
post_msg(vmbus, &alloffers_msg, sizeof(alloffers_msg));
}
static bool complete_offer(VMBus *vmbus)
{
VMBusChannel *chan;
QTAILQ_FOREACH(chan, &vmbus->channel_list, link) {
if (chan->offer_state == VMOFFER_SENDING) {
chan->offer_state = VMOFFER_SENT;
goto next_offer;
}
}
/*
* no transitioning channels found so this is completing the terminator
* message, and vmbus can move to the next state
*/
return true;
next_offer:
/* try to mark another channel for offering */
QTAILQ_FOREACH(chan, &vmbus->channel_list, link) {
if (chan->offer_state == VMOFFER_INIT) {
chan->offer_state = VMOFFER_SENDING;
break;
}
}
/*
* if an offer has been sent there are more offers or the terminator yet to
* send, so no state transition for vmbus
*/
return false;
}
static void handle_gpadl_header(VMBus *vmbus, vmbus_message_gpadl_header *msg,
uint32_t msglen)
{
VMBusGpadl *gpadl;
uint32_t num_gfns, i;
/* must include at least one gpa range */
if (msglen < sizeof(*msg) + sizeof(msg->range[0]) ||
!vmbus_initialized(vmbus)) {
return;
}
num_gfns = (msg->range_buflen - msg->rangecount * sizeof(msg->range[0])) /
sizeof(msg->range[0].pfn_array[0]);
trace_vmbus_gpadl_header(msg->gpadl_id, num_gfns);
/*
* In theory the GPADL_HEADER message can define a GPADL with multiple GPA
* ranges each with arbitrary size and alignment. However in practice only
* single-range page-aligned GPADLs have been observed so just ignore
* anything else and simplify things greatly.
*/
if (msg->rangecount != 1 || msg->range[0].byte_offset ||
(msg->range[0].byte_count != (num_gfns << TARGET_PAGE_BITS))) {
return;
}
/* ignore requests to create already existing GPADLs */
if (find_gpadl(vmbus, msg->gpadl_id)) {
return;
}
gpadl = create_gpadl(vmbus, msg->gpadl_id, msg->child_relid, num_gfns);
for (i = 0; i < num_gfns &&
(void *)&msg->range[0].pfn_array[i + 1] <= (void *)msg + msglen;
i++) {
gpadl->gfns[gpadl->seen_gfns++] = msg->range[0].pfn_array[i];
}
if (gpadl_full(gpadl)) {
vmbus->state = VMBUS_CREATE_GPADL;
}
}
static void handle_gpadl_body(VMBus *vmbus, vmbus_message_gpadl_body *msg,
uint32_t msglen)
{
VMBusGpadl *gpadl;
uint32_t num_gfns_left, i;
if (msglen < sizeof(*msg) || !vmbus_initialized(vmbus)) {
return;
}
trace_vmbus_gpadl_body(msg->gpadl_id);
gpadl = find_gpadl(vmbus, msg->gpadl_id);
if (!gpadl) {
return;
}
num_gfns_left = gpadl->num_gfns - gpadl->seen_gfns;
assert(num_gfns_left);
for (i = 0; i < num_gfns_left &&
(void *)&msg->pfn_array[i + 1] <= (void *)msg + msglen; i++) {
gpadl->gfns[gpadl->seen_gfns++] = msg->pfn_array[i];
}
if (gpadl_full(gpadl)) {
vmbus->state = VMBUS_CREATE_GPADL;
}
}
static void send_create_gpadl(VMBus *vmbus)
{
VMBusGpadl *gpadl;
QTAILQ_FOREACH(gpadl, &vmbus->gpadl_list, link) {
if (gpadl_full(gpadl) && gpadl->state == VMGPADL_INIT) {
struct vmbus_message_gpadl_created msg = {
.header.message_type = VMBUS_MSG_GPADL_CREATED,
.gpadl_id = gpadl->id,
.child_relid = gpadl->child_relid,
};
trace_vmbus_gpadl_created(gpadl->id);
post_msg(vmbus, &msg, sizeof(msg));
return;
}
}
assert(false);
}
static bool complete_create_gpadl(VMBus *vmbus)
{
VMBusGpadl *gpadl;
QTAILQ_FOREACH(gpadl, &vmbus->gpadl_list, link) {
if (gpadl_full(gpadl) && gpadl->state == VMGPADL_INIT) {
gpadl->state = VMGPADL_ALIVE;
return true;
}
}
assert(false);
return false;
}
static void handle_gpadl_teardown(VMBus *vmbus,
vmbus_message_gpadl_teardown *msg,
uint32_t msglen)
{
VMBusGpadl *gpadl;
if (msglen < sizeof(*msg) || !vmbus_initialized(vmbus)) {
return;
}
trace_vmbus_gpadl_teardown(msg->gpadl_id);
gpadl = find_gpadl(vmbus, msg->gpadl_id);
if (!gpadl || gpadl->state == VMGPADL_TORNDOWN) {
return;
}
gpadl->state = VMGPADL_TEARINGDOWN;
vmbus->state = VMBUS_TEARDOWN_GPADL;
}
static void send_teardown_gpadl(VMBus *vmbus)
{
VMBusGpadl *gpadl;
QTAILQ_FOREACH(gpadl, &vmbus->gpadl_list, link) {
if (gpadl->state == VMGPADL_TEARINGDOWN) {
struct vmbus_message_gpadl_torndown msg = {
.header.message_type = VMBUS_MSG_GPADL_TORNDOWN,
.gpadl_id = gpadl->id,
};
trace_vmbus_gpadl_torndown(gpadl->id);
post_msg(vmbus, &msg, sizeof(msg));
return;
}
}
assert(false);
}
static bool complete_teardown_gpadl(VMBus *vmbus)
{
VMBusGpadl *gpadl;
QTAILQ_FOREACH(gpadl, &vmbus->gpadl_list, link) {
if (gpadl->state == VMGPADL_TEARINGDOWN) {
gpadl->state = VMGPADL_TORNDOWN;
vmbus_put_gpadl(gpadl);
return true;
}
}
assert(false);
return false;
}
static void handle_open_channel(VMBus *vmbus, vmbus_message_open_channel *msg,
uint32_t msglen)
{
VMBusChannel *chan;
if (msglen < sizeof(*msg) || !vmbus_initialized(vmbus)) {
return;
}
trace_vmbus_open_channel(msg->child_relid, msg->ring_buffer_gpadl_id,
msg->target_vp);
chan = find_channel(vmbus, msg->child_relid);
if (!chan || chan->state != VMCHAN_INIT) {
return;
}
chan->ringbuf_gpadl = msg->ring_buffer_gpadl_id;
chan->ringbuf_send_offset = msg->ring_buffer_offset;
chan->target_vp = msg->target_vp;
chan->open_id = msg->open_id;
open_channel(chan);
chan->state = VMCHAN_OPENING;
vmbus->state = VMBUS_OPEN_CHANNEL;
}
static void send_open_channel(VMBus *vmbus)
{
VMBusChannel *chan;
QTAILQ_FOREACH(chan, &vmbus->channel_list, link) {
if (chan->state == VMCHAN_OPENING) {
struct vmbus_message_open_result msg = {
.header.message_type = VMBUS_MSG_OPENCHANNEL_RESULT,
.child_relid = chan->id,
.open_id = chan->open_id,
.status = !vmbus_channel_is_open(chan),
};
trace_vmbus_channel_open(chan->id, msg.status);
post_msg(vmbus, &msg, sizeof(msg));
return;
}
}
assert(false);
}
static bool complete_open_channel(VMBus *vmbus)
{
VMBusChannel *chan;
QTAILQ_FOREACH(chan, &vmbus->channel_list, link) {
if (chan->state == VMCHAN_OPENING) {
if (vmbus_channel_is_open(chan)) {
chan->state = VMCHAN_OPEN;
/*
* simulate guest notification of ringbuffer space made
* available, for the channel protocols where the host
* initiates the communication
*/
vmbus_channel_notify_host(chan);
} else {
chan->state = VMCHAN_INIT;
}
return true;
}
}
assert(false);
return false;
}
static void vdev_reset_on_close(VMBusDevice *vdev)
{
uint16_t i;
for (i = 0; i < vdev->num_channels; i++) {
if (vmbus_channel_is_open(&vdev->channels[i])) {
return;
}
}
/* all channels closed -- reset device */
device_cold_reset(DEVICE(vdev));
}
static void handle_close_channel(VMBus *vmbus, vmbus_message_close_channel *msg,
uint32_t msglen)
{
VMBusChannel *chan;
if (msglen < sizeof(*msg) || !vmbus_initialized(vmbus)) {
return;
}
trace_vmbus_close_channel(msg->child_relid);
chan = find_channel(vmbus, msg->child_relid);
if (!chan) {
return;
}
close_channel(chan);
chan->state = VMCHAN_INIT;
vdev_reset_on_close(chan->dev);
}
static void handle_unload(VMBus *vmbus, void *msg, uint32_t msglen)
{
vmbus->state = VMBUS_UNLOAD;
}
static void send_unload(VMBus *vmbus)
{
vmbus_message_header msg = {
.message_type = VMBUS_MSG_UNLOAD_RESPONSE,
};
qemu_mutex_lock(&vmbus->rx_queue_lock);
vmbus->rx_queue_size = 0;
qemu_mutex_unlock(&vmbus->rx_queue_lock);
post_msg(vmbus, &msg, sizeof(msg));
return;
}
static bool complete_unload(VMBus *vmbus)
{
vmbus_reset_all(vmbus);
return true;
}
static void process_message(VMBus *vmbus)
{
struct hyperv_post_message_input *hv_msg;
struct vmbus_message_header *msg;
void *msgdata;
uint32_t msglen;
qemu_mutex_lock(&vmbus->rx_queue_lock);
if (!vmbus->rx_queue_size) {
goto unlock;
}
hv_msg = &vmbus->rx_queue[vmbus->rx_queue_head];
msglen = hv_msg->payload_size;
if (msglen < sizeof(*msg)) {
goto out;
}
msgdata = hv_msg->payload;
msg = msgdata;
trace_vmbus_process_incoming_message(msg->message_type);
switch (msg->message_type) {
case VMBUS_MSG_INITIATE_CONTACT:
handle_initiate_contact(vmbus, msgdata, msglen);
break;
case VMBUS_MSG_REQUESTOFFERS:
handle_request_offers(vmbus, msgdata, msglen);
break;
case VMBUS_MSG_GPADL_HEADER:
handle_gpadl_header(vmbus, msgdata, msglen);
break;
case VMBUS_MSG_GPADL_BODY:
handle_gpadl_body(vmbus, msgdata, msglen);
break;
case VMBUS_MSG_GPADL_TEARDOWN:
handle_gpadl_teardown(vmbus, msgdata, msglen);
break;
case VMBUS_MSG_OPENCHANNEL:
handle_open_channel(vmbus, msgdata, msglen);
break;
case VMBUS_MSG_CLOSECHANNEL:
handle_close_channel(vmbus, msgdata, msglen);
break;
case VMBUS_MSG_UNLOAD:
handle_unload(vmbus, msgdata, msglen);
break;
default:
error_report("unknown message type %#x", msg->message_type);
break;
}
out:
vmbus->rx_queue_size--;
vmbus->rx_queue_head++;
vmbus->rx_queue_head %= HV_MSG_QUEUE_LEN;
vmbus_resched(vmbus);
unlock:
qemu_mutex_unlock(&vmbus->rx_queue_lock);
}
static const struct {
void (*run)(VMBus *vmbus);
bool (*complete)(VMBus *vmbus);
} state_runner[] = {
[VMBUS_LISTEN] = {process_message, NULL},
[VMBUS_HANDSHAKE] = {send_handshake, NULL},
[VMBUS_OFFER] = {send_offer, complete_offer},
[VMBUS_CREATE_GPADL] = {send_create_gpadl, complete_create_gpadl},
[VMBUS_TEARDOWN_GPADL] = {send_teardown_gpadl, complete_teardown_gpadl},
[VMBUS_OPEN_CHANNEL] = {send_open_channel, complete_open_channel},
[VMBUS_UNLOAD] = {send_unload, complete_unload},
};
static void vmbus_do_run(VMBus *vmbus)
{
if (vmbus->msg_in_progress) {
return;
}
assert(vmbus->state < VMBUS_STATE_MAX);
assert(state_runner[vmbus->state].run);
state_runner[vmbus->state].run(vmbus);
}
static void vmbus_run(void *opaque)
{
VMBus *vmbus = opaque;
/* make sure no recursion happens (e.g. due to recursive aio_poll()) */
if (vmbus->in_progress) {
return;
}
vmbus->in_progress = true;
/*
* FIXME: if vmbus_resched() is called from within vmbus_do_run(), it
* should go *after* the code that can result in aio_poll; otherwise
* reschedules can be missed. No idea how to enforce that.
*/
vmbus_do_run(vmbus);
vmbus->in_progress = false;
}
static void vmbus_msg_cb(void *data, int status)
{
VMBus *vmbus = data;
bool (*complete)(VMBus *vmbus);
assert(vmbus->msg_in_progress);
trace_vmbus_msg_cb(status);
if (status == -EAGAIN) {
goto out;
}
if (status) {
error_report("message delivery fatal failure: %d; aborting vmbus",
status);
vmbus_reset_all(vmbus);
return;
}
assert(vmbus->state < VMBUS_STATE_MAX);
complete = state_runner[vmbus->state].complete;
if (!complete || complete(vmbus)) {
vmbus->state = VMBUS_LISTEN;
}
out:
vmbus->msg_in_progress = false;
vmbus_resched(vmbus);
}
static void vmbus_resched(VMBus *vmbus)
{
aio_bh_schedule_oneshot(qemu_get_aio_context(), vmbus_run, vmbus);
}
static void vmbus_signal_event(EventNotifier *e)
{
VMBusChannel *chan;
VMBus *vmbus = container_of(e, VMBus, notifier);
unsigned long *int_map;
hwaddr addr, len;
bool is_dirty = false;
if (!event_notifier_test_and_clear(e)) {
return;
}
trace_vmbus_signal_event();
if (!vmbus->int_page_gpa) {
return;
}
addr = vmbus->int_page_gpa + TARGET_PAGE_SIZE / 2;
len = TARGET_PAGE_SIZE / 2;
int_map = cpu_physical_memory_map(addr, &len, 1);
if (len != TARGET_PAGE_SIZE / 2) {
goto unmap;
}
QTAILQ_FOREACH(chan, &vmbus->channel_list, link) {
if (bitmap_test_and_clear_atomic(int_map, chan->id, 1)) {
if (!vmbus_channel_is_open(chan)) {
continue;
}
vmbus_channel_notify_host(chan);
is_dirty = true;
}
}
unmap:
cpu_physical_memory_unmap(int_map, len, 1, is_dirty);
}
static void vmbus_dev_realize(DeviceState *dev, Error **errp)
{
VMBusDevice *vdev = VMBUS_DEVICE(dev);
VMBusDeviceClass *vdc = VMBUS_DEVICE_GET_CLASS(vdev);
VMBus *vmbus = VMBUS(qdev_get_parent_bus(dev));
BusChild *child;
Error *err = NULL;
char idstr[UUID_STR_LEN];
assert(!qemu_uuid_is_null(&vdev->instanceid));
if (!qemu_uuid_is_null(&vdc->instanceid)) {
/* Class wants to only have a single instance with a fixed UUID */
if (!qemu_uuid_is_equal(&vdev->instanceid, &vdc->instanceid)) {
error_setg(&err, "instance id can't be changed");
goto error_out;
}
}
/* Check for instance id collision for this class id */
QTAILQ_FOREACH(child, &BUS(vmbus)->children, sibling) {
VMBusDevice *child_dev = VMBUS_DEVICE(child->child);
if (child_dev == vdev) {
continue;
}
if (qemu_uuid_is_equal(&child_dev->instanceid, &vdev->instanceid)) {
qemu_uuid_unparse(&vdev->instanceid, idstr);
error_setg(&err, "duplicate vmbus device instance id %s", idstr);
goto error_out;
}
}
vdev->dma_as = &address_space_memory;
create_channels(vmbus, vdev, &err);
if (err) {
goto error_out;
}
if (vdc->vmdev_realize) {
vdc->vmdev_realize(vdev, &err);
if (err) {
goto err_vdc_realize;
}
}
return;
err_vdc_realize:
free_channels(vdev);
error_out:
error_propagate(errp, err);
}
static void vmbus_dev_reset(DeviceState *dev)
{
uint16_t i;
VMBusDevice *vdev = VMBUS_DEVICE(dev);
VMBusDeviceClass *vdc = VMBUS_DEVICE_GET_CLASS(vdev);
if (vdev->channels) {
for (i = 0; i < vdev->num_channels; i++) {
VMBusChannel *chan = &vdev->channels[i];
close_channel(chan);
chan->state = VMCHAN_INIT;
}
}
if (vdc->vmdev_reset) {
vdc->vmdev_reset(vdev);
}
}
static void vmbus_dev_unrealize(DeviceState *dev)
{
VMBusDevice *vdev = VMBUS_DEVICE(dev);
VMBusDeviceClass *vdc = VMBUS_DEVICE_GET_CLASS(vdev);
if (vdc->vmdev_unrealize) {
vdc->vmdev_unrealize(vdev);
}
free_channels(vdev);
}
static Property vmbus_dev_props[] = {
DEFINE_PROP_UUID("instanceid", VMBusDevice, instanceid),
DEFINE_PROP_END_OF_LIST()
};
static void vmbus_dev_class_init(ObjectClass *klass, void *data)
{
DeviceClass *kdev = DEVICE_CLASS(klass);
device_class_set_props(kdev, vmbus_dev_props);
kdev->bus_type = TYPE_VMBUS;
kdev->realize = vmbus_dev_realize;
kdev->unrealize = vmbus_dev_unrealize;
kdev->reset = vmbus_dev_reset;
}
static void vmbus_dev_instance_init(Object *obj)
{
VMBusDevice *vdev = VMBUS_DEVICE(obj);
VMBusDeviceClass *vdc = VMBUS_DEVICE_GET_CLASS(vdev);
if (!qemu_uuid_is_null(&vdc->instanceid)) {
/* Class wants to only have a single instance with a fixed UUID */
vdev->instanceid = vdc->instanceid;
}
}
const VMStateDescription vmstate_vmbus_dev = {
.name = TYPE_VMBUS_DEVICE,
.version_id = 0,
.minimum_version_id = 0,
.fields = (const VMStateField[]) {
VMSTATE_UINT8_ARRAY(instanceid.data, VMBusDevice, 16),
VMSTATE_UINT16(num_channels, VMBusDevice),
VMSTATE_STRUCT_VARRAY_POINTER_UINT16(channels, VMBusDevice,
num_channels, vmstate_channel,
VMBusChannel),
VMSTATE_END_OF_LIST()
}
};
/* vmbus generic device base */
static const TypeInfo vmbus_dev_type_info = {
.name = TYPE_VMBUS_DEVICE,
.parent = TYPE_DEVICE,
.abstract = true,
.instance_size = sizeof(VMBusDevice),
.class_size = sizeof(VMBusDeviceClass),
.class_init = vmbus_dev_class_init,
.instance_init = vmbus_dev_instance_init,
};
static void vmbus_realize(BusState *bus, Error **errp)
{
int ret = 0;
VMBus *vmbus = VMBUS(bus);
qemu_mutex_init(&vmbus->rx_queue_lock);
QTAILQ_INIT(&vmbus->gpadl_list);
QTAILQ_INIT(&vmbus->channel_list);
ret = hyperv_set_msg_handler(VMBUS_MESSAGE_CONNECTION_ID,
vmbus_recv_message, vmbus);
if (ret != 0) {
error_setg(errp, "hyperv set message handler failed: %d", ret);
goto error_out;
}
ret = event_notifier_init(&vmbus->notifier, 0);
if (ret != 0) {
error_setg(errp, "event notifier failed to init with %d", ret);
goto remove_msg_handler;
}
event_notifier_set_handler(&vmbus->notifier, vmbus_signal_event);
ret = hyperv_set_event_flag_handler(VMBUS_EVENT_CONNECTION_ID,
&vmbus->notifier);
if (ret != 0) {
error_setg(errp, "hyperv set event handler failed with %d", ret);
goto clear_event_notifier;
}
return;
clear_event_notifier:
event_notifier_cleanup(&vmbus->notifier);
remove_msg_handler:
hyperv_set_msg_handler(VMBUS_MESSAGE_CONNECTION_ID, NULL, NULL);
error_out:
qemu_mutex_destroy(&vmbus->rx_queue_lock);
}
static void vmbus_unrealize(BusState *bus)
{
VMBus *vmbus = VMBUS(bus);
hyperv_set_msg_handler(VMBUS_MESSAGE_CONNECTION_ID, NULL, NULL);
hyperv_set_event_flag_handler(VMBUS_EVENT_CONNECTION_ID, NULL);
event_notifier_cleanup(&vmbus->notifier);
qemu_mutex_destroy(&vmbus->rx_queue_lock);
}
static void vmbus_reset_hold(Object *obj)
{
vmbus_deinit(VMBUS(obj));
}
static char *vmbus_get_dev_path(DeviceState *dev)
{
BusState *bus = qdev_get_parent_bus(dev);
return qdev_get_dev_path(bus->parent);
}
static char *vmbus_get_fw_dev_path(DeviceState *dev)
{
VMBusDevice *vdev = VMBUS_DEVICE(dev);
char uuid[UUID_STR_LEN];
qemu_uuid_unparse(&vdev->instanceid, uuid);
return g_strdup_printf("%s@%s", qdev_fw_name(dev), uuid);
}
static void vmbus_class_init(ObjectClass *klass, void *data)
{
BusClass *k = BUS_CLASS(klass);
ResettableClass *rc = RESETTABLE_CLASS(klass);
k->get_dev_path = vmbus_get_dev_path;
k->get_fw_dev_path = vmbus_get_fw_dev_path;
k->realize = vmbus_realize;
k->unrealize = vmbus_unrealize;
rc->phases.hold = vmbus_reset_hold;
}
static int vmbus_pre_load(void *opaque)
{
VMBusChannel *chan;
VMBus *vmbus = VMBUS(opaque);
/*
* channel IDs allocated by the source will come in the migration stream
* for each channel, so clean up the ones allocated at realize
*/
QTAILQ_FOREACH(chan, &vmbus->channel_list, link) {
unregister_chan_id(chan);
}
return 0;
}
static int vmbus_post_load(void *opaque, int version_id)
{
int ret;
VMBus *vmbus = VMBUS(opaque);
VMBusGpadl *gpadl;
VMBusChannel *chan;
ret = vmbus_init(vmbus);
if (ret) {
return ret;
}
QTAILQ_FOREACH(gpadl, &vmbus->gpadl_list, link) {
gpadl->vmbus = vmbus;
gpadl->refcount = 1;
}
/*
* reopening channels depends on initialized vmbus so it's done here
* instead of channel_post_load()
*/
QTAILQ_FOREACH(chan, &vmbus->channel_list, link) {
if (chan->state == VMCHAN_OPENING || chan->state == VMCHAN_OPEN) {
open_channel(chan);
}
if (chan->state != VMCHAN_OPEN) {
continue;
}
if (!vmbus_channel_is_open(chan)) {
/* reopen failed, abort loading */
return -1;
}
/* resume processing on the guest side if it missed the notification */
hyperv_sint_route_set_sint(chan->notify_route);
/* ditto on the host side */
vmbus_channel_notify_host(chan);
}
vmbus_resched(vmbus);
return 0;
}
static const VMStateDescription vmstate_post_message_input = {
.name = "vmbus/hyperv_post_message_input",
.version_id = 0,
.minimum_version_id = 0,
.fields = (const VMStateField[]) {
/*
* skip connection_id and message_type as they are validated before
* queueing and ignored on dequeueing
*/
VMSTATE_UINT32(payload_size, struct hyperv_post_message_input),
VMSTATE_UINT8_ARRAY(payload, struct hyperv_post_message_input,
HV_MESSAGE_PAYLOAD_SIZE),
VMSTATE_END_OF_LIST()
}
};
static bool vmbus_rx_queue_needed(void *opaque)
{
VMBus *vmbus = VMBUS(opaque);
return vmbus->rx_queue_size;
}
static const VMStateDescription vmstate_rx_queue = {
.name = "vmbus/rx_queue",
.version_id = 0,
.minimum_version_id = 0,
.needed = vmbus_rx_queue_needed,
.fields = (const VMStateField[]) {
VMSTATE_UINT8(rx_queue_head, VMBus),
VMSTATE_UINT8(rx_queue_size, VMBus),
VMSTATE_STRUCT_ARRAY(rx_queue, VMBus,
HV_MSG_QUEUE_LEN, 0,
vmstate_post_message_input,
struct hyperv_post_message_input),
VMSTATE_END_OF_LIST()
}
};
static const VMStateDescription vmstate_vmbus = {
.name = TYPE_VMBUS,
.version_id = 0,
.minimum_version_id = 0,
.pre_load = vmbus_pre_load,
.post_load = vmbus_post_load,
.fields = (const VMStateField[]) {
VMSTATE_UINT8(state, VMBus),
VMSTATE_UINT32(version, VMBus),
VMSTATE_UINT32(target_vp, VMBus),
VMSTATE_UINT64(int_page_gpa, VMBus),
VMSTATE_QTAILQ_V(gpadl_list, VMBus, 0,
vmstate_gpadl, VMBusGpadl, link),
VMSTATE_END_OF_LIST()
},
.subsections = (const VMStateDescription * const []) {
&vmstate_rx_queue,
NULL
}
};
static const TypeInfo vmbus_type_info = {
.name = TYPE_VMBUS,
.parent = TYPE_BUS,
.instance_size = sizeof(VMBus),
.class_init = vmbus_class_init,
};
static void vmbus_bridge_realize(DeviceState *dev, Error **errp)
{
VMBusBridge *bridge = VMBUS_BRIDGE(dev);
/*
* here there's at least one vmbus bridge that is being realized, so
* vmbus_bridge_find can only return NULL if it's not unique
*/
if (!vmbus_bridge_find()) {
error_setg(errp, "there can be at most one %s in the system",
TYPE_VMBUS_BRIDGE);
return;
}
if (!hyperv_is_synic_enabled()) {
error_report("VMBus requires usable Hyper-V SynIC and VP_INDEX");
return;
}
if (!hyperv_are_vmbus_recommended_features_enabled()) {
warn_report("VMBus enabled without the recommended set of Hyper-V features: "
"hv-stimer, hv-vapic and hv-runtime. "
"Some Windows versions might not boot or enable the VMBus device");
}
bridge->bus = VMBUS(qbus_new(TYPE_VMBUS, dev, "vmbus"));
}
static char *vmbus_bridge_ofw_unit_address(const SysBusDevice *dev)
{
/* there can be only one VMBus */
return g_strdup("0");
}
static const VMStateDescription vmstate_vmbus_bridge = {
.name = TYPE_VMBUS_BRIDGE,
.version_id = 0,
.minimum_version_id = 0,
.fields = (const VMStateField[]) {
VMSTATE_STRUCT_POINTER(bus, VMBusBridge, vmstate_vmbus, VMBus),
VMSTATE_END_OF_LIST()
},
};
static Property vmbus_bridge_props[] = {
DEFINE_PROP_UINT8("irq", VMBusBridge, irq, 7),
DEFINE_PROP_END_OF_LIST()
};
static void vmbus_bridge_class_init(ObjectClass *klass, void *data)
{
DeviceClass *k = DEVICE_CLASS(klass);
SysBusDeviceClass *sk = SYS_BUS_DEVICE_CLASS(klass);
k->realize = vmbus_bridge_realize;
k->fw_name = "vmbus";
sk->explicit_ofw_unit_address = vmbus_bridge_ofw_unit_address;
set_bit(DEVICE_CATEGORY_BRIDGE, k->categories);
k->vmsd = &vmstate_vmbus_bridge;
device_class_set_props(k, vmbus_bridge_props);
/* override SysBusDevice's default */
k->user_creatable = true;
}
static const TypeInfo vmbus_bridge_type_info = {
.name = TYPE_VMBUS_BRIDGE,
.parent = TYPE_SYS_BUS_DEVICE,
.instance_size = sizeof(VMBusBridge),
.class_init = vmbus_bridge_class_init,
};
static void vmbus_register_types(void)
{
type_register_static(&vmbus_bridge_type_info);
type_register_static(&vmbus_dev_type_info);
type_register_static(&vmbus_type_info);
}
type_init(vmbus_register_types)