2023-01-01 18:54:41 +01:00
|
|
|
/*
|
|
|
|
* QEMU Xen backend support
|
|
|
|
*
|
|
|
|
* Copyright © 2022 Amazon.com, Inc. or its affiliates. All Rights Reserved.
|
|
|
|
*
|
|
|
|
* Authors: David Woodhouse <dwmw2@infradead.org>
|
|
|
|
*
|
|
|
|
* This work is licensed under the terms of the GNU GPL, version 2 or later.
|
|
|
|
* See the COPYING file in the top-level directory.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#ifndef QEMU_XEN_BACKEND_OPS_H
|
|
|
|
#define QEMU_XEN_BACKEND_OPS_H
|
|
|
|
|
2023-01-02 12:05:16 +01:00
|
|
|
#include "hw/xen/xen.h"
|
|
|
|
#include "hw/xen/interface/xen.h"
|
|
|
|
#include "hw/xen/interface/io/xenbus.h"
|
|
|
|
|
2023-01-01 18:54:41 +01:00
|
|
|
/*
|
|
|
|
* For the time being, these operations map fairly closely to the API of
|
|
|
|
* the actual Xen libraries, e.g. libxenevtchn. As we complete the migration
|
|
|
|
* from XenLegacyDevice back ends to the new XenDevice model, they may
|
|
|
|
* evolve to slightly higher-level APIs.
|
|
|
|
*
|
|
|
|
* The internal emulations do not emulate the Xen APIs entirely faithfully;
|
|
|
|
* only enough to be used by the Xen backend devices. For example, only one
|
|
|
|
* event channel can be bound to each handle, since that's sufficient for
|
|
|
|
* the device support (only the true Xen HVM backend uses more). And the
|
|
|
|
* behaviour of unmask() and pending() is different too because the device
|
|
|
|
* backends don't care.
|
|
|
|
*/
|
|
|
|
|
|
|
|
typedef struct xenevtchn_handle xenevtchn_handle;
|
|
|
|
typedef int xenevtchn_port_or_error_t;
|
|
|
|
typedef uint32_t evtchn_port_t;
|
2023-01-01 22:31:37 +01:00
|
|
|
typedef uint16_t domid_t;
|
|
|
|
typedef uint32_t grant_ref_t;
|
|
|
|
|
|
|
|
#define XEN_PAGE_SHIFT 12
|
|
|
|
#define XEN_PAGE_SIZE (1UL << XEN_PAGE_SHIFT)
|
|
|
|
#define XEN_PAGE_MASK (~(XEN_PAGE_SIZE - 1))
|
2023-01-01 18:54:41 +01:00
|
|
|
|
2023-01-02 12:05:16 +01:00
|
|
|
#ifndef xen_rmb
|
|
|
|
#define xen_rmb() smp_rmb()
|
|
|
|
#endif
|
|
|
|
#ifndef xen_wmb
|
|
|
|
#define xen_wmb() smp_wmb()
|
|
|
|
#endif
|
|
|
|
#ifndef xen_mb
|
|
|
|
#define xen_mb() smp_mb()
|
|
|
|
#endif
|
|
|
|
|
2023-01-01 18:54:41 +01:00
|
|
|
struct evtchn_backend_ops {
|
|
|
|
xenevtchn_handle *(*open)(void);
|
|
|
|
int (*bind_interdomain)(xenevtchn_handle *xc, uint32_t domid,
|
|
|
|
evtchn_port_t guest_port);
|
|
|
|
int (*unbind)(xenevtchn_handle *xc, evtchn_port_t port);
|
|
|
|
int (*close)(struct xenevtchn_handle *xc);
|
|
|
|
int (*get_fd)(struct xenevtchn_handle *xc);
|
|
|
|
int (*notify)(struct xenevtchn_handle *xc, evtchn_port_t port);
|
|
|
|
int (*unmask)(struct xenevtchn_handle *xc, evtchn_port_t port);
|
|
|
|
int (*pending)(struct xenevtchn_handle *xc);
|
|
|
|
};
|
|
|
|
|
|
|
|
extern struct evtchn_backend_ops *xen_evtchn_ops;
|
|
|
|
|
|
|
|
static inline xenevtchn_handle *qemu_xen_evtchn_open(void)
|
|
|
|
{
|
|
|
|
if (!xen_evtchn_ops) {
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
return xen_evtchn_ops->open();
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int qemu_xen_evtchn_bind_interdomain(xenevtchn_handle *xc,
|
|
|
|
uint32_t domid,
|
|
|
|
evtchn_port_t guest_port)
|
|
|
|
{
|
|
|
|
if (!xen_evtchn_ops) {
|
|
|
|
return -ENOSYS;
|
|
|
|
}
|
|
|
|
return xen_evtchn_ops->bind_interdomain(xc, domid, guest_port);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int qemu_xen_evtchn_unbind(xenevtchn_handle *xc,
|
|
|
|
evtchn_port_t port)
|
|
|
|
{
|
|
|
|
if (!xen_evtchn_ops) {
|
|
|
|
return -ENOSYS;
|
|
|
|
}
|
|
|
|
return xen_evtchn_ops->unbind(xc, port);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int qemu_xen_evtchn_close(xenevtchn_handle *xc)
|
|
|
|
{
|
|
|
|
if (!xen_evtchn_ops) {
|
|
|
|
return -ENOSYS;
|
|
|
|
}
|
|
|
|
return xen_evtchn_ops->close(xc);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int qemu_xen_evtchn_fd(xenevtchn_handle *xc)
|
|
|
|
{
|
|
|
|
if (!xen_evtchn_ops) {
|
|
|
|
return -ENOSYS;
|
|
|
|
}
|
|
|
|
return xen_evtchn_ops->get_fd(xc);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int qemu_xen_evtchn_notify(xenevtchn_handle *xc,
|
|
|
|
evtchn_port_t port)
|
|
|
|
{
|
|
|
|
if (!xen_evtchn_ops) {
|
|
|
|
return -ENOSYS;
|
|
|
|
}
|
|
|
|
return xen_evtchn_ops->notify(xc, port);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int qemu_xen_evtchn_unmask(xenevtchn_handle *xc,
|
|
|
|
evtchn_port_t port)
|
|
|
|
{
|
|
|
|
if (!xen_evtchn_ops) {
|
|
|
|
return -ENOSYS;
|
|
|
|
}
|
|
|
|
return xen_evtchn_ops->unmask(xc, port);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int qemu_xen_evtchn_pending(xenevtchn_handle *xc)
|
|
|
|
{
|
|
|
|
if (!xen_evtchn_ops) {
|
|
|
|
return -ENOSYS;
|
|
|
|
}
|
|
|
|
return xen_evtchn_ops->pending(xc);
|
|
|
|
}
|
|
|
|
|
2023-01-01 22:31:37 +01:00
|
|
|
typedef struct xengntdev_handle xengnttab_handle;
|
|
|
|
|
|
|
|
typedef struct XenGrantCopySegment {
|
|
|
|
union {
|
|
|
|
void *virt;
|
|
|
|
struct {
|
|
|
|
uint32_t ref;
|
|
|
|
off_t offset;
|
|
|
|
} foreign;
|
|
|
|
} source, dest;
|
|
|
|
size_t len;
|
|
|
|
} XenGrantCopySegment;
|
|
|
|
|
|
|
|
#define XEN_GNTTAB_OP_FEATURE_MAP_MULTIPLE (1U << 0)
|
|
|
|
|
|
|
|
struct gnttab_backend_ops {
|
|
|
|
uint32_t features;
|
|
|
|
xengnttab_handle *(*open)(void);
|
|
|
|
int (*close)(xengnttab_handle *xgt);
|
|
|
|
int (*grant_copy)(xengnttab_handle *xgt, bool to_domain, uint32_t domid,
|
|
|
|
XenGrantCopySegment *segs, uint32_t nr_segs,
|
|
|
|
Error **errp);
|
|
|
|
int (*set_max_grants)(xengnttab_handle *xgt, uint32_t nr_grants);
|
|
|
|
void *(*map_refs)(xengnttab_handle *xgt, uint32_t count, uint32_t domid,
|
|
|
|
uint32_t *refs, int prot);
|
2023-01-10 01:03:49 +01:00
|
|
|
int (*unmap)(xengnttab_handle *xgt, void *start_address, uint32_t *refs,
|
|
|
|
uint32_t count);
|
2023-01-01 22:31:37 +01:00
|
|
|
};
|
|
|
|
|
|
|
|
extern struct gnttab_backend_ops *xen_gnttab_ops;
|
|
|
|
|
|
|
|
static inline bool qemu_xen_gnttab_can_map_multi(void)
|
|
|
|
{
|
|
|
|
return xen_gnttab_ops &&
|
|
|
|
!!(xen_gnttab_ops->features & XEN_GNTTAB_OP_FEATURE_MAP_MULTIPLE);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline xengnttab_handle *qemu_xen_gnttab_open(void)
|
|
|
|
{
|
|
|
|
if (!xen_gnttab_ops) {
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
return xen_gnttab_ops->open();
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int qemu_xen_gnttab_close(xengnttab_handle *xgt)
|
|
|
|
{
|
|
|
|
if (!xen_gnttab_ops) {
|
|
|
|
return -ENOSYS;
|
|
|
|
}
|
|
|
|
return xen_gnttab_ops->close(xgt);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int qemu_xen_gnttab_grant_copy(xengnttab_handle *xgt,
|
|
|
|
bool to_domain, uint32_t domid,
|
|
|
|
XenGrantCopySegment *segs,
|
|
|
|
uint32_t nr_segs, Error **errp)
|
|
|
|
{
|
|
|
|
if (!xen_gnttab_ops) {
|
|
|
|
return -ENOSYS;
|
|
|
|
}
|
|
|
|
|
|
|
|
return xen_gnttab_ops->grant_copy(xgt, to_domain, domid, segs, nr_segs,
|
|
|
|
errp);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int qemu_xen_gnttab_set_max_grants(xengnttab_handle *xgt,
|
|
|
|
uint32_t nr_grants)
|
|
|
|
{
|
|
|
|
if (!xen_gnttab_ops) {
|
|
|
|
return -ENOSYS;
|
|
|
|
}
|
|
|
|
return xen_gnttab_ops->set_max_grants(xgt, nr_grants);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void *qemu_xen_gnttab_map_refs(xengnttab_handle *xgt,
|
|
|
|
uint32_t count, uint32_t domid,
|
|
|
|
uint32_t *refs, int prot)
|
|
|
|
{
|
|
|
|
if (!xen_gnttab_ops) {
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
return xen_gnttab_ops->map_refs(xgt, count, domid, refs, prot);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int qemu_xen_gnttab_unmap(xengnttab_handle *xgt,
|
2023-01-10 01:03:49 +01:00
|
|
|
void *start_address, uint32_t *refs,
|
2023-01-01 22:31:37 +01:00
|
|
|
uint32_t count)
|
|
|
|
{
|
|
|
|
if (!xen_gnttab_ops) {
|
|
|
|
return -ENOSYS;
|
|
|
|
}
|
2023-01-10 01:03:49 +01:00
|
|
|
return xen_gnttab_ops->unmap(xgt, start_address, refs, count);
|
2023-01-01 22:31:37 +01:00
|
|
|
}
|
|
|
|
|
2023-01-02 02:13:46 +01:00
|
|
|
struct foreignmem_backend_ops {
|
|
|
|
void *(*map)(uint32_t dom, void *addr, int prot, size_t pages,
|
|
|
|
xen_pfn_t *pfns, int *errs);
|
|
|
|
int (*unmap)(void *addr, size_t pages);
|
|
|
|
};
|
|
|
|
|
|
|
|
extern struct foreignmem_backend_ops *xen_foreignmem_ops;
|
|
|
|
|
|
|
|
static inline void *qemu_xen_foreignmem_map(uint32_t dom, void *addr, int prot,
|
|
|
|
size_t pages, xen_pfn_t *pfns,
|
|
|
|
int *errs)
|
|
|
|
{
|
|
|
|
if (!xen_foreignmem_ops) {
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
return xen_foreignmem_ops->map(dom, addr, prot, pages, pfns, errs);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int qemu_xen_foreignmem_unmap(void *addr, size_t pages)
|
|
|
|
{
|
|
|
|
if (!xen_foreignmem_ops) {
|
|
|
|
return -ENOSYS;
|
|
|
|
}
|
|
|
|
return xen_foreignmem_ops->unmap(addr, pages);
|
|
|
|
}
|
|
|
|
|
2023-01-02 12:05:16 +01:00
|
|
|
typedef void (*xs_watch_fn)(void *opaque, const char *path);
|
|
|
|
|
|
|
|
struct qemu_xs_handle;
|
|
|
|
struct qemu_xs_watch;
|
|
|
|
typedef uint32_t xs_transaction_t;
|
|
|
|
|
|
|
|
#define XBT_NULL 0
|
|
|
|
|
|
|
|
#define XS_PERM_NONE 0x00
|
|
|
|
#define XS_PERM_READ 0x01
|
|
|
|
#define XS_PERM_WRITE 0x02
|
|
|
|
|
|
|
|
struct xenstore_backend_ops {
|
|
|
|
struct qemu_xs_handle *(*open)(void);
|
|
|
|
void (*close)(struct qemu_xs_handle *h);
|
|
|
|
char *(*get_domain_path)(struct qemu_xs_handle *h, unsigned int domid);
|
|
|
|
char **(*directory)(struct qemu_xs_handle *h, xs_transaction_t t,
|
|
|
|
const char *path, unsigned int *num);
|
|
|
|
void *(*read)(struct qemu_xs_handle *h, xs_transaction_t t,
|
|
|
|
const char *path, unsigned int *len);
|
|
|
|
bool (*write)(struct qemu_xs_handle *h, xs_transaction_t t,
|
|
|
|
const char *path, const void *data, unsigned int len);
|
|
|
|
bool (*create)(struct qemu_xs_handle *h, xs_transaction_t t,
|
|
|
|
unsigned int owner, unsigned int domid,
|
|
|
|
unsigned int perms, const char *path);
|
|
|
|
bool (*destroy)(struct qemu_xs_handle *h, xs_transaction_t t,
|
|
|
|
const char *path);
|
|
|
|
struct qemu_xs_watch *(*watch)(struct qemu_xs_handle *h, const char *path,
|
|
|
|
xs_watch_fn fn, void *opaque);
|
|
|
|
void (*unwatch)(struct qemu_xs_handle *h, struct qemu_xs_watch *w);
|
|
|
|
xs_transaction_t (*transaction_start)(struct qemu_xs_handle *h);
|
|
|
|
bool (*transaction_end)(struct qemu_xs_handle *h, xs_transaction_t t,
|
|
|
|
bool abort);
|
|
|
|
};
|
|
|
|
|
|
|
|
extern struct xenstore_backend_ops *xen_xenstore_ops;
|
|
|
|
|
|
|
|
static inline struct qemu_xs_handle *qemu_xen_xs_open(void)
|
|
|
|
{
|
|
|
|
if (!xen_xenstore_ops) {
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
return xen_xenstore_ops->open();
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void qemu_xen_xs_close(struct qemu_xs_handle *h)
|
|
|
|
{
|
|
|
|
if (!xen_xenstore_ops) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
xen_xenstore_ops->close(h);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline char *qemu_xen_xs_get_domain_path(struct qemu_xs_handle *h,
|
|
|
|
unsigned int domid)
|
|
|
|
{
|
|
|
|
if (!xen_xenstore_ops) {
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
return xen_xenstore_ops->get_domain_path(h, domid);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline char **qemu_xen_xs_directory(struct qemu_xs_handle *h,
|
|
|
|
xs_transaction_t t, const char *path,
|
|
|
|
unsigned int *num)
|
|
|
|
{
|
|
|
|
if (!xen_xenstore_ops) {
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
return xen_xenstore_ops->directory(h, t, path, num);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void *qemu_xen_xs_read(struct qemu_xs_handle *h,
|
|
|
|
xs_transaction_t t, const char *path,
|
|
|
|
unsigned int *len)
|
|
|
|
{
|
|
|
|
if (!xen_xenstore_ops) {
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
return xen_xenstore_ops->read(h, t, path, len);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline bool qemu_xen_xs_write(struct qemu_xs_handle *h,
|
|
|
|
xs_transaction_t t, const char *path,
|
|
|
|
const void *data, unsigned int len)
|
|
|
|
{
|
|
|
|
if (!xen_xenstore_ops) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
return xen_xenstore_ops->write(h, t, path, data, len);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline bool qemu_xen_xs_create(struct qemu_xs_handle *h,
|
|
|
|
xs_transaction_t t, unsigned int owner,
|
|
|
|
unsigned int domid, unsigned int perms,
|
|
|
|
const char *path)
|
|
|
|
{
|
|
|
|
if (!xen_xenstore_ops) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
return xen_xenstore_ops->create(h, t, owner, domid, perms, path);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline bool qemu_xen_xs_destroy(struct qemu_xs_handle *h,
|
|
|
|
xs_transaction_t t, const char *path)
|
|
|
|
{
|
|
|
|
if (!xen_xenstore_ops) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
return xen_xenstore_ops->destroy(h, t, path);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline struct qemu_xs_watch *qemu_xen_xs_watch(struct qemu_xs_handle *h,
|
|
|
|
const char *path,
|
|
|
|
xs_watch_fn fn,
|
|
|
|
void *opaque)
|
|
|
|
{
|
|
|
|
if (!xen_xenstore_ops) {
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
return xen_xenstore_ops->watch(h, path, fn, opaque);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void qemu_xen_xs_unwatch(struct qemu_xs_handle *h,
|
|
|
|
struct qemu_xs_watch *w)
|
|
|
|
{
|
|
|
|
if (!xen_xenstore_ops) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
xen_xenstore_ops->unwatch(h, w);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline xs_transaction_t qemu_xen_xs_transaction_start(struct qemu_xs_handle *h)
|
|
|
|
{
|
|
|
|
if (!xen_xenstore_ops) {
|
|
|
|
return XBT_NULL;
|
|
|
|
}
|
|
|
|
return xen_xenstore_ops->transaction_start(h);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline bool qemu_xen_xs_transaction_end(struct qemu_xs_handle *h,
|
|
|
|
xs_transaction_t t, bool abort)
|
|
|
|
{
|
|
|
|
if (!xen_xenstore_ops) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
return xen_xenstore_ops->transaction_end(h, t, abort);
|
|
|
|
}
|
|
|
|
|
2023-01-01 18:54:41 +01:00
|
|
|
void setup_xen_backend_ops(void);
|
|
|
|
|
|
|
|
#endif /* QEMU_XEN_BACKEND_OPS_H */
|