108a64818e
Currently we have 2 types of vhost backends in QEMU: vhost kernel and vhost-user. The above patch provides a generic device for vDPA purpose, this vDPA device exposes to user space a non-vendor-specific configuration interface for setting up a vhost HW accelerator, this patch set introduces a third vhost backend called vhost-vdpa based on the vDPA interface. Vhost-vdpa usage: qemu-system-x86_64 -cpu host -enable-kvm \ ...... -netdev type=vhost-vdpa,vhostdev=/dev/vhost-vdpa-id,id=vhost-vdpa0 \ -device virtio-net-pci,netdev=vhost-vdpa0,page-per-vq=on \ Signed-off-by: Lingshan zhu <lingshan.zhu@intel.com> Signed-off-by: Tiwei Bie <tiwei.bie@intel.com> Signed-off-by: Cindy Lu <lulu@redhat.com> Signed-off-by: Jason Wang <jasowang@redhat.com> Message-Id: <20200701145538.22333-14-lulu@redhat.com> Reviewed-by: Michael S. Tsirkin <mst@redhat.com> Signed-off-by: Michael S. Tsirkin <mst@redhat.com> Acked-by: Jason Wang <jasowang@redhat.com>
148 lines
4.5 KiB
C
148 lines
4.5 KiB
C
#ifndef VHOST_H
|
|
#define VHOST_H
|
|
|
|
#include "hw/virtio/vhost-backend.h"
|
|
#include "hw/virtio/virtio.h"
|
|
#include "exec/memory.h"
|
|
|
|
/* Generic structures common for any vhost based device. */
|
|
|
|
struct vhost_inflight {
|
|
int fd;
|
|
void *addr;
|
|
uint64_t size;
|
|
uint64_t offset;
|
|
uint16_t queue_size;
|
|
};
|
|
|
|
struct vhost_virtqueue {
|
|
int kick;
|
|
int call;
|
|
void *desc;
|
|
void *avail;
|
|
void *used;
|
|
int num;
|
|
unsigned long long desc_phys;
|
|
unsigned desc_size;
|
|
unsigned long long avail_phys;
|
|
unsigned avail_size;
|
|
unsigned long long used_phys;
|
|
unsigned used_size;
|
|
EventNotifier masked_notifier;
|
|
struct vhost_dev *dev;
|
|
};
|
|
|
|
typedef unsigned long vhost_log_chunk_t;
|
|
#define VHOST_LOG_PAGE 0x1000
|
|
#define VHOST_LOG_BITS (8 * sizeof(vhost_log_chunk_t))
|
|
#define VHOST_LOG_CHUNK (VHOST_LOG_PAGE * VHOST_LOG_BITS)
|
|
#define VHOST_INVALID_FEATURE_BIT (0xff)
|
|
|
|
struct vhost_log {
|
|
unsigned long long size;
|
|
int refcnt;
|
|
int fd;
|
|
vhost_log_chunk_t *log;
|
|
};
|
|
|
|
struct vhost_dev;
|
|
struct vhost_iommu {
|
|
struct vhost_dev *hdev;
|
|
MemoryRegion *mr;
|
|
hwaddr iommu_offset;
|
|
IOMMUNotifier n;
|
|
QLIST_ENTRY(vhost_iommu) iommu_next;
|
|
};
|
|
|
|
typedef struct VhostDevConfigOps {
|
|
/* Vhost device config space changed callback
|
|
*/
|
|
int (*vhost_dev_config_notifier)(struct vhost_dev *dev);
|
|
} VhostDevConfigOps;
|
|
|
|
struct vhost_memory;
|
|
struct vhost_dev {
|
|
VirtIODevice *vdev;
|
|
MemoryListener memory_listener;
|
|
MemoryListener iommu_listener;
|
|
struct vhost_memory *mem;
|
|
int n_mem_sections;
|
|
MemoryRegionSection *mem_sections;
|
|
int n_tmp_sections;
|
|
MemoryRegionSection *tmp_sections;
|
|
struct vhost_virtqueue *vqs;
|
|
int nvqs;
|
|
/* the first virtqueue which would be used by this vhost dev */
|
|
int vq_index;
|
|
uint64_t features;
|
|
uint64_t acked_features;
|
|
uint64_t backend_features;
|
|
uint64_t protocol_features;
|
|
uint64_t max_queues;
|
|
bool started;
|
|
bool log_enabled;
|
|
uint64_t log_size;
|
|
Error *migration_blocker;
|
|
const VhostOps *vhost_ops;
|
|
void *opaque;
|
|
struct vhost_log *log;
|
|
QLIST_ENTRY(vhost_dev) entry;
|
|
QLIST_HEAD(, vhost_iommu) iommu_list;
|
|
IOMMUNotifier n;
|
|
const VhostDevConfigOps *config_ops;
|
|
};
|
|
|
|
struct vhost_net {
|
|
struct vhost_dev dev;
|
|
struct vhost_virtqueue vqs[2];
|
|
int backend;
|
|
NetClientState *nc;
|
|
};
|
|
|
|
int vhost_dev_init(struct vhost_dev *hdev, void *opaque,
|
|
VhostBackendType backend_type,
|
|
uint32_t busyloop_timeout);
|
|
void vhost_dev_cleanup(struct vhost_dev *hdev);
|
|
int vhost_dev_start(struct vhost_dev *hdev, VirtIODevice *vdev);
|
|
void vhost_dev_stop(struct vhost_dev *hdev, VirtIODevice *vdev);
|
|
int vhost_dev_enable_notifiers(struct vhost_dev *hdev, VirtIODevice *vdev);
|
|
void vhost_dev_disable_notifiers(struct vhost_dev *hdev, VirtIODevice *vdev);
|
|
|
|
/* Test and clear masked event pending status.
|
|
* Should be called after unmask to avoid losing events.
|
|
*/
|
|
bool vhost_virtqueue_pending(struct vhost_dev *hdev, int n);
|
|
|
|
/* Mask/unmask events from this vq.
|
|
*/
|
|
void vhost_virtqueue_mask(struct vhost_dev *hdev, VirtIODevice *vdev, int n,
|
|
bool mask);
|
|
uint64_t vhost_get_features(struct vhost_dev *hdev, const int *feature_bits,
|
|
uint64_t features);
|
|
void vhost_ack_features(struct vhost_dev *hdev, const int *feature_bits,
|
|
uint64_t features);
|
|
bool vhost_has_free_slot(void);
|
|
|
|
int vhost_net_set_backend(struct vhost_dev *hdev,
|
|
struct vhost_vring_file *file);
|
|
|
|
int vhost_device_iotlb_miss(struct vhost_dev *dev, uint64_t iova, int write);
|
|
int vhost_dev_get_config(struct vhost_dev *dev, uint8_t *config,
|
|
uint32_t config_len);
|
|
int vhost_dev_set_config(struct vhost_dev *dev, const uint8_t *data,
|
|
uint32_t offset, uint32_t size, uint32_t flags);
|
|
/* notifier callback in case vhost device config space changed
|
|
*/
|
|
void vhost_dev_set_config_notifier(struct vhost_dev *dev,
|
|
const VhostDevConfigOps *ops);
|
|
|
|
void vhost_dev_reset_inflight(struct vhost_inflight *inflight);
|
|
void vhost_dev_free_inflight(struct vhost_inflight *inflight);
|
|
void vhost_dev_save_inflight(struct vhost_inflight *inflight, QEMUFile *f);
|
|
int vhost_dev_load_inflight(struct vhost_inflight *inflight, QEMUFile *f);
|
|
int vhost_dev_set_inflight(struct vhost_dev *dev,
|
|
struct vhost_inflight *inflight);
|
|
int vhost_dev_get_inflight(struct vhost_dev *dev, uint16_t queue_size,
|
|
struct vhost_inflight *inflight);
|
|
#endif
|