vdpa: add asid parameter to vhost_vdpa_dma_map/unmap
So the caller can choose which ASID is destined. No need to update the batch functions as they will always be called from memory listener updates at the moment. Memory listener updates will always update ASID 0, as it's the passthrough ASID. All vhost devices's ASID are 0 at this moment. Signed-off-by: Eugenio Pérez <eperezma@redhat.com> Acked-by: Jason Wang <jasowang@redhat.com> Message-Id: <20221215113144.322011-10-eperezma@redhat.com> Reviewed-by: Michael S. Tsirkin <mst@redhat.com> Signed-off-by: Michael S. Tsirkin <mst@redhat.com>
This commit is contained in:
parent
273e0003f0
commit
cd831ed5c4
@ -30,8 +30,8 @@ vhost_user_write(uint32_t req, uint32_t flags) "req:%d flags:0x%"PRIx32""
|
|||||||
vhost_user_create_notifier(int idx, void *n) "idx:%d n:%p"
|
vhost_user_create_notifier(int idx, void *n) "idx:%d n:%p"
|
||||||
|
|
||||||
# vhost-vdpa.c
|
# vhost-vdpa.c
|
||||||
vhost_vdpa_dma_map(void *vdpa, int fd, uint32_t msg_type, uint64_t iova, uint64_t size, uint64_t uaddr, uint8_t perm, uint8_t type) "vdpa:%p fd: %d msg_type: %"PRIu32" iova: 0x%"PRIx64" size: 0x%"PRIx64" uaddr: 0x%"PRIx64" perm: 0x%"PRIx8" type: %"PRIu8
|
vhost_vdpa_dma_map(void *vdpa, int fd, uint32_t msg_type, uint32_t asid, uint64_t iova, uint64_t size, uint64_t uaddr, uint8_t perm, uint8_t type) "vdpa:%p fd: %d msg_type: %"PRIu32" asid: %"PRIu32" iova: 0x%"PRIx64" size: 0x%"PRIx64" uaddr: 0x%"PRIx64" perm: 0x%"PRIx8" type: %"PRIu8
|
||||||
vhost_vdpa_dma_unmap(void *vdpa, int fd, uint32_t msg_type, uint64_t iova, uint64_t size, uint8_t type) "vdpa:%p fd: %d msg_type: %"PRIu32" iova: 0x%"PRIx64" size: 0x%"PRIx64" type: %"PRIu8
|
vhost_vdpa_dma_unmap(void *vdpa, int fd, uint32_t msg_type, uint32_t asid, uint64_t iova, uint64_t size, uint8_t type) "vdpa:%p fd: %d msg_type: %"PRIu32" asid: %"PRIu32" iova: 0x%"PRIx64" size: 0x%"PRIx64" type: %"PRIu8
|
||||||
vhost_vdpa_listener_begin_batch(void *v, int fd, uint32_t msg_type, uint8_t type) "vdpa:%p fd: %d msg_type: %"PRIu32" type: %"PRIu8
|
vhost_vdpa_listener_begin_batch(void *v, int fd, uint32_t msg_type, uint8_t type) "vdpa:%p fd: %d msg_type: %"PRIu32" type: %"PRIu8
|
||||||
vhost_vdpa_listener_commit(void *v, int fd, uint32_t msg_type, uint8_t type) "vdpa:%p fd: %d msg_type: %"PRIu32" type: %"PRIu8
|
vhost_vdpa_listener_commit(void *v, int fd, uint32_t msg_type, uint8_t type) "vdpa:%p fd: %d msg_type: %"PRIu32" type: %"PRIu8
|
||||||
vhost_vdpa_listener_region_add(void *vdpa, uint64_t iova, uint64_t llend, void *vaddr, bool readonly) "vdpa: %p iova 0x%"PRIx64" llend 0x%"PRIx64" vaddr: %p read-only: %d"
|
vhost_vdpa_listener_region_add(void *vdpa, uint64_t iova, uint64_t llend, void *vaddr, bool readonly) "vdpa: %p iova 0x%"PRIx64" llend 0x%"PRIx64" vaddr: %p read-only: %d"
|
||||||
|
@ -72,22 +72,28 @@ static bool vhost_vdpa_listener_skipped_section(MemoryRegionSection *section,
|
|||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
int vhost_vdpa_dma_map(struct vhost_vdpa *v, hwaddr iova, hwaddr size,
|
/*
|
||||||
void *vaddr, bool readonly)
|
* The caller must set asid = 0 if the device does not support asid.
|
||||||
|
* This is not an ABI break since it is set to 0 by the initializer anyway.
|
||||||
|
*/
|
||||||
|
int vhost_vdpa_dma_map(struct vhost_vdpa *v, uint32_t asid, hwaddr iova,
|
||||||
|
hwaddr size, void *vaddr, bool readonly)
|
||||||
{
|
{
|
||||||
struct vhost_msg_v2 msg = {};
|
struct vhost_msg_v2 msg = {};
|
||||||
int fd = v->device_fd;
|
int fd = v->device_fd;
|
||||||
int ret = 0;
|
int ret = 0;
|
||||||
|
|
||||||
msg.type = v->msg_type;
|
msg.type = v->msg_type;
|
||||||
|
msg.asid = asid;
|
||||||
msg.iotlb.iova = iova;
|
msg.iotlb.iova = iova;
|
||||||
msg.iotlb.size = size;
|
msg.iotlb.size = size;
|
||||||
msg.iotlb.uaddr = (uint64_t)(uintptr_t)vaddr;
|
msg.iotlb.uaddr = (uint64_t)(uintptr_t)vaddr;
|
||||||
msg.iotlb.perm = readonly ? VHOST_ACCESS_RO : VHOST_ACCESS_RW;
|
msg.iotlb.perm = readonly ? VHOST_ACCESS_RO : VHOST_ACCESS_RW;
|
||||||
msg.iotlb.type = VHOST_IOTLB_UPDATE;
|
msg.iotlb.type = VHOST_IOTLB_UPDATE;
|
||||||
|
|
||||||
trace_vhost_vdpa_dma_map(v, fd, msg.type, msg.iotlb.iova, msg.iotlb.size,
|
trace_vhost_vdpa_dma_map(v, fd, msg.type, msg.asid, msg.iotlb.iova,
|
||||||
msg.iotlb.uaddr, msg.iotlb.perm, msg.iotlb.type);
|
msg.iotlb.size, msg.iotlb.uaddr, msg.iotlb.perm,
|
||||||
|
msg.iotlb.type);
|
||||||
|
|
||||||
if (write(fd, &msg, sizeof(msg)) != sizeof(msg)) {
|
if (write(fd, &msg, sizeof(msg)) != sizeof(msg)) {
|
||||||
error_report("failed to write, fd=%d, errno=%d (%s)",
|
error_report("failed to write, fd=%d, errno=%d (%s)",
|
||||||
@ -98,18 +104,24 @@ int vhost_vdpa_dma_map(struct vhost_vdpa *v, hwaddr iova, hwaddr size,
|
|||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
int vhost_vdpa_dma_unmap(struct vhost_vdpa *v, hwaddr iova, hwaddr size)
|
/*
|
||||||
|
* The caller must set asid = 0 if the device does not support asid.
|
||||||
|
* This is not an ABI break since it is set to 0 by the initializer anyway.
|
||||||
|
*/
|
||||||
|
int vhost_vdpa_dma_unmap(struct vhost_vdpa *v, uint32_t asid, hwaddr iova,
|
||||||
|
hwaddr size)
|
||||||
{
|
{
|
||||||
struct vhost_msg_v2 msg = {};
|
struct vhost_msg_v2 msg = {};
|
||||||
int fd = v->device_fd;
|
int fd = v->device_fd;
|
||||||
int ret = 0;
|
int ret = 0;
|
||||||
|
|
||||||
msg.type = v->msg_type;
|
msg.type = v->msg_type;
|
||||||
|
msg.asid = asid;
|
||||||
msg.iotlb.iova = iova;
|
msg.iotlb.iova = iova;
|
||||||
msg.iotlb.size = size;
|
msg.iotlb.size = size;
|
||||||
msg.iotlb.type = VHOST_IOTLB_INVALIDATE;
|
msg.iotlb.type = VHOST_IOTLB_INVALIDATE;
|
||||||
|
|
||||||
trace_vhost_vdpa_dma_unmap(v, fd, msg.type, msg.iotlb.iova,
|
trace_vhost_vdpa_dma_unmap(v, fd, msg.type, msg.asid, msg.iotlb.iova,
|
||||||
msg.iotlb.size, msg.iotlb.type);
|
msg.iotlb.size, msg.iotlb.type);
|
||||||
|
|
||||||
if (write(fd, &msg, sizeof(msg)) != sizeof(msg)) {
|
if (write(fd, &msg, sizeof(msg)) != sizeof(msg)) {
|
||||||
@ -229,8 +241,8 @@ static void vhost_vdpa_listener_region_add(MemoryListener *listener,
|
|||||||
}
|
}
|
||||||
|
|
||||||
vhost_vdpa_iotlb_batch_begin_once(v);
|
vhost_vdpa_iotlb_batch_begin_once(v);
|
||||||
ret = vhost_vdpa_dma_map(v, iova, int128_get64(llsize),
|
ret = vhost_vdpa_dma_map(v, VHOST_VDPA_GUEST_PA_ASID, iova,
|
||||||
vaddr, section->readonly);
|
int128_get64(llsize), vaddr, section->readonly);
|
||||||
if (ret) {
|
if (ret) {
|
||||||
error_report("vhost vdpa map fail!");
|
error_report("vhost vdpa map fail!");
|
||||||
goto fail_map;
|
goto fail_map;
|
||||||
@ -303,7 +315,8 @@ static void vhost_vdpa_listener_region_del(MemoryListener *listener,
|
|||||||
vhost_iova_tree_remove(v->iova_tree, *result);
|
vhost_iova_tree_remove(v->iova_tree, *result);
|
||||||
}
|
}
|
||||||
vhost_vdpa_iotlb_batch_begin_once(v);
|
vhost_vdpa_iotlb_batch_begin_once(v);
|
||||||
ret = vhost_vdpa_dma_unmap(v, iova, int128_get64(llsize));
|
ret = vhost_vdpa_dma_unmap(v, VHOST_VDPA_GUEST_PA_ASID, iova,
|
||||||
|
int128_get64(llsize));
|
||||||
if (ret) {
|
if (ret) {
|
||||||
error_report("vhost_vdpa dma unmap error!");
|
error_report("vhost_vdpa dma unmap error!");
|
||||||
}
|
}
|
||||||
@ -869,7 +882,7 @@ static void vhost_vdpa_svq_unmap_ring(struct vhost_vdpa *v, hwaddr addr)
|
|||||||
}
|
}
|
||||||
|
|
||||||
size = ROUND_UP(result->size, qemu_real_host_page_size());
|
size = ROUND_UP(result->size, qemu_real_host_page_size());
|
||||||
r = vhost_vdpa_dma_unmap(v, result->iova, size);
|
r = vhost_vdpa_dma_unmap(v, v->address_space_id, result->iova, size);
|
||||||
if (unlikely(r < 0)) {
|
if (unlikely(r < 0)) {
|
||||||
error_report("Unable to unmap SVQ vring: %s (%d)", g_strerror(-r), -r);
|
error_report("Unable to unmap SVQ vring: %s (%d)", g_strerror(-r), -r);
|
||||||
return;
|
return;
|
||||||
@ -909,7 +922,8 @@ static bool vhost_vdpa_svq_map_ring(struct vhost_vdpa *v, DMAMap *needle,
|
|||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
r = vhost_vdpa_dma_map(v, needle->iova, needle->size + 1,
|
r = vhost_vdpa_dma_map(v, v->address_space_id, needle->iova,
|
||||||
|
needle->size + 1,
|
||||||
(void *)(uintptr_t)needle->translated_addr,
|
(void *)(uintptr_t)needle->translated_addr,
|
||||||
needle->perm == IOMMU_RO);
|
needle->perm == IOMMU_RO);
|
||||||
if (unlikely(r != 0)) {
|
if (unlikely(r != 0)) {
|
||||||
|
@ -19,6 +19,12 @@
|
|||||||
#include "hw/virtio/virtio.h"
|
#include "hw/virtio/virtio.h"
|
||||||
#include "standard-headers/linux/vhost_types.h"
|
#include "standard-headers/linux/vhost_types.h"
|
||||||
|
|
||||||
|
/*
|
||||||
|
* ASID dedicated to map guest's addresses. If SVQ is disabled it maps GPA to
|
||||||
|
* qemu's IOVA. If SVQ is enabled it maps also the SVQ vring here
|
||||||
|
*/
|
||||||
|
#define VHOST_VDPA_GUEST_PA_ASID 0
|
||||||
|
|
||||||
typedef struct VhostVDPAHostNotifier {
|
typedef struct VhostVDPAHostNotifier {
|
||||||
MemoryRegion mr;
|
MemoryRegion mr;
|
||||||
void *addr;
|
void *addr;
|
||||||
@ -29,6 +35,7 @@ typedef struct vhost_vdpa {
|
|||||||
int index;
|
int index;
|
||||||
uint32_t msg_type;
|
uint32_t msg_type;
|
||||||
bool iotlb_batch_begin_sent;
|
bool iotlb_batch_begin_sent;
|
||||||
|
uint32_t address_space_id;
|
||||||
MemoryListener listener;
|
MemoryListener listener;
|
||||||
struct vhost_vdpa_iova_range iova_range;
|
struct vhost_vdpa_iova_range iova_range;
|
||||||
uint64_t acked_features;
|
uint64_t acked_features;
|
||||||
@ -42,8 +49,9 @@ typedef struct vhost_vdpa {
|
|||||||
VhostVDPAHostNotifier notifier[VIRTIO_QUEUE_MAX];
|
VhostVDPAHostNotifier notifier[VIRTIO_QUEUE_MAX];
|
||||||
} VhostVDPA;
|
} VhostVDPA;
|
||||||
|
|
||||||
int vhost_vdpa_dma_map(struct vhost_vdpa *v, hwaddr iova, hwaddr size,
|
int vhost_vdpa_dma_map(struct vhost_vdpa *v, uint32_t asid, hwaddr iova,
|
||||||
void *vaddr, bool readonly);
|
hwaddr size, void *vaddr, bool readonly);
|
||||||
int vhost_vdpa_dma_unmap(struct vhost_vdpa *v, hwaddr iova, hwaddr size);
|
int vhost_vdpa_dma_unmap(struct vhost_vdpa *v, uint32_t asid, hwaddr iova,
|
||||||
|
hwaddr size);
|
||||||
|
|
||||||
#endif
|
#endif
|
||||||
|
@ -259,7 +259,7 @@ static void vhost_vdpa_cvq_unmap_buf(struct vhost_vdpa *v, void *addr)
|
|||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
r = vhost_vdpa_dma_unmap(v, map->iova, map->size + 1);
|
r = vhost_vdpa_dma_unmap(v, v->address_space_id, map->iova, map->size + 1);
|
||||||
if (unlikely(r != 0)) {
|
if (unlikely(r != 0)) {
|
||||||
error_report("Device cannot unmap: %s(%d)", g_strerror(r), r);
|
error_report("Device cannot unmap: %s(%d)", g_strerror(r), r);
|
||||||
}
|
}
|
||||||
@ -299,8 +299,8 @@ static int vhost_vdpa_cvq_map_buf(struct vhost_vdpa *v, void *buf, size_t size,
|
|||||||
return r;
|
return r;
|
||||||
}
|
}
|
||||||
|
|
||||||
r = vhost_vdpa_dma_map(v, map.iova, vhost_vdpa_net_cvq_cmd_page_len(), buf,
|
r = vhost_vdpa_dma_map(v, v->address_space_id, map.iova,
|
||||||
!write);
|
vhost_vdpa_net_cvq_cmd_page_len(), buf, !write);
|
||||||
if (unlikely(r < 0)) {
|
if (unlikely(r < 0)) {
|
||||||
goto dma_map_err;
|
goto dma_map_err;
|
||||||
}
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user