#ifndef QEMU_HW_XEN_COMMON_H #define QEMU_HW_XEN_COMMON_H /* * If we have new enough libxenctrl then we do not want/need these compat * interfaces, despite what the user supplied cflags might say. They * must be undefined before including xenctrl.h */ #undef XC_WANT_COMPAT_EVTCHN_API #undef XC_WANT_COMPAT_GNTTAB_API #undef XC_WANT_COMPAT_MAP_FOREIGN_API #include #include #include #include "hw/hw.h" #include "hw/xen/xen.h" #include "hw/pci/pci.h" #include "qemu/queue.h" #include "hw/xen/trace.h" extern xc_interface *xen_xc; /* * We don't support Xen prior to 4.2.0. */ #if CONFIG_XEN_CTRL_INTERFACE_VERSION < 40900 typedef xc_interface xendevicemodel_handle; static inline xendevicemodel_handle *xendevicemodel_open( struct xentoollog_logger *logger, unsigned int open_flags) { return xen_xc; } #if CONFIG_XEN_CTRL_INTERFACE_VERSION >= 40500 static inline int xendevicemodel_create_ioreq_server( xendevicemodel_handle *dmod, domid_t domid, int handle_bufioreq, ioservid_t *id) { return xc_hvm_create_ioreq_server(dmod, domid, handle_bufioreq, id); } static inline int xendevicemodel_get_ioreq_server_info( xendevicemodel_handle *dmod, domid_t domid, ioservid_t id, xen_pfn_t *ioreq_pfn, xen_pfn_t *bufioreq_pfn, evtchn_port_t *bufioreq_port) { return xc_hvm_get_ioreq_server_info(dmod, domid, id, ioreq_pfn, bufioreq_pfn, bufioreq_port); } static inline int xendevicemodel_map_io_range_to_ioreq_server( xendevicemodel_handle *dmod, domid_t domid, ioservid_t id, int is_mmio, uint64_t start, uint64_t end) { return xc_hvm_map_io_range_to_ioreq_server(dmod, domid, id, is_mmio, start, end); } static inline int xendevicemodel_unmap_io_range_from_ioreq_server( xendevicemodel_handle *dmod, domid_t domid, ioservid_t id, int is_mmio, uint64_t start, uint64_t end) { return xc_hvm_unmap_io_range_from_ioreq_server(dmod, domid, id, is_mmio, start, end); } static inline int xendevicemodel_map_pcidev_to_ioreq_server( xendevicemodel_handle *dmod, domid_t domid, ioservid_t id, uint16_t segment, uint8_t bus, uint8_t device, uint8_t function) { return xc_hvm_map_pcidev_to_ioreq_server(dmod, domid, id, segment, bus, device, function); } static inline int xendevicemodel_unmap_pcidev_from_ioreq_server( xendevicemodel_handle *dmod, domid_t domid, ioservid_t id, uint16_t segment, uint8_t bus, uint8_t device, uint8_t function) { return xc_hvm_unmap_pcidev_from_ioreq_server(dmod, domid, id, segment, bus, device, function); } static inline int xendevicemodel_destroy_ioreq_server( xendevicemodel_handle *dmod, domid_t domid, ioservid_t id) { return xc_hvm_destroy_ioreq_server(dmod, domid, id); } static inline int xendevicemodel_set_ioreq_server_state( xendevicemodel_handle *dmod, domid_t domid, ioservid_t id, int enabled) { return xc_hvm_set_ioreq_server_state(dmod, domid, id, enabled); } #endif /* CONFIG_XEN_CTRL_INTERFACE_VERSION >= 40500 */ static inline int xendevicemodel_set_pci_intx_level( xendevicemodel_handle *dmod, domid_t domid, uint16_t segment, uint8_t bus, uint8_t device, uint8_t intx, unsigned int level) { return xc_hvm_set_pci_intx_level(dmod, domid, segment, bus, device, intx, level); } static inline int xendevicemodel_set_isa_irq_level( xendevicemodel_handle *dmod, domid_t domid, uint8_t irq, unsigned int level) { return xc_hvm_set_isa_irq_level(dmod, domid, irq, level); } static inline int xendevicemodel_set_pci_link_route( xendevicemodel_handle *dmod, domid_t domid, uint8_t link, uint8_t irq) { return xc_hvm_set_pci_link_route(dmod, domid, link, irq); } static inline int xendevicemodel_inject_msi( xendevicemodel_handle *dmod, domid_t domid, uint64_t msi_addr, uint32_t msi_data) { return xc_hvm_inject_msi(dmod, domid, msi_addr, msi_data); } static inline int xendevicemodel_track_dirty_vram( xendevicemodel_handle *dmod, domid_t domid, uint64_t first_pfn, uint32_t nr, unsigned long *dirty_bitmap) { return xc_hvm_track_dirty_vram(dmod, domid, first_pfn, nr, dirty_bitmap); } static inline int xendevicemodel_modified_memory( xendevicemodel_handle *dmod, domid_t domid, uint64_t first_pfn, uint32_t nr) { return xc_hvm_modified_memory(dmod, domid, first_pfn, nr); } static inline int xendevicemodel_set_mem_type( xendevicemodel_handle *dmod, domid_t domid, hvmmem_type_t mem_type, uint64_t first_pfn, uint32_t nr) { return xc_hvm_set_mem_type(dmod, domid, mem_type, first_pfn, nr); } static inline int xendevicemodel_restrict( xendevicemodel_handle *dmod, domid_t domid) { errno = ENOTTY; return -1; } #else /* CONFIG_XEN_CTRL_INTERFACE_VERSION >= 40900 */ #undef XC_WANT_COMPAT_DEVICEMODEL_API #include #endif extern xendevicemodel_handle *xen_dmod; static inline int xen_set_mem_type(domid_t domid, hvmmem_type_t type, uint64_t first_pfn, uint32_t nr) { return xendevicemodel_set_mem_type(xen_dmod, domid, type, first_pfn, nr); } static inline int xen_set_pci_intx_level(domid_t domid, uint16_t segment, uint8_t bus, uint8_t device, uint8_t intx, unsigned int level) { return xendevicemodel_set_pci_intx_level(xen_dmod, domid, segment, bus, device, intx, level); } static inline int xen_set_pci_link_route(domid_t domid, uint8_t link, uint8_t irq) { return xendevicemodel_set_pci_link_route(xen_dmod, domid, link, irq); } static inline int xen_inject_msi(domid_t domid, uint64_t msi_addr, uint32_t msi_data) { return xendevicemodel_inject_msi(xen_dmod, domid, msi_addr, msi_data); } static inline int xen_set_isa_irq_level(domid_t domid, uint8_t irq, unsigned int level) { return xendevicemodel_set_isa_irq_level(xen_dmod, domid, irq, level); } static inline int xen_track_dirty_vram(domid_t domid, uint64_t first_pfn, uint32_t nr, unsigned long *bitmap) { return xendevicemodel_track_dirty_vram(xen_dmod, domid, first_pfn, nr, bitmap); } static inline int xen_modified_memory(domid_t domid, uint64_t first_pfn, uint32_t nr) { return xendevicemodel_modified_memory(xen_dmod, domid, first_pfn, nr); } static inline int xen_restrict(domid_t domid) { int rc = xendevicemodel_restrict(xen_dmod, domid); trace_xen_domid_restrict(errno); if (errno == ENOTTY) { return 0; } return rc; } /* Xen 4.2 through 4.6 */ #if CONFIG_XEN_CTRL_INTERFACE_VERSION < 40701 typedef xc_interface xenforeignmemory_handle; typedef xc_evtchn xenevtchn_handle; typedef xc_gnttab xengnttab_handle; #define xenevtchn_open(l, f) xc_evtchn_open(l, f); #define xenevtchn_close(h) xc_evtchn_close(h) #define xenevtchn_fd(h) xc_evtchn_fd(h) #define xenevtchn_pending(h) xc_evtchn_pending(h) #define xenevtchn_notify(h, p) xc_evtchn_notify(h, p) #define xenevtchn_bind_interdomain(h, d, p) xc_evtchn_bind_interdomain(h, d, p) #define xenevtchn_unmask(h, p) xc_evtchn_unmask(h, p) #define xenevtchn_unbind(h, p) xc_evtchn_unbind(h, p) #define xengnttab_open(l, f) xc_gnttab_open(l, f) #define xengnttab_close(h) xc_gnttab_close(h) #define xengnttab_set_max_grants(h, n) xc_gnttab_set_max_grants(h, n) #define xengnttab_map_grant_ref(h, d, r, p) xc_gnttab_map_grant_ref(h, d, r, p) #define xengnttab_unmap(h, a, n) xc_gnttab_munmap(h, a, n) #define xengnttab_map_grant_refs(h, c, d, r, p) \ xc_gnttab_map_grant_refs(h, c, d, r, p) #define xengnttab_map_domain_grant_refs(h, c, d, r, p) \ xc_gnttab_map_domain_grant_refs(h, c, d, r, p) #define xenforeignmemory_open(l, f) xen_xc #define xenforeignmemory_close(h) static inline void *xenforeignmemory_map(xc_interface *h, uint32_t dom, int prot, size_t pages, const xen_pfn_t arr[/*pages*/], int err[/*pages*/]) { if (err) return xc_map_foreign_bulk(h, dom, prot, arr, err, pages); else return xc_map_foreign_pages(h, dom, prot, arr, pages); } #define xenforeignmemory_unmap(h, p, s) munmap(p, s * XC_PAGE_SIZE) #else /* CONFIG_XEN_CTRL_INTERFACE_VERSION >= 40701 */ #include #include #include #endif extern xenforeignmemory_handle *xen_fmem; void destroy_hvm_domain(bool reboot); /* shutdown/destroy current domain because of an error */ void xen_shutdown_fatal_error(const char *fmt, ...) GCC_FMT_ATTR(1, 2); #ifdef HVM_PARAM_VMPORT_REGS_PFN static inline int xen_get_vmport_regs_pfn(xc_interface *xc, domid_t dom, xen_pfn_t *vmport_regs_pfn) { int rc; uint64_t value; rc = xc_hvm_param_get(xc, dom, HVM_PARAM_VMPORT_REGS_PFN, &value); if (rc >= 0) { *vmport_regs_pfn = (xen_pfn_t) value; } return rc; } #else static inline int xen_get_vmport_regs_pfn(xc_interface *xc, domid_t dom, xen_pfn_t *vmport_regs_pfn) { return -ENOSYS; } #endif /* Xen before 4.6 */ #if CONFIG_XEN_CTRL_INTERFACE_VERSION < 40600 #ifndef HVM_IOREQSRV_BUFIOREQ_ATOMIC #define HVM_IOREQSRV_BUFIOREQ_ATOMIC 2 #endif #endif static inline int xen_get_default_ioreq_server_info(domid_t dom, xen_pfn_t *ioreq_pfn, xen_pfn_t *bufioreq_pfn, evtchn_port_t *bufioreq_evtchn) { unsigned long param; int rc; rc = xc_get_hvm_param(xen_xc, dom, HVM_PARAM_IOREQ_PFN, ¶m); if (rc < 0) { fprintf(stderr, "failed to get HVM_PARAM_IOREQ_PFN\n"); return -1; } *ioreq_pfn = param; rc = xc_get_hvm_param(xen_xc, dom, HVM_PARAM_BUFIOREQ_PFN, ¶m); if (rc < 0) { fprintf(stderr, "failed to get HVM_PARAM_BUFIOREQ_PFN\n"); return -1; } *bufioreq_pfn = param; rc = xc_get_hvm_param(xen_xc, dom, HVM_PARAM_BUFIOREQ_EVTCHN, ¶m); if (rc < 0) { fprintf(stderr, "failed to get HVM_PARAM_BUFIOREQ_EVTCHN\n"); return -1; } *bufioreq_evtchn = param; return 0; } /* Xen before 4.5 */ #if CONFIG_XEN_CTRL_INTERFACE_VERSION < 40500 #ifndef HVM_PARAM_BUFIOREQ_EVTCHN #define HVM_PARAM_BUFIOREQ_EVTCHN 26 #endif #define IOREQ_TYPE_PCI_CONFIG 2 typedef uint16_t ioservid_t; static inline void xen_map_memory_section(domid_t dom, ioservid_t ioservid, MemoryRegionSection *section) { } static inline void xen_unmap_memory_section(domid_t dom, ioservid_t ioservid, MemoryRegionSection *section) { } static inline void xen_map_io_section(domid_t dom, ioservid_t ioservid, MemoryRegionSection *section) { } static inline void xen_unmap_io_section(domid_t dom, ioservid_t ioservid, MemoryRegionSection *section) { } static inline void xen_map_pcidev(domid_t dom, ioservid_t ioservid, PCIDevice *pci_dev) { } static inline void xen_unmap_pcidev(domid_t dom, ioservid_t ioservid, PCIDevice *pci_dev) { } static inline void xen_create_ioreq_server(domid_t dom, ioservid_t *ioservid) { } static inline void xen_destroy_ioreq_server(domid_t dom, ioservid_t ioservid) { } static inline int xen_get_ioreq_server_info(domid_t dom, ioservid_t ioservid, xen_pfn_t *ioreq_pfn, xen_pfn_t *bufioreq_pfn, evtchn_port_t *bufioreq_evtchn) { return xen_get_default_ioreq_server_info(dom, ioreq_pfn, bufioreq_pfn, bufioreq_evtchn); } static inline int xen_set_ioreq_server_state(domid_t dom, ioservid_t ioservid, bool enable) { return 0; } /* Xen 4.5 */ #else static bool use_default_ioreq_server; static inline void xen_map_memory_section(domid_t dom, ioservid_t ioservid, MemoryRegionSection *section) { hwaddr start_addr = section->offset_within_address_space; ram_addr_t size = int128_get64(section->size); hwaddr end_addr = start_addr + size - 1; if (use_default_ioreq_server) { return; } trace_xen_map_mmio_range(ioservid, start_addr, end_addr); xendevicemodel_map_io_range_to_ioreq_server(xen_dmod, dom, ioservid, 1, start_addr, end_addr); } static inline void xen_unmap_memory_section(domid_t dom, ioservid_t ioservid, MemoryRegionSection *section) { hwaddr start_addr = section->offset_within_address_space; ram_addr_t size = int128_get64(section->size); hwaddr end_addr = start_addr + size - 1; if (use_default_ioreq_server) { return; } trace_xen_unmap_mmio_range(ioservid, start_addr, end_addr); xendevicemodel_unmap_io_range_from_ioreq_server(xen_dmod, dom, ioservid, 1, start_addr, end_addr); } static inline void xen_map_io_section(domid_t dom, ioservid_t ioservid, MemoryRegionSection *section) { hwaddr start_addr = section->offset_within_address_space; ram_addr_t size = int128_get64(section->size); hwaddr end_addr = start_addr + size - 1; if (use_default_ioreq_server) { return; } trace_xen_map_portio_range(ioservid, start_addr, end_addr); xendevicemodel_map_io_range_to_ioreq_server(xen_dmod, dom, ioservid, 0, start_addr, end_addr); } static inline void xen_unmap_io_section(domid_t dom, ioservid_t ioservid, MemoryRegionSection *section) { hwaddr start_addr = section->offset_within_address_space; ram_addr_t size = int128_get64(section->size); hwaddr end_addr = start_addr + size - 1; if (use_default_ioreq_server) { return; } trace_xen_unmap_portio_range(ioservid, start_addr, end_addr); xendevicemodel_unmap_io_range_from_ioreq_server(xen_dmod, dom, ioservid, 0, start_addr, end_addr); } static inline void xen_map_pcidev(domid_t dom, ioservid_t ioservid, PCIDevice *pci_dev) { if (use_default_ioreq_server) { return; } trace_xen_map_pcidev(ioservid, pci_bus_num(pci_dev->bus), PCI_SLOT(pci_dev->devfn), PCI_FUNC(pci_dev->devfn)); xendevicemodel_map_pcidev_to_ioreq_server(xen_dmod, dom, ioservid, 0, pci_bus_num(pci_dev->bus), PCI_SLOT(pci_dev->devfn), PCI_FUNC(pci_dev->devfn)); } static inline void xen_unmap_pcidev(domid_t dom, ioservid_t ioservid, PCIDevice *pci_dev) { if (use_default_ioreq_server) { return; } trace_xen_unmap_pcidev(ioservid, pci_bus_num(pci_dev->bus), PCI_SLOT(pci_dev->devfn), PCI_FUNC(pci_dev->devfn)); xendevicemodel_unmap_pcidev_from_ioreq_server(xen_dmod, dom, ioservid, 0, pci_bus_num(pci_dev->bus), PCI_SLOT(pci_dev->devfn), PCI_FUNC(pci_dev->devfn)); } static inline void xen_create_ioreq_server(domid_t dom, ioservid_t *ioservid) { int rc = xendevicemodel_create_ioreq_server(xen_dmod, dom, HVM_IOREQSRV_BUFIOREQ_ATOMIC, ioservid); if (rc == 0) { trace_xen_ioreq_server_create(*ioservid); return; } *ioservid = 0; use_default_ioreq_server = true; trace_xen_default_ioreq_server(); } static inline void xen_destroy_ioreq_server(domid_t dom, ioservid_t ioservid) { if (use_default_ioreq_server) { return; } trace_xen_ioreq_server_destroy(ioservid); xendevicemodel_destroy_ioreq_server(xen_dmod, dom, ioservid); } static inline int xen_get_ioreq_server_info(domid_t dom, ioservid_t ioservid, xen_pfn_t *ioreq_pfn, xen_pfn_t *bufioreq_pfn, evtchn_port_t *bufioreq_evtchn) { if (use_default_ioreq_server) { return xen_get_default_ioreq_server_info(dom, ioreq_pfn, bufioreq_pfn, bufioreq_evtchn); } return xendevicemodel_get_ioreq_server_info(xen_dmod, dom, ioservid, ioreq_pfn, bufioreq_pfn, bufioreq_evtchn); } static inline int xen_set_ioreq_server_state(domid_t dom, ioservid_t ioservid, bool enable) { if (use_default_ioreq_server) { return 0; } trace_xen_ioreq_server_state(ioservid, enable); return xendevicemodel_set_ioreq_server_state(xen_dmod, dom, ioservid, enable); } #endif #if CONFIG_XEN_CTRL_INTERFACE_VERSION < 40600 static inline int xen_xc_domain_add_to_physmap(xc_interface *xch, uint32_t domid, unsigned int space, unsigned long idx, xen_pfn_t gpfn) { return xc_domain_add_to_physmap(xch, domid, space, idx, gpfn); } #else static inline int xen_xc_domain_add_to_physmap(xc_interface *xch, uint32_t domid, unsigned int space, unsigned long idx, xen_pfn_t gpfn) { /* In Xen 4.6 rc is -1 and errno contains the error value. */ int rc = xc_domain_add_to_physmap(xch, domid, space, idx, gpfn); if (rc == -1) return errno; return rc; } #endif #ifdef CONFIG_XEN_PV_DOMAIN_BUILD #if CONFIG_XEN_CTRL_INTERFACE_VERSION < 40700 static inline int xen_domain_create(xc_interface *xc, uint32_t ssidref, xen_domain_handle_t handle, uint32_t flags, uint32_t *pdomid) { return xc_domain_create(xc, ssidref, handle, flags, pdomid); } #else static inline int xen_domain_create(xc_interface *xc, uint32_t ssidref, xen_domain_handle_t handle, uint32_t flags, uint32_t *pdomid) { return xc_domain_create(xc, ssidref, handle, flags, pdomid, NULL); } #endif #endif /* Xen before 4.8 */ #if CONFIG_XEN_CTRL_INTERFACE_VERSION < 40800 typedef void *xengnttab_grant_copy_segment_t; static inline int xengnttab_grant_copy(xengnttab_handle *xgt, uint32_t count, xengnttab_grant_copy_segment_t *segs) { return -ENOSYS; } #endif #endif /* QEMU_HW_XEN_COMMON_H */