3c5fd80743
If the kernel irqchip has been disabled, we don't want the
{add,release}_adapter_routes routines to call any kvm_irqchip_*
interfaces, as they may rely on an irqchip actually having been
created. Just take a quick exit in that case instead. If you are
trying to use irqfd without a kernel irqchip, we will fail with
an error.
Also initialize routes->gsi[] with -1 in the virtio-ccw handling,
to make sure we don't trip over other errors, either. (Nobody
else uses the gsi array in that structure.)
Fixes: d426d9fba8
("s390x/virtio-ccw: wire up irq routing and irqfds")
Reviewed-by: Thomas Huth <thuth@redhat.com>
Acked-by: Christian Borntraeger <borntraeger@de.ibm.com>
Message-Id: <20200117111147.5006-1-cohuck@redhat.com>
Signed-off-by: Cornelia Huck <cohuck@redhat.com>
1242 lines
39 KiB
C
1242 lines
39 KiB
C
/*
|
|
* virtio ccw target implementation
|
|
*
|
|
* Copyright 2012,2015 IBM Corp.
|
|
* Author(s): Cornelia Huck <cornelia.huck@de.ibm.com>
|
|
* Pierre Morel <pmorel@linux.vnet.ibm.com>
|
|
*
|
|
* This work is licensed under the terms of the GNU GPL, version 2 or (at
|
|
* your option) any later version. See the COPYING file in the top-level
|
|
* directory.
|
|
*/
|
|
|
|
#include "qemu/osdep.h"
|
|
#include "qapi/error.h"
|
|
#include "sysemu/kvm.h"
|
|
#include "net/net.h"
|
|
#include "hw/virtio/virtio.h"
|
|
#include "migration/qemu-file-types.h"
|
|
#include "hw/virtio/virtio-net.h"
|
|
#include "hw/sysbus.h"
|
|
#include "qemu/bitops.h"
|
|
#include "qemu/error-report.h"
|
|
#include "qemu/module.h"
|
|
#include "hw/virtio/virtio-access.h"
|
|
#include "hw/virtio/virtio-bus.h"
|
|
#include "hw/s390x/adapter.h"
|
|
#include "hw/s390x/s390_flic.h"
|
|
|
|
#include "hw/s390x/ioinst.h"
|
|
#include "hw/s390x/css.h"
|
|
#include "virtio-ccw.h"
|
|
#include "trace.h"
|
|
#include "hw/s390x/css-bridge.h"
|
|
#include "hw/s390x/s390-virtio-ccw.h"
|
|
|
|
#define NR_CLASSIC_INDICATOR_BITS 64
|
|
|
|
static int virtio_ccw_dev_post_load(void *opaque, int version_id)
|
|
{
|
|
VirtioCcwDevice *dev = VIRTIO_CCW_DEVICE(opaque);
|
|
CcwDevice *ccw_dev = CCW_DEVICE(dev);
|
|
CCWDeviceClass *ck = CCW_DEVICE_GET_CLASS(ccw_dev);
|
|
|
|
ccw_dev->sch->driver_data = dev;
|
|
if (ccw_dev->sch->thinint_active) {
|
|
dev->routes.adapter.adapter_id = css_get_adapter_id(
|
|
CSS_IO_ADAPTER_VIRTIO,
|
|
dev->thinint_isc);
|
|
}
|
|
/* Re-fill subch_id after loading the subchannel states.*/
|
|
if (ck->refill_ids) {
|
|
ck->refill_ids(ccw_dev);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
typedef struct VirtioCcwDeviceTmp {
|
|
VirtioCcwDevice *parent;
|
|
uint16_t config_vector;
|
|
} VirtioCcwDeviceTmp;
|
|
|
|
static int virtio_ccw_dev_tmp_pre_save(void *opaque)
|
|
{
|
|
VirtioCcwDeviceTmp *tmp = opaque;
|
|
VirtioCcwDevice *dev = tmp->parent;
|
|
VirtIODevice *vdev = virtio_bus_get_device(&dev->bus);
|
|
|
|
tmp->config_vector = vdev->config_vector;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int virtio_ccw_dev_tmp_post_load(void *opaque, int version_id)
|
|
{
|
|
VirtioCcwDeviceTmp *tmp = opaque;
|
|
VirtioCcwDevice *dev = tmp->parent;
|
|
VirtIODevice *vdev = virtio_bus_get_device(&dev->bus);
|
|
|
|
vdev->config_vector = tmp->config_vector;
|
|
return 0;
|
|
}
|
|
|
|
const VMStateDescription vmstate_virtio_ccw_dev_tmp = {
|
|
.name = "s390_virtio_ccw_dev_tmp",
|
|
.pre_save = virtio_ccw_dev_tmp_pre_save,
|
|
.post_load = virtio_ccw_dev_tmp_post_load,
|
|
.fields = (VMStateField[]) {
|
|
VMSTATE_UINT16(config_vector, VirtioCcwDeviceTmp),
|
|
VMSTATE_END_OF_LIST()
|
|
}
|
|
};
|
|
|
|
const VMStateDescription vmstate_virtio_ccw_dev = {
|
|
.name = "s390_virtio_ccw_dev",
|
|
.version_id = 1,
|
|
.minimum_version_id = 1,
|
|
.post_load = virtio_ccw_dev_post_load,
|
|
.fields = (VMStateField[]) {
|
|
VMSTATE_CCW_DEVICE(parent_obj, VirtioCcwDevice),
|
|
VMSTATE_PTR_TO_IND_ADDR(indicators, VirtioCcwDevice),
|
|
VMSTATE_PTR_TO_IND_ADDR(indicators2, VirtioCcwDevice),
|
|
VMSTATE_PTR_TO_IND_ADDR(summary_indicator, VirtioCcwDevice),
|
|
/*
|
|
* Ugly hack because VirtIODevice does not migrate itself.
|
|
* This also makes legacy via vmstate_save_state possible.
|
|
*/
|
|
VMSTATE_WITH_TMP(VirtioCcwDevice, VirtioCcwDeviceTmp,
|
|
vmstate_virtio_ccw_dev_tmp),
|
|
VMSTATE_STRUCT(routes, VirtioCcwDevice, 1, vmstate_adapter_routes,
|
|
AdapterRoutes),
|
|
VMSTATE_UINT8(thinint_isc, VirtioCcwDevice),
|
|
VMSTATE_INT32(revision, VirtioCcwDevice),
|
|
VMSTATE_END_OF_LIST()
|
|
}
|
|
};
|
|
|
|
static void virtio_ccw_bus_new(VirtioBusState *bus, size_t bus_size,
|
|
VirtioCcwDevice *dev);
|
|
|
|
VirtIODevice *virtio_ccw_get_vdev(SubchDev *sch)
|
|
{
|
|
VirtIODevice *vdev = NULL;
|
|
VirtioCcwDevice *dev = sch->driver_data;
|
|
|
|
if (dev) {
|
|
vdev = virtio_bus_get_device(&dev->bus);
|
|
}
|
|
return vdev;
|
|
}
|
|
|
|
static void virtio_ccw_start_ioeventfd(VirtioCcwDevice *dev)
|
|
{
|
|
virtio_bus_start_ioeventfd(&dev->bus);
|
|
}
|
|
|
|
static void virtio_ccw_stop_ioeventfd(VirtioCcwDevice *dev)
|
|
{
|
|
virtio_bus_stop_ioeventfd(&dev->bus);
|
|
}
|
|
|
|
static bool virtio_ccw_ioeventfd_enabled(DeviceState *d)
|
|
{
|
|
VirtioCcwDevice *dev = VIRTIO_CCW_DEVICE(d);
|
|
|
|
return (dev->flags & VIRTIO_CCW_FLAG_USE_IOEVENTFD) != 0;
|
|
}
|
|
|
|
static int virtio_ccw_ioeventfd_assign(DeviceState *d, EventNotifier *notifier,
|
|
int n, bool assign)
|
|
{
|
|
VirtioCcwDevice *dev = VIRTIO_CCW_DEVICE(d);
|
|
CcwDevice *ccw_dev = CCW_DEVICE(dev);
|
|
SubchDev *sch = ccw_dev->sch;
|
|
uint32_t sch_id = (css_build_subchannel_id(sch) << 16) | sch->schid;
|
|
|
|
return s390_assign_subch_ioeventfd(notifier, sch_id, n, assign);
|
|
}
|
|
|
|
/* Communication blocks used by several channel commands. */
|
|
typedef struct VqInfoBlockLegacy {
|
|
uint64_t queue;
|
|
uint32_t align;
|
|
uint16_t index;
|
|
uint16_t num;
|
|
} QEMU_PACKED VqInfoBlockLegacy;
|
|
|
|
typedef struct VqInfoBlock {
|
|
uint64_t desc;
|
|
uint32_t res0;
|
|
uint16_t index;
|
|
uint16_t num;
|
|
uint64_t avail;
|
|
uint64_t used;
|
|
} QEMU_PACKED VqInfoBlock;
|
|
|
|
typedef struct VqConfigBlock {
|
|
uint16_t index;
|
|
uint16_t num_max;
|
|
} QEMU_PACKED VqConfigBlock;
|
|
|
|
typedef struct VirtioFeatDesc {
|
|
uint32_t features;
|
|
uint8_t index;
|
|
} QEMU_PACKED VirtioFeatDesc;
|
|
|
|
typedef struct VirtioThinintInfo {
|
|
hwaddr summary_indicator;
|
|
hwaddr device_indicator;
|
|
uint64_t ind_bit;
|
|
uint8_t isc;
|
|
} QEMU_PACKED VirtioThinintInfo;
|
|
|
|
typedef struct VirtioRevInfo {
|
|
uint16_t revision;
|
|
uint16_t length;
|
|
uint8_t data[0];
|
|
} QEMU_PACKED VirtioRevInfo;
|
|
|
|
/* Specify where the virtqueues for the subchannel are in guest memory. */
|
|
static int virtio_ccw_set_vqs(SubchDev *sch, VqInfoBlock *info,
|
|
VqInfoBlockLegacy *linfo)
|
|
{
|
|
VirtIODevice *vdev = virtio_ccw_get_vdev(sch);
|
|
uint16_t index = info ? info->index : linfo->index;
|
|
uint16_t num = info ? info->num : linfo->num;
|
|
uint64_t desc = info ? info->desc : linfo->queue;
|
|
|
|
if (index >= VIRTIO_QUEUE_MAX) {
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* Current code in virtio.c relies on 4K alignment. */
|
|
if (linfo && desc && (linfo->align != 4096)) {
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (!vdev) {
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (info) {
|
|
virtio_queue_set_rings(vdev, index, desc, info->avail, info->used);
|
|
} else {
|
|
virtio_queue_set_addr(vdev, index, desc);
|
|
}
|
|
if (!desc) {
|
|
virtio_queue_set_vector(vdev, index, VIRTIO_NO_VECTOR);
|
|
} else {
|
|
if (info) {
|
|
/* virtio-1 allows changing the ring size. */
|
|
if (virtio_queue_get_max_num(vdev, index) < num) {
|
|
/* Fail if we exceed the maximum number. */
|
|
return -EINVAL;
|
|
}
|
|
virtio_queue_set_num(vdev, index, num);
|
|
} else if (virtio_queue_get_num(vdev, index) > num) {
|
|
/* Fail if we don't have a big enough queue. */
|
|
return -EINVAL;
|
|
}
|
|
/* We ignore possible increased num for legacy for compatibility. */
|
|
virtio_queue_set_vector(vdev, index, index);
|
|
}
|
|
/* tell notify handler in case of config change */
|
|
vdev->config_vector = VIRTIO_QUEUE_MAX;
|
|
return 0;
|
|
}
|
|
|
|
static void virtio_ccw_reset_virtio(VirtioCcwDevice *dev, VirtIODevice *vdev)
|
|
{
|
|
CcwDevice *ccw_dev = CCW_DEVICE(dev);
|
|
|
|
virtio_ccw_stop_ioeventfd(dev);
|
|
virtio_reset(vdev);
|
|
if (dev->indicators) {
|
|
release_indicator(&dev->routes.adapter, dev->indicators);
|
|
dev->indicators = NULL;
|
|
}
|
|
if (dev->indicators2) {
|
|
release_indicator(&dev->routes.adapter, dev->indicators2);
|
|
dev->indicators2 = NULL;
|
|
}
|
|
if (dev->summary_indicator) {
|
|
release_indicator(&dev->routes.adapter, dev->summary_indicator);
|
|
dev->summary_indicator = NULL;
|
|
}
|
|
ccw_dev->sch->thinint_active = false;
|
|
}
|
|
|
|
static int virtio_ccw_handle_set_vq(SubchDev *sch, CCW1 ccw, bool check_len,
|
|
bool is_legacy)
|
|
{
|
|
int ret;
|
|
VqInfoBlock info;
|
|
VqInfoBlockLegacy linfo;
|
|
size_t info_len = is_legacy ? sizeof(linfo) : sizeof(info);
|
|
|
|
if (check_len) {
|
|
if (ccw.count != info_len) {
|
|
return -EINVAL;
|
|
}
|
|
} else if (ccw.count < info_len) {
|
|
/* Can't execute command. */
|
|
return -EINVAL;
|
|
}
|
|
if (!ccw.cda) {
|
|
return -EFAULT;
|
|
}
|
|
if (is_legacy) {
|
|
ccw_dstream_read(&sch->cds, linfo);
|
|
linfo.queue = be64_to_cpu(linfo.queue);
|
|
linfo.align = be32_to_cpu(linfo.align);
|
|
linfo.index = be16_to_cpu(linfo.index);
|
|
linfo.num = be16_to_cpu(linfo.num);
|
|
ret = virtio_ccw_set_vqs(sch, NULL, &linfo);
|
|
} else {
|
|
ccw_dstream_read(&sch->cds, info);
|
|
info.desc = be64_to_cpu(info.desc);
|
|
info.index = be16_to_cpu(info.index);
|
|
info.num = be16_to_cpu(info.num);
|
|
info.avail = be64_to_cpu(info.avail);
|
|
info.used = be64_to_cpu(info.used);
|
|
ret = virtio_ccw_set_vqs(sch, &info, NULL);
|
|
}
|
|
sch->curr_status.scsw.count = 0;
|
|
return ret;
|
|
}
|
|
|
|
static int virtio_ccw_cb(SubchDev *sch, CCW1 ccw)
|
|
{
|
|
int ret;
|
|
VirtioRevInfo revinfo;
|
|
uint8_t status;
|
|
VirtioFeatDesc features;
|
|
hwaddr indicators;
|
|
VqConfigBlock vq_config;
|
|
VirtioCcwDevice *dev = sch->driver_data;
|
|
VirtIODevice *vdev = virtio_ccw_get_vdev(sch);
|
|
bool check_len;
|
|
int len;
|
|
VirtioThinintInfo thinint;
|
|
|
|
if (!dev) {
|
|
return -EINVAL;
|
|
}
|
|
|
|
trace_virtio_ccw_interpret_ccw(sch->cssid, sch->ssid, sch->schid,
|
|
ccw.cmd_code);
|
|
check_len = !((ccw.flags & CCW_FLAG_SLI) && !(ccw.flags & CCW_FLAG_DC));
|
|
|
|
if (dev->force_revision_1 && dev->revision < 0 &&
|
|
ccw.cmd_code != CCW_CMD_SET_VIRTIO_REV) {
|
|
/*
|
|
* virtio-1 drivers must start with negotiating to a revision >= 1,
|
|
* so post a command reject for all other commands
|
|
*/
|
|
return -ENOSYS;
|
|
}
|
|
|
|
/* Look at the command. */
|
|
switch (ccw.cmd_code) {
|
|
case CCW_CMD_SET_VQ:
|
|
ret = virtio_ccw_handle_set_vq(sch, ccw, check_len, dev->revision < 1);
|
|
break;
|
|
case CCW_CMD_VDEV_RESET:
|
|
virtio_ccw_reset_virtio(dev, vdev);
|
|
ret = 0;
|
|
break;
|
|
case CCW_CMD_READ_FEAT:
|
|
if (check_len) {
|
|
if (ccw.count != sizeof(features)) {
|
|
ret = -EINVAL;
|
|
break;
|
|
}
|
|
} else if (ccw.count < sizeof(features)) {
|
|
/* Can't execute command. */
|
|
ret = -EINVAL;
|
|
break;
|
|
}
|
|
if (!ccw.cda) {
|
|
ret = -EFAULT;
|
|
} else {
|
|
VirtioDeviceClass *vdc = VIRTIO_DEVICE_GET_CLASS(vdev);
|
|
|
|
ccw_dstream_advance(&sch->cds, sizeof(features.features));
|
|
ccw_dstream_read(&sch->cds, features.index);
|
|
if (features.index == 0) {
|
|
if (dev->revision >= 1) {
|
|
/* Don't offer legacy features for modern devices. */
|
|
features.features = (uint32_t)
|
|
(vdev->host_features & ~vdc->legacy_features);
|
|
} else {
|
|
features.features = (uint32_t)vdev->host_features;
|
|
}
|
|
} else if ((features.index == 1) && (dev->revision >= 1)) {
|
|
/*
|
|
* Only offer feature bits beyond 31 if the guest has
|
|
* negotiated at least revision 1.
|
|
*/
|
|
features.features = (uint32_t)(vdev->host_features >> 32);
|
|
} else {
|
|
/* Return zeroes if the guest supports more feature bits. */
|
|
features.features = 0;
|
|
}
|
|
ccw_dstream_rewind(&sch->cds);
|
|
features.features = cpu_to_le32(features.features);
|
|
ccw_dstream_write(&sch->cds, features.features);
|
|
sch->curr_status.scsw.count = ccw.count - sizeof(features);
|
|
ret = 0;
|
|
}
|
|
break;
|
|
case CCW_CMD_WRITE_FEAT:
|
|
if (check_len) {
|
|
if (ccw.count != sizeof(features)) {
|
|
ret = -EINVAL;
|
|
break;
|
|
}
|
|
} else if (ccw.count < sizeof(features)) {
|
|
/* Can't execute command. */
|
|
ret = -EINVAL;
|
|
break;
|
|
}
|
|
if (!ccw.cda) {
|
|
ret = -EFAULT;
|
|
} else {
|
|
ccw_dstream_read(&sch->cds, features);
|
|
features.features = le32_to_cpu(features.features);
|
|
if (features.index == 0) {
|
|
virtio_set_features(vdev,
|
|
(vdev->guest_features & 0xffffffff00000000ULL) |
|
|
features.features);
|
|
} else if ((features.index == 1) && (dev->revision >= 1)) {
|
|
/*
|
|
* If the guest did not negotiate at least revision 1,
|
|
* we did not offer it any feature bits beyond 31. Such a
|
|
* guest passing us any bit here is therefore buggy.
|
|
*/
|
|
virtio_set_features(vdev,
|
|
(vdev->guest_features & 0x00000000ffffffffULL) |
|
|
((uint64_t)features.features << 32));
|
|
} else {
|
|
/*
|
|
* If the guest supports more feature bits, assert that it
|
|
* passes us zeroes for those we don't support.
|
|
*/
|
|
if (features.features) {
|
|
qemu_log_mask(LOG_GUEST_ERROR,
|
|
"Guest bug: features[%i]=%x (expected 0)",
|
|
features.index, features.features);
|
|
/* XXX: do a unit check here? */
|
|
}
|
|
}
|
|
sch->curr_status.scsw.count = ccw.count - sizeof(features);
|
|
ret = 0;
|
|
}
|
|
break;
|
|
case CCW_CMD_READ_CONF:
|
|
if (check_len) {
|
|
if (ccw.count > vdev->config_len) {
|
|
ret = -EINVAL;
|
|
break;
|
|
}
|
|
}
|
|
len = MIN(ccw.count, vdev->config_len);
|
|
if (!ccw.cda) {
|
|
ret = -EFAULT;
|
|
} else {
|
|
virtio_bus_get_vdev_config(&dev->bus, vdev->config);
|
|
ccw_dstream_write_buf(&sch->cds, vdev->config, len);
|
|
sch->curr_status.scsw.count = ccw.count - len;
|
|
ret = 0;
|
|
}
|
|
break;
|
|
case CCW_CMD_WRITE_CONF:
|
|
if (check_len) {
|
|
if (ccw.count > vdev->config_len) {
|
|
ret = -EINVAL;
|
|
break;
|
|
}
|
|
}
|
|
len = MIN(ccw.count, vdev->config_len);
|
|
if (!ccw.cda) {
|
|
ret = -EFAULT;
|
|
} else {
|
|
ret = ccw_dstream_read_buf(&sch->cds, vdev->config, len);
|
|
if (!ret) {
|
|
virtio_bus_set_vdev_config(&dev->bus, vdev->config);
|
|
sch->curr_status.scsw.count = ccw.count - len;
|
|
}
|
|
}
|
|
break;
|
|
case CCW_CMD_READ_STATUS:
|
|
if (check_len) {
|
|
if (ccw.count != sizeof(status)) {
|
|
ret = -EINVAL;
|
|
break;
|
|
}
|
|
} else if (ccw.count < sizeof(status)) {
|
|
/* Can't execute command. */
|
|
ret = -EINVAL;
|
|
break;
|
|
}
|
|
if (!ccw.cda) {
|
|
ret = -EFAULT;
|
|
} else {
|
|
address_space_stb(&address_space_memory, ccw.cda, vdev->status,
|
|
MEMTXATTRS_UNSPECIFIED, NULL);
|
|
sch->curr_status.scsw.count = ccw.count - sizeof(vdev->status);
|
|
ret = 0;
|
|
}
|
|
break;
|
|
case CCW_CMD_WRITE_STATUS:
|
|
if (check_len) {
|
|
if (ccw.count != sizeof(status)) {
|
|
ret = -EINVAL;
|
|
break;
|
|
}
|
|
} else if (ccw.count < sizeof(status)) {
|
|
/* Can't execute command. */
|
|
ret = -EINVAL;
|
|
break;
|
|
}
|
|
if (!ccw.cda) {
|
|
ret = -EFAULT;
|
|
} else {
|
|
ccw_dstream_read(&sch->cds, status);
|
|
if (!(status & VIRTIO_CONFIG_S_DRIVER_OK)) {
|
|
virtio_ccw_stop_ioeventfd(dev);
|
|
}
|
|
if (virtio_set_status(vdev, status) == 0) {
|
|
if (vdev->status == 0) {
|
|
virtio_ccw_reset_virtio(dev, vdev);
|
|
}
|
|
if (status & VIRTIO_CONFIG_S_DRIVER_OK) {
|
|
virtio_ccw_start_ioeventfd(dev);
|
|
}
|
|
sch->curr_status.scsw.count = ccw.count - sizeof(status);
|
|
ret = 0;
|
|
} else {
|
|
/* Trigger a command reject. */
|
|
ret = -ENOSYS;
|
|
}
|
|
}
|
|
break;
|
|
case CCW_CMD_SET_IND:
|
|
if (check_len) {
|
|
if (ccw.count != sizeof(indicators)) {
|
|
ret = -EINVAL;
|
|
break;
|
|
}
|
|
} else if (ccw.count < sizeof(indicators)) {
|
|
/* Can't execute command. */
|
|
ret = -EINVAL;
|
|
break;
|
|
}
|
|
if (sch->thinint_active) {
|
|
/* Trigger a command reject. */
|
|
ret = -ENOSYS;
|
|
break;
|
|
}
|
|
if (virtio_get_num_queues(vdev) > NR_CLASSIC_INDICATOR_BITS) {
|
|
/* More queues than indicator bits --> trigger a reject */
|
|
ret = -ENOSYS;
|
|
break;
|
|
}
|
|
if (!ccw.cda) {
|
|
ret = -EFAULT;
|
|
} else {
|
|
ccw_dstream_read(&sch->cds, indicators);
|
|
indicators = be64_to_cpu(indicators);
|
|
dev->indicators = get_indicator(indicators, sizeof(uint64_t));
|
|
sch->curr_status.scsw.count = ccw.count - sizeof(indicators);
|
|
ret = 0;
|
|
}
|
|
break;
|
|
case CCW_CMD_SET_CONF_IND:
|
|
if (check_len) {
|
|
if (ccw.count != sizeof(indicators)) {
|
|
ret = -EINVAL;
|
|
break;
|
|
}
|
|
} else if (ccw.count < sizeof(indicators)) {
|
|
/* Can't execute command. */
|
|
ret = -EINVAL;
|
|
break;
|
|
}
|
|
if (!ccw.cda) {
|
|
ret = -EFAULT;
|
|
} else {
|
|
ccw_dstream_read(&sch->cds, indicators);
|
|
indicators = be64_to_cpu(indicators);
|
|
dev->indicators2 = get_indicator(indicators, sizeof(uint64_t));
|
|
sch->curr_status.scsw.count = ccw.count - sizeof(indicators);
|
|
ret = 0;
|
|
}
|
|
break;
|
|
case CCW_CMD_READ_VQ_CONF:
|
|
if (check_len) {
|
|
if (ccw.count != sizeof(vq_config)) {
|
|
ret = -EINVAL;
|
|
break;
|
|
}
|
|
} else if (ccw.count < sizeof(vq_config)) {
|
|
/* Can't execute command. */
|
|
ret = -EINVAL;
|
|
break;
|
|
}
|
|
if (!ccw.cda) {
|
|
ret = -EFAULT;
|
|
} else {
|
|
ccw_dstream_read(&sch->cds, vq_config.index);
|
|
vq_config.index = be16_to_cpu(vq_config.index);
|
|
if (vq_config.index >= VIRTIO_QUEUE_MAX) {
|
|
ret = -EINVAL;
|
|
break;
|
|
}
|
|
vq_config.num_max = virtio_queue_get_num(vdev,
|
|
vq_config.index);
|
|
vq_config.num_max = cpu_to_be16(vq_config.num_max);
|
|
ccw_dstream_write(&sch->cds, vq_config.num_max);
|
|
sch->curr_status.scsw.count = ccw.count - sizeof(vq_config);
|
|
ret = 0;
|
|
}
|
|
break;
|
|
case CCW_CMD_SET_IND_ADAPTER:
|
|
if (check_len) {
|
|
if (ccw.count != sizeof(thinint)) {
|
|
ret = -EINVAL;
|
|
break;
|
|
}
|
|
} else if (ccw.count < sizeof(thinint)) {
|
|
/* Can't execute command. */
|
|
ret = -EINVAL;
|
|
break;
|
|
}
|
|
if (!ccw.cda) {
|
|
ret = -EFAULT;
|
|
} else if (dev->indicators && !sch->thinint_active) {
|
|
/* Trigger a command reject. */
|
|
ret = -ENOSYS;
|
|
} else {
|
|
if (ccw_dstream_read(&sch->cds, thinint)) {
|
|
ret = -EFAULT;
|
|
} else {
|
|
thinint.ind_bit = be64_to_cpu(thinint.ind_bit);
|
|
thinint.summary_indicator =
|
|
be64_to_cpu(thinint.summary_indicator);
|
|
thinint.device_indicator =
|
|
be64_to_cpu(thinint.device_indicator);
|
|
|
|
dev->summary_indicator =
|
|
get_indicator(thinint.summary_indicator, sizeof(uint8_t));
|
|
dev->indicators =
|
|
get_indicator(thinint.device_indicator,
|
|
thinint.ind_bit / 8 + 1);
|
|
dev->thinint_isc = thinint.isc;
|
|
dev->routes.adapter.ind_offset = thinint.ind_bit;
|
|
dev->routes.adapter.summary_offset = 7;
|
|
dev->routes.adapter.adapter_id = css_get_adapter_id(
|
|
CSS_IO_ADAPTER_VIRTIO,
|
|
dev->thinint_isc);
|
|
sch->thinint_active = ((dev->indicators != NULL) &&
|
|
(dev->summary_indicator != NULL));
|
|
sch->curr_status.scsw.count = ccw.count - sizeof(thinint);
|
|
ret = 0;
|
|
}
|
|
}
|
|
break;
|
|
case CCW_CMD_SET_VIRTIO_REV:
|
|
len = sizeof(revinfo);
|
|
if (ccw.count < len) {
|
|
ret = -EINVAL;
|
|
break;
|
|
}
|
|
if (!ccw.cda) {
|
|
ret = -EFAULT;
|
|
break;
|
|
}
|
|
ccw_dstream_read_buf(&sch->cds, &revinfo, 4);
|
|
revinfo.revision = be16_to_cpu(revinfo.revision);
|
|
revinfo.length = be16_to_cpu(revinfo.length);
|
|
if (ccw.count < len + revinfo.length ||
|
|
(check_len && ccw.count > len + revinfo.length)) {
|
|
ret = -EINVAL;
|
|
break;
|
|
}
|
|
/*
|
|
* Once we start to support revisions with additional data, we'll
|
|
* need to fetch it here. Nothing to do for now, though.
|
|
*/
|
|
if (dev->revision >= 0 ||
|
|
revinfo.revision > virtio_ccw_rev_max(dev) ||
|
|
(dev->force_revision_1 && !revinfo.revision)) {
|
|
ret = -ENOSYS;
|
|
break;
|
|
}
|
|
ret = 0;
|
|
dev->revision = revinfo.revision;
|
|
break;
|
|
default:
|
|
ret = -ENOSYS;
|
|
break;
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
static void virtio_sch_disable_cb(SubchDev *sch)
|
|
{
|
|
VirtioCcwDevice *dev = sch->driver_data;
|
|
|
|
dev->revision = -1;
|
|
}
|
|
|
|
static void virtio_ccw_device_realize(VirtioCcwDevice *dev, Error **errp)
|
|
{
|
|
VirtIOCCWDeviceClass *k = VIRTIO_CCW_DEVICE_GET_CLASS(dev);
|
|
CcwDevice *ccw_dev = CCW_DEVICE(dev);
|
|
CCWDeviceClass *ck = CCW_DEVICE_GET_CLASS(ccw_dev);
|
|
SubchDev *sch;
|
|
Error *err = NULL;
|
|
int i;
|
|
|
|
sch = css_create_sch(ccw_dev->devno, errp);
|
|
if (!sch) {
|
|
return;
|
|
}
|
|
if (!virtio_ccw_rev_max(dev) && dev->force_revision_1) {
|
|
error_setg(&err, "Invalid value of property max_rev "
|
|
"(is %d expected >= 1)", virtio_ccw_rev_max(dev));
|
|
goto out_err;
|
|
}
|
|
|
|
sch->driver_data = dev;
|
|
sch->ccw_cb = virtio_ccw_cb;
|
|
sch->disable_cb = virtio_sch_disable_cb;
|
|
sch->id.reserved = 0xff;
|
|
sch->id.cu_type = VIRTIO_CCW_CU_TYPE;
|
|
sch->do_subchannel_work = do_subchannel_work_virtual;
|
|
ccw_dev->sch = sch;
|
|
dev->indicators = NULL;
|
|
dev->revision = -1;
|
|
for (i = 0; i < ADAPTER_ROUTES_MAX_GSI; i++) {
|
|
dev->routes.gsi[i] = -1;
|
|
}
|
|
css_sch_build_virtual_schib(sch, 0, VIRTIO_CCW_CHPID_TYPE);
|
|
|
|
trace_virtio_ccw_new_device(
|
|
sch->cssid, sch->ssid, sch->schid, sch->devno,
|
|
ccw_dev->devno.valid ? "user-configured" : "auto-configured");
|
|
|
|
if (kvm_enabled() && !kvm_eventfds_enabled()) {
|
|
dev->flags &= ~VIRTIO_CCW_FLAG_USE_IOEVENTFD;
|
|
}
|
|
|
|
if (k->realize) {
|
|
k->realize(dev, &err);
|
|
if (err) {
|
|
goto out_err;
|
|
}
|
|
}
|
|
|
|
ck->realize(ccw_dev, &err);
|
|
if (err) {
|
|
goto out_err;
|
|
}
|
|
|
|
return;
|
|
|
|
out_err:
|
|
error_propagate(errp, err);
|
|
css_subch_assign(sch->cssid, sch->ssid, sch->schid, sch->devno, NULL);
|
|
ccw_dev->sch = NULL;
|
|
g_free(sch);
|
|
}
|
|
|
|
static void virtio_ccw_device_unrealize(VirtioCcwDevice *dev, Error **errp)
|
|
{
|
|
VirtIOCCWDeviceClass *dc = VIRTIO_CCW_DEVICE_GET_CLASS(dev);
|
|
CcwDevice *ccw_dev = CCW_DEVICE(dev);
|
|
SubchDev *sch = ccw_dev->sch;
|
|
|
|
if (dc->unrealize) {
|
|
dc->unrealize(dev, errp);
|
|
}
|
|
|
|
if (sch) {
|
|
css_subch_assign(sch->cssid, sch->ssid, sch->schid, sch->devno, NULL);
|
|
g_free(sch);
|
|
ccw_dev->sch = NULL;
|
|
}
|
|
if (dev->indicators) {
|
|
release_indicator(&dev->routes.adapter, dev->indicators);
|
|
dev->indicators = NULL;
|
|
}
|
|
}
|
|
|
|
/* DeviceState to VirtioCcwDevice. Note: used on datapath,
|
|
* be careful and test performance if you change this.
|
|
*/
|
|
static inline VirtioCcwDevice *to_virtio_ccw_dev_fast(DeviceState *d)
|
|
{
|
|
CcwDevice *ccw_dev = to_ccw_dev_fast(d);
|
|
|
|
return container_of(ccw_dev, VirtioCcwDevice, parent_obj);
|
|
}
|
|
|
|
static uint8_t virtio_set_ind_atomic(SubchDev *sch, uint64_t ind_loc,
|
|
uint8_t to_be_set)
|
|
{
|
|
uint8_t ind_old, ind_new;
|
|
hwaddr len = 1;
|
|
uint8_t *ind_addr;
|
|
|
|
ind_addr = cpu_physical_memory_map(ind_loc, &len, 1);
|
|
if (!ind_addr) {
|
|
error_report("%s(%x.%x.%04x): unable to access indicator",
|
|
__func__, sch->cssid, sch->ssid, sch->schid);
|
|
return -1;
|
|
}
|
|
do {
|
|
ind_old = *ind_addr;
|
|
ind_new = ind_old | to_be_set;
|
|
} while (atomic_cmpxchg(ind_addr, ind_old, ind_new) != ind_old);
|
|
trace_virtio_ccw_set_ind(ind_loc, ind_old, ind_new);
|
|
cpu_physical_memory_unmap(ind_addr, len, 1, len);
|
|
|
|
return ind_old;
|
|
}
|
|
|
|
static void virtio_ccw_notify(DeviceState *d, uint16_t vector)
|
|
{
|
|
VirtioCcwDevice *dev = to_virtio_ccw_dev_fast(d);
|
|
CcwDevice *ccw_dev = to_ccw_dev_fast(d);
|
|
SubchDev *sch = ccw_dev->sch;
|
|
uint64_t indicators;
|
|
|
|
if (vector == VIRTIO_NO_VECTOR) {
|
|
return;
|
|
}
|
|
/*
|
|
* vector < VIRTIO_QUEUE_MAX: notification for a virtqueue
|
|
* vector == VIRTIO_QUEUE_MAX: configuration change notification
|
|
* bits beyond that are unused and should never be notified for
|
|
*/
|
|
assert(vector <= VIRTIO_QUEUE_MAX);
|
|
|
|
if (vector < VIRTIO_QUEUE_MAX) {
|
|
if (!dev->indicators) {
|
|
return;
|
|
}
|
|
if (sch->thinint_active) {
|
|
/*
|
|
* In the adapter interrupt case, indicators points to a
|
|
* memory area that may be (way) larger than 64 bit and
|
|
* ind_bit indicates the start of the indicators in a big
|
|
* endian notation.
|
|
*/
|
|
uint64_t ind_bit = dev->routes.adapter.ind_offset;
|
|
|
|
virtio_set_ind_atomic(sch, dev->indicators->addr +
|
|
(ind_bit + vector) / 8,
|
|
0x80 >> ((ind_bit + vector) % 8));
|
|
if (!virtio_set_ind_atomic(sch, dev->summary_indicator->addr,
|
|
0x01)) {
|
|
css_adapter_interrupt(CSS_IO_ADAPTER_VIRTIO, dev->thinint_isc);
|
|
}
|
|
} else {
|
|
assert(vector < NR_CLASSIC_INDICATOR_BITS);
|
|
indicators = address_space_ldq(&address_space_memory,
|
|
dev->indicators->addr,
|
|
MEMTXATTRS_UNSPECIFIED,
|
|
NULL);
|
|
indicators |= 1ULL << vector;
|
|
address_space_stq(&address_space_memory, dev->indicators->addr,
|
|
indicators, MEMTXATTRS_UNSPECIFIED, NULL);
|
|
css_conditional_io_interrupt(sch);
|
|
}
|
|
} else {
|
|
if (!dev->indicators2) {
|
|
return;
|
|
}
|
|
indicators = address_space_ldq(&address_space_memory,
|
|
dev->indicators2->addr,
|
|
MEMTXATTRS_UNSPECIFIED,
|
|
NULL);
|
|
indicators |= 1ULL;
|
|
address_space_stq(&address_space_memory, dev->indicators2->addr,
|
|
indicators, MEMTXATTRS_UNSPECIFIED, NULL);
|
|
css_conditional_io_interrupt(sch);
|
|
}
|
|
}
|
|
|
|
static void virtio_ccw_reset(DeviceState *d)
|
|
{
|
|
VirtioCcwDevice *dev = VIRTIO_CCW_DEVICE(d);
|
|
VirtIODevice *vdev = virtio_bus_get_device(&dev->bus);
|
|
VirtIOCCWDeviceClass *vdc = VIRTIO_CCW_DEVICE_GET_CLASS(dev);
|
|
|
|
virtio_ccw_reset_virtio(dev, vdev);
|
|
if (vdc->parent_reset) {
|
|
vdc->parent_reset(d);
|
|
}
|
|
}
|
|
|
|
static void virtio_ccw_vmstate_change(DeviceState *d, bool running)
|
|
{
|
|
VirtioCcwDevice *dev = VIRTIO_CCW_DEVICE(d);
|
|
|
|
if (running) {
|
|
virtio_ccw_start_ioeventfd(dev);
|
|
} else {
|
|
virtio_ccw_stop_ioeventfd(dev);
|
|
}
|
|
}
|
|
|
|
static bool virtio_ccw_query_guest_notifiers(DeviceState *d)
|
|
{
|
|
CcwDevice *dev = CCW_DEVICE(d);
|
|
|
|
return !!(dev->sch->curr_status.pmcw.flags & PMCW_FLAGS_MASK_ENA);
|
|
}
|
|
|
|
static int virtio_ccw_get_mappings(VirtioCcwDevice *dev)
|
|
{
|
|
int r;
|
|
CcwDevice *ccw_dev = CCW_DEVICE(dev);
|
|
|
|
if (!ccw_dev->sch->thinint_active) {
|
|
return -EINVAL;
|
|
}
|
|
|
|
r = map_indicator(&dev->routes.adapter, dev->summary_indicator);
|
|
if (r) {
|
|
return r;
|
|
}
|
|
r = map_indicator(&dev->routes.adapter, dev->indicators);
|
|
if (r) {
|
|
return r;
|
|
}
|
|
dev->routes.adapter.summary_addr = dev->summary_indicator->map;
|
|
dev->routes.adapter.ind_addr = dev->indicators->map;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int virtio_ccw_setup_irqroutes(VirtioCcwDevice *dev, int nvqs)
|
|
{
|
|
int i;
|
|
VirtIODevice *vdev = virtio_bus_get_device(&dev->bus);
|
|
int ret;
|
|
S390FLICState *fs = s390_get_flic();
|
|
S390FLICStateClass *fsc = s390_get_flic_class(fs);
|
|
|
|
ret = virtio_ccw_get_mappings(dev);
|
|
if (ret) {
|
|
return ret;
|
|
}
|
|
for (i = 0; i < nvqs; i++) {
|
|
if (!virtio_queue_get_num(vdev, i)) {
|
|
break;
|
|
}
|
|
}
|
|
dev->routes.num_routes = i;
|
|
return fsc->add_adapter_routes(fs, &dev->routes);
|
|
}
|
|
|
|
static void virtio_ccw_release_irqroutes(VirtioCcwDevice *dev, int nvqs)
|
|
{
|
|
S390FLICState *fs = s390_get_flic();
|
|
S390FLICStateClass *fsc = s390_get_flic_class(fs);
|
|
|
|
fsc->release_adapter_routes(fs, &dev->routes);
|
|
}
|
|
|
|
static int virtio_ccw_add_irqfd(VirtioCcwDevice *dev, int n)
|
|
{
|
|
VirtIODevice *vdev = virtio_bus_get_device(&dev->bus);
|
|
VirtQueue *vq = virtio_get_queue(vdev, n);
|
|
EventNotifier *notifier = virtio_queue_get_guest_notifier(vq);
|
|
|
|
return kvm_irqchip_add_irqfd_notifier_gsi(kvm_state, notifier, NULL,
|
|
dev->routes.gsi[n]);
|
|
}
|
|
|
|
static void virtio_ccw_remove_irqfd(VirtioCcwDevice *dev, int n)
|
|
{
|
|
VirtIODevice *vdev = virtio_bus_get_device(&dev->bus);
|
|
VirtQueue *vq = virtio_get_queue(vdev, n);
|
|
EventNotifier *notifier = virtio_queue_get_guest_notifier(vq);
|
|
int ret;
|
|
|
|
ret = kvm_irqchip_remove_irqfd_notifier_gsi(kvm_state, notifier,
|
|
dev->routes.gsi[n]);
|
|
assert(ret == 0);
|
|
}
|
|
|
|
static int virtio_ccw_set_guest_notifier(VirtioCcwDevice *dev, int n,
|
|
bool assign, bool with_irqfd)
|
|
{
|
|
VirtIODevice *vdev = virtio_bus_get_device(&dev->bus);
|
|
VirtQueue *vq = virtio_get_queue(vdev, n);
|
|
EventNotifier *notifier = virtio_queue_get_guest_notifier(vq);
|
|
VirtioDeviceClass *k = VIRTIO_DEVICE_GET_CLASS(vdev);
|
|
|
|
if (assign) {
|
|
int r = event_notifier_init(notifier, 0);
|
|
|
|
if (r < 0) {
|
|
return r;
|
|
}
|
|
virtio_queue_set_guest_notifier_fd_handler(vq, true, with_irqfd);
|
|
if (with_irqfd) {
|
|
r = virtio_ccw_add_irqfd(dev, n);
|
|
if (r) {
|
|
virtio_queue_set_guest_notifier_fd_handler(vq, false,
|
|
with_irqfd);
|
|
return r;
|
|
}
|
|
}
|
|
/*
|
|
* We do not support individual masking for channel devices, so we
|
|
* need to manually trigger any guest masking callbacks here.
|
|
*/
|
|
if (k->guest_notifier_mask && vdev->use_guest_notifier_mask) {
|
|
k->guest_notifier_mask(vdev, n, false);
|
|
}
|
|
/* get lost events and re-inject */
|
|
if (k->guest_notifier_pending &&
|
|
k->guest_notifier_pending(vdev, n)) {
|
|
event_notifier_set(notifier);
|
|
}
|
|
} else {
|
|
if (k->guest_notifier_mask && vdev->use_guest_notifier_mask) {
|
|
k->guest_notifier_mask(vdev, n, true);
|
|
}
|
|
if (with_irqfd) {
|
|
virtio_ccw_remove_irqfd(dev, n);
|
|
}
|
|
virtio_queue_set_guest_notifier_fd_handler(vq, false, with_irqfd);
|
|
event_notifier_cleanup(notifier);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int virtio_ccw_set_guest_notifiers(DeviceState *d, int nvqs,
|
|
bool assigned)
|
|
{
|
|
VirtioCcwDevice *dev = VIRTIO_CCW_DEVICE(d);
|
|
VirtIODevice *vdev = virtio_bus_get_device(&dev->bus);
|
|
CcwDevice *ccw_dev = CCW_DEVICE(d);
|
|
bool with_irqfd = ccw_dev->sch->thinint_active && kvm_irqfds_enabled();
|
|
int r, n;
|
|
|
|
if (with_irqfd && assigned) {
|
|
/* irq routes need to be set up before assigning irqfds */
|
|
r = virtio_ccw_setup_irqroutes(dev, nvqs);
|
|
if (r < 0) {
|
|
goto irqroute_error;
|
|
}
|
|
}
|
|
for (n = 0; n < nvqs; n++) {
|
|
if (!virtio_queue_get_num(vdev, n)) {
|
|
break;
|
|
}
|
|
r = virtio_ccw_set_guest_notifier(dev, n, assigned, with_irqfd);
|
|
if (r < 0) {
|
|
goto assign_error;
|
|
}
|
|
}
|
|
if (with_irqfd && !assigned) {
|
|
/* release irq routes after irqfds have been released */
|
|
virtio_ccw_release_irqroutes(dev, nvqs);
|
|
}
|
|
return 0;
|
|
|
|
assign_error:
|
|
while (--n >= 0) {
|
|
virtio_ccw_set_guest_notifier(dev, n, !assigned, false);
|
|
}
|
|
irqroute_error:
|
|
if (with_irqfd && assigned) {
|
|
virtio_ccw_release_irqroutes(dev, nvqs);
|
|
}
|
|
return r;
|
|
}
|
|
|
|
static void virtio_ccw_save_queue(DeviceState *d, int n, QEMUFile *f)
|
|
{
|
|
VirtioCcwDevice *dev = VIRTIO_CCW_DEVICE(d);
|
|
VirtIODevice *vdev = virtio_bus_get_device(&dev->bus);
|
|
|
|
qemu_put_be16(f, virtio_queue_vector(vdev, n));
|
|
}
|
|
|
|
static int virtio_ccw_load_queue(DeviceState *d, int n, QEMUFile *f)
|
|
{
|
|
VirtioCcwDevice *dev = VIRTIO_CCW_DEVICE(d);
|
|
VirtIODevice *vdev = virtio_bus_get_device(&dev->bus);
|
|
uint16_t vector;
|
|
|
|
qemu_get_be16s(f, &vector);
|
|
virtio_queue_set_vector(vdev, n , vector);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void virtio_ccw_save_config(DeviceState *d, QEMUFile *f)
|
|
{
|
|
VirtioCcwDevice *dev = VIRTIO_CCW_DEVICE(d);
|
|
vmstate_save_state(f, &vmstate_virtio_ccw_dev, dev, NULL);
|
|
}
|
|
|
|
static int virtio_ccw_load_config(DeviceState *d, QEMUFile *f)
|
|
{
|
|
VirtioCcwDevice *dev = VIRTIO_CCW_DEVICE(d);
|
|
return vmstate_load_state(f, &vmstate_virtio_ccw_dev, dev, 1);
|
|
}
|
|
|
|
static void virtio_ccw_pre_plugged(DeviceState *d, Error **errp)
|
|
{
|
|
VirtioCcwDevice *dev = VIRTIO_CCW_DEVICE(d);
|
|
VirtIODevice *vdev = virtio_bus_get_device(&dev->bus);
|
|
|
|
if (dev->max_rev >= 1) {
|
|
virtio_add_feature(&vdev->host_features, VIRTIO_F_VERSION_1);
|
|
}
|
|
}
|
|
|
|
/* This is called by virtio-bus just after the device is plugged. */
|
|
static void virtio_ccw_device_plugged(DeviceState *d, Error **errp)
|
|
{
|
|
VirtioCcwDevice *dev = VIRTIO_CCW_DEVICE(d);
|
|
VirtIODevice *vdev = virtio_bus_get_device(&dev->bus);
|
|
CcwDevice *ccw_dev = CCW_DEVICE(d);
|
|
SubchDev *sch = ccw_dev->sch;
|
|
int n = virtio_get_num_queues(vdev);
|
|
S390FLICState *flic = s390_get_flic();
|
|
|
|
if (!virtio_has_feature(vdev->host_features, VIRTIO_F_VERSION_1)) {
|
|
dev->max_rev = 0;
|
|
}
|
|
|
|
if (virtio_get_num_queues(vdev) > VIRTIO_QUEUE_MAX) {
|
|
error_setg(errp, "The number of virtqueues %d "
|
|
"exceeds virtio limit %d", n,
|
|
VIRTIO_QUEUE_MAX);
|
|
return;
|
|
}
|
|
if (virtio_get_num_queues(vdev) > flic->adapter_routes_max_batch) {
|
|
error_setg(errp, "The number of virtqueues %d "
|
|
"exceeds flic adapter route limit %d", n,
|
|
flic->adapter_routes_max_batch);
|
|
return;
|
|
}
|
|
|
|
sch->id.cu_model = virtio_bus_get_vdev_id(&dev->bus);
|
|
|
|
|
|
css_generate_sch_crws(sch->cssid, sch->ssid, sch->schid,
|
|
d->hotplugged, 1);
|
|
}
|
|
|
|
static void virtio_ccw_device_unplugged(DeviceState *d)
|
|
{
|
|
VirtioCcwDevice *dev = VIRTIO_CCW_DEVICE(d);
|
|
|
|
virtio_ccw_stop_ioeventfd(dev);
|
|
}
|
|
/**************** Virtio-ccw Bus Device Descriptions *******************/
|
|
|
|
static void virtio_ccw_busdev_realize(DeviceState *dev, Error **errp)
|
|
{
|
|
VirtioCcwDevice *_dev = (VirtioCcwDevice *)dev;
|
|
|
|
virtio_ccw_bus_new(&_dev->bus, sizeof(_dev->bus), _dev);
|
|
virtio_ccw_device_realize(_dev, errp);
|
|
}
|
|
|
|
static void virtio_ccw_busdev_unrealize(DeviceState *dev, Error **errp)
|
|
{
|
|
VirtioCcwDevice *_dev = (VirtioCcwDevice *)dev;
|
|
|
|
virtio_ccw_device_unrealize(_dev, errp);
|
|
}
|
|
|
|
static void virtio_ccw_busdev_unplug(HotplugHandler *hotplug_dev,
|
|
DeviceState *dev, Error **errp)
|
|
{
|
|
VirtioCcwDevice *_dev = to_virtio_ccw_dev_fast(dev);
|
|
|
|
virtio_ccw_stop_ioeventfd(_dev);
|
|
}
|
|
|
|
static void virtio_ccw_device_class_init(ObjectClass *klass, void *data)
|
|
{
|
|
DeviceClass *dc = DEVICE_CLASS(klass);
|
|
CCWDeviceClass *k = CCW_DEVICE_CLASS(dc);
|
|
VirtIOCCWDeviceClass *vdc = VIRTIO_CCW_DEVICE_CLASS(klass);
|
|
|
|
k->unplug = virtio_ccw_busdev_unplug;
|
|
dc->realize = virtio_ccw_busdev_realize;
|
|
dc->unrealize = virtio_ccw_busdev_unrealize;
|
|
dc->bus_type = TYPE_VIRTUAL_CSS_BUS;
|
|
device_class_set_parent_reset(dc, virtio_ccw_reset, &vdc->parent_reset);
|
|
}
|
|
|
|
static const TypeInfo virtio_ccw_device_info = {
|
|
.name = TYPE_VIRTIO_CCW_DEVICE,
|
|
.parent = TYPE_CCW_DEVICE,
|
|
.instance_size = sizeof(VirtioCcwDevice),
|
|
.class_init = virtio_ccw_device_class_init,
|
|
.class_size = sizeof(VirtIOCCWDeviceClass),
|
|
.abstract = true,
|
|
};
|
|
|
|
/* virtio-ccw-bus */
|
|
|
|
static void virtio_ccw_bus_new(VirtioBusState *bus, size_t bus_size,
|
|
VirtioCcwDevice *dev)
|
|
{
|
|
DeviceState *qdev = DEVICE(dev);
|
|
char virtio_bus_name[] = "virtio-bus";
|
|
|
|
qbus_create_inplace(bus, bus_size, TYPE_VIRTIO_CCW_BUS,
|
|
qdev, virtio_bus_name);
|
|
}
|
|
|
|
static void virtio_ccw_bus_class_init(ObjectClass *klass, void *data)
|
|
{
|
|
VirtioBusClass *k = VIRTIO_BUS_CLASS(klass);
|
|
BusClass *bus_class = BUS_CLASS(klass);
|
|
|
|
bus_class->max_dev = 1;
|
|
k->notify = virtio_ccw_notify;
|
|
k->vmstate_change = virtio_ccw_vmstate_change;
|
|
k->query_guest_notifiers = virtio_ccw_query_guest_notifiers;
|
|
k->set_guest_notifiers = virtio_ccw_set_guest_notifiers;
|
|
k->save_queue = virtio_ccw_save_queue;
|
|
k->load_queue = virtio_ccw_load_queue;
|
|
k->save_config = virtio_ccw_save_config;
|
|
k->load_config = virtio_ccw_load_config;
|
|
k->pre_plugged = virtio_ccw_pre_plugged;
|
|
k->device_plugged = virtio_ccw_device_plugged;
|
|
k->device_unplugged = virtio_ccw_device_unplugged;
|
|
k->ioeventfd_enabled = virtio_ccw_ioeventfd_enabled;
|
|
k->ioeventfd_assign = virtio_ccw_ioeventfd_assign;
|
|
}
|
|
|
|
static const TypeInfo virtio_ccw_bus_info = {
|
|
.name = TYPE_VIRTIO_CCW_BUS,
|
|
.parent = TYPE_VIRTIO_BUS,
|
|
.instance_size = sizeof(VirtioCcwBusState),
|
|
.class_init = virtio_ccw_bus_class_init,
|
|
};
|
|
|
|
static void virtio_ccw_register(void)
|
|
{
|
|
type_register_static(&virtio_ccw_bus_info);
|
|
type_register_static(&virtio_ccw_device_info);
|
|
}
|
|
|
|
type_init(virtio_ccw_register)
|