ba28e0ff4d
Replace qemu_bh_new_guarded() by virtio_bh_new_guarded() so the bus and device use the same guard. Otherwise the DMA-reentrancy protection can be bypassed: $ cat << EOF | qemu-system-i386 -display none -nodefaults \ -machine q35,accel=qtest \ -m 512M \ -device virtio-gpu \ -qtest stdio outl 0xcf8 0x80000820 outl 0xcfc 0xe0004000 outl 0xcf8 0x80000804 outw 0xcfc 0x06 write 0xe0004030 0x4 0x024000e0 write 0xe0004028 0x1 0xff write 0xe0004020 0x4 0x00009300 write 0xe000401c 0x1 0x01 write 0x101 0x1 0x04 write 0x103 0x1 0x1c write 0x9301c8 0x1 0x18 write 0x105 0x1 0x1c write 0x107 0x1 0x1c write 0x109 0x1 0x1c write 0x10b 0x1 0x00 write 0x10d 0x1 0x00 write 0x10f 0x1 0x00 write 0x111 0x1 0x00 write 0x113 0x1 0x00 write 0x115 0x1 0x00 write 0x117 0x1 0x00 write 0x119 0x1 0x00 write 0x11b 0x1 0x00 write 0x11d 0x1 0x00 write 0x11f 0x1 0x00 write 0x121 0x1 0x00 write 0x123 0x1 0x00 write 0x125 0x1 0x00 write 0x127 0x1 0x00 write 0x129 0x1 0x00 write 0x12b 0x1 0x00 write 0x12d 0x1 0x00 write 0x12f 0x1 0x00 write 0x131 0x1 0x00 write 0x133 0x1 0x00 write 0x135 0x1 0x00 write 0x137 0x1 0x00 write 0x139 0x1 0x00 write 0xe0007003 0x1 0x00 EOF ... ================================================================= ==276099==ERROR: AddressSanitizer: heap-use-after-free on address 0x60d000011178 at pc 0x562cc3b736c7 bp 0x7ffed49dee60 sp 0x7ffed49dee58 READ of size 8 at 0x60d000011178 thread T0 #0 0x562cc3b736c6 in virtio_gpu_ctrl_response hw/display/virtio-gpu.c:180:42 #1 0x562cc3b7c40b in virtio_gpu_ctrl_response_nodata hw/display/virtio-gpu.c:192:5 #2 0x562cc3b7c40b in virtio_gpu_simple_process_cmd hw/display/virtio-gpu.c:1015:13 #3 0x562cc3b82873 in virtio_gpu_process_cmdq hw/display/virtio-gpu.c:1050:9 #4 0x562cc4a85514 in aio_bh_call util/async.c:169:5 #5 0x562cc4a85c52 in aio_bh_poll util/async.c:216:13 #6 0x562cc4a1a79b in aio_dispatch util/aio-posix.c:423:5 #7 0x562cc4a8a2da in aio_ctx_dispatch util/async.c:358:5 #8 0x7f36840547a8 in g_main_context_dispatch (/lib/x86_64-linux-gnu/libglib-2.0.so.0+0x547a8) #9 0x562cc4a8b753 in glib_pollfds_poll util/main-loop.c:290:9 #10 0x562cc4a8b753 in os_host_main_loop_wait util/main-loop.c:313:5 #11 0x562cc4a8b753 in main_loop_wait util/main-loop.c:592:11 #12 0x562cc3938186 in qemu_main_loop system/runstate.c:782:9 #13 0x562cc43b7af5 in qemu_default_main system/main.c:37:14 #14 0x7f3683a6c189 in __libc_start_call_main csu/../sysdeps/nptl/libc_start_call_main.h:58:16 #15 0x7f3683a6c244 in __libc_start_main csu/../csu/libc-start.c:381:3 #16 0x562cc2a58ac0 in _start (qemu-system-i386+0x231bac0) 0x60d000011178 is located 56 bytes inside of 136-byte region [0x60d000011140,0x60d0000111c8) freed by thread T0 here: #0 0x562cc2adb662 in __interceptor_free (qemu-system-i386+0x239e662) #1 0x562cc3b86b21 in virtio_gpu_reset hw/display/virtio-gpu.c:1524:9 #2 0x562cc416e20e in virtio_reset hw/virtio/virtio.c:2145:9 #3 0x562cc37c5644 in virtio_pci_reset hw/virtio/virtio-pci.c:2249:5 #4 0x562cc4233758 in memory_region_write_accessor system/memory.c:497:5 #5 0x562cc4232eea in access_with_adjusted_size system/memory.c:573:18 previously allocated by thread T0 here: #0 0x562cc2adb90e in malloc (qemu-system-i386+0x239e90e) #1 0x7f368405a678 in g_malloc (/lib/x86_64-linux-gnu/libglib-2.0.so.0+0x5a678) #2 0x562cc4163ffc in virtqueue_split_pop hw/virtio/virtio.c:1612:12 #3 0x562cc4163ffc in virtqueue_pop hw/virtio/virtio.c:1783:16 #4 0x562cc3b91a95 in virtio_gpu_handle_ctrl hw/display/virtio-gpu.c:1112:15 #5 0x562cc4a85514 in aio_bh_call util/async.c:169:5 #6 0x562cc4a85c52 in aio_bh_poll util/async.c:216:13 #7 0x562cc4a1a79b in aio_dispatch util/aio-posix.c:423:5 SUMMARY: AddressSanitizer: heap-use-after-free hw/display/virtio-gpu.c:180:42 in virtio_gpu_ctrl_response With this change, the same reproducer triggers: qemu-system-i386: warning: Blocked re-entrant IO on MemoryRegion: virtio-pci-common-virtio-gpu at addr: 0x6 Fixes: CVE-2024-3446 Cc: qemu-stable@nongnu.org Reported-by: Alexander Bulekov <alxndr@bu.edu> Reported-by: Yongkang Jia <kangel@zju.edu.cn> Reported-by: Xiao Lei <nop.leixiao@gmail.com> Reported-by: Yiming Tao <taoym@zju.edu.cn> Buglink: https://bugs.launchpad.net/qemu/+bug/1888606 Reviewed-by: Gerd Hoffmann <kraxel@redhat.com> Acked-by: Michael S. Tsirkin <mst@redhat.com> Signed-off-by: Philippe Mathieu-Daudé <philmd@linaro.org> Reviewed-by: Michael S. Tsirkin <mst@redhat.com> Message-Id: <20240409105537.18308-3-philmd@linaro.org>
1704 lines
53 KiB
C
1704 lines
53 KiB
C
/*
|
|
* Virtio GPU Device
|
|
*
|
|
* Copyright Red Hat, Inc. 2013-2014
|
|
*
|
|
* Authors:
|
|
* Dave Airlie <airlied@redhat.com>
|
|
* Gerd Hoffmann <kraxel@redhat.com>
|
|
*
|
|
* This work is licensed under the terms of the GNU GPL, version 2 or later.
|
|
* See the COPYING file in the top-level directory.
|
|
*/
|
|
|
|
#include "qemu/osdep.h"
|
|
#include "qemu/units.h"
|
|
#include "qemu/iov.h"
|
|
#include "sysemu/cpus.h"
|
|
#include "ui/console.h"
|
|
#include "ui/rect.h"
|
|
#include "trace.h"
|
|
#include "sysemu/dma.h"
|
|
#include "sysemu/sysemu.h"
|
|
#include "hw/virtio/virtio.h"
|
|
#include "migration/qemu-file-types.h"
|
|
#include "hw/virtio/virtio-gpu.h"
|
|
#include "hw/virtio/virtio-gpu-bswap.h"
|
|
#include "hw/virtio/virtio-gpu-pixman.h"
|
|
#include "hw/virtio/virtio-bus.h"
|
|
#include "hw/qdev-properties.h"
|
|
#include "qemu/log.h"
|
|
#include "qemu/module.h"
|
|
#include "qapi/error.h"
|
|
#include "qemu/error-report.h"
|
|
|
|
#define VIRTIO_GPU_VM_VERSION 1
|
|
|
|
static struct virtio_gpu_simple_resource *
|
|
virtio_gpu_find_check_resource(VirtIOGPU *g, uint32_t resource_id,
|
|
bool require_backing,
|
|
const char *caller, uint32_t *error);
|
|
|
|
static void virtio_gpu_reset_bh(void *opaque);
|
|
|
|
void virtio_gpu_update_cursor_data(VirtIOGPU *g,
|
|
struct virtio_gpu_scanout *s,
|
|
uint32_t resource_id)
|
|
{
|
|
struct virtio_gpu_simple_resource *res;
|
|
uint32_t pixels;
|
|
void *data;
|
|
|
|
res = virtio_gpu_find_check_resource(g, resource_id, false,
|
|
__func__, NULL);
|
|
if (!res) {
|
|
return;
|
|
}
|
|
|
|
if (res->blob_size) {
|
|
if (res->blob_size < (s->current_cursor->width *
|
|
s->current_cursor->height * 4)) {
|
|
return;
|
|
}
|
|
data = res->blob;
|
|
} else {
|
|
if (pixman_image_get_width(res->image) != s->current_cursor->width ||
|
|
pixman_image_get_height(res->image) != s->current_cursor->height) {
|
|
return;
|
|
}
|
|
data = pixman_image_get_data(res->image);
|
|
}
|
|
|
|
pixels = s->current_cursor->width * s->current_cursor->height;
|
|
memcpy(s->current_cursor->data, data,
|
|
pixels * sizeof(uint32_t));
|
|
}
|
|
|
|
static void update_cursor(VirtIOGPU *g, struct virtio_gpu_update_cursor *cursor)
|
|
{
|
|
struct virtio_gpu_scanout *s;
|
|
VirtIOGPUClass *vgc = VIRTIO_GPU_GET_CLASS(g);
|
|
bool move = cursor->hdr.type == VIRTIO_GPU_CMD_MOVE_CURSOR;
|
|
|
|
if (cursor->pos.scanout_id >= g->parent_obj.conf.max_outputs) {
|
|
return;
|
|
}
|
|
s = &g->parent_obj.scanout[cursor->pos.scanout_id];
|
|
|
|
trace_virtio_gpu_update_cursor(cursor->pos.scanout_id,
|
|
cursor->pos.x,
|
|
cursor->pos.y,
|
|
move ? "move" : "update",
|
|
cursor->resource_id);
|
|
|
|
if (!move) {
|
|
if (!s->current_cursor) {
|
|
s->current_cursor = cursor_alloc(64, 64);
|
|
}
|
|
|
|
s->current_cursor->hot_x = cursor->hot_x;
|
|
s->current_cursor->hot_y = cursor->hot_y;
|
|
|
|
if (cursor->resource_id > 0) {
|
|
vgc->update_cursor_data(g, s, cursor->resource_id);
|
|
}
|
|
dpy_cursor_define(s->con, s->current_cursor);
|
|
|
|
s->cursor = *cursor;
|
|
} else {
|
|
s->cursor.pos.x = cursor->pos.x;
|
|
s->cursor.pos.y = cursor->pos.y;
|
|
}
|
|
dpy_mouse_set(s->con, cursor->pos.x, cursor->pos.y,
|
|
cursor->resource_id ? 1 : 0);
|
|
}
|
|
|
|
struct virtio_gpu_simple_resource *
|
|
virtio_gpu_find_resource(VirtIOGPU *g, uint32_t resource_id)
|
|
{
|
|
struct virtio_gpu_simple_resource *res;
|
|
|
|
QTAILQ_FOREACH(res, &g->reslist, next) {
|
|
if (res->resource_id == resource_id) {
|
|
return res;
|
|
}
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
static struct virtio_gpu_simple_resource *
|
|
virtio_gpu_find_check_resource(VirtIOGPU *g, uint32_t resource_id,
|
|
bool require_backing,
|
|
const char *caller, uint32_t *error)
|
|
{
|
|
struct virtio_gpu_simple_resource *res;
|
|
|
|
res = virtio_gpu_find_resource(g, resource_id);
|
|
if (!res) {
|
|
qemu_log_mask(LOG_GUEST_ERROR, "%s: invalid resource specified %d\n",
|
|
caller, resource_id);
|
|
if (error) {
|
|
*error = VIRTIO_GPU_RESP_ERR_INVALID_RESOURCE_ID;
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
if (require_backing) {
|
|
if (!res->iov || (!res->image && !res->blob)) {
|
|
qemu_log_mask(LOG_GUEST_ERROR, "%s: no backing storage %d\n",
|
|
caller, resource_id);
|
|
if (error) {
|
|
*error = VIRTIO_GPU_RESP_ERR_UNSPEC;
|
|
}
|
|
return NULL;
|
|
}
|
|
}
|
|
|
|
return res;
|
|
}
|
|
|
|
void virtio_gpu_ctrl_response(VirtIOGPU *g,
|
|
struct virtio_gpu_ctrl_command *cmd,
|
|
struct virtio_gpu_ctrl_hdr *resp,
|
|
size_t resp_len)
|
|
{
|
|
size_t s;
|
|
|
|
if (cmd->cmd_hdr.flags & VIRTIO_GPU_FLAG_FENCE) {
|
|
resp->flags |= VIRTIO_GPU_FLAG_FENCE;
|
|
resp->fence_id = cmd->cmd_hdr.fence_id;
|
|
resp->ctx_id = cmd->cmd_hdr.ctx_id;
|
|
}
|
|
virtio_gpu_ctrl_hdr_bswap(resp);
|
|
s = iov_from_buf(cmd->elem.in_sg, cmd->elem.in_num, 0, resp, resp_len);
|
|
if (s != resp_len) {
|
|
qemu_log_mask(LOG_GUEST_ERROR,
|
|
"%s: response size incorrect %zu vs %zu\n",
|
|
__func__, s, resp_len);
|
|
}
|
|
virtqueue_push(cmd->vq, &cmd->elem, s);
|
|
virtio_notify(VIRTIO_DEVICE(g), cmd->vq);
|
|
cmd->finished = true;
|
|
}
|
|
|
|
void virtio_gpu_ctrl_response_nodata(VirtIOGPU *g,
|
|
struct virtio_gpu_ctrl_command *cmd,
|
|
enum virtio_gpu_ctrl_type type)
|
|
{
|
|
struct virtio_gpu_ctrl_hdr resp;
|
|
|
|
memset(&resp, 0, sizeof(resp));
|
|
resp.type = type;
|
|
virtio_gpu_ctrl_response(g, cmd, &resp, sizeof(resp));
|
|
}
|
|
|
|
void virtio_gpu_get_display_info(VirtIOGPU *g,
|
|
struct virtio_gpu_ctrl_command *cmd)
|
|
{
|
|
struct virtio_gpu_resp_display_info display_info;
|
|
|
|
trace_virtio_gpu_cmd_get_display_info();
|
|
memset(&display_info, 0, sizeof(display_info));
|
|
display_info.hdr.type = VIRTIO_GPU_RESP_OK_DISPLAY_INFO;
|
|
virtio_gpu_base_fill_display_info(VIRTIO_GPU_BASE(g), &display_info);
|
|
virtio_gpu_ctrl_response(g, cmd, &display_info.hdr,
|
|
sizeof(display_info));
|
|
}
|
|
|
|
void virtio_gpu_get_edid(VirtIOGPU *g,
|
|
struct virtio_gpu_ctrl_command *cmd)
|
|
{
|
|
struct virtio_gpu_resp_edid edid;
|
|
struct virtio_gpu_cmd_get_edid get_edid;
|
|
VirtIOGPUBase *b = VIRTIO_GPU_BASE(g);
|
|
|
|
VIRTIO_GPU_FILL_CMD(get_edid);
|
|
virtio_gpu_bswap_32(&get_edid, sizeof(get_edid));
|
|
|
|
if (get_edid.scanout >= b->conf.max_outputs) {
|
|
cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_PARAMETER;
|
|
return;
|
|
}
|
|
|
|
trace_virtio_gpu_cmd_get_edid(get_edid.scanout);
|
|
memset(&edid, 0, sizeof(edid));
|
|
edid.hdr.type = VIRTIO_GPU_RESP_OK_EDID;
|
|
virtio_gpu_base_generate_edid(VIRTIO_GPU_BASE(g), get_edid.scanout, &edid);
|
|
virtio_gpu_ctrl_response(g, cmd, &edid.hdr, sizeof(edid));
|
|
}
|
|
|
|
static uint32_t calc_image_hostmem(pixman_format_code_t pformat,
|
|
uint32_t width, uint32_t height)
|
|
{
|
|
/* Copied from pixman/pixman-bits-image.c, skip integer overflow check.
|
|
* pixman_image_create_bits will fail in case it overflow.
|
|
*/
|
|
|
|
int bpp = PIXMAN_FORMAT_BPP(pformat);
|
|
int stride = ((width * bpp + 0x1f) >> 5) * sizeof(uint32_t);
|
|
return height * stride;
|
|
}
|
|
|
|
#ifdef WIN32
|
|
static void
|
|
win32_pixman_image_destroy(pixman_image_t *image, void *data)
|
|
{
|
|
HANDLE handle = data;
|
|
|
|
qemu_win32_map_free(pixman_image_get_data(image), handle, &error_warn);
|
|
}
|
|
#endif
|
|
|
|
static void virtio_gpu_resource_create_2d(VirtIOGPU *g,
|
|
struct virtio_gpu_ctrl_command *cmd)
|
|
{
|
|
pixman_format_code_t pformat;
|
|
struct virtio_gpu_simple_resource *res;
|
|
struct virtio_gpu_resource_create_2d c2d;
|
|
|
|
VIRTIO_GPU_FILL_CMD(c2d);
|
|
virtio_gpu_bswap_32(&c2d, sizeof(c2d));
|
|
trace_virtio_gpu_cmd_res_create_2d(c2d.resource_id, c2d.format,
|
|
c2d.width, c2d.height);
|
|
|
|
if (c2d.resource_id == 0) {
|
|
qemu_log_mask(LOG_GUEST_ERROR, "%s: resource id 0 is not allowed\n",
|
|
__func__);
|
|
cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_RESOURCE_ID;
|
|
return;
|
|
}
|
|
|
|
res = virtio_gpu_find_resource(g, c2d.resource_id);
|
|
if (res) {
|
|
qemu_log_mask(LOG_GUEST_ERROR, "%s: resource already exists %d\n",
|
|
__func__, c2d.resource_id);
|
|
cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_RESOURCE_ID;
|
|
return;
|
|
}
|
|
|
|
res = g_new0(struct virtio_gpu_simple_resource, 1);
|
|
|
|
res->width = c2d.width;
|
|
res->height = c2d.height;
|
|
res->format = c2d.format;
|
|
res->resource_id = c2d.resource_id;
|
|
|
|
pformat = virtio_gpu_get_pixman_format(c2d.format);
|
|
if (!pformat) {
|
|
qemu_log_mask(LOG_GUEST_ERROR,
|
|
"%s: host couldn't handle guest format %d\n",
|
|
__func__, c2d.format);
|
|
g_free(res);
|
|
cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_PARAMETER;
|
|
return;
|
|
}
|
|
|
|
res->hostmem = calc_image_hostmem(pformat, c2d.width, c2d.height);
|
|
if (res->hostmem + g->hostmem < g->conf_max_hostmem) {
|
|
void *bits = NULL;
|
|
#ifdef WIN32
|
|
bits = qemu_win32_map_alloc(res->hostmem, &res->handle, &error_warn);
|
|
if (!bits) {
|
|
goto end;
|
|
}
|
|
#endif
|
|
res->image = pixman_image_create_bits(
|
|
pformat,
|
|
c2d.width,
|
|
c2d.height,
|
|
bits, c2d.height ? res->hostmem / c2d.height : 0);
|
|
#ifdef WIN32
|
|
if (res->image) {
|
|
pixman_image_set_destroy_function(res->image, win32_pixman_image_destroy, res->handle);
|
|
}
|
|
#endif
|
|
}
|
|
|
|
#ifdef WIN32
|
|
end:
|
|
#endif
|
|
if (!res->image) {
|
|
qemu_log_mask(LOG_GUEST_ERROR,
|
|
"%s: resource creation failed %d %d %d\n",
|
|
__func__, c2d.resource_id, c2d.width, c2d.height);
|
|
g_free(res);
|
|
cmd->error = VIRTIO_GPU_RESP_ERR_OUT_OF_MEMORY;
|
|
return;
|
|
}
|
|
|
|
QTAILQ_INSERT_HEAD(&g->reslist, res, next);
|
|
g->hostmem += res->hostmem;
|
|
}
|
|
|
|
static void virtio_gpu_resource_create_blob(VirtIOGPU *g,
|
|
struct virtio_gpu_ctrl_command *cmd)
|
|
{
|
|
struct virtio_gpu_simple_resource *res;
|
|
struct virtio_gpu_resource_create_blob cblob;
|
|
int ret;
|
|
|
|
VIRTIO_GPU_FILL_CMD(cblob);
|
|
virtio_gpu_create_blob_bswap(&cblob);
|
|
trace_virtio_gpu_cmd_res_create_blob(cblob.resource_id, cblob.size);
|
|
|
|
if (cblob.resource_id == 0) {
|
|
qemu_log_mask(LOG_GUEST_ERROR, "%s: resource id 0 is not allowed\n",
|
|
__func__);
|
|
cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_RESOURCE_ID;
|
|
return;
|
|
}
|
|
|
|
if (cblob.blob_mem != VIRTIO_GPU_BLOB_MEM_GUEST &&
|
|
cblob.blob_flags != VIRTIO_GPU_BLOB_FLAG_USE_SHAREABLE) {
|
|
qemu_log_mask(LOG_GUEST_ERROR, "%s: invalid memory type\n",
|
|
__func__);
|
|
cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_PARAMETER;
|
|
return;
|
|
}
|
|
|
|
if (virtio_gpu_find_resource(g, cblob.resource_id)) {
|
|
qemu_log_mask(LOG_GUEST_ERROR, "%s: resource already exists %d\n",
|
|
__func__, cblob.resource_id);
|
|
cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_RESOURCE_ID;
|
|
return;
|
|
}
|
|
|
|
res = g_new0(struct virtio_gpu_simple_resource, 1);
|
|
res->resource_id = cblob.resource_id;
|
|
res->blob_size = cblob.size;
|
|
|
|
ret = virtio_gpu_create_mapping_iov(g, cblob.nr_entries, sizeof(cblob),
|
|
cmd, &res->addrs, &res->iov,
|
|
&res->iov_cnt);
|
|
if (ret != 0) {
|
|
cmd->error = VIRTIO_GPU_RESP_ERR_UNSPEC;
|
|
g_free(res);
|
|
return;
|
|
}
|
|
|
|
virtio_gpu_init_udmabuf(res);
|
|
QTAILQ_INSERT_HEAD(&g->reslist, res, next);
|
|
}
|
|
|
|
static void virtio_gpu_disable_scanout(VirtIOGPU *g, int scanout_id)
|
|
{
|
|
struct virtio_gpu_scanout *scanout = &g->parent_obj.scanout[scanout_id];
|
|
struct virtio_gpu_simple_resource *res;
|
|
|
|
if (scanout->resource_id == 0) {
|
|
return;
|
|
}
|
|
|
|
res = virtio_gpu_find_resource(g, scanout->resource_id);
|
|
if (res) {
|
|
res->scanout_bitmask &= ~(1 << scanout_id);
|
|
}
|
|
|
|
dpy_gfx_replace_surface(scanout->con, NULL);
|
|
scanout->resource_id = 0;
|
|
scanout->ds = NULL;
|
|
scanout->width = 0;
|
|
scanout->height = 0;
|
|
}
|
|
|
|
static void virtio_gpu_resource_destroy(VirtIOGPU *g,
|
|
struct virtio_gpu_simple_resource *res,
|
|
Error **errp)
|
|
{
|
|
int i;
|
|
|
|
if (res->scanout_bitmask) {
|
|
for (i = 0; i < g->parent_obj.conf.max_outputs; i++) {
|
|
if (res->scanout_bitmask & (1 << i)) {
|
|
virtio_gpu_disable_scanout(g, i);
|
|
}
|
|
}
|
|
}
|
|
|
|
qemu_pixman_image_unref(res->image);
|
|
virtio_gpu_cleanup_mapping(g, res);
|
|
QTAILQ_REMOVE(&g->reslist, res, next);
|
|
g->hostmem -= res->hostmem;
|
|
g_free(res);
|
|
}
|
|
|
|
static void virtio_gpu_resource_unref(VirtIOGPU *g,
|
|
struct virtio_gpu_ctrl_command *cmd)
|
|
{
|
|
struct virtio_gpu_simple_resource *res;
|
|
struct virtio_gpu_resource_unref unref;
|
|
|
|
VIRTIO_GPU_FILL_CMD(unref);
|
|
virtio_gpu_bswap_32(&unref, sizeof(unref));
|
|
trace_virtio_gpu_cmd_res_unref(unref.resource_id);
|
|
|
|
res = virtio_gpu_find_resource(g, unref.resource_id);
|
|
if (!res) {
|
|
qemu_log_mask(LOG_GUEST_ERROR, "%s: illegal resource specified %d\n",
|
|
__func__, unref.resource_id);
|
|
cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_RESOURCE_ID;
|
|
return;
|
|
}
|
|
/*
|
|
* virtio_gpu_resource_destroy does not set any errors, so pass a NULL errp
|
|
* to ignore them.
|
|
*/
|
|
virtio_gpu_resource_destroy(g, res, NULL);
|
|
}
|
|
|
|
static void virtio_gpu_transfer_to_host_2d(VirtIOGPU *g,
|
|
struct virtio_gpu_ctrl_command *cmd)
|
|
{
|
|
struct virtio_gpu_simple_resource *res;
|
|
int h, bpp;
|
|
uint32_t src_offset, dst_offset, stride;
|
|
pixman_format_code_t format;
|
|
struct virtio_gpu_transfer_to_host_2d t2d;
|
|
void *img_data;
|
|
|
|
VIRTIO_GPU_FILL_CMD(t2d);
|
|
virtio_gpu_t2d_bswap(&t2d);
|
|
trace_virtio_gpu_cmd_res_xfer_toh_2d(t2d.resource_id);
|
|
|
|
res = virtio_gpu_find_check_resource(g, t2d.resource_id, true,
|
|
__func__, &cmd->error);
|
|
if (!res || res->blob) {
|
|
return;
|
|
}
|
|
|
|
if (t2d.r.x > res->width ||
|
|
t2d.r.y > res->height ||
|
|
t2d.r.width > res->width ||
|
|
t2d.r.height > res->height ||
|
|
t2d.r.x + t2d.r.width > res->width ||
|
|
t2d.r.y + t2d.r.height > res->height) {
|
|
qemu_log_mask(LOG_GUEST_ERROR, "%s: transfer bounds outside resource"
|
|
" bounds for resource %d: %d %d %d %d vs %d %d\n",
|
|
__func__, t2d.resource_id, t2d.r.x, t2d.r.y,
|
|
t2d.r.width, t2d.r.height, res->width, res->height);
|
|
cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_PARAMETER;
|
|
return;
|
|
}
|
|
|
|
format = pixman_image_get_format(res->image);
|
|
bpp = DIV_ROUND_UP(PIXMAN_FORMAT_BPP(format), 8);
|
|
stride = pixman_image_get_stride(res->image);
|
|
img_data = pixman_image_get_data(res->image);
|
|
|
|
if (t2d.r.x || t2d.r.width != pixman_image_get_width(res->image)) {
|
|
for (h = 0; h < t2d.r.height; h++) {
|
|
src_offset = t2d.offset + stride * h;
|
|
dst_offset = (t2d.r.y + h) * stride + (t2d.r.x * bpp);
|
|
|
|
iov_to_buf(res->iov, res->iov_cnt, src_offset,
|
|
(uint8_t *)img_data + dst_offset,
|
|
t2d.r.width * bpp);
|
|
}
|
|
} else {
|
|
src_offset = t2d.offset;
|
|
dst_offset = t2d.r.y * stride + t2d.r.x * bpp;
|
|
iov_to_buf(res->iov, res->iov_cnt, src_offset,
|
|
(uint8_t *)img_data + dst_offset,
|
|
stride * t2d.r.height);
|
|
}
|
|
}
|
|
|
|
static void virtio_gpu_resource_flush(VirtIOGPU *g,
|
|
struct virtio_gpu_ctrl_command *cmd)
|
|
{
|
|
struct virtio_gpu_simple_resource *res;
|
|
struct virtio_gpu_resource_flush rf;
|
|
struct virtio_gpu_scanout *scanout;
|
|
QemuRect flush_rect;
|
|
bool within_bounds = false;
|
|
bool update_submitted = false;
|
|
int i;
|
|
|
|
VIRTIO_GPU_FILL_CMD(rf);
|
|
virtio_gpu_bswap_32(&rf, sizeof(rf));
|
|
trace_virtio_gpu_cmd_res_flush(rf.resource_id,
|
|
rf.r.width, rf.r.height, rf.r.x, rf.r.y);
|
|
|
|
res = virtio_gpu_find_check_resource(g, rf.resource_id, false,
|
|
__func__, &cmd->error);
|
|
if (!res) {
|
|
return;
|
|
}
|
|
|
|
if (res->blob) {
|
|
for (i = 0; i < g->parent_obj.conf.max_outputs; i++) {
|
|
scanout = &g->parent_obj.scanout[i];
|
|
if (scanout->resource_id == res->resource_id &&
|
|
rf.r.x < scanout->x + scanout->width &&
|
|
rf.r.x + rf.r.width >= scanout->x &&
|
|
rf.r.y < scanout->y + scanout->height &&
|
|
rf.r.y + rf.r.height >= scanout->y) {
|
|
within_bounds = true;
|
|
|
|
if (console_has_gl(scanout->con)) {
|
|
dpy_gl_update(scanout->con, 0, 0, scanout->width,
|
|
scanout->height);
|
|
update_submitted = true;
|
|
}
|
|
}
|
|
}
|
|
|
|
if (update_submitted) {
|
|
return;
|
|
}
|
|
if (!within_bounds) {
|
|
qemu_log_mask(LOG_GUEST_ERROR, "%s: flush bounds outside scanouts"
|
|
" bounds for flush %d: %d %d %d %d\n",
|
|
__func__, rf.resource_id, rf.r.x, rf.r.y,
|
|
rf.r.width, rf.r.height);
|
|
cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_PARAMETER;
|
|
return;
|
|
}
|
|
}
|
|
|
|
if (!res->blob &&
|
|
(rf.r.x > res->width ||
|
|
rf.r.y > res->height ||
|
|
rf.r.width > res->width ||
|
|
rf.r.height > res->height ||
|
|
rf.r.x + rf.r.width > res->width ||
|
|
rf.r.y + rf.r.height > res->height)) {
|
|
qemu_log_mask(LOG_GUEST_ERROR, "%s: flush bounds outside resource"
|
|
" bounds for resource %d: %d %d %d %d vs %d %d\n",
|
|
__func__, rf.resource_id, rf.r.x, rf.r.y,
|
|
rf.r.width, rf.r.height, res->width, res->height);
|
|
cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_PARAMETER;
|
|
return;
|
|
}
|
|
|
|
qemu_rect_init(&flush_rect, rf.r.x, rf.r.y, rf.r.width, rf.r.height);
|
|
for (i = 0; i < g->parent_obj.conf.max_outputs; i++) {
|
|
QemuRect rect;
|
|
|
|
if (!(res->scanout_bitmask & (1 << i))) {
|
|
continue;
|
|
}
|
|
scanout = &g->parent_obj.scanout[i];
|
|
|
|
qemu_rect_init(&rect, scanout->x, scanout->y,
|
|
scanout->width, scanout->height);
|
|
|
|
/* work out the area we need to update for each console */
|
|
if (qemu_rect_intersect(&flush_rect, &rect, &rect)) {
|
|
qemu_rect_translate(&rect, -scanout->x, -scanout->y);
|
|
dpy_gfx_update(g->parent_obj.scanout[i].con,
|
|
rect.x, rect.y, rect.width, rect.height);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void virtio_unref_resource(pixman_image_t *image, void *data)
|
|
{
|
|
pixman_image_unref(data);
|
|
}
|
|
|
|
static void virtio_gpu_update_scanout(VirtIOGPU *g,
|
|
uint32_t scanout_id,
|
|
struct virtio_gpu_simple_resource *res,
|
|
struct virtio_gpu_framebuffer *fb,
|
|
struct virtio_gpu_rect *r)
|
|
{
|
|
struct virtio_gpu_simple_resource *ores;
|
|
struct virtio_gpu_scanout *scanout;
|
|
|
|
scanout = &g->parent_obj.scanout[scanout_id];
|
|
ores = virtio_gpu_find_resource(g, scanout->resource_id);
|
|
if (ores) {
|
|
ores->scanout_bitmask &= ~(1 << scanout_id);
|
|
}
|
|
|
|
res->scanout_bitmask |= (1 << scanout_id);
|
|
scanout->resource_id = res->resource_id;
|
|
scanout->x = r->x;
|
|
scanout->y = r->y;
|
|
scanout->width = r->width;
|
|
scanout->height = r->height;
|
|
scanout->fb = *fb;
|
|
}
|
|
|
|
static bool virtio_gpu_do_set_scanout(VirtIOGPU *g,
|
|
uint32_t scanout_id,
|
|
struct virtio_gpu_framebuffer *fb,
|
|
struct virtio_gpu_simple_resource *res,
|
|
struct virtio_gpu_rect *r,
|
|
uint32_t *error)
|
|
{
|
|
struct virtio_gpu_scanout *scanout;
|
|
uint8_t *data;
|
|
|
|
scanout = &g->parent_obj.scanout[scanout_id];
|
|
|
|
if (r->x > fb->width ||
|
|
r->y > fb->height ||
|
|
r->width < 16 ||
|
|
r->height < 16 ||
|
|
r->width > fb->width ||
|
|
r->height > fb->height ||
|
|
r->x + r->width > fb->width ||
|
|
r->y + r->height > fb->height) {
|
|
qemu_log_mask(LOG_GUEST_ERROR, "%s: illegal scanout %d bounds for"
|
|
" resource %d, rect (%d,%d)+%d,%d, fb %d %d\n",
|
|
__func__, scanout_id, res->resource_id,
|
|
r->x, r->y, r->width, r->height,
|
|
fb->width, fb->height);
|
|
*error = VIRTIO_GPU_RESP_ERR_INVALID_PARAMETER;
|
|
return false;
|
|
}
|
|
|
|
g->parent_obj.enable = 1;
|
|
|
|
if (res->blob) {
|
|
if (console_has_gl(scanout->con)) {
|
|
if (!virtio_gpu_update_dmabuf(g, scanout_id, res, fb, r)) {
|
|
virtio_gpu_update_scanout(g, scanout_id, res, fb, r);
|
|
} else {
|
|
*error = VIRTIO_GPU_RESP_ERR_OUT_OF_MEMORY;
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
data = res->blob;
|
|
} else {
|
|
data = (uint8_t *)pixman_image_get_data(res->image);
|
|
}
|
|
|
|
/* create a surface for this scanout */
|
|
if ((res->blob && !console_has_gl(scanout->con)) ||
|
|
!scanout->ds ||
|
|
surface_data(scanout->ds) != data + fb->offset ||
|
|
scanout->width != r->width ||
|
|
scanout->height != r->height) {
|
|
pixman_image_t *rect;
|
|
void *ptr = data + fb->offset;
|
|
rect = pixman_image_create_bits(fb->format, r->width, r->height,
|
|
ptr, fb->stride);
|
|
|
|
if (res->image) {
|
|
pixman_image_ref(res->image);
|
|
pixman_image_set_destroy_function(rect, virtio_unref_resource,
|
|
res->image);
|
|
}
|
|
|
|
/* realloc the surface ptr */
|
|
scanout->ds = qemu_create_displaysurface_pixman(rect);
|
|
#ifdef WIN32
|
|
qemu_displaysurface_win32_set_handle(scanout->ds, res->handle, fb->offset);
|
|
#endif
|
|
|
|
pixman_image_unref(rect);
|
|
dpy_gfx_replace_surface(g->parent_obj.scanout[scanout_id].con,
|
|
scanout->ds);
|
|
}
|
|
|
|
virtio_gpu_update_scanout(g, scanout_id, res, fb, r);
|
|
return true;
|
|
}
|
|
|
|
static void virtio_gpu_set_scanout(VirtIOGPU *g,
|
|
struct virtio_gpu_ctrl_command *cmd)
|
|
{
|
|
struct virtio_gpu_simple_resource *res;
|
|
struct virtio_gpu_framebuffer fb = { 0 };
|
|
struct virtio_gpu_set_scanout ss;
|
|
|
|
VIRTIO_GPU_FILL_CMD(ss);
|
|
virtio_gpu_bswap_32(&ss, sizeof(ss));
|
|
trace_virtio_gpu_cmd_set_scanout(ss.scanout_id, ss.resource_id,
|
|
ss.r.width, ss.r.height, ss.r.x, ss.r.y);
|
|
|
|
if (ss.scanout_id >= g->parent_obj.conf.max_outputs) {
|
|
qemu_log_mask(LOG_GUEST_ERROR, "%s: illegal scanout id specified %d",
|
|
__func__, ss.scanout_id);
|
|
cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_SCANOUT_ID;
|
|
return;
|
|
}
|
|
|
|
if (ss.resource_id == 0) {
|
|
virtio_gpu_disable_scanout(g, ss.scanout_id);
|
|
return;
|
|
}
|
|
|
|
res = virtio_gpu_find_check_resource(g, ss.resource_id, true,
|
|
__func__, &cmd->error);
|
|
if (!res) {
|
|
return;
|
|
}
|
|
|
|
fb.format = pixman_image_get_format(res->image);
|
|
fb.bytes_pp = DIV_ROUND_UP(PIXMAN_FORMAT_BPP(fb.format), 8);
|
|
fb.width = pixman_image_get_width(res->image);
|
|
fb.height = pixman_image_get_height(res->image);
|
|
fb.stride = pixman_image_get_stride(res->image);
|
|
fb.offset = ss.r.x * fb.bytes_pp + ss.r.y * fb.stride;
|
|
|
|
virtio_gpu_do_set_scanout(g, ss.scanout_id,
|
|
&fb, res, &ss.r, &cmd->error);
|
|
}
|
|
|
|
static void virtio_gpu_set_scanout_blob(VirtIOGPU *g,
|
|
struct virtio_gpu_ctrl_command *cmd)
|
|
{
|
|
struct virtio_gpu_simple_resource *res;
|
|
struct virtio_gpu_framebuffer fb = { 0 };
|
|
struct virtio_gpu_set_scanout_blob ss;
|
|
uint64_t fbend;
|
|
|
|
VIRTIO_GPU_FILL_CMD(ss);
|
|
virtio_gpu_scanout_blob_bswap(&ss);
|
|
trace_virtio_gpu_cmd_set_scanout_blob(ss.scanout_id, ss.resource_id,
|
|
ss.r.width, ss.r.height, ss.r.x,
|
|
ss.r.y);
|
|
|
|
if (ss.scanout_id >= g->parent_obj.conf.max_outputs) {
|
|
qemu_log_mask(LOG_GUEST_ERROR, "%s: illegal scanout id specified %d",
|
|
__func__, ss.scanout_id);
|
|
cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_SCANOUT_ID;
|
|
return;
|
|
}
|
|
|
|
if (ss.resource_id == 0) {
|
|
virtio_gpu_disable_scanout(g, ss.scanout_id);
|
|
return;
|
|
}
|
|
|
|
res = virtio_gpu_find_check_resource(g, ss.resource_id, true,
|
|
__func__, &cmd->error);
|
|
if (!res) {
|
|
return;
|
|
}
|
|
|
|
fb.format = virtio_gpu_get_pixman_format(ss.format);
|
|
if (!fb.format) {
|
|
qemu_log_mask(LOG_GUEST_ERROR,
|
|
"%s: host couldn't handle guest format %d\n",
|
|
__func__, ss.format);
|
|
cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_PARAMETER;
|
|
return;
|
|
}
|
|
|
|
fb.bytes_pp = DIV_ROUND_UP(PIXMAN_FORMAT_BPP(fb.format), 8);
|
|
fb.width = ss.width;
|
|
fb.height = ss.height;
|
|
fb.stride = ss.strides[0];
|
|
fb.offset = ss.offsets[0] + ss.r.x * fb.bytes_pp + ss.r.y * fb.stride;
|
|
|
|
fbend = fb.offset;
|
|
fbend += fb.stride * (ss.r.height - 1);
|
|
fbend += fb.bytes_pp * ss.r.width;
|
|
if (fbend > res->blob_size) {
|
|
qemu_log_mask(LOG_GUEST_ERROR,
|
|
"%s: fb end out of range\n",
|
|
__func__);
|
|
cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_PARAMETER;
|
|
return;
|
|
}
|
|
|
|
virtio_gpu_do_set_scanout(g, ss.scanout_id,
|
|
&fb, res, &ss.r, &cmd->error);
|
|
}
|
|
|
|
int virtio_gpu_create_mapping_iov(VirtIOGPU *g,
|
|
uint32_t nr_entries, uint32_t offset,
|
|
struct virtio_gpu_ctrl_command *cmd,
|
|
uint64_t **addr, struct iovec **iov,
|
|
uint32_t *niov)
|
|
{
|
|
struct virtio_gpu_mem_entry *ents;
|
|
size_t esize, s;
|
|
int e, v;
|
|
|
|
if (nr_entries > 16384) {
|
|
qemu_log_mask(LOG_GUEST_ERROR,
|
|
"%s: nr_entries is too big (%d > 16384)\n",
|
|
__func__, nr_entries);
|
|
return -1;
|
|
}
|
|
|
|
esize = sizeof(*ents) * nr_entries;
|
|
ents = g_malloc(esize);
|
|
s = iov_to_buf(cmd->elem.out_sg, cmd->elem.out_num,
|
|
offset, ents, esize);
|
|
if (s != esize) {
|
|
qemu_log_mask(LOG_GUEST_ERROR,
|
|
"%s: command data size incorrect %zu vs %zu\n",
|
|
__func__, s, esize);
|
|
g_free(ents);
|
|
return -1;
|
|
}
|
|
|
|
*iov = NULL;
|
|
if (addr) {
|
|
*addr = NULL;
|
|
}
|
|
for (e = 0, v = 0; e < nr_entries; e++) {
|
|
uint64_t a = le64_to_cpu(ents[e].addr);
|
|
uint32_t l = le32_to_cpu(ents[e].length);
|
|
hwaddr len;
|
|
void *map;
|
|
|
|
do {
|
|
len = l;
|
|
map = dma_memory_map(VIRTIO_DEVICE(g)->dma_as, a, &len,
|
|
DMA_DIRECTION_TO_DEVICE,
|
|
MEMTXATTRS_UNSPECIFIED);
|
|
if (!map) {
|
|
qemu_log_mask(LOG_GUEST_ERROR, "%s: failed to map MMIO memory for"
|
|
" element %d\n", __func__, e);
|
|
virtio_gpu_cleanup_mapping_iov(g, *iov, v);
|
|
g_free(ents);
|
|
*iov = NULL;
|
|
if (addr) {
|
|
g_free(*addr);
|
|
*addr = NULL;
|
|
}
|
|
return -1;
|
|
}
|
|
|
|
if (!(v % 16)) {
|
|
*iov = g_renew(struct iovec, *iov, v + 16);
|
|
if (addr) {
|
|
*addr = g_renew(uint64_t, *addr, v + 16);
|
|
}
|
|
}
|
|
(*iov)[v].iov_base = map;
|
|
(*iov)[v].iov_len = len;
|
|
if (addr) {
|
|
(*addr)[v] = a;
|
|
}
|
|
|
|
a += len;
|
|
l -= len;
|
|
v += 1;
|
|
} while (l > 0);
|
|
}
|
|
*niov = v;
|
|
|
|
g_free(ents);
|
|
return 0;
|
|
}
|
|
|
|
void virtio_gpu_cleanup_mapping_iov(VirtIOGPU *g,
|
|
struct iovec *iov, uint32_t count)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < count; i++) {
|
|
dma_memory_unmap(VIRTIO_DEVICE(g)->dma_as,
|
|
iov[i].iov_base, iov[i].iov_len,
|
|
DMA_DIRECTION_TO_DEVICE,
|
|
iov[i].iov_len);
|
|
}
|
|
g_free(iov);
|
|
}
|
|
|
|
void virtio_gpu_cleanup_mapping(VirtIOGPU *g,
|
|
struct virtio_gpu_simple_resource *res)
|
|
{
|
|
virtio_gpu_cleanup_mapping_iov(g, res->iov, res->iov_cnt);
|
|
res->iov = NULL;
|
|
res->iov_cnt = 0;
|
|
g_free(res->addrs);
|
|
res->addrs = NULL;
|
|
|
|
if (res->blob) {
|
|
virtio_gpu_fini_udmabuf(res);
|
|
}
|
|
}
|
|
|
|
static void
|
|
virtio_gpu_resource_attach_backing(VirtIOGPU *g,
|
|
struct virtio_gpu_ctrl_command *cmd)
|
|
{
|
|
struct virtio_gpu_simple_resource *res;
|
|
struct virtio_gpu_resource_attach_backing ab;
|
|
int ret;
|
|
|
|
VIRTIO_GPU_FILL_CMD(ab);
|
|
virtio_gpu_bswap_32(&ab, sizeof(ab));
|
|
trace_virtio_gpu_cmd_res_back_attach(ab.resource_id);
|
|
|
|
res = virtio_gpu_find_resource(g, ab.resource_id);
|
|
if (!res) {
|
|
qemu_log_mask(LOG_GUEST_ERROR, "%s: illegal resource specified %d\n",
|
|
__func__, ab.resource_id);
|
|
cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_RESOURCE_ID;
|
|
return;
|
|
}
|
|
|
|
if (res->iov) {
|
|
cmd->error = VIRTIO_GPU_RESP_ERR_UNSPEC;
|
|
return;
|
|
}
|
|
|
|
ret = virtio_gpu_create_mapping_iov(g, ab.nr_entries, sizeof(ab), cmd,
|
|
&res->addrs, &res->iov, &res->iov_cnt);
|
|
if (ret != 0) {
|
|
cmd->error = VIRTIO_GPU_RESP_ERR_UNSPEC;
|
|
return;
|
|
}
|
|
}
|
|
|
|
static void
|
|
virtio_gpu_resource_detach_backing(VirtIOGPU *g,
|
|
struct virtio_gpu_ctrl_command *cmd)
|
|
{
|
|
struct virtio_gpu_simple_resource *res;
|
|
struct virtio_gpu_resource_detach_backing detach;
|
|
|
|
VIRTIO_GPU_FILL_CMD(detach);
|
|
virtio_gpu_bswap_32(&detach, sizeof(detach));
|
|
trace_virtio_gpu_cmd_res_back_detach(detach.resource_id);
|
|
|
|
res = virtio_gpu_find_check_resource(g, detach.resource_id, true,
|
|
__func__, &cmd->error);
|
|
if (!res) {
|
|
return;
|
|
}
|
|
virtio_gpu_cleanup_mapping(g, res);
|
|
}
|
|
|
|
void virtio_gpu_simple_process_cmd(VirtIOGPU *g,
|
|
struct virtio_gpu_ctrl_command *cmd)
|
|
{
|
|
VIRTIO_GPU_FILL_CMD(cmd->cmd_hdr);
|
|
virtio_gpu_ctrl_hdr_bswap(&cmd->cmd_hdr);
|
|
|
|
switch (cmd->cmd_hdr.type) {
|
|
case VIRTIO_GPU_CMD_GET_DISPLAY_INFO:
|
|
virtio_gpu_get_display_info(g, cmd);
|
|
break;
|
|
case VIRTIO_GPU_CMD_GET_EDID:
|
|
virtio_gpu_get_edid(g, cmd);
|
|
break;
|
|
case VIRTIO_GPU_CMD_RESOURCE_CREATE_2D:
|
|
virtio_gpu_resource_create_2d(g, cmd);
|
|
break;
|
|
case VIRTIO_GPU_CMD_RESOURCE_CREATE_BLOB:
|
|
if (!virtio_gpu_blob_enabled(g->parent_obj.conf)) {
|
|
cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_PARAMETER;
|
|
break;
|
|
}
|
|
virtio_gpu_resource_create_blob(g, cmd);
|
|
break;
|
|
case VIRTIO_GPU_CMD_RESOURCE_UNREF:
|
|
virtio_gpu_resource_unref(g, cmd);
|
|
break;
|
|
case VIRTIO_GPU_CMD_RESOURCE_FLUSH:
|
|
virtio_gpu_resource_flush(g, cmd);
|
|
break;
|
|
case VIRTIO_GPU_CMD_TRANSFER_TO_HOST_2D:
|
|
virtio_gpu_transfer_to_host_2d(g, cmd);
|
|
break;
|
|
case VIRTIO_GPU_CMD_SET_SCANOUT:
|
|
virtio_gpu_set_scanout(g, cmd);
|
|
break;
|
|
case VIRTIO_GPU_CMD_SET_SCANOUT_BLOB:
|
|
if (!virtio_gpu_blob_enabled(g->parent_obj.conf)) {
|
|
cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_PARAMETER;
|
|
break;
|
|
}
|
|
virtio_gpu_set_scanout_blob(g, cmd);
|
|
break;
|
|
case VIRTIO_GPU_CMD_RESOURCE_ATTACH_BACKING:
|
|
virtio_gpu_resource_attach_backing(g, cmd);
|
|
break;
|
|
case VIRTIO_GPU_CMD_RESOURCE_DETACH_BACKING:
|
|
virtio_gpu_resource_detach_backing(g, cmd);
|
|
break;
|
|
default:
|
|
cmd->error = VIRTIO_GPU_RESP_ERR_UNSPEC;
|
|
break;
|
|
}
|
|
if (!cmd->finished) {
|
|
if (!g->parent_obj.renderer_blocked) {
|
|
virtio_gpu_ctrl_response_nodata(g, cmd, cmd->error ? cmd->error :
|
|
VIRTIO_GPU_RESP_OK_NODATA);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void virtio_gpu_handle_ctrl_cb(VirtIODevice *vdev, VirtQueue *vq)
|
|
{
|
|
VirtIOGPU *g = VIRTIO_GPU(vdev);
|
|
qemu_bh_schedule(g->ctrl_bh);
|
|
}
|
|
|
|
static void virtio_gpu_handle_cursor_cb(VirtIODevice *vdev, VirtQueue *vq)
|
|
{
|
|
VirtIOGPU *g = VIRTIO_GPU(vdev);
|
|
qemu_bh_schedule(g->cursor_bh);
|
|
}
|
|
|
|
void virtio_gpu_process_cmdq(VirtIOGPU *g)
|
|
{
|
|
struct virtio_gpu_ctrl_command *cmd;
|
|
VirtIOGPUClass *vgc = VIRTIO_GPU_GET_CLASS(g);
|
|
|
|
if (g->processing_cmdq) {
|
|
return;
|
|
}
|
|
g->processing_cmdq = true;
|
|
while (!QTAILQ_EMPTY(&g->cmdq)) {
|
|
cmd = QTAILQ_FIRST(&g->cmdq);
|
|
|
|
if (g->parent_obj.renderer_blocked) {
|
|
break;
|
|
}
|
|
|
|
/* process command */
|
|
vgc->process_cmd(g, cmd);
|
|
|
|
QTAILQ_REMOVE(&g->cmdq, cmd, next);
|
|
if (virtio_gpu_stats_enabled(g->parent_obj.conf)) {
|
|
g->stats.requests++;
|
|
}
|
|
|
|
if (!cmd->finished) {
|
|
QTAILQ_INSERT_TAIL(&g->fenceq, cmd, next);
|
|
g->inflight++;
|
|
if (virtio_gpu_stats_enabled(g->parent_obj.conf)) {
|
|
if (g->stats.max_inflight < g->inflight) {
|
|
g->stats.max_inflight = g->inflight;
|
|
}
|
|
fprintf(stderr, "inflight: %3d (+)\r", g->inflight);
|
|
}
|
|
} else {
|
|
g_free(cmd);
|
|
}
|
|
}
|
|
g->processing_cmdq = false;
|
|
}
|
|
|
|
static void virtio_gpu_process_fenceq(VirtIOGPU *g)
|
|
{
|
|
struct virtio_gpu_ctrl_command *cmd, *tmp;
|
|
|
|
QTAILQ_FOREACH_SAFE(cmd, &g->fenceq, next, tmp) {
|
|
trace_virtio_gpu_fence_resp(cmd->cmd_hdr.fence_id);
|
|
virtio_gpu_ctrl_response_nodata(g, cmd, VIRTIO_GPU_RESP_OK_NODATA);
|
|
QTAILQ_REMOVE(&g->fenceq, cmd, next);
|
|
g_free(cmd);
|
|
g->inflight--;
|
|
if (virtio_gpu_stats_enabled(g->parent_obj.conf)) {
|
|
fprintf(stderr, "inflight: %3d (-)\r", g->inflight);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void virtio_gpu_handle_gl_flushed(VirtIOGPUBase *b)
|
|
{
|
|
VirtIOGPU *g = container_of(b, VirtIOGPU, parent_obj);
|
|
|
|
virtio_gpu_process_fenceq(g);
|
|
virtio_gpu_process_cmdq(g);
|
|
}
|
|
|
|
static void virtio_gpu_handle_ctrl(VirtIODevice *vdev, VirtQueue *vq)
|
|
{
|
|
VirtIOGPU *g = VIRTIO_GPU(vdev);
|
|
struct virtio_gpu_ctrl_command *cmd;
|
|
|
|
if (!virtio_queue_ready(vq)) {
|
|
return;
|
|
}
|
|
|
|
cmd = virtqueue_pop(vq, sizeof(struct virtio_gpu_ctrl_command));
|
|
while (cmd) {
|
|
cmd->vq = vq;
|
|
cmd->error = 0;
|
|
cmd->finished = false;
|
|
QTAILQ_INSERT_TAIL(&g->cmdq, cmd, next);
|
|
cmd = virtqueue_pop(vq, sizeof(struct virtio_gpu_ctrl_command));
|
|
}
|
|
|
|
virtio_gpu_process_cmdq(g);
|
|
}
|
|
|
|
static void virtio_gpu_ctrl_bh(void *opaque)
|
|
{
|
|
VirtIOGPU *g = opaque;
|
|
VirtIOGPUClass *vgc = VIRTIO_GPU_GET_CLASS(g);
|
|
|
|
vgc->handle_ctrl(VIRTIO_DEVICE(g), g->ctrl_vq);
|
|
}
|
|
|
|
static void virtio_gpu_handle_cursor(VirtIODevice *vdev, VirtQueue *vq)
|
|
{
|
|
VirtIOGPU *g = VIRTIO_GPU(vdev);
|
|
VirtQueueElement *elem;
|
|
size_t s;
|
|
struct virtio_gpu_update_cursor cursor_info;
|
|
|
|
if (!virtio_queue_ready(vq)) {
|
|
return;
|
|
}
|
|
for (;;) {
|
|
elem = virtqueue_pop(vq, sizeof(VirtQueueElement));
|
|
if (!elem) {
|
|
break;
|
|
}
|
|
|
|
s = iov_to_buf(elem->out_sg, elem->out_num, 0,
|
|
&cursor_info, sizeof(cursor_info));
|
|
if (s != sizeof(cursor_info)) {
|
|
qemu_log_mask(LOG_GUEST_ERROR,
|
|
"%s: cursor size incorrect %zu vs %zu\n",
|
|
__func__, s, sizeof(cursor_info));
|
|
} else {
|
|
virtio_gpu_bswap_32(&cursor_info, sizeof(cursor_info));
|
|
update_cursor(g, &cursor_info);
|
|
}
|
|
virtqueue_push(vq, elem, 0);
|
|
virtio_notify(vdev, vq);
|
|
g_free(elem);
|
|
}
|
|
}
|
|
|
|
static void virtio_gpu_cursor_bh(void *opaque)
|
|
{
|
|
VirtIOGPU *g = opaque;
|
|
virtio_gpu_handle_cursor(&g->parent_obj.parent_obj, g->cursor_vq);
|
|
}
|
|
|
|
static const VMStateDescription vmstate_virtio_gpu_scanout = {
|
|
.name = "virtio-gpu-one-scanout",
|
|
.version_id = 2,
|
|
.minimum_version_id = 1,
|
|
.fields = (const VMStateField[]) {
|
|
VMSTATE_UINT32(resource_id, struct virtio_gpu_scanout),
|
|
VMSTATE_UINT32(width, struct virtio_gpu_scanout),
|
|
VMSTATE_UINT32(height, struct virtio_gpu_scanout),
|
|
VMSTATE_INT32(x, struct virtio_gpu_scanout),
|
|
VMSTATE_INT32(y, struct virtio_gpu_scanout),
|
|
VMSTATE_UINT32(cursor.resource_id, struct virtio_gpu_scanout),
|
|
VMSTATE_UINT32(cursor.hot_x, struct virtio_gpu_scanout),
|
|
VMSTATE_UINT32(cursor.hot_y, struct virtio_gpu_scanout),
|
|
VMSTATE_UINT32(cursor.pos.x, struct virtio_gpu_scanout),
|
|
VMSTATE_UINT32(cursor.pos.y, struct virtio_gpu_scanout),
|
|
VMSTATE_UINT32_V(fb.format, struct virtio_gpu_scanout, 2),
|
|
VMSTATE_UINT32_V(fb.bytes_pp, struct virtio_gpu_scanout, 2),
|
|
VMSTATE_UINT32_V(fb.width, struct virtio_gpu_scanout, 2),
|
|
VMSTATE_UINT32_V(fb.height, struct virtio_gpu_scanout, 2),
|
|
VMSTATE_UINT32_V(fb.stride, struct virtio_gpu_scanout, 2),
|
|
VMSTATE_UINT32_V(fb.offset, struct virtio_gpu_scanout, 2),
|
|
VMSTATE_END_OF_LIST()
|
|
},
|
|
};
|
|
|
|
static const VMStateDescription vmstate_virtio_gpu_scanouts = {
|
|
.name = "virtio-gpu-scanouts",
|
|
.version_id = 1,
|
|
.fields = (const VMStateField[]) {
|
|
VMSTATE_INT32(parent_obj.enable, struct VirtIOGPU),
|
|
VMSTATE_UINT32_EQUAL(parent_obj.conf.max_outputs,
|
|
struct VirtIOGPU, NULL),
|
|
VMSTATE_STRUCT_VARRAY_UINT32(parent_obj.scanout, struct VirtIOGPU,
|
|
parent_obj.conf.max_outputs, 1,
|
|
vmstate_virtio_gpu_scanout,
|
|
struct virtio_gpu_scanout),
|
|
VMSTATE_END_OF_LIST()
|
|
},
|
|
};
|
|
|
|
static int virtio_gpu_save(QEMUFile *f, void *opaque, size_t size,
|
|
const VMStateField *field, JSONWriter *vmdesc)
|
|
{
|
|
VirtIOGPU *g = opaque;
|
|
struct virtio_gpu_simple_resource *res;
|
|
int i;
|
|
|
|
/* in 2d mode we should never find unprocessed commands here */
|
|
assert(QTAILQ_EMPTY(&g->cmdq));
|
|
|
|
QTAILQ_FOREACH(res, &g->reslist, next) {
|
|
if (res->blob_size) {
|
|
continue;
|
|
}
|
|
qemu_put_be32(f, res->resource_id);
|
|
qemu_put_be32(f, res->width);
|
|
qemu_put_be32(f, res->height);
|
|
qemu_put_be32(f, res->format);
|
|
qemu_put_be32(f, res->iov_cnt);
|
|
for (i = 0; i < res->iov_cnt; i++) {
|
|
qemu_put_be64(f, res->addrs[i]);
|
|
qemu_put_be32(f, res->iov[i].iov_len);
|
|
}
|
|
qemu_put_buffer(f, (void *)pixman_image_get_data(res->image),
|
|
pixman_image_get_stride(res->image) * res->height);
|
|
}
|
|
qemu_put_be32(f, 0); /* end of list */
|
|
|
|
return vmstate_save_state(f, &vmstate_virtio_gpu_scanouts, g, NULL);
|
|
}
|
|
|
|
static bool virtio_gpu_load_restore_mapping(VirtIOGPU *g,
|
|
struct virtio_gpu_simple_resource *res)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < res->iov_cnt; i++) {
|
|
hwaddr len = res->iov[i].iov_len;
|
|
res->iov[i].iov_base =
|
|
dma_memory_map(VIRTIO_DEVICE(g)->dma_as, res->addrs[i], &len,
|
|
DMA_DIRECTION_TO_DEVICE, MEMTXATTRS_UNSPECIFIED);
|
|
|
|
if (!res->iov[i].iov_base || len != res->iov[i].iov_len) {
|
|
/* Clean up the half-a-mapping we just created... */
|
|
if (res->iov[i].iov_base) {
|
|
dma_memory_unmap(VIRTIO_DEVICE(g)->dma_as, res->iov[i].iov_base,
|
|
len, DMA_DIRECTION_TO_DEVICE, 0);
|
|
}
|
|
/* ...and the mappings for previous loop iterations */
|
|
res->iov_cnt = i;
|
|
virtio_gpu_cleanup_mapping(g, res);
|
|
return false;
|
|
}
|
|
}
|
|
|
|
QTAILQ_INSERT_HEAD(&g->reslist, res, next);
|
|
g->hostmem += res->hostmem;
|
|
return true;
|
|
}
|
|
|
|
static int virtio_gpu_load(QEMUFile *f, void *opaque, size_t size,
|
|
const VMStateField *field)
|
|
{
|
|
VirtIOGPU *g = opaque;
|
|
struct virtio_gpu_simple_resource *res;
|
|
uint32_t resource_id, pformat;
|
|
void *bits = NULL;
|
|
int i;
|
|
|
|
g->hostmem = 0;
|
|
|
|
resource_id = qemu_get_be32(f);
|
|
while (resource_id != 0) {
|
|
res = virtio_gpu_find_resource(g, resource_id);
|
|
if (res) {
|
|
return -EINVAL;
|
|
}
|
|
|
|
res = g_new0(struct virtio_gpu_simple_resource, 1);
|
|
res->resource_id = resource_id;
|
|
res->width = qemu_get_be32(f);
|
|
res->height = qemu_get_be32(f);
|
|
res->format = qemu_get_be32(f);
|
|
res->iov_cnt = qemu_get_be32(f);
|
|
|
|
/* allocate */
|
|
pformat = virtio_gpu_get_pixman_format(res->format);
|
|
if (!pformat) {
|
|
g_free(res);
|
|
return -EINVAL;
|
|
}
|
|
|
|
res->hostmem = calc_image_hostmem(pformat, res->width, res->height);
|
|
#ifdef WIN32
|
|
bits = qemu_win32_map_alloc(res->hostmem, &res->handle, &error_warn);
|
|
if (!bits) {
|
|
g_free(res);
|
|
return -EINVAL;
|
|
}
|
|
#endif
|
|
res->image = pixman_image_create_bits(
|
|
pformat,
|
|
res->width, res->height,
|
|
bits, res->height ? res->hostmem / res->height : 0);
|
|
if (!res->image) {
|
|
g_free(res);
|
|
return -EINVAL;
|
|
}
|
|
#ifdef WIN32
|
|
pixman_image_set_destroy_function(res->image, win32_pixman_image_destroy, res->handle);
|
|
#endif
|
|
|
|
res->addrs = g_new(uint64_t, res->iov_cnt);
|
|
res->iov = g_new(struct iovec, res->iov_cnt);
|
|
|
|
/* read data */
|
|
for (i = 0; i < res->iov_cnt; i++) {
|
|
res->addrs[i] = qemu_get_be64(f);
|
|
res->iov[i].iov_len = qemu_get_be32(f);
|
|
}
|
|
qemu_get_buffer(f, (void *)pixman_image_get_data(res->image),
|
|
pixman_image_get_stride(res->image) * res->height);
|
|
|
|
if (!virtio_gpu_load_restore_mapping(g, res)) {
|
|
pixman_image_unref(res->image);
|
|
g_free(res);
|
|
return -EINVAL;
|
|
}
|
|
|
|
resource_id = qemu_get_be32(f);
|
|
}
|
|
|
|
/* load & apply scanout state */
|
|
vmstate_load_state(f, &vmstate_virtio_gpu_scanouts, g, 1);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int virtio_gpu_blob_save(QEMUFile *f, void *opaque, size_t size,
|
|
const VMStateField *field, JSONWriter *vmdesc)
|
|
{
|
|
VirtIOGPU *g = opaque;
|
|
struct virtio_gpu_simple_resource *res;
|
|
int i;
|
|
|
|
/* in 2d mode we should never find unprocessed commands here */
|
|
assert(QTAILQ_EMPTY(&g->cmdq));
|
|
|
|
QTAILQ_FOREACH(res, &g->reslist, next) {
|
|
if (!res->blob_size) {
|
|
continue;
|
|
}
|
|
assert(!res->image);
|
|
qemu_put_be32(f, res->resource_id);
|
|
qemu_put_be32(f, res->blob_size);
|
|
qemu_put_be32(f, res->iov_cnt);
|
|
for (i = 0; i < res->iov_cnt; i++) {
|
|
qemu_put_be64(f, res->addrs[i]);
|
|
qemu_put_be32(f, res->iov[i].iov_len);
|
|
}
|
|
}
|
|
qemu_put_be32(f, 0); /* end of list */
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int virtio_gpu_blob_load(QEMUFile *f, void *opaque, size_t size,
|
|
const VMStateField *field)
|
|
{
|
|
VirtIOGPU *g = opaque;
|
|
struct virtio_gpu_simple_resource *res;
|
|
uint32_t resource_id;
|
|
int i;
|
|
|
|
resource_id = qemu_get_be32(f);
|
|
while (resource_id != 0) {
|
|
res = virtio_gpu_find_resource(g, resource_id);
|
|
if (res) {
|
|
return -EINVAL;
|
|
}
|
|
|
|
res = g_new0(struct virtio_gpu_simple_resource, 1);
|
|
res->resource_id = resource_id;
|
|
res->blob_size = qemu_get_be32(f);
|
|
res->iov_cnt = qemu_get_be32(f);
|
|
res->addrs = g_new(uint64_t, res->iov_cnt);
|
|
res->iov = g_new(struct iovec, res->iov_cnt);
|
|
|
|
/* read data */
|
|
for (i = 0; i < res->iov_cnt; i++) {
|
|
res->addrs[i] = qemu_get_be64(f);
|
|
res->iov[i].iov_len = qemu_get_be32(f);
|
|
}
|
|
|
|
if (!virtio_gpu_load_restore_mapping(g, res)) {
|
|
g_free(res);
|
|
return -EINVAL;
|
|
}
|
|
|
|
virtio_gpu_init_udmabuf(res);
|
|
|
|
resource_id = qemu_get_be32(f);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int virtio_gpu_post_load(void *opaque, int version_id)
|
|
{
|
|
VirtIOGPU *g = opaque;
|
|
struct virtio_gpu_scanout *scanout;
|
|
struct virtio_gpu_simple_resource *res;
|
|
int i;
|
|
|
|
for (i = 0; i < g->parent_obj.conf.max_outputs; i++) {
|
|
scanout = &g->parent_obj.scanout[i];
|
|
if (!scanout->resource_id) {
|
|
continue;
|
|
}
|
|
|
|
res = virtio_gpu_find_resource(g, scanout->resource_id);
|
|
if (!res) {
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (scanout->fb.format != 0) {
|
|
uint32_t error = 0;
|
|
struct virtio_gpu_rect r = {
|
|
.x = scanout->x,
|
|
.y = scanout->y,
|
|
.width = scanout->width,
|
|
.height = scanout->height
|
|
};
|
|
|
|
if (!virtio_gpu_do_set_scanout(g, i, &scanout->fb, res, &r, &error)) {
|
|
return -EINVAL;
|
|
}
|
|
} else {
|
|
/* legacy v1 migration support */
|
|
if (!res->image) {
|
|
return -EINVAL;
|
|
}
|
|
scanout->ds = qemu_create_displaysurface_pixman(res->image);
|
|
#ifdef WIN32
|
|
qemu_displaysurface_win32_set_handle(scanout->ds, res->handle, 0);
|
|
#endif
|
|
dpy_gfx_replace_surface(scanout->con, scanout->ds);
|
|
}
|
|
|
|
dpy_gfx_update_full(scanout->con);
|
|
if (scanout->cursor.resource_id) {
|
|
update_cursor(g, &scanout->cursor);
|
|
}
|
|
res->scanout_bitmask |= (1 << i);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
void virtio_gpu_device_realize(DeviceState *qdev, Error **errp)
|
|
{
|
|
VirtIODevice *vdev = VIRTIO_DEVICE(qdev);
|
|
VirtIOGPU *g = VIRTIO_GPU(qdev);
|
|
|
|
if (virtio_gpu_blob_enabled(g->parent_obj.conf)) {
|
|
if (!virtio_gpu_rutabaga_enabled(g->parent_obj.conf) &&
|
|
!virtio_gpu_have_udmabuf()) {
|
|
error_setg(errp, "need rutabaga or udmabuf for blob resources");
|
|
return;
|
|
}
|
|
|
|
if (virtio_gpu_virgl_enabled(g->parent_obj.conf)) {
|
|
error_setg(errp, "blobs and virgl are not compatible (yet)");
|
|
return;
|
|
}
|
|
}
|
|
|
|
if (!virtio_gpu_base_device_realize(qdev,
|
|
virtio_gpu_handle_ctrl_cb,
|
|
virtio_gpu_handle_cursor_cb,
|
|
errp)) {
|
|
return;
|
|
}
|
|
|
|
g->ctrl_vq = virtio_get_queue(vdev, 0);
|
|
g->cursor_vq = virtio_get_queue(vdev, 1);
|
|
g->ctrl_bh = virtio_bh_new_guarded(qdev, virtio_gpu_ctrl_bh, g);
|
|
g->cursor_bh = virtio_bh_new_guarded(qdev, virtio_gpu_cursor_bh, g);
|
|
g->reset_bh = qemu_bh_new(virtio_gpu_reset_bh, g);
|
|
qemu_cond_init(&g->reset_cond);
|
|
QTAILQ_INIT(&g->reslist);
|
|
QTAILQ_INIT(&g->cmdq);
|
|
QTAILQ_INIT(&g->fenceq);
|
|
}
|
|
|
|
static void virtio_gpu_device_unrealize(DeviceState *qdev)
|
|
{
|
|
VirtIOGPU *g = VIRTIO_GPU(qdev);
|
|
|
|
g_clear_pointer(&g->ctrl_bh, qemu_bh_delete);
|
|
g_clear_pointer(&g->cursor_bh, qemu_bh_delete);
|
|
g_clear_pointer(&g->reset_bh, qemu_bh_delete);
|
|
qemu_cond_destroy(&g->reset_cond);
|
|
virtio_gpu_base_device_unrealize(qdev);
|
|
}
|
|
|
|
static void virtio_gpu_reset_bh(void *opaque)
|
|
{
|
|
VirtIOGPU *g = VIRTIO_GPU(opaque);
|
|
VirtIOGPUClass *vgc = VIRTIO_GPU_GET_CLASS(g);
|
|
struct virtio_gpu_simple_resource *res, *tmp;
|
|
uint32_t resource_id;
|
|
Error *local_err = NULL;
|
|
int i = 0;
|
|
|
|
QTAILQ_FOREACH_SAFE(res, &g->reslist, next, tmp) {
|
|
resource_id = res->resource_id;
|
|
vgc->resource_destroy(g, res, &local_err);
|
|
if (local_err) {
|
|
error_append_hint(&local_err, "%s: %s resource_destroy"
|
|
"for resource_id = %"PRIu32" failed.\n",
|
|
__func__, object_get_typename(OBJECT(g)),
|
|
resource_id);
|
|
/* error_report_err frees the error object for us */
|
|
error_report_err(local_err);
|
|
local_err = NULL;
|
|
}
|
|
}
|
|
|
|
for (i = 0; i < g->parent_obj.conf.max_outputs; i++) {
|
|
dpy_gfx_replace_surface(g->parent_obj.scanout[i].con, NULL);
|
|
}
|
|
|
|
g->reset_finished = true;
|
|
qemu_cond_signal(&g->reset_cond);
|
|
}
|
|
|
|
void virtio_gpu_reset(VirtIODevice *vdev)
|
|
{
|
|
VirtIOGPU *g = VIRTIO_GPU(vdev);
|
|
struct virtio_gpu_ctrl_command *cmd;
|
|
|
|
if (qemu_in_vcpu_thread()) {
|
|
g->reset_finished = false;
|
|
qemu_bh_schedule(g->reset_bh);
|
|
while (!g->reset_finished) {
|
|
qemu_cond_wait_bql(&g->reset_cond);
|
|
}
|
|
} else {
|
|
aio_bh_call(g->reset_bh);
|
|
}
|
|
|
|
while (!QTAILQ_EMPTY(&g->cmdq)) {
|
|
cmd = QTAILQ_FIRST(&g->cmdq);
|
|
QTAILQ_REMOVE(&g->cmdq, cmd, next);
|
|
g_free(cmd);
|
|
}
|
|
|
|
while (!QTAILQ_EMPTY(&g->fenceq)) {
|
|
cmd = QTAILQ_FIRST(&g->fenceq);
|
|
QTAILQ_REMOVE(&g->fenceq, cmd, next);
|
|
g->inflight--;
|
|
g_free(cmd);
|
|
}
|
|
|
|
virtio_gpu_base_reset(VIRTIO_GPU_BASE(vdev));
|
|
}
|
|
|
|
static void
|
|
virtio_gpu_get_config(VirtIODevice *vdev, uint8_t *config)
|
|
{
|
|
VirtIOGPUBase *g = VIRTIO_GPU_BASE(vdev);
|
|
|
|
memcpy(config, &g->virtio_config, sizeof(g->virtio_config));
|
|
}
|
|
|
|
static void
|
|
virtio_gpu_set_config(VirtIODevice *vdev, const uint8_t *config)
|
|
{
|
|
VirtIOGPUBase *g = VIRTIO_GPU_BASE(vdev);
|
|
const struct virtio_gpu_config *vgconfig =
|
|
(const struct virtio_gpu_config *)config;
|
|
|
|
if (vgconfig->events_clear) {
|
|
g->virtio_config.events_read &= ~vgconfig->events_clear;
|
|
}
|
|
}
|
|
|
|
static bool virtio_gpu_blob_state_needed(void *opaque)
|
|
{
|
|
VirtIOGPU *g = VIRTIO_GPU(opaque);
|
|
|
|
return virtio_gpu_blob_enabled(g->parent_obj.conf);
|
|
}
|
|
|
|
const VMStateDescription vmstate_virtio_gpu_blob_state = {
|
|
.name = "virtio-gpu/blob",
|
|
.minimum_version_id = VIRTIO_GPU_VM_VERSION,
|
|
.version_id = VIRTIO_GPU_VM_VERSION,
|
|
.needed = virtio_gpu_blob_state_needed,
|
|
.fields = (const VMStateField[]){
|
|
{
|
|
.name = "virtio-gpu/blob",
|
|
.info = &(const VMStateInfo) {
|
|
.name = "blob",
|
|
.get = virtio_gpu_blob_load,
|
|
.put = virtio_gpu_blob_save,
|
|
},
|
|
.flags = VMS_SINGLE,
|
|
} /* device */,
|
|
VMSTATE_END_OF_LIST()
|
|
},
|
|
};
|
|
|
|
/*
|
|
* For historical reasons virtio_gpu does not adhere to virtio migration
|
|
* scheme as described in doc/virtio-migration.txt, in a sense that no
|
|
* save/load callback are provided to the core. Instead the device data
|
|
* is saved/loaded after the core data.
|
|
*
|
|
* Because of this we need a special vmsd.
|
|
*/
|
|
static const VMStateDescription vmstate_virtio_gpu = {
|
|
.name = "virtio-gpu",
|
|
.minimum_version_id = VIRTIO_GPU_VM_VERSION,
|
|
.version_id = VIRTIO_GPU_VM_VERSION,
|
|
.fields = (const VMStateField[]) {
|
|
VMSTATE_VIRTIO_DEVICE /* core */,
|
|
{
|
|
.name = "virtio-gpu",
|
|
.info = &(const VMStateInfo) {
|
|
.name = "virtio-gpu",
|
|
.get = virtio_gpu_load,
|
|
.put = virtio_gpu_save,
|
|
},
|
|
.flags = VMS_SINGLE,
|
|
} /* device */,
|
|
VMSTATE_END_OF_LIST()
|
|
},
|
|
.subsections = (const VMStateDescription * const []) {
|
|
&vmstate_virtio_gpu_blob_state,
|
|
NULL
|
|
},
|
|
.post_load = virtio_gpu_post_load,
|
|
};
|
|
|
|
static Property virtio_gpu_properties[] = {
|
|
VIRTIO_GPU_BASE_PROPERTIES(VirtIOGPU, parent_obj.conf),
|
|
DEFINE_PROP_SIZE("max_hostmem", VirtIOGPU, conf_max_hostmem,
|
|
256 * MiB),
|
|
DEFINE_PROP_BIT("blob", VirtIOGPU, parent_obj.conf.flags,
|
|
VIRTIO_GPU_FLAG_BLOB_ENABLED, false),
|
|
DEFINE_PROP_SIZE("hostmem", VirtIOGPU, parent_obj.conf.hostmem, 0),
|
|
DEFINE_PROP_END_OF_LIST(),
|
|
};
|
|
|
|
static void virtio_gpu_class_init(ObjectClass *klass, void *data)
|
|
{
|
|
DeviceClass *dc = DEVICE_CLASS(klass);
|
|
VirtioDeviceClass *vdc = VIRTIO_DEVICE_CLASS(klass);
|
|
VirtIOGPUClass *vgc = VIRTIO_GPU_CLASS(klass);
|
|
VirtIOGPUBaseClass *vgbc = &vgc->parent;
|
|
|
|
vgc->handle_ctrl = virtio_gpu_handle_ctrl;
|
|
vgc->process_cmd = virtio_gpu_simple_process_cmd;
|
|
vgc->update_cursor_data = virtio_gpu_update_cursor_data;
|
|
vgc->resource_destroy = virtio_gpu_resource_destroy;
|
|
vgbc->gl_flushed = virtio_gpu_handle_gl_flushed;
|
|
|
|
vdc->realize = virtio_gpu_device_realize;
|
|
vdc->unrealize = virtio_gpu_device_unrealize;
|
|
vdc->reset = virtio_gpu_reset;
|
|
vdc->get_config = virtio_gpu_get_config;
|
|
vdc->set_config = virtio_gpu_set_config;
|
|
|
|
dc->vmsd = &vmstate_virtio_gpu;
|
|
device_class_set_props(dc, virtio_gpu_properties);
|
|
}
|
|
|
|
static const TypeInfo virtio_gpu_info = {
|
|
.name = TYPE_VIRTIO_GPU,
|
|
.parent = TYPE_VIRTIO_GPU_BASE,
|
|
.instance_size = sizeof(VirtIOGPU),
|
|
.class_size = sizeof(VirtIOGPUClass),
|
|
.class_init = virtio_gpu_class_init,
|
|
};
|
|
module_obj(TYPE_VIRTIO_GPU);
|
|
module_kconfig(VIRTIO_GPU);
|
|
|
|
static void virtio_register_types(void)
|
|
{
|
|
type_register_static(&virtio_gpu_info);
|
|
}
|
|
|
|
type_init(virtio_register_types)
|