2010-05-22 18:52:39 +02:00
|
|
|
# Trace events for debugging and performance instrumentation
|
|
|
|
#
|
|
|
|
# This file is processed by the tracetool script during the build.
|
|
|
|
#
|
|
|
|
# To add a new trace event:
|
|
|
|
#
|
|
|
|
# 1. Choose a name for the trace event. Declare its arguments and format
|
|
|
|
# string.
|
|
|
|
#
|
|
|
|
# 2. Call the trace event from code using trace_##name, e.g. multiwrite_cb() ->
|
|
|
|
# trace_multiwrite_cb(). The source file must #include "trace.h".
|
|
|
|
#
|
|
|
|
# Format of a trace event:
|
|
|
|
#
|
2010-05-24 12:32:09 +02:00
|
|
|
# [disable] <name>(<type1> <arg1>[, <type2> <arg2>] ...) "<format-string>"
|
2010-05-22 18:52:39 +02:00
|
|
|
#
|
2011-08-31 09:25:35 +02:00
|
|
|
# Example: g_malloc(size_t size) "size %zu"
|
2010-05-22 18:52:39 +02:00
|
|
|
#
|
2010-05-24 12:32:09 +02:00
|
|
|
# The "disable" keyword will build without the trace event.
|
|
|
|
#
|
2010-05-22 18:52:39 +02:00
|
|
|
# The <name> must be a valid as a C function name.
|
|
|
|
#
|
|
|
|
# Types should be standard C types. Use void * for pointers because the trace
|
|
|
|
# system may not have the necessary headers included.
|
|
|
|
#
|
|
|
|
# The <format-string> should be a sprintf()-compatible format string.
|
2010-05-22 19:09:25 +02:00
|
|
|
|
2010-08-11 13:45:11 +02:00
|
|
|
# ioport.c
|
2015-10-16 15:08:34 +02:00
|
|
|
cpu_in(unsigned int addr, char size, unsigned int val) "addr %#x(%c) value %u"
|
|
|
|
cpu_out(unsigned int addr, char size, unsigned int val) "addr %#x(%c) value %u"
|
2010-08-11 13:46:03 +02:00
|
|
|
|
|
|
|
# balloon.c
|
|
|
|
# Since requests are raised via monitor, not many tracepoints are needed.
|
2011-08-31 20:31:58 +02:00
|
|
|
balloon_event(void *opaque, unsigned long addr) "opaque %p addr %lu"
|
2010-10-20 18:41:28 +02:00
|
|
|
|
2010-11-16 13:20:25 +01:00
|
|
|
# vl.c
|
2011-08-31 20:31:58 +02:00
|
|
|
vm_state_notify(int running, int reason) "running %d reason %d"
|
2013-03-08 11:42:24 +01:00
|
|
|
load_file(const char *name, const char *path) "name %s location %s"
|
2013-03-22 09:26:59 +01:00
|
|
|
runstate_set(int new_state) "new state %d"
|
2014-03-06 21:03:36 +01:00
|
|
|
system_wakeup_request(int reason) "reason=%d"
|
2014-06-21 20:43:03 +02:00
|
|
|
qemu_system_shutdown_request(void) ""
|
|
|
|
qemu_system_powerdown_request(void) ""
|
2010-12-06 17:08:01 +01:00
|
|
|
|
2011-01-19 09:49:50 +01:00
|
|
|
# spice-qemu-char.c
|
2016-12-12 23:17:59 +01:00
|
|
|
spice_vmc_write(ssize_t out, int len) "spice wrote %zd of requested %d"
|
2011-08-31 20:31:58 +02:00
|
|
|
spice_vmc_read(int bytes, int len) "spice read %d of requested %d"
|
|
|
|
spice_vmc_register_interface(void *scd) "spice vmc registered interface %p"
|
|
|
|
spice_vmc_unregister_interface(void *scd) "spice vmc unregistered interface %p"
|
2012-12-05 16:15:34 +01:00
|
|
|
spice_vmc_event(int event) "spice vmc event %d"
|
2011-02-17 23:45:07 +01:00
|
|
|
|
2011-09-23 09:23:06 +02:00
|
|
|
# monitor.c
|
monitor: Simplify event throttling
The event throttling state machine is hard to understand. I'm not
sure it's entirely correct. Rewrite it in a more straightforward
manner:
State 1: No event sent recently (less than evconf->rate ns ago)
Invariant: evstate->timer is not pending, evstate->qdict is null
On event: send event, arm timer, goto state 2
State 2: Event sent recently, no additional event being delayed
Invariant: evstate->timer is pending, evstate->qdict is null
On event: store it in evstate->qdict, goto state 3
On timer: goto state 1
State 3: Event sent recently, additional event being delayed
Invariant: evstate->timer is pending, evstate->qdict is non-null
On event: store it in evstate->qdict, goto state 3
On timer: send evstate->qdict, clear evstate->qdict,
arm timer, goto state 2
Signed-off-by: Markus Armbruster <armbru@redhat.com>
Message-Id: <1444921716-9511-3-git-send-email-armbru@redhat.com>
Reviewed-by: Eric Blake <eblake@redhat.com>
2015-10-15 17:08:31 +02:00
|
|
|
monitor_protocol_event_handler(uint32_t event, void *qdict) "event=%d data=%p"
|
2012-06-14 19:12:57 +02:00
|
|
|
monitor_protocol_event_emit(uint32_t event, void *data) "event=%d data=%p"
|
monitor: Simplify event throttling
The event throttling state machine is hard to understand. I'm not
sure it's entirely correct. Rewrite it in a more straightforward
manner:
State 1: No event sent recently (less than evconf->rate ns ago)
Invariant: evstate->timer is not pending, evstate->qdict is null
On event: send event, arm timer, goto state 2
State 2: Event sent recently, no additional event being delayed
Invariant: evstate->timer is pending, evstate->qdict is null
On event: store it in evstate->qdict, goto state 3
On timer: goto state 1
State 3: Event sent recently, additional event being delayed
Invariant: evstate->timer is pending, evstate->qdict is non-null
On event: store it in evstate->qdict, goto state 3
On timer: send evstate->qdict, clear evstate->qdict,
arm timer, goto state 2
Signed-off-by: Markus Armbruster <armbru@redhat.com>
Message-Id: <1444921716-9511-3-git-send-email-armbru@redhat.com>
Reviewed-by: Eric Blake <eblake@redhat.com>
2015-10-15 17:08:31 +02:00
|
|
|
monitor_protocol_event_queue(uint32_t event, void *qdict, uint64_t rate) "event=%d data=%p rate=%" PRId64
|
2011-10-16 00:56:45 +02:00
|
|
|
|
2011-11-24 12:15:28 +01:00
|
|
|
# dma-helpers.c
|
2016-05-23 14:54:05 +02:00
|
|
|
dma_blk_io(void *dbs, void *bs, int64_t offset, bool to_dev) "dbs=%p bs=%p offset=%" PRId64 " to_dev=%d"
|
2011-11-24 12:15:28 +01:00
|
|
|
dma_aio_cancel(void *dbs) "dbs=%p"
|
|
|
|
dma_complete(void *dbs, int ret, void *cb) "dbs=%p ret=%d cb=%p"
|
2014-10-07 13:59:18 +02:00
|
|
|
dma_blk_cb(void *dbs, int ret) "dbs=%p ret=%d"
|
2011-11-24 12:15:28 +01:00
|
|
|
dma_map_wait(void *dbs) "dbs=%p"
|
2012-03-11 17:11:26 +01:00
|
|
|
|
2013-03-29 05:27:05 +01:00
|
|
|
# kvm-all.c
|
2013-09-04 12:26:25 +02:00
|
|
|
kvm_ioctl(int type, void *arg) "type 0x%x, arg %p"
|
|
|
|
kvm_vm_ioctl(int type, void *arg) "type 0x%x, arg %p"
|
|
|
|
kvm_vcpu_ioctl(int cpu_index, int type, void *arg) "cpu_index %d, type 0x%x, arg %p"
|
2013-03-29 05:27:52 +01:00
|
|
|
kvm_run_exit(int cpu_index, uint32_t reason) "cpu_index %d, reason %d"
|
2014-02-26 18:20:00 +01:00
|
|
|
kvm_device_ioctl(int fd, int type, void *arg) "dev fd %d, type 0x%x, arg %p"
|
2014-05-09 10:06:46 +02:00
|
|
|
kvm_failed_reg_get(uint64_t id, const char *msg) "Warning: Unable to retrieve ONEREG %" PRIu64 " from KVM: %s"
|
|
|
|
kvm_failed_reg_set(uint64_t id, const char *msg) "Warning: Unable to set ONEREG %" PRIu64 " to KVM: %s"
|
2016-07-14 07:56:35 +02:00
|
|
|
kvm_irqchip_commit_routes(void) ""
|
|
|
|
kvm_irqchip_add_msi_route(int virq) "Adding MSI route virq=%d"
|
|
|
|
kvm_irqchip_update_msi_route(int virq) "Updating MSI route virq=%d"
|
2013-03-29 05:27:52 +01:00
|
|
|
|
2014-08-01 18:08:57 +02:00
|
|
|
# TCG related tracing (mostly disabled by default)
|
|
|
|
# cpu-exec.c
|
|
|
|
disable exec_tb(void *tb, uintptr_t pc) "tb:%p pc=0x%"PRIxPTR
|
|
|
|
disable exec_tb_nocache(void *tb, uintptr_t pc) "tb:%p pc=0x%"PRIxPTR
|
2016-04-21 14:58:23 +02:00
|
|
|
disable exec_tb_exit(void *last_tb, unsigned int flags) "tb:%p flags=%x"
|
2014-08-01 18:08:57 +02:00
|
|
|
|
|
|
|
# translate-all.c
|
|
|
|
translate_block(void *tb, uintptr_t pc, uint8_t *tb_code) "tb:%p, pc:0x%"PRIxPTR", tb_code:%p"
|
|
|
|
|
2013-07-28 14:57:22 +02:00
|
|
|
# memory.c
|
2016-03-02 21:12:54 +01:00
|
|
|
memory_region_ops_read(int cpu_index, void *mr, uint64_t addr, uint64_t value, unsigned size) "cpu %d mr %p addr %#"PRIx64" value %#"PRIx64" size %u"
|
|
|
|
memory_region_ops_write(int cpu_index, void *mr, uint64_t addr, uint64_t value, unsigned size) "cpu %d mr %p addr %#"PRIx64" value %#"PRIx64" size %u"
|
|
|
|
memory_region_subpage_read(int cpu_index, void *mr, uint64_t offset, uint64_t value, unsigned size) "cpu %d mr %p offset %#"PRIx64" value %#"PRIx64" size %u"
|
|
|
|
memory_region_subpage_write(int cpu_index, void *mr, uint64_t offset, uint64_t value, unsigned size) "cpu %d mr %p offset %#"PRIx64" value %#"PRIx64" size %u"
|
2016-03-02 21:12:55 +01:00
|
|
|
memory_region_tb_read(int cpu_index, uint64_t addr, uint64_t value, unsigned size) "cpu %d addr %#"PRIx64" value %#"PRIx64" size %u"
|
|
|
|
memory_region_tb_write(int cpu_index, uint64_t addr, uint64_t value, unsigned size) "cpu %d addr %#"PRIx64" value %#"PRIx64" size %u"
|
memory: Don't use memcpy for ram_device regions
With a vfio assigned device we lay down a base MemoryRegion registered
as an IO region, giving us read & write accessors. If the region
supports mmap, we lay down a higher priority sub-region MemoryRegion
on top of the base layer initialized as a RAM device pointer to the
mmap. Finally, if we have any quirks for the device (ie. address
ranges that need additional virtualization support), we put another IO
sub-region on top of the mmap MemoryRegion. When this is flattened,
we now potentially have sub-page mmap MemoryRegions exposed which
cannot be directly mapped through KVM.
This is as expected, but a subtle detail of this is that we end up
with two different access mechanisms through QEMU. If we disable the
mmap MemoryRegion, we make use of the IO MemoryRegion and service
accesses using pread and pwrite to the vfio device file descriptor.
If the mmap MemoryRegion is enabled and results in one of these
sub-page gaps, QEMU handles the access as RAM, using memcpy to the
mmap. Using either pread/pwrite or the mmap directly should be
correct, but using memcpy causes us problems. I expect that not only
does memcpy not necessarily honor the original width and alignment in
performing a copy, but it potentially also uses processor instructions
not intended for MMIO spaces. It turns out that this has been a
problem for Realtek NIC assignment, which has such a quirk that
creates a sub-page mmap MemoryRegion access.
To resolve this, we disable memory_access_is_direct() for ram_device
regions since QEMU assumes that it can use memcpy for those regions.
Instead we access through MemoryRegionOps, which replaces the memcpy
with simple de-references of standard sizes to the host memory.
With this patch we attempt to provide unrestricted access to the RAM
device, allowing byte through qword access as well as unaligned
access. The assumption here is that accesses initiated by the VM are
driven by a device specific driver, which knows the device
capabilities. If unaligned accesses are not supported by the device,
we don't want them to work in a VM by performing multiple aligned
accesses to compose the unaligned access. A down-side of this
philosophy is that the xp command from the monitor attempts to use
the largest available access weidth, unaware of the underlying
device. Using memcpy had this same restriction, but at least now an
operator can dump individual registers, even if blocks of device
memory may result in access widths beyond the capabilities of a
given device (RTL NICs only support up to dword).
Reported-by: Thorsten Kohfeldt <thorsten.kohfeldt@gmx.de>
Signed-off-by: Alex Williamson <alex.williamson@redhat.com>
Acked-by: Paolo Bonzini <pbonzini@redhat.com>
2016-10-31 16:53:03 +01:00
|
|
|
memory_region_ram_device_read(int cpu_index, void *mr, uint64_t addr, uint64_t value, unsigned size) "cpu %d mr %p addr %#"PRIx64" value %#"PRIx64" size %u"
|
|
|
|
memory_region_ram_device_write(int cpu_index, void *mr, uint64_t addr, uint64_t value, unsigned size) "cpu %d mr %p addr %#"PRIx64" value %#"PRIx64" size %u"
|
2013-07-28 14:57:22 +02:00
|
|
|
|
2016-06-09 19:31:47 +02:00
|
|
|
### Guest events, keep at bottom
|
|
|
|
|
2016-09-19 14:55:13 +02:00
|
|
|
|
|
|
|
## vCPU
|
|
|
|
|
|
|
|
# Hot-plug a new virtual (guest) CPU
|
|
|
|
#
|
2016-10-05 14:03:29 +02:00
|
|
|
# Mode: user, softmmu
|
2016-09-19 14:55:13 +02:00
|
|
|
# Targets: all
|
|
|
|
vcpu guest_cpu_enter(void)
|
|
|
|
|
2016-12-26 22:24:46 +01:00
|
|
|
# Hot-unplug a virtual (guest) CPU
|
|
|
|
#
|
|
|
|
# Mode: user, softmmu
|
|
|
|
# Targets: all
|
|
|
|
vcpu guest_cpu_exit(void)
|
|
|
|
|
2016-09-19 14:55:18 +02:00
|
|
|
# Reset the state of a virtual (guest) CPU
|
|
|
|
#
|
2016-10-05 14:03:29 +02:00
|
|
|
# Mode: user, softmmu
|
2016-09-19 14:55:18 +02:00
|
|
|
# Targets: all
|
|
|
|
vcpu guest_cpu_reset(void)
|
|
|
|
|
2016-06-09 19:31:47 +02:00
|
|
|
# @vaddr: Access' virtual address.
|
|
|
|
# @info : Access' information (see below).
|
|
|
|
#
|
|
|
|
# Start virtual memory access (before any potential access violation).
|
|
|
|
#
|
|
|
|
# Does not include memory accesses performed by devices.
|
|
|
|
#
|
|
|
|
# Access information can be parsed as:
|
|
|
|
#
|
|
|
|
# struct mem_info {
|
|
|
|
# uint8_t size_shift : 2; /* interpreted as "1 << size_shift" bytes */
|
|
|
|
# bool sign_extend: 1; /* sign-extended */
|
|
|
|
# uint8_t endianness : 1; /* 0: little, 1: big */
|
|
|
|
# bool store : 1; /* wheter it's a store operation */
|
|
|
|
# };
|
|
|
|
#
|
2016-09-22 20:40:21 +02:00
|
|
|
# Mode: user, softmmu
|
2016-06-09 19:31:47 +02:00
|
|
|
# Targets: TCG(all)
|
|
|
|
disable vcpu tcg guest_mem_before(TCGv vaddr, uint8_t info) "info=%d", "vaddr=0x%016"PRIx64" info=%d"
|
2016-06-21 15:52:04 +02:00
|
|
|
|
|
|
|
# @num: System call number.
|
|
|
|
# @arg*: System call argument value.
|
|
|
|
#
|
|
|
|
# Start executing a guest system call in syscall emulation mode.
|
|
|
|
#
|
2016-09-22 20:40:21 +02:00
|
|
|
# Mode: user
|
2016-06-21 15:52:04 +02:00
|
|
|
# Targets: TCG(all)
|
|
|
|
disable vcpu guest_user_syscall(uint64_t num, uint64_t arg1, uint64_t arg2, uint64_t arg3, uint64_t arg4, uint64_t arg5, uint64_t arg6, uint64_t arg7, uint64_t arg8) "num=0x%016"PRIx64" arg1=0x%016"PRIx64" arg2=0x%016"PRIx64" arg3=0x%016"PRIx64" arg4=0x%016"PRIx64" arg5=0x%016"PRIx64" arg6=0x%016"PRIx64" arg7=0x%016"PRIx64" arg8=0x%016"PRIx64
|
|
|
|
|
|
|
|
# @num: System call number.
|
|
|
|
# @ret: System call result value.
|
|
|
|
#
|
|
|
|
# Finish executing a guest system call in syscall emulation mode.
|
|
|
|
#
|
2016-09-22 20:40:21 +02:00
|
|
|
# Mode: user
|
2016-06-21 15:52:04 +02:00
|
|
|
# Targets: TCG(all)
|
|
|
|
disable vcpu guest_user_syscall_ret(uint64_t num, uint64_t ret) "num=0x%016"PRIx64" ret=0x%016"PRIx64
|