tracing: remove TCG memory access tracing
If you really want to trace all memory operations TCG plugins gives you a more flexible interface for doing so. Signed-off-by: Alex Bennée <alex.bennee@linaro.org> Reviewed-by: Richard Henderson <richard.henderson@linaro.org> Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org> Cc: Luis Vilanova <vilanova@imperial.ac.uk> Cc: Stefan Hajnoczi <stefanha@redhat.com> Message-Id: <20220204204335.1689602-19-alex.bennee@linaro.org>
This commit is contained in:
parent
d9a6bad542
commit
c51e51005b
@ -13,14 +13,6 @@
|
|||||||
* See the COPYING file in the top-level directory.
|
* See the COPYING file in the top-level directory.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
static void atomic_trace_rmw_pre(CPUArchState *env, target_ulong addr,
|
|
||||||
MemOpIdx oi)
|
|
||||||
{
|
|
||||||
CPUState *cpu = env_cpu(env);
|
|
||||||
|
|
||||||
trace_guest_rmw_before_exec(cpu, addr, oi);
|
|
||||||
}
|
|
||||||
|
|
||||||
static void atomic_trace_rmw_post(CPUArchState *env, target_ulong addr,
|
static void atomic_trace_rmw_post(CPUArchState *env, target_ulong addr,
|
||||||
MemOpIdx oi)
|
MemOpIdx oi)
|
||||||
{
|
{
|
||||||
@ -28,24 +20,12 @@ static void atomic_trace_rmw_post(CPUArchState *env, target_ulong addr,
|
|||||||
}
|
}
|
||||||
|
|
||||||
#if HAVE_ATOMIC128
|
#if HAVE_ATOMIC128
|
||||||
static void atomic_trace_ld_pre(CPUArchState *env, target_ulong addr,
|
|
||||||
MemOpIdx oi)
|
|
||||||
{
|
|
||||||
trace_guest_ld_before_exec(env_cpu(env), addr, oi);
|
|
||||||
}
|
|
||||||
|
|
||||||
static void atomic_trace_ld_post(CPUArchState *env, target_ulong addr,
|
static void atomic_trace_ld_post(CPUArchState *env, target_ulong addr,
|
||||||
MemOpIdx oi)
|
MemOpIdx oi)
|
||||||
{
|
{
|
||||||
qemu_plugin_vcpu_mem_cb(env_cpu(env), addr, oi, QEMU_PLUGIN_MEM_R);
|
qemu_plugin_vcpu_mem_cb(env_cpu(env), addr, oi, QEMU_PLUGIN_MEM_R);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void atomic_trace_st_pre(CPUArchState *env, target_ulong addr,
|
|
||||||
MemOpIdx oi)
|
|
||||||
{
|
|
||||||
trace_guest_st_before_exec(env_cpu(env), addr, oi);
|
|
||||||
}
|
|
||||||
|
|
||||||
static void atomic_trace_st_post(CPUArchState *env, target_ulong addr,
|
static void atomic_trace_st_post(CPUArchState *env, target_ulong addr,
|
||||||
MemOpIdx oi)
|
MemOpIdx oi)
|
||||||
{
|
{
|
||||||
|
@ -77,7 +77,6 @@ ABI_TYPE ATOMIC_NAME(cmpxchg)(CPUArchState *env, target_ulong addr,
|
|||||||
PAGE_READ | PAGE_WRITE, retaddr);
|
PAGE_READ | PAGE_WRITE, retaddr);
|
||||||
DATA_TYPE ret;
|
DATA_TYPE ret;
|
||||||
|
|
||||||
atomic_trace_rmw_pre(env, addr, oi);
|
|
||||||
#if DATA_SIZE == 16
|
#if DATA_SIZE == 16
|
||||||
ret = atomic16_cmpxchg(haddr, cmpv, newv);
|
ret = atomic16_cmpxchg(haddr, cmpv, newv);
|
||||||
#else
|
#else
|
||||||
@ -97,7 +96,6 @@ ABI_TYPE ATOMIC_NAME(ld)(CPUArchState *env, target_ulong addr,
|
|||||||
PAGE_READ, retaddr);
|
PAGE_READ, retaddr);
|
||||||
DATA_TYPE val;
|
DATA_TYPE val;
|
||||||
|
|
||||||
atomic_trace_ld_pre(env, addr, oi);
|
|
||||||
val = atomic16_read(haddr);
|
val = atomic16_read(haddr);
|
||||||
ATOMIC_MMU_CLEANUP;
|
ATOMIC_MMU_CLEANUP;
|
||||||
atomic_trace_ld_post(env, addr, oi);
|
atomic_trace_ld_post(env, addr, oi);
|
||||||
@ -110,7 +108,6 @@ void ATOMIC_NAME(st)(CPUArchState *env, target_ulong addr, ABI_TYPE val,
|
|||||||
DATA_TYPE *haddr = atomic_mmu_lookup(env, addr, oi, DATA_SIZE,
|
DATA_TYPE *haddr = atomic_mmu_lookup(env, addr, oi, DATA_SIZE,
|
||||||
PAGE_WRITE, retaddr);
|
PAGE_WRITE, retaddr);
|
||||||
|
|
||||||
atomic_trace_st_pre(env, addr, oi);
|
|
||||||
atomic16_set(haddr, val);
|
atomic16_set(haddr, val);
|
||||||
ATOMIC_MMU_CLEANUP;
|
ATOMIC_MMU_CLEANUP;
|
||||||
atomic_trace_st_post(env, addr, oi);
|
atomic_trace_st_post(env, addr, oi);
|
||||||
@ -124,7 +121,6 @@ ABI_TYPE ATOMIC_NAME(xchg)(CPUArchState *env, target_ulong addr, ABI_TYPE val,
|
|||||||
PAGE_READ | PAGE_WRITE, retaddr);
|
PAGE_READ | PAGE_WRITE, retaddr);
|
||||||
DATA_TYPE ret;
|
DATA_TYPE ret;
|
||||||
|
|
||||||
atomic_trace_rmw_pre(env, addr, oi);
|
|
||||||
ret = qatomic_xchg__nocheck(haddr, val);
|
ret = qatomic_xchg__nocheck(haddr, val);
|
||||||
ATOMIC_MMU_CLEANUP;
|
ATOMIC_MMU_CLEANUP;
|
||||||
atomic_trace_rmw_post(env, addr, oi);
|
atomic_trace_rmw_post(env, addr, oi);
|
||||||
@ -138,7 +134,6 @@ ABI_TYPE ATOMIC_NAME(X)(CPUArchState *env, target_ulong addr, \
|
|||||||
DATA_TYPE *haddr = atomic_mmu_lookup(env, addr, oi, DATA_SIZE, \
|
DATA_TYPE *haddr = atomic_mmu_lookup(env, addr, oi, DATA_SIZE, \
|
||||||
PAGE_READ | PAGE_WRITE, retaddr); \
|
PAGE_READ | PAGE_WRITE, retaddr); \
|
||||||
DATA_TYPE ret; \
|
DATA_TYPE ret; \
|
||||||
atomic_trace_rmw_pre(env, addr, oi); \
|
|
||||||
ret = qatomic_##X(haddr, val); \
|
ret = qatomic_##X(haddr, val); \
|
||||||
ATOMIC_MMU_CLEANUP; \
|
ATOMIC_MMU_CLEANUP; \
|
||||||
atomic_trace_rmw_post(env, addr, oi); \
|
atomic_trace_rmw_post(env, addr, oi); \
|
||||||
@ -171,7 +166,6 @@ ABI_TYPE ATOMIC_NAME(X)(CPUArchState *env, target_ulong addr, \
|
|||||||
XDATA_TYPE *haddr = atomic_mmu_lookup(env, addr, oi, DATA_SIZE, \
|
XDATA_TYPE *haddr = atomic_mmu_lookup(env, addr, oi, DATA_SIZE, \
|
||||||
PAGE_READ | PAGE_WRITE, retaddr); \
|
PAGE_READ | PAGE_WRITE, retaddr); \
|
||||||
XDATA_TYPE cmp, old, new, val = xval; \
|
XDATA_TYPE cmp, old, new, val = xval; \
|
||||||
atomic_trace_rmw_pre(env, addr, oi); \
|
|
||||||
smp_mb(); \
|
smp_mb(); \
|
||||||
cmp = qatomic_read__nocheck(haddr); \
|
cmp = qatomic_read__nocheck(haddr); \
|
||||||
do { \
|
do { \
|
||||||
@ -216,7 +210,6 @@ ABI_TYPE ATOMIC_NAME(cmpxchg)(CPUArchState *env, target_ulong addr,
|
|||||||
PAGE_READ | PAGE_WRITE, retaddr);
|
PAGE_READ | PAGE_WRITE, retaddr);
|
||||||
DATA_TYPE ret;
|
DATA_TYPE ret;
|
||||||
|
|
||||||
atomic_trace_rmw_pre(env, addr, oi);
|
|
||||||
#if DATA_SIZE == 16
|
#if DATA_SIZE == 16
|
||||||
ret = atomic16_cmpxchg(haddr, BSWAP(cmpv), BSWAP(newv));
|
ret = atomic16_cmpxchg(haddr, BSWAP(cmpv), BSWAP(newv));
|
||||||
#else
|
#else
|
||||||
@ -236,7 +229,6 @@ ABI_TYPE ATOMIC_NAME(ld)(CPUArchState *env, target_ulong addr,
|
|||||||
PAGE_READ, retaddr);
|
PAGE_READ, retaddr);
|
||||||
DATA_TYPE val;
|
DATA_TYPE val;
|
||||||
|
|
||||||
atomic_trace_ld_pre(env, addr, oi);
|
|
||||||
val = atomic16_read(haddr);
|
val = atomic16_read(haddr);
|
||||||
ATOMIC_MMU_CLEANUP;
|
ATOMIC_MMU_CLEANUP;
|
||||||
atomic_trace_ld_post(env, addr, oi);
|
atomic_trace_ld_post(env, addr, oi);
|
||||||
@ -249,7 +241,6 @@ void ATOMIC_NAME(st)(CPUArchState *env, target_ulong addr, ABI_TYPE val,
|
|||||||
DATA_TYPE *haddr = atomic_mmu_lookup(env, addr, oi, DATA_SIZE,
|
DATA_TYPE *haddr = atomic_mmu_lookup(env, addr, oi, DATA_SIZE,
|
||||||
PAGE_WRITE, retaddr);
|
PAGE_WRITE, retaddr);
|
||||||
|
|
||||||
atomic_trace_st_pre(env, addr, oi);
|
|
||||||
val = BSWAP(val);
|
val = BSWAP(val);
|
||||||
atomic16_set(haddr, val);
|
atomic16_set(haddr, val);
|
||||||
ATOMIC_MMU_CLEANUP;
|
ATOMIC_MMU_CLEANUP;
|
||||||
@ -264,7 +255,6 @@ ABI_TYPE ATOMIC_NAME(xchg)(CPUArchState *env, target_ulong addr, ABI_TYPE val,
|
|||||||
PAGE_READ | PAGE_WRITE, retaddr);
|
PAGE_READ | PAGE_WRITE, retaddr);
|
||||||
ABI_TYPE ret;
|
ABI_TYPE ret;
|
||||||
|
|
||||||
atomic_trace_rmw_pre(env, addr, oi);
|
|
||||||
ret = qatomic_xchg__nocheck(haddr, BSWAP(val));
|
ret = qatomic_xchg__nocheck(haddr, BSWAP(val));
|
||||||
ATOMIC_MMU_CLEANUP;
|
ATOMIC_MMU_CLEANUP;
|
||||||
atomic_trace_rmw_post(env, addr, oi);
|
atomic_trace_rmw_post(env, addr, oi);
|
||||||
@ -278,7 +268,6 @@ ABI_TYPE ATOMIC_NAME(X)(CPUArchState *env, target_ulong addr, \
|
|||||||
DATA_TYPE *haddr = atomic_mmu_lookup(env, addr, oi, DATA_SIZE, \
|
DATA_TYPE *haddr = atomic_mmu_lookup(env, addr, oi, DATA_SIZE, \
|
||||||
PAGE_READ | PAGE_WRITE, retaddr); \
|
PAGE_READ | PAGE_WRITE, retaddr); \
|
||||||
DATA_TYPE ret; \
|
DATA_TYPE ret; \
|
||||||
atomic_trace_rmw_pre(env, addr, oi); \
|
|
||||||
ret = qatomic_##X(haddr, BSWAP(val)); \
|
ret = qatomic_##X(haddr, BSWAP(val)); \
|
||||||
ATOMIC_MMU_CLEANUP; \
|
ATOMIC_MMU_CLEANUP; \
|
||||||
atomic_trace_rmw_post(env, addr, oi); \
|
atomic_trace_rmw_post(env, addr, oi); \
|
||||||
@ -308,7 +297,6 @@ ABI_TYPE ATOMIC_NAME(X)(CPUArchState *env, target_ulong addr, \
|
|||||||
XDATA_TYPE *haddr = atomic_mmu_lookup(env, addr, oi, DATA_SIZE, \
|
XDATA_TYPE *haddr = atomic_mmu_lookup(env, addr, oi, DATA_SIZE, \
|
||||||
PAGE_READ | PAGE_WRITE, retaddr); \
|
PAGE_READ | PAGE_WRITE, retaddr); \
|
||||||
XDATA_TYPE ldo, ldn, old, new, val = xval; \
|
XDATA_TYPE ldo, ldn, old, new, val = xval; \
|
||||||
atomic_trace_rmw_pre(env, addr, oi); \
|
|
||||||
smp_mb(); \
|
smp_mb(); \
|
||||||
ldn = qatomic_read__nocheck(haddr); \
|
ldn = qatomic_read__nocheck(haddr); \
|
||||||
do { \
|
do { \
|
||||||
|
@ -2140,7 +2140,6 @@ static inline uint64_t cpu_load_helper(CPUArchState *env, abi_ptr addr,
|
|||||||
{
|
{
|
||||||
uint64_t ret;
|
uint64_t ret;
|
||||||
|
|
||||||
trace_guest_ld_before_exec(env_cpu(env), addr, oi);
|
|
||||||
ret = full_load(env, addr, oi, retaddr);
|
ret = full_load(env, addr, oi, retaddr);
|
||||||
qemu_plugin_vcpu_mem_cb(env_cpu(env), addr, oi, QEMU_PLUGIN_MEM_R);
|
qemu_plugin_vcpu_mem_cb(env_cpu(env), addr, oi, QEMU_PLUGIN_MEM_R);
|
||||||
return ret;
|
return ret;
|
||||||
@ -2487,7 +2486,6 @@ static inline void cpu_store_helper(CPUArchState *env, target_ulong addr,
|
|||||||
uint64_t val, MemOpIdx oi, uintptr_t ra,
|
uint64_t val, MemOpIdx oi, uintptr_t ra,
|
||||||
FullStoreHelper *full_store)
|
FullStoreHelper *full_store)
|
||||||
{
|
{
|
||||||
trace_guest_st_before_exec(env_cpu(env), addr, oi);
|
|
||||||
full_store(env, addr, val, oi, ra);
|
full_store(env, addr, val, oi, ra);
|
||||||
qemu_plugin_vcpu_mem_cb(env_cpu(env), addr, oi, QEMU_PLUGIN_MEM_W);
|
qemu_plugin_vcpu_mem_cb(env_cpu(env), addr, oi, QEMU_PLUGIN_MEM_W);
|
||||||
}
|
}
|
||||||
|
@ -250,7 +250,6 @@ uint8_t cpu_ldb_mmu(CPUArchState *env, abi_ptr addr,
|
|||||||
uint8_t ret;
|
uint8_t ret;
|
||||||
|
|
||||||
validate_memop(oi, MO_UB);
|
validate_memop(oi, MO_UB);
|
||||||
trace_guest_ld_before_exec(env_cpu(env), addr, oi);
|
|
||||||
haddr = cpu_mmu_lookup(env, addr, oi, ra, MMU_DATA_LOAD);
|
haddr = cpu_mmu_lookup(env, addr, oi, ra, MMU_DATA_LOAD);
|
||||||
ret = ldub_p(haddr);
|
ret = ldub_p(haddr);
|
||||||
clear_helper_retaddr();
|
clear_helper_retaddr();
|
||||||
@ -265,7 +264,6 @@ uint16_t cpu_ldw_be_mmu(CPUArchState *env, abi_ptr addr,
|
|||||||
uint16_t ret;
|
uint16_t ret;
|
||||||
|
|
||||||
validate_memop(oi, MO_BEUW);
|
validate_memop(oi, MO_BEUW);
|
||||||
trace_guest_ld_before_exec(env_cpu(env), addr, oi);
|
|
||||||
haddr = cpu_mmu_lookup(env, addr, oi, ra, MMU_DATA_LOAD);
|
haddr = cpu_mmu_lookup(env, addr, oi, ra, MMU_DATA_LOAD);
|
||||||
ret = lduw_be_p(haddr);
|
ret = lduw_be_p(haddr);
|
||||||
clear_helper_retaddr();
|
clear_helper_retaddr();
|
||||||
@ -280,7 +278,6 @@ uint32_t cpu_ldl_be_mmu(CPUArchState *env, abi_ptr addr,
|
|||||||
uint32_t ret;
|
uint32_t ret;
|
||||||
|
|
||||||
validate_memop(oi, MO_BEUL);
|
validate_memop(oi, MO_BEUL);
|
||||||
trace_guest_ld_before_exec(env_cpu(env), addr, oi);
|
|
||||||
haddr = cpu_mmu_lookup(env, addr, oi, ra, MMU_DATA_LOAD);
|
haddr = cpu_mmu_lookup(env, addr, oi, ra, MMU_DATA_LOAD);
|
||||||
ret = ldl_be_p(haddr);
|
ret = ldl_be_p(haddr);
|
||||||
clear_helper_retaddr();
|
clear_helper_retaddr();
|
||||||
@ -295,7 +292,6 @@ uint64_t cpu_ldq_be_mmu(CPUArchState *env, abi_ptr addr,
|
|||||||
uint64_t ret;
|
uint64_t ret;
|
||||||
|
|
||||||
validate_memop(oi, MO_BEUQ);
|
validate_memop(oi, MO_BEUQ);
|
||||||
trace_guest_ld_before_exec(env_cpu(env), addr, oi);
|
|
||||||
haddr = cpu_mmu_lookup(env, addr, oi, ra, MMU_DATA_LOAD);
|
haddr = cpu_mmu_lookup(env, addr, oi, ra, MMU_DATA_LOAD);
|
||||||
ret = ldq_be_p(haddr);
|
ret = ldq_be_p(haddr);
|
||||||
clear_helper_retaddr();
|
clear_helper_retaddr();
|
||||||
@ -310,7 +306,6 @@ uint16_t cpu_ldw_le_mmu(CPUArchState *env, abi_ptr addr,
|
|||||||
uint16_t ret;
|
uint16_t ret;
|
||||||
|
|
||||||
validate_memop(oi, MO_LEUW);
|
validate_memop(oi, MO_LEUW);
|
||||||
trace_guest_ld_before_exec(env_cpu(env), addr, oi);
|
|
||||||
haddr = cpu_mmu_lookup(env, addr, oi, ra, MMU_DATA_LOAD);
|
haddr = cpu_mmu_lookup(env, addr, oi, ra, MMU_DATA_LOAD);
|
||||||
ret = lduw_le_p(haddr);
|
ret = lduw_le_p(haddr);
|
||||||
clear_helper_retaddr();
|
clear_helper_retaddr();
|
||||||
@ -325,7 +320,6 @@ uint32_t cpu_ldl_le_mmu(CPUArchState *env, abi_ptr addr,
|
|||||||
uint32_t ret;
|
uint32_t ret;
|
||||||
|
|
||||||
validate_memop(oi, MO_LEUL);
|
validate_memop(oi, MO_LEUL);
|
||||||
trace_guest_ld_before_exec(env_cpu(env), addr, oi);
|
|
||||||
haddr = cpu_mmu_lookup(env, addr, oi, ra, MMU_DATA_LOAD);
|
haddr = cpu_mmu_lookup(env, addr, oi, ra, MMU_DATA_LOAD);
|
||||||
ret = ldl_le_p(haddr);
|
ret = ldl_le_p(haddr);
|
||||||
clear_helper_retaddr();
|
clear_helper_retaddr();
|
||||||
@ -340,7 +334,6 @@ uint64_t cpu_ldq_le_mmu(CPUArchState *env, abi_ptr addr,
|
|||||||
uint64_t ret;
|
uint64_t ret;
|
||||||
|
|
||||||
validate_memop(oi, MO_LEUQ);
|
validate_memop(oi, MO_LEUQ);
|
||||||
trace_guest_ld_before_exec(env_cpu(env), addr, oi);
|
|
||||||
haddr = cpu_mmu_lookup(env, addr, oi, ra, MMU_DATA_LOAD);
|
haddr = cpu_mmu_lookup(env, addr, oi, ra, MMU_DATA_LOAD);
|
||||||
ret = ldq_le_p(haddr);
|
ret = ldq_le_p(haddr);
|
||||||
clear_helper_retaddr();
|
clear_helper_retaddr();
|
||||||
@ -354,7 +347,6 @@ void cpu_stb_mmu(CPUArchState *env, abi_ptr addr, uint8_t val,
|
|||||||
void *haddr;
|
void *haddr;
|
||||||
|
|
||||||
validate_memop(oi, MO_UB);
|
validate_memop(oi, MO_UB);
|
||||||
trace_guest_st_before_exec(env_cpu(env), addr, oi);
|
|
||||||
haddr = cpu_mmu_lookup(env, addr, oi, ra, MMU_DATA_STORE);
|
haddr = cpu_mmu_lookup(env, addr, oi, ra, MMU_DATA_STORE);
|
||||||
stb_p(haddr, val);
|
stb_p(haddr, val);
|
||||||
clear_helper_retaddr();
|
clear_helper_retaddr();
|
||||||
@ -367,7 +359,6 @@ void cpu_stw_be_mmu(CPUArchState *env, abi_ptr addr, uint16_t val,
|
|||||||
void *haddr;
|
void *haddr;
|
||||||
|
|
||||||
validate_memop(oi, MO_BEUW);
|
validate_memop(oi, MO_BEUW);
|
||||||
trace_guest_st_before_exec(env_cpu(env), addr, oi);
|
|
||||||
haddr = cpu_mmu_lookup(env, addr, oi, ra, MMU_DATA_STORE);
|
haddr = cpu_mmu_lookup(env, addr, oi, ra, MMU_DATA_STORE);
|
||||||
stw_be_p(haddr, val);
|
stw_be_p(haddr, val);
|
||||||
clear_helper_retaddr();
|
clear_helper_retaddr();
|
||||||
@ -380,7 +371,6 @@ void cpu_stl_be_mmu(CPUArchState *env, abi_ptr addr, uint32_t val,
|
|||||||
void *haddr;
|
void *haddr;
|
||||||
|
|
||||||
validate_memop(oi, MO_BEUL);
|
validate_memop(oi, MO_BEUL);
|
||||||
trace_guest_st_before_exec(env_cpu(env), addr, oi);
|
|
||||||
haddr = cpu_mmu_lookup(env, addr, oi, ra, MMU_DATA_STORE);
|
haddr = cpu_mmu_lookup(env, addr, oi, ra, MMU_DATA_STORE);
|
||||||
stl_be_p(haddr, val);
|
stl_be_p(haddr, val);
|
||||||
clear_helper_retaddr();
|
clear_helper_retaddr();
|
||||||
@ -393,7 +383,6 @@ void cpu_stq_be_mmu(CPUArchState *env, abi_ptr addr, uint64_t val,
|
|||||||
void *haddr;
|
void *haddr;
|
||||||
|
|
||||||
validate_memop(oi, MO_BEUQ);
|
validate_memop(oi, MO_BEUQ);
|
||||||
trace_guest_st_before_exec(env_cpu(env), addr, oi);
|
|
||||||
haddr = cpu_mmu_lookup(env, addr, oi, ra, MMU_DATA_STORE);
|
haddr = cpu_mmu_lookup(env, addr, oi, ra, MMU_DATA_STORE);
|
||||||
stq_be_p(haddr, val);
|
stq_be_p(haddr, val);
|
||||||
clear_helper_retaddr();
|
clear_helper_retaddr();
|
||||||
@ -406,7 +395,6 @@ void cpu_stw_le_mmu(CPUArchState *env, abi_ptr addr, uint16_t val,
|
|||||||
void *haddr;
|
void *haddr;
|
||||||
|
|
||||||
validate_memop(oi, MO_LEUW);
|
validate_memop(oi, MO_LEUW);
|
||||||
trace_guest_st_before_exec(env_cpu(env), addr, oi);
|
|
||||||
haddr = cpu_mmu_lookup(env, addr, oi, ra, MMU_DATA_STORE);
|
haddr = cpu_mmu_lookup(env, addr, oi, ra, MMU_DATA_STORE);
|
||||||
stw_le_p(haddr, val);
|
stw_le_p(haddr, val);
|
||||||
clear_helper_retaddr();
|
clear_helper_retaddr();
|
||||||
@ -419,7 +407,6 @@ void cpu_stl_le_mmu(CPUArchState *env, abi_ptr addr, uint32_t val,
|
|||||||
void *haddr;
|
void *haddr;
|
||||||
|
|
||||||
validate_memop(oi, MO_LEUL);
|
validate_memop(oi, MO_LEUL);
|
||||||
trace_guest_st_before_exec(env_cpu(env), addr, oi);
|
|
||||||
haddr = cpu_mmu_lookup(env, addr, oi, ra, MMU_DATA_STORE);
|
haddr = cpu_mmu_lookup(env, addr, oi, ra, MMU_DATA_STORE);
|
||||||
stl_le_p(haddr, val);
|
stl_le_p(haddr, val);
|
||||||
clear_helper_retaddr();
|
clear_helper_retaddr();
|
||||||
@ -432,7 +419,6 @@ void cpu_stq_le_mmu(CPUArchState *env, abi_ptr addr, uint64_t val,
|
|||||||
void *haddr;
|
void *haddr;
|
||||||
|
|
||||||
validate_memop(oi, MO_LEUQ);
|
validate_memop(oi, MO_LEUQ);
|
||||||
trace_guest_st_before_exec(env_cpu(env), addr, oi);
|
|
||||||
haddr = cpu_mmu_lookup(env, addr, oi, ra, MMU_DATA_STORE);
|
haddr = cpu_mmu_lookup(env, addr, oi, ra, MMU_DATA_STORE);
|
||||||
stq_le_p(haddr, val);
|
stq_le_p(haddr, val);
|
||||||
clear_helper_retaddr();
|
clear_helper_retaddr();
|
||||||
|
@ -27,7 +27,6 @@
|
|||||||
#include "tcg/tcg.h"
|
#include "tcg/tcg.h"
|
||||||
#include "tcg/tcg-op.h"
|
#include "tcg/tcg-op.h"
|
||||||
#include "tcg/tcg-mo.h"
|
#include "tcg/tcg-mo.h"
|
||||||
#include "trace-tcg.h"
|
|
||||||
#include "exec/plugin-gen.h"
|
#include "exec/plugin-gen.h"
|
||||||
|
|
||||||
/* Reduce the number of ifdefs below. This assumes that all uses of
|
/* Reduce the number of ifdefs below. This assumes that all uses of
|
||||||
@ -2877,7 +2876,6 @@ void tcg_gen_qemu_ld_i32(TCGv_i32 val, TCGv addr, TCGArg idx, MemOp memop)
|
|||||||
tcg_gen_req_mo(TCG_MO_LD_LD | TCG_MO_ST_LD);
|
tcg_gen_req_mo(TCG_MO_LD_LD | TCG_MO_ST_LD);
|
||||||
memop = tcg_canonicalize_memop(memop, 0, 0);
|
memop = tcg_canonicalize_memop(memop, 0, 0);
|
||||||
oi = make_memop_idx(memop, idx);
|
oi = make_memop_idx(memop, idx);
|
||||||
trace_guest_ld_before_tcg(tcg_ctx->cpu, cpu_env, addr, oi);
|
|
||||||
|
|
||||||
orig_memop = memop;
|
orig_memop = memop;
|
||||||
if (!TCG_TARGET_HAS_MEMORY_BSWAP && (memop & MO_BSWAP)) {
|
if (!TCG_TARGET_HAS_MEMORY_BSWAP && (memop & MO_BSWAP)) {
|
||||||
@ -2916,7 +2914,6 @@ void tcg_gen_qemu_st_i32(TCGv_i32 val, TCGv addr, TCGArg idx, MemOp memop)
|
|||||||
tcg_gen_req_mo(TCG_MO_LD_ST | TCG_MO_ST_ST);
|
tcg_gen_req_mo(TCG_MO_LD_ST | TCG_MO_ST_ST);
|
||||||
memop = tcg_canonicalize_memop(memop, 0, 1);
|
memop = tcg_canonicalize_memop(memop, 0, 1);
|
||||||
oi = make_memop_idx(memop, idx);
|
oi = make_memop_idx(memop, idx);
|
||||||
trace_guest_st_before_tcg(tcg_ctx->cpu, cpu_env, addr, oi);
|
|
||||||
|
|
||||||
if (!TCG_TARGET_HAS_MEMORY_BSWAP && (memop & MO_BSWAP)) {
|
if (!TCG_TARGET_HAS_MEMORY_BSWAP && (memop & MO_BSWAP)) {
|
||||||
swap = tcg_temp_new_i32();
|
swap = tcg_temp_new_i32();
|
||||||
@ -2965,7 +2962,6 @@ void tcg_gen_qemu_ld_i64(TCGv_i64 val, TCGv addr, TCGArg idx, MemOp memop)
|
|||||||
tcg_gen_req_mo(TCG_MO_LD_LD | TCG_MO_ST_LD);
|
tcg_gen_req_mo(TCG_MO_LD_LD | TCG_MO_ST_LD);
|
||||||
memop = tcg_canonicalize_memop(memop, 1, 0);
|
memop = tcg_canonicalize_memop(memop, 1, 0);
|
||||||
oi = make_memop_idx(memop, idx);
|
oi = make_memop_idx(memop, idx);
|
||||||
trace_guest_ld_before_tcg(tcg_ctx->cpu, cpu_env, addr, oi);
|
|
||||||
|
|
||||||
orig_memop = memop;
|
orig_memop = memop;
|
||||||
if (!TCG_TARGET_HAS_MEMORY_BSWAP && (memop & MO_BSWAP)) {
|
if (!TCG_TARGET_HAS_MEMORY_BSWAP && (memop & MO_BSWAP)) {
|
||||||
@ -3013,7 +3009,6 @@ void tcg_gen_qemu_st_i64(TCGv_i64 val, TCGv addr, TCGArg idx, MemOp memop)
|
|||||||
tcg_gen_req_mo(TCG_MO_LD_ST | TCG_MO_ST_ST);
|
tcg_gen_req_mo(TCG_MO_LD_ST | TCG_MO_ST_ST);
|
||||||
memop = tcg_canonicalize_memop(memop, 1, 1);
|
memop = tcg_canonicalize_memop(memop, 1, 1);
|
||||||
oi = make_memop_idx(memop, idx);
|
oi = make_memop_idx(memop, idx);
|
||||||
trace_guest_st_before_tcg(tcg_ctx->cpu, cpu_env, addr, oi);
|
|
||||||
|
|
||||||
if (!TCG_TARGET_HAS_MEMORY_BSWAP && (memop & MO_BSWAP)) {
|
if (!TCG_TARGET_HAS_MEMORY_BSWAP && (memop & MO_BSWAP)) {
|
||||||
swap = tcg_temp_new_i64();
|
swap = tcg_temp_new_i64();
|
||||||
|
14
trace-events
14
trace-events
@ -117,20 +117,6 @@ vcpu guest_cpu_exit(void)
|
|||||||
# Targets: all
|
# Targets: all
|
||||||
vcpu guest_cpu_reset(void)
|
vcpu guest_cpu_reset(void)
|
||||||
|
|
||||||
# tcg/tcg-op.c
|
|
||||||
|
|
||||||
# @vaddr: Access' virtual address.
|
|
||||||
# @memopidx: Access' information (see below).
|
|
||||||
#
|
|
||||||
# Start virtual memory access (before any potential access violation).
|
|
||||||
# Does not include memory accesses performed by devices.
|
|
||||||
#
|
|
||||||
# Mode: user, softmmu
|
|
||||||
# Targets: TCG(all)
|
|
||||||
vcpu tcg guest_ld_before(TCGv vaddr, uint32_t memopidx) "info=%d", "vaddr=0x%016"PRIx64" memopidx=0x%x"
|
|
||||||
vcpu tcg guest_st_before(TCGv vaddr, uint32_t memopidx) "info=%d", "vaddr=0x%016"PRIx64" memopidx=0x%x"
|
|
||||||
vcpu tcg guest_rmw_before(TCGv vaddr, uint32_t memopidx) "info=%d", "vaddr=0x%016"PRIx64" memopidx=0x%x"
|
|
||||||
|
|
||||||
# include/user/syscall-trace.h
|
# include/user/syscall-trace.h
|
||||||
|
|
||||||
# @num: System call number.
|
# @num: System call number.
|
||||||
|
Loading…
Reference in New Issue
Block a user