3b2992e42f
For w64, some entries need 'uintptr_t' instead of 'unsigned long'. For other host systems, both data types are identical, so nothing changes. Signed-off-by: Stefan Weil <sw@weilnetz.de>
231 lines
9.9 KiB
C
231 lines
9.9 KiB
C
/*
|
|
* common defines for all CPUs
|
|
*
|
|
* Copyright (c) 2003 Fabrice Bellard
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with this library; if not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
#ifndef CPU_DEFS_H
|
|
#define CPU_DEFS_H
|
|
|
|
#ifndef NEED_CPU_H
|
|
#error cpu.h included from common code
|
|
#endif
|
|
|
|
#include "config.h"
|
|
#include <setjmp.h>
|
|
#include <inttypes.h>
|
|
#include <signal.h>
|
|
#include "osdep.h"
|
|
#include "qemu-queue.h"
|
|
#include "targphys.h"
|
|
|
|
#ifndef TARGET_LONG_BITS
|
|
#error TARGET_LONG_BITS must be defined before including this header
|
|
#endif
|
|
|
|
#define TARGET_LONG_SIZE (TARGET_LONG_BITS / 8)
|
|
|
|
typedef int16_t target_short __attribute__ ((aligned(TARGET_SHORT_ALIGNMENT)));
|
|
typedef uint16_t target_ushort __attribute__((aligned(TARGET_SHORT_ALIGNMENT)));
|
|
typedef int32_t target_int __attribute__((aligned(TARGET_INT_ALIGNMENT)));
|
|
typedef uint32_t target_uint __attribute__((aligned(TARGET_INT_ALIGNMENT)));
|
|
typedef int64_t target_llong __attribute__((aligned(TARGET_LLONG_ALIGNMENT)));
|
|
typedef uint64_t target_ullong __attribute__((aligned(TARGET_LLONG_ALIGNMENT)));
|
|
/* target_ulong is the type of a virtual address */
|
|
#if TARGET_LONG_SIZE == 4
|
|
typedef int32_t target_long __attribute__((aligned(TARGET_LONG_ALIGNMENT)));
|
|
typedef uint32_t target_ulong __attribute__((aligned(TARGET_LONG_ALIGNMENT)));
|
|
#define TARGET_FMT_lx "%08x"
|
|
#define TARGET_FMT_ld "%d"
|
|
#define TARGET_FMT_lu "%u"
|
|
#elif TARGET_LONG_SIZE == 8
|
|
typedef int64_t target_long __attribute__((aligned(TARGET_LONG_ALIGNMENT)));
|
|
typedef uint64_t target_ulong __attribute__((aligned(TARGET_LONG_ALIGNMENT)));
|
|
#define TARGET_FMT_lx "%016" PRIx64
|
|
#define TARGET_FMT_ld "%" PRId64
|
|
#define TARGET_FMT_lu "%" PRIu64
|
|
#else
|
|
#error TARGET_LONG_SIZE undefined
|
|
#endif
|
|
|
|
#define EXCP_INTERRUPT 0x10000 /* async interruption */
|
|
#define EXCP_HLT 0x10001 /* hlt instruction reached */
|
|
#define EXCP_DEBUG 0x10002 /* cpu stopped after a breakpoint or singlestep */
|
|
#define EXCP_HALTED 0x10003 /* cpu is halted (waiting for external event) */
|
|
|
|
#define TB_JMP_CACHE_BITS 12
|
|
#define TB_JMP_CACHE_SIZE (1 << TB_JMP_CACHE_BITS)
|
|
|
|
/* Only the bottom TB_JMP_PAGE_BITS of the jump cache hash bits vary for
|
|
addresses on the same page. The top bits are the same. This allows
|
|
TLB invalidation to quickly clear a subset of the hash table. */
|
|
#define TB_JMP_PAGE_BITS (TB_JMP_CACHE_BITS / 2)
|
|
#define TB_JMP_PAGE_SIZE (1 << TB_JMP_PAGE_BITS)
|
|
#define TB_JMP_ADDR_MASK (TB_JMP_PAGE_SIZE - 1)
|
|
#define TB_JMP_PAGE_MASK (TB_JMP_CACHE_SIZE - TB_JMP_PAGE_SIZE)
|
|
|
|
#if !defined(CONFIG_USER_ONLY)
|
|
#define CPU_TLB_BITS 8
|
|
#define CPU_TLB_SIZE (1 << CPU_TLB_BITS)
|
|
|
|
#if HOST_LONG_BITS == 32 && TARGET_LONG_BITS == 32
|
|
#define CPU_TLB_ENTRY_BITS 4
|
|
#else
|
|
#define CPU_TLB_ENTRY_BITS 5
|
|
#endif
|
|
|
|
typedef struct CPUTLBEntry {
|
|
/* bit TARGET_LONG_BITS to TARGET_PAGE_BITS : virtual address
|
|
bit TARGET_PAGE_BITS-1..4 : Nonzero for accesses that should not
|
|
go directly to ram.
|
|
bit 3 : indicates that the entry is invalid
|
|
bit 2..0 : zero
|
|
*/
|
|
target_ulong addr_read;
|
|
target_ulong addr_write;
|
|
target_ulong addr_code;
|
|
/* Addend to virtual address to get host address. IO accesses
|
|
use the corresponding iotlb value. */
|
|
uintptr_t addend;
|
|
/* padding to get a power of two size */
|
|
uint8_t dummy[(1 << CPU_TLB_ENTRY_BITS) -
|
|
(sizeof(target_ulong) * 3 +
|
|
((-sizeof(target_ulong) * 3) & (sizeof(uintptr_t) - 1)) +
|
|
sizeof(uintptr_t))];
|
|
} CPUTLBEntry;
|
|
|
|
extern int CPUTLBEntry_wrong_size[sizeof(CPUTLBEntry) == (1 << CPU_TLB_ENTRY_BITS) ? 1 : -1];
|
|
|
|
#define CPU_COMMON_TLB \
|
|
/* The meaning of the MMU modes is defined in the target code. */ \
|
|
CPUTLBEntry tlb_table[NB_MMU_MODES][CPU_TLB_SIZE]; \
|
|
target_phys_addr_t iotlb[NB_MMU_MODES][CPU_TLB_SIZE]; \
|
|
target_ulong tlb_flush_addr; \
|
|
target_ulong tlb_flush_mask;
|
|
|
|
#else
|
|
|
|
#define CPU_COMMON_TLB
|
|
|
|
#endif
|
|
|
|
|
|
#ifdef HOST_WORDS_BIGENDIAN
|
|
typedef struct icount_decr_u16 {
|
|
uint16_t high;
|
|
uint16_t low;
|
|
} icount_decr_u16;
|
|
#else
|
|
typedef struct icount_decr_u16 {
|
|
uint16_t low;
|
|
uint16_t high;
|
|
} icount_decr_u16;
|
|
#endif
|
|
|
|
struct kvm_run;
|
|
struct KVMState;
|
|
struct qemu_work_item;
|
|
|
|
typedef struct CPUBreakpoint {
|
|
target_ulong pc;
|
|
int flags; /* BP_* */
|
|
QTAILQ_ENTRY(CPUBreakpoint) entry;
|
|
} CPUBreakpoint;
|
|
|
|
typedef struct CPUWatchpoint {
|
|
target_ulong vaddr;
|
|
target_ulong len_mask;
|
|
int flags; /* BP_* */
|
|
QTAILQ_ENTRY(CPUWatchpoint) entry;
|
|
} CPUWatchpoint;
|
|
|
|
#ifdef _WIN32
|
|
#define CPU_COMMON_THREAD \
|
|
void *hThread;
|
|
|
|
#else
|
|
#define CPU_COMMON_THREAD
|
|
#endif
|
|
|
|
#define CPU_TEMP_BUF_NLONGS 128
|
|
#define CPU_COMMON \
|
|
struct TranslationBlock *current_tb; /* currently executing TB */ \
|
|
/* soft mmu support */ \
|
|
/* in order to avoid passing too many arguments to the MMIO \
|
|
helpers, we store some rarely used information in the CPU \
|
|
context) */ \
|
|
uintptr_t mem_io_pc; /* host pc at which the memory was \
|
|
accessed */ \
|
|
target_ulong mem_io_vaddr; /* target virtual addr at which the \
|
|
memory was accessed */ \
|
|
uint32_t halted; /* Nonzero if the CPU is in suspend state */ \
|
|
uint32_t interrupt_request; \
|
|
volatile sig_atomic_t exit_request; \
|
|
CPU_COMMON_TLB \
|
|
struct TranslationBlock *tb_jmp_cache[TB_JMP_CACHE_SIZE]; \
|
|
/* buffer for temporaries in the code generator */ \
|
|
long temp_buf[CPU_TEMP_BUF_NLONGS]; \
|
|
\
|
|
int64_t icount_extra; /* Instructions until next timer event. */ \
|
|
/* Number of cycles left, with interrupt flag in high bit. \
|
|
This allows a single read-compare-cbranch-write sequence to test \
|
|
for both decrementer underflow and exceptions. */ \
|
|
union { \
|
|
uint32_t u32; \
|
|
icount_decr_u16 u16; \
|
|
} icount_decr; \
|
|
uint32_t can_do_io; /* nonzero if memory mapped IO is safe. */ \
|
|
\
|
|
/* from this point: preserved by CPU reset */ \
|
|
/* ice debug support */ \
|
|
QTAILQ_HEAD(breakpoints_head, CPUBreakpoint) breakpoints; \
|
|
int singlestep_enabled; \
|
|
\
|
|
QTAILQ_HEAD(watchpoints_head, CPUWatchpoint) watchpoints; \
|
|
CPUWatchpoint *watchpoint_hit; \
|
|
\
|
|
struct GDBRegisterState *gdb_regs; \
|
|
\
|
|
/* Core interrupt code */ \
|
|
jmp_buf jmp_env; \
|
|
int exception_index; \
|
|
\
|
|
CPUArchState *next_cpu; /* next CPU sharing TB cache */ \
|
|
int cpu_index; /* CPU index (informative) */ \
|
|
uint32_t host_tid; /* host thread ID */ \
|
|
int numa_node; /* NUMA node this cpu is belonging to */ \
|
|
int nr_cores; /* number of cores within this CPU package */ \
|
|
int nr_threads;/* number of threads within this CPU */ \
|
|
int running; /* Nonzero if cpu is currently running(usermode). */ \
|
|
int thread_id; \
|
|
/* user data */ \
|
|
void *opaque; \
|
|
\
|
|
uint32_t created; \
|
|
uint32_t stop; /* Stop request */ \
|
|
uint32_t stopped; /* Artificially stopped */ \
|
|
struct QemuThread *thread; \
|
|
CPU_COMMON_THREAD \
|
|
struct QemuCond *halt_cond; \
|
|
int thread_kicked; \
|
|
struct qemu_work_item *queued_work_first, *queued_work_last; \
|
|
const char *cpu_model_str; \
|
|
struct KVMState *kvm_state; \
|
|
struct kvm_run *kvm_run; \
|
|
int kvm_fd; \
|
|
int kvm_vcpu_dirty;
|
|
|
|
#endif
|