2007-11-17 18:14:51 +01:00
|
|
|
#ifndef QEMU_TIMER_H
|
|
|
|
#define QEMU_TIMER_H
|
|
|
|
|
2010-03-29 21:24:00 +02:00
|
|
|
#include "qemu-common.h"
|
|
|
|
|
2007-11-17 18:14:51 +01:00
|
|
|
/* timers */
|
|
|
|
|
|
|
|
typedef struct QEMUClock QEMUClock;
|
|
|
|
typedef void QEMUTimerCB(void *opaque);
|
|
|
|
|
|
|
|
/* The real time clock should be used only for stuff which does not
|
|
|
|
change the virtual machine state, as it is run even if the virtual
|
|
|
|
machine is stopped. The real time clock has a frequency of 1000
|
|
|
|
Hz. */
|
|
|
|
extern QEMUClock *rt_clock;
|
|
|
|
|
|
|
|
/* The virtual clock is only run during the emulation. It is stopped
|
|
|
|
when the virtual machine is stopped. Virtual timers use a high
|
|
|
|
precision clock, usually cpu cycles (use ticks_per_sec). */
|
|
|
|
extern QEMUClock *vm_clock;
|
|
|
|
|
2009-09-15 13:36:04 +02:00
|
|
|
/* The host clock should be use for device models that emulate accurate
|
|
|
|
real time sources. It will continue to run when the virtual machine
|
|
|
|
is suspended, and it will reflect system time changes the host may
|
|
|
|
undergo (e.g. due to NTP). The host clock has the same precision as
|
|
|
|
the virtual clock. */
|
|
|
|
extern QEMUClock *host_clock;
|
|
|
|
|
2007-11-17 18:14:51 +01:00
|
|
|
int64_t qemu_get_clock(QEMUClock *clock);
|
2010-01-26 09:31:46 +01:00
|
|
|
int64_t qemu_get_clock_ns(QEMUClock *clock);
|
2010-03-10 11:38:55 +01:00
|
|
|
void qemu_clock_enable(QEMUClock *clock, int enabled);
|
2007-11-17 18:14:51 +01:00
|
|
|
|
|
|
|
QEMUTimer *qemu_new_timer(QEMUClock *clock, QEMUTimerCB *cb, void *opaque);
|
|
|
|
void qemu_free_timer(QEMUTimer *ts);
|
|
|
|
void qemu_del_timer(QEMUTimer *ts);
|
|
|
|
void qemu_mod_timer(QEMUTimer *ts, int64_t expire_time);
|
|
|
|
int qemu_timer_pending(QEMUTimer *ts);
|
2009-08-03 11:56:01 +02:00
|
|
|
int qemu_timer_expired(QEMUTimer *timer_head, int64_t current_time);
|
2007-11-17 18:14:51 +01:00
|
|
|
|
2010-03-10 11:38:55 +01:00
|
|
|
void qemu_run_all_timers(void);
|
|
|
|
int qemu_alarm_pending(void);
|
|
|
|
int64_t qemu_next_deadline(void);
|
|
|
|
void configure_alarms(char const *opt);
|
|
|
|
void configure_icount(const char *option);
|
|
|
|
int qemu_calculate_timeout(void);
|
|
|
|
void init_clocks(void);
|
|
|
|
int init_timer_alarm(void);
|
|
|
|
void quit_timers(void);
|
|
|
|
|
2009-09-11 17:28:26 +02:00
|
|
|
static inline int64_t get_ticks_per_sec(void)
|
|
|
|
{
|
|
|
|
return 1000000000LL;
|
|
|
|
}
|
2007-11-17 18:14:51 +01:00
|
|
|
|
2010-03-10 11:38:55 +01:00
|
|
|
|
2007-11-17 18:14:51 +01:00
|
|
|
void qemu_get_timer(QEMUFile *f, QEMUTimer *ts);
|
|
|
|
void qemu_put_timer(QEMUFile *f, QEMUTimer *ts);
|
|
|
|
|
|
|
|
/* ptimer.c */
|
|
|
|
typedef struct ptimer_state ptimer_state;
|
|
|
|
typedef void (*ptimer_cb)(void *opaque);
|
|
|
|
|
|
|
|
ptimer_state *ptimer_init(QEMUBH *bh);
|
|
|
|
void ptimer_set_period(ptimer_state *s, int64_t period);
|
|
|
|
void ptimer_set_freq(ptimer_state *s, uint32_t freq);
|
|
|
|
void ptimer_set_limit(ptimer_state *s, uint64_t limit, int reload);
|
|
|
|
uint64_t ptimer_get_count(ptimer_state *s);
|
|
|
|
void ptimer_set_count(ptimer_state *s, uint64_t count);
|
|
|
|
void ptimer_run(ptimer_state *s, int oneshot);
|
|
|
|
void ptimer_stop(ptimer_state *s);
|
|
|
|
void qemu_put_ptimer(QEMUFile *f, ptimer_state *s);
|
|
|
|
void qemu_get_ptimer(QEMUFile *f, ptimer_state *s);
|
|
|
|
|
2010-03-29 21:24:00 +02:00
|
|
|
/* icount */
|
|
|
|
int64_t qemu_icount_round(int64_t count);
|
|
|
|
extern int64_t qemu_icount;
|
|
|
|
extern int use_icount;
|
|
|
|
extern int icount_time_shift;
|
|
|
|
extern int64_t qemu_icount_bias;
|
|
|
|
int64_t cpu_get_icount(void);
|
|
|
|
|
|
|
|
/*******************************************/
|
|
|
|
/* host CPU ticks (if available) */
|
|
|
|
|
|
|
|
#if defined(_ARCH_PPC)
|
|
|
|
|
|
|
|
static inline int64_t cpu_get_real_ticks(void)
|
|
|
|
{
|
|
|
|
int64_t retval;
|
|
|
|
#ifdef _ARCH_PPC64
|
|
|
|
/* This reads timebase in one 64bit go and includes Cell workaround from:
|
|
|
|
http://ozlabs.org/pipermail/linuxppc-dev/2006-October/027052.html
|
|
|
|
*/
|
|
|
|
__asm__ __volatile__ ("mftb %0\n\t"
|
|
|
|
"cmpwi %0,0\n\t"
|
|
|
|
"beq- $-8"
|
|
|
|
: "=r" (retval));
|
|
|
|
#else
|
|
|
|
/* http://ozlabs.org/pipermail/linuxppc-dev/1999-October/003889.html */
|
|
|
|
unsigned long junk;
|
|
|
|
__asm__ __volatile__ ("mftbu %1\n\t"
|
|
|
|
"mftb %L0\n\t"
|
|
|
|
"mftbu %0\n\t"
|
|
|
|
"cmpw %0,%1\n\t"
|
|
|
|
"bne $-16"
|
|
|
|
: "=r" (retval), "=r" (junk));
|
|
|
|
#endif
|
|
|
|
return retval;
|
|
|
|
}
|
|
|
|
|
|
|
|
#elif defined(__i386__)
|
|
|
|
|
|
|
|
static inline int64_t cpu_get_real_ticks(void)
|
|
|
|
{
|
|
|
|
int64_t val;
|
|
|
|
asm volatile ("rdtsc" : "=A" (val));
|
|
|
|
return val;
|
|
|
|
}
|
|
|
|
|
|
|
|
#elif defined(__x86_64__)
|
|
|
|
|
|
|
|
static inline int64_t cpu_get_real_ticks(void)
|
|
|
|
{
|
|
|
|
uint32_t low,high;
|
|
|
|
int64_t val;
|
|
|
|
asm volatile("rdtsc" : "=a" (low), "=d" (high));
|
|
|
|
val = high;
|
|
|
|
val <<= 32;
|
|
|
|
val |= low;
|
|
|
|
return val;
|
|
|
|
}
|
|
|
|
|
|
|
|
#elif defined(__hppa__)
|
|
|
|
|
|
|
|
static inline int64_t cpu_get_real_ticks(void)
|
|
|
|
{
|
|
|
|
int val;
|
|
|
|
asm volatile ("mfctl %%cr16, %0" : "=r"(val));
|
|
|
|
return val;
|
|
|
|
}
|
|
|
|
|
|
|
|
#elif defined(__ia64)
|
|
|
|
|
|
|
|
static inline int64_t cpu_get_real_ticks(void)
|
|
|
|
{
|
|
|
|
int64_t val;
|
|
|
|
asm volatile ("mov %0 = ar.itc" : "=r"(val) :: "memory");
|
|
|
|
return val;
|
|
|
|
}
|
|
|
|
|
|
|
|
#elif defined(__s390__)
|
|
|
|
|
|
|
|
static inline int64_t cpu_get_real_ticks(void)
|
|
|
|
{
|
|
|
|
int64_t val;
|
|
|
|
asm volatile("stck 0(%1)" : "=m" (val) : "a" (&val) : "cc");
|
|
|
|
return val;
|
|
|
|
}
|
|
|
|
|
|
|
|
#elif defined(__sparc_v8plus__) || defined(__sparc_v8plusa__) || defined(__sparc_v9__)
|
|
|
|
|
|
|
|
static inline int64_t cpu_get_real_ticks (void)
|
|
|
|
{
|
|
|
|
#if defined(_LP64)
|
|
|
|
uint64_t rval;
|
|
|
|
asm volatile("rd %%tick,%0" : "=r"(rval));
|
|
|
|
return rval;
|
|
|
|
#else
|
|
|
|
union {
|
|
|
|
uint64_t i64;
|
|
|
|
struct {
|
|
|
|
uint32_t high;
|
|
|
|
uint32_t low;
|
|
|
|
} i32;
|
|
|
|
} rval;
|
|
|
|
asm volatile("rd %%tick,%1; srlx %1,32,%0"
|
|
|
|
: "=r"(rval.i32.high), "=r"(rval.i32.low));
|
|
|
|
return rval.i64;
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
|
|
|
#elif defined(__mips__) && \
|
|
|
|
((defined(__mips_isa_rev) && __mips_isa_rev >= 2) || defined(__linux__))
|
|
|
|
/*
|
|
|
|
* binutils wants to use rdhwr only on mips32r2
|
|
|
|
* but as linux kernel emulate it, it's fine
|
|
|
|
* to use it.
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
#define MIPS_RDHWR(rd, value) { \
|
|
|
|
__asm__ __volatile__ (".set push\n\t" \
|
|
|
|
".set mips32r2\n\t" \
|
|
|
|
"rdhwr %0, "rd"\n\t" \
|
|
|
|
".set pop" \
|
|
|
|
: "=r" (value)); \
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int64_t cpu_get_real_ticks(void)
|
|
|
|
{
|
|
|
|
/* On kernels >= 2.6.25 rdhwr <reg>, $2 and $3 are emulated */
|
|
|
|
uint32_t count;
|
|
|
|
static uint32_t cyc_per_count = 0;
|
|
|
|
|
|
|
|
if (!cyc_per_count) {
|
|
|
|
MIPS_RDHWR("$3", cyc_per_count);
|
|
|
|
}
|
|
|
|
|
|
|
|
MIPS_RDHWR("$2", count);
|
|
|
|
return (int64_t)(count * cyc_per_count);
|
|
|
|
}
|
|
|
|
|
|
|
|
#else
|
|
|
|
/* The host CPU doesn't have an easily accessible cycle counter.
|
|
|
|
Just return a monotonically increasing value. This will be
|
|
|
|
totally wrong, but hopefully better than nothing. */
|
|
|
|
static inline int64_t cpu_get_real_ticks (void)
|
|
|
|
{
|
|
|
|
static int64_t ticks = 0;
|
|
|
|
return ticks++;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#ifdef NEED_CPU_H
|
|
|
|
/* Deterministic execution requires that IO only be performed on the last
|
|
|
|
instruction of a TB so that interrupts take effect immediately. */
|
|
|
|
static inline int can_do_io(CPUState *env)
|
|
|
|
{
|
|
|
|
if (!use_icount)
|
|
|
|
return 1;
|
|
|
|
|
|
|
|
/* If not executing code then assume we are ok. */
|
|
|
|
if (!env->current_tb)
|
|
|
|
return 1;
|
|
|
|
|
|
|
|
return env->can_do_io != 0;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2007-11-17 18:14:51 +01:00
|
|
|
#endif
|