• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 #ifndef QEMU_TIMER_H
2 #define QEMU_TIMER_H
3 
4 #include "qemu-common.h"
5 #include <time.h>
6 #include <sys/time.h>
7 
8 #ifdef _WIN32
9 #include <windows.h>
10 #endif
11 
12 /* timers */
13 
14 #define SCALE_MS 1000000
15 #define SCALE_US 1000
16 #define SCALE_NS 1
17 
18 typedef struct QEMUClock QEMUClock;
19 typedef void QEMUTimerCB(void *opaque);
20 
21 /* The real time clock should be used only for stuff which does not
22    change the virtual machine state, as it is run even if the virtual
23    machine is stopped. The real time clock has a frequency of 1000
24    Hz. */
25 extern QEMUClock *rt_clock;
26 
27 /* The virtual clock is only run during the emulation. It is stopped
28    when the virtual machine is stopped. Virtual timers use a high
29    precision clock, usually cpu cycles (use ticks_per_sec). */
30 extern QEMUClock *vm_clock;
31 
32 /* The host clock should be use for device models that emulate accurate
33    real time sources. It will continue to run when the virtual machine
34    is suspended, and it will reflect system time changes the host may
35    undergo (e.g. due to NTP). The host clock has the same precision as
36    the virtual clock. */
37 extern QEMUClock *host_clock;
38 
39 int64_t qemu_get_clock(QEMUClock *clock);
40 int64_t qemu_get_clock_ns(QEMUClock *clock);
41 void qemu_clock_enable(QEMUClock *clock, int enabled);
42 void qemu_clock_warp(QEMUClock *clock);
43 
44 QEMUTimer *qemu_new_timer(QEMUClock *clock, int scale,
45                           QEMUTimerCB *cb, void *opaque);
46 
47 void qemu_free_timer(QEMUTimer *ts);
48 void qemu_del_timer(QEMUTimer *ts);
49 void qemu_mod_timer(QEMUTimer *ts, int64_t expire_time);
50 int qemu_timer_pending(QEMUTimer *ts);
51 int qemu_timer_expired(QEMUTimer *timer_head, int64_t current_time);
52 int qemu_timer_alarm_pending(void);
53 
54 void qemu_run_all_timers(void);
55 int qemu_alarm_pending(void);
56 int64_t qemu_next_icount_deadline(void);
57 int64_t qemu_next_deadline(void);
58 void configure_alarms(char const *opt);
59 void configure_icount(const char *option);
60 int qemu_calculate_timeout(void);
61 void init_clocks(void);
62 int init_timer_alarm(void);
63 void quit_timers(void);
64 
65 int64_t cpu_get_ticks(void);
66 void cpu_enable_ticks(void);
67 void cpu_disable_ticks(void);
68 
qemu_new_timer_ns(QEMUClock * clock,QEMUTimerCB * cb,void * opaque)69 static inline QEMUTimer *qemu_new_timer_ns(QEMUClock *clock, QEMUTimerCB *cb,
70                                            void *opaque)
71 {
72     return qemu_new_timer(clock, SCALE_NS, cb, opaque);
73 }
74 
qemu_new_timer_ms(QEMUClock * clock,QEMUTimerCB * cb,void * opaque)75 static inline QEMUTimer *qemu_new_timer_ms(QEMUClock *clock, QEMUTimerCB *cb,
76                                            void *opaque)
77 {
78     return qemu_new_timer(clock, SCALE_MS, cb, opaque);
79 }
80 
qemu_get_clock_ms(QEMUClock * clock)81 static inline int64_t qemu_get_clock_ms(QEMUClock *clock)
82 {
83     return qemu_get_clock_ns(clock) / SCALE_MS;
84 }
85 
get_ticks_per_sec(void)86 static inline int64_t get_ticks_per_sec(void)
87 {
88     return 1000000000LL;
89 }
90 
91 /* real time host monotonic timer */
get_clock_realtime(void)92 static inline int64_t get_clock_realtime(void)
93 {
94     struct timeval tv;
95 
96     gettimeofday(&tv, NULL);
97     return tv.tv_sec * 1000000000LL + (tv.tv_usec * 1000);
98 }
99 
100 /* Warning: don't insert tracepoints into these functions, they are
101    also used by simpletrace backend and tracepoints would cause
102    an infinite recursion! */
103 #ifdef _WIN32
104 extern int64_t clock_freq;
105 
get_clock(void)106 static inline int64_t get_clock(void)
107 {
108     LARGE_INTEGER ti;
109     QueryPerformanceCounter(&ti);
110     return muldiv64(ti.QuadPart, get_ticks_per_sec(), clock_freq);
111 }
112 
113 #else
114 
115 extern int use_rt_clock;
116 
get_clock(void)117 static inline int64_t get_clock(void)
118 {
119 #if defined(__linux__) || (defined(__FreeBSD__) && __FreeBSD_version >= 500000) \
120     || defined(__DragonFly__) || defined(__FreeBSD_kernel__)
121     if (use_rt_clock) {
122         struct timespec ts;
123         clock_gettime(CLOCK_MONOTONIC, &ts);
124         return ts.tv_sec * 1000000000LL + ts.tv_nsec;
125     } else
126 #endif
127     {
128         /* XXX: using gettimeofday leads to problems if the date
129            changes, so it should be avoided. */
130         return get_clock_realtime();
131     }
132 }
133 #endif
134 
135 void qemu_get_timer(QEMUFile *f, QEMUTimer *ts);
136 void qemu_put_timer(QEMUFile *f, QEMUTimer *ts);
137 
138 /* ptimer.c */
139 typedef struct ptimer_state ptimer_state;
140 typedef void (*ptimer_cb)(void *opaque);
141 
142 ptimer_state *ptimer_init(QEMUBH *bh);
143 void ptimer_set_period(ptimer_state *s, int64_t period);
144 void ptimer_set_freq(ptimer_state *s, uint32_t freq);
145 void ptimer_set_limit(ptimer_state *s, uint64_t limit, int reload);
146 uint64_t ptimer_get_count(ptimer_state *s);
147 void ptimer_set_count(ptimer_state *s, uint64_t count);
148 void ptimer_run(ptimer_state *s, int oneshot);
149 void ptimer_stop(ptimer_state *s);
150 void qemu_put_ptimer(QEMUFile *f, ptimer_state *s);
151 void qemu_get_ptimer(QEMUFile *f, ptimer_state *s);
152 
153 /* icount */
154 int64_t qemu_icount_round(int64_t count);
155 extern int64_t qemu_icount;
156 extern int use_icount;
157 extern int icount_time_shift;
158 extern int64_t qemu_icount_bias;
159 int64_t cpu_get_icount(void);
160 
161 /*******************************************/
162 /* host CPU ticks (if available) */
163 
164 #if defined(_ARCH_PPC)
165 
cpu_get_real_ticks(void)166 static inline int64_t cpu_get_real_ticks(void)
167 {
168     int64_t retval;
169 #ifdef _ARCH_PPC64
170     /* This reads timebase in one 64bit go and includes Cell workaround from:
171        http://ozlabs.org/pipermail/linuxppc-dev/2006-October/027052.html
172     */
173     __asm__ __volatile__ ("mftb    %0\n\t"
174                           "cmpwi   %0,0\n\t"
175                           "beq-    $-8"
176                           : "=r" (retval));
177 #else
178     /* http://ozlabs.org/pipermail/linuxppc-dev/1999-October/003889.html */
179     unsigned long junk;
180     __asm__ __volatile__ ("mfspr   %1,269\n\t"  /* mftbu */
181                           "mfspr   %L0,268\n\t" /* mftb */
182                           "mfspr   %0,269\n\t"  /* mftbu */
183                           "cmpw    %0,%1\n\t"
184                           "bne     $-16"
185                           : "=r" (retval), "=r" (junk));
186 #endif
187     return retval;
188 }
189 
190 #elif defined(__i386__)
191 
cpu_get_real_ticks(void)192 static inline int64_t cpu_get_real_ticks(void)
193 {
194     int64_t val;
195     asm volatile ("rdtsc" : "=A" (val));
196     return val;
197 }
198 
199 #elif defined(__x86_64__)
200 
cpu_get_real_ticks(void)201 static inline int64_t cpu_get_real_ticks(void)
202 {
203     uint32_t low,high;
204     int64_t val;
205     asm volatile("rdtsc" : "=a" (low), "=d" (high));
206     val = high;
207     val <<= 32;
208     val |= low;
209     return val;
210 }
211 
212 #elif defined(__hppa__)
213 
cpu_get_real_ticks(void)214 static inline int64_t cpu_get_real_ticks(void)
215 {
216     int val;
217     asm volatile ("mfctl %%cr16, %0" : "=r"(val));
218     return val;
219 }
220 
221 #elif defined(__ia64)
222 
cpu_get_real_ticks(void)223 static inline int64_t cpu_get_real_ticks(void)
224 {
225     int64_t val;
226     asm volatile ("mov %0 = ar.itc" : "=r"(val) :: "memory");
227     return val;
228 }
229 
230 #elif defined(__s390__)
231 
cpu_get_real_ticks(void)232 static inline int64_t cpu_get_real_ticks(void)
233 {
234     int64_t val;
235     asm volatile("stck 0(%1)" : "=m" (val) : "a" (&val) : "cc");
236     return val;
237 }
238 
239 #elif defined(__sparc_v8plus__) || defined(__sparc_v8plusa__) || defined(__sparc_v9__)
240 
cpu_get_real_ticks(void)241 static inline int64_t cpu_get_real_ticks (void)
242 {
243 #if defined(_LP64)
244     uint64_t        rval;
245     asm volatile("rd %%tick,%0" : "=r"(rval));
246     return rval;
247 #else
248     union {
249         uint64_t i64;
250         struct {
251             uint32_t high;
252             uint32_t low;
253         }       i32;
254     } rval;
255     asm volatile("rd %%tick,%1; srlx %1,32,%0"
256                  : "=r"(rval.i32.high), "=r"(rval.i32.low));
257     return rval.i64;
258 #endif
259 }
260 
261 #elif defined(__mips__) && \
262     ((defined(__mips_isa_rev) && __mips_isa_rev >= 2) || defined(__linux__))
263 /*
264  * binutils wants to use rdhwr only on mips32r2
265  * but as linux kernel emulate it, it's fine
266  * to use it.
267  *
268  */
269 #define MIPS_RDHWR(rd, value) {                         \
270         __asm__ __volatile__ (".set   push\n\t"         \
271                               ".set mips32r2\n\t"       \
272                               "rdhwr  %0, "rd"\n\t"     \
273                               ".set   pop"              \
274                               : "=r" (value));          \
275     }
276 
cpu_get_real_ticks(void)277 static inline int64_t cpu_get_real_ticks(void)
278 {
279     /* On kernels >= 2.6.25 rdhwr <reg>, $2 and $3 are emulated */
280     uint32_t count;
281     static uint32_t cyc_per_count = 0;
282 
283     if (!cyc_per_count) {
284         MIPS_RDHWR("$3", cyc_per_count);
285     }
286 
287     MIPS_RDHWR("$2", count);
288     return (int64_t)(count * cyc_per_count);
289 }
290 
291 #elif defined(__alpha__)
292 
cpu_get_real_ticks(void)293 static inline int64_t cpu_get_real_ticks(void)
294 {
295     uint64_t cc;
296     uint32_t cur, ofs;
297 
298     asm volatile("rpcc %0" : "=r"(cc));
299     cur = cc;
300     ofs = cc >> 32;
301     return cur - ofs;
302 }
303 
304 #else
305 /* The host CPU doesn't have an easily accessible cycle counter.
306    Just return a monotonically increasing value.  This will be
307    totally wrong, but hopefully better than nothing.  */
cpu_get_real_ticks(void)308 static inline int64_t cpu_get_real_ticks (void)
309 {
310     static int64_t ticks = 0;
311     return ticks++;
312 }
313 #endif
314 
315 #ifdef NEED_CPU_H
316 /* Deterministic execution requires that IO only be performed on the last
317    instruction of a TB so that interrupts take effect immediately.  */
can_do_io(CPUState * env)318 static inline int can_do_io(CPUState *env)
319 {
320     if (!use_icount)
321         return 1;
322 
323     /* If not executing code then assume we are ok.  */
324     if (!env->current_tb)
325         return 1;
326 
327     return env->can_do_io != 0;
328 }
329 #endif
330 
331 #ifdef CONFIG_PROFILER
profile_getclock(void)332 static inline int64_t profile_getclock(void)
333 {
334     return cpu_get_real_ticks();
335 }
336 
337 extern int64_t qemu_time, qemu_time_start;
338 extern int64_t tlb_flush_time;
339 extern int64_t dev_time;
340 #endif
341 
342 #endif
343