blob: 67ca72e045cec2bd1d23e3d62e583507143c8761 [file] [log] [blame]
pbrook87ecb682007-11-17 17:14:51 +00001#ifndef QEMU_TIMER_H
2#define QEMU_TIMER_H
3
Blue Swirl29e922b2010-03-29 19:24:00 +00004#include "qemu-common.h"
Paolo Bonzini44a9b352011-09-12 16:44:30 +02005#include "main-loop.h"
Jan Kiszka691a0c92011-06-20 14:06:27 +02006#include "notify.h"
Blue Swirlc57c8462010-10-23 15:24:07 +00007#include <time.h>
8#include <sys/time.h>
9
10#ifdef _WIN32
11#include <windows.h>
Blue Swirlc57c8462010-10-23 15:24:07 +000012#endif
Blue Swirl29e922b2010-03-29 19:24:00 +000013
pbrook87ecb682007-11-17 17:14:51 +000014/* timers */
15
Paolo Bonzini0ce1b942011-03-11 16:46:02 +010016#define SCALE_MS 1000000
17#define SCALE_US 1000
18#define SCALE_NS 1
19
pbrook87ecb682007-11-17 17:14:51 +000020typedef struct QEMUClock QEMUClock;
21typedef void QEMUTimerCB(void *opaque);
22
23/* The real time clock should be used only for stuff which does not
24 change the virtual machine state, as it is run even if the virtual
25 machine is stopped. The real time clock has a frequency of 1000
26 Hz. */
27extern QEMUClock *rt_clock;
28
29/* The virtual clock is only run during the emulation. It is stopped
30 when the virtual machine is stopped. Virtual timers use a high
31 precision clock, usually cpu cycles (use ticks_per_sec). */
32extern QEMUClock *vm_clock;
33
Jan Kiszka21d5d122009-09-15 13:36:04 +020034/* The host clock should be use for device models that emulate accurate
35 real time sources. It will continue to run when the virtual machine
36 is suspended, and it will reflect system time changes the host may
37 undergo (e.g. due to NTP). The host clock has the same precision as
38 the virtual clock. */
39extern QEMUClock *host_clock;
40
Paolo Bonzini41c872b2010-01-26 10:31:46 +020041int64_t qemu_get_clock_ns(QEMUClock *clock);
Paolo Bonzinidc2dfcf2011-09-12 15:50:16 +020042int64_t qemu_clock_has_timers(QEMUClock *clock);
43int64_t qemu_clock_expired(QEMUClock *clock);
44int64_t qemu_clock_deadline(QEMUClock *clock);
Paolo Bonzinidb1a4972010-03-10 11:38:55 +010045void qemu_clock_enable(QEMUClock *clock, int enabled);
Paolo Bonziniab33fcd2011-04-13 10:03:44 +020046void qemu_clock_warp(QEMUClock *clock);
pbrook87ecb682007-11-17 17:14:51 +000047
Jan Kiszka691a0c92011-06-20 14:06:27 +020048void qemu_register_clock_reset_notifier(QEMUClock *clock, Notifier *notifier);
49void qemu_unregister_clock_reset_notifier(QEMUClock *clock,
50 Notifier *notifier);
51
Paolo Bonzini4a998742011-03-11 16:33:58 +010052QEMUTimer *qemu_new_timer(QEMUClock *clock, int scale,
53 QEMUTimerCB *cb, void *opaque);
pbrook87ecb682007-11-17 17:14:51 +000054void qemu_free_timer(QEMUTimer *ts);
55void qemu_del_timer(QEMUTimer *ts);
Paolo Bonzini2ff68d02011-09-12 16:21:44 +020056void qemu_mod_timer_ns(QEMUTimer *ts, int64_t expire_time);
pbrook87ecb682007-11-17 17:14:51 +000057void qemu_mod_timer(QEMUTimer *ts, int64_t expire_time);
58int qemu_timer_pending(QEMUTimer *ts);
Stefano Stabellini2430ffe2009-08-03 10:56:01 +010059int qemu_timer_expired(QEMUTimer *timer_head, int64_t current_time);
Paolo Bonzini2ff68d02011-09-12 16:21:44 +020060uint64_t qemu_timer_expire_time_ns(QEMUTimer *ts);
pbrook87ecb682007-11-17 17:14:51 +000061
Paolo Bonzinidb1a4972010-03-10 11:38:55 +010062void qemu_run_all_timers(void);
63int qemu_alarm_pending(void);
Paolo Bonzinidb1a4972010-03-10 11:38:55 +010064void configure_alarms(char const *opt);
Paolo Bonzinidb1a4972010-03-10 11:38:55 +010065int qemu_calculate_timeout(void);
66void init_clocks(void);
67int init_timer_alarm(void);
Paolo Bonzinidb1a4972010-03-10 11:38:55 +010068
Blue Swirl70c3b552011-03-27 15:45:39 +000069int64_t cpu_get_ticks(void);
70void cpu_enable_ticks(void);
71void cpu_disable_ticks(void);
72
Paolo Bonzini0ce1b942011-03-11 16:46:02 +010073static inline QEMUTimer *qemu_new_timer_ns(QEMUClock *clock, QEMUTimerCB *cb,
74 void *opaque)
75{
Paolo Bonzini4a998742011-03-11 16:33:58 +010076 return qemu_new_timer(clock, SCALE_NS, cb, opaque);
Paolo Bonzini0ce1b942011-03-11 16:46:02 +010077}
78
79static inline QEMUTimer *qemu_new_timer_ms(QEMUClock *clock, QEMUTimerCB *cb,
80 void *opaque)
81{
Paolo Bonzini4a998742011-03-11 16:33:58 +010082 return qemu_new_timer(clock, SCALE_MS, cb, opaque);
Paolo Bonzini0ce1b942011-03-11 16:46:02 +010083}
84
85static inline int64_t qemu_get_clock_ms(QEMUClock *clock)
86{
87 return qemu_get_clock_ns(clock) / SCALE_MS;
88}
89
Anthony Liguori274dfed2009-09-11 10:28:26 -050090static inline int64_t get_ticks_per_sec(void)
91{
92 return 1000000000LL;
93}
pbrook87ecb682007-11-17 17:14:51 +000094
Blue Swirlc57c8462010-10-23 15:24:07 +000095/* real time host monotonic timer */
96static inline int64_t get_clock_realtime(void)
97{
98 struct timeval tv;
99
100 gettimeofday(&tv, NULL);
101 return tv.tv_sec * 1000000000LL + (tv.tv_usec * 1000);
102}
103
104/* Warning: don't insert tracepoints into these functions, they are
105 also used by simpletrace backend and tracepoints would cause
106 an infinite recursion! */
107#ifdef _WIN32
108extern int64_t clock_freq;
109
110static inline int64_t get_clock(void)
111{
112 LARGE_INTEGER ti;
113 QueryPerformanceCounter(&ti);
114 return muldiv64(ti.QuadPart, get_ticks_per_sec(), clock_freq);
115}
116
117#else
118
119extern int use_rt_clock;
120
121static inline int64_t get_clock(void)
122{
123#if defined(__linux__) || (defined(__FreeBSD__) && __FreeBSD_version >= 500000) \
124 || defined(__DragonFly__) || defined(__FreeBSD_kernel__)
125 if (use_rt_clock) {
126 struct timespec ts;
127 clock_gettime(CLOCK_MONOTONIC, &ts);
128 return ts.tv_sec * 1000000000LL + ts.tv_nsec;
129 } else
130#endif
131 {
132 /* XXX: using gettimeofday leads to problems if the date
133 changes, so it should be avoided. */
134 return get_clock_realtime();
135 }
136}
137#endif
Paolo Bonzinidb1a4972010-03-10 11:38:55 +0100138
pbrook87ecb682007-11-17 17:14:51 +0000139void qemu_get_timer(QEMUFile *f, QEMUTimer *ts);
140void qemu_put_timer(QEMUFile *f, QEMUTimer *ts);
141
142/* ptimer.c */
143typedef struct ptimer_state ptimer_state;
144typedef void (*ptimer_cb)(void *opaque);
145
146ptimer_state *ptimer_init(QEMUBH *bh);
147void ptimer_set_period(ptimer_state *s, int64_t period);
148void ptimer_set_freq(ptimer_state *s, uint32_t freq);
149void ptimer_set_limit(ptimer_state *s, uint64_t limit, int reload);
150uint64_t ptimer_get_count(ptimer_state *s);
151void ptimer_set_count(ptimer_state *s, uint64_t count);
152void ptimer_run(ptimer_state *s, int oneshot);
153void ptimer_stop(ptimer_state *s);
pbrook87ecb682007-11-17 17:14:51 +0000154
Blue Swirl29e922b2010-03-29 19:24:00 +0000155/* icount */
Blue Swirl29e922b2010-03-29 19:24:00 +0000156int64_t cpu_get_icount(void);
Paolo Bonzini946fb272011-09-12 13:57:37 +0200157int64_t cpu_get_clock(void);
Blue Swirl29e922b2010-03-29 19:24:00 +0000158
159/*******************************************/
160/* host CPU ticks (if available) */
161
162#if defined(_ARCH_PPC)
163
164static inline int64_t cpu_get_real_ticks(void)
165{
166 int64_t retval;
167#ifdef _ARCH_PPC64
168 /* This reads timebase in one 64bit go and includes Cell workaround from:
169 http://ozlabs.org/pipermail/linuxppc-dev/2006-October/027052.html
170 */
171 __asm__ __volatile__ ("mftb %0\n\t"
172 "cmpwi %0,0\n\t"
173 "beq- $-8"
174 : "=r" (retval));
175#else
176 /* http://ozlabs.org/pipermail/linuxppc-dev/1999-October/003889.html */
177 unsigned long junk;
Alexander Graf4a9590f2010-04-03 11:37:26 +0200178 __asm__ __volatile__ ("mfspr %1,269\n\t" /* mftbu */
179 "mfspr %L0,268\n\t" /* mftb */
180 "mfspr %0,269\n\t" /* mftbu */
Blue Swirl29e922b2010-03-29 19:24:00 +0000181 "cmpw %0,%1\n\t"
182 "bne $-16"
183 : "=r" (retval), "=r" (junk));
184#endif
185 return retval;
186}
187
188#elif defined(__i386__)
189
190static inline int64_t cpu_get_real_ticks(void)
191{
192 int64_t val;
193 asm volatile ("rdtsc" : "=A" (val));
194 return val;
195}
196
197#elif defined(__x86_64__)
198
199static inline int64_t cpu_get_real_ticks(void)
200{
201 uint32_t low,high;
202 int64_t val;
203 asm volatile("rdtsc" : "=a" (low), "=d" (high));
204 val = high;
205 val <<= 32;
206 val |= low;
207 return val;
208}
209
210#elif defined(__hppa__)
211
212static inline int64_t cpu_get_real_ticks(void)
213{
214 int val;
215 asm volatile ("mfctl %%cr16, %0" : "=r"(val));
216 return val;
217}
218
219#elif defined(__ia64)
220
221static inline int64_t cpu_get_real_ticks(void)
222{
223 int64_t val;
224 asm volatile ("mov %0 = ar.itc" : "=r"(val) :: "memory");
225 return val;
226}
227
228#elif defined(__s390__)
229
230static inline int64_t cpu_get_real_ticks(void)
231{
232 int64_t val;
233 asm volatile("stck 0(%1)" : "=m" (val) : "a" (&val) : "cc");
234 return val;
235}
236
237#elif defined(__sparc_v8plus__) || defined(__sparc_v8plusa__) || defined(__sparc_v9__)
238
239static inline int64_t cpu_get_real_ticks (void)
240{
241#if defined(_LP64)
242 uint64_t rval;
243 asm volatile("rd %%tick,%0" : "=r"(rval));
244 return rval;
245#else
246 union {
247 uint64_t i64;
248 struct {
249 uint32_t high;
250 uint32_t low;
251 } i32;
252 } rval;
253 asm volatile("rd %%tick,%1; srlx %1,32,%0"
254 : "=r"(rval.i32.high), "=r"(rval.i32.low));
255 return rval.i64;
256#endif
257}
258
259#elif defined(__mips__) && \
260 ((defined(__mips_isa_rev) && __mips_isa_rev >= 2) || defined(__linux__))
261/*
262 * binutils wants to use rdhwr only on mips32r2
263 * but as linux kernel emulate it, it's fine
264 * to use it.
265 *
266 */
267#define MIPS_RDHWR(rd, value) { \
268 __asm__ __volatile__ (".set push\n\t" \
269 ".set mips32r2\n\t" \
270 "rdhwr %0, "rd"\n\t" \
271 ".set pop" \
272 : "=r" (value)); \
273 }
274
275static inline int64_t cpu_get_real_ticks(void)
276{
277 /* On kernels >= 2.6.25 rdhwr <reg>, $2 and $3 are emulated */
278 uint32_t count;
279 static uint32_t cyc_per_count = 0;
280
281 if (!cyc_per_count) {
282 MIPS_RDHWR("$3", cyc_per_count);
283 }
284
285 MIPS_RDHWR("$2", count);
286 return (int64_t)(count * cyc_per_count);
287}
288
Richard Henderson14a60632010-04-12 16:19:26 -0700289#elif defined(__alpha__)
290
291static inline int64_t cpu_get_real_ticks(void)
292{
293 uint64_t cc;
294 uint32_t cur, ofs;
295
296 asm volatile("rpcc %0" : "=r"(cc));
297 cur = cc;
298 ofs = cc >> 32;
299 return cur - ofs;
300}
301
Blue Swirl29e922b2010-03-29 19:24:00 +0000302#else
303/* The host CPU doesn't have an easily accessible cycle counter.
304 Just return a monotonically increasing value. This will be
305 totally wrong, but hopefully better than nothing. */
306static inline int64_t cpu_get_real_ticks (void)
307{
308 static int64_t ticks = 0;
309 return ticks++;
310}
311#endif
312
Richard Henderson2d8ebcf2010-04-17 16:25:10 +0000313#ifdef CONFIG_PROFILER
314static inline int64_t profile_getclock(void)
315{
316 return cpu_get_real_ticks();
317}
318
319extern int64_t qemu_time, qemu_time_start;
320extern int64_t tlb_flush_time;
321extern int64_t dev_time;
322#endif
323
pbrook87ecb682007-11-17 17:14:51 +0000324#endif