 c57c846a80
			
		
	
	
		c57c846a80
		
	
	
	
	
		
			
			Move timer init functions to a new file, qemu-timer-common.c. Make other critical timer functions inlined to preserve performance in qemu-timer.c, also move muldiv64() (used by the inline functions) to qemu-timer.h. Adjust block/raw-posix.c and simpletrace.c to use get_clock() directly. Remove a similar/duplicate definition in qemu-tool.c. Adjust hw/omap_clk.c to include qemu-timer.h because muldiv64() is used there. After this change, tracing can be used also for user code and simpletrace on Win32. Cc: Stefan Hajnoczi <stefanha@linux.vnet.ibm.com> Acked-by: Stefan Hajnoczi <stefanha@linux.vnet.ibm.com> Signed-off-by: Blue Swirl <blauwirbel@gmail.com>
		
			
				
	
	
		
			1118 lines
		
	
	
		
			28 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			1118 lines
		
	
	
		
			28 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * QEMU System Emulator
 | |
|  *
 | |
|  * Copyright (c) 2003-2008 Fabrice Bellard
 | |
|  *
 | |
|  * Permission is hereby granted, free of charge, to any person obtaining a copy
 | |
|  * of this software and associated documentation files (the "Software"), to deal
 | |
|  * in the Software without restriction, including without limitation the rights
 | |
|  * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
 | |
|  * copies of the Software, and to permit persons to whom the Software is
 | |
|  * furnished to do so, subject to the following conditions:
 | |
|  *
 | |
|  * The above copyright notice and this permission notice shall be included in
 | |
|  * all copies or substantial portions of the Software.
 | |
|  *
 | |
|  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
 | |
|  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
 | |
|  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
 | |
|  * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
 | |
|  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
 | |
|  * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
 | |
|  * THE SOFTWARE.
 | |
|  */
 | |
| 
 | |
| #include "sysemu.h"
 | |
| #include "net.h"
 | |
| #include "monitor.h"
 | |
| #include "console.h"
 | |
| 
 | |
| #include "hw/hw.h"
 | |
| 
 | |
| #include <unistd.h>
 | |
| #include <fcntl.h>
 | |
| #include <time.h>
 | |
| #include <errno.h>
 | |
| #include <sys/time.h>
 | |
| #include <signal.h>
 | |
| #ifdef __FreeBSD__
 | |
| #include <sys/param.h>
 | |
| #endif
 | |
| 
 | |
| #ifdef __linux__
 | |
| #include <sys/ioctl.h>
 | |
| #include <linux/rtc.h>
 | |
| /* For the benefit of older linux systems which don't supply it,
 | |
|    we use a local copy of hpet.h. */
 | |
| /* #include <linux/hpet.h> */
 | |
| #include "hpet.h"
 | |
| #endif
 | |
| 
 | |
| #ifdef _WIN32
 | |
| #include <windows.h>
 | |
| #include <mmsystem.h>
 | |
| #endif
 | |
| 
 | |
| #include "qemu-timer.h"
 | |
| 
 | |
| /* Conversion factor from emulated instructions to virtual clock ticks.  */
 | |
| int icount_time_shift;
 | |
| /* Arbitrarily pick 1MIPS as the minimum allowable speed.  */
 | |
| #define MAX_ICOUNT_SHIFT 10
 | |
| /* Compensate for varying guest execution speed.  */
 | |
| int64_t qemu_icount_bias;
 | |
| static QEMUTimer *icount_rt_timer;
 | |
| static QEMUTimer *icount_vm_timer;
 | |
| 
 | |
| /***********************************************************/
 | |
| /* guest cycle counter */
 | |
| 
 | |
| typedef struct TimersState {
 | |
|     int64_t cpu_ticks_prev;
 | |
|     int64_t cpu_ticks_offset;
 | |
|     int64_t cpu_clock_offset;
 | |
|     int32_t cpu_ticks_enabled;
 | |
|     int64_t dummy;
 | |
| } TimersState;
 | |
| 
 | |
| TimersState timers_state;
 | |
| 
 | |
| /* return the host CPU cycle counter and handle stop/restart */
 | |
| int64_t cpu_get_ticks(void)
 | |
| {
 | |
|     if (use_icount) {
 | |
|         return cpu_get_icount();
 | |
|     }
 | |
|     if (!timers_state.cpu_ticks_enabled) {
 | |
|         return timers_state.cpu_ticks_offset;
 | |
|     } else {
 | |
|         int64_t ticks;
 | |
|         ticks = cpu_get_real_ticks();
 | |
|         if (timers_state.cpu_ticks_prev > ticks) {
 | |
|             /* Note: non increasing ticks may happen if the host uses
 | |
|                software suspend */
 | |
|             timers_state.cpu_ticks_offset += timers_state.cpu_ticks_prev - ticks;
 | |
|         }
 | |
|         timers_state.cpu_ticks_prev = ticks;
 | |
|         return ticks + timers_state.cpu_ticks_offset;
 | |
|     }
 | |
| }
 | |
| 
 | |
| /* return the host CPU monotonic timer and handle stop/restart */
 | |
| static int64_t cpu_get_clock(void)
 | |
| {
 | |
|     int64_t ti;
 | |
|     if (!timers_state.cpu_ticks_enabled) {
 | |
|         return timers_state.cpu_clock_offset;
 | |
|     } else {
 | |
|         ti = get_clock();
 | |
|         return ti + timers_state.cpu_clock_offset;
 | |
|     }
 | |
| }
 | |
| 
 | |
| #ifndef CONFIG_IOTHREAD
 | |
| static int64_t qemu_icount_delta(void)
 | |
| {
 | |
|     if (!use_icount) {
 | |
|         return 5000 * (int64_t) 1000000;
 | |
|     } else if (use_icount == 1) {
 | |
|         /* When not using an adaptive execution frequency
 | |
|            we tend to get badly out of sync with real time,
 | |
|            so just delay for a reasonable amount of time.  */
 | |
|         return 0;
 | |
|     } else {
 | |
|         return cpu_get_icount() - cpu_get_clock();
 | |
|     }
 | |
| }
 | |
| #endif
 | |
| 
 | |
| /* enable cpu_get_ticks() */
 | |
| void cpu_enable_ticks(void)
 | |
| {
 | |
|     if (!timers_state.cpu_ticks_enabled) {
 | |
|         timers_state.cpu_ticks_offset -= cpu_get_real_ticks();
 | |
|         timers_state.cpu_clock_offset -= get_clock();
 | |
|         timers_state.cpu_ticks_enabled = 1;
 | |
|     }
 | |
| }
 | |
| 
 | |
| /* disable cpu_get_ticks() : the clock is stopped. You must not call
 | |
|    cpu_get_ticks() after that.  */
 | |
| void cpu_disable_ticks(void)
 | |
| {
 | |
|     if (timers_state.cpu_ticks_enabled) {
 | |
|         timers_state.cpu_ticks_offset = cpu_get_ticks();
 | |
|         timers_state.cpu_clock_offset = cpu_get_clock();
 | |
|         timers_state.cpu_ticks_enabled = 0;
 | |
|     }
 | |
| }
 | |
| 
 | |
| /***********************************************************/
 | |
| /* timers */
 | |
| 
 | |
| #define QEMU_CLOCK_REALTIME 0
 | |
| #define QEMU_CLOCK_VIRTUAL  1
 | |
| #define QEMU_CLOCK_HOST     2
 | |
| 
 | |
| struct QEMUClock {
 | |
|     int type;
 | |
|     int enabled;
 | |
|     /* XXX: add frequency */
 | |
| };
 | |
| 
 | |
| struct QEMUTimer {
 | |
|     QEMUClock *clock;
 | |
|     int64_t expire_time;
 | |
|     QEMUTimerCB *cb;
 | |
|     void *opaque;
 | |
|     struct QEMUTimer *next;
 | |
| };
 | |
| 
 | |
| struct qemu_alarm_timer {
 | |
|     char const *name;
 | |
|     int (*start)(struct qemu_alarm_timer *t);
 | |
|     void (*stop)(struct qemu_alarm_timer *t);
 | |
|     void (*rearm)(struct qemu_alarm_timer *t);
 | |
|     void *priv;
 | |
| 
 | |
|     char expired;
 | |
|     char pending;
 | |
| };
 | |
| 
 | |
| static struct qemu_alarm_timer *alarm_timer;
 | |
| 
 | |
| int qemu_alarm_pending(void)
 | |
| {
 | |
|     return alarm_timer->pending;
 | |
| }
 | |
| 
 | |
| static inline int alarm_has_dynticks(struct qemu_alarm_timer *t)
 | |
| {
 | |
|     return !!t->rearm;
 | |
| }
 | |
| 
 | |
| static void qemu_rearm_alarm_timer(struct qemu_alarm_timer *t)
 | |
| {
 | |
|     if (!alarm_has_dynticks(t))
 | |
|         return;
 | |
| 
 | |
|     t->rearm(t);
 | |
| }
 | |
| 
 | |
| /* TODO: MIN_TIMER_REARM_US should be optimized */
 | |
| #define MIN_TIMER_REARM_US 250
 | |
| 
 | |
| #ifdef _WIN32
 | |
| 
 | |
| struct qemu_alarm_win32 {
 | |
|     MMRESULT timerId;
 | |
|     unsigned int period;
 | |
| } alarm_win32_data = {0, 0};
 | |
| 
 | |
| static int win32_start_timer(struct qemu_alarm_timer *t);
 | |
| static void win32_stop_timer(struct qemu_alarm_timer *t);
 | |
| static void win32_rearm_timer(struct qemu_alarm_timer *t);
 | |
| 
 | |
| #else
 | |
| 
 | |
| static int unix_start_timer(struct qemu_alarm_timer *t);
 | |
| static void unix_stop_timer(struct qemu_alarm_timer *t);
 | |
| 
 | |
| #ifdef __linux__
 | |
| 
 | |
| static int dynticks_start_timer(struct qemu_alarm_timer *t);
 | |
| static void dynticks_stop_timer(struct qemu_alarm_timer *t);
 | |
| static void dynticks_rearm_timer(struct qemu_alarm_timer *t);
 | |
| 
 | |
| static int hpet_start_timer(struct qemu_alarm_timer *t);
 | |
| static void hpet_stop_timer(struct qemu_alarm_timer *t);
 | |
| 
 | |
| static int rtc_start_timer(struct qemu_alarm_timer *t);
 | |
| static void rtc_stop_timer(struct qemu_alarm_timer *t);
 | |
| 
 | |
| #endif /* __linux__ */
 | |
| 
 | |
| #endif /* _WIN32 */
 | |
| 
 | |
| /* Correlation between real and virtual time is always going to be
 | |
|    fairly approximate, so ignore small variation.
 | |
|    When the guest is idle real and virtual time will be aligned in
 | |
|    the IO wait loop.  */
 | |
| #define ICOUNT_WOBBLE (get_ticks_per_sec() / 10)
 | |
| 
 | |
| static void icount_adjust(void)
 | |
| {
 | |
|     int64_t cur_time;
 | |
|     int64_t cur_icount;
 | |
|     int64_t delta;
 | |
|     static int64_t last_delta;
 | |
|     /* If the VM is not running, then do nothing.  */
 | |
|     if (!vm_running)
 | |
|         return;
 | |
| 
 | |
|     cur_time = cpu_get_clock();
 | |
|     cur_icount = qemu_get_clock(vm_clock);
 | |
|     delta = cur_icount - cur_time;
 | |
|     /* FIXME: This is a very crude algorithm, somewhat prone to oscillation.  */
 | |
|     if (delta > 0
 | |
|         && last_delta + ICOUNT_WOBBLE < delta * 2
 | |
|         && icount_time_shift > 0) {
 | |
|         /* The guest is getting too far ahead.  Slow time down.  */
 | |
|         icount_time_shift--;
 | |
|     }
 | |
|     if (delta < 0
 | |
|         && last_delta - ICOUNT_WOBBLE > delta * 2
 | |
|         && icount_time_shift < MAX_ICOUNT_SHIFT) {
 | |
|         /* The guest is getting too far behind.  Speed time up.  */
 | |
|         icount_time_shift++;
 | |
|     }
 | |
|     last_delta = delta;
 | |
|     qemu_icount_bias = cur_icount - (qemu_icount << icount_time_shift);
 | |
| }
 | |
| 
 | |
| static void icount_adjust_rt(void * opaque)
 | |
| {
 | |
|     qemu_mod_timer(icount_rt_timer,
 | |
|                    qemu_get_clock(rt_clock) + 1000);
 | |
|     icount_adjust();
 | |
| }
 | |
| 
 | |
| static void icount_adjust_vm(void * opaque)
 | |
| {
 | |
|     qemu_mod_timer(icount_vm_timer,
 | |
|                    qemu_get_clock(vm_clock) + get_ticks_per_sec() / 10);
 | |
|     icount_adjust();
 | |
| }
 | |
| 
 | |
| int64_t qemu_icount_round(int64_t count)
 | |
| {
 | |
|     return (count + (1 << icount_time_shift) - 1) >> icount_time_shift;
 | |
| }
 | |
| 
 | |
| static struct qemu_alarm_timer alarm_timers[] = {
 | |
| #ifndef _WIN32
 | |
| #ifdef __linux__
 | |
|     {"dynticks", dynticks_start_timer,
 | |
|      dynticks_stop_timer, dynticks_rearm_timer, NULL},
 | |
|     /* HPET - if available - is preferred */
 | |
|     {"hpet", hpet_start_timer, hpet_stop_timer, NULL, NULL},
 | |
|     /* ...otherwise try RTC */
 | |
|     {"rtc", rtc_start_timer, rtc_stop_timer, NULL, NULL},
 | |
| #endif
 | |
|     {"unix", unix_start_timer, unix_stop_timer, NULL, NULL},
 | |
| #else
 | |
|     {"dynticks", win32_start_timer,
 | |
|      win32_stop_timer, win32_rearm_timer, &alarm_win32_data},
 | |
|     {"win32", win32_start_timer,
 | |
|      win32_stop_timer, NULL, &alarm_win32_data},
 | |
| #endif
 | |
|     {NULL, }
 | |
| };
 | |
| 
 | |
| static void show_available_alarms(void)
 | |
| {
 | |
|     int i;
 | |
| 
 | |
|     printf("Available alarm timers, in order of precedence:\n");
 | |
|     for (i = 0; alarm_timers[i].name; i++)
 | |
|         printf("%s\n", alarm_timers[i].name);
 | |
| }
 | |
| 
 | |
| void configure_alarms(char const *opt)
 | |
| {
 | |
|     int i;
 | |
|     int cur = 0;
 | |
|     int count = ARRAY_SIZE(alarm_timers) - 1;
 | |
|     char *arg;
 | |
|     char *name;
 | |
|     struct qemu_alarm_timer tmp;
 | |
| 
 | |
|     if (!strcmp(opt, "?")) {
 | |
|         show_available_alarms();
 | |
|         exit(0);
 | |
|     }
 | |
| 
 | |
|     arg = qemu_strdup(opt);
 | |
| 
 | |
|     /* Reorder the array */
 | |
|     name = strtok(arg, ",");
 | |
|     while (name) {
 | |
|         for (i = 0; i < count && alarm_timers[i].name; i++) {
 | |
|             if (!strcmp(alarm_timers[i].name, name))
 | |
|                 break;
 | |
|         }
 | |
| 
 | |
|         if (i == count) {
 | |
|             fprintf(stderr, "Unknown clock %s\n", name);
 | |
|             goto next;
 | |
|         }
 | |
| 
 | |
|         if (i < cur)
 | |
|             /* Ignore */
 | |
|             goto next;
 | |
| 
 | |
| 	/* Swap */
 | |
|         tmp = alarm_timers[i];
 | |
|         alarm_timers[i] = alarm_timers[cur];
 | |
|         alarm_timers[cur] = tmp;
 | |
| 
 | |
|         cur++;
 | |
| next:
 | |
|         name = strtok(NULL, ",");
 | |
|     }
 | |
| 
 | |
|     qemu_free(arg);
 | |
| 
 | |
|     if (cur) {
 | |
|         /* Disable remaining timers */
 | |
|         for (i = cur; i < count; i++)
 | |
|             alarm_timers[i].name = NULL;
 | |
|     } else {
 | |
|         show_available_alarms();
 | |
|         exit(1);
 | |
|     }
 | |
| }
 | |
| 
 | |
| #define QEMU_NUM_CLOCKS 3
 | |
| 
 | |
| QEMUClock *rt_clock;
 | |
| QEMUClock *vm_clock;
 | |
| QEMUClock *host_clock;
 | |
| 
 | |
| static QEMUTimer *active_timers[QEMU_NUM_CLOCKS];
 | |
| 
 | |
| static QEMUClock *qemu_new_clock(int type)
 | |
| {
 | |
|     QEMUClock *clock;
 | |
|     clock = qemu_mallocz(sizeof(QEMUClock));
 | |
|     clock->type = type;
 | |
|     clock->enabled = 1;
 | |
|     return clock;
 | |
| }
 | |
| 
 | |
| void qemu_clock_enable(QEMUClock *clock, int enabled)
 | |
| {
 | |
|     clock->enabled = enabled;
 | |
| }
 | |
| 
 | |
| QEMUTimer *qemu_new_timer(QEMUClock *clock, QEMUTimerCB *cb, void *opaque)
 | |
| {
 | |
|     QEMUTimer *ts;
 | |
| 
 | |
|     ts = qemu_mallocz(sizeof(QEMUTimer));
 | |
|     ts->clock = clock;
 | |
|     ts->cb = cb;
 | |
|     ts->opaque = opaque;
 | |
|     return ts;
 | |
| }
 | |
| 
 | |
| void qemu_free_timer(QEMUTimer *ts)
 | |
| {
 | |
|     qemu_free(ts);
 | |
| }
 | |
| 
 | |
| /* stop a timer, but do not dealloc it */
 | |
| void qemu_del_timer(QEMUTimer *ts)
 | |
| {
 | |
|     QEMUTimer **pt, *t;
 | |
| 
 | |
|     /* NOTE: this code must be signal safe because
 | |
|        qemu_timer_expired() can be called from a signal. */
 | |
|     pt = &active_timers[ts->clock->type];
 | |
|     for(;;) {
 | |
|         t = *pt;
 | |
|         if (!t)
 | |
|             break;
 | |
|         if (t == ts) {
 | |
|             *pt = t->next;
 | |
|             break;
 | |
|         }
 | |
|         pt = &t->next;
 | |
|     }
 | |
| }
 | |
| 
 | |
| /* modify the current timer so that it will be fired when current_time
 | |
|    >= expire_time. The corresponding callback will be called. */
 | |
| void qemu_mod_timer(QEMUTimer *ts, int64_t expire_time)
 | |
| {
 | |
|     QEMUTimer **pt, *t;
 | |
| 
 | |
|     qemu_del_timer(ts);
 | |
| 
 | |
|     /* add the timer in the sorted list */
 | |
|     /* NOTE: this code must be signal safe because
 | |
|        qemu_timer_expired() can be called from a signal. */
 | |
|     pt = &active_timers[ts->clock->type];
 | |
|     for(;;) {
 | |
|         t = *pt;
 | |
|         if (!t)
 | |
|             break;
 | |
|         if (t->expire_time > expire_time)
 | |
|             break;
 | |
|         pt = &t->next;
 | |
|     }
 | |
|     ts->expire_time = expire_time;
 | |
|     ts->next = *pt;
 | |
|     *pt = ts;
 | |
| 
 | |
|     /* Rearm if necessary  */
 | |
|     if (pt == &active_timers[ts->clock->type]) {
 | |
|         if (!alarm_timer->pending) {
 | |
|             qemu_rearm_alarm_timer(alarm_timer);
 | |
|         }
 | |
|         /* Interrupt execution to force deadline recalculation.  */
 | |
|         if (use_icount)
 | |
|             qemu_notify_event();
 | |
|     }
 | |
| }
 | |
| 
 | |
| int qemu_timer_pending(QEMUTimer *ts)
 | |
| {
 | |
|     QEMUTimer *t;
 | |
|     for(t = active_timers[ts->clock->type]; t != NULL; t = t->next) {
 | |
|         if (t == ts)
 | |
|             return 1;
 | |
|     }
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| int qemu_timer_expired(QEMUTimer *timer_head, int64_t current_time)
 | |
| {
 | |
|     if (!timer_head)
 | |
|         return 0;
 | |
|     return (timer_head->expire_time <= current_time);
 | |
| }
 | |
| 
 | |
| static void qemu_run_timers(QEMUClock *clock)
 | |
| {
 | |
|     QEMUTimer **ptimer_head, *ts;
 | |
|     int64_t current_time;
 | |
|    
 | |
|     if (!clock->enabled)
 | |
|         return;
 | |
| 
 | |
|     current_time = qemu_get_clock (clock);
 | |
|     ptimer_head = &active_timers[clock->type];
 | |
|     for(;;) {
 | |
|         ts = *ptimer_head;
 | |
|         if (!ts || ts->expire_time > current_time)
 | |
|             break;
 | |
|         /* remove timer from the list before calling the callback */
 | |
|         *ptimer_head = ts->next;
 | |
|         ts->next = NULL;
 | |
| 
 | |
|         /* run the callback (the timer list can be modified) */
 | |
|         ts->cb(ts->opaque);
 | |
|     }
 | |
| }
 | |
| 
 | |
| int64_t qemu_get_clock(QEMUClock *clock)
 | |
| {
 | |
|     switch(clock->type) {
 | |
|     case QEMU_CLOCK_REALTIME:
 | |
|         return get_clock() / 1000000;
 | |
|     default:
 | |
|     case QEMU_CLOCK_VIRTUAL:
 | |
|         if (use_icount) {
 | |
|             return cpu_get_icount();
 | |
|         } else {
 | |
|             return cpu_get_clock();
 | |
|         }
 | |
|     case QEMU_CLOCK_HOST:
 | |
|         return get_clock_realtime();
 | |
|     }
 | |
| }
 | |
| 
 | |
| int64_t qemu_get_clock_ns(QEMUClock *clock)
 | |
| {
 | |
|     switch(clock->type) {
 | |
|     case QEMU_CLOCK_REALTIME:
 | |
|         return get_clock();
 | |
|     default:
 | |
|     case QEMU_CLOCK_VIRTUAL:
 | |
|         if (use_icount) {
 | |
|             return cpu_get_icount();
 | |
|         } else {
 | |
|             return cpu_get_clock();
 | |
|         }
 | |
|     case QEMU_CLOCK_HOST:
 | |
|         return get_clock_realtime();
 | |
|     }
 | |
| }
 | |
| 
 | |
| void init_clocks(void)
 | |
| {
 | |
|     rt_clock = qemu_new_clock(QEMU_CLOCK_REALTIME);
 | |
|     vm_clock = qemu_new_clock(QEMU_CLOCK_VIRTUAL);
 | |
|     host_clock = qemu_new_clock(QEMU_CLOCK_HOST);
 | |
| 
 | |
|     rtc_clock = host_clock;
 | |
| }
 | |
| 
 | |
| /* save a timer */
 | |
| void qemu_put_timer(QEMUFile *f, QEMUTimer *ts)
 | |
| {
 | |
|     uint64_t expire_time;
 | |
| 
 | |
|     if (qemu_timer_pending(ts)) {
 | |
|         expire_time = ts->expire_time;
 | |
|     } else {
 | |
|         expire_time = -1;
 | |
|     }
 | |
|     qemu_put_be64(f, expire_time);
 | |
| }
 | |
| 
 | |
| void qemu_get_timer(QEMUFile *f, QEMUTimer *ts)
 | |
| {
 | |
|     uint64_t expire_time;
 | |
| 
 | |
|     expire_time = qemu_get_be64(f);
 | |
|     if (expire_time != -1) {
 | |
|         qemu_mod_timer(ts, expire_time);
 | |
|     } else {
 | |
|         qemu_del_timer(ts);
 | |
|     }
 | |
| }
 | |
| 
 | |
| static const VMStateDescription vmstate_timers = {
 | |
|     .name = "timer",
 | |
|     .version_id = 2,
 | |
|     .minimum_version_id = 1,
 | |
|     .minimum_version_id_old = 1,
 | |
|     .fields      = (VMStateField []) {
 | |
|         VMSTATE_INT64(cpu_ticks_offset, TimersState),
 | |
|         VMSTATE_INT64(dummy, TimersState),
 | |
|         VMSTATE_INT64_V(cpu_clock_offset, TimersState, 2),
 | |
|         VMSTATE_END_OF_LIST()
 | |
|     }
 | |
| };
 | |
| 
 | |
| void configure_icount(const char *option)
 | |
| {
 | |
|     vmstate_register(NULL, 0, &vmstate_timers, &timers_state);
 | |
|     if (!option)
 | |
|         return;
 | |
| 
 | |
|     if (strcmp(option, "auto") != 0) {
 | |
|         icount_time_shift = strtol(option, NULL, 0);
 | |
|         use_icount = 1;
 | |
|         return;
 | |
|     }
 | |
| 
 | |
|     use_icount = 2;
 | |
| 
 | |
|     /* 125MIPS seems a reasonable initial guess at the guest speed.
 | |
|        It will be corrected fairly quickly anyway.  */
 | |
|     icount_time_shift = 3;
 | |
| 
 | |
|     /* Have both realtime and virtual time triggers for speed adjustment.
 | |
|        The realtime trigger catches emulated time passing too slowly,
 | |
|        the virtual time trigger catches emulated time passing too fast.
 | |
|        Realtime triggers occur even when idle, so use them less frequently
 | |
|        than VM triggers.  */
 | |
|     icount_rt_timer = qemu_new_timer(rt_clock, icount_adjust_rt, NULL);
 | |
|     qemu_mod_timer(icount_rt_timer,
 | |
|                    qemu_get_clock(rt_clock) + 1000);
 | |
|     icount_vm_timer = qemu_new_timer(vm_clock, icount_adjust_vm, NULL);
 | |
|     qemu_mod_timer(icount_vm_timer,
 | |
|                    qemu_get_clock(vm_clock) + get_ticks_per_sec() / 10);
 | |
| }
 | |
| 
 | |
| void qemu_run_all_timers(void)
 | |
| {
 | |
|     alarm_timer->pending = 0;
 | |
| 
 | |
|     /* rearm timer, if not periodic */
 | |
|     if (alarm_timer->expired) {
 | |
|         alarm_timer->expired = 0;
 | |
|         qemu_rearm_alarm_timer(alarm_timer);
 | |
|     }
 | |
| 
 | |
|     /* vm time timers */
 | |
|     if (vm_running) {
 | |
|         qemu_run_timers(vm_clock);
 | |
|     }
 | |
| 
 | |
|     qemu_run_timers(rt_clock);
 | |
|     qemu_run_timers(host_clock);
 | |
| }
 | |
| 
 | |
| #ifdef _WIN32
 | |
| static void CALLBACK host_alarm_handler(UINT uTimerID, UINT uMsg,
 | |
|                                         DWORD_PTR dwUser, DWORD_PTR dw1,
 | |
|                                         DWORD_PTR dw2)
 | |
| #else
 | |
| static void host_alarm_handler(int host_signum)
 | |
| #endif
 | |
| {
 | |
|     struct qemu_alarm_timer *t = alarm_timer;
 | |
|     if (!t)
 | |
| 	return;
 | |
| 
 | |
| #if 0
 | |
| #define DISP_FREQ 1000
 | |
|     {
 | |
|         static int64_t delta_min = INT64_MAX;
 | |
|         static int64_t delta_max, delta_cum, last_clock, delta, ti;
 | |
|         static int count;
 | |
|         ti = qemu_get_clock(vm_clock);
 | |
|         if (last_clock != 0) {
 | |
|             delta = ti - last_clock;
 | |
|             if (delta < delta_min)
 | |
|                 delta_min = delta;
 | |
|             if (delta > delta_max)
 | |
|                 delta_max = delta;
 | |
|             delta_cum += delta;
 | |
|             if (++count == DISP_FREQ) {
 | |
|                 printf("timer: min=%" PRId64 " us max=%" PRId64 " us avg=%" PRId64 " us avg_freq=%0.3f Hz\n",
 | |
|                        muldiv64(delta_min, 1000000, get_ticks_per_sec()),
 | |
|                        muldiv64(delta_max, 1000000, get_ticks_per_sec()),
 | |
|                        muldiv64(delta_cum, 1000000 / DISP_FREQ, get_ticks_per_sec()),
 | |
|                        (double)get_ticks_per_sec() / ((double)delta_cum / DISP_FREQ));
 | |
|                 count = 0;
 | |
|                 delta_min = INT64_MAX;
 | |
|                 delta_max = 0;
 | |
|                 delta_cum = 0;
 | |
|             }
 | |
|         }
 | |
|         last_clock = ti;
 | |
|     }
 | |
| #endif
 | |
|     if (alarm_has_dynticks(t) ||
 | |
|         (!use_icount &&
 | |
|             qemu_timer_expired(active_timers[QEMU_CLOCK_VIRTUAL],
 | |
|                                qemu_get_clock(vm_clock))) ||
 | |
|         qemu_timer_expired(active_timers[QEMU_CLOCK_REALTIME],
 | |
|                            qemu_get_clock(rt_clock)) ||
 | |
|         qemu_timer_expired(active_timers[QEMU_CLOCK_HOST],
 | |
|                            qemu_get_clock(host_clock))) {
 | |
| 
 | |
|         t->expired = alarm_has_dynticks(t);
 | |
|         t->pending = 1;
 | |
|         qemu_notify_event();
 | |
|     }
 | |
| }
 | |
| 
 | |
| int64_t qemu_next_deadline(void)
 | |
| {
 | |
|     /* To avoid problems with overflow limit this to 2^32.  */
 | |
|     int64_t delta = INT32_MAX;
 | |
| 
 | |
|     if (active_timers[QEMU_CLOCK_VIRTUAL]) {
 | |
|         delta = active_timers[QEMU_CLOCK_VIRTUAL]->expire_time -
 | |
|                      qemu_get_clock(vm_clock);
 | |
|     }
 | |
|     if (active_timers[QEMU_CLOCK_HOST]) {
 | |
|         int64_t hdelta = active_timers[QEMU_CLOCK_HOST]->expire_time -
 | |
|                  qemu_get_clock(host_clock);
 | |
|         if (hdelta < delta)
 | |
|             delta = hdelta;
 | |
|     }
 | |
| 
 | |
|     if (delta < 0)
 | |
|         delta = 0;
 | |
| 
 | |
|     return delta;
 | |
| }
 | |
| 
 | |
| #ifndef _WIN32
 | |
| 
 | |
| #if defined(__linux__)
 | |
| 
 | |
| #define RTC_FREQ 1024
 | |
| 
 | |
| static uint64_t qemu_next_deadline_dyntick(void)
 | |
| {
 | |
|     int64_t delta;
 | |
|     int64_t rtdelta;
 | |
| 
 | |
|     if (use_icount)
 | |
|         delta = INT32_MAX;
 | |
|     else
 | |
|         delta = (qemu_next_deadline() + 999) / 1000;
 | |
| 
 | |
|     if (active_timers[QEMU_CLOCK_REALTIME]) {
 | |
|         rtdelta = (active_timers[QEMU_CLOCK_REALTIME]->expire_time -
 | |
|                  qemu_get_clock(rt_clock))*1000;
 | |
|         if (rtdelta < delta)
 | |
|             delta = rtdelta;
 | |
|     }
 | |
| 
 | |
|     if (delta < MIN_TIMER_REARM_US)
 | |
|         delta = MIN_TIMER_REARM_US;
 | |
| 
 | |
|     return delta;
 | |
| }
 | |
| 
 | |
| static void enable_sigio_timer(int fd)
 | |
| {
 | |
|     struct sigaction act;
 | |
| 
 | |
|     /* timer signal */
 | |
|     sigfillset(&act.sa_mask);
 | |
|     act.sa_flags = 0;
 | |
|     act.sa_handler = host_alarm_handler;
 | |
| 
 | |
|     sigaction(SIGIO, &act, NULL);
 | |
|     fcntl_setfl(fd, O_ASYNC);
 | |
|     fcntl(fd, F_SETOWN, getpid());
 | |
| }
 | |
| 
 | |
| static int hpet_start_timer(struct qemu_alarm_timer *t)
 | |
| {
 | |
|     struct hpet_info info;
 | |
|     int r, fd;
 | |
| 
 | |
|     fd = qemu_open("/dev/hpet", O_RDONLY);
 | |
|     if (fd < 0)
 | |
|         return -1;
 | |
| 
 | |
|     /* Set frequency */
 | |
|     r = ioctl(fd, HPET_IRQFREQ, RTC_FREQ);
 | |
|     if (r < 0) {
 | |
|         fprintf(stderr, "Could not configure '/dev/hpet' to have a 1024Hz timer. This is not a fatal\n"
 | |
|                 "error, but for better emulation accuracy type:\n"
 | |
|                 "'echo 1024 > /proc/sys/dev/hpet/max-user-freq' as root.\n");
 | |
|         goto fail;
 | |
|     }
 | |
| 
 | |
|     /* Check capabilities */
 | |
|     r = ioctl(fd, HPET_INFO, &info);
 | |
|     if (r < 0)
 | |
|         goto fail;
 | |
| 
 | |
|     /* Enable periodic mode */
 | |
|     r = ioctl(fd, HPET_EPI, 0);
 | |
|     if (info.hi_flags && (r < 0))
 | |
|         goto fail;
 | |
| 
 | |
|     /* Enable interrupt */
 | |
|     r = ioctl(fd, HPET_IE_ON, 0);
 | |
|     if (r < 0)
 | |
|         goto fail;
 | |
| 
 | |
|     enable_sigio_timer(fd);
 | |
|     t->priv = (void *)(long)fd;
 | |
| 
 | |
|     return 0;
 | |
| fail:
 | |
|     close(fd);
 | |
|     return -1;
 | |
| }
 | |
| 
 | |
| static void hpet_stop_timer(struct qemu_alarm_timer *t)
 | |
| {
 | |
|     int fd = (long)t->priv;
 | |
| 
 | |
|     close(fd);
 | |
| }
 | |
| 
 | |
| static int rtc_start_timer(struct qemu_alarm_timer *t)
 | |
| {
 | |
|     int rtc_fd;
 | |
|     unsigned long current_rtc_freq = 0;
 | |
| 
 | |
|     TFR(rtc_fd = qemu_open("/dev/rtc", O_RDONLY));
 | |
|     if (rtc_fd < 0)
 | |
|         return -1;
 | |
|     ioctl(rtc_fd, RTC_IRQP_READ, ¤t_rtc_freq);
 | |
|     if (current_rtc_freq != RTC_FREQ &&
 | |
|         ioctl(rtc_fd, RTC_IRQP_SET, RTC_FREQ) < 0) {
 | |
|         fprintf(stderr, "Could not configure '/dev/rtc' to have a 1024 Hz timer. This is not a fatal\n"
 | |
|                 "error, but for better emulation accuracy either use a 2.6 host Linux kernel or\n"
 | |
|                 "type 'echo 1024 > /proc/sys/dev/rtc/max-user-freq' as root.\n");
 | |
|         goto fail;
 | |
|     }
 | |
|     if (ioctl(rtc_fd, RTC_PIE_ON, 0) < 0) {
 | |
|     fail:
 | |
|         close(rtc_fd);
 | |
|         return -1;
 | |
|     }
 | |
| 
 | |
|     enable_sigio_timer(rtc_fd);
 | |
| 
 | |
|     t->priv = (void *)(long)rtc_fd;
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static void rtc_stop_timer(struct qemu_alarm_timer *t)
 | |
| {
 | |
|     int rtc_fd = (long)t->priv;
 | |
| 
 | |
|     close(rtc_fd);
 | |
| }
 | |
| 
 | |
| static int dynticks_start_timer(struct qemu_alarm_timer *t)
 | |
| {
 | |
|     struct sigevent ev;
 | |
|     timer_t host_timer;
 | |
|     struct sigaction act;
 | |
| 
 | |
|     sigfillset(&act.sa_mask);
 | |
|     act.sa_flags = 0;
 | |
|     act.sa_handler = host_alarm_handler;
 | |
| 
 | |
|     sigaction(SIGALRM, &act, NULL);
 | |
| 
 | |
|     /* 
 | |
|      * Initialize ev struct to 0 to avoid valgrind complaining
 | |
|      * about uninitialized data in timer_create call
 | |
|      */
 | |
|     memset(&ev, 0, sizeof(ev));
 | |
|     ev.sigev_value.sival_int = 0;
 | |
|     ev.sigev_notify = SIGEV_SIGNAL;
 | |
|     ev.sigev_signo = SIGALRM;
 | |
| 
 | |
|     if (timer_create(CLOCK_REALTIME, &ev, &host_timer)) {
 | |
|         perror("timer_create");
 | |
| 
 | |
|         /* disable dynticks */
 | |
|         fprintf(stderr, "Dynamic Ticks disabled\n");
 | |
| 
 | |
|         return -1;
 | |
|     }
 | |
| 
 | |
|     t->priv = (void *)(long)host_timer;
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static void dynticks_stop_timer(struct qemu_alarm_timer *t)
 | |
| {
 | |
|     timer_t host_timer = (timer_t)(long)t->priv;
 | |
| 
 | |
|     timer_delete(host_timer);
 | |
| }
 | |
| 
 | |
| static void dynticks_rearm_timer(struct qemu_alarm_timer *t)
 | |
| {
 | |
|     timer_t host_timer = (timer_t)(long)t->priv;
 | |
|     struct itimerspec timeout;
 | |
|     int64_t nearest_delta_us = INT64_MAX;
 | |
|     int64_t current_us;
 | |
| 
 | |
|     assert(alarm_has_dynticks(t));
 | |
|     if (!active_timers[QEMU_CLOCK_REALTIME] &&
 | |
|         !active_timers[QEMU_CLOCK_VIRTUAL] &&
 | |
|         !active_timers[QEMU_CLOCK_HOST])
 | |
|         return;
 | |
| 
 | |
|     nearest_delta_us = qemu_next_deadline_dyntick();
 | |
| 
 | |
|     /* check whether a timer is already running */
 | |
|     if (timer_gettime(host_timer, &timeout)) {
 | |
|         perror("gettime");
 | |
|         fprintf(stderr, "Internal timer error: aborting\n");
 | |
|         exit(1);
 | |
|     }
 | |
|     current_us = timeout.it_value.tv_sec * 1000000 + timeout.it_value.tv_nsec/1000;
 | |
|     if (current_us && current_us <= nearest_delta_us)
 | |
|         return;
 | |
| 
 | |
|     timeout.it_interval.tv_sec = 0;
 | |
|     timeout.it_interval.tv_nsec = 0; /* 0 for one-shot timer */
 | |
|     timeout.it_value.tv_sec =  nearest_delta_us / 1000000;
 | |
|     timeout.it_value.tv_nsec = (nearest_delta_us % 1000000) * 1000;
 | |
|     if (timer_settime(host_timer, 0 /* RELATIVE */, &timeout, NULL)) {
 | |
|         perror("settime");
 | |
|         fprintf(stderr, "Internal timer error: aborting\n");
 | |
|         exit(1);
 | |
|     }
 | |
| }
 | |
| 
 | |
| #endif /* defined(__linux__) */
 | |
| 
 | |
| static int unix_start_timer(struct qemu_alarm_timer *t)
 | |
| {
 | |
|     struct sigaction act;
 | |
|     struct itimerval itv;
 | |
|     int err;
 | |
| 
 | |
|     /* timer signal */
 | |
|     sigfillset(&act.sa_mask);
 | |
|     act.sa_flags = 0;
 | |
|     act.sa_handler = host_alarm_handler;
 | |
| 
 | |
|     sigaction(SIGALRM, &act, NULL);
 | |
| 
 | |
|     itv.it_interval.tv_sec = 0;
 | |
|     /* for i386 kernel 2.6 to get 1 ms */
 | |
|     itv.it_interval.tv_usec = 999;
 | |
|     itv.it_value.tv_sec = 0;
 | |
|     itv.it_value.tv_usec = 10 * 1000;
 | |
| 
 | |
|     err = setitimer(ITIMER_REAL, &itv, NULL);
 | |
|     if (err)
 | |
|         return -1;
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static void unix_stop_timer(struct qemu_alarm_timer *t)
 | |
| {
 | |
|     struct itimerval itv;
 | |
| 
 | |
|     memset(&itv, 0, sizeof(itv));
 | |
|     setitimer(ITIMER_REAL, &itv, NULL);
 | |
| }
 | |
| 
 | |
| #endif /* !defined(_WIN32) */
 | |
| 
 | |
| 
 | |
| #ifdef _WIN32
 | |
| 
 | |
| static int win32_start_timer(struct qemu_alarm_timer *t)
 | |
| {
 | |
|     TIMECAPS tc;
 | |
|     struct qemu_alarm_win32 *data = t->priv;
 | |
|     UINT flags;
 | |
| 
 | |
|     memset(&tc, 0, sizeof(tc));
 | |
|     timeGetDevCaps(&tc, sizeof(tc));
 | |
| 
 | |
|     data->period = tc.wPeriodMin;
 | |
|     timeBeginPeriod(data->period);
 | |
| 
 | |
|     flags = TIME_CALLBACK_FUNCTION;
 | |
|     if (alarm_has_dynticks(t))
 | |
|         flags |= TIME_ONESHOT;
 | |
|     else
 | |
|         flags |= TIME_PERIODIC;
 | |
| 
 | |
|     data->timerId = timeSetEvent(1,         // interval (ms)
 | |
|                         data->period,       // resolution
 | |
|                         host_alarm_handler, // function
 | |
|                         (DWORD)t,           // parameter
 | |
|                         flags);
 | |
| 
 | |
|     if (!data->timerId) {
 | |
|         fprintf(stderr, "Failed to initialize win32 alarm timer: %ld\n",
 | |
|                 GetLastError());
 | |
|         timeEndPeriod(data->period);
 | |
|         return -1;
 | |
|     }
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static void win32_stop_timer(struct qemu_alarm_timer *t)
 | |
| {
 | |
|     struct qemu_alarm_win32 *data = t->priv;
 | |
| 
 | |
|     timeKillEvent(data->timerId);
 | |
|     timeEndPeriod(data->period);
 | |
| }
 | |
| 
 | |
| static void win32_rearm_timer(struct qemu_alarm_timer *t)
 | |
| {
 | |
|     struct qemu_alarm_win32 *data = t->priv;
 | |
| 
 | |
|     assert(alarm_has_dynticks(t));
 | |
|     if (!active_timers[QEMU_CLOCK_REALTIME] &&
 | |
|         !active_timers[QEMU_CLOCK_VIRTUAL] &&
 | |
|         !active_timers[QEMU_CLOCK_HOST])
 | |
|         return;
 | |
| 
 | |
|     timeKillEvent(data->timerId);
 | |
| 
 | |
|     data->timerId = timeSetEvent(1,
 | |
|                         data->period,
 | |
|                         host_alarm_handler,
 | |
|                         (DWORD)t,
 | |
|                         TIME_ONESHOT | TIME_CALLBACK_FUNCTION);
 | |
| 
 | |
|     if (!data->timerId) {
 | |
|         fprintf(stderr, "Failed to re-arm win32 alarm timer %ld\n",
 | |
|                 GetLastError());
 | |
| 
 | |
|         timeEndPeriod(data->period);
 | |
|         exit(1);
 | |
|     }
 | |
| }
 | |
| 
 | |
| #endif /* _WIN32 */
 | |
| 
 | |
| static void alarm_timer_on_change_state_rearm(void *opaque, int running, int reason)
 | |
| {
 | |
|     if (running)
 | |
|         qemu_rearm_alarm_timer((struct qemu_alarm_timer *) opaque);
 | |
| }
 | |
| 
 | |
| int init_timer_alarm(void)
 | |
| {
 | |
|     struct qemu_alarm_timer *t = NULL;
 | |
|     int i, err = -1;
 | |
| 
 | |
|     for (i = 0; alarm_timers[i].name; i++) {
 | |
|         t = &alarm_timers[i];
 | |
| 
 | |
|         err = t->start(t);
 | |
|         if (!err)
 | |
|             break;
 | |
|     }
 | |
| 
 | |
|     if (err) {
 | |
|         err = -ENOENT;
 | |
|         goto fail;
 | |
|     }
 | |
| 
 | |
|     /* first event is at time 0 */
 | |
|     t->pending = 1;
 | |
|     alarm_timer = t;
 | |
|     qemu_add_vm_change_state_handler(alarm_timer_on_change_state_rearm, t);
 | |
| 
 | |
|     return 0;
 | |
| 
 | |
| fail:
 | |
|     return err;
 | |
| }
 | |
| 
 | |
| void quit_timers(void)
 | |
| {
 | |
|     struct qemu_alarm_timer *t = alarm_timer;
 | |
|     alarm_timer = NULL;
 | |
|     t->stop(t);
 | |
| }
 | |
| 
 | |
| int qemu_calculate_timeout(void)
 | |
| {
 | |
| #ifndef CONFIG_IOTHREAD
 | |
|     int timeout;
 | |
| 
 | |
|     if (!vm_running)
 | |
|         timeout = 5000;
 | |
|     else {
 | |
|      /* XXX: use timeout computed from timers */
 | |
|         int64_t add;
 | |
|         int64_t delta;
 | |
|         /* Advance virtual time to the next event.  */
 | |
| 	delta = qemu_icount_delta();
 | |
|         if (delta > 0) {
 | |
|             /* If virtual time is ahead of real time then just
 | |
|                wait for IO.  */
 | |
|             timeout = (delta + 999999) / 1000000;
 | |
|         } else {
 | |
|             /* Wait for either IO to occur or the next
 | |
|                timer event.  */
 | |
|             add = qemu_next_deadline();
 | |
|             /* We advance the timer before checking for IO.
 | |
|                Limit the amount we advance so that early IO
 | |
|                activity won't get the guest too far ahead.  */
 | |
|             if (add > 10000000)
 | |
|                 add = 10000000;
 | |
|             delta += add;
 | |
|             qemu_icount += qemu_icount_round (add);
 | |
|             timeout = delta / 1000000;
 | |
|             if (timeout < 0)
 | |
|                 timeout = 0;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     return timeout;
 | |
| #else /* CONFIG_IOTHREAD */
 | |
|     return 1000;
 | |
| #endif
 | |
| }
 | |
| 
 |