[34] | 1 | /****************************************************************************** |
---|
| 2 | * arch/x86/time.c |
---|
| 3 | * |
---|
| 4 | * Per-CPU time calibration and management. |
---|
| 5 | * |
---|
| 6 | * Copyright (c) 2002-2005, K A Fraser |
---|
| 7 | * |
---|
| 8 | * Portions from Linux are: |
---|
| 9 | * Copyright (c) 1991, 1992, 1995 Linus Torvalds |
---|
| 10 | */ |
---|
| 11 | |
---|
| 12 | #include <xen/config.h> |
---|
| 13 | #include <xen/errno.h> |
---|
| 14 | #include <xen/event.h> |
---|
| 15 | #include <xen/sched.h> |
---|
| 16 | #include <xen/lib.h> |
---|
| 17 | #include <xen/config.h> |
---|
| 18 | #include <xen/init.h> |
---|
| 19 | #include <xen/time.h> |
---|
| 20 | #include <xen/timer.h> |
---|
| 21 | #include <xen/smp.h> |
---|
| 22 | #include <xen/irq.h> |
---|
| 23 | #include <xen/softirq.h> |
---|
| 24 | #include <asm/io.h> |
---|
| 25 | #include <asm/msr.h> |
---|
| 26 | #include <asm/mpspec.h> |
---|
| 27 | #include <asm/processor.h> |
---|
| 28 | #include <asm/fixmap.h> |
---|
| 29 | #include <asm/mc146818rtc.h> |
---|
| 30 | #include <asm/div64.h> |
---|
| 31 | #include <asm/hpet.h> |
---|
| 32 | #include <io_ports.h> |
---|
| 33 | |
---|
| 34 | /* opt_hpet_force: If true, force HPET configuration via PCI space. */ |
---|
| 35 | /* NB. This is a gross hack. Mainly useful for HPET testing. */ |
---|
| 36 | static int opt_hpet_force = 0; |
---|
| 37 | boolean_param("hpet_force", opt_hpet_force); |
---|
| 38 | |
---|
| 39 | #define EPOCH MILLISECS(1000) |
---|
| 40 | |
---|
| 41 | unsigned long cpu_khz; /* CPU clock frequency in kHz. */ |
---|
| 42 | unsigned long hpet_address; |
---|
| 43 | DEFINE_SPINLOCK(rtc_lock); |
---|
| 44 | volatile unsigned long jiffies; |
---|
| 45 | static u32 wc_sec, wc_nsec; /* UTC time at last 'time update'. */ |
---|
| 46 | static DEFINE_SPINLOCK(wc_lock); |
---|
| 47 | |
---|
| 48 | struct time_scale { |
---|
| 49 | int shift; |
---|
| 50 | u32 mul_frac; |
---|
| 51 | }; |
---|
| 52 | |
---|
| 53 | struct cpu_time { |
---|
| 54 | u64 local_tsc_stamp; |
---|
| 55 | s_time_t stime_local_stamp; |
---|
| 56 | s_time_t stime_master_stamp; |
---|
| 57 | struct time_scale tsc_scale; |
---|
| 58 | struct timer calibration_timer; |
---|
| 59 | }; |
---|
| 60 | |
---|
| 61 | static DEFINE_PER_CPU(struct cpu_time, cpu_time); |
---|
| 62 | |
---|
| 63 | /* |
---|
| 64 | * Protected by platform_timer_lock, which must be acquired with interrupts |
---|
| 65 | * disabled because pit_overflow() is called from PIT ch0 interrupt context. |
---|
| 66 | */ |
---|
| 67 | static s_time_t stime_platform_stamp; |
---|
| 68 | static u64 platform_timer_stamp; |
---|
| 69 | static struct time_scale platform_timer_scale; |
---|
| 70 | static DEFINE_SPINLOCK(platform_timer_lock); |
---|
| 71 | static u64 (*read_platform_count)(void); |
---|
| 72 | |
---|
| 73 | /* |
---|
| 74 | * Folding 16-bit PIT into 64-bit software counter is a really critical |
---|
| 75 | * operation! We therefore do it directly in PIT ch0 interrupt handler, |
---|
| 76 | * based on this flag. |
---|
| 77 | */ |
---|
| 78 | static int using_pit; |
---|
| 79 | static void pit_overflow(void); |
---|
| 80 | |
---|
| 81 | /* |
---|
| 82 | * 32-bit division of integer dividend and integer divisor yielding |
---|
| 83 | * 32-bit fractional quotient. |
---|
| 84 | */ |
---|
| 85 | static inline u32 div_frac(u32 dividend, u32 divisor) |
---|
| 86 | { |
---|
| 87 | u32 quotient, remainder; |
---|
| 88 | ASSERT(dividend < divisor); |
---|
| 89 | __asm__ ( |
---|
| 90 | "divl %4" |
---|
| 91 | : "=a" (quotient), "=d" (remainder) |
---|
| 92 | : "0" (0), "1" (dividend), "r" (divisor) ); |
---|
| 93 | return quotient; |
---|
| 94 | } |
---|
| 95 | |
---|
| 96 | /* |
---|
| 97 | * 32-bit multiplication of multiplicand and fractional multiplier |
---|
| 98 | * yielding 32-bit product (radix point at same position as in multiplicand). |
---|
| 99 | */ |
---|
| 100 | static inline u32 mul_frac(u32 multiplicand, u32 multiplier) |
---|
| 101 | { |
---|
| 102 | u32 product_int, product_frac; |
---|
| 103 | __asm__ ( |
---|
| 104 | "mul %3" |
---|
| 105 | : "=a" (product_frac), "=d" (product_int) |
---|
| 106 | : "0" (multiplicand), "r" (multiplier) ); |
---|
| 107 | return product_int; |
---|
| 108 | } |
---|
| 109 | |
---|
| 110 | /* |
---|
| 111 | * Scale a 64-bit delta by scaling and multiplying by a 32-bit fraction, |
---|
| 112 | * yielding a 64-bit result. |
---|
| 113 | */ |
---|
| 114 | static inline u64 scale_delta(u64 delta, struct time_scale *scale) |
---|
| 115 | { |
---|
| 116 | u64 product; |
---|
| 117 | #ifdef CONFIG_X86_32 |
---|
| 118 | u32 tmp1, tmp2; |
---|
| 119 | #endif |
---|
| 120 | |
---|
| 121 | if ( scale->shift < 0 ) |
---|
| 122 | delta >>= -scale->shift; |
---|
| 123 | else |
---|
| 124 | delta <<= scale->shift; |
---|
| 125 | |
---|
| 126 | #ifdef CONFIG_X86_32 |
---|
| 127 | __asm__ ( |
---|
| 128 | "mul %5 ; " |
---|
| 129 | "mov %4,%%eax ; " |
---|
| 130 | "mov %%edx,%4 ; " |
---|
| 131 | "mul %5 ; " |
---|
| 132 | "xor %5,%5 ; " |
---|
| 133 | "add %4,%%eax ; " |
---|
| 134 | "adc %5,%%edx ; " |
---|
| 135 | : "=A" (product), "=r" (tmp1), "=r" (tmp2) |
---|
| 136 | : "a" ((u32)delta), "1" ((u32)(delta >> 32)), "2" (scale->mul_frac) ); |
---|
| 137 | #else |
---|
| 138 | __asm__ ( |
---|
| 139 | "mul %%rdx ; shrd $32,%%rdx,%%rax" |
---|
| 140 | : "=a" (product) : "0" (delta), "d" ((u64)scale->mul_frac) ); |
---|
| 141 | #endif |
---|
| 142 | |
---|
| 143 | return product; |
---|
| 144 | } |
---|
| 145 | |
---|
| 146 | void timer_interrupt(int irq, void *dev_id, struct cpu_user_regs *regs) |
---|
| 147 | { |
---|
| 148 | ASSERT(local_irq_is_enabled()); |
---|
| 149 | |
---|
| 150 | /* Update jiffies counter. */ |
---|
| 151 | (*(volatile unsigned long *)&jiffies)++; |
---|
| 152 | |
---|
| 153 | /* Rough hack to allow accurate timers to sort-of-work with no APIC. */ |
---|
| 154 | if ( !cpu_has_apic ) |
---|
| 155 | raise_softirq(TIMER_SOFTIRQ); |
---|
| 156 | |
---|
| 157 | if ( using_pit ) |
---|
| 158 | pit_overflow(); |
---|
| 159 | } |
---|
| 160 | |
---|
| 161 | static struct irqaction irq0 = { timer_interrupt, "timer", NULL}; |
---|
| 162 | |
---|
| 163 | /* ------ Calibrate the TSC ------- |
---|
| 164 | * Return processor ticks per second / CALIBRATE_FRAC. |
---|
| 165 | */ |
---|
| 166 | |
---|
| 167 | #define CLOCK_TICK_RATE 1193180 /* system crystal frequency (Hz) */ |
---|
| 168 | #define CALIBRATE_FRAC 20 /* calibrate over 50ms */ |
---|
| 169 | #define CALIBRATE_LATCH ((CLOCK_TICK_RATE+(CALIBRATE_FRAC/2))/CALIBRATE_FRAC) |
---|
| 170 | |
---|
| 171 | static u64 calibrate_boot_tsc(void) |
---|
| 172 | { |
---|
| 173 | u64 start, end; |
---|
| 174 | unsigned long count; |
---|
| 175 | |
---|
| 176 | /* Set the Gate high, disable speaker */ |
---|
| 177 | outb((inb(0x61) & ~0x02) | 0x01, 0x61); |
---|
| 178 | |
---|
| 179 | /* |
---|
| 180 | * Now let's take care of CTC channel 2 |
---|
| 181 | * |
---|
| 182 | * Set the Gate high, program CTC channel 2 for mode 0, (interrupt on |
---|
| 183 | * terminal count mode), binary count, load 5 * LATCH count, (LSB and MSB) |
---|
| 184 | * to begin countdown. |
---|
| 185 | */ |
---|
| 186 | outb(0xb0, PIT_MODE); /* binary, mode 0, LSB/MSB, Ch 2 */ |
---|
| 187 | outb(CALIBRATE_LATCH & 0xff, PIT_CH2); /* LSB of count */ |
---|
| 188 | outb(CALIBRATE_LATCH >> 8, PIT_CH2); /* MSB of count */ |
---|
| 189 | |
---|
| 190 | rdtscll(start); |
---|
| 191 | for ( count = 0; (inb(0x61) & 0x20) == 0; count++ ) |
---|
| 192 | continue; |
---|
| 193 | rdtscll(end); |
---|
| 194 | |
---|
| 195 | /* Error if the CTC doesn't behave itself. */ |
---|
| 196 | if ( count == 0 ) |
---|
| 197 | return 0; |
---|
| 198 | |
---|
| 199 | return ((end - start) * (u64)CALIBRATE_FRAC); |
---|
| 200 | } |
---|
| 201 | |
---|
| 202 | static void set_time_scale(struct time_scale *ts, u64 ticks_per_sec) |
---|
| 203 | { |
---|
| 204 | u64 tps64 = ticks_per_sec; |
---|
| 205 | u32 tps32; |
---|
| 206 | int shift = 0; |
---|
| 207 | |
---|
| 208 | while ( tps64 > (MILLISECS(1000)*2) ) |
---|
| 209 | { |
---|
| 210 | tps64 >>= 1; |
---|
| 211 | shift--; |
---|
| 212 | } |
---|
| 213 | |
---|
| 214 | tps32 = (u32)tps64; |
---|
| 215 | while ( tps32 < (u32)MILLISECS(1000) ) |
---|
| 216 | { |
---|
| 217 | tps32 <<= 1; |
---|
| 218 | shift++; |
---|
| 219 | } |
---|
| 220 | |
---|
| 221 | ts->mul_frac = div_frac(MILLISECS(1000), tps32); |
---|
| 222 | ts->shift = shift; |
---|
| 223 | } |
---|
| 224 | |
---|
| 225 | static atomic_t tsc_calibrate_gang = ATOMIC_INIT(0); |
---|
| 226 | static unsigned int tsc_calibrate_status = 0; |
---|
| 227 | |
---|
| 228 | void calibrate_tsc_bp(void) |
---|
| 229 | { |
---|
| 230 | while ( atomic_read(&tsc_calibrate_gang) != (num_booting_cpus() - 1) ) |
---|
| 231 | mb(); |
---|
| 232 | |
---|
| 233 | outb(CALIBRATE_LATCH & 0xff, PIT_CH2); |
---|
| 234 | outb(CALIBRATE_LATCH >> 8, PIT_CH2); |
---|
| 235 | |
---|
| 236 | tsc_calibrate_status = 1; |
---|
| 237 | wmb(); |
---|
| 238 | |
---|
| 239 | while ( (inb(0x61) & 0x20) == 0 ) |
---|
| 240 | continue; |
---|
| 241 | |
---|
| 242 | tsc_calibrate_status = 2; |
---|
| 243 | wmb(); |
---|
| 244 | |
---|
| 245 | while ( atomic_read(&tsc_calibrate_gang) != 0 ) |
---|
| 246 | mb(); |
---|
| 247 | } |
---|
| 248 | |
---|
| 249 | void calibrate_tsc_ap(void) |
---|
| 250 | { |
---|
| 251 | u64 t1, t2, ticks_per_sec; |
---|
| 252 | |
---|
| 253 | atomic_inc(&tsc_calibrate_gang); |
---|
| 254 | |
---|
| 255 | while ( tsc_calibrate_status < 1 ) |
---|
| 256 | mb(); |
---|
| 257 | |
---|
| 258 | rdtscll(t1); |
---|
| 259 | |
---|
| 260 | while ( tsc_calibrate_status < 2 ) |
---|
| 261 | mb(); |
---|
| 262 | |
---|
| 263 | rdtscll(t2); |
---|
| 264 | |
---|
| 265 | ticks_per_sec = (t2 - t1) * (u64)CALIBRATE_FRAC; |
---|
| 266 | set_time_scale(&this_cpu(cpu_time).tsc_scale, ticks_per_sec); |
---|
| 267 | |
---|
| 268 | atomic_dec(&tsc_calibrate_gang); |
---|
| 269 | } |
---|
| 270 | |
---|
| 271 | static char *freq_string(u64 freq) |
---|
| 272 | { |
---|
| 273 | static char s[20]; |
---|
| 274 | unsigned int x, y; |
---|
| 275 | y = (unsigned int)do_div(freq, 1000000) / 1000; |
---|
| 276 | x = (unsigned int)freq; |
---|
| 277 | snprintf(s, sizeof(s), "%u.%03uMHz", x, y); |
---|
| 278 | return s; |
---|
| 279 | } |
---|
| 280 | |
---|
| 281 | /************************************************************ |
---|
| 282 | * PLATFORM TIMER 1: PROGRAMMABLE INTERVAL TIMER (LEGACY PIT) |
---|
| 283 | */ |
---|
| 284 | |
---|
| 285 | /* Protected by platform_timer_lock. */ |
---|
| 286 | static u64 pit_counter64; |
---|
| 287 | static u16 pit_stamp; |
---|
| 288 | |
---|
| 289 | static u16 pit_read_counter(void) |
---|
| 290 | { |
---|
| 291 | u16 count; |
---|
| 292 | ASSERT(spin_is_locked(&platform_timer_lock)); |
---|
| 293 | outb(0x80, PIT_MODE); |
---|
| 294 | count = inb(PIT_CH2); |
---|
| 295 | count |= inb(PIT_CH2) << 8; |
---|
| 296 | return count; |
---|
| 297 | } |
---|
| 298 | |
---|
| 299 | static void pit_overflow(void) |
---|
| 300 | { |
---|
| 301 | u16 counter; |
---|
| 302 | |
---|
| 303 | spin_lock_irq(&platform_timer_lock); |
---|
| 304 | counter = pit_read_counter(); |
---|
| 305 | pit_counter64 += (u16)(pit_stamp - counter); |
---|
| 306 | pit_stamp = counter; |
---|
| 307 | spin_unlock_irq(&platform_timer_lock); |
---|
| 308 | } |
---|
| 309 | |
---|
| 310 | static u64 read_pit_count(void) |
---|
| 311 | { |
---|
| 312 | return pit_counter64 + (u16)(pit_stamp - pit_read_counter()); |
---|
| 313 | } |
---|
| 314 | |
---|
| 315 | static void init_pit(void) |
---|
| 316 | { |
---|
| 317 | read_platform_count = read_pit_count; |
---|
| 318 | |
---|
| 319 | pit_overflow(); |
---|
| 320 | platform_timer_stamp = pit_counter64; |
---|
| 321 | set_time_scale(&platform_timer_scale, CLOCK_TICK_RATE); |
---|
| 322 | |
---|
| 323 | printk("Platform timer is %s PIT\n", freq_string(CLOCK_TICK_RATE)); |
---|
| 324 | using_pit = 1; |
---|
| 325 | } |
---|
| 326 | |
---|
| 327 | /************************************************************ |
---|
| 328 | * PLATFORM TIMER 2: HIGH PRECISION EVENT TIMER (HPET) |
---|
| 329 | */ |
---|
| 330 | |
---|
| 331 | /* Protected by platform_timer_lock. */ |
---|
| 332 | static u64 hpet_counter64, hpet_overflow_period; |
---|
| 333 | static u32 hpet_stamp; |
---|
| 334 | static struct timer hpet_overflow_timer; |
---|
| 335 | |
---|
| 336 | static void hpet_overflow(void *unused) |
---|
| 337 | { |
---|
| 338 | u32 counter; |
---|
| 339 | |
---|
| 340 | spin_lock_irq(&platform_timer_lock); |
---|
| 341 | counter = hpet_read32(HPET_COUNTER); |
---|
| 342 | hpet_counter64 += (u32)(counter - hpet_stamp); |
---|
| 343 | hpet_stamp = counter; |
---|
| 344 | spin_unlock_irq(&platform_timer_lock); |
---|
| 345 | |
---|
| 346 | set_timer(&hpet_overflow_timer, NOW() + hpet_overflow_period); |
---|
| 347 | } |
---|
| 348 | |
---|
| 349 | static u64 read_hpet_count(void) |
---|
| 350 | { |
---|
| 351 | return hpet_counter64 + (u32)(hpet_read32(HPET_COUNTER) - hpet_stamp); |
---|
| 352 | } |
---|
| 353 | |
---|
| 354 | static int init_hpet(void) |
---|
| 355 | { |
---|
| 356 | u64 hpet_rate; |
---|
| 357 | u32 hpet_id, hpet_period, cfg; |
---|
| 358 | int i; |
---|
| 359 | |
---|
| 360 | if ( (hpet_address == 0) && opt_hpet_force ) |
---|
| 361 | { |
---|
| 362 | outl(0x800038a0, 0xcf8); |
---|
| 363 | outl(0xff000001, 0xcfc); |
---|
| 364 | outl(0x800038a0, 0xcf8); |
---|
| 365 | hpet_address = inl(0xcfc) & 0xfffffffe; |
---|
| 366 | printk("WARNING: Forcibly enabled HPET at %#lx.\n", hpet_address); |
---|
| 367 | } |
---|
| 368 | |
---|
| 369 | if ( hpet_address == 0 ) |
---|
| 370 | return 0; |
---|
| 371 | |
---|
| 372 | set_fixmap_nocache(FIX_HPET_BASE, hpet_address); |
---|
| 373 | |
---|
| 374 | hpet_id = hpet_read32(HPET_ID); |
---|
| 375 | if ( hpet_id == 0 ) |
---|
| 376 | { |
---|
| 377 | printk("BAD HPET vendor id.\n"); |
---|
| 378 | return 0; |
---|
| 379 | } |
---|
| 380 | |
---|
| 381 | /* Check for sane period (100ps <= period <= 100ns). */ |
---|
| 382 | hpet_period = hpet_read32(HPET_PERIOD); |
---|
| 383 | if ( (hpet_period > 100000000) || (hpet_period < 100000) ) |
---|
| 384 | { |
---|
| 385 | printk("BAD HPET period %u.\n", hpet_period); |
---|
| 386 | return 0; |
---|
| 387 | } |
---|
| 388 | |
---|
| 389 | cfg = hpet_read32(HPET_CFG); |
---|
| 390 | cfg &= ~(HPET_CFG_ENABLE | HPET_CFG_LEGACY); |
---|
| 391 | hpet_write32(cfg, HPET_CFG); |
---|
| 392 | |
---|
| 393 | for ( i = 0; i <= ((hpet_id >> 8) & 31); i++ ) |
---|
| 394 | { |
---|
| 395 | cfg = hpet_read32(HPET_T0_CFG + i*0x20); |
---|
| 396 | cfg &= ~HPET_TN_ENABLE; |
---|
| 397 | hpet_write32(cfg & ~HPET_TN_ENABLE, HPET_T0_CFG); |
---|
| 398 | } |
---|
| 399 | |
---|
| 400 | cfg = hpet_read32(HPET_CFG); |
---|
| 401 | cfg |= HPET_CFG_ENABLE; |
---|
| 402 | hpet_write32(cfg, HPET_CFG); |
---|
| 403 | |
---|
| 404 | read_platform_count = read_hpet_count; |
---|
| 405 | |
---|
| 406 | hpet_rate = 1000000000000000ULL; /* 10^15 */ |
---|
| 407 | (void)do_div(hpet_rate, hpet_period); |
---|
| 408 | set_time_scale(&platform_timer_scale, hpet_rate); |
---|
| 409 | |
---|
| 410 | /* Trigger overflow avoidance roughly when counter increments 2^31. */ |
---|
| 411 | if ( (hpet_rate >> 31) != 0 ) |
---|
| 412 | { |
---|
| 413 | hpet_overflow_period = MILLISECS(1000); |
---|
| 414 | (void)do_div(hpet_overflow_period, (u32)(hpet_rate >> 31) + 1); |
---|
| 415 | } |
---|
| 416 | else |
---|
| 417 | { |
---|
| 418 | hpet_overflow_period = MILLISECS(1000) << 31; |
---|
| 419 | (void)do_div(hpet_overflow_period, (u32)hpet_rate); |
---|
| 420 | } |
---|
| 421 | |
---|
| 422 | init_timer(&hpet_overflow_timer, hpet_overflow, NULL, 0); |
---|
| 423 | hpet_overflow(NULL); |
---|
| 424 | platform_timer_stamp = hpet_counter64; |
---|
| 425 | |
---|
| 426 | printk("Platform timer is %s HPET\n", freq_string(hpet_rate)); |
---|
| 427 | |
---|
| 428 | return 1; |
---|
| 429 | } |
---|
| 430 | |
---|
| 431 | /************************************************************ |
---|
| 432 | * PLATFORM TIMER 3: IBM 'CYCLONE' TIMER |
---|
| 433 | */ |
---|
| 434 | |
---|
| 435 | int use_cyclone; |
---|
| 436 | |
---|
| 437 | /* |
---|
| 438 | * Although the counter is read via a 64-bit register, I believe it is actually |
---|
| 439 | * a 40-bit counter. Since this will wrap, I read only the low 32 bits and |
---|
| 440 | * periodically fold into a 64-bit software counter, just as for PIT and HPET. |
---|
| 441 | */ |
---|
| 442 | #define CYCLONE_CBAR_ADDR 0xFEB00CD0 |
---|
| 443 | #define CYCLONE_PMCC_OFFSET 0x51A0 |
---|
| 444 | #define CYCLONE_MPMC_OFFSET 0x51D0 |
---|
| 445 | #define CYCLONE_MPCS_OFFSET 0x51A8 |
---|
| 446 | #define CYCLONE_TIMER_FREQ 100000000 |
---|
| 447 | |
---|
| 448 | /* Protected by platform_timer_lock. */ |
---|
| 449 | static u64 cyclone_counter64; |
---|
| 450 | static u32 cyclone_stamp; |
---|
| 451 | static struct timer cyclone_overflow_timer; |
---|
| 452 | static volatile u32 *cyclone_timer; /* Cyclone MPMC0 register */ |
---|
| 453 | |
---|
| 454 | static void cyclone_overflow(void *unused) |
---|
| 455 | { |
---|
| 456 | u32 counter; |
---|
| 457 | |
---|
| 458 | spin_lock_irq(&platform_timer_lock); |
---|
| 459 | counter = *cyclone_timer; |
---|
| 460 | cyclone_counter64 += (u32)(counter - cyclone_stamp); |
---|
| 461 | cyclone_stamp = counter; |
---|
| 462 | spin_unlock_irq(&platform_timer_lock); |
---|
| 463 | |
---|
| 464 | set_timer(&cyclone_overflow_timer, NOW() + MILLISECS(20000)); |
---|
| 465 | } |
---|
| 466 | |
---|
| 467 | static u64 read_cyclone_count(void) |
---|
| 468 | { |
---|
| 469 | return cyclone_counter64 + (u32)(*cyclone_timer - cyclone_stamp); |
---|
| 470 | } |
---|
| 471 | |
---|
| 472 | static volatile u32 *map_cyclone_reg(unsigned long regaddr) |
---|
| 473 | { |
---|
| 474 | unsigned long pageaddr = regaddr & PAGE_MASK; |
---|
| 475 | unsigned long offset = regaddr & ~PAGE_MASK; |
---|
| 476 | set_fixmap_nocache(FIX_CYCLONE_TIMER, pageaddr); |
---|
| 477 | return (volatile u32 *)(fix_to_virt(FIX_CYCLONE_TIMER) + offset); |
---|
| 478 | } |
---|
| 479 | |
---|
| 480 | static int init_cyclone(void) |
---|
| 481 | { |
---|
| 482 | u32 base; |
---|
| 483 | |
---|
| 484 | if ( !use_cyclone ) |
---|
| 485 | return 0; |
---|
| 486 | |
---|
| 487 | /* Find base address. */ |
---|
| 488 | base = *(map_cyclone_reg(CYCLONE_CBAR_ADDR)); |
---|
| 489 | if ( base == 0 ) |
---|
| 490 | { |
---|
| 491 | printk(KERN_ERR "Cyclone: Could not find valid CBAR value.\n"); |
---|
| 492 | return 0; |
---|
| 493 | } |
---|
| 494 | |
---|
| 495 | /* Enable timer and map the counter register. */ |
---|
| 496 | *(map_cyclone_reg(base + CYCLONE_PMCC_OFFSET)) = 1; |
---|
| 497 | *(map_cyclone_reg(base + CYCLONE_MPCS_OFFSET)) = 1; |
---|
| 498 | cyclone_timer = map_cyclone_reg(base + CYCLONE_MPMC_OFFSET); |
---|
| 499 | |
---|
| 500 | read_platform_count = read_cyclone_count; |
---|
| 501 | |
---|
| 502 | init_timer(&cyclone_overflow_timer, cyclone_overflow, NULL, 0); |
---|
| 503 | cyclone_overflow(NULL); |
---|
| 504 | platform_timer_stamp = cyclone_counter64; |
---|
| 505 | set_time_scale(&platform_timer_scale, CYCLONE_TIMER_FREQ); |
---|
| 506 | |
---|
| 507 | printk("Platform timer is %s IBM Cyclone\n", |
---|
| 508 | freq_string(CYCLONE_TIMER_FREQ)); |
---|
| 509 | |
---|
| 510 | return 1; |
---|
| 511 | } |
---|
| 512 | |
---|
| 513 | /************************************************************ |
---|
| 514 | * GENERIC PLATFORM TIMER INFRASTRUCTURE |
---|
| 515 | */ |
---|
| 516 | |
---|
| 517 | static s_time_t __read_platform_stime(u64 platform_time) |
---|
| 518 | { |
---|
| 519 | u64 diff = platform_time - platform_timer_stamp; |
---|
| 520 | ASSERT(spin_is_locked(&platform_timer_lock)); |
---|
| 521 | return (stime_platform_stamp + scale_delta(diff, &platform_timer_scale)); |
---|
| 522 | } |
---|
| 523 | |
---|
| 524 | static s_time_t read_platform_stime(void) |
---|
| 525 | { |
---|
| 526 | u64 counter; |
---|
| 527 | s_time_t stime; |
---|
| 528 | |
---|
| 529 | spin_lock_irq(&platform_timer_lock); |
---|
| 530 | counter = read_platform_count(); |
---|
| 531 | stime = __read_platform_stime(counter); |
---|
| 532 | spin_unlock_irq(&platform_timer_lock); |
---|
| 533 | |
---|
| 534 | return stime; |
---|
| 535 | } |
---|
| 536 | |
---|
| 537 | static void platform_time_calibration(void) |
---|
| 538 | { |
---|
| 539 | u64 counter; |
---|
| 540 | s_time_t stamp; |
---|
| 541 | |
---|
| 542 | spin_lock_irq(&platform_timer_lock); |
---|
| 543 | counter = read_platform_count(); |
---|
| 544 | stamp = __read_platform_stime(counter); |
---|
| 545 | stime_platform_stamp = stamp; |
---|
| 546 | platform_timer_stamp = counter; |
---|
| 547 | spin_unlock_irq(&platform_timer_lock); |
---|
| 548 | } |
---|
| 549 | |
---|
| 550 | static void init_platform_timer(void) |
---|
| 551 | { |
---|
| 552 | if ( !init_cyclone() && !init_hpet() ) |
---|
| 553 | init_pit(); |
---|
| 554 | } |
---|
| 555 | |
---|
| 556 | |
---|
| 557 | /*************************************************************************** |
---|
| 558 | * CMOS Timer functions |
---|
| 559 | ***************************************************************************/ |
---|
| 560 | |
---|
| 561 | /* Converts Gregorian date to seconds since 1970-01-01 00:00:00. |
---|
| 562 | * Assumes input in normal date format, i.e. 1980-12-31 23:59:59 |
---|
| 563 | * => year=1980, mon=12, day=31, hour=23, min=59, sec=59. |
---|
| 564 | * |
---|
| 565 | * [For the Julian calendar (which was used in Russia before 1917, |
---|
| 566 | * Britain & colonies before 1752, anywhere else before 1582, |
---|
| 567 | * and is still in use by some communities) leave out the |
---|
| 568 | * -year/100+year/400 terms, and add 10.] |
---|
| 569 | * |
---|
| 570 | * This algorithm was first published by Gauss (I think). |
---|
| 571 | * |
---|
| 572 | * WARNING: this function will overflow on 2106-02-07 06:28:16 on |
---|
| 573 | * machines were long is 32-bit! (However, as time_t is signed, we |
---|
| 574 | * will already get problems at other places on 2038-01-19 03:14:08) |
---|
| 575 | */ |
---|
| 576 | unsigned long |
---|
| 577 | mktime (unsigned int year, unsigned int mon, |
---|
| 578 | unsigned int day, unsigned int hour, |
---|
| 579 | unsigned int min, unsigned int sec) |
---|
| 580 | { |
---|
| 581 | /* 1..12 -> 11,12,1..10: put Feb last since it has a leap day. */ |
---|
| 582 | if ( 0 >= (int) (mon -= 2) ) |
---|
| 583 | { |
---|
| 584 | mon += 12; |
---|
| 585 | year -= 1; |
---|
| 586 | } |
---|
| 587 | |
---|
| 588 | return ((((unsigned long)(year/4 - year/100 + year/400 + 367*mon/12 + day)+ |
---|
| 589 | year*365 - 719499 |
---|
| 590 | )*24 + hour /* now have hours */ |
---|
| 591 | )*60 + min /* now have minutes */ |
---|
| 592 | )*60 + sec; /* finally seconds */ |
---|
| 593 | } |
---|
| 594 | |
---|
| 595 | static unsigned long __get_cmos_time(void) |
---|
| 596 | { |
---|
| 597 | unsigned int year, mon, day, hour, min, sec; |
---|
| 598 | |
---|
| 599 | sec = CMOS_READ(RTC_SECONDS); |
---|
| 600 | min = CMOS_READ(RTC_MINUTES); |
---|
| 601 | hour = CMOS_READ(RTC_HOURS); |
---|
| 602 | day = CMOS_READ(RTC_DAY_OF_MONTH); |
---|
| 603 | mon = CMOS_READ(RTC_MONTH); |
---|
| 604 | year = CMOS_READ(RTC_YEAR); |
---|
| 605 | |
---|
| 606 | if ( !(CMOS_READ(RTC_CONTROL) & RTC_DM_BINARY) || RTC_ALWAYS_BCD ) |
---|
| 607 | { |
---|
| 608 | BCD_TO_BIN(sec); |
---|
| 609 | BCD_TO_BIN(min); |
---|
| 610 | BCD_TO_BIN(hour); |
---|
| 611 | BCD_TO_BIN(day); |
---|
| 612 | BCD_TO_BIN(mon); |
---|
| 613 | BCD_TO_BIN(year); |
---|
| 614 | } |
---|
| 615 | |
---|
| 616 | if ( (year += 1900) < 1970 ) |
---|
| 617 | year += 100; |
---|
| 618 | |
---|
| 619 | return mktime(year, mon, day, hour, min, sec); |
---|
| 620 | } |
---|
| 621 | |
---|
| 622 | static unsigned long get_cmos_time(void) |
---|
| 623 | { |
---|
| 624 | unsigned long res, flags; |
---|
| 625 | int i; |
---|
| 626 | |
---|
| 627 | spin_lock_irqsave(&rtc_lock, flags); |
---|
| 628 | |
---|
| 629 | /* read RTC exactly on falling edge of update flag */ |
---|
| 630 | for ( i = 0 ; i < 1000000 ; i++ ) /* may take up to 1 second... */ |
---|
| 631 | if ( (CMOS_READ(RTC_FREQ_SELECT) & RTC_UIP) ) |
---|
| 632 | break; |
---|
| 633 | for ( i = 0 ; i < 1000000 ; i++ ) /* must try at least 2.228 ms */ |
---|
| 634 | if ( !(CMOS_READ(RTC_FREQ_SELECT) & RTC_UIP) ) |
---|
| 635 | break; |
---|
| 636 | |
---|
| 637 | res = __get_cmos_time(); |
---|
| 638 | |
---|
| 639 | spin_unlock_irqrestore(&rtc_lock, flags); |
---|
| 640 | return res; |
---|
| 641 | } |
---|
| 642 | |
---|
| 643 | /*************************************************************************** |
---|
| 644 | * System Time |
---|
| 645 | ***************************************************************************/ |
---|
| 646 | |
---|
| 647 | s_time_t get_s_time(void) |
---|
| 648 | { |
---|
| 649 | struct cpu_time *t = &this_cpu(cpu_time); |
---|
| 650 | u64 tsc, delta; |
---|
| 651 | s_time_t now; |
---|
| 652 | |
---|
| 653 | rdtscll(tsc); |
---|
| 654 | delta = tsc - t->local_tsc_stamp; |
---|
| 655 | now = t->stime_local_stamp + scale_delta(delta, &t->tsc_scale); |
---|
| 656 | |
---|
| 657 | return now; |
---|
| 658 | } |
---|
| 659 | |
---|
| 660 | static inline void version_update_begin(u32 *version) |
---|
| 661 | { |
---|
| 662 | /* Explicitly OR with 1 just in case version number gets out of sync. */ |
---|
| 663 | *version = (*version + 1) | 1; |
---|
| 664 | wmb(); |
---|
| 665 | } |
---|
| 666 | |
---|
| 667 | static inline void version_update_end(u32 *version) |
---|
| 668 | { |
---|
| 669 | wmb(); |
---|
| 670 | (*version)++; |
---|
| 671 | } |
---|
| 672 | |
---|
| 673 | void update_vcpu_system_time(struct vcpu *v) |
---|
| 674 | { |
---|
| 675 | struct cpu_time *t; |
---|
| 676 | struct vcpu_time_info *u; |
---|
| 677 | |
---|
| 678 | if ( v->vcpu_info == NULL ) |
---|
| 679 | return; |
---|
| 680 | |
---|
| 681 | t = &this_cpu(cpu_time); |
---|
| 682 | u = &vcpu_info(v, time); |
---|
| 683 | |
---|
| 684 | if ( u->tsc_timestamp == t->local_tsc_stamp ) |
---|
| 685 | return; |
---|
| 686 | |
---|
| 687 | version_update_begin(&u->version); |
---|
| 688 | |
---|
| 689 | u->tsc_timestamp = t->local_tsc_stamp; |
---|
| 690 | u->system_time = t->stime_local_stamp; |
---|
| 691 | u->tsc_to_system_mul = t->tsc_scale.mul_frac; |
---|
| 692 | u->tsc_shift = (s8)t->tsc_scale.shift; |
---|
| 693 | |
---|
| 694 | version_update_end(&u->version); |
---|
| 695 | } |
---|
| 696 | |
---|
| 697 | void update_domain_wallclock_time(struct domain *d) |
---|
| 698 | { |
---|
| 699 | spin_lock(&wc_lock); |
---|
| 700 | version_update_begin(&shared_info(d, wc_version)); |
---|
| 701 | shared_info(d, wc_sec) = wc_sec + d->time_offset_seconds; |
---|
| 702 | shared_info(d, wc_nsec) = wc_nsec; |
---|
| 703 | version_update_end(&shared_info(d, wc_version)); |
---|
| 704 | spin_unlock(&wc_lock); |
---|
| 705 | } |
---|
| 706 | |
---|
| 707 | /* Set clock to <secs,usecs> after 00:00:00 UTC, 1 January, 1970. */ |
---|
| 708 | void do_settime(unsigned long secs, unsigned long nsecs, u64 system_time_base) |
---|
| 709 | { |
---|
| 710 | u64 x; |
---|
| 711 | u32 y, _wc_sec, _wc_nsec; |
---|
| 712 | struct domain *d; |
---|
| 713 | |
---|
| 714 | x = (secs * 1000000000ULL) + (u64)nsecs - system_time_base; |
---|
| 715 | y = do_div(x, 1000000000); |
---|
| 716 | |
---|
| 717 | spin_lock(&wc_lock); |
---|
| 718 | wc_sec = _wc_sec = (u32)x; |
---|
| 719 | wc_nsec = _wc_nsec = (u32)y; |
---|
| 720 | spin_unlock(&wc_lock); |
---|
| 721 | |
---|
| 722 | rcu_read_lock(&domlist_read_lock); |
---|
| 723 | for_each_domain ( d ) |
---|
| 724 | update_domain_wallclock_time(d); |
---|
| 725 | rcu_read_unlock(&domlist_read_lock); |
---|
| 726 | } |
---|
| 727 | |
---|
| 728 | static void local_time_calibration(void *unused) |
---|
| 729 | { |
---|
| 730 | struct cpu_time *t = &this_cpu(cpu_time); |
---|
| 731 | |
---|
| 732 | /* |
---|
| 733 | * System timestamps, extrapolated from local and master oscillators, |
---|
| 734 | * taken during this calibration and the previous calibration. |
---|
| 735 | */ |
---|
| 736 | s_time_t prev_local_stime, curr_local_stime; |
---|
| 737 | s_time_t prev_master_stime, curr_master_stime; |
---|
| 738 | |
---|
| 739 | /* TSC timestamps taken during this calibration and prev calibration. */ |
---|
| 740 | u64 prev_tsc, curr_tsc; |
---|
| 741 | |
---|
| 742 | /* |
---|
| 743 | * System time and TSC ticks elapsed during the previous calibration |
---|
| 744 | * 'epoch'. These values are down-shifted to fit in 32 bits. |
---|
| 745 | */ |
---|
| 746 | u64 stime_elapsed64, tsc_elapsed64; |
---|
| 747 | u32 stime_elapsed32, tsc_elapsed32; |
---|
| 748 | |
---|
| 749 | /* The accumulated error in the local estimate. */ |
---|
| 750 | u64 local_stime_err; |
---|
| 751 | |
---|
| 752 | /* Error correction to slow down a fast local clock. */ |
---|
| 753 | u32 error_factor = 0; |
---|
| 754 | |
---|
| 755 | /* Calculated TSC shift to ensure 32-bit scale multiplier. */ |
---|
| 756 | int tsc_shift = 0; |
---|
| 757 | |
---|
| 758 | /* The overall calibration scale multiplier. */ |
---|
| 759 | u32 calibration_mul_frac; |
---|
| 760 | |
---|
| 761 | prev_tsc = t->local_tsc_stamp; |
---|
| 762 | prev_local_stime = t->stime_local_stamp; |
---|
| 763 | prev_master_stime = t->stime_master_stamp; |
---|
| 764 | |
---|
| 765 | /* Disable IRQs to get 'instantaneous' current timestamps. */ |
---|
| 766 | local_irq_disable(); |
---|
| 767 | rdtscll(curr_tsc); |
---|
| 768 | curr_local_stime = get_s_time(); |
---|
| 769 | curr_master_stime = read_platform_stime(); |
---|
| 770 | local_irq_enable(); |
---|
| 771 | |
---|
| 772 | #if 0 |
---|
| 773 | printk("PRE%d: tsc=%"PRIu64" stime=%"PRIu64" master=%"PRIu64"\n", |
---|
| 774 | smp_processor_id(), prev_tsc, prev_local_stime, prev_master_stime); |
---|
| 775 | printk("CUR%d: tsc=%"PRIu64" stime=%"PRIu64" master=%"PRIu64 |
---|
| 776 | " -> %"PRId64"\n", |
---|
| 777 | smp_processor_id(), curr_tsc, curr_local_stime, curr_master_stime, |
---|
| 778 | curr_master_stime - curr_local_stime); |
---|
| 779 | #endif |
---|
| 780 | |
---|
| 781 | /* Local time warps forward if it lags behind master time. */ |
---|
| 782 | if ( curr_local_stime < curr_master_stime ) |
---|
| 783 | curr_local_stime = curr_master_stime; |
---|
| 784 | |
---|
| 785 | stime_elapsed64 = curr_master_stime - prev_master_stime; |
---|
| 786 | tsc_elapsed64 = curr_tsc - prev_tsc; |
---|
| 787 | |
---|
| 788 | /* |
---|
| 789 | * Weirdness can happen if we lose sync with the platform timer. |
---|
| 790 | * We could be smarter here: resync platform timer with local timer? |
---|
| 791 | */ |
---|
| 792 | if ( ((s64)stime_elapsed64 < (EPOCH / 2)) ) |
---|
| 793 | goto out; |
---|
| 794 | |
---|
| 795 | /* |
---|
| 796 | * Calculate error-correction factor. This only slows down a fast local |
---|
| 797 | * clock (slow clocks are warped forwards). The scale factor is clamped |
---|
| 798 | * to >= 0.5. |
---|
| 799 | */ |
---|
| 800 | if ( curr_local_stime != curr_master_stime ) |
---|
| 801 | { |
---|
| 802 | local_stime_err = curr_local_stime - curr_master_stime; |
---|
| 803 | if ( local_stime_err > EPOCH ) |
---|
| 804 | local_stime_err = EPOCH; |
---|
| 805 | error_factor = div_frac(EPOCH, EPOCH + (u32)local_stime_err); |
---|
| 806 | } |
---|
| 807 | |
---|
| 808 | /* |
---|
| 809 | * We require 0 < stime_elapsed < 2^31. |
---|
| 810 | * This allows us to binary shift a 32-bit tsc_elapsed such that: |
---|
| 811 | * stime_elapsed < tsc_elapsed <= 2*stime_elapsed |
---|
| 812 | */ |
---|
| 813 | while ( ((u32)stime_elapsed64 != stime_elapsed64) || |
---|
| 814 | ((s32)stime_elapsed64 < 0) ) |
---|
| 815 | { |
---|
| 816 | stime_elapsed64 >>= 1; |
---|
| 817 | tsc_elapsed64 >>= 1; |
---|
| 818 | } |
---|
| 819 | |
---|
| 820 | /* stime_master_diff now fits in a 32-bit word. */ |
---|
| 821 | stime_elapsed32 = (u32)stime_elapsed64; |
---|
| 822 | |
---|
| 823 | /* tsc_elapsed <= 2*stime_elapsed */ |
---|
| 824 | while ( tsc_elapsed64 > (stime_elapsed32 * 2) ) |
---|
| 825 | { |
---|
| 826 | tsc_elapsed64 >>= 1; |
---|
| 827 | tsc_shift--; |
---|
| 828 | } |
---|
| 829 | |
---|
| 830 | /* Local difference must now fit in 32 bits. */ |
---|
| 831 | ASSERT((u32)tsc_elapsed64 == tsc_elapsed64); |
---|
| 832 | tsc_elapsed32 = (u32)tsc_elapsed64; |
---|
| 833 | |
---|
| 834 | /* tsc_elapsed > stime_elapsed */ |
---|
| 835 | ASSERT(tsc_elapsed32 != 0); |
---|
| 836 | while ( tsc_elapsed32 <= stime_elapsed32 ) |
---|
| 837 | { |
---|
| 838 | tsc_elapsed32 <<= 1; |
---|
| 839 | tsc_shift++; |
---|
| 840 | } |
---|
| 841 | |
---|
| 842 | calibration_mul_frac = div_frac(stime_elapsed32, tsc_elapsed32); |
---|
| 843 | if ( error_factor != 0 ) |
---|
| 844 | calibration_mul_frac = mul_frac(calibration_mul_frac, error_factor); |
---|
| 845 | |
---|
| 846 | #if 0 |
---|
| 847 | printk("---%d: %08x %08x %d\n", smp_processor_id(), |
---|
| 848 | error_factor, calibration_mul_frac, tsc_shift); |
---|
| 849 | #endif |
---|
| 850 | |
---|
| 851 | /* Record new timestamp information. */ |
---|
| 852 | t->tsc_scale.mul_frac = calibration_mul_frac; |
---|
| 853 | t->tsc_scale.shift = tsc_shift; |
---|
| 854 | t->local_tsc_stamp = curr_tsc; |
---|
| 855 | t->stime_local_stamp = curr_local_stime; |
---|
| 856 | t->stime_master_stamp = curr_master_stime; |
---|
| 857 | |
---|
| 858 | update_vcpu_system_time(current); |
---|
| 859 | |
---|
| 860 | out: |
---|
| 861 | set_timer(&t->calibration_timer, NOW() + EPOCH); |
---|
| 862 | |
---|
| 863 | if ( smp_processor_id() == 0 ) |
---|
| 864 | platform_time_calibration(); |
---|
| 865 | } |
---|
| 866 | |
---|
| 867 | void init_percpu_time(void) |
---|
| 868 | { |
---|
| 869 | struct cpu_time *t = &this_cpu(cpu_time); |
---|
| 870 | unsigned long flags; |
---|
| 871 | s_time_t now; |
---|
| 872 | |
---|
| 873 | local_irq_save(flags); |
---|
| 874 | rdtscll(t->local_tsc_stamp); |
---|
| 875 | now = (smp_processor_id() == 0) ? 0 : read_platform_stime(); |
---|
| 876 | local_irq_restore(flags); |
---|
| 877 | |
---|
| 878 | t->stime_master_stamp = now; |
---|
| 879 | t->stime_local_stamp = now; |
---|
| 880 | |
---|
| 881 | init_timer(&t->calibration_timer, local_time_calibration, |
---|
| 882 | NULL, smp_processor_id()); |
---|
| 883 | set_timer(&t->calibration_timer, NOW() + EPOCH); |
---|
| 884 | } |
---|
| 885 | |
---|
| 886 | /* Late init function (after all CPUs are booted). */ |
---|
| 887 | int __init init_xen_time(void) |
---|
| 888 | { |
---|
| 889 | wc_sec = get_cmos_time(); |
---|
| 890 | |
---|
| 891 | local_irq_disable(); |
---|
| 892 | |
---|
| 893 | init_percpu_time(); |
---|
| 894 | |
---|
| 895 | stime_platform_stamp = 0; |
---|
| 896 | init_platform_timer(); |
---|
| 897 | |
---|
| 898 | local_irq_enable(); |
---|
| 899 | |
---|
| 900 | return 0; |
---|
| 901 | } |
---|
| 902 | |
---|
| 903 | |
---|
| 904 | /* Early init function. */ |
---|
| 905 | void __init early_time_init(void) |
---|
| 906 | { |
---|
| 907 | u64 tmp = calibrate_boot_tsc(); |
---|
| 908 | |
---|
| 909 | set_time_scale(&per_cpu(cpu_time, 0).tsc_scale, tmp); |
---|
| 910 | |
---|
| 911 | do_div(tmp, 1000); |
---|
| 912 | cpu_khz = (unsigned long)tmp; |
---|
| 913 | printk("Detected %lu.%03lu MHz processor.\n", |
---|
| 914 | cpu_khz / 1000, cpu_khz % 1000); |
---|
| 915 | |
---|
| 916 | setup_irq(0, &irq0); |
---|
| 917 | } |
---|
| 918 | |
---|
| 919 | void send_timer_event(struct vcpu *v) |
---|
| 920 | { |
---|
| 921 | send_guest_vcpu_virq(v, VIRQ_TIMER); |
---|
| 922 | } |
---|
| 923 | |
---|
| 924 | /* Return secs after 00:00:00 localtime, 1 January, 1970. */ |
---|
| 925 | unsigned long get_localtime(struct domain *d) |
---|
| 926 | { |
---|
| 927 | return wc_sec + (wc_nsec + NOW()) / 1000000000ULL |
---|
| 928 | + d->time_offset_seconds; |
---|
| 929 | } |
---|
| 930 | |
---|
| 931 | /* |
---|
| 932 | * Local variables: |
---|
| 933 | * mode: C |
---|
| 934 | * c-set-style: "BSD" |
---|
| 935 | * c-basic-offset: 4 |
---|
| 936 | * tab-width: 4 |
---|
| 937 | * indent-tabs-mode: nil |
---|
| 938 | * End: |
---|
| 939 | */ |
---|