1 | /* |
---|
2 | * Copyright (C) 2001 MandrakeSoft S.A. |
---|
3 | * |
---|
4 | * MandrakeSoft S.A. |
---|
5 | * 43, rue d'Aboukir |
---|
6 | * 75002 Paris - France |
---|
7 | * http://www.linux-mandrake.com/ |
---|
8 | * http://www.mandrakesoft.com/ |
---|
9 | * |
---|
10 | * This library is free software; you can redistribute it and/or |
---|
11 | * modify it under the terms of the GNU Lesser General Public |
---|
12 | * License as published by the Free Software Foundation; either |
---|
13 | * version 2 of the License, or (at your option) any later version. |
---|
14 | * |
---|
15 | * This library is distributed in the hope that it will be useful, |
---|
16 | * but WITHOUT ANY WARRANTY; without even the implied warranty of |
---|
17 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
---|
18 | * Lesser General Public License for more details. |
---|
19 | * |
---|
20 | * You should have received a copy of the GNU Lesser General Public |
---|
21 | * License along with this library; if not, write to the Free Software |
---|
22 | * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA |
---|
23 | * |
---|
24 | * Yunhong Jiang <yunhong.jiang@intel.com> |
---|
25 | * Ported to xen by using virtual IRQ line. |
---|
26 | */ |
---|
27 | |
---|
28 | #include <xen/config.h> |
---|
29 | #include <xen/types.h> |
---|
30 | #include <xen/mm.h> |
---|
31 | #include <xen/xmalloc.h> |
---|
32 | #include <xen/lib.h> |
---|
33 | #include <xen/errno.h> |
---|
34 | #include <xen/sched.h> |
---|
35 | #include <public/hvm/ioreq.h> |
---|
36 | #include <asm/hvm/io.h> |
---|
37 | #include <asm/hvm/vpic.h> |
---|
38 | #include <asm/hvm/vlapic.h> |
---|
39 | #include <asm/hvm/support.h> |
---|
40 | #include <asm/current.h> |
---|
41 | #include <asm/event.h> |
---|
42 | |
---|
43 | /* HACK: Route IRQ0 only to VCPU0 to prevent time jumps. */ |
---|
44 | #define IRQ0_SPECIAL_ROUTING 1 |
---|
45 | |
---|
46 | #if defined(__ia64__) |
---|
47 | #define opt_hvm_debug_level opt_vmx_debug_level |
---|
48 | #endif |
---|
49 | |
---|
50 | static void vioapic_deliver(struct hvm_hw_vioapic *vioapic, int irq); |
---|
51 | |
---|
52 | static unsigned long vioapic_read_indirect(struct hvm_hw_vioapic *vioapic, |
---|
53 | unsigned long addr, |
---|
54 | unsigned long length) |
---|
55 | { |
---|
56 | unsigned long result = 0; |
---|
57 | |
---|
58 | switch ( vioapic->ioregsel ) |
---|
59 | { |
---|
60 | case VIOAPIC_REG_VERSION: |
---|
61 | result = ((((VIOAPIC_NUM_PINS-1) & 0xff) << 16) |
---|
62 | | (VIOAPIC_VERSION_ID & 0xff)); |
---|
63 | break; |
---|
64 | |
---|
65 | #if !VIOAPIC_IS_IOSAPIC |
---|
66 | case VIOAPIC_REG_APIC_ID: |
---|
67 | case VIOAPIC_REG_ARB_ID: |
---|
68 | result = ((vioapic->id & 0xf) << 24); |
---|
69 | break; |
---|
70 | #endif |
---|
71 | |
---|
72 | default: |
---|
73 | { |
---|
74 | uint32_t redir_index = (vioapic->ioregsel - 0x10) >> 1; |
---|
75 | uint64_t redir_content; |
---|
76 | |
---|
77 | if ( redir_index >= VIOAPIC_NUM_PINS ) |
---|
78 | { |
---|
79 | gdprintk(XENLOG_WARNING, "apic_mem_readl:undefined ioregsel %x\n", |
---|
80 | vioapic->ioregsel); |
---|
81 | break; |
---|
82 | } |
---|
83 | |
---|
84 | redir_content = vioapic->redirtbl[redir_index].bits; |
---|
85 | result = (vioapic->ioregsel & 0x1)? |
---|
86 | (redir_content >> 32) & 0xffffffff : |
---|
87 | redir_content & 0xffffffff; |
---|
88 | break; |
---|
89 | } |
---|
90 | } |
---|
91 | |
---|
92 | return result; |
---|
93 | } |
---|
94 | |
---|
95 | static unsigned long vioapic_read(struct vcpu *v, |
---|
96 | unsigned long addr, |
---|
97 | unsigned long length) |
---|
98 | { |
---|
99 | struct hvm_hw_vioapic *vioapic = domain_vioapic(v->domain); |
---|
100 | uint32_t result; |
---|
101 | |
---|
102 | HVM_DBG_LOG(DBG_LEVEL_IOAPIC, "vioapic_read addr %lx\n", addr); |
---|
103 | |
---|
104 | addr &= 0xff; |
---|
105 | |
---|
106 | switch ( addr ) |
---|
107 | { |
---|
108 | case VIOAPIC_REG_SELECT: |
---|
109 | result = vioapic->ioregsel; |
---|
110 | break; |
---|
111 | |
---|
112 | case VIOAPIC_REG_WINDOW: |
---|
113 | result = vioapic_read_indirect(vioapic, addr, length); |
---|
114 | break; |
---|
115 | |
---|
116 | default: |
---|
117 | result = 0; |
---|
118 | break; |
---|
119 | } |
---|
120 | |
---|
121 | return result; |
---|
122 | } |
---|
123 | |
---|
124 | static void vioapic_write_redirent( |
---|
125 | struct hvm_hw_vioapic *vioapic, unsigned int idx, int top_word, uint32_t val) |
---|
126 | { |
---|
127 | struct domain *d = vioapic_domain(vioapic); |
---|
128 | struct hvm_irq *hvm_irq = &d->arch.hvm_domain.irq; |
---|
129 | union vioapic_redir_entry *pent, ent; |
---|
130 | |
---|
131 | spin_lock(&d->arch.hvm_domain.irq_lock); |
---|
132 | |
---|
133 | pent = &vioapic->redirtbl[idx]; |
---|
134 | ent = *pent; |
---|
135 | |
---|
136 | if ( top_word ) |
---|
137 | { |
---|
138 | /* Contains only the dest_id. */ |
---|
139 | ent.bits = (uint32_t)ent.bits | ((uint64_t)val << 32); |
---|
140 | } |
---|
141 | else |
---|
142 | { |
---|
143 | /* Remote IRR and Delivery Status are read-only. */ |
---|
144 | ent.bits = ((ent.bits >> 32) << 32) | val; |
---|
145 | ent.fields.delivery_status = 0; |
---|
146 | ent.fields.remote_irr = pent->fields.remote_irr; |
---|
147 | } |
---|
148 | |
---|
149 | *pent = ent; |
---|
150 | |
---|
151 | if ( (ent.fields.trig_mode == VIOAPIC_LEVEL_TRIG) && |
---|
152 | !ent.fields.mask && |
---|
153 | !ent.fields.remote_irr && |
---|
154 | hvm_irq->gsi_assert_count[idx] ) |
---|
155 | { |
---|
156 | pent->fields.remote_irr = 1; |
---|
157 | vioapic_deliver(vioapic, idx); |
---|
158 | } |
---|
159 | |
---|
160 | spin_unlock(&d->arch.hvm_domain.irq_lock); |
---|
161 | } |
---|
162 | |
---|
163 | static void vioapic_write_indirect( |
---|
164 | struct hvm_hw_vioapic *vioapic, unsigned long addr, |
---|
165 | unsigned long length, unsigned long val) |
---|
166 | { |
---|
167 | switch ( vioapic->ioregsel ) |
---|
168 | { |
---|
169 | case VIOAPIC_REG_VERSION: |
---|
170 | /* Writes are ignored. */ |
---|
171 | break; |
---|
172 | |
---|
173 | #if !VIOAPIC_IS_IOSAPIC |
---|
174 | case VIOAPIC_REG_APIC_ID: |
---|
175 | vioapic->id = (val >> 24) & 0xf; |
---|
176 | break; |
---|
177 | |
---|
178 | case VIOAPIC_REG_ARB_ID: |
---|
179 | break; |
---|
180 | #endif |
---|
181 | |
---|
182 | default: |
---|
183 | { |
---|
184 | uint32_t redir_index = (vioapic->ioregsel - 0x10) >> 1; |
---|
185 | |
---|
186 | HVM_DBG_LOG(DBG_LEVEL_IOAPIC, "vioapic_write_indirect " |
---|
187 | "change redir index %x val %lx\n", |
---|
188 | redir_index, val); |
---|
189 | |
---|
190 | if ( redir_index >= VIOAPIC_NUM_PINS ) |
---|
191 | { |
---|
192 | gdprintk(XENLOG_WARNING, "vioapic_write_indirect " |
---|
193 | "error register %x\n", vioapic->ioregsel); |
---|
194 | break; |
---|
195 | } |
---|
196 | |
---|
197 | vioapic_write_redirent( |
---|
198 | vioapic, redir_index, vioapic->ioregsel&1, val); |
---|
199 | break; |
---|
200 | } |
---|
201 | } |
---|
202 | } |
---|
203 | |
---|
204 | static void vioapic_write(struct vcpu *v, |
---|
205 | unsigned long addr, |
---|
206 | unsigned long length, |
---|
207 | unsigned long val) |
---|
208 | { |
---|
209 | struct hvm_hw_vioapic *vioapic = domain_vioapic(v->domain); |
---|
210 | |
---|
211 | addr &= 0xff; |
---|
212 | |
---|
213 | switch ( addr ) |
---|
214 | { |
---|
215 | case VIOAPIC_REG_SELECT: |
---|
216 | vioapic->ioregsel = val; |
---|
217 | break; |
---|
218 | |
---|
219 | case VIOAPIC_REG_WINDOW: |
---|
220 | vioapic_write_indirect(vioapic, addr, length, val); |
---|
221 | break; |
---|
222 | |
---|
223 | #if VIOAPIC_IS_IOSAPIC |
---|
224 | case VIOAPIC_REG_EOI: |
---|
225 | vioapic_update_EOI(v->domain, val); |
---|
226 | break; |
---|
227 | #endif |
---|
228 | |
---|
229 | default: |
---|
230 | break; |
---|
231 | } |
---|
232 | } |
---|
233 | |
---|
234 | static int vioapic_range(struct vcpu *v, unsigned long addr) |
---|
235 | { |
---|
236 | struct hvm_hw_vioapic *vioapic = domain_vioapic(v->domain); |
---|
237 | |
---|
238 | return ((addr >= vioapic->base_address && |
---|
239 | (addr < vioapic->base_address + VIOAPIC_MEM_LENGTH))); |
---|
240 | } |
---|
241 | |
---|
242 | struct hvm_mmio_handler vioapic_mmio_handler = { |
---|
243 | .check_handler = vioapic_range, |
---|
244 | .read_handler = vioapic_read, |
---|
245 | .write_handler = vioapic_write |
---|
246 | }; |
---|
247 | |
---|
248 | static void ioapic_inj_irq( |
---|
249 | struct hvm_hw_vioapic *vioapic, |
---|
250 | struct vlapic *target, |
---|
251 | uint8_t vector, |
---|
252 | uint8_t trig_mode, |
---|
253 | uint8_t delivery_mode) |
---|
254 | { |
---|
255 | HVM_DBG_LOG(DBG_LEVEL_IOAPIC, "ioapic_inj_irq " |
---|
256 | "irq %d trig %d delive mode %d\n", |
---|
257 | vector, trig_mode, delivery_mode); |
---|
258 | |
---|
259 | switch ( delivery_mode ) |
---|
260 | { |
---|
261 | case dest_Fixed: |
---|
262 | case dest_LowestPrio: |
---|
263 | if ( vlapic_set_irq(target, vector, trig_mode) ) |
---|
264 | vcpu_kick(vlapic_vcpu(target)); |
---|
265 | break; |
---|
266 | default: |
---|
267 | gdprintk(XENLOG_WARNING, "error delivery mode %d\n", delivery_mode); |
---|
268 | break; |
---|
269 | } |
---|
270 | } |
---|
271 | |
---|
272 | static uint32_t ioapic_get_delivery_bitmask( |
---|
273 | struct hvm_hw_vioapic *vioapic, uint16_t dest, uint8_t dest_mode) |
---|
274 | { |
---|
275 | uint32_t mask = 0; |
---|
276 | struct vcpu *v; |
---|
277 | |
---|
278 | HVM_DBG_LOG(DBG_LEVEL_IOAPIC, "ioapic_get_delivery_bitmask " |
---|
279 | "dest %d dest_mode %d\n", dest, dest_mode); |
---|
280 | |
---|
281 | if ( dest_mode == 0 ) /* Physical mode. */ |
---|
282 | { |
---|
283 | if ( dest == 0xFF ) /* Broadcast. */ |
---|
284 | { |
---|
285 | for_each_vcpu ( vioapic_domain(vioapic), v ) |
---|
286 | mask |= 1 << v->vcpu_id; |
---|
287 | goto out; |
---|
288 | } |
---|
289 | |
---|
290 | for_each_vcpu ( vioapic_domain(vioapic), v ) |
---|
291 | { |
---|
292 | if ( VLAPIC_ID(vcpu_vlapic(v)) == dest ) |
---|
293 | { |
---|
294 | mask = 1 << v->vcpu_id; |
---|
295 | break; |
---|
296 | } |
---|
297 | } |
---|
298 | } |
---|
299 | else if ( dest != 0 ) /* Logical mode, MDA non-zero. */ |
---|
300 | { |
---|
301 | for_each_vcpu ( vioapic_domain(vioapic), v ) |
---|
302 | if ( vlapic_match_logical_addr(vcpu_vlapic(v), dest) ) |
---|
303 | mask |= 1 << v->vcpu_id; |
---|
304 | } |
---|
305 | |
---|
306 | out: |
---|
307 | HVM_DBG_LOG(DBG_LEVEL_IOAPIC, "ioapic_get_delivery_bitmask mask %x\n", |
---|
308 | mask); |
---|
309 | return mask; |
---|
310 | } |
---|
311 | |
---|
312 | static inline int pit_channel0_enabled(void) |
---|
313 | { |
---|
314 | PITState *pit = ¤t->domain->arch.hvm_domain.pl_time.vpit; |
---|
315 | struct periodic_time *pt = &pit->pt[0]; |
---|
316 | return pt->enabled; |
---|
317 | } |
---|
318 | |
---|
319 | static void vioapic_deliver(struct hvm_hw_vioapic *vioapic, int irq) |
---|
320 | { |
---|
321 | uint16_t dest = vioapic->redirtbl[irq].fields.dest_id; |
---|
322 | uint8_t dest_mode = vioapic->redirtbl[irq].fields.dest_mode; |
---|
323 | uint8_t delivery_mode = vioapic->redirtbl[irq].fields.delivery_mode; |
---|
324 | uint8_t vector = vioapic->redirtbl[irq].fields.vector; |
---|
325 | uint8_t trig_mode = vioapic->redirtbl[irq].fields.trig_mode; |
---|
326 | uint32_t deliver_bitmask; |
---|
327 | struct vlapic *target; |
---|
328 | struct vcpu *v; |
---|
329 | |
---|
330 | ASSERT(spin_is_locked(&vioapic_domain(vioapic)->arch.hvm_domain.irq_lock)); |
---|
331 | |
---|
332 | HVM_DBG_LOG(DBG_LEVEL_IOAPIC, |
---|
333 | "dest=%x dest_mode=%x delivery_mode=%x " |
---|
334 | "vector=%x trig_mode=%x\n", |
---|
335 | dest, dest_mode, delivery_mode, vector, trig_mode); |
---|
336 | |
---|
337 | deliver_bitmask = ioapic_get_delivery_bitmask(vioapic, dest, dest_mode); |
---|
338 | if ( !deliver_bitmask ) |
---|
339 | { |
---|
340 | HVM_DBG_LOG(DBG_LEVEL_IOAPIC, "ioapic deliver " |
---|
341 | "no target on destination\n"); |
---|
342 | return; |
---|
343 | } |
---|
344 | |
---|
345 | switch ( delivery_mode ) |
---|
346 | { |
---|
347 | case dest_LowestPrio: |
---|
348 | { |
---|
349 | #ifdef IRQ0_SPECIAL_ROUTING |
---|
350 | /* Force round-robin to pick VCPU 0 */ |
---|
351 | if ( (irq == hvm_isa_irq_to_gsi(0)) && pit_channel0_enabled() ) |
---|
352 | { |
---|
353 | v = vioapic_domain(vioapic)->vcpu[0]; |
---|
354 | target = v ? vcpu_vlapic(v) : NULL; |
---|
355 | } |
---|
356 | else |
---|
357 | #endif |
---|
358 | target = apic_round_robin(vioapic_domain(vioapic), |
---|
359 | vector, deliver_bitmask); |
---|
360 | if ( target != NULL ) |
---|
361 | { |
---|
362 | ioapic_inj_irq(vioapic, target, vector, trig_mode, delivery_mode); |
---|
363 | } |
---|
364 | else |
---|
365 | { |
---|
366 | HVM_DBG_LOG(DBG_LEVEL_IOAPIC, "null round robin: " |
---|
367 | "mask=%x vector=%x delivery_mode=%x\n", |
---|
368 | deliver_bitmask, vector, dest_LowestPrio); |
---|
369 | } |
---|
370 | break; |
---|
371 | } |
---|
372 | |
---|
373 | case dest_Fixed: |
---|
374 | case dest_ExtINT: |
---|
375 | { |
---|
376 | uint8_t bit; |
---|
377 | for ( bit = 0; deliver_bitmask != 0; bit++ ) |
---|
378 | { |
---|
379 | if ( !(deliver_bitmask & (1 << bit)) ) |
---|
380 | continue; |
---|
381 | deliver_bitmask &= ~(1 << bit); |
---|
382 | #ifdef IRQ0_SPECIAL_ROUTING |
---|
383 | /* Do not deliver timer interrupts to VCPU != 0 */ |
---|
384 | if ( (irq == hvm_isa_irq_to_gsi(0)) && pit_channel0_enabled() ) |
---|
385 | v = vioapic_domain(vioapic)->vcpu[0]; |
---|
386 | else |
---|
387 | #endif |
---|
388 | v = vioapic_domain(vioapic)->vcpu[bit]; |
---|
389 | if ( v != NULL ) |
---|
390 | { |
---|
391 | target = vcpu_vlapic(v); |
---|
392 | ioapic_inj_irq(vioapic, target, vector, |
---|
393 | trig_mode, delivery_mode); |
---|
394 | } |
---|
395 | } |
---|
396 | break; |
---|
397 | } |
---|
398 | |
---|
399 | case dest_SMI: |
---|
400 | case dest_NMI: |
---|
401 | case dest_INIT: |
---|
402 | case dest__reserved_2: |
---|
403 | default: |
---|
404 | gdprintk(XENLOG_WARNING, "Unsupported delivery mode %d\n", |
---|
405 | delivery_mode); |
---|
406 | break; |
---|
407 | } |
---|
408 | } |
---|
409 | |
---|
410 | void vioapic_irq_positive_edge(struct domain *d, unsigned int irq) |
---|
411 | { |
---|
412 | struct hvm_hw_vioapic *vioapic = domain_vioapic(d); |
---|
413 | union vioapic_redir_entry *ent; |
---|
414 | |
---|
415 | HVM_DBG_LOG(DBG_LEVEL_IOAPIC, "ioapic_irq_positive_edge irq %x", irq); |
---|
416 | |
---|
417 | ASSERT(irq < VIOAPIC_NUM_PINS); |
---|
418 | ASSERT(spin_is_locked(&d->arch.hvm_domain.irq_lock)); |
---|
419 | |
---|
420 | ent = &vioapic->redirtbl[irq]; |
---|
421 | if ( ent->fields.mask ) |
---|
422 | return; |
---|
423 | |
---|
424 | if ( ent->fields.trig_mode == VIOAPIC_EDGE_TRIG ) |
---|
425 | { |
---|
426 | vioapic_deliver(vioapic, irq); |
---|
427 | } |
---|
428 | else if ( !ent->fields.remote_irr ) |
---|
429 | { |
---|
430 | ent->fields.remote_irr = 1; |
---|
431 | vioapic_deliver(vioapic, irq); |
---|
432 | } |
---|
433 | } |
---|
434 | |
---|
435 | static int get_eoi_gsi(struct hvm_hw_vioapic *vioapic, int vector) |
---|
436 | { |
---|
437 | int i; |
---|
438 | |
---|
439 | for ( i = 0; i < VIOAPIC_NUM_PINS; i++ ) |
---|
440 | if ( vioapic->redirtbl[i].fields.vector == vector ) |
---|
441 | return i; |
---|
442 | |
---|
443 | return -1; |
---|
444 | } |
---|
445 | |
---|
446 | void vioapic_update_EOI(struct domain *d, int vector) |
---|
447 | { |
---|
448 | struct hvm_hw_vioapic *vioapic = domain_vioapic(d); |
---|
449 | struct hvm_irq *hvm_irq = &d->arch.hvm_domain.irq; |
---|
450 | union vioapic_redir_entry *ent; |
---|
451 | int gsi; |
---|
452 | |
---|
453 | spin_lock(&d->arch.hvm_domain.irq_lock); |
---|
454 | |
---|
455 | if ( (gsi = get_eoi_gsi(vioapic, vector)) == -1 ) |
---|
456 | { |
---|
457 | gdprintk(XENLOG_WARNING, "Can't find redir item for %d EOI\n", vector); |
---|
458 | goto out; |
---|
459 | } |
---|
460 | |
---|
461 | ent = &vioapic->redirtbl[gsi]; |
---|
462 | |
---|
463 | ent->fields.remote_irr = 0; |
---|
464 | if ( (ent->fields.trig_mode == VIOAPIC_LEVEL_TRIG) && |
---|
465 | !ent->fields.mask && |
---|
466 | hvm_irq->gsi_assert_count[gsi] ) |
---|
467 | { |
---|
468 | ent->fields.remote_irr = 1; |
---|
469 | vioapic_deliver(vioapic, gsi); |
---|
470 | } |
---|
471 | |
---|
472 | out: |
---|
473 | spin_unlock(&d->arch.hvm_domain.irq_lock); |
---|
474 | } |
---|
475 | |
---|
476 | #ifdef HVM_DEBUG_SUSPEND |
---|
477 | static void ioapic_info(struct hvm_hw_vioapic *s) |
---|
478 | { |
---|
479 | int i; |
---|
480 | printk("*****ioapic state:*****\n"); |
---|
481 | printk("ioapic 0x%x.\n", s->ioregsel); |
---|
482 | printk("ioapic 0x%x.\n", s->id); |
---|
483 | printk("ioapic 0x%lx.\n", s->base_address); |
---|
484 | for (i = 0; i < VIOAPIC_NUM_PINS; i++) { |
---|
485 | printk("ioapic redirtbl[%d]:0x%"PRIx64"\n", i, s->redirtbl[i].bits); |
---|
486 | } |
---|
487 | |
---|
488 | } |
---|
489 | #else |
---|
490 | static void ioapic_info(struct hvm_hw_vioapic *s) |
---|
491 | { |
---|
492 | } |
---|
493 | #endif |
---|
494 | |
---|
495 | |
---|
496 | static int ioapic_save(struct domain *d, hvm_domain_context_t *h) |
---|
497 | { |
---|
498 | struct hvm_hw_vioapic *s = domain_vioapic(d); |
---|
499 | ioapic_info(s); |
---|
500 | |
---|
501 | /* save io-apic state*/ |
---|
502 | return ( hvm_save_entry(IOAPIC, 0, h, s) ); |
---|
503 | } |
---|
504 | |
---|
505 | static int ioapic_load(struct domain *d, hvm_domain_context_t *h) |
---|
506 | { |
---|
507 | struct hvm_hw_vioapic *s = domain_vioapic(d); |
---|
508 | |
---|
509 | /* restore ioapic state */ |
---|
510 | if ( hvm_load_entry(IOAPIC, h, s) != 0 ) |
---|
511 | return -EINVAL; |
---|
512 | |
---|
513 | ioapic_info(s); |
---|
514 | return 0; |
---|
515 | } |
---|
516 | |
---|
517 | HVM_REGISTER_SAVE_RESTORE(IOAPIC, ioapic_save, ioapic_load, 1, HVMSR_PER_DOM); |
---|
518 | |
---|
519 | void vioapic_init(struct domain *d) |
---|
520 | { |
---|
521 | struct hvm_hw_vioapic *vioapic = domain_vioapic(d); |
---|
522 | int i; |
---|
523 | |
---|
524 | memset(vioapic, 0, sizeof(*vioapic)); |
---|
525 | for ( i = 0; i < VIOAPIC_NUM_PINS; i++ ) |
---|
526 | vioapic->redirtbl[i].fields.mask = 1; |
---|
527 | vioapic->base_address = VIOAPIC_DEFAULT_BASE_ADDRESS; |
---|
528 | } |
---|