1 | /* |
---|
2 | * This program is free software; you can redistribute it and/or modify |
---|
3 | * it under the terms of the GNU General Public License as published by |
---|
4 | * the Free Software Foundation; either version 2 of the License, or |
---|
5 | * (at your option) any later version. |
---|
6 | * |
---|
7 | * This program is distributed in the hope that it will be useful, |
---|
8 | * but WITHOUT ANY WARRANTY; without even the implied warranty of |
---|
9 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
---|
10 | * GNU General Public License for more details. |
---|
11 | * |
---|
12 | * You should have received a copy of the GNU General Public License |
---|
13 | * along with this program; if not, write to the Free Software |
---|
14 | * Foundation, 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. |
---|
15 | * |
---|
16 | * Copyright (C) IBM Corp. 2005, 2006 |
---|
17 | * |
---|
18 | * Authors: Hollis Blanchard <hollisb@us.ibm.com> |
---|
19 | * Jimi Xenidis <jimix@watson.ibm.com> |
---|
20 | * Ryan Harper <ryanh@us.ibm.com> |
---|
21 | */ |
---|
22 | |
---|
23 | #include <xen/config.h> |
---|
24 | #include <xen/mm.h> |
---|
25 | #include <xen/shadow.h> |
---|
26 | #include <xen/kernel.h> |
---|
27 | #include <xen/sched.h> |
---|
28 | #include <xen/perfc.h> |
---|
29 | #include <asm/init.h> |
---|
30 | #include <asm/page.h> |
---|
31 | #include <asm/platform.h> |
---|
32 | #include <asm/string.h> |
---|
33 | #include <asm/platform.h> |
---|
34 | #include <public/arch-powerpc.h> |
---|
35 | |
---|
36 | #ifdef VERBOSE |
---|
37 | #define MEM_LOG(_f, _a...) \ |
---|
38 | printk("DOM%u: (file=mm.c, line=%d) " _f "\n", \ |
---|
39 | current->domain->domain_id , __LINE__ , ## _a ) |
---|
40 | #else |
---|
41 | #define MEM_LOG(_f, _a...) ((void)0) |
---|
42 | #endif |
---|
43 | |
---|
44 | /* Frame table and its size in pages. */ |
---|
45 | struct page_info *frame_table; |
---|
46 | unsigned long max_page; |
---|
47 | unsigned long total_pages; |
---|
48 | |
---|
49 | /* machine to phys mapping to used by all domains */ |
---|
50 | unsigned long *machine_phys_mapping; |
---|
51 | |
---|
52 | void __init init_frametable(void) |
---|
53 | { |
---|
54 | unsigned long p; |
---|
55 | unsigned long nr_pages; |
---|
56 | int i; |
---|
57 | |
---|
58 | nr_pages = PFN_UP(max_page * sizeof(struct page_info)); |
---|
59 | |
---|
60 | p = alloc_boot_pages(nr_pages, 1); |
---|
61 | if (p == 0) |
---|
62 | panic("Not enough memory for frame table\n"); |
---|
63 | |
---|
64 | frame_table = (struct page_info *)(p << PAGE_SHIFT); |
---|
65 | for (i = 0; i < nr_pages; i += 1) |
---|
66 | clear_page((void *)((p + i) << PAGE_SHIFT)); |
---|
67 | } |
---|
68 | |
---|
69 | /* Array of PFNs, indexed by MFN. */ |
---|
70 | void __init init_machine_to_phys_table(void) |
---|
71 | { |
---|
72 | unsigned long p; |
---|
73 | unsigned long nr_pages; |
---|
74 | int i; |
---|
75 | |
---|
76 | nr_pages = PFN_UP(max_page * sizeof(unsigned long)); |
---|
77 | |
---|
78 | p = alloc_boot_pages(nr_pages, 1); |
---|
79 | if (p == 0) |
---|
80 | panic("Not enough memory for machine phys mapping table\n"); |
---|
81 | |
---|
82 | machine_phys_mapping = (unsigned long *)(p << PAGE_SHIFT); |
---|
83 | for (i = 0; i < nr_pages; i += 1) |
---|
84 | clear_page((void *)((p + i) << PAGE_SHIFT)); |
---|
85 | } |
---|
86 | |
---|
87 | void share_xen_page_with_guest( |
---|
88 | struct page_info *page, struct domain *d, int readonly) |
---|
89 | { |
---|
90 | if ( page_get_owner(page) == d ) |
---|
91 | return; |
---|
92 | |
---|
93 | /* this causes us to leak pages in the Domain and reuslts in |
---|
94 | * Zombie domains, I think we are missing a piece, until we find |
---|
95 | * it we disable the following code */ |
---|
96 | set_gpfn_from_mfn(page_to_mfn(page), INVALID_M2P_ENTRY); |
---|
97 | |
---|
98 | spin_lock(&d->page_alloc_lock); |
---|
99 | |
---|
100 | /* The incremented type count pins as writable or read-only. */ |
---|
101 | page->u.inuse.type_info = (readonly ? PGT_none : PGT_writable_page); |
---|
102 | page->u.inuse.type_info |= PGT_validated | 1; |
---|
103 | |
---|
104 | page_set_owner(page, d); |
---|
105 | wmb(); /* install valid domain ptr before updating refcnt. */ |
---|
106 | ASSERT(page->count_info == 0); |
---|
107 | |
---|
108 | /* Only add to the allocation list if the domain isn't dying. */ |
---|
109 | if ( !d->is_dying ) |
---|
110 | { |
---|
111 | page->count_info |= PGC_allocated | 1; |
---|
112 | if ( unlikely(d->xenheap_pages++ == 0) ) |
---|
113 | get_knownalive_domain(d); |
---|
114 | list_add_tail(&page->list, &d->xenpage_list); |
---|
115 | } |
---|
116 | |
---|
117 | spin_unlock(&d->page_alloc_lock); |
---|
118 | } |
---|
119 | |
---|
120 | void share_xen_page_with_privileged_guests( |
---|
121 | struct page_info *page, int readonly) |
---|
122 | { |
---|
123 | unimplemented(); |
---|
124 | } |
---|
125 | |
---|
126 | static ulong foreign_to_mfn(struct domain *d, ulong pfn) |
---|
127 | { |
---|
128 | |
---|
129 | pfn -= 1UL << cpu_foreign_map_order(); |
---|
130 | |
---|
131 | BUG_ON(pfn >= d->arch.foreign_mfn_count); |
---|
132 | |
---|
133 | return d->arch.foreign_mfns[pfn]; |
---|
134 | } |
---|
135 | |
---|
136 | static int set_foreign(struct domain *d, ulong pfn, ulong mfn) |
---|
137 | { |
---|
138 | pfn -= 1UL << cpu_foreign_map_order(); |
---|
139 | |
---|
140 | BUG_ON(pfn >= d->arch.foreign_mfn_count); |
---|
141 | d->arch.foreign_mfns[pfn] = mfn; |
---|
142 | |
---|
143 | return 0; |
---|
144 | } |
---|
145 | |
---|
146 | static int create_grant_va_mapping( |
---|
147 | unsigned long va, unsigned long frame, struct vcpu *v) |
---|
148 | { |
---|
149 | if (v->domain->domain_id != 0) { |
---|
150 | printk("only Dom0 can map a grant entry\n"); |
---|
151 | BUG(); |
---|
152 | return GNTST_permission_denied; |
---|
153 | } |
---|
154 | set_foreign(v->domain, va >> PAGE_SHIFT, frame); |
---|
155 | return GNTST_okay; |
---|
156 | } |
---|
157 | |
---|
158 | static int destroy_grant_va_mapping( |
---|
159 | unsigned long addr, unsigned long frame, struct domain *d) |
---|
160 | { |
---|
161 | if (d->domain_id != 0) { |
---|
162 | printk("only Dom0 can map a grant entry\n"); |
---|
163 | BUG(); |
---|
164 | return GNTST_permission_denied; |
---|
165 | } |
---|
166 | set_foreign(d, addr >> PAGE_SHIFT, ~0UL); |
---|
167 | return GNTST_okay; |
---|
168 | } |
---|
169 | |
---|
170 | int create_grant_host_mapping( |
---|
171 | unsigned long addr, unsigned long frame, unsigned int flags) |
---|
172 | { |
---|
173 | if (flags & GNTMAP_application_map) { |
---|
174 | printk("%s: GNTMAP_application_map not supported\n", __func__); |
---|
175 | BUG(); |
---|
176 | return GNTST_general_error; |
---|
177 | } |
---|
178 | if (flags & GNTMAP_contains_pte) { |
---|
179 | printk("%s: GNTMAP_contains_pte not supported\n", __func__); |
---|
180 | BUG(); |
---|
181 | return GNTST_general_error; |
---|
182 | } |
---|
183 | return create_grant_va_mapping(addr, frame, current); |
---|
184 | } |
---|
185 | |
---|
186 | int destroy_grant_host_mapping( |
---|
187 | unsigned long addr, unsigned long frame, unsigned int flags) |
---|
188 | { |
---|
189 | if (flags & GNTMAP_contains_pte) { |
---|
190 | printk("%s: GNTMAP_contains_pte not supported\n", __func__); |
---|
191 | BUG(); |
---|
192 | return GNTST_general_error; |
---|
193 | } |
---|
194 | |
---|
195 | /* may have force the remove here */ |
---|
196 | return destroy_grant_va_mapping(addr, frame, current->domain); |
---|
197 | } |
---|
198 | |
---|
199 | int steal_page(struct domain *d, struct page_info *page, unsigned int memflags) |
---|
200 | { |
---|
201 | panic("%s called\n", __func__); |
---|
202 | return 1; |
---|
203 | } |
---|
204 | |
---|
205 | void put_page_type(struct page_info *page) |
---|
206 | { |
---|
207 | unsigned long nx, x, y = page->u.inuse.type_info; |
---|
208 | |
---|
209 | do { |
---|
210 | x = y; |
---|
211 | nx = x - 1; |
---|
212 | |
---|
213 | ASSERT((x & PGT_count_mask) != 0); |
---|
214 | |
---|
215 | /* |
---|
216 | * The page should always be validated while a reference is held. The |
---|
217 | * exception is during domain destruction, when we forcibly invalidate |
---|
218 | * page-table pages if we detect a referential loop. |
---|
219 | * See domain.c:relinquish_list(). |
---|
220 | */ |
---|
221 | ASSERT((x & PGT_validated) || page_get_owner(page)->is_dying); |
---|
222 | |
---|
223 | if ( unlikely((nx & PGT_count_mask) == 0) ) |
---|
224 | { |
---|
225 | /* Record TLB information for flush later. */ |
---|
226 | page->tlbflush_timestamp = tlbflush_current_time(); |
---|
227 | } |
---|
228 | } |
---|
229 | while ( unlikely((y = cmpxchg(&page->u.inuse.type_info, x, nx)) != x) ); |
---|
230 | } |
---|
231 | |
---|
232 | |
---|
233 | int get_page_type(struct page_info *page, unsigned long type) |
---|
234 | { |
---|
235 | unsigned long nx, x, y = page->u.inuse.type_info; |
---|
236 | |
---|
237 | ASSERT(!(type & ~PGT_type_mask)); |
---|
238 | |
---|
239 | again: |
---|
240 | do { |
---|
241 | x = y; |
---|
242 | nx = x + 1; |
---|
243 | if ( unlikely((nx & PGT_count_mask) == 0) ) |
---|
244 | { |
---|
245 | MEM_LOG("Type count overflow on pfn %lx", page_to_mfn(page)); |
---|
246 | return 0; |
---|
247 | } |
---|
248 | else if ( unlikely((x & PGT_count_mask) == 0) ) |
---|
249 | { |
---|
250 | if ( (x & PGT_type_mask) != type ) |
---|
251 | { |
---|
252 | /* |
---|
253 | * On type change we check to flush stale TLB entries. This |
---|
254 | * may be unnecessary (e.g., page was GDT/LDT) but those |
---|
255 | * circumstances should be very rare. |
---|
256 | */ |
---|
257 | cpumask_t mask = |
---|
258 | page_get_owner(page)->domain_dirty_cpumask; |
---|
259 | tlbflush_filter(mask, page->tlbflush_timestamp); |
---|
260 | |
---|
261 | if ( unlikely(!cpus_empty(mask)) ) |
---|
262 | { |
---|
263 | perfc_incr(need_flush_tlb_flush); |
---|
264 | flush_tlb_mask(mask); |
---|
265 | } |
---|
266 | |
---|
267 | /* We lose existing type, back pointer, and validity. */ |
---|
268 | nx &= ~(PGT_type_mask | PGT_validated); |
---|
269 | nx |= type; |
---|
270 | |
---|
271 | /* No special validation needed for writable pages. */ |
---|
272 | /* Page tables and GDT/LDT need to be scanned for validity. */ |
---|
273 | if ( type == PGT_writable_page ) |
---|
274 | nx |= PGT_validated; |
---|
275 | } |
---|
276 | } |
---|
277 | else if ( unlikely((x & PGT_type_mask) != type) ) |
---|
278 | { |
---|
279 | return 0; |
---|
280 | } |
---|
281 | else if ( unlikely(!(x & PGT_validated)) ) |
---|
282 | { |
---|
283 | /* Someone else is updating validation of this page. Wait... */ |
---|
284 | while ( (y = page->u.inuse.type_info) == x ) |
---|
285 | cpu_relax(); |
---|
286 | goto again; |
---|
287 | } |
---|
288 | } |
---|
289 | while ( unlikely((y = cmpxchg(&page->u.inuse.type_info, x, nx)) != x) ); |
---|
290 | |
---|
291 | if ( unlikely(!(nx & PGT_validated)) ) |
---|
292 | { |
---|
293 | /* Noone else is updating simultaneously. */ |
---|
294 | __set_bit(_PGT_validated, &page->u.inuse.type_info); |
---|
295 | } |
---|
296 | |
---|
297 | return 1; |
---|
298 | } |
---|
299 | |
---|
300 | long arch_memory_op(int op, XEN_GUEST_HANDLE(void) arg) |
---|
301 | { |
---|
302 | printk("%s: no PPC specific memory ops\n", __func__); |
---|
303 | return -ENOSYS; |
---|
304 | } |
---|
305 | |
---|
306 | extern void copy_page(void *dp, void *sp) |
---|
307 | { |
---|
308 | if (on_systemsim()) { |
---|
309 | systemsim_memcpy(dp, sp, PAGE_SIZE); |
---|
310 | } else { |
---|
311 | memcpy(dp, sp, PAGE_SIZE); |
---|
312 | } |
---|
313 | } |
---|
314 | |
---|
315 | /* Allocate (rma_nrpages - nrpages) more memory for domain in proper size. */ |
---|
316 | uint allocate_extents(struct domain *d, uint nrpages, uint rma_nrpages) |
---|
317 | { |
---|
318 | struct page_info *pg; |
---|
319 | ulong mfn; |
---|
320 | ulong gpfn = rma_nrpages; /* starting PFN at end of RMA */ |
---|
321 | uint ext_order; |
---|
322 | uint ext_nrpages; |
---|
323 | uint total_nrpages; |
---|
324 | int i; |
---|
325 | |
---|
326 | ext_order = cpu_extent_order(); |
---|
327 | ext_nrpages = 1 << ext_order; |
---|
328 | |
---|
329 | total_nrpages = rma_nrpages; |
---|
330 | |
---|
331 | /* We only allocate in nr_extsz chunks so if you are not divisible |
---|
332 | * you get more than you asked for. */ |
---|
333 | while (total_nrpages < nrpages) { |
---|
334 | pg = alloc_domheap_pages(d, ext_order, 0); |
---|
335 | if (pg == NULL) |
---|
336 | return total_nrpages; |
---|
337 | |
---|
338 | /* Build p2m mapping for newly allocated extent. */ |
---|
339 | mfn = page_to_mfn(pg); |
---|
340 | for (i = 0; i < (1 << ext_order); i++) |
---|
341 | guest_physmap_add_page(d, gpfn + i, mfn + i); |
---|
342 | |
---|
343 | /* Bump starting PFN by extent size pages. */ |
---|
344 | gpfn += ext_nrpages; |
---|
345 | |
---|
346 | total_nrpages += ext_nrpages; |
---|
347 | } |
---|
348 | |
---|
349 | return total_nrpages; |
---|
350 | } |
---|
351 | |
---|
352 | int allocate_rma(struct domain *d, unsigned int order) |
---|
353 | { |
---|
354 | struct vcpu *v; |
---|
355 | ulong rma_base; |
---|
356 | ulong rma_sz; |
---|
357 | ulong mfn; |
---|
358 | int i; |
---|
359 | |
---|
360 | if (d->arch.rma_page) |
---|
361 | return -EINVAL; |
---|
362 | |
---|
363 | d->arch.rma_page = alloc_domheap_pages(d, order, 0); |
---|
364 | if (d->arch.rma_page == NULL) { |
---|
365 | gdprintk(XENLOG_INFO, "Could not allocate order=%d RMA for domain %u\n", |
---|
366 | order, d->domain_id); |
---|
367 | return -ENOMEM; |
---|
368 | } |
---|
369 | d->arch.rma_order = order; |
---|
370 | |
---|
371 | rma_base = page_to_maddr(d->arch.rma_page); |
---|
372 | rma_sz = rma_size(d->arch.rma_order); |
---|
373 | |
---|
374 | BUG_ON(rma_base & (rma_sz - 1)); /* check alignment */ |
---|
375 | |
---|
376 | printk("allocated RMA for Dom[%d]: 0x%lx[0x%lx]\n", |
---|
377 | d->domain_id, rma_base, rma_sz); |
---|
378 | |
---|
379 | mfn = page_to_mfn(d->arch.rma_page); |
---|
380 | |
---|
381 | for (i = 0; i < (1 << d->arch.rma_order); i++ ) { |
---|
382 | d->arch.rma_page[i].count_info |= PGC_page_RMA; |
---|
383 | clear_page((void *)page_to_maddr(&d->arch.rma_page[i])); |
---|
384 | |
---|
385 | /* Set up p2m mapping for RMA. */ |
---|
386 | guest_physmap_add_page(d, i, mfn+i); |
---|
387 | } |
---|
388 | |
---|
389 | /* shared_info uses last page of RMA */ |
---|
390 | d->shared_info = (shared_info_t *) (rma_base + rma_sz - PAGE_SIZE); |
---|
391 | |
---|
392 | /* if there are already running vcpus, adjust v->vcpu_info */ |
---|
393 | /* XXX untested */ |
---|
394 | for_each_vcpu(d, v) { |
---|
395 | v->vcpu_info = &d->shared_info->vcpu_info[v->vcpu_id]; |
---|
396 | } |
---|
397 | |
---|
398 | return 0; |
---|
399 | } |
---|
400 | |
---|
401 | void free_rma_check(struct page_info *page) |
---|
402 | { |
---|
403 | if (test_bit(_PGC_page_RMA, &page->count_info)) { |
---|
404 | if (!page_get_owner(page)->is_dying) { |
---|
405 | panic("Attempt to free an RMA page: 0x%lx\n", page_to_mfn(page)); |
---|
406 | } else { |
---|
407 | clear_bit(_PGC_page_RMA, &page->count_info); |
---|
408 | } |
---|
409 | } |
---|
410 | } |
---|
411 | |
---|
412 | ulong pfn2mfn(struct domain *d, ulong pfn, int *type) |
---|
413 | { |
---|
414 | ulong mfn = INVALID_MFN; |
---|
415 | int t = PFN_TYPE_NONE; |
---|
416 | ulong foreign_map_pfn = 1UL << cpu_foreign_map_order(); |
---|
417 | |
---|
418 | /* quick tests first */ |
---|
419 | if (pfn & foreign_map_pfn) { |
---|
420 | t = PFN_TYPE_FOREIGN; |
---|
421 | mfn = foreign_to_mfn(d, pfn); |
---|
422 | } else if (pfn >= max_page && pfn < |
---|
423 | (max_page + nr_grant_frames(d->grant_table))) { |
---|
424 | /* XXX access d->grant_table->nr_grant_frames without lock. |
---|
425 | * Currently on powerpc dynamic expanding grant table is |
---|
426 | * inhibited by setting max_nr_grant_frames = INITIAL_NR_GRANT_FRAMES |
---|
427 | * so that this access is safe. |
---|
428 | */ |
---|
429 | /* Its a grant table access */ |
---|
430 | t = PFN_TYPE_GNTTAB; |
---|
431 | mfn = gnttab_shared_mfn(d, d->grant_table, (pfn - max_page)); |
---|
432 | } else if (d->is_privileged && platform_io_mfn(pfn)) { |
---|
433 | t = PFN_TYPE_IO; |
---|
434 | mfn = pfn; |
---|
435 | } else { |
---|
436 | if (pfn < d->arch.p2m_entries) { |
---|
437 | t = PFN_TYPE_LOGICAL; |
---|
438 | mfn = d->arch.p2m[pfn]; |
---|
439 | } |
---|
440 | #ifdef DEBUG |
---|
441 | if (t != PFN_TYPE_NONE && d->is_dying && |
---|
442 | page_get_owner(mfn_to_page(mfn)) != d) { |
---|
443 | printk("%s: page type: %d owner Dom[%d]:%p expected Dom[%d]:%p\n", |
---|
444 | __func__, t, |
---|
445 | page_get_owner(mfn_to_page(mfn))->domain_id, |
---|
446 | page_get_owner(mfn_to_page(mfn)), |
---|
447 | d->domain_id, d); |
---|
448 | BUG(); |
---|
449 | } |
---|
450 | #endif |
---|
451 | } |
---|
452 | |
---|
453 | if (t == PFN_TYPE_NONE) { |
---|
454 | /* This hack allows dom0 to map all memory, necessary to |
---|
455 | * initialize domU state. */ |
---|
456 | if (d->is_privileged && mfn_valid(pfn)) { |
---|
457 | struct page_info *pg; |
---|
458 | |
---|
459 | /* page better be allocated to some domain but not the caller */ |
---|
460 | pg = mfn_to_page(pfn); |
---|
461 | if (!(pg->count_info & PGC_allocated)) |
---|
462 | panic("Foreign page: 0x%lx is not owned by any domain\n", |
---|
463 | mfn); |
---|
464 | if (page_get_owner(pg) == d) |
---|
465 | panic("Foreign page: 0x%lx is owned by this domain\n", |
---|
466 | mfn); |
---|
467 | |
---|
468 | t = PFN_TYPE_FOREIGN; |
---|
469 | mfn = pfn; |
---|
470 | } |
---|
471 | } |
---|
472 | |
---|
473 | if (mfn == INVALID_MFN) { |
---|
474 | printk("%s: Dom[%d] pfn 0x%lx is not a valid page\n", |
---|
475 | __func__, d->domain_id, pfn); |
---|
476 | } |
---|
477 | |
---|
478 | if (type) |
---|
479 | *type = t; |
---|
480 | |
---|
481 | return mfn; |
---|
482 | } |
---|
483 | |
---|
484 | unsigned long mfn_to_gmfn(struct domain *d, unsigned long mfn) |
---|
485 | { |
---|
486 | struct page_info *pg = mfn_to_page(mfn); |
---|
487 | ulong gnttab_mfn; |
---|
488 | |
---|
489 | /* is this our mfn? */ |
---|
490 | if (page_get_owner(pg) != d) |
---|
491 | return INVALID_M2P_ENTRY; |
---|
492 | |
---|
493 | /* XXX access d->grant_table->nr_grant_frames without lock. |
---|
494 | * Currently on powerpc dynamic expanding grant table is |
---|
495 | * inhibited by setting max_nr_grant_frames = INITIAL_NR_GRANT_FRAMES |
---|
496 | * so that this access is safe. |
---|
497 | */ |
---|
498 | /* grant? */ |
---|
499 | gnttab_mfn = gnttab_shared_mfn(d, d->grant_table, 0); |
---|
500 | if (mfn >= gnttab_mfn && mfn < |
---|
501 | (gnttab_mfn + nr_grant_frames(d->grant_table))) |
---|
502 | return max_page + (mfn - gnttab_mfn); |
---|
503 | |
---|
504 | /* IO? */ |
---|
505 | if (d->is_privileged && platform_io_mfn(mfn)) |
---|
506 | return mfn; |
---|
507 | |
---|
508 | /* check m2p table */ |
---|
509 | return get_gpfn_from_mfn(mfn); |
---|
510 | } |
---|
511 | |
---|
512 | /* NB: caller holds d->page_alloc lock, sets d->max_pages = new_max */ |
---|
513 | int guest_physmap_max_mem_pages(struct domain *d, unsigned long new_max_pages) |
---|
514 | { |
---|
515 | u32 *p2m_array = NULL; |
---|
516 | u32 *p2m_old = NULL; |
---|
517 | ulong i; |
---|
518 | |
---|
519 | /* XXX We probably could, but right now we don't shrink the p2m array. |
---|
520 | * NB: d->max_pages >= d->arch.p2m_entries */ |
---|
521 | if (new_max_pages < d->max_pages) { |
---|
522 | printk("Can't shrink DOM%d max memory pages\n", d->domain_id); |
---|
523 | return -EINVAL; |
---|
524 | } |
---|
525 | |
---|
526 | /* Allocate one u32 per page. */ |
---|
527 | p2m_array = xmalloc_array(u32, new_max_pages); |
---|
528 | if (p2m_array == NULL) |
---|
529 | return -ENOMEM; |
---|
530 | |
---|
531 | /* Copy old mappings into new array. */ |
---|
532 | if (d->arch.p2m != NULL) { |
---|
533 | /* XXX This could take a long time; we should use a continuation. */ |
---|
534 | memcpy(p2m_array, d->arch.p2m, d->arch.p2m_entries * sizeof(u32)); |
---|
535 | p2m_old = d->arch.p2m; |
---|
536 | } |
---|
537 | |
---|
538 | /* Mark new mfns as invalid. */ |
---|
539 | for (i = d->arch.p2m_entries; i < new_max_pages; i++) |
---|
540 | p2m_array[i] = INVALID_MFN; |
---|
541 | |
---|
542 | /* Set new p2m pointer and size. */ |
---|
543 | d->arch.p2m = p2m_array; |
---|
544 | d->arch.p2m_entries = new_max_pages; |
---|
545 | |
---|
546 | /* Free old p2m array if present. */ |
---|
547 | if (p2m_old) |
---|
548 | xfree(p2m_old); |
---|
549 | |
---|
550 | return 0; |
---|
551 | } |
---|
552 | |
---|
553 | void guest_physmap_add_page( |
---|
554 | struct domain *d, unsigned long gpfn, unsigned long mfn) |
---|
555 | { |
---|
556 | if (page_get_owner(mfn_to_page(mfn)) != d) { |
---|
557 | printk("Won't map foreign MFN 0x%lx for DOM%d\n", mfn, d->domain_id); |
---|
558 | return; |
---|
559 | } |
---|
560 | |
---|
561 | /* Check that pfn is within guest table. */ |
---|
562 | if (gpfn >= d->arch.p2m_entries) { |
---|
563 | printk("Won't map invalid PFN 0x%lx for DOM%d\n", gpfn, d->domain_id); |
---|
564 | return; |
---|
565 | } |
---|
566 | |
---|
567 | /* Warn if there is an existing mapping. */ |
---|
568 | /* XXX: probably shouldn't let this happen, but |
---|
569 | current interface doesn't throw errors. =( */ |
---|
570 | if (d->arch.p2m[gpfn] != INVALID_MFN) |
---|
571 | printk("Ack! PFN aliased. PFN%lx, old MFN=%x, new MFN=%lx\n", |
---|
572 | gpfn, d->arch.p2m[gpfn], mfn); |
---|
573 | |
---|
574 | /* PFN and MFN ok, map in p2m table. */ |
---|
575 | d->arch.p2m[gpfn] = mfn; |
---|
576 | |
---|
577 | /* Map in m2p table. */ |
---|
578 | set_gpfn_from_mfn(mfn, gpfn); |
---|
579 | } |
---|
580 | |
---|
581 | void guest_physmap_remove_page( |
---|
582 | struct domain *d, unsigned long gpfn, unsigned long mfn) |
---|
583 | { |
---|
584 | if (page_get_owner(mfn_to_page(mfn)) != d) { |
---|
585 | printk("Won't unmap foreign MFN 0x%lx for DOM%d\n", mfn, d->domain_id); |
---|
586 | return; |
---|
587 | } |
---|
588 | |
---|
589 | /* check that pfn is within guest table */ |
---|
590 | if (gpfn >= d->arch.p2m_entries) { |
---|
591 | printk("Won't unmap invalid PFN 0x%lx for DOM%d\n", gpfn, d->domain_id); |
---|
592 | return; |
---|
593 | } |
---|
594 | |
---|
595 | /* PFN and MFN ok, unmap from p2m table. */ |
---|
596 | d->arch.p2m[gpfn] = INVALID_MFN; |
---|
597 | |
---|
598 | /* Unmap from m2p table. */ |
---|
599 | set_gpfn_from_mfn(mfn, INVALID_M2P_ENTRY); |
---|
600 | } |
---|
601 | |
---|
602 | void shadow_drop_references( |
---|
603 | struct domain *d, struct page_info *page) |
---|
604 | { |
---|
605 | } |
---|