@@ -1072,10 +1072,9 @@ int map_pages_to_xen(unsigned long virt,
return create_xen_entries(INSERT, virt, _mfn(mfn), nr_mfns, flags);
}
-int populate_pt_range(unsigned long virt, unsigned long mfn,
- unsigned long nr_mfns)
+int populate_pt_range(unsigned long virt, unsigned long nr_mfns)
{
- return create_xen_entries(RESERVE, virt, _mfn(mfn), nr_mfns, 0);
+ return create_xen_entries(RESERVE, virt, INVALID_MFN, nr_mfns, 0);
}
int destroy_xen_mappings(unsigned long v, unsigned long e)
@@ -5007,10 +5007,9 @@ int map_pages_to_xen(
return 0;
}
-int populate_pt_range(unsigned long virt, unsigned long mfn,
- unsigned long nr_mfns)
+int populate_pt_range(unsigned long virt, unsigned long nr_mfns)
{
- return map_pages_to_xen(virt, mfn, nr_mfns, MAP_SMALL_PAGES);
+ return map_pages_to_xen(virt, mfn_x(INVALID_MFN), nr_mfns, MAP_SMALL_PAGES);
}
/*
@@ -42,7 +42,7 @@ void __init vm_init_type(enum vmap_region type, void *start, void *end)
bitmap_fill(vm_bitmap(type), vm_low[type]);
/* Populate page tables for the bitmap if necessary. */
- populate_pt_range(va, 0, vm_low[type] - nr);
+ populate_pt_range(va, vm_low[type] - nr);
}
static void *vm_alloc(unsigned int nr, unsigned int align,
@@ -175,8 +175,7 @@ int destroy_xen_mappings(unsigned long v, unsigned long e);
* Create only non-leaf page table entries for the
* page range in Xen virtual address space.
*/
-int populate_pt_range(unsigned long virt, unsigned long mfn,
- unsigned long nr_mfns);
+int populate_pt_range(unsigned long virt, unsigned long nr_mfns);
/* Claim handling */
unsigned long domain_adjust_tot_pages(struct domain *d, long pages);
int domain_set_outstanding_pages(struct domain *d, unsigned long pages);