#include <linux/sched.h>
#include <linux/mm.h>
#include <linux/pm.h>
+#include <linux/memblock.h>
#include <asm/elf.h>
#include <asm/vdso.h>
#include <asm/xen/hypervisor.h>
#include <asm/xen/hypercall.h>
+#include <xen/xen.h>
#include <xen/page.h>
#include <xen/interface/callback.h>
#include <xen/interface/memory.h>
static __init void xen_add_extra_mem(unsigned long pages)
{
u64 size = (u64)pages * PAGE_SIZE;
+ u64 extra_start = xen_extra_mem_start + xen_extra_mem_size;
if (!pages)
return;
- e820_add_region(xen_extra_mem_start + xen_extra_mem_size, size, E820_RAM);
+ e820_add_region(extra_start, size, E820_RAM);
sanitize_e820_map(e820.map, ARRAY_SIZE(e820.map), &e820.nr_map);
- reserve_early(xen_extra_mem_start + xen_extra_mem_size,
- xen_extra_mem_start + xen_extra_mem_size + size,
- "XEN EXTRA");
+ memblock_x86_reserve_range(extra_start, extra_start + size, "XEN EXTRA");
xen_extra_mem_size += size;
+
+ xen_max_p2m_pfn = PFN_DOWN(extra_start + size);
}
static unsigned long __init xen_release_chunk(phys_addr_t start_addr,
unsigned long extra_pages = 0;
unsigned long extra_limit;
int i;
+ int op;
max_pfn = min(MAX_DOMAIN_PAGES, max_pfn);
mem_end = PFN_PHYS(max_pfn);
memmap.nr_entries = E820MAX;
set_xen_guest_handle(memmap.buffer, map);
- rc = HYPERVISOR_memory_op(XENMEM_memory_map, &memmap);
+ op = xen_initial_domain() ?
+ XENMEM_machine_memory_map :
+ XENMEM_memory_map;
+ rc = HYPERVISOR_memory_op(op, &memmap);
if (rc == -ENOSYS) {
memmap.nr_entries = 1;
map[0].addr = 0ULL;
unsigned long long end = map[i].addr + map[i].size;
if (map[i].type == E820_RAM) {
- if (end > mem_end) {
+ if (map[i].addr < mem_end && end > mem_end) {
/* Truncate region to max_mem. */
- map[i].size -= end - mem_end;
+ u64 delta = end - mem_end;
+
+ map[i].size -= delta;
+ extra_pages += PFN_DOWN(delta);
- extra_pages += PFN_DOWN(end - mem_end);
+ end = mem_end;
}
- } else if (map[i].type != E820_RAM)
+ }
+
+ if (end > xen_extra_mem_start)
xen_extra_mem_start = end;
+ /* If region is non-RAM or below mem_end, add what remains */
if ((map[i].type != E820_RAM || map[i].addr < mem_end) &&
map[i].size > 0)
e820_add_region(map[i].addr, map[i].size, map[i].type);
* Even though this is normal, usable memory under Xen, reserve
* ISA memory anyway because too many things think they can poke
* about in there.
+ *
+ * In a dom0 kernel, this region is identity mapped with the
+ * hardware ISA area, so it really is out of bounds.
*/
e820_add_region(ISA_START_ADDRESS, ISA_END_ADDRESS - ISA_START_ADDRESS,
E820_RESERVED);
* - xen_start_info
* See comment above "struct start_info" in <xen/interface/xen.h>
*/
- reserve_early(__pa(xen_start_info->mfn_list),
+ memblock_x86_reserve_range(__pa(xen_start_info->mfn_list),
__pa(xen_start_info->pt_base),
"XEN START INFO");
else
extra_pages = 0;
- xen_add_extra_mem(extra_pages);
+ if (!xen_initial_domain())
+ xen_add_extra_mem(extra_pages);
return "Xen";
}
pm_idle = xen_idle;
- paravirt_disable_iospace();
-
fiddle_vdso();
}