}
early_param("mem", early_mem);
-/*
- * clip_mem_range() - remove memblock memory between @min and @max until
- * we meet the limit in 'memory_limit'.
- */
-static void __init clip_mem_range(u64 min, u64 max)
-{
- u64 mem_size, to_remove;
- int i;
-
-again:
- mem_size = memblock_phys_mem_size();
- if (mem_size <= memory_limit || max <= min)
- return;
-
- to_remove = mem_size - memory_limit;
-
- for (i = memblock.memory.cnt - 1; i >= 0; i--) {
- struct memblock_region *r = memblock.memory.regions + i;
- u64 start = max(min, r->base);
- u64 end = min(max, r->base + r->size);
-
- if (start >= max || end <= min)
- continue;
-
- if (end > min) {
- u64 size = min(to_remove, end - max(start, min));
-
- memblock_remove(end - size, size);
- } else {
- memblock_remove(start, min(max - start, to_remove));
- }
- goto again;
- }
-}
-
void __init arm64_memblock_init(void)
{
const s64 linear_region_size = -(s64)PAGE_OFFSET;
if (memblock_end_of_DRAM() > linear_region_size)
memblock_remove(0, memblock_end_of_DRAM() - linear_region_size);
+ /*
+ * Apply the memory limit if it was set. Since the kernel may be loaded
+ * high up in memory, add back the kernel region that must be accessible
+ * via the linear mapping.
+ */
if (memory_limit != (phys_addr_t)ULLONG_MAX) {
- u64 kbase = round_down(__pa(_text), MIN_KIMG_ALIGN);
- u64 kend = PAGE_ALIGN(__pa(_end));
- u64 const sz_4g = 0x100000000UL;
-
- /*
- * Clip memory in order of preference:
- * - above the kernel and above 4 GB
- * - between 4 GB and the start of the kernel (if the kernel
- * is loaded high in memory)
- * - between the kernel and 4 GB (if the kernel is loaded
- * low in memory)
- * - below 4 GB
- */
- clip_mem_range(max(sz_4g, kend), ULLONG_MAX);
- clip_mem_range(sz_4g, kbase);
- clip_mem_range(kend, sz_4g);
- clip_mem_range(0, min(kbase, sz_4g));
+ memblock_enforce_memory_limit(memory_limit);
+ memblock_add(__pa(_text), (u64)(_end - _text));
}
/*