Lines Matching +full:max +full:- +full:virtual +full:- +full:functions
1 // SPDX-License-Identifier: GPL-2.0-only
5 * Copyright (C) 1995-2005 Russell King
24 #include <linux/dma-direct.h>
25 #include <linux/dma-map-ops.h>
40 #include <asm/kernel-pgtable.h>
50 #include <asm/xen/swiotlb-xen.h>
58 s64 memstart_addr __ro_after_init = -1;
63 * and ZONE_DMA32. By default ZONE_DMA covers the 32-bit addressable memory
64 * unless restricted on specific platforms (e.g. 30-bit on Raspberry Pi 4).
65 * In such case, ZONE_DMA32 covers the rest of the 32-bit addressable memory,
89 * in the virtual address space.
118 return min(zone_limit, memblock_end_of_DRAM() - 1) + 1; in max_zone_phys()
134 * Information we get from firmware (e.g. DT dma-ranges) describe DMA in zone_sizes_init()
136 * Some of them rely on DMA zone in low 32-bit memory. Keep low RAM in zone_sizes_init()
187 s64 linear_region_size = PAGE_END - _PAGE_OFFSET(vabits_actual); in arm64_memblock_init()
190 * Corner case: 52-bit VA capable systems running KVM in nVHE mode may in arm64_memblock_init()
212 if ((memblock_end_of_DRAM() - memstart_addr) > linear_region_size) in arm64_memblock_init()
224 memstart_addr = round_up(memblock_end_of_DRAM() - linear_region_size, in arm64_memblock_init()
230 * If we are running with a 52-bit kernel VA config on a system that in arm64_memblock_init()
232 * memory in the 48-bit addressable part of the linear region, i.e., in arm64_memblock_init()
237 memstart_addr -= _PAGE_OFFSET(vabits_actual) - _PAGE_OFFSET(52); in arm64_memblock_init()
246 memblock_add(__pa_symbol(_text), (u64)(_end - _text)); in arm64_memblock_init()
253 * Otherwise, this is a no-op in arm64_memblock_init()
256 u64 size = PAGE_ALIGN(phys_initrd_start + phys_initrd_size) - base; in arm64_memblock_init()
269 "initrd not fully accessible via the linear mapping -- please check your bootloader ...\n")) { in arm64_memblock_init()
283 s64 range = linear_region_size - in arm64_memblock_init()
293 memstart_addr -= ARM64_MEMSTART_ALIGN * in arm64_memblock_init()
302 memblock_reserve(__pa_symbol(_stext), _end - _stext); in arm64_memblock_init()
304 /* the generic initrd code expects virtual addresses */ in arm64_memblock_init()
314 unsigned long min, max; in bootmem_init() local
317 max = PFN_DOWN(memblock_end_of_DRAM()); in bootmem_init()
319 early_memtest(min << PAGE_SHIFT, max << PAGE_SHIFT); in bootmem_init()
321 max_pfn = max_low_pfn = max; in bootmem_init()
392 * scratch using the virtual address range and page size. in arch_mm_preinit()
416 memblock_free(lm_init_begin, lm_init_end - lm_init_begin); in free_initmem()
442 * Choose a random page-aligned base address for a window of 'size' bytes which
443 * entirely contains the interval [start, end - 1].
449 if ((end - start) >= size) in random_bounding_box()
452 max_pgoff = (size - (end - start)) / PAGE_SIZE; in random_bounding_box()
455 return start - pgoff * PAGE_SIZE; in random_bounding_box()
460 * image and other modules. References using PREL32 relocations have a +/-2G
464 * Modules may directly branch to functions and code within the kernel text,
465 * and to functions and code within other modules. These branches will use
466 * CALL26/JUMP26 relocations with a +/-128M range. Without PLTs, we must ensure
482 u64 kernel_size = kernel_end - kernel_start; in module_init_limits()
493 module_direct_base = kernel_end - SZ_128M; in module_init_limits()
495 module_plt_base = kernel_end - SZ_2G; in module_init_limits()
498 u64 max = kernel_end; in module_init_limits() local
503 module_direct_base = random_bounding_box(SZ_128M, min, max); in module_init_limits()
506 max = module_direct_base + SZ_128M; in module_init_limits()
510 module_plt_base = random_bounding_box(SZ_2G, min, max); in module_init_limits()
513 pr_info("%llu pages in range for non-PLT usage", in module_init_limits()
514 module_direct_base ? (SZ_128M - kernel_size) / PAGE_SIZE : 0); in module_init_limits()
516 module_plt_base ? (SZ_2G - kernel_size) / PAGE_SIZE : 0); in module_init_limits()