/linux/arch/x86/include/asm/ |
H A D | pgtable_64_types.h | 113 # define VMALLOC_START vmalloc_base macro 124 * For KMSAN builds, VMALLOC_START..VMEMORY_END is 4 times bigger than 125 * VMALLOC_START..VMALLOC_END (see below). 127 #define VMEMORY_END (VMALLOC_START + (VMALLOC_SIZE_TB << 40) - 1) 137 * 1st quarter: VMALLOC_START to VMALLOC_END - new vmalloc area 148 #define VMALLOC_END (VMALLOC_START + VMALLOC_QUARTER_SIZE - 1) 157 #define KMSAN_VMALLOC_SHADOW_START (VMALLOC_START + KMSAN_VMALLOC_SHADOW_OFFSET) 158 #define KMSAN_VMALLOC_ORIGIN_START (VMALLOC_START + KMSAN_VMALLOC_ORIGIN_OFFSET)
|
H A D | pgtable_32_areas.h | 20 #define VMALLOC_START ((unsigned long)high_memory + VMALLOC_OFFSET) macro 47 #define MODULES_VADDR VMALLOC_START
|
/linux/arch/m68k/include/asm/ |
H A D | pgtable_mm.h | 91 #define VMALLOC_START 0x0f800000 macro 94 #define VMALLOC_START 0xd0000000 macro 98 #define VMALLOC_START (((unsigned long) high_memory + VMALLOC_OFFSET) & ~(VMALLOC_OFFSET-1)) macro 109 #define VMALLOC_START (((unsigned long) high_memory + VMALLOC_OFFSET) & ~(VMALLOC_OFFSET-1)) macro
|
H A D | pgtable_no.h | 44 #define VMALLOC_START 0 macro
|
/linux/arch/arc/include/asm/ |
H A D | processor.h | 87 #define VMALLOC_START (PAGE_OFFSET - (CONFIG_ARC_KVADDR_SIZE << 20)) macro 92 #define VMALLOC_END (VMALLOC_START + VMALLOC_SIZE) 94 #define USER_KERNEL_GUTTER (VMALLOC_START - TASK_SIZE)
|
/linux/arch/powerpc/include/asm/nohash/32/ |
H A D | pgtable.h | 83 #define IOREMAP_START VMALLOC_START 98 * from VMALLOC_START). For this reason we have ioremap_bot to check when 105 #define VMALLOC_START (((ALIGN((long)high_memory, PPC_PIN_SIZE) + VMALLOC_OFFSET) & ~(VMALLOC_OFFSET-1))) macro 107 #define VMALLOC_START ((((long)high_memory + VMALLOC_OFFSET) & ~(VMALLOC_OFFSET-1))) macro
|
/linux/arch/xtensa/mm/ |
H A D | kasan_init.c | 75 BUILD_BUG_ON(VMALLOC_START < KASAN_START_VADDR); in kasan_init() 81 populate(kasan_mem_to_shadow((void *)VMALLOC_START), in kasan_init()
|
H A D | init.c | 93 VMALLOC_START, VMALLOC_END, in print_vm_layout() 94 (VMALLOC_END - VMALLOC_START) >> 20, in print_vm_layout()
|
/linux/arch/xtensa/include/asm/ |
H A D | pgtable.h | 68 #define VMALLOC_START (XCHAL_KSEG_CACHED_VADDR - 0x10000000) macro 69 #define VMALLOC_END (VMALLOC_START + 0x07FEFFFF) 70 #define TLBTEMP_BASE_1 (VMALLOC_START + 0x08000000) 80 #define VMALLOC_START __XTENSA_UL_CONST(0) macro
|
/linux/fs/proc/ |
H A D | kcore.c | 231 if (VMALLOC_START > ent->addr) { in kclist_add_private() 232 if (VMALLOC_START - ent->addr < ent->size) in kclist_add_private() 233 ent->size = VMALLOC_START - ent->addr; in kclist_add_private() 712 if (MODULES_VADDR != VMALLOC_START && MODULES_END != VMALLOC_END) { in add_modules_range() 733 kclist_add(&kcore_vmalloc, (void *)VMALLOC_START, in proc_kcore_init() 734 VMALLOC_END - VMALLOC_START, KCORE_VMALLOC); in proc_kcore_init()
|
/linux/arch/powerpc/mm/ |
H A D | mem.c | 417 start = VMALLOC_START; in execmem_arch_setup() 434 .start = VMALLOC_START, in execmem_arch_setup() 440 .start = VMALLOC_START, in execmem_arch_setup()
|
/linux/include/linux/ |
H A D | ioremap.h | 15 #define IOREMAP_START VMALLOC_START
|
/linux/arch/csky/include/asm/ |
H A D | memory.h | 13 #define VMALLOC_START (PAGE_OFFSET + LOWMEM_LIMIT + (PAGE_SIZE * 8)) macro
|
/linux/Documentation/arch/xtensa/ |
H A D | mmu.rst | 83 | VMALLOC area | VMALLOC_START 0xc0000000 128MB - 64KB 126 | VMALLOC area | VMALLOC_START 0xa0000000 128MB - 64KB 170 | VMALLOC area | VMALLOC_START 0x90000000 128MB - 64KB
|
/linux/arch/arm/mm/ |
H A D | iomap.c | 40 if ((unsigned long)addr >= VMALLOC_START && in pci_iounmap()
|
/linux/arch/parisc/include/asm/ |
H A D | fixmap.h | 54 #define VMALLOC_START ((unsigned long)parisc_vmalloc_start) macro
|
/linux/lib/ |
H A D | test_debug_virtual.c | 27 va = (void *)VMALLOC_START; in test_debug_virtual_init()
|
/linux/arch/s390/boot/ |
H A D | startup.c | 31 unsigned long __bootdata_preserved(VMALLOC_START); 416 VMALLOC_START = VMALLOC_END - vmalloc_size; in setup_kernel_memory_layout() 417 boot_debug("vmalloc area: 0x%016lx-0x%016lx\n", VMALLOC_START, VMALLOC_END); in setup_kernel_memory_layout() 419 __memcpy_real_area = round_down(VMALLOC_START - MEMCPY_REAL_SIZE, PAGE_SIZE); in setup_kernel_memory_layout()
|
/linux/arch/arm/include/asm/ |
H A D | pgtable-nommu.h | 62 #define VMALLOC_START 0UL macro
|
/linux/arch/hexagon/include/asm/ |
H A D | mem-layout.h | 62 #define VMALLOC_START ((unsigned long) __va(high_memory + VMALLOC_OFFSET)) macro 68 * Create the space between VMALLOC_START and FIXADDR_TOP backwards
|
/linux/arch/mips/include/asm/ |
H A D | pgtable-64.h | 39 * fault address - VMALLOC_START. 141 #define VMALLOC_START (MAP_BASE + (2 * PAGE_SIZE)) macro 148 VMALLOC_START != CKSSEG
|
/linux/arch/x86/mm/ |
H A D | kasan_init_64.c | 392 kasan_mem_to_shadow((void *)VMALLOC_START)); in kasan_init() 401 kasan_mem_to_shadow((void *)VMALLOC_START), in kasan_init() 405 kasan_mem_to_shadow((void *)VMALLOC_START), in kasan_init()
|
/linux/arch/sh/mm/ |
H A D | init.c | 355 (unsigned long)VMALLOC_START, VMALLOC_END, in mem_init() 356 (VMALLOC_END - VMALLOC_START) >> 20, in mem_init()
|
/linux/scripts/gdb/linux/ |
H A D | vmalloc.py | 20 return addr >= pg_ops.VMALLOC_START and addr < pg_ops.VMALLOC_END
|
/linux/arch/riscv/include/asm/ |
H A D | pgtable.h | 45 #define VMALLOC_START (PAGE_OFFSET - VMALLOC_SIZE) macro 63 #define MODULES_VADDR VMALLOC_START 87 #define VMEMMAP_END VMALLOC_START 88 #define VMEMMAP_START (VMALLOC_START - VMEMMAP_SIZE) 1075 #define VMALLOC_START _AC(0, UL) macro
|