| /linux/arch/x86/kernel/ |
| H A D | head64.c | 103 p4d_p = (p4dval_t *)((pgd & PTE_PFN_MASK) + __START_KERNEL_map - phys_base); in __early_make_pgtable() 112 *pgd_p = (pgdval_t)p4d_p - __START_KERNEL_map + phys_base + _KERNPG_TABLE; in __early_make_pgtable() 118 pud_p = (pudval_t *)((p4d & PTE_PFN_MASK) + __START_KERNEL_map - phys_base); in __early_make_pgtable() 127 *p4d_p = (p4dval_t)pud_p - __START_KERNEL_map + phys_base + _KERNPG_TABLE; in __early_make_pgtable() 133 pmd_p = (pmdval_t *)((pud & PTE_PFN_MASK) + __START_KERNEL_map - phys_base); in __early_make_pgtable() 142 *pud_p = (pudval_t)pmd_p - __START_KERNEL_map + phys_base + _KERNPG_TABLE; in __early_make_pgtable() 228 BUILD_BUG_ON(MODULES_VADDR < __START_KERNEL_map); in x86_64_start_kernel() 229 BUILD_BUG_ON(MODULES_VADDR - __START_KERNEL_map < KERNEL_IMAGE_SIZE); in x86_64_start_kernel() 231 BUILD_BUG_ON((__START_KERNEL_map & ~PMD_MASK) != 0); in x86_64_start_kernel()
|
| H A D | head_64.S | 183 addq $(init_top_pgt - __START_KERNEL_map), %rax 605 .quad level3_kernel_pgt - __START_KERNEL_map + _PAGE_TABLE_NOENC 626 .quad level3_kernel_pgt - __START_KERNEL_map + _PAGE_TABLE_NOENC 633 .quad level2_kernel_pgt - __START_KERNEL_map + _KERNPG_TABLE_NOENC 634 .quad level2_fixmap_pgt - __START_KERNEL_map + _PAGE_TABLE_NOENC 660 .quad level1_fixmap_pgt + (pgtno << PAGE_SHIFT) - __START_KERNEL_map \
|
| H A D | vmlinux.lds.S | 18 #define LOAD_OFFSET __START_KERNEL_map
|
| H A D | setup.c | 838 __START_KERNEL_map, in dump_kernel_offset()
|
| /linux/arch/x86/mm/ |
| H A D | physaddr.c | 17 unsigned long y = x - __START_KERNEL_map; in __phys_addr() 25 x = y + (__START_KERNEL_map - PAGE_OFFSET); in __phys_addr() 38 unsigned long y = x - __START_KERNEL_map; in __virt_addr_valid() 47 x = y + (__START_KERNEL_map - PAGE_OFFSET); in __virt_addr_valid()
|
| H A D | kaslr.c | 98 BUILD_BUG_ON(vaddr_end > __START_KERNEL_map); in kernel_randomize_memory()
|
| H A D | kasan_init_64.c | 199 p4d += __START_KERNEL_map - phys_base; in early_p4d_offset() 422 kasan_mem_to_shadow((void *)__START_KERNEL_map)); in kasan_init()
|
| H A D | init_64.c | 446 unsigned long vaddr = __START_KERNEL_map; in cleanup_highmap() 447 unsigned long vaddr_end = __START_KERNEL_map + KERNEL_IMAGE_SIZE; in cleanup_highmap() 457 vaddr_end = __START_KERNEL_map + (max_pfn_mapped << PAGE_SHIFT); in cleanup_highmap()
|
| H A D | dump_pagetables.c | 104 [HIGH_KERNEL_NR] = { __START_KERNEL_map, "High Kernel Mapping" },
|
| /linux/arch/x86/platform/pvh/ |
| H A D | head.S | 10 #define _pa(x) ((x) - __START_KERNEL_map) 190 addq $__START_KERNEL_map, %rax 259 .quad pvh_level3_ident_pgt - __START_KERNEL_map + _KERNPG_TABLE_NOENC 261 .quad pvh_level3_ident_pgt - __START_KERNEL_map + _KERNPG_TABLE_NOENC 264 .quad pvh_level3_kernel_pgt - __START_KERNEL_map + _PAGE_TABLE_NOENC 268 .quad pvh_level2_ident_pgt - __START_KERNEL_map + _KERNPG_TABLE_NOENC 285 .quad pvh_level2_kernel_pgt - __START_KERNEL_map + _KERNPG_TABLE_NOENC
|
| /linux/arch/x86/include/asm/ |
| H A D | page_64.h | 25 unsigned long y = x - __START_KERNEL_map; in __phys_addr_nodebug() 28 x = y + ((x > y) ? phys_base : (__START_KERNEL_map - PAGE_OFFSET)); in __phys_addr_nodebug() 41 unsigned long y = x - __START_KERNEL_map; in __phys_addr_symbol()
|
| H A D | kmsan.h | 66 unsigned long y = x - __START_KERNEL_map; in kmsan_virt_addr_valid() 76 x = y + (__START_KERNEL_map - PAGE_OFFSET); in kmsan_virt_addr_valid()
|
| H A D | pgtable_64.h | 275 L4_START_KERNEL = l4_index(__START_KERNEL_map) 277 L3_START_KERNEL = pud_index(__START_KERNEL_map)
|
| H A D | page_32_types.h | 20 #define __START_KERNEL_map __PAGE_OFFSET macro
|
| H A D | page_types.h | 34 #define __START_KERNEL (__START_KERNEL_map + LOAD_PHYSICAL_ADDR)
|
| H A D | page_64_types.h | 46 #define __START_KERNEL_map _AC(0xffffffff80000000, UL) macro
|
| H A D | pgtable_64_types.h | 171 #define MODULES_VADDR (__START_KERNEL_map + KERNEL_IMAGE_SIZE)
|
| /linux/arch/x86/boot/startup/ |
| H A D | map_kernel.c | 112 phys_base = load_delta = __START_KERNEL_map + p2v_offset; in __startup_64() 127 pgd[pgd_index(__START_KERNEL_map)] += load_delta; in __startup_64() 133 pgd[pgd_index(__START_KERNEL_map)] = (pgdval_t)p4d | _PAGE_TABLE; in __startup_64()
|
| /linux/arch/x86/platform/efi/ |
| H A D | efi_thunk_64.S | 55 movq $__START_KERNEL_map, %rax
|
| /linux/arch/x86/xen/ |
| H A D | mmu_pv.c | 1221 if (addr >= __START_KERNEL_map) { in xen_pagetable_p2m_free() 1525 if (xen_start_info->mfn_list < __START_KERNEL_map && in xen_make_pte_init() 1679 return (void *)(paddr + __START_KERNEL_map); in __ka() 1772 if (xen_start_info->mfn_list < __START_KERNEL_map) in xen_setup_kernel_pagetable() 1785 init_top_pgt[pgd_index(__START_KERNEL_map)].pgd = in xen_setup_kernel_pagetable() 1804 l3 = m2v(pgd[pgd_index(__START_KERNEL_map)].pgd); in xen_setup_kernel_pagetable() 1805 l2 = m2v(l3[pud_index(__START_KERNEL_map)].pud); in xen_setup_kernel_pagetable() 1834 if (i && i < pgd_index(__START_KERNEL_map)) in xen_setup_kernel_pagetable() 2043 if (xen_start_info->mfn_list < __START_KERNEL_map) { in xen_relocate_p2m()
|
| H A D | xen-head.S | 150 ELFNOTE(Xen, XEN_ELFNOTE_VIRT_BASE, _ASM_PTR __START_KERNEL_map)
|
| H A D | setup.c | 769 if (xen_start_info->mfn_list >= __START_KERNEL_map) { in xen_reserve_xen_mfnlist()
|
| /linux/arch/x86/kernel/acpi/ |
| H A D | madt_wakeup.c | 111 info.offset = __START_KERNEL_map - phys_base; in acpi_mp_setup_reset()
|
| /linux/arch/x86/boot/compressed/ |
| H A D | misc.c | 219 map = delta - __START_KERNEL_map; in handle_relocations()
|
| /linux/arch/x86/mm/pat/ |
| H A D | set_memory.c | 1957 __START_KERNEL_map - phys_base; in cpa_process_alias()
|