| /linux/arch/powerpc/include/asm/ |
| H A D | kasan.h | 34 #define KASAN_SHADOW_END (-(-KASAN_SHADOW_START >> KASAN_SHADOW_SCALE_SHIFT)) macro 41 #define KASAN_SHADOW_END 0xc00fc00000000000UL macro 51 #define KASAN_SHADOW_END 0xc000200000000000UL macro
|
| /linux/arch/arm64/mm/ |
| H A D | kasan_init.c | 202 KASAN_SHADOW_END - (1UL << (64 - KASAN_SHADOW_SCALE_SHIFT))); in kasan_early_init() 205 BUILD_BUG_ON(!IS_ALIGNED(KASAN_SHADOW_END, SHADOW_ALIGN)); in kasan_early_init() 223 kasan_pgd_populate(KASAN_SHADOW_START, KASAN_SHADOW_END, NUMA_NO_NODE, in kasan_early_init() 335 if (!root_level_aligned(KASAN_SHADOW_END)) in kasan_init_shadow() 336 clone_next_level(KASAN_SHADOW_END, tmp_pg_dir, pud[1]); in kasan_init_shadow() 340 clear_shadow(KASAN_SHADOW_START, KASAN_SHADOW_END); in kasan_init_shadow() 350 (void *)KASAN_SHADOW_END); in kasan_init_shadow()
|
| H A D | ptdump.c | 386 { KASAN_SHADOW_END, "Kasan shadow end" }, in ptdump_init()
|
| /linux/arch/powerpc/mm/kasan/ |
| H A D | init_32.c | 108 kasan_update_early_region(KASAN_SHADOW_START, KASAN_SHADOW_END, in kasan_remap_early_shadow_ro() 131 ret = kasan_init_shadow_page_tables(KASAN_SHADOW_START, KASAN_SHADOW_END); in kasan_mmu_init() 156 ret = kasan_init_shadow_page_tables(KASAN_SHADOW_START, KASAN_SHADOW_END); in kasan_init() 180 unsigned long end = KASAN_SHADOW_END; in kasan_early_init()
|
| H A D | init_book3e_64.c | 90 BUILD_BUG_ON(!IS_ALIGNED(KASAN_SHADOW_END, PGDIR_SIZE)); in kasan_early_init() 104 for (addr = KASAN_SHADOW_START; addr != KASAN_SHADOW_END; addr += PGDIR_SIZE) in kasan_early_init() 124 flush_tlb_kernel_range(KASAN_SHADOW_START, KASAN_SHADOW_END); in kasan_init()
|
| /linux/arch/riscv/mm/ |
| H A D | kasan_init.c | 327 KASAN_SHADOW_END - (1UL << (64 - KASAN_SHADOW_SCALE_SHIFT))); in kasan_early_init() 356 KASAN_SHADOW_START, KASAN_SHADOW_END); in kasan_early_init() 364 KASAN_SHADOW_START, KASAN_SHADOW_END); in kasan_swapper_init() 465 ptr = (p4d_t *)pgd_page_vaddr(pgdp_get(pgd_offset_k(KASAN_SHADOW_END))); in create_tmp_mapping() 467 set_pgd(&tmp_pg_dir[pgd_index(KASAN_SHADOW_END)], in create_tmp_mapping() 476 ptr = (pud_t *)p4d_page_vaddr(p4dp_get(base_p4d + p4d_index(KASAN_SHADOW_END))); in create_tmp_mapping() 478 set_p4d(&base_p4d[p4d_index(KASAN_SHADOW_END)], in create_tmp_mapping() 492 KASAN_SHADOW_START, KASAN_SHADOW_END); in kasan_init()
|
| H A D | ptdump.c | 439 address_markers[KASAN_SHADOW_END_NR].start_address = KASAN_SHADOW_END; in ptdump_init()
|
| /linux/arch/arm/include/asm/ |
| H A D | kasan_def.h | 76 #define KASAN_SHADOW_END ((UL(1) << (32 - KASAN_SHADOW_SCALE_SHIFT)) \ macro 78 #define KASAN_SHADOW_START ((KASAN_SHADOW_END >> 3) + KASAN_SHADOW_OFFSET)
|
| /linux/arch/riscv/include/asm/ |
| H A D | kasan.h | 33 #define KASAN_SHADOW_START ((KASAN_SHADOW_END - KASAN_SHADOW_SIZE) & PGDIR_MASK) 34 #define KASAN_SHADOW_END MODULES_LOWEST_VADDR macro
|
| /linux/arch/arm/mm/ |
| H A D | kasan_init.c | 180 BUILD_BUG_ON((KASAN_SHADOW_END - (1UL << 29)) != KASAN_SHADOW_OFFSET); in kasan_early_init() 185 kasan_pgd_populate(KASAN_SHADOW_START, KASAN_SHADOW_END, true); in kasan_early_init() 232 pgd_index(KASAN_SHADOW_END)); in kasan_init() 242 clear_pgds(KASAN_SHADOW_START, KASAN_SHADOW_END); in kasan_init()
|
| H A D | dump.c | 24 { KASAN_SHADOW_END, "Kasan shadow end"},
|
| /linux/arch/loongarch/mm/ |
| H A D | kasan_init.c | 86 if (unlikely(addr > KASAN_SHADOW_END) || in kasan_shadow_to_mem() 248 BUILD_BUG_ON(!IS_ALIGNED(KASAN_SHADOW_END + 1, PGDIR_SIZE)); in kasan_early_init() 278 if (KASAN_SHADOW_END < vm_map_base) { in kasan_init() 295 clear_pgds(KASAN_SHADOW_START, KASAN_SHADOW_END); in kasan_init() 298 kasan_pgd_populate(KASAN_SHADOW_START, KASAN_SHADOW_END, NUMA_NO_NODE, true); in kasan_init()
|
| /linux/arch/x86/mm/ |
| H A D | kasan_init_64.c | 234 unsigned long end = KASAN_SHADOW_END; in kasan_map_early_shadow() 364 ptr = (void *)pgd_page_vaddr(*pgd_offset_k(KASAN_SHADOW_END)); in kasan_init() 366 set_pgd(&early_top_pgt[pgd_index(KASAN_SHADOW_END)], in kasan_init() 373 clear_pgds(KASAN_SHADOW_START & PGDIR_MASK, KASAN_SHADOW_END); in kasan_init() 429 (void *)KASAN_SHADOW_END); in kasan_init()
|
| H A D | dump_pagetables.c | 528 address_markers[KASAN_SHADOW_END_NR].start_address = KASAN_SHADOW_END; in pt_dump_init()
|
| /linux/arch/s390/include/asm/ |
| H A D | kasan.h | 14 #define KASAN_SHADOW_END (KASAN_SHADOW_START + KASAN_SHADOW_SIZE) macro
|
| /linux/arch/um/include/asm/ |
| H A D | kasan.h | 23 #define KASAN_SHADOW_END (KASAN_SHADOW_START + KASAN_SHADOW_SIZE) macro
|
| /linux/arch/xtensa/include/asm/ |
| H A D | kasan.h | 22 #define KASAN_SHADOW_END (KASAN_SHADOW_START + KASAN_SHADOW_SIZE) macro
|
| /linux/arch/x86/include/asm/ |
| H A D | kasan.h | 22 #define KASAN_SHADOW_END (KASAN_SHADOW_START + \ macro
|
| /linux/arch/arm64/include/asm/ |
| H A D | memory.h | 103 #define KASAN_SHADOW_END ((UL(1) << (64 - KASAN_SHADOW_SCALE_SHIFT)) + KASAN_SHADOW_OFFSET) macro 104 #define _KASAN_SHADOW_START(va) (KASAN_SHADOW_END - (UL(1) << ((va) - KASAN_SHADOW_SCALE_SHIFT)))
|
| /linux/arch/loongarch/include/asm/ |
| H A D | kasan.h | 62 #define KASAN_SHADOW_END (round_up(KASAN_SHADOW_START + KASAN_SHADOW_SIZE, PGDIR_SIZE) - 1) macro
|
| /linux/scripts/gdb/linux/ |
| H A D | mm.py | 289 … self.KASAN_SHADOW_END = (1 << (64 - self.KASAN_SHADOW_SCALE_SHIFT)) + self.KASAN_SHADOW_OFFSET 290 …self.PAGE_END = self.KASAN_SHADOW_END - (1 << (self.vabits_actual - self.KASAN_SHADOW_SCALE_SHIFT))
|
| /linux/arch/s390/mm/ |
| H A D | dump_pagetables.c | 351 rc |= add_marker(KASAN_SHADOW_START, KASAN_SHADOW_END, "Kasan Shadow"); in pt_dump_init()
|
| /linux/arch/powerpc/mm/ptdump/ |
| H A D | ptdump.c | 298 address_markers[i++].start_address = KASAN_SHADOW_END; in populate_markers()
|
| /linux/mm/kasan/ |
| H A D | kasan.h | 319 addr < (void *)KASAN_SHADOW_END; in addr_in_shadow()
|
| /linux/arch/s390/boot/ |
| H A D | startup.c | 368 boot_debug("KASAN shadow area: 0x%016lx-0x%016lx\n", KASAN_SHADOW_START, KASAN_SHADOW_END); in setup_kernel_memory_layout()
|