| /linux/arch/arm64/include/asm/ |
| H A D | memory.h | 43 #define VA_BITS (CONFIG_ARM64_VA_BITS) macro 45 #define PAGE_OFFSET (_PAGE_OFFSET(VA_BITS)) 56 #if VA_BITS > 48 63 #define VA_BITS_MIN (VA_BITS) 235 #if VA_BITS > 48 239 #define vabits_actual ((u64)VA_BITS)
|
| H A D | pgtable-hwdef.h | 85 #define PTRS_PER_PGD (1 << (VA_BITS - PGDIR_SHIFT))
|
| H A D | processor.h | 57 #define TASK_SIZE_MAX (UL(1) << VA_BITS)
|
| H A D | pgtable.h | 24 #if VA_BITS == VA_BITS_MIN 948 return vabits_actual == VA_BITS; in pgtable_l4_enabled() 1076 return vabits_actual == VA_BITS; in pgtable_l5_enabled()
|
| /linux/arch/sparc/include/asm/ |
| H A D | processor_64.h | 23 #define VA_BITS 44 macro 25 #define VPTE_SIZE (1UL << (VA_BITS - PAGE_SHIFT + 3)) 27 #define VPTE_SIZE (1 << (VA_BITS - PAGE_SHIFT + 3))
|
| /linux/arch/powerpc/include/asm/book3s/64/ |
| H A D | mmu-hash.h | 577 #define VA_BITS 68 macro 579 #define ESID_BITS (VA_BITS - (SID_SHIFT + CONTEXT_BITS)) 580 #define ESID_BITS_1T (VA_BITS - (SID_SHIFT_1T + CONTEXT_BITS)) 650 #define VSID_BITS_256M (VA_BITS - SID_SHIFT) 658 #define VSID_BITS_1T (VA_BITS - SID_SHIFT_1T) 780 unsigned long va_bits = VA_BITS; in get_vsid()
|
| /linux/arch/loongarch/include/asm/ |
| H A D | processor.h | 33 #define TASK_SIZE64 (0x1UL << ((cpu_vabits > VA_BITS) ? VA_BITS : cpu_vabits))
|
| H A D | kasan.h | 19 #define XRANGE_SHADOW_SHIFT min(cpu_vabits, VA_BITS)
|
| H A D | pgtable.h | 47 #define VA_BITS 32 macro 49 #define VA_BITS (PGDIR_SHIFT + (PAGE_SHIFT - PTRLOG)) macro
|
| /linux/arch/riscv/include/asm/ |
| H A D | kasan.h | 28 #define KASAN_SHADOW_SIZE (UL(1) << ((VA_BITS - 1) - KASAN_SHADOW_SCALE_SHIFT))
|
| H A D | pgtable.h | 78 #define VA_BITS (pgtable_l5_enabled ? \ macro 81 #define VA_BITS VA_BITS_SV32 macro 85 (VA_BITS - PAGE_SHIFT - 1 + STRUCT_PAGE_MAX_SHIFT) 127 #define MMAP_VA_BITS_64 ((VA_BITS >= VA_BITS_SV48) ? VA_BITS_SV48 : VA_BITS) 1226 #define KERN_VIRT_START (-(BIT(VA_BITS)) + TASK_SIZE)
|
| /linux/arch/riscv/kernel/ |
| H A D | vmcore_info.c | 13 VMCOREINFO_NUMBER(VA_BITS); in arch_crash_save_vmcoreinfo()
|
| /linux/scripts/gdb/linux/ |
| H A D | mm.py | 219 self.VA_BITS = constants.LX_CONFIG_ARM64_VA_BITS 220 if self.VA_BITS > 48: 229 self.VA_BITS_MIN = self.VA_BITS 230 self.vabits_actual = self.VA_BITS 264 self.PAGE_OFFSET = self._PAGE_OFFSET(self.VA_BITS)
|
| /linux/arch/arm64/kernel/ |
| H A D | vmcore_info.c | 22 VMCOREINFO_NUMBER(VA_BITS); in arch_crash_save_vmcoreinfo()
|
| /linux/arch/arm64/mm/ |
| H A D | kasan_init.c | 203 BUILD_BUG_ON(!IS_ALIGNED(_KASAN_SHADOW_START(VA_BITS), SHADOW_ALIGN)); in kasan_early_init() 246 u64 vabits = IS_ENABLED(CONFIG_ARM64_64K_PAGES) ? VA_BITS in root_level_idx()
|
| H A D | proc.S | 509 mov x9, #64 - VA_BITS
|
| /linux/Documentation/arch/arm64/ |
| H A D | memory.rst | 62 Most code in the kernel should not need to consider the VA_BITS, for 66 VA_BITS constant the *maximum* VA space size
|
| /linux/arch/arm64/kernel/pi/ |
| H A D | map_kernel.c | 247 int va_bits = VA_BITS; in early_map_kernel()
|
| /linux/arch/riscv/mm/ |
| H A D | fault.c | 41 current->comm, PAGE_SIZE / SZ_1K, VA_BITS, in show_pte()
|
| /linux/arch/csky/ |
| H A D | Kconfig | 184 # VA_BITS - PAGE_SHIFT - 3
|
| /linux/Documentation/admin-guide/kdump/ |
| H A D | vmcoreinfo.rst | 419 VA_BITS section in ARM64 566 VA_BITS section in RISCV64
|
| /linux/arch/powerpc/platforms/pseries/ |
| H A D | lpar.c | 1934 unsigned long va_bits = VA_BITS; in vsid_unscramble()
|
| /linux/arch/arm64/kvm/ |
| H A D | mmu.c | 632 if ((base ^ io_map_base) & BIT(VA_BITS - 1)) in __hyp_alloc_private_va_range()
|
| /linux/arch/riscv/ |
| H A D | Kconfig | 270 # VA_BITS - PAGE_SHIFT - 3
|
| /linux/drivers/iommu/arm/arm-smmu-v3/ |
| H A D | arm-smmu-v3.c | 2613 pgtbl_cfg.ias = min_t(unsigned long, ias, VA_BITS); in arm_smmu_domain_finalise()
|