Searched refs:L2_SIZE (Results 1 – 11 of 11) sorted by relevance
| /src/sys/arm64/include/ |
| H A D | pte.h | 196 #define L2_SIZE (UINT64_C(1) << L2_SHIFT) macro 197 #define L2_OFFSET (L2_SIZE - 1) 214 #define PMAP_MAPDEV_EARLY_SIZE (L2_SIZE * 20) 247 #define L2C_SIZE (L2C_ENTRIES * L2_SIZE)
|
| /src/sys/arm64/arm64/ |
| H A D | minidump_machdep.c | 176 for (va = VM_MIN_KERNEL_ADDRESS; va < kva_end; va += L2_SIZE) { in cpu_minidumpsys() 191 va += L1_SIZE - L2_SIZE; in cpu_minidumpsys() 288 for (va = VM_MIN_KERNEL_ADDRESS; va < kva_end; va += L2_SIZE) { in cpu_minidumpsys() 312 tmpbuffer[j] = (pa + i * L2_SIZE + in cpu_minidumpsys() 326 va += L1_SIZE - L2_SIZE; in cpu_minidumpsys()
|
| H A D | pmap.c | 307 #define PMAP_PREINIT_MAPPING_SIZE (PMAP_PREINIT_MAPPING_COUNT * L2_SIZE) 1132 if ((physmap[i + 1] - state->pa) < L2_SIZE) in pmap_bootstrap_l2_block() 1141 (physmap[i + 1] - state->pa) >= L2_SIZE; in pmap_bootstrap_l2_block() 1142 state->va += L2_SIZE, state->pa += L2_SIZE) { in pmap_bootstrap_l2_block() 1289 (physmap[i + 1] - bs_state.pa) >= L2_SIZE) { in pmap_bootstrap_dmap() 1334 for (; bs_state.va < VM_MAX_KERNEL_ADDRESS; bs_state.va += L2_SIZE) in pmap_bootstrap_l3() 1362 preinit_map_va = roundup2(bs_state.freemempos, L2_SIZE); in pmap_bootstrap() 1437 pa = rounddown2(end_pa - L2_SIZE, L2_SIZE); in pmap_bootstrap_allocate_san_l2() 1438 for (; pa >= start_pa && va < eva; va += L2_SIZE, pa -= L2_SIZE) { in pmap_bootstrap_allocate_san_l2() 1446 pa += L2_SIZE; in pmap_bootstrap_allocate_san_l2() [all …]
|
| H A D | locore.S | 654 ldr x7, =(3 * L2_SIZE - 1) 793 ldr x9, =(L2_SIZE) 813 ldr x9, =(L2_SIZE)
|
| /src/sys/riscv/include/ |
| H A D | pte.h | 57 #define L2_SIZE (1UL << L2_SHIFT) macro 58 #define L2_OFFSET (L2_SIZE - 1)
|
| H A D | vmparam.h | 250 #define PMAP_MAPDEV_EARLY_SIZE (4 * L2_SIZE)
|
| /src/sys/riscv/riscv/ |
| H A D | pmap.c | 633 if (roundup(pa, L2_SIZE) + L2_SIZE > endpa) in pmap_bootstrap_dmap() 685 pa += L2_SIZE; in pmap_bootstrap_dmap() 686 va += L2_SIZE; in pmap_bootstrap_dmap() 701 while (pa + L2_SIZE - 1 < endpa) { in pmap_bootstrap_dmap() 716 pa += L2_SIZE; in pmap_bootstrap_dmap() 717 va += L2_SIZE; in pmap_bootstrap_dmap() 814 nkernl2 = howmany(howmany(kernlen, L2_SIZE), Ln_ENTRIES); in pmap_create_pagetables() 845 if (freemempos < roundup2(kernend, L2_SIZE)) in pmap_create_pagetables() 846 freemempos = roundup2(kernend, L2_SIZE); in pmap_create_pagetables() 862 for (pa = kernstart; pa < kernend; pa += L2_SIZE, slot++) { in pmap_create_pagetables() [all …]
|
| H A D | minidump_machdep.c | 183 for (va = VM_MIN_KERNEL_ADDRESS; va < kva_max; va += L2_SIZE) { in cpu_minidumpsys() 283 for (va = VM_MIN_KERNEL_ADDRESS; va < kva_max; va += L2_SIZE) { in cpu_minidumpsys()
|
| H A D | machdep.c | 527 physmem_exclude_region(mr->mr_start, L2_SIZE, in fdt_physmem_hardware_region_cb() 552 min(p->md_pages * EFI_PAGE_SIZE, L2_SIZE), in efi_exclude_sbi_pmp_cb()
|
| /src/sys/arm64/vmm/ |
| H A D | vmm_arm64.c | 310 next_hyp_va = roundup2(vmm_base + hyp_code_len, L2_SIZE); in vmmops_modinit() 323 next_hyp_va += L2_SIZE; in vmmops_modinit() 413 if (vmm_base > (L2_SIZE + PAGE_SIZE)) { in vmmops_modinit() 419 vmm_base -= L2_SIZE + PAGE_SIZE; in vmmops_modinit() 420 vmm_base = rounddown2(vmm_base, L2_SIZE); in vmmops_modinit() 428 if (vmm_base > L2_SIZE) in vmmops_modinit() 429 vmem_add(el2_mem_alloc, L2_SIZE, vmm_base - L2_SIZE, in vmmops_modinit()
|
| /src/sys/kern/ |
| H A D | subr_devmap.c | 287 if (size >= L2_SIZE && (pa & L2_OFFSET) == 0) in pmap_mapdev_attr() 288 va = kva_alloc_aligned(size, L2_SIZE); in pmap_mapdev_attr()
|