Searched refs:L3_SIZE (Results 1 – 7 of 7) sorted by relevance
| /src/sys/arm64/include/ |
| H A D | pte.h | 203 #define L3_SIZE (1 << L3_SHIFT) macro 204 #define L3_OFFSET (L3_SIZE - 1) 250 #define L3C_SIZE (L3C_ENTRIES * L3_SIZE)
|
| /src/sys/riscv/include/ |
| H A D | pte.h | 62 #define L3_SIZE (1UL << L3_SHIFT) macro 63 #define L3_OFFSET (L3_SIZE - 1)
|
| /src/lib/libkvm/ |
| H A D | kvm_riscv.h | 44 #define RISCV_L3_OFFSET (L3_SIZE - 1) 69 _Static_assert(L3_SIZE == RISCV_L3_SIZE, "L3_SIZE mismatch");
|
| /src/sys/arm64/arm64/ |
| H A D | pmap.c | 1182 if (physmap[i + 1] - state->pa < L3_SIZE) in pmap_bootstrap_l3_page() 1191 physmap[i + 1] - state->pa >= L3_SIZE; in pmap_bootstrap_l3_page() 1192 state->va += L3_SIZE, state->pa += L3_SIZE) { in pmap_bootstrap_l3_page() 2000 pmap_s1_invalidate_strided(pmap, sva, eva, L3_SIZE, final_only); in pmap_s1_invalidate_range() 2236 sva = (sva & ~L3_OFFSET) + L3_SIZE; in pmap_kmapped_range() 2574 va += L3_SIZE; in pmap_qenter() 3894 for (mt = m, tva = va; mt < &m[L3C_ENTRIES]; mt++, tva += L3_SIZE) { in pmap_pv_insert_l3c() 3900 tva -= L3_SIZE; in pmap_pv_insert_l3c() 4113 L3_SIZE) { in pmap_remove_l3c() 4160 for (l3 = pmap_l2_to_l3(&l2e, sva); sva != eva; l3++, sva += L3_SIZE) { in pmap_remove_l3_range() [all …]
|
| /src/sys/riscv/riscv/ |
| H A D | pmap.c | 626 pa = rounddown(physmap[idx], L3_SIZE); in pmap_bootstrap_dmap() 661 pa += L3_SIZE; in pmap_bootstrap_dmap() 662 va += L3_SIZE; in pmap_bootstrap_dmap() 745 pa += L3_SIZE; in pmap_bootstrap_dmap() 746 va += L3_SIZE; in pmap_bootstrap_dmap() 825 ndevmapl3 = howmany(howmany(PMAP_MAPDEV_EARLY_SIZE, L3_SIZE), in pmap_create_pagetables() 1410 va += L3_SIZE; in pmap_qenter() 2688 sva += L3_SIZE) { in pmap_remove() 2699 sva += L3_SIZE; in pmap_remove() 2885 sva += L3_SIZE) { in pmap_protect() [all …]
|
| /src/sys/arm64/vmm/ |
| H A D | vmm_mmu.c | 397 va_next = (va + L3_SIZE) & ~L3_OFFSET; in vmmpmap_remove()
|
| H A D | vmm_hyp.c | 718 #define TLBI_VA_L3_INCR (L3_SIZE >> TLBI_VA_SHIFT) in s2_tlbi_range()
|