| /src/sys/arm/include/ |
| H A D | cpu.h | 336 tlb_flush_local(vm_offset_t va) in tlb_flush_local() argument 339 KASSERT((va & PAGE_MASK) == 0, ("%s: va %#x not aligned", __func__, va)); in tlb_flush_local() 342 _CP15_TLBIMVA(va | CPU_ASID_KERNEL); in tlb_flush_local() 348 tlb_flush_range_local(vm_offset_t va, vm_size_t size) in tlb_flush_range_local() argument 350 vm_offset_t eva = va + size; in tlb_flush_range_local() 352 KASSERT((va & PAGE_MASK) == 0, ("%s: va %#x not aligned", __func__, va)); in tlb_flush_range_local() 357 for (; va < eva; va += PAGE_SIZE) in tlb_flush_range_local() 358 _CP15_TLBIMVA(va | CPU_ASID_KERNEL); in tlb_flush_range_local() 390 tlb_flush(vm_offset_t va) in tlb_flush() argument 393 KASSERT((va & PAGE_MASK) == 0, ("%s: va %#x not aligned", __func__, va)); in tlb_flush() [all …]
|
| H A D | pmap_var.h | 83 pte1_index(vm_offset_t va) in pte1_index() argument 86 return (va >> PTE1_SHIFT); in pte1_index() 90 pte1_ptr(pt1_entry_t *pt1, vm_offset_t va) in pte1_ptr() argument 93 return (pt1 + pte1_index(va)); in pte1_ptr() 97 pte1_trunc(vm_offset_t va) in pte1_trunc() argument 100 return (va & PTE1_FRAME); in pte1_trunc() 104 pte1_roundup(vm_offset_t va) in pte1_roundup() argument 107 return ((va + PTE1_OFFSET) & PTE1_FRAME); in pte1_roundup() 393 pt2tab_index(vm_offset_t va) in pt2tab_index() argument 396 return (va >> PT2TAB_SHIFT); in pt2tab_index() [all …]
|
| /src/sys/powerpc/booke/ |
| H A D | pmap_64.c | 134 static pte_t *ptbl_alloc(pmap_t pmap, vm_offset_t va, 181 pte_find(pmap_t pmap, vm_offset_t va) in pte_find() argument 189 pdir_l1 = pmap->pm_root[PG_ROOT_IDX(va)]; in pte_find() 192 pdir = pdir_l1[PDIR_L1_IDX(va)]; in pte_find() 195 ptbl = pdir[PDIR_IDX(va)]; in pte_find() 197 return ((ptbl != NULL) ? &ptbl[PTBL_IDX(va)] : NULL); in pte_find() 204 vm_offset_t va; in pte_find_next() local 211 va = *pva; in pte_find_next() 212 i = PG_ROOT_IDX(va); in pte_find_next() 213 j = PDIR_L1_IDX(va); in pte_find_next() [all …]
|
| H A D | pmap.c | 567 pv_insert(pmap_t pmap, vm_offset_t va, vm_page_t m) in pv_insert() argument 580 pve->pv_va = va; in pv_insert() 593 pv_remove(pmap_t pmap, vm_offset_t va, vm_page_t m) in pv_remove() argument 605 if ((pmap == pve->pv_pmap) && (va == pve->pv_va)) { in pv_remove() 1012 mmu_booke_extract(pmap_t pmap, vm_offset_t va) in mmu_booke_extract() argument 1017 pa = pte_vatopa(pmap, va); in mmu_booke_extract() 1028 mmu_booke_kextract(vm_offset_t va) in mmu_booke_kextract() argument 1035 if (va >= DMAP_BASE_ADDRESS && va <= DMAP_MAX_ADDRESS) in mmu_booke_kextract() 1036 return (DMAP_TO_PHYS(va)); in mmu_booke_kextract() 1039 if (va >= VM_MIN_KERNEL_ADDRESS && va <= VM_MAX_KERNEL_ADDRESS) in mmu_booke_kextract() [all …]
|
| H A D | pmap_32.c | 300 vm_offset_t va; in ptbl_free() local 329 va = ((vm_offset_t)ptbl + (i * PAGE_SIZE)); in ptbl_free() 330 pa = pte_vatopa(kernel_pmap, va); in ptbl_free() 334 mmu_booke_kremove(va); in ptbl_free() 429 pte_remove(pmap_t pmap, vm_offset_t va, uint8_t flags) in pte_remove() argument 431 unsigned int pdir_idx = PDIR_IDX(va); in pte_remove() 432 unsigned int ptbl_idx = PTBL_IDX(va); in pte_remove() 463 pv_remove(pmap, va, m); in pte_remove() 470 pv_remove(pmap, va, m); in pte_remove() 478 tlb0_flush_entry(va); in pte_remove() [all …]
|
| /src/sys/arm64/iommu/ |
| H A D | iommu_pmap.c | 81 #define smmu_l0_index(va) (((va) >> IOMMU_L0_SHIFT) & IOMMU_L0_ADDR_MASK) argument 82 #define smmu_l1_index(va) (((va) >> IOMMU_L1_SHIFT) & IOMMU_Ln_ADDR_MASK) argument 83 #define smmu_l2_index(va) (((va) >> IOMMU_L2_SHIFT) & IOMMU_Ln_ADDR_MASK) argument 84 #define smmu_l3_index(va) (((va) >> IOMMU_L3_SHIFT) & IOMMU_Ln_ADDR_MASK) argument 87 static void _smmu_pmap_unwire_l3(struct smmu_pmap *pmap, vm_offset_t va, 104 smmu_pmap_l0(struct smmu_pmap *pmap, vm_offset_t va) in smmu_pmap_l0() argument 107 return (&pmap->sp_l0[smmu_l0_index(va)]); in smmu_pmap_l0() 111 smmu_pmap_l0_to_l1(pd_entry_t *l0, vm_offset_t va) in smmu_pmap_l0_to_l1() argument 116 return (&l1[smmu_l1_index(va)]); in smmu_pmap_l0_to_l1() 120 smmu_pmap_l1(struct smmu_pmap *pmap, vm_offset_t va) in smmu_pmap_l1() argument [all …]
|
| /src/sys/arm/arm/ |
| H A D | pmap-v6.c | 329 vm_offset_t va); 330 static int pmap_enter_pte1(pmap_t pmap, vm_offset_t va, pt1_entry_t pte1, 332 void cache_icache_sync_fresh(vm_offset_t va, vm_paddr_t pa, vm_size_t size); 651 pt2map_pt2pg(vm_offset_t va) in pt2map_pt2pg() argument 654 va &= ~(NPT2_IN_PG * PTE1_SIZE - 1); in pt2map_pt2pg() 655 return ((vm_offset_t)pt2map_entry(va)); in pt2map_pt2pg() 722 #define KERNEL_V2P(va) \ argument 723 ((vm_paddr_t)((va) - KERNVIRTADDR + arm_physmem_kernaddr)) 913 pmap_preboot_pt2pg_setup(vm_offset_t va) in pmap_preboot_pt2pg_setup() argument 919 pte2p = kern_pt2tab_entry(va); in pmap_preboot_pt2pg_setup() [all …]
|
| /src/sys/arm64/vmm/ |
| H A D | vmm_mmu.c | 158 vmmpmap_l1_table(vm_offset_t va) in vmmpmap_l1_table() argument 166 l0e = atomic_load_64(&l0[pmap_l0_index(va)]); in vmmpmap_l1_table() 178 rv = atomic_cmpset_64(&l0[pmap_l0_index(va)], l0e, new_l0e); in vmmpmap_l1_table() 197 vmmpmap_l2_table(vm_offset_t va) in vmmpmap_l2_table() argument 203 l1 = vmmpmap_l1_table(va); in vmmpmap_l2_table() 209 l1e = atomic_load_64(&l1[pmap_l1_index(va)]); in vmmpmap_l2_table() 221 rv = atomic_cmpset_64(&l1[pmap_l1_index(va)], l1e, new_l1e); in vmmpmap_l2_table() 240 vmmpmap_l3_table(vm_offset_t va) in vmmpmap_l3_table() argument 246 l2 = vmmpmap_l2_table(va); in vmmpmap_l3_table() 252 l2e = atomic_load_64(&l2[pmap_l2_index(va)]); in vmmpmap_l3_table() [all …]
|
| /src/sys/arm64/arm64/ |
| H A D | pmap.c | 318 vm_offset_t va; member 471 static void pmap_pvh_free(struct md_page *pvh, pmap_t pmap, vm_offset_t va); 473 vm_offset_t va); 475 static void pmap_abort_ptp(pmap_t pmap, vm_offset_t va, vm_page_t mpte); 478 static int pmap_change_props_locked(vm_offset_t va, vm_size_t size, 480 static bool pmap_copy_l3c(pmap_t pmap, pt_entry_t *l3p, vm_offset_t va, 482 static pt_entry_t *pmap_demote_l1(pmap_t pmap, pt_entry_t *l1, vm_offset_t va); 484 vm_offset_t va, struct rwlock **lockp); 485 static pt_entry_t *pmap_demote_l2(pmap_t pmap, pt_entry_t *l2, vm_offset_t va); 486 static bool pmap_demote_l2c(pmap_t pmap, pt_entry_t *l2p, vm_offset_t va); [all …]
|
| /src/sys/riscv/riscv/ |
| H A D | pmap.c | 317 static void pmap_pvh_free(struct md_page *pvh, pmap_t pmap, vm_offset_t va); 319 vm_offset_t va); 320 static bool pmap_demote_l1(pmap_t pmap, pd_entry_t *l1, vm_offset_t va); 321 static bool pmap_demote_l2(pmap_t pmap, pd_entry_t *l2, vm_offset_t va); 323 vm_offset_t va, struct rwlock **lockp); 324 static int pmap_enter_l2(pmap_t pmap, vm_offset_t va, pd_entry_t new_l2, 326 static vm_page_t pmap_enter_quick_locked(pmap_t pmap, vm_offset_t va, 330 static bool pmap_try_insert_pv_entry(pmap_t pmap, vm_offset_t va, 336 static void _pmap_unwire_ptp(pmap_t pmap, vm_offset_t va, vm_page_t m, 340 static int pmap_change_attr_locked(vm_offset_t va, vm_size_t size, int mode); [all …]
|
| H A D | cbo.c | 36 cbo_zicbom_cpu_dcache_wbinv_range(vm_offset_t va, vm_size_t len) in cbo_zicbom_cpu_dcache_wbinv_range() argument 45 va &= ~(dcache_line_size - 1); in cbo_zicbom_cpu_dcache_wbinv_range() 46 for (addr = va; addr < va + len; addr += dcache_line_size) in cbo_zicbom_cpu_dcache_wbinv_range() 52 cbo_zicbom_cpu_dcache_inv_range(vm_offset_t va, vm_size_t len) in cbo_zicbom_cpu_dcache_inv_range() argument 63 va &= ~(dcache_line_size - 1); in cbo_zicbom_cpu_dcache_inv_range() 64 for (addr = va; addr < va + len; addr += dcache_line_size) in cbo_zicbom_cpu_dcache_inv_range() 70 cbo_zicbom_cpu_dcache_wb_range(vm_offset_t va, vm_size_t len) in cbo_zicbom_cpu_dcache_wb_range() argument 83 va &= ~(dcache_line_size - 1); in cbo_zicbom_cpu_dcache_wb_range() 84 for (addr = va; addr < va + len; addr += dcache_line_size) in cbo_zicbom_cpu_dcache_wb_range()
|
| /src/sys/i386/i386/ |
| H A D | pmap.c | 187 #define vtopte(va) (PTmap + i386_btop(va)) argument 237 vm_offset_t va; member 289 static void pmap_pv_demote_pde(pmap_t pmap, vm_offset_t va, vm_paddr_t pa); 290 static bool pmap_pv_insert_pde(pmap_t pmap, vm_offset_t va, pd_entry_t pde, 293 static void pmap_pv_promote_pde(pmap_t pmap, vm_offset_t va, vm_paddr_t pa); 295 static void pmap_pvh_free(struct md_page *pvh, pmap_t pmap, vm_offset_t va); 297 vm_offset_t va); 300 static void pmap_abort_ptp(pmap_t pmap, vm_offset_t va, vm_page_t mpte); 301 static bool pmap_demote_pde(pmap_t pmap, pd_entry_t *pde, vm_offset_t va); 302 static int pmap_enter_4mpage(pmap_t pmap, vm_offset_t va, vm_page_t m, [all …]
|
| /src/lib/libkvm/ |
| H A D | kvm_minidump_amd64.c | 51 #define VA_OFF(vm, va) \ argument 52 (VM_IS_V1(vm) ? ((va) & (AMD64_PAGE_SIZE - 1)) : ((va) & AMD64_PAGE_MASK)) 184 _amd64_minidump_vatop_v1(kvm_t *kd, kvaddr_t va, off_t *pa) in _amd64_minidump_vatop_v1() argument 194 offset = va & AMD64_PAGE_MASK; in _amd64_minidump_vatop_v1() 196 if (va >= vm->hdr.kernbase) { in _amd64_minidump_vatop_v1() 197 pteindex = (va - vm->hdr.kernbase) >> AMD64_PAGE_SHIFT; in _amd64_minidump_vatop_v1() 216 } else if (va >= vm->hdr.dmapbase && va < vm->hdr.dmapend) { in _amd64_minidump_vatop_v1() 217 a = (va - vm->hdr.dmapbase) & ~AMD64_PAGE_MASK; in _amd64_minidump_vatop_v1() 222 (uintmax_t)va); in _amd64_minidump_vatop_v1() 230 (uintmax_t)va); in _amd64_minidump_vatop_v1() [all …]
|
| H A D | kvm_minidump_i386.c | 146 _i386_minidump_vatop_pae(kvm_t *kd, kvaddr_t va, off_t *pa) in _i386_minidump_vatop_pae() argument 156 offset = va & I386_PAGE_MASK; in _i386_minidump_vatop_pae() 158 if (va >= vm->hdr.kernbase) { in _i386_minidump_vatop_pae() 159 pteindex = (va - vm->hdr.kernbase) >> I386_PAGE_SHIFT; in _i386_minidump_vatop_pae() 181 (uintmax_t)va); in _i386_minidump_vatop_pae() 186 _kvm_err(kd, 0, "invalid address (0x%jx)", (uintmax_t)va); in _i386_minidump_vatop_pae() 191 _i386_minidump_vatop(kvm_t *kd, kvaddr_t va, off_t *pa) in _i386_minidump_vatop() argument 201 offset = va & I386_PAGE_MASK; in _i386_minidump_vatop() 203 if (va >= vm->hdr.kernbase) { in _i386_minidump_vatop() 204 pteindex = (va - vm->hdr.kernbase) >> I386_PAGE_SHIFT; in _i386_minidump_vatop() [all …]
|
| /src/sys/amd64/amd64/ |
| H A D | pmap.c | 533 vm_offset_t va; member 596 #define PMAP_ADDRESS_IN_LARGEMAP(va) ((va) >= kva_layout.lm_low && \ argument 597 (va) < kva_layout.lm_high) 639 static pt_entry_t pmap_pkru_get(pmap_t pmap, vm_offset_t va); 1330 static void pmap_pv_demote_pde(pmap_t pmap, vm_offset_t va, vm_paddr_t pa, 1332 static bool pmap_pv_insert_pde(pmap_t pmap, vm_offset_t va, pd_entry_t pde, 1335 static void pmap_pv_promote_pde(pmap_t pmap, vm_offset_t va, vm_paddr_t pa, 1338 static void pmap_pvh_free(struct md_page *pvh, pmap_t pmap, vm_offset_t va); 1340 vm_offset_t va); 1342 static void pmap_abort_ptp(pmap_t pmap, vm_offset_t va, vm_page_t mpte); [all …]
|
| /src/sys/kern/ |
| H A D | subr_devmap.c | 223 vm_offset_t va; in devmap_vtop() local 228 va = (vm_offset_t)vpva; in devmap_vtop() 230 if (va >= pd->pd_va && va + size <= pd->pd_va + pd->pd_size) in devmap_vtop() 231 return ((vm_paddr_t)(pd->pd_pa + (va - pd->pd_va))); in devmap_vtop() 258 vm_offset_t va, offset; in pmap_mapdev_attr() local 281 va = akva_devmap_vaddr; in pmap_mapdev_attr() 282 KASSERT(va >= (VM_MAX_KERNEL_ADDRESS - PMAP_MAPDEV_EARLY_SIZE), in pmap_mapdev_attr() 288 va = kva_alloc_aligned(size, L2_SIZE); in pmap_mapdev_attr() 290 va = kva_alloc_aligned(size, L3C_SIZE); in pmap_mapdev_attr() 293 va = kva_alloc(size); in pmap_mapdev_attr() [all …]
|
| /src/sys/powerpc/aim/ |
| H A D | mmu_radix.c | 101 static void pmap_pte_walk(pml1_entry_t *l1, vm_offset_t va); 191 vm_offset_t va, uint16_t ap) in radix_tlbie() argument 195 MPASS((va & PAGE_MASK) == 0); in radix_tlbie() 198 rb = va | is | ap; in radix_tlbie() 204 radix_tlbie_fixup(uint32_t pid, vm_offset_t va, int ap) in radix_tlbie_fixup() argument 209 TLBIEL_INVAL_PAGE, 0, 0, va, ap); in radix_tlbie_fixup() 212 TLBIEL_INVAL_PAGE, pid, 0, va, ap); in radix_tlbie_fixup() 216 radix_tlbie_invlpg_user_4k(uint32_t pid, vm_offset_t va) in radix_tlbie_invlpg_user_4k() argument 220 TLBIEL_INVAL_PAGE, pid, 0, va, TLBIE_ACTUAL_PAGE_4K); in radix_tlbie_invlpg_user_4k() 221 radix_tlbie_fixup(pid, va, TLBIE_ACTUAL_PAGE_4K); in radix_tlbie_invlpg_user_4k() [all …]
|
| H A D | mmu_oea.c | 324 void moea_dumpsys_map(vm_paddr_t pa, size_t sz, void **va); 457 tlbie(vm_offset_t va) in tlbie() argument 462 __asm __volatile("tlbie %0" :: "r"(va)); in tlbie() 470 vm_offset_t va; in tlbia() local 472 for (va = 0; va < 0x00040000; va += 0x00001000) { in tlbia() 473 __asm __volatile("tlbie %0" :: "r"(va)); in tlbia() 481 va_to_sr(u_int *sr, vm_offset_t va) in va_to_sr() argument 483 return (sr[(uintptr_t)va >> ADDR_SR_SHFT]); in va_to_sr() 536 moea_pte_match(struct pte *pt, u_int sr, vm_offset_t va, int which) in moea_pte_match() argument 540 ((va >> ADDR_API_SHFT) & PTE_API) | which); in moea_pte_match() [all …]
|
| /src/sys/fs/smbfs/ |
| H A D | smbfs_node.c | 301 struct vattr va; in smbfs_inactive() local 309 VOP_GETATTR(vp, &va, cred); in smbfs_inactive() 351 smbfs_attr_cachelookup(struct vnode *vp, struct vattr *va) in smbfs_attr_cachelookup() argument 360 va->va_type = vp->v_type; /* vnode type (for create) */ in smbfs_attr_cachelookup() 361 va->va_flags = 0; /* flags defined for file */ in smbfs_attr_cachelookup() 363 va->va_mode = smp->sm_file_mode; /* files access mode and type */ in smbfs_attr_cachelookup() 365 va->va_mode &= ~(S_IWUSR|S_IWGRP|S_IWOTH); in smbfs_attr_cachelookup() 366 va->va_flags |= UF_READONLY; in smbfs_attr_cachelookup() 369 va->va_mode = smp->sm_dir_mode; /* files access mode and type */ in smbfs_attr_cachelookup() 372 va->va_size = np->n_size; in smbfs_attr_cachelookup() [all …]
|
| /src/contrib/tcsh/ |
| H A D | tc.printf.c | 371 va_list va; in xsnprintf() local 372 va_start(va, fmt); in xsnprintf() 376 count = doprnt(xaddchar, fmt, va); in xsnprintf() 377 va_end(va); in xsnprintf() 387 va_list va; in xprintf() local 388 va_start(va, fmt); in xprintf() 389 count = doprnt(xputchar, fmt, va); in xprintf() 390 va_end(va); in xprintf() 395 xvprintf(const char *fmt, va_list va) in xvprintf() argument 397 return doprnt(xputchar, fmt, va); in xvprintf() [all …]
|
| /src/sys/dev/xdma/ |
| H A D | xdma_iommu.c | 62 xdma_iommu_remove_entry(xdma_channel_t *xchan, vm_offset_t va) in xdma_iommu_remove_entry() argument 68 va &= ~(PAGE_SIZE - 1); in xdma_iommu_remove_entry() 69 pmap_remove(&xio->p, va, va + PAGE_SIZE); in xdma_iommu_remove_entry() 71 XDMA_IOMMU_REMOVE(xio->dev, xio, va); in xdma_iommu_remove_entry() 73 vmem_free(xio->vmem, va, PAGE_SIZE); in xdma_iommu_remove_entry() 77 xdma_iommu_enter(struct xdma_iommu *xio, vm_offset_t va, in xdma_iommu_enter() argument 90 pmap_enter(p, va, m, prot, prot | PMAP_ENTER_WIRED, 0); in xdma_iommu_enter() 92 XDMA_IOMMU_ENTER(xio->dev, xio, va, pa); in xdma_iommu_enter() 94 va += PAGE_SIZE; in xdma_iommu_enter() 100 xdma_iommu_add_entry(xdma_channel_t *xchan, vm_offset_t *va, in xdma_iommu_add_entry() argument [all …]
|
| /src/contrib/llvm-project/lld/MachO/Arch/ |
| H A D | ARM64Common.h | 26 void relocateOne(uint8_t *loc, const Reloc &, uint64_t va, 46 uint64_t va) { in encodeBranch26() argument 47 checkInt(loc, r, va, 28); in encodeBranch26() 50 llvm::support::endian::write32le(loc, base | bitField(va, 2, 26, 0)); in encodeBranch26() 54 uint64_t va) { in encodeBranch26() argument 55 checkInt(loc, d, va, 28); in encodeBranch26() 56 llvm::support::endian::write32le(loc, base | bitField(va, 2, 26, 0)); in encodeBranch26() 65 uint64_t va) { in encodePage21() argument 66 checkInt(loc, r, va, 35); in encodePage21() 67 llvm::support::endian::write32le(loc, base | bitField(va, 12, 2, 29) | in encodePage21() [all …]
|
| /src/contrib/netbsd-tests/fs/puffs/h_dtfs/ |
| H A D | dtfs_vnops.c | 110 const struct vattr *va, const struct puffs_cred *pcr) in dtfs_node_setattr() argument 116 if (va->va_flags != PUFFS_VNOVAL) in dtfs_node_setattr() 119 if (va->va_uid != PUFFS_VNOVAL || va->va_gid != PUFFS_VNOVAL) { in dtfs_node_setattr() 121 va->va_uid, va->va_gid, pcr); in dtfs_node_setattr() 126 if (va->va_mode != PUFFS_VNOVAL) { in dtfs_node_setattr() 128 pn->pn_va.va_type, va->va_mode, pcr); in dtfs_node_setattr() 133 if ((va->va_atime.tv_sec != PUFFS_VNOVAL in dtfs_node_setattr() 134 && va->va_atime.tv_nsec != PUFFS_VNOVAL) in dtfs_node_setattr() 135 || (va->va_mtime.tv_sec != PUFFS_VNOVAL in dtfs_node_setattr() 136 && va->va_mtime.tv_nsec != PUFFS_VNOVAL)) { in dtfs_node_setattr() [all …]
|
| /src/sys/riscv/thead/ |
| H A D | thead.c | 56 thead_cpu_dcache_wbinv_range(vm_offset_t va, vm_size_t len) in thead_cpu_dcache_wbinv_range() argument 58 register vm_offset_t t0 __asm("t0") = rounddown(va, dcache_line_size); in thead_cpu_dcache_wbinv_range() 60 for (; t0 < va + len; t0 += dcache_line_size) { in thead_cpu_dcache_wbinv_range() 68 thead_cpu_dcache_inv_range(vm_offset_t va, vm_size_t len) in thead_cpu_dcache_inv_range() argument 70 register vm_offset_t t0 __asm("t0") = rounddown(va, dcache_line_size); in thead_cpu_dcache_inv_range() 72 for (; t0 < va + len; t0 += dcache_line_size) { in thead_cpu_dcache_inv_range() 80 thead_cpu_dcache_wb_range(vm_offset_t va, vm_size_t len) in thead_cpu_dcache_wb_range() argument 82 register vm_offset_t t0 __asm("t0") = rounddown(va, dcache_line_size); in thead_cpu_dcache_wb_range() 84 for (; t0 < va + len; t0 += dcache_line_size) { in thead_cpu_dcache_wb_range()
|
| /src/sys/dev/smbios/ |
| H A D | smbios.c | 194 void *va; in smbios_probe() local 201 va = pmap_mapbios(pa, size); in smbios_probe() 202 if (va == NULL) { in smbios_probe() 207 if (smbios_cksum(va)) { in smbios_probe() 212 pmap_unmapbios(va, size); in smbios_probe() 220 void *va; in smbios_attach() local 228 va = pmap_mapbios(pa, size); in smbios_attach() 229 if (va == NULL) { in smbios_attach() 233 sc->is_eps3 = smbios_eps3(va); in smbios_attach() 236 sc->eps3 = va; in smbios_attach() [all …]
|