Home
last modified time | relevance | path

Searched refs:va (Results 1 – 25 of 434) sorted by relevance

12345678910>>...18

/src/sys/arm/include/
H A Dcpu.h336 tlb_flush_local(vm_offset_t va) in tlb_flush_local() argument
339 KASSERT((va & PAGE_MASK) == 0, ("%s: va %#x not aligned", __func__, va)); in tlb_flush_local()
342 _CP15_TLBIMVA(va | CPU_ASID_KERNEL); in tlb_flush_local()
348 tlb_flush_range_local(vm_offset_t va, vm_size_t size) in tlb_flush_range_local() argument
350 vm_offset_t eva = va + size; in tlb_flush_range_local()
352 KASSERT((va & PAGE_MASK) == 0, ("%s: va %#x not aligned", __func__, va)); in tlb_flush_range_local()
357 for (; va < eva; va += PAGE_SIZE) in tlb_flush_range_local()
358 _CP15_TLBIMVA(va | CPU_ASID_KERNEL); in tlb_flush_range_local()
390 tlb_flush(vm_offset_t va) in tlb_flush() argument
393 KASSERT((va & PAGE_MASK) == 0, ("%s: va %#x not aligned", __func__, va)); in tlb_flush()
[all …]
H A Dpmap_var.h83 pte1_index(vm_offset_t va) in pte1_index() argument
86 return (va >> PTE1_SHIFT); in pte1_index()
90 pte1_ptr(pt1_entry_t *pt1, vm_offset_t va) in pte1_ptr() argument
93 return (pt1 + pte1_index(va)); in pte1_ptr()
97 pte1_trunc(vm_offset_t va) in pte1_trunc() argument
100 return (va & PTE1_FRAME); in pte1_trunc()
104 pte1_roundup(vm_offset_t va) in pte1_roundup() argument
107 return ((va + PTE1_OFFSET) & PTE1_FRAME); in pte1_roundup()
393 pt2tab_index(vm_offset_t va) in pt2tab_index() argument
396 return (va >> PT2TAB_SHIFT); in pt2tab_index()
[all …]
/src/sys/powerpc/booke/
H A Dpmap_64.c134 static pte_t *ptbl_alloc(pmap_t pmap, vm_offset_t va,
181 pte_find(pmap_t pmap, vm_offset_t va) in pte_find() argument
189 pdir_l1 = pmap->pm_root[PG_ROOT_IDX(va)]; in pte_find()
192 pdir = pdir_l1[PDIR_L1_IDX(va)]; in pte_find()
195 ptbl = pdir[PDIR_IDX(va)]; in pte_find()
197 return ((ptbl != NULL) ? &ptbl[PTBL_IDX(va)] : NULL); in pte_find()
204 vm_offset_t va; in pte_find_next() local
211 va = *pva; in pte_find_next()
212 i = PG_ROOT_IDX(va); in pte_find_next()
213 j = PDIR_L1_IDX(va); in pte_find_next()
[all …]
H A Dpmap.c567 pv_insert(pmap_t pmap, vm_offset_t va, vm_page_t m) in pv_insert() argument
580 pve->pv_va = va; in pv_insert()
593 pv_remove(pmap_t pmap, vm_offset_t va, vm_page_t m) in pv_remove() argument
605 if ((pmap == pve->pv_pmap) && (va == pve->pv_va)) { in pv_remove()
1012 mmu_booke_extract(pmap_t pmap, vm_offset_t va) in mmu_booke_extract() argument
1017 pa = pte_vatopa(pmap, va); in mmu_booke_extract()
1028 mmu_booke_kextract(vm_offset_t va) in mmu_booke_kextract() argument
1035 if (va >= DMAP_BASE_ADDRESS && va <= DMAP_MAX_ADDRESS) in mmu_booke_kextract()
1036 return (DMAP_TO_PHYS(va)); in mmu_booke_kextract()
1039 if (va >= VM_MIN_KERNEL_ADDRESS && va <= VM_MAX_KERNEL_ADDRESS) in mmu_booke_kextract()
[all …]
H A Dpmap_32.c300 vm_offset_t va; in ptbl_free() local
329 va = ((vm_offset_t)ptbl + (i * PAGE_SIZE)); in ptbl_free()
330 pa = pte_vatopa(kernel_pmap, va); in ptbl_free()
334 mmu_booke_kremove(va); in ptbl_free()
429 pte_remove(pmap_t pmap, vm_offset_t va, uint8_t flags) in pte_remove() argument
431 unsigned int pdir_idx = PDIR_IDX(va); in pte_remove()
432 unsigned int ptbl_idx = PTBL_IDX(va); in pte_remove()
463 pv_remove(pmap, va, m); in pte_remove()
470 pv_remove(pmap, va, m); in pte_remove()
478 tlb0_flush_entry(va); in pte_remove()
[all …]
/src/sys/arm64/iommu/
H A Diommu_pmap.c81 #define smmu_l0_index(va) (((va) >> IOMMU_L0_SHIFT) & IOMMU_L0_ADDR_MASK) argument
82 #define smmu_l1_index(va) (((va) >> IOMMU_L1_SHIFT) & IOMMU_Ln_ADDR_MASK) argument
83 #define smmu_l2_index(va) (((va) >> IOMMU_L2_SHIFT) & IOMMU_Ln_ADDR_MASK) argument
84 #define smmu_l3_index(va) (((va) >> IOMMU_L3_SHIFT) & IOMMU_Ln_ADDR_MASK) argument
87 static void _smmu_pmap_unwire_l3(struct smmu_pmap *pmap, vm_offset_t va,
104 smmu_pmap_l0(struct smmu_pmap *pmap, vm_offset_t va) in smmu_pmap_l0() argument
107 return (&pmap->sp_l0[smmu_l0_index(va)]); in smmu_pmap_l0()
111 smmu_pmap_l0_to_l1(pd_entry_t *l0, vm_offset_t va) in smmu_pmap_l0_to_l1() argument
116 return (&l1[smmu_l1_index(va)]); in smmu_pmap_l0_to_l1()
120 smmu_pmap_l1(struct smmu_pmap *pmap, vm_offset_t va) in smmu_pmap_l1() argument
[all …]
/src/sys/arm/arm/
H A Dpmap-v6.c329 vm_offset_t va);
330 static int pmap_enter_pte1(pmap_t pmap, vm_offset_t va, pt1_entry_t pte1,
332 void cache_icache_sync_fresh(vm_offset_t va, vm_paddr_t pa, vm_size_t size);
651 pt2map_pt2pg(vm_offset_t va) in pt2map_pt2pg() argument
654 va &= ~(NPT2_IN_PG * PTE1_SIZE - 1); in pt2map_pt2pg()
655 return ((vm_offset_t)pt2map_entry(va)); in pt2map_pt2pg()
722 #define KERNEL_V2P(va) \ argument
723 ((vm_paddr_t)((va) - KERNVIRTADDR + arm_physmem_kernaddr))
913 pmap_preboot_pt2pg_setup(vm_offset_t va) in pmap_preboot_pt2pg_setup() argument
919 pte2p = kern_pt2tab_entry(va); in pmap_preboot_pt2pg_setup()
[all …]
/src/sys/arm64/vmm/
H A Dvmm_mmu.c158 vmmpmap_l1_table(vm_offset_t va) in vmmpmap_l1_table() argument
166 l0e = atomic_load_64(&l0[pmap_l0_index(va)]); in vmmpmap_l1_table()
178 rv = atomic_cmpset_64(&l0[pmap_l0_index(va)], l0e, new_l0e); in vmmpmap_l1_table()
197 vmmpmap_l2_table(vm_offset_t va) in vmmpmap_l2_table() argument
203 l1 = vmmpmap_l1_table(va); in vmmpmap_l2_table()
209 l1e = atomic_load_64(&l1[pmap_l1_index(va)]); in vmmpmap_l2_table()
221 rv = atomic_cmpset_64(&l1[pmap_l1_index(va)], l1e, new_l1e); in vmmpmap_l2_table()
240 vmmpmap_l3_table(vm_offset_t va) in vmmpmap_l3_table() argument
246 l2 = vmmpmap_l2_table(va); in vmmpmap_l3_table()
252 l2e = atomic_load_64(&l2[pmap_l2_index(va)]); in vmmpmap_l3_table()
[all …]
/src/sys/arm64/arm64/
H A Dpmap.c318 vm_offset_t va; member
471 static void pmap_pvh_free(struct md_page *pvh, pmap_t pmap, vm_offset_t va);
473 vm_offset_t va);
475 static void pmap_abort_ptp(pmap_t pmap, vm_offset_t va, vm_page_t mpte);
478 static int pmap_change_props_locked(vm_offset_t va, vm_size_t size,
480 static bool pmap_copy_l3c(pmap_t pmap, pt_entry_t *l3p, vm_offset_t va,
482 static pt_entry_t *pmap_demote_l1(pmap_t pmap, pt_entry_t *l1, vm_offset_t va);
484 vm_offset_t va, struct rwlock **lockp);
485 static pt_entry_t *pmap_demote_l2(pmap_t pmap, pt_entry_t *l2, vm_offset_t va);
486 static bool pmap_demote_l2c(pmap_t pmap, pt_entry_t *l2p, vm_offset_t va);
[all …]
/src/sys/riscv/riscv/
H A Dpmap.c317 static void pmap_pvh_free(struct md_page *pvh, pmap_t pmap, vm_offset_t va);
319 vm_offset_t va);
320 static bool pmap_demote_l1(pmap_t pmap, pd_entry_t *l1, vm_offset_t va);
321 static bool pmap_demote_l2(pmap_t pmap, pd_entry_t *l2, vm_offset_t va);
323 vm_offset_t va, struct rwlock **lockp);
324 static int pmap_enter_l2(pmap_t pmap, vm_offset_t va, pd_entry_t new_l2,
326 static vm_page_t pmap_enter_quick_locked(pmap_t pmap, vm_offset_t va,
330 static bool pmap_try_insert_pv_entry(pmap_t pmap, vm_offset_t va,
336 static void _pmap_unwire_ptp(pmap_t pmap, vm_offset_t va, vm_page_t m,
340 static int pmap_change_attr_locked(vm_offset_t va, vm_size_t size, int mode);
[all …]
H A Dcbo.c36 cbo_zicbom_cpu_dcache_wbinv_range(vm_offset_t va, vm_size_t len) in cbo_zicbom_cpu_dcache_wbinv_range() argument
45 va &= ~(dcache_line_size - 1); in cbo_zicbom_cpu_dcache_wbinv_range()
46 for (addr = va; addr < va + len; addr += dcache_line_size) in cbo_zicbom_cpu_dcache_wbinv_range()
52 cbo_zicbom_cpu_dcache_inv_range(vm_offset_t va, vm_size_t len) in cbo_zicbom_cpu_dcache_inv_range() argument
63 va &= ~(dcache_line_size - 1); in cbo_zicbom_cpu_dcache_inv_range()
64 for (addr = va; addr < va + len; addr += dcache_line_size) in cbo_zicbom_cpu_dcache_inv_range()
70 cbo_zicbom_cpu_dcache_wb_range(vm_offset_t va, vm_size_t len) in cbo_zicbom_cpu_dcache_wb_range() argument
83 va &= ~(dcache_line_size - 1); in cbo_zicbom_cpu_dcache_wb_range()
84 for (addr = va; addr < va + len; addr += dcache_line_size) in cbo_zicbom_cpu_dcache_wb_range()
/src/sys/i386/i386/
H A Dpmap.c187 #define vtopte(va) (PTmap + i386_btop(va)) argument
237 vm_offset_t va; member
289 static void pmap_pv_demote_pde(pmap_t pmap, vm_offset_t va, vm_paddr_t pa);
290 static bool pmap_pv_insert_pde(pmap_t pmap, vm_offset_t va, pd_entry_t pde,
293 static void pmap_pv_promote_pde(pmap_t pmap, vm_offset_t va, vm_paddr_t pa);
295 static void pmap_pvh_free(struct md_page *pvh, pmap_t pmap, vm_offset_t va);
297 vm_offset_t va);
300 static void pmap_abort_ptp(pmap_t pmap, vm_offset_t va, vm_page_t mpte);
301 static bool pmap_demote_pde(pmap_t pmap, pd_entry_t *pde, vm_offset_t va);
302 static int pmap_enter_4mpage(pmap_t pmap, vm_offset_t va, vm_page_t m,
[all …]
/src/lib/libkvm/
H A Dkvm_minidump_amd64.c51 #define VA_OFF(vm, va) \ argument
52 (VM_IS_V1(vm) ? ((va) & (AMD64_PAGE_SIZE - 1)) : ((va) & AMD64_PAGE_MASK))
184 _amd64_minidump_vatop_v1(kvm_t *kd, kvaddr_t va, off_t *pa) in _amd64_minidump_vatop_v1() argument
194 offset = va & AMD64_PAGE_MASK; in _amd64_minidump_vatop_v1()
196 if (va >= vm->hdr.kernbase) { in _amd64_minidump_vatop_v1()
197 pteindex = (va - vm->hdr.kernbase) >> AMD64_PAGE_SHIFT; in _amd64_minidump_vatop_v1()
216 } else if (va >= vm->hdr.dmapbase && va < vm->hdr.dmapend) { in _amd64_minidump_vatop_v1()
217 a = (va - vm->hdr.dmapbase) & ~AMD64_PAGE_MASK; in _amd64_minidump_vatop_v1()
222 (uintmax_t)va); in _amd64_minidump_vatop_v1()
230 (uintmax_t)va); in _amd64_minidump_vatop_v1()
[all …]
H A Dkvm_minidump_i386.c146 _i386_minidump_vatop_pae(kvm_t *kd, kvaddr_t va, off_t *pa) in _i386_minidump_vatop_pae() argument
156 offset = va & I386_PAGE_MASK; in _i386_minidump_vatop_pae()
158 if (va >= vm->hdr.kernbase) { in _i386_minidump_vatop_pae()
159 pteindex = (va - vm->hdr.kernbase) >> I386_PAGE_SHIFT; in _i386_minidump_vatop_pae()
181 (uintmax_t)va); in _i386_minidump_vatop_pae()
186 _kvm_err(kd, 0, "invalid address (0x%jx)", (uintmax_t)va); in _i386_minidump_vatop_pae()
191 _i386_minidump_vatop(kvm_t *kd, kvaddr_t va, off_t *pa) in _i386_minidump_vatop() argument
201 offset = va & I386_PAGE_MASK; in _i386_minidump_vatop()
203 if (va >= vm->hdr.kernbase) { in _i386_minidump_vatop()
204 pteindex = (va - vm->hdr.kernbase) >> I386_PAGE_SHIFT; in _i386_minidump_vatop()
[all …]
/src/sys/amd64/amd64/
H A Dpmap.c533 vm_offset_t va; member
596 #define PMAP_ADDRESS_IN_LARGEMAP(va) ((va) >= kva_layout.lm_low && \ argument
597 (va) < kva_layout.lm_high)
639 static pt_entry_t pmap_pkru_get(pmap_t pmap, vm_offset_t va);
1330 static void pmap_pv_demote_pde(pmap_t pmap, vm_offset_t va, vm_paddr_t pa,
1332 static bool pmap_pv_insert_pde(pmap_t pmap, vm_offset_t va, pd_entry_t pde,
1335 static void pmap_pv_promote_pde(pmap_t pmap, vm_offset_t va, vm_paddr_t pa,
1338 static void pmap_pvh_free(struct md_page *pvh, pmap_t pmap, vm_offset_t va);
1340 vm_offset_t va);
1342 static void pmap_abort_ptp(pmap_t pmap, vm_offset_t va, vm_page_t mpte);
[all …]
/src/sys/kern/
H A Dsubr_devmap.c223 vm_offset_t va; in devmap_vtop() local
228 va = (vm_offset_t)vpva; in devmap_vtop()
230 if (va >= pd->pd_va && va + size <= pd->pd_va + pd->pd_size) in devmap_vtop()
231 return ((vm_paddr_t)(pd->pd_pa + (va - pd->pd_va))); in devmap_vtop()
258 vm_offset_t va, offset; in pmap_mapdev_attr() local
281 va = akva_devmap_vaddr; in pmap_mapdev_attr()
282 KASSERT(va >= (VM_MAX_KERNEL_ADDRESS - PMAP_MAPDEV_EARLY_SIZE), in pmap_mapdev_attr()
288 va = kva_alloc_aligned(size, L2_SIZE); in pmap_mapdev_attr()
290 va = kva_alloc_aligned(size, L3C_SIZE); in pmap_mapdev_attr()
293 va = kva_alloc(size); in pmap_mapdev_attr()
[all …]
/src/sys/powerpc/aim/
H A Dmmu_radix.c101 static void pmap_pte_walk(pml1_entry_t *l1, vm_offset_t va);
191 vm_offset_t va, uint16_t ap) in radix_tlbie() argument
195 MPASS((va & PAGE_MASK) == 0); in radix_tlbie()
198 rb = va | is | ap; in radix_tlbie()
204 radix_tlbie_fixup(uint32_t pid, vm_offset_t va, int ap) in radix_tlbie_fixup() argument
209 TLBIEL_INVAL_PAGE, 0, 0, va, ap); in radix_tlbie_fixup()
212 TLBIEL_INVAL_PAGE, pid, 0, va, ap); in radix_tlbie_fixup()
216 radix_tlbie_invlpg_user_4k(uint32_t pid, vm_offset_t va) in radix_tlbie_invlpg_user_4k() argument
220 TLBIEL_INVAL_PAGE, pid, 0, va, TLBIE_ACTUAL_PAGE_4K); in radix_tlbie_invlpg_user_4k()
221 radix_tlbie_fixup(pid, va, TLBIE_ACTUAL_PAGE_4K); in radix_tlbie_invlpg_user_4k()
[all …]
H A Dmmu_oea.c324 void moea_dumpsys_map(vm_paddr_t pa, size_t sz, void **va);
457 tlbie(vm_offset_t va) in tlbie() argument
462 __asm __volatile("tlbie %0" :: "r"(va)); in tlbie()
470 vm_offset_t va; in tlbia() local
472 for (va = 0; va < 0x00040000; va += 0x00001000) { in tlbia()
473 __asm __volatile("tlbie %0" :: "r"(va)); in tlbia()
481 va_to_sr(u_int *sr, vm_offset_t va) in va_to_sr() argument
483 return (sr[(uintptr_t)va >> ADDR_SR_SHFT]); in va_to_sr()
536 moea_pte_match(struct pte *pt, u_int sr, vm_offset_t va, int which) in moea_pte_match() argument
540 ((va >> ADDR_API_SHFT) & PTE_API) | which); in moea_pte_match()
[all …]
/src/sys/fs/smbfs/
H A Dsmbfs_node.c301 struct vattr va; in smbfs_inactive() local
309 VOP_GETATTR(vp, &va, cred); in smbfs_inactive()
351 smbfs_attr_cachelookup(struct vnode *vp, struct vattr *va) in smbfs_attr_cachelookup() argument
360 va->va_type = vp->v_type; /* vnode type (for create) */ in smbfs_attr_cachelookup()
361 va->va_flags = 0; /* flags defined for file */ in smbfs_attr_cachelookup()
363 va->va_mode = smp->sm_file_mode; /* files access mode and type */ in smbfs_attr_cachelookup()
365 va->va_mode &= ~(S_IWUSR|S_IWGRP|S_IWOTH); in smbfs_attr_cachelookup()
366 va->va_flags |= UF_READONLY; in smbfs_attr_cachelookup()
369 va->va_mode = smp->sm_dir_mode; /* files access mode and type */ in smbfs_attr_cachelookup()
372 va->va_size = np->n_size; in smbfs_attr_cachelookup()
[all …]
/src/contrib/tcsh/
H A Dtc.printf.c371 va_list va; in xsnprintf() local
372 va_start(va, fmt); in xsnprintf()
376 count = doprnt(xaddchar, fmt, va); in xsnprintf()
377 va_end(va); in xsnprintf()
387 va_list va; in xprintf() local
388 va_start(va, fmt); in xprintf()
389 count = doprnt(xputchar, fmt, va); in xprintf()
390 va_end(va); in xprintf()
395 xvprintf(const char *fmt, va_list va) in xvprintf() argument
397 return doprnt(xputchar, fmt, va); in xvprintf()
[all …]
/src/sys/dev/xdma/
H A Dxdma_iommu.c62 xdma_iommu_remove_entry(xdma_channel_t *xchan, vm_offset_t va) in xdma_iommu_remove_entry() argument
68 va &= ~(PAGE_SIZE - 1); in xdma_iommu_remove_entry()
69 pmap_remove(&xio->p, va, va + PAGE_SIZE); in xdma_iommu_remove_entry()
71 XDMA_IOMMU_REMOVE(xio->dev, xio, va); in xdma_iommu_remove_entry()
73 vmem_free(xio->vmem, va, PAGE_SIZE); in xdma_iommu_remove_entry()
77 xdma_iommu_enter(struct xdma_iommu *xio, vm_offset_t va, in xdma_iommu_enter() argument
90 pmap_enter(p, va, m, prot, prot | PMAP_ENTER_WIRED, 0); in xdma_iommu_enter()
92 XDMA_IOMMU_ENTER(xio->dev, xio, va, pa); in xdma_iommu_enter()
94 va += PAGE_SIZE; in xdma_iommu_enter()
100 xdma_iommu_add_entry(xdma_channel_t *xchan, vm_offset_t *va, in xdma_iommu_add_entry() argument
[all …]
/src/contrib/llvm-project/lld/MachO/Arch/
H A DARM64Common.h26 void relocateOne(uint8_t *loc, const Reloc &, uint64_t va,
46 uint64_t va) { in encodeBranch26() argument
47 checkInt(loc, r, va, 28); in encodeBranch26()
50 llvm::support::endian::write32le(loc, base | bitField(va, 2, 26, 0)); in encodeBranch26()
54 uint64_t va) { in encodeBranch26() argument
55 checkInt(loc, d, va, 28); in encodeBranch26()
56 llvm::support::endian::write32le(loc, base | bitField(va, 2, 26, 0)); in encodeBranch26()
65 uint64_t va) { in encodePage21() argument
66 checkInt(loc, r, va, 35); in encodePage21()
67 llvm::support::endian::write32le(loc, base | bitField(va, 12, 2, 29) | in encodePage21()
[all …]
/src/contrib/netbsd-tests/fs/puffs/h_dtfs/
H A Ddtfs_vnops.c110 const struct vattr *va, const struct puffs_cred *pcr) in dtfs_node_setattr() argument
116 if (va->va_flags != PUFFS_VNOVAL) in dtfs_node_setattr()
119 if (va->va_uid != PUFFS_VNOVAL || va->va_gid != PUFFS_VNOVAL) { in dtfs_node_setattr()
121 va->va_uid, va->va_gid, pcr); in dtfs_node_setattr()
126 if (va->va_mode != PUFFS_VNOVAL) { in dtfs_node_setattr()
128 pn->pn_va.va_type, va->va_mode, pcr); in dtfs_node_setattr()
133 if ((va->va_atime.tv_sec != PUFFS_VNOVAL in dtfs_node_setattr()
134 && va->va_atime.tv_nsec != PUFFS_VNOVAL) in dtfs_node_setattr()
135 || (va->va_mtime.tv_sec != PUFFS_VNOVAL in dtfs_node_setattr()
136 && va->va_mtime.tv_nsec != PUFFS_VNOVAL)) { in dtfs_node_setattr()
[all …]
/src/sys/riscv/thead/
H A Dthead.c56 thead_cpu_dcache_wbinv_range(vm_offset_t va, vm_size_t len) in thead_cpu_dcache_wbinv_range() argument
58 register vm_offset_t t0 __asm("t0") = rounddown(va, dcache_line_size); in thead_cpu_dcache_wbinv_range()
60 for (; t0 < va + len; t0 += dcache_line_size) { in thead_cpu_dcache_wbinv_range()
68 thead_cpu_dcache_inv_range(vm_offset_t va, vm_size_t len) in thead_cpu_dcache_inv_range() argument
70 register vm_offset_t t0 __asm("t0") = rounddown(va, dcache_line_size); in thead_cpu_dcache_inv_range()
72 for (; t0 < va + len; t0 += dcache_line_size) { in thead_cpu_dcache_inv_range()
80 thead_cpu_dcache_wb_range(vm_offset_t va, vm_size_t len) in thead_cpu_dcache_wb_range() argument
82 register vm_offset_t t0 __asm("t0") = rounddown(va, dcache_line_size); in thead_cpu_dcache_wb_range()
84 for (; t0 < va + len; t0 += dcache_line_size) { in thead_cpu_dcache_wb_range()
/src/sys/dev/smbios/
H A Dsmbios.c194 void *va; in smbios_probe() local
201 va = pmap_mapbios(pa, size); in smbios_probe()
202 if (va == NULL) { in smbios_probe()
207 if (smbios_cksum(va)) { in smbios_probe()
212 pmap_unmapbios(va, size); in smbios_probe()
220 void *va; in smbios_attach() local
228 va = pmap_mapbios(pa, size); in smbios_attach()
229 if (va == NULL) { in smbios_attach()
233 sc->is_eps3 = smbios_eps3(va); in smbios_attach()
236 sc->eps3 = va; in smbios_attach()
[all …]

12345678910>>...18