Home
last modified time | relevance | path

Searched refs:pfns (Results 1 – 25 of 31) sorted by relevance

12

/linux/drivers/iommu/iommufd/
H A Dpages.c269 unsigned long *pfns; member
280 batch->pfns[0] = 0; in batch_clear()
298 batch->pfns[0] = batch->pfns[batch->end - 1] + in batch_clear_carry()
311 batch->pfns[0] += skip_pfns; in batch_skip_carry()
319 const size_t elmsz = sizeof(*batch->pfns) + sizeof(*batch->npfns); in __batch_init()
322 batch->pfns = temp_kmalloc(&size, backup, backup_len); in __batch_init()
323 if (!batch->pfns) in __batch_init()
328 batch->npfns = (u32 *)(batch->pfns + batch->array_size); in __batch_init()
346 if (batch->pfns != backup) in batch_destroy()
347 kfree(batch->pfns); in batch_destroy()
[all …]
H A Dselftest.c129 struct xarray pfns; member
356 ent = xa_load(&mock->pfns, cur / MOCK_IO_PAGE_SIZE); in mock_test_and_clear_dirty()
366 old = xa_store(&mock->pfns, cur / MOCK_IO_PAGE_SIZE, in mock_test_and_clear_dirty()
391 ent = xa_load(&mock->pfns, iova / MOCK_IO_PAGE_SIZE); in mock_domain_read_and_clear_dirty()
489 xa_init(&mock->pfns); in mock_domain_alloc_paging_flags()
500 WARN_ON(!xa_empty(&mock->pfns)); in mock_domain_free()
533 old = xa_store(&mock->pfns, iova / MOCK_IO_PAGE_SIZE, in mock_domain_map_pages()
540 xa_erase(&mock->pfns, in mock_domain_map_pages()
572 ent = xa_erase(&mock->pfns, iova / MOCK_IO_PAGE_SIZE); in mock_domain_unmap_pages()
612 ent = xa_load(&mock->pfns, iova / MOCK_IO_PAGE_SIZE); in mock_domain_iova_to_phys()
[all …]
/linux/drivers/xen/
H A Dxlate_mmu.c193 xen_pfn_t *pfns; member
201 info->pfns[info->idx++] = gfn; in setup_balloon_gfn()
218 xen_pfn_t *pfns; in xen_xlate_map_ballooned_pages() local
230 pfns = kcalloc(nr_grant_frames, sizeof(pfns[0]), GFP_KERNEL); in xen_xlate_map_ballooned_pages()
231 if (!pfns) { in xen_xlate_map_ballooned_pages()
240 kfree(pfns); in xen_xlate_map_ballooned_pages()
244 data.pfns = pfns; in xen_xlate_map_ballooned_pages()
254 kfree(pfns); in xen_xlate_map_ballooned_pages()
259 *gfns = pfns; in xen_xlate_map_ballooned_pages()
H A Dprivcmd.c741 xen_pfn_t *pfns = NULL; in privcmd_ioctl_mmap_resource() local
776 pfns = kcalloc(kdata.num, sizeof(*pfns), GFP_KERNEL | __GFP_NOWARN); in privcmd_ioctl_mmap_resource()
777 if (!pfns) { in privcmd_ioctl_mmap_resource()
797 pfns[i] = pfn + (i % XEN_PFN_PER_PAGE); in privcmd_ioctl_mmap_resource()
804 set_xen_guest_handle(xdata.frame_list, pfns); in privcmd_ioctl_mmap_resource()
819 int num, *errs = (int *)pfns; in privcmd_ioctl_mmap_resource()
821 BUILD_BUG_ON(sizeof(*errs) > sizeof(*pfns)); in privcmd_ioctl_mmap_resource()
824 pfns, kdata.num, errs, in privcmd_ioctl_mmap_resource()
843 kfree(pfns); in privcmd_ioctl_mmap_resource()
/linux/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_hmm.c176 unsigned long *pfns; in amdgpu_hmm_range_get_pages() local
183 pfns = kvmalloc_array(npages, sizeof(*pfns), GFP_KERNEL); in amdgpu_hmm_range_get_pages()
184 if (unlikely(!pfns)) { in amdgpu_hmm_range_get_pages()
193 hmm_range->hmm_pfns = pfns; in amdgpu_hmm_range_get_pages()
222 hmm_range->hmm_pfns = pfns; in amdgpu_hmm_range_get_pages()
229 kvfree(pfns); in amdgpu_hmm_range_get_pages()
H A Damdgpu_umc.h187 uint64_t pa_addr, uint64_t *pfns, int len);
H A Damdgpu_umc.c510 uint64_t pa_addr, uint64_t *pfns, int len) in amdgpu_umc_lookup_bad_pages_in_a_row() argument
530 pfns[i] = err_data.err_addr[i].retired_page; in amdgpu_umc_lookup_bad_pages_in_a_row()
/linux/drivers/gpu/drm/
H A Ddrm_gpusvm.c705 unsigned long *pfns; in drm_gpusvm_check_pages() local
711 pfns = kvmalloc_array(npages, sizeof(*pfns), GFP_KERNEL); in drm_gpusvm_check_pages()
712 if (!pfns) in drm_gpusvm_check_pages()
716 hmm_range.hmm_pfns = pfns; in drm_gpusvm_check_pages()
734 if (!(pfns[i] & HMM_PFN_VALID)) { in drm_gpusvm_check_pages()
738 i += 0x1 << drm_gpusvm_hmm_pfn_to_order(pfns[i], i, npages); in drm_gpusvm_check_pages()
742 kvfree(pfns); in drm_gpusvm_check_pages()
1282 unsigned long *pfns; in drm_gpusvm_get_pages() local
1295 pfns = kvmalloc_array(npages, sizeof(*pfns), GFP_KERNEL); in drm_gpusvm_get_pages()
1296 if (!pfns) in drm_gpusvm_get_pages()
[all …]
/linux/mm/
H A Dhmm.c772 unsigned long *pfns = map->pfn_list; in hmm_dma_map_pfn() local
773 struct page *page = hmm_pfn_to_page(pfns[idx]); in hmm_dma_map_pfn()
774 phys_addr_t paddr = hmm_pfn_to_phys(pfns[idx]); in hmm_dma_map_pfn()
780 if ((pfns[idx] & HMM_PFN_DMA_MAPPED) && in hmm_dma_map_pfn()
781 !(pfns[idx] & HMM_PFN_P2PDMA_BUS)) { in hmm_dma_map_pfn()
810 pfns[idx] |= HMM_PFN_P2PDMA; in hmm_dma_map_pfn()
813 pfns[idx] |= HMM_PFN_P2PDMA_BUS | HMM_PFN_DMA_MAPPED; in hmm_dma_map_pfn()
846 pfns[idx] |= HMM_PFN_DMA_MAPPED; in hmm_dma_map_pfn()
849 pfns[idx] &= ~HMM_PFN_P2PDMA; in hmm_dma_map_pfn()
868 unsigned long *pfns = map->pfn_list; in hmm_dma_unmap_pfn() local
[all …]
H A Dsparse.c195 unsigned long pfns; in subsection_map_init() local
197 pfns = min(nr_pages, PAGES_PER_SECTION in subsection_map_init()
200 subsection_mask_set(ms->usage->subsection_map, pfn, pfns); in subsection_map_init()
203 pfns, subsection_map_index(pfn), in subsection_map_init()
204 subsection_map_index(pfn + pfns - 1)); in subsection_map_init()
206 pfn += pfns; in subsection_map_init()
207 nr_pages -= pfns; in subsection_map_init()
/linux/tools/mm/
H A Dthpmaps229 def thp_parse(vma, kpageflags, ranges, indexes, vfns, pfns, anons, heads):
399 pfns = pmes & PM_PFN_MASK
400 pfns = pfns[present]
404 pfn_vec = cont_ranges_all([pfns], [pfns])[0]
411 pfns = pfns[thps]
416 ranges = cont_ranges_all([vfns, pfns], [indexes, vfns, pfns])
418 thpstats = thp_parse(vma, kpageflags, ranges, indexes, vfns, pfns, anons, heads)
/linux/drivers/gpu/drm/nouveau/
H A Dnouveau_svm.h34 void nouveau_pfns_free(u64 *pfns);
36 unsigned long addr, u64 *pfns, unsigned long npages);
H A Dnouveau_dmem.c658 dma_addr_t *dma_addrs, u64 *pfns) in nouveau_dmem_migrate_chunk() argument
665 args->src[i], dma_addrs + nr_dma, pfns + i); in nouveau_dmem_migrate_chunk()
674 nouveau_pfns_map(svmm, args->vma->vm_mm, args->start, pfns, i); in nouveau_dmem_migrate_chunk()
700 u64 *pfns; in nouveau_dmem_migrate_vma() local
717 pfns = nouveau_pfns_alloc(max); in nouveau_dmem_migrate_vma()
718 if (!pfns) in nouveau_dmem_migrate_vma()
733 pfns); in nouveau_dmem_migrate_vma()
739 nouveau_pfns_free(pfns); in nouveau_dmem_migrate_vma()
H A Dnouveau_svm.c893 nouveau_pfns_to_args(void *pfns) in nouveau_pfns_to_args() argument
895 return container_of(pfns, struct nouveau_pfnmap_args, p.phys); in nouveau_pfns_to_args()
915 nouveau_pfns_free(u64 *pfns) in nouveau_pfns_free() argument
917 struct nouveau_pfnmap_args *args = nouveau_pfns_to_args(pfns); in nouveau_pfns_free()
924 unsigned long addr, u64 *pfns, unsigned long npages) in nouveau_pfns_map() argument
926 struct nouveau_pfnmap_args *args = nouveau_pfns_to_args(pfns); in nouveau_pfns_map()
/linux/drivers/virtio/
H A Dvirtio_balloon.c108 __virtio32 pfns[VIRTIO_BALLOON_ARRAY_PFNS_MAX]; member
188 sg_init_one(&sg, vb->pfns, sizeof(vb->pfns[0]) * vb->num_pfns); in tell_host()
227 __virtio32 pfns[], struct page *page) in set_page_pfns() argument
238 pfns[i] = cpu_to_virtio32(vb->vdev, in set_page_pfns()
250 num = min(num, ARRAY_SIZE(vb->pfns)); in fill_balloon()
275 set_page_pfns(vb, vb->pfns + vb->num_pfns, page); in fill_balloon()
314 num = min(num, ARRAY_SIZE(vb->pfns)); in leak_balloon()
324 set_page_pfns(vb, vb->pfns + vb->num_pfns, page); in leak_balloon()
863 set_page_pfns(vb, vb->pfns, newpage); in virtballoon_migratepage()
868 set_page_pfns(vb, vb->pfns, page); in virtballoon_migratepage()
/linux/drivers/gpu/drm/i915/gem/
H A Di915_gem_pages.c336 unsigned long stack[32], *pfns = stack, i; in i915_gem_object_map_pfn() local
345 pfns = kvmalloc_array(n_pfn, sizeof(*pfns), GFP_KERNEL); in i915_gem_object_map_pfn()
346 if (!pfns) in i915_gem_object_map_pfn()
352 pfns[i++] = (iomap + addr) >> PAGE_SHIFT; in i915_gem_object_map_pfn()
353 vaddr = vmap_pfn(pfns, n_pfn, pgprot_writecombine(PAGE_KERNEL_IO)); in i915_gem_object_map_pfn()
354 if (pfns != stack) in i915_gem_object_map_pfn()
355 kvfree(pfns); in i915_gem_object_map_pfn()
/linux/arch/powerpc/kvm/
H A Dbook3s_hv_uvmem.c230 unsigned long *pfns; member
255 p->pfns = vcalloc(slot->npages, sizeof(*p->pfns)); in kvmppc_uvmem_slot_init()
256 if (!p->pfns) { in kvmppc_uvmem_slot_init()
280 vfree(p->pfns); in kvmppc_uvmem_slot_free()
299 p->pfns[index] = uvmem_pfn | flag; in kvmppc_mark_gfn()
301 p->pfns[index] = flag; in kvmppc_mark_gfn()
342 if (p->pfns[index] & KVMPPC_GFN_UVMEM_PFN) { in kvmppc_gfn_is_uvmem_pfn()
344 *uvmem_pfn = p->pfns[index] & in kvmppc_gfn_is_uvmem_pfn()
382 if (!(p->pfns[index] & KVMPPC_GFN_FLAG_MASK)) { in kvmppc_next_nontransitioned_gfn()
/linux/lib/
H A Dtest_hmm.c210 unsigned long *pfns = range->hmm_pfns; in dmirror_do_fault() local
215 pfn++, pfns++) { in dmirror_do_fault()
223 WARN_ON(*pfns & HMM_PFN_ERROR); in dmirror_do_fault()
224 WARN_ON(!(*pfns & HMM_PFN_VALID)); in dmirror_do_fault()
226 page = hmm_pfn_to_page(*pfns); in dmirror_do_fault()
230 if (*pfns & HMM_PFN_WRITE) in dmirror_do_fault()
333 unsigned long pfns[32]; in dmirror_fault() local
336 .hmm_pfns = pfns, in dmirror_fault()
350 range.end = min(addr + (ARRAY_SIZE(pfns) << PAGE_SHIFT), end); in dmirror_fault()
1147 unsigned long pfns[32]; in dmirror_snapshot() local
[all …]
/linux/Documentation/translations/zh_CN/mm/
H A Dhmm.rst165 /* Use pfns array content to update device page table,
179 的故障或快照策略,而不必为 pfns 数组中的每个条目设置它们。
193 range->pfns[index_of_write] = HMM_PFN_REQ_WRITE;
H A Dmemory-model.rst115 为给定的pfns范围执行足够的内存热插拔来开启 :c:func:`pfn_to_page`,
/linux/drivers/iommu/
H A Diova.c575 unsigned long pfns[IOVA_MAG_SIZE]; member
626 struct iova *iova = private_find_iova(iovad, mag->pfns[i]); in iova_magazine_free_pfns()
657 for (i = mag->size - 1; mag->pfns[i] > limit_pfn; i--) in iova_magazine_pop()
662 pfn = mag->pfns[i]; in iova_magazine_pop()
663 mag->pfns[i] = mag->pfns[--mag->size]; in iova_magazine_pop()
670 mag->pfns[mag->size++] = pfn; in iova_magazine_push()
/linux/arch/s390/include/asm/
H A Dgmap.h137 void s390_uv_destroy_pfns(unsigned long count, unsigned long *pfns);
/linux/include/xen/
H A Dxen-ops.h188 int xen_xlate_map_ballooned_pages(xen_pfn_t **pfns, void **vaddr,
/linux/include/linux/
H A Dvmalloc.h215 void *vmap_pfn(unsigned long *pfns, unsigned int count, pgprot_t prot);
/linux/arch/s390/mm/
H A Dgmap.c2338 unsigned long pfns[GATHER_GET_PAGES]; member
2350 p->pfns[p->count] = phys_to_pfn(pte_val(pte)); in s390_gather_pages()
2366 void s390_uv_destroy_pfns(unsigned long count, unsigned long *pfns) in s390_uv_destroy_pfns() argument
2372 folio = pfn_folio(pfns[i]); in s390_uv_destroy_pfns()
2408 s390_uv_destroy_pfns(state.count, state.pfns); in __s390_uv_destroy_range()

12