Home
last modified time | relevance | path

Searched refs:kvm_pfn_t (Results 1 – 25 of 36) sorted by relevance

12

/linux/virt/kvm/
H A Dpfncache.c99 static void *gpc_map(kvm_pfn_t pfn) in gpc_map()
111 static void gpc_unmap(kvm_pfn_t pfn, void *khva) in gpc_unmap()
155 static kvm_pfn_t hva_to_pfn_retry(struct gfn_to_pfn_cache *gpc) in hva_to_pfn_retry()
159 kvm_pfn_t new_pfn = KVM_PFN_ERR_FAULT; in hva_to_pfn_retry()
261 kvm_pfn_t old_pfn; in __kvm_gpc_refresh()
450 kvm_pfn_t old_pfn; in kvm_gpc_deactivate()
H A Dguest_memfd.c53 static inline kvm_pfn_t folio_file_pfn(struct folio *folio, pgoff_t index) in folio_file_pfn()
67 kvm_pfn_t pfn = folio_file_pfn(folio, index); in __kvm_gmem_prepare_folio()
530 kvm_pfn_t pfn = page_to_pfn(page); in kvm_gmem_free_folio()
755 pgoff_t index, kvm_pfn_t *pfn, in __kvm_gmem_get_pfn()
790 gfn_t gfn, kvm_pfn_t *pfn, struct page **page, in kvm_gmem_get_pfn()
831 kvm_pfn_t pfn; in __kvm_gmem_populate()
H A Dkvm_mm.h56 kvm_pfn_t hva_to_pfn(struct kvm_follow_pfn *kfp);
H A Dkvm_main.c2834 static kvm_pfn_t kvm_resolve_pfn(struct kvm_follow_pfn *kfp, struct page *page, in kvm_resolve_pfn()
2837 kvm_pfn_t pfn; in kvm_resolve_pfn()
2858 static bool hva_to_pfn_fast(struct kvm_follow_pfn *kfp, kvm_pfn_t *pfn) in hva_to_pfn_fast()
2890 static int hva_to_pfn_slow(struct kvm_follow_pfn *kfp, kvm_pfn_t *pfn) in hva_to_pfn_slow()
2947 struct kvm_follow_pfn *kfp, kvm_pfn_t *p_pfn) in hva_to_pfn_remapped()
2992 kvm_pfn_t hva_to_pfn(struct kvm_follow_pfn *kfp) in hva_to_pfn()
2995 kvm_pfn_t pfn; in hva_to_pfn()
3037 static kvm_pfn_t kvm_follow_pfn(struct kvm_follow_pfn *kfp) in kvm_follow_pfn()
3056 kvm_pfn_t __kvm_faultin_pfn(const struct kvm_memory_slot *slot, gfn_t gfn, in __kvm_faultin_pfn()
/linux/include/linux/
H A Dkvm_types.h74 typedef hfn_t kvm_pfn_t; typedef
94 kvm_pfn_t pfn;
H A Dkvm_host.h107 static inline bool is_error_pfn(kvm_pfn_t pfn) in is_error_pfn()
116 static inline bool is_sigpending_pfn(kvm_pfn_t pfn) in is_sigpending_pfn()
126 static inline bool is_error_noslot_pfn(kvm_pfn_t pfn) in is_error_noslot_pfn()
132 static inline bool is_noslot_pfn(kvm_pfn_t pfn) in is_noslot_pfn()
295 kvm_pfn_t pfn;
296 kvm_pfn_t gfn;
1303 kvm_pfn_t __kvm_faultin_pfn(const struct kvm_memory_slot *slot, gfn_t gfn,
1307 static inline kvm_pfn_t kvm_faultin_pfn(struct kvm_vcpu *vcpu, gfn_t gfn, in kvm_faultin_pfn()
1914 static inline hpa_t pfn_to_hpa(kvm_pfn_t pfn) in pfn_to_hpa()
2542 gfn_t gfn, kvm_pfn_t *pfn, struct page **page,
[all …]
/linux/arch/x86/kvm/svm/
H A Dsvm.h897 int sev_gmem_prepare(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order);
898 void sev_gmem_invalidate(kvm_pfn_t start, kvm_pfn_t end);
899 int sev_gmem_max_mapping_level(struct kvm *kvm, kvm_pfn_t pfn, bool is_private);
923 static inline int sev_gmem_prepare(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order) in sev_gmem_prepare()
927 static inline void sev_gmem_invalidate(kvm_pfn_t start, kvm_pfn_t end) {} in sev_gmem_invalidate()
928 static inline int sev_gmem_max_mapping_level(struct kvm *kvm, kvm_pfn_t pfn, bool is_private) in sev_gmem_max_mapping_level()
H A Dsev.c2280 static int sev_gmem_post_populate(struct kvm *kvm, gfn_t gfn, kvm_pfn_t pfn, in sev_gmem_post_populate()
3702 static int snp_rmptable_psmash(kvm_pfn_t pfn) in snp_rmptable_psmash()
3967 kvm_pfn_t pfn; in sev_snp_init_protected_guest_state()
4850 kvm_pfn_t pfn; in sev_handle_rmp_fault()
4936 static bool is_pfn_range_shared(kvm_pfn_t start, kvm_pfn_t end) in is_pfn_range_shared()
4938 kvm_pfn_t pfn = start; in is_pfn_range_shared()
4971 static bool is_large_rmp_possible(struct kvm *kvm, kvm_pfn_t pfn, int order) in is_large_rmp_possible()
4973 kvm_pfn_t pfn_aligned = ALIGN_DOWN(pfn, PTRS_PER_PMD); in is_large_rmp_possible()
4987 int sev_gmem_prepare(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order) in sev_gmem_prepare()
4990 kvm_pfn_t pfn_aligned; in sev_gmem_prepare()
[all …]
/linux/arch/x86/kvm/mmu/
H A Dspte.c107 static bool __kvm_is_mmio_pfn(kvm_pfn_t pfn) in __kvm_is_mmio_pfn()
128 static bool kvm_is_mmio_pfn(kvm_pfn_t pfn, int *is_host_mmio) in kvm_is_mmio_pfn()
188 unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn, in make_spte()
H A Dspte.h363 static inline kvm_pfn_t spte_to_pfn(u64 pte) in spte_to_pfn()
545 unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn,
H A Dmmu_internal.h281 kvm_pfn_t pfn;
/linux/arch/powerpc/kvm/
H A De500_mmu_host.c167 kvm_pfn_t pfn; in kvmppc_map_magic()
169 pfn = (kvm_pfn_t)virt_to_phys((void *)shared_page) >> PAGE_SHIFT; in kvmppc_map_magic()
250 kvm_pfn_t pfn, unsigned int wimg, in kvmppc_e500_tlbe_setup()
307 kvm_pfn_t pfn = tlbe->pfn; in kvmppc_e500_setup_stlbe()
H A Dbook3s.c424 kvm_pfn_t kvmppc_gpa_to_pfn(struct kvm_vcpu *vcpu, gpa_t gpa, bool writing, in kvmppc_gpa_to_pfn()
437 kvm_pfn_t pfn; in kvmppc_gpa_to_pfn()
439 pfn = (kvm_pfn_t)virt_to_phys((void*)shared_page) >> PAGE_SHIFT; in kvmppc_gpa_to_pfn()
H A De500.h43 kvm_pfn_t pfn; /* valid only for TLB0, except briefly */
H A Dtrace_pr.h32 TP_PROTO(int rflags, ulong hpteg, ulong va, kvm_pfn_t hpaddr,
H A Dbook3s_32_mmu_host.c132 kvm_pfn_t hpaddr; in kvmppc_mmu_map_page()
H A Dbook3s_64_mmu_host.c75 kvm_pfn_t hpaddr; in kvmppc_mmu_map_page()
/linux/arch/riscv/kvm/
H A Dmmu.c422 kvm_pfn_t *hfnp, gpa_t *gpa) in transparent_hugepage_adjust()
424 kvm_pfn_t hfn = *hfnp; in transparent_hugepage_adjust()
453 kvm_pfn_t hfn; in kvm_riscv_mmu_map()
/linux/arch/s390/kvm/
H A Ddat.h393 static inline union pte _pte(kvm_pfn_t pfn, bool writable, bool dirty, bool special) in _pte()
407 static inline union crste _crste_fc0(kvm_pfn_t pfn, int tt) in _crste_fc0()
431 static inline union crste _crste_fc1(kvm_pfn_t pfn, int tt, bool writable, bool dirty) in _crste_fc1()
490 kvm_pfn_t pfn; /* Host PFN */
H A Dgmap.h104 kvm_pfn_t pfn, int level, bool wr);
/linux/arch/arm64/kvm/
H A Dmmu.c1403 unsigned long hva, kvm_pfn_t *pfnp, in transparent_hugepage_adjust()
1406 kvm_pfn_t pfn = *pfnp; in transparent_hugepage_adjust()
1470 static void sanitise_mte_tags(struct kvm *kvm, kvm_pfn_t pfn, in sanitise_mte_tags()
1579 kvm_pfn_t pfn; in gmem_abort()
1659 kvm_pfn_t pfn; in user_mem_abort()
/linux/arch/x86/kvm/
H A Dmmu.h258 int kvm_tdp_mmu_map_private_pfn(struct kvm_vcpu *vcpu, gfn_t gfn, kvm_pfn_t pfn);
/linux/arch/x86/kvm/vmx/
H A Dx86_ops.h156 int tdx_gmem_max_mapping_level(struct kvm *kvm, kvm_pfn_t pfn, bool is_private);
/linux/arch/x86/include/asm/
H A Dkvm_host.h1962 int (*gmem_prepare)(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order);
1963 void (*gmem_invalidate)(kvm_pfn_t start, kvm_pfn_t end);
1964 int (*gmem_max_mapping_level)(struct kvm *kvm, kvm_pfn_t pfn, bool is_private);
/linux/arch/arm64/include/asm/
H A Dkvm_pgtable.h151 static inline kvm_pfn_t kvm_pte_to_pfn(kvm_pte_t pte) in kvm_pte_to_pfn()

12