| /linux/virt/kvm/ |
| H A D | pfncache.c | 99 static void *gpc_map(kvm_pfn_t pfn) in gpc_map() 111 static void gpc_unmap(kvm_pfn_t pfn, void *khva) in gpc_unmap() 155 static kvm_pfn_t hva_to_pfn_retry(struct gfn_to_pfn_cache *gpc) in hva_to_pfn_retry() 159 kvm_pfn_t new_pfn = KVM_PFN_ERR_FAULT; in hva_to_pfn_retry() 261 kvm_pfn_t old_pfn; in __kvm_gpc_refresh() 450 kvm_pfn_t old_pfn; in kvm_gpc_deactivate()
|
| H A D | guest_memfd.c | 53 static inline kvm_pfn_t folio_file_pfn(struct folio *folio, pgoff_t index) in folio_file_pfn() 67 kvm_pfn_t pfn = folio_file_pfn(folio, index); in __kvm_gmem_prepare_folio() 530 kvm_pfn_t pfn = page_to_pfn(page); in kvm_gmem_free_folio() 755 pgoff_t index, kvm_pfn_t *pfn, in __kvm_gmem_get_pfn() 790 gfn_t gfn, kvm_pfn_t *pfn, struct page **page, in kvm_gmem_get_pfn() 831 kvm_pfn_t pfn; in __kvm_gmem_populate()
|
| H A D | kvm_mm.h | 56 kvm_pfn_t hva_to_pfn(struct kvm_follow_pfn *kfp);
|
| H A D | kvm_main.c | 2834 static kvm_pfn_t kvm_resolve_pfn(struct kvm_follow_pfn *kfp, struct page *page, in kvm_resolve_pfn() 2837 kvm_pfn_t pfn; in kvm_resolve_pfn() 2858 static bool hva_to_pfn_fast(struct kvm_follow_pfn *kfp, kvm_pfn_t *pfn) in hva_to_pfn_fast() 2890 static int hva_to_pfn_slow(struct kvm_follow_pfn *kfp, kvm_pfn_t *pfn) in hva_to_pfn_slow() 2947 struct kvm_follow_pfn *kfp, kvm_pfn_t *p_pfn) in hva_to_pfn_remapped() 2992 kvm_pfn_t hva_to_pfn(struct kvm_follow_pfn *kfp) in hva_to_pfn() 2995 kvm_pfn_t pfn; in hva_to_pfn() 3037 static kvm_pfn_t kvm_follow_pfn(struct kvm_follow_pfn *kfp) in kvm_follow_pfn() 3056 kvm_pfn_t __kvm_faultin_pfn(const struct kvm_memory_slot *slot, gfn_t gfn, in __kvm_faultin_pfn()
|
| /linux/include/linux/ |
| H A D | kvm_types.h | 74 typedef hfn_t kvm_pfn_t; typedef 94 kvm_pfn_t pfn;
|
| H A D | kvm_host.h | 107 static inline bool is_error_pfn(kvm_pfn_t pfn) in is_error_pfn() 116 static inline bool is_sigpending_pfn(kvm_pfn_t pfn) in is_sigpending_pfn() 126 static inline bool is_error_noslot_pfn(kvm_pfn_t pfn) in is_error_noslot_pfn() 132 static inline bool is_noslot_pfn(kvm_pfn_t pfn) in is_noslot_pfn() 295 kvm_pfn_t pfn; 296 kvm_pfn_t gfn; 1303 kvm_pfn_t __kvm_faultin_pfn(const struct kvm_memory_slot *slot, gfn_t gfn, 1307 static inline kvm_pfn_t kvm_faultin_pfn(struct kvm_vcpu *vcpu, gfn_t gfn, in kvm_faultin_pfn() 1914 static inline hpa_t pfn_to_hpa(kvm_pfn_t pfn) in pfn_to_hpa() 2542 gfn_t gfn, kvm_pfn_t *pfn, struct page **page, [all …]
|
| /linux/arch/x86/kvm/svm/ |
| H A D | svm.h | 897 int sev_gmem_prepare(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order); 898 void sev_gmem_invalidate(kvm_pfn_t start, kvm_pfn_t end); 899 int sev_gmem_max_mapping_level(struct kvm *kvm, kvm_pfn_t pfn, bool is_private); 923 static inline int sev_gmem_prepare(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order) in sev_gmem_prepare() 927 static inline void sev_gmem_invalidate(kvm_pfn_t start, kvm_pfn_t end) {} in sev_gmem_invalidate() 928 static inline int sev_gmem_max_mapping_level(struct kvm *kvm, kvm_pfn_t pfn, bool is_private) in sev_gmem_max_mapping_level()
|
| H A D | sev.c | 2280 static int sev_gmem_post_populate(struct kvm *kvm, gfn_t gfn, kvm_pfn_t pfn, in sev_gmem_post_populate() 3702 static int snp_rmptable_psmash(kvm_pfn_t pfn) in snp_rmptable_psmash() 3967 kvm_pfn_t pfn; in sev_snp_init_protected_guest_state() 4850 kvm_pfn_t pfn; in sev_handle_rmp_fault() 4936 static bool is_pfn_range_shared(kvm_pfn_t start, kvm_pfn_t end) in is_pfn_range_shared() 4938 kvm_pfn_t pfn = start; in is_pfn_range_shared() 4971 static bool is_large_rmp_possible(struct kvm *kvm, kvm_pfn_t pfn, int order) in is_large_rmp_possible() 4973 kvm_pfn_t pfn_aligned = ALIGN_DOWN(pfn, PTRS_PER_PMD); in is_large_rmp_possible() 4987 int sev_gmem_prepare(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order) in sev_gmem_prepare() 4990 kvm_pfn_t pfn_aligned; in sev_gmem_prepare() [all …]
|
| /linux/arch/x86/kvm/mmu/ |
| H A D | spte.c | 107 static bool __kvm_is_mmio_pfn(kvm_pfn_t pfn) in __kvm_is_mmio_pfn() 128 static bool kvm_is_mmio_pfn(kvm_pfn_t pfn, int *is_host_mmio) in kvm_is_mmio_pfn() 188 unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn, in make_spte()
|
| H A D | spte.h | 363 static inline kvm_pfn_t spte_to_pfn(u64 pte) in spte_to_pfn() 545 unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn,
|
| H A D | mmu_internal.h | 281 kvm_pfn_t pfn;
|
| /linux/arch/powerpc/kvm/ |
| H A D | e500_mmu_host.c | 167 kvm_pfn_t pfn; in kvmppc_map_magic() 169 pfn = (kvm_pfn_t)virt_to_phys((void *)shared_page) >> PAGE_SHIFT; in kvmppc_map_magic() 250 kvm_pfn_t pfn, unsigned int wimg, in kvmppc_e500_tlbe_setup() 307 kvm_pfn_t pfn = tlbe->pfn; in kvmppc_e500_setup_stlbe()
|
| H A D | book3s.c | 424 kvm_pfn_t kvmppc_gpa_to_pfn(struct kvm_vcpu *vcpu, gpa_t gpa, bool writing, in kvmppc_gpa_to_pfn() 437 kvm_pfn_t pfn; in kvmppc_gpa_to_pfn() 439 pfn = (kvm_pfn_t)virt_to_phys((void*)shared_page) >> PAGE_SHIFT; in kvmppc_gpa_to_pfn()
|
| H A D | e500.h | 43 kvm_pfn_t pfn; /* valid only for TLB0, except briefly */
|
| H A D | trace_pr.h | 32 TP_PROTO(int rflags, ulong hpteg, ulong va, kvm_pfn_t hpaddr,
|
| H A D | book3s_32_mmu_host.c | 132 kvm_pfn_t hpaddr; in kvmppc_mmu_map_page()
|
| H A D | book3s_64_mmu_host.c | 75 kvm_pfn_t hpaddr; in kvmppc_mmu_map_page()
|
| /linux/arch/riscv/kvm/ |
| H A D | mmu.c | 422 kvm_pfn_t *hfnp, gpa_t *gpa) in transparent_hugepage_adjust() 424 kvm_pfn_t hfn = *hfnp; in transparent_hugepage_adjust() 453 kvm_pfn_t hfn; in kvm_riscv_mmu_map()
|
| /linux/arch/s390/kvm/ |
| H A D | dat.h | 393 static inline union pte _pte(kvm_pfn_t pfn, bool writable, bool dirty, bool special) in _pte() 407 static inline union crste _crste_fc0(kvm_pfn_t pfn, int tt) in _crste_fc0() 431 static inline union crste _crste_fc1(kvm_pfn_t pfn, int tt, bool writable, bool dirty) in _crste_fc1() 490 kvm_pfn_t pfn; /* Host PFN */
|
| H A D | gmap.h | 104 kvm_pfn_t pfn, int level, bool wr);
|
| /linux/arch/arm64/kvm/ |
| H A D | mmu.c | 1403 unsigned long hva, kvm_pfn_t *pfnp, in transparent_hugepage_adjust() 1406 kvm_pfn_t pfn = *pfnp; in transparent_hugepage_adjust() 1470 static void sanitise_mte_tags(struct kvm *kvm, kvm_pfn_t pfn, in sanitise_mte_tags() 1579 kvm_pfn_t pfn; in gmem_abort() 1659 kvm_pfn_t pfn; in user_mem_abort()
|
| /linux/arch/x86/kvm/ |
| H A D | mmu.h | 258 int kvm_tdp_mmu_map_private_pfn(struct kvm_vcpu *vcpu, gfn_t gfn, kvm_pfn_t pfn);
|
| /linux/arch/x86/kvm/vmx/ |
| H A D | x86_ops.h | 156 int tdx_gmem_max_mapping_level(struct kvm *kvm, kvm_pfn_t pfn, bool is_private);
|
| /linux/arch/x86/include/asm/ |
| H A D | kvm_host.h | 1962 int (*gmem_prepare)(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order); 1963 void (*gmem_invalidate)(kvm_pfn_t start, kvm_pfn_t end); 1964 int (*gmem_max_mapping_level)(struct kvm *kvm, kvm_pfn_t pfn, bool is_private);
|
| /linux/arch/arm64/include/asm/ |
| H A D | kvm_pgtable.h | 151 static inline kvm_pfn_t kvm_pte_to_pfn(kvm_pte_t pte) in kvm_pte_to_pfn()
|