Lines Matching defs:gfn
215 static void mark_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, u64 gfn, in mark_mmio_spte()
239 static bool set_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn, in set_mmio_spte()
709 static void kvm_mmu_page_set_gfn(struct kvm_mmu_page *sp, int index, gfn_t gfn) in kvm_mmu_page_set_gfn()
727 static struct kvm_lpage_info *lpage_info_slot(gfn_t gfn, in lpage_info_slot()
738 gfn_t gfn, int count) in update_gfn_disallow_lpage_count()
750 void kvm_mmu_gfn_disallow_lpage(struct kvm_memory_slot *slot, gfn_t gfn) in kvm_mmu_gfn_disallow_lpage()
755 void kvm_mmu_gfn_allow_lpage(struct kvm_memory_slot *slot, gfn_t gfn) in kvm_mmu_gfn_allow_lpage()
764 gfn_t gfn; in account_shadowed() local
794 gfn_t gfn; in unaccount_shadowed() local
815 gfn_to_memslot_dirty_bitmap(struct kvm_vcpu *vcpu, gfn_t gfn, in gfn_to_memslot_dirty_bitmap()
941 static struct kvm_rmap_head *__gfn_to_rmap(gfn_t gfn, int level, in __gfn_to_rmap()
950 static struct kvm_rmap_head *gfn_to_rmap(struct kvm *kvm, gfn_t gfn, in gfn_to_rmap()
969 static int rmap_add(struct kvm_vcpu *vcpu, u64 *spte, gfn_t gfn) in rmap_add()
983 gfn_t gfn; in rmap_remove() local
1293 struct kvm_memory_slot *slot, u64 gfn) in kvm_mmu_slot_gfn_write_protect()
1311 static bool rmap_write_protect(struct kvm_vcpu *vcpu, u64 gfn) in rmap_write_protect()
1336 struct kvm_memory_slot *slot, gfn_t gfn, int level, in kvm_unmap_rmapp()
1343 struct kvm_memory_slot *slot, gfn_t gfn, int level, in kvm_set_pte_rmapp()
1392 gfn_t gfn; member
1458 gfn_t gfn, in kvm_handle_hva_range()
1503 gfn_t gfn, int level, in kvm_handle_hva()
1535 struct kvm_memory_slot *slot, gfn_t gfn, int level, in kvm_age_rmapp()
1550 struct kvm_memory_slot *slot, gfn_t gfn, in kvm_test_age_rmapp()
1564 static void rmap_recycle(struct kvm_vcpu *vcpu, u64 *spte, gfn_t gfn) in rmap_recycle()
1639 static unsigned kvm_page_table_hashfn(gfn_t gfn) in kvm_page_table_hashfn()
1897 static bool kvm_sync_pages(struct kvm_vcpu *vcpu, gfn_t gfn, in kvm_sync_pages()
2032 gfn_t gfn, in kvm_mmu_get_page()
2479 int kvm_mmu_unprotect_page(struct kvm *kvm, gfn_t gfn) in kvm_mmu_unprotect_page()
2510 bool mmu_need_write_protect(struct kvm_vcpu *vcpu, gfn_t gfn, in mmu_need_write_protect()
2573 gfn_t gfn, kvm_pfn_t pfn, bool speculative, in set_spte()
2600 gfn_t gfn, kvm_pfn_t pfn, bool speculative, in mmu_set_spte()
2673 static kvm_pfn_t pte_prefetch_gfn_to_pfn(struct kvm_vcpu *vcpu, gfn_t gfn, in pte_prefetch_gfn_to_pfn()
2693 gfn_t gfn; in direct_pte_prefetch_many() local
2756 static int host_pfn_mapping_level(struct kvm_vcpu *vcpu, gfn_t gfn, in host_pfn_mapping_level()
2783 int kvm_mmu_hugepage_adjust(struct kvm_vcpu *vcpu, gfn_t gfn, in kvm_mmu_hugepage_adjust()
2839 void disallowed_hugepage_adjust(u64 spte, gfn_t gfn, int cur_level, in disallowed_hugepage_adjust()
2872 gfn_t gfn = gpa >> PAGE_SHIFT; in __direct_map() local
2923 static int kvm_handle_bad_page(struct kvm_vcpu *vcpu, gfn_t gfn, kvm_pfn_t pfn) in kvm_handle_bad_page()
2941 static bool handle_abnormal_pfn(struct kvm_vcpu *vcpu, gva_t gva, gfn_t gfn, in handle_abnormal_pfn()
2999 gfn_t gfn; in fast_pf_fix_direct_spte() local
3226 static hpa_t mmu_alloc_root(struct kvm_vcpu *vcpu, gfn_t gfn, gva_t gva, in mmu_alloc_root()
3576 gfn_t gfn = get_mmio_spte_gfn(spte); in handle_mmio_page_fault() local
3598 u32 error_code, gfn_t gfn) in page_fault_handle_page_track()
3632 gfn_t gfn) in kvm_arch_setup_async_pf()
3645 static bool try_async_pf(struct kvm_vcpu *vcpu, bool prefault, gfn_t gfn, in try_async_pf()
3684 gfn_t gfn = gpa >> PAGE_SHIFT; in direct_page_fault() local
3919 static bool sync_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn, in sync_mmio_spte()
4962 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_mmu_pte_write() local