Lines Matching defs:gfn
287 gfn_t gfn = kvm_mmu_page_get_gfn(sp, spte_index(sptep)); in kvm_flush_remote_tlbs_sptep() local
292 static void mark_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, u64 gfn, in mark_mmio_spte()
755 gfn_t gfn, unsigned int access) in kvm_mmu_page_set_translation()
776 gfn_t gfn = kvm_mmu_page_get_gfn(sp, index); in kvm_mmu_page_set_access() local
785 static struct kvm_lpage_info *lpage_info_slot(gfn_t gfn, in lpage_info_slot()
803 gfn_t gfn, int count) in update_gfn_disallow_lpage_count()
817 void kvm_mmu_gfn_disallow_lpage(const struct kvm_memory_slot *slot, gfn_t gfn) in kvm_mmu_gfn_disallow_lpage()
822 void kvm_mmu_gfn_allow_lpage(const struct kvm_memory_slot *slot, gfn_t gfn) in kvm_mmu_gfn_allow_lpage()
831 gfn_t gfn; in account_shadowed() local
879 gfn_t gfn; in unaccount_shadowed() local
908 gfn_t gfn, in gfn_to_memslot_dirty_bitmap()
1086 static struct kvm_rmap_head *gfn_to_rmap(gfn_t gfn, int level, in gfn_to_rmap()
1100 gfn_t gfn; in rmap_remove() local
1415 struct kvm_memory_slot *slot, u64 gfn, in kvm_mmu_slot_gfn_write_protect()
1436 static bool kvm_vcpu_write_protect_gfn(struct kvm_vcpu *vcpu, u64 gfn) in kvm_vcpu_write_protect_gfn()
1451 struct kvm_memory_slot *slot, gfn_t gfn, int level, in kvm_zap_rmap()
1458 struct kvm_memory_slot *slot, gfn_t gfn, int level, in kvm_set_pte_rmap()
1503 gfn_t gfn; member
1613 struct kvm_memory_slot *slot, gfn_t gfn, int level, in kvm_age_rmap()
1627 struct kvm_memory_slot *slot, gfn_t gfn, in kvm_test_age_rmap()
1644 u64 *spte, gfn_t gfn, unsigned int access) in __rmap_add()
1666 u64 *spte, gfn_t gfn, unsigned int access) in rmap_add()
1749 static unsigned kvm_page_table_hashfn(gfn_t gfn) in kvm_page_table_hashfn()
2153 gfn_t gfn, in kvm_mmu_find_shadow_page()
2238 gfn_t gfn, in kvm_mmu_alloc_shadow_page()
2275 gfn_t gfn, in __kvm_mmu_get_shadow_page()
2295 gfn_t gfn, in kvm_mmu_get_shadow_page()
2354 u64 *sptep, gfn_t gfn, in kvm_mmu_get_child_sp()
2757 int kvm_mmu_unprotect_page(struct kvm *kvm, gfn_t gfn) in kvm_mmu_unprotect_page()
2806 gfn_t gfn, bool can_unsync, bool prefetch) in mmu_try_to_unsync_pages()
2907 u64 *sptep, unsigned int pte_access, gfn_t gfn, in mmu_set_spte()
2985 gfn_t gfn; in direct_pte_prefetch_many() local
3082 static int host_pfn_mapping_level(struct kvm *kvm, gfn_t gfn, in host_pfn_mapping_level()
3148 gfn_t gfn, int max_level, bool is_private) in __kvm_mmu_max_mapping_level()
3171 const struct kvm_memory_slot *slot, gfn_t gfn, in kvm_mmu_max_mapping_level()
3281 static void kvm_send_hwpoison_signal(struct kvm_memory_slot *slot, gfn_t gfn) in kvm_send_hwpoison_signal()
3670 static hpa_t mmu_alloc_root(struct kvm_vcpu *vcpu, gfn_t gfn, int quadrant, in mmu_alloc_root()
4187 gfn_t gfn = get_mmio_spte_gfn(spte); in handle_mmio_page_fault() local
4250 gfn_t gfn) in kvm_arch_setup_async_pf()
4799 static bool sync_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn, in sync_mmio_spte()
5784 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_mmu_track_write() local
6483 gfn_t gfn; in shadow_mmu_get_sp_for_split() local
6514 gfn_t gfn; in shadow_mmu_split_huge_page() local
6556 gfn_t gfn; in shadow_mmu_try_split_huge_page() local
7322 static bool hugepage_test_mixed(struct kvm_memory_slot *slot, gfn_t gfn, in hugepage_test_mixed()
7328 static void hugepage_clear_mixed(struct kvm_memory_slot *slot, gfn_t gfn, in hugepage_clear_mixed()
7334 static void hugepage_set_mixed(struct kvm_memory_slot *slot, gfn_t gfn, in hugepage_set_mixed()
7341 gfn_t gfn, int level, unsigned long attrs) in hugepage_has_attrs()
7382 gfn_t gfn = gfn_round_for_level(range->start, level); in kvm_arch_post_set_memory_attributes() local
7440 gfn_t gfn; in kvm_mmu_init_memslot_memory_attributes() local