Lines Matching defs:gfn
286 gfn_t gfn = kvm_mmu_page_get_gfn(sp, spte_index(sptep)); in kvm_flush_remote_tlbs_sptep() local
291 static void mark_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, u64 gfn, in mark_mmio_spte()
676 gfn_t gfn, unsigned int access) in kvm_mmu_page_set_translation()
697 gfn_t gfn = kvm_mmu_page_get_gfn(sp, index); in kvm_mmu_page_set_access() local
706 static struct kvm_lpage_info *lpage_info_slot(gfn_t gfn, in lpage_info_slot()
724 gfn_t gfn, int count) in update_gfn_disallow_lpage_count()
738 void kvm_mmu_gfn_disallow_lpage(const struct kvm_memory_slot *slot, gfn_t gfn) in kvm_mmu_gfn_disallow_lpage()
743 void kvm_mmu_gfn_allow_lpage(const struct kvm_memory_slot *slot, gfn_t gfn) in kvm_mmu_gfn_allow_lpage()
752 gfn_t gfn; in account_shadowed() local
809 gfn_t gfn; in unaccount_shadowed() local
838 gfn_t gfn, in gfn_to_memslot_dirty_bitmap()
1170 static struct kvm_rmap_head *gfn_to_rmap(gfn_t gfn, int level, in gfn_to_rmap()
1184 gfn_t gfn; in rmap_remove() local
1471 struct kvm_memory_slot *slot, u64 gfn, in kvm_mmu_slot_gfn_write_protect()
1492 static bool kvm_vcpu_write_protect_gfn(struct kvm_vcpu *vcpu, u64 gfn) in kvm_vcpu_write_protect_gfn()
1515 gfn_t gfn; member
1677 u64 *spte, gfn_t gfn, unsigned int access) in __rmap_add()
1699 u64 *spte, gfn_t gfn, unsigned int access) in rmap_add()
1715 gfn_t gfn; in kvm_rmap_age_gfn_range() local
1825 static unsigned kvm_page_table_hashfn(gfn_t gfn) in kvm_page_table_hashfn()
2231 gfn_t gfn, in kvm_mmu_find_shadow_page()
2316 gfn_t gfn, in kvm_mmu_alloc_shadow_page()
2353 gfn_t gfn, in __kvm_mmu_get_shadow_page()
2373 gfn_t gfn, in kvm_mmu_get_shadow_page()
2432 u64 *sptep, gfn_t gfn, in kvm_mmu_get_child_sp()
2898 gfn_t gfn, bool synchronizing, bool prefetch) in mmu_try_to_unsync_pages()
2999 u64 *sptep, unsigned int pte_access, gfn_t gfn, in mmu_set_spte()
3070 static bool kvm_mmu_prefetch_sptes(struct kvm_vcpu *vcpu, gfn_t gfn, u64 *sptep, in kvm_mmu_prefetch_sptes()
3111 gfn_t gfn = kvm_mmu_page_get_gfn(sp, spte_index(start)); in direct_pte_prefetch_many() local
3195 static int host_pfn_mapping_level(struct kvm *kvm, gfn_t gfn, in host_pfn_mapping_level()
3261 gfn_t gfn, int max_level, bool is_private) in __kvm_mmu_max_mapping_level()
3284 const struct kvm_memory_slot *slot, gfn_t gfn) in kvm_mmu_max_mapping_level()
3393 static void kvm_send_hwpoison_signal(struct kvm_memory_slot *slot, gfn_t gfn) in kvm_send_hwpoison_signal()
3813 static hpa_t mmu_alloc_root(struct kvm_vcpu *vcpu, gfn_t gfn, int quadrant, in mmu_alloc_root()
4343 gfn_t gfn = get_mmio_spte_gfn(spte); in handle_mmio_page_fault() local
5090 static bool sync_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn, in sync_mmio_spte()
6081 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_mmu_track_write() local
6855 gfn_t gfn; in shadow_mmu_get_sp_for_split() local
6886 gfn_t gfn; in shadow_mmu_split_huge_page() local
6928 gfn_t gfn; in shadow_mmu_try_split_huge_page() local
7218 gfn_t gfn = slot->base_gfn + i; in kvm_mmu_zap_memslot_pages_and_flush() local
7673 static bool hugepage_test_mixed(struct kvm_memory_slot *slot, gfn_t gfn, in hugepage_test_mixed()
7679 static void hugepage_clear_mixed(struct kvm_memory_slot *slot, gfn_t gfn, in hugepage_clear_mixed()
7685 static void hugepage_set_mixed(struct kvm_memory_slot *slot, gfn_t gfn, in hugepage_set_mixed()
7755 gfn_t gfn, int level, unsigned long attrs) in hugepage_has_attrs()
7796 gfn_t gfn = gfn_round_for_level(range->start, level); in kvm_arch_post_set_memory_attributes() local
7855 gfn_t gfn; in kvm_mmu_init_memslot_memory_attributes() local