Lines Matching defs:spte
172 #define for_each_shadow_entry_lockless(_vcpu, _addr, _walker, spte) \ argument
205 static bool is_mmio_spte(u64 spte) in is_mmio_spte()
210 static gfn_t get_mmio_spte_gfn(u64 spte) in get_mmio_spte_gfn()
215 static unsigned get_mmio_spte_access(u64 spte) in get_mmio_spte_access()
298 static void __set_spte(u64 *sptep, u64 spte) in __set_spte()
303 static void __update_clear_spte_fast(u64 *sptep, u64 spte) in __update_clear_spte_fast()
308 static u64 __update_clear_spte_slow(u64 *sptep, u64 spte) in __update_clear_spte_slow()
318 static bool __check_direct_spte_mmio_pf(u64 spte) in __check_direct_spte_mmio_pf()
329 u64 spte; member
332 static void count_spte_clear(u64 *sptep, u64 spte) in count_spte_clear()
344 static void __set_spte(u64 *sptep, u64 spte) in __set_spte()
363 static void __update_clear_spte_fast(u64 *sptep, u64 spte) in __update_clear_spte_fast()
382 static u64 __update_clear_spte_slow(u64 *sptep, u64 spte) in __update_clear_spte_slow()
408 union split_spte spte, *orig = (union split_spte *)sptep; in __get_spte_lockless() local
428 static bool __check_direct_spte_mmio_pf(u64 spte) in __check_direct_spte_mmio_pf()
446 static bool spte_has_volatile_bits(u64 spte) in spte_has_volatile_bits()
811 static int pte_list_add(struct kvm_vcpu *vcpu, u64 *spte, in pte_list_add()
845 static u64 *pte_list_next(unsigned long *pte_list, u64 *spte) in pte_list_next()
893 static void pte_list_remove(u64 *spte, unsigned long *pte_list) in pte_list_remove()
980 static int rmap_add(struct kvm_vcpu *vcpu, u64 *spte, gfn_t gfn) in rmap_add()
991 static u64 *rmap_next(struct kvm *kvm, unsigned long *rmapp, u64 *spte) in rmap_next()
996 static void rmap_remove(struct kvm *kvm, u64 *spte) in rmap_remove()
1018 u64 *spte; in kvm_mmu_rmap_write_protect() local
1066 u64 *spte; in kvm_unmap_rmapp() local
1082 u64 *spte, new_spte; in kvm_set_pte_rmapp() local
1166 u64 *spte; in kvm_age_rmapp() local
1197 u64 *spte; in kvm_test_age_rmapp() local
1225 static void rmap_recycle(struct kvm_vcpu *vcpu, u64 *spte, gfn_t gfn) in rmap_recycle()
1353 static void mark_unsync(u64 *spte) in mark_unsync()
1378 struct kvm_mmu_page *sp, u64 *spte, in nonpaging_update_pte()
1668 static void clear_sp_write_flooding_count(u64 *spte) in clear_sp_write_flooding_count()
1776 u64 spte) in __shadow_walk_next()
1794 u64 spte; in link_shadow_page() local
1833 u64 *spte) in mmu_page_zap_pte()
2202 u64 spte, entry = *sptep; in set_spte() local
2410 u64 *spte, *start = NULL; in __direct_pte_prefetch() local
2910 static bool check_direct_spte_mmio_pf(u64 spte) in check_direct_spte_mmio_pf()
2918 u64 spte = 0ull; in walk_shadow_page_get_mmio_spte() local
2936 u64 spte; in handle_mmio_page_fault_common() local
3481 struct kvm_mmu_page *sp, u64 *spte, in mmu_pte_write_new_pte()
3558 static bool detect_write_flooding(struct kvm_mmu_page *sp, u64 *spte) in detect_write_flooding()
3601 u64 *spte; in get_written_sptes() local
3637 u64 entry, gentry, *spte; in kvm_mmu_pte_write() local
4005 u64 spte; in kvm_mmu_get_spte_hierarchy() local