• Home
  • Raw
  • Download

Lines Matching defs:spte

240 #define for_each_shadow_entry_lockless(_vcpu, _addr, _walker, spte)	\  argument
354 static bool is_mmio_spte(u64 spte) in is_mmio_spte()
375 static inline bool spte_ad_enabled(u64 spte) in spte_ad_enabled()
381 static inline bool spte_ad_need_write_protect(u64 spte) in spte_ad_need_write_protect()
392 static inline u64 spte_shadow_accessed_mask(u64 spte) in spte_shadow_accessed_mask()
398 static inline u64 spte_shadow_dirty_mask(u64 spte) in spte_shadow_dirty_mask()
404 static inline bool is_access_track_spte(u64 spte) in is_access_track_spte()
458 static u64 get_mmio_spte_generation(u64 spte) in get_mmio_spte_generation()
484 static gfn_t get_mmio_spte_gfn(u64 spte) in get_mmio_spte_gfn()
494 static unsigned get_mmio_spte_access(u64 spte) in get_mmio_spte_access()
510 static bool check_mmio_spte(struct kvm_vcpu *vcpu, u64 spte) in check_mmio_spte()
639 static bool is_executable_pte(u64 spte) in is_executable_pte()
657 static void __set_spte(u64 *sptep, u64 spte) in __set_spte()
662 static void __update_clear_spte_fast(u64 *sptep, u64 spte) in __update_clear_spte_fast()
667 static u64 __update_clear_spte_slow(u64 *sptep, u64 spte) in __update_clear_spte_slow()
682 u64 spte; member
685 static void count_spte_clear(u64 *sptep, u64 spte) in count_spte_clear()
697 static void __set_spte(u64 *sptep, u64 spte) in __set_spte()
716 static void __update_clear_spte_fast(u64 *sptep, u64 spte) in __update_clear_spte_fast()
735 static u64 __update_clear_spte_slow(u64 *sptep, u64 spte) in __update_clear_spte_slow()
772 union split_spte spte, *orig = (union split_spte *)sptep; in __get_spte_lockless() local
793 static bool spte_can_locklessly_be_made_writable(u64 spte) in spte_can_locklessly_be_made_writable()
799 static bool spte_has_volatile_bits(u64 spte) in spte_has_volatile_bits()
823 static bool is_accessed_spte(u64 spte) in is_accessed_spte()
831 static bool is_dirty_spte(u64 spte) in is_dirty_spte()
973 static u64 mark_spte_for_access_track(u64 spte) in mark_spte_for_access_track()
1002 static u64 restore_acc_track_spte(u64 spte) in restore_acc_track_spte()
1022 u64 spte = mmu_spte_get_lockless(sptep); in mmu_spte_age() local
1381 static int pte_list_add(struct kvm_vcpu *vcpu, u64 *spte, in pte_list_add()
1438 static void __pte_list_remove(u64 *spte, struct kvm_rmap_head *rmap_head) in __pte_list_remove()
1508 static int rmap_add(struct kvm_vcpu *vcpu, u64 *spte, gfn_t gfn) in rmap_add()
1519 static void rmap_remove(struct kvm *kvm, u64 *spte) in rmap_remove()
1652 u64 spte = *sptep; in spte_write_protect() local
1683 u64 spte = *sptep; in spte_clear_dirty() local
1725 u64 spte = *sptep; in spte_set_dirty() local
2100 static void rmap_recycle(struct kvm_vcpu *vcpu, u64 *spte, gfn_t gfn) in rmap_recycle()
2222 static void mark_unsync(u64 *spte) in mark_unsync()
2548 static void clear_sp_write_flooding_count(u64 *spte) in clear_sp_write_flooding_count()
2692 u64 spte) in __shadow_walk_next()
2711 u64 spte; in link_shadow_page() local
2754 u64 *spte) in mmu_page_zap_pte()
3059 u64 spte = 0; in set_spte() local
3267 u64 *spte, *start = NULL; in __direct_pte_prefetch() local
3311 u64 spte = *it.sptep; in disallowed_hugepage_adjust() local
3524 static bool is_access_allowed(u32 fault_err_code, u64 spte) in is_access_allowed()
3547 u64 spte = 0ull; in fast_page_fault() local
4026 static bool is_shadow_zero_bits_set(struct kvm_mmu *mmu, u64 spte, int level) in is_shadow_zero_bits_set()
4051 u64 sptes[PT64_ROOT_MAX_LEVEL], spte = 0ull; in walk_shadow_page_get_mmio_spte() local
4094 u64 spte; in handle_mmio_page_fault() local
4149 u64 spte; in shadow_page_table_clear_flood() local
5398 u64 *spte; in get_written_sptes() local
5433 u64 entry, gentry, *spte; in kvm_mmu_pte_write() local