/kernel/linux/linux-5.10/arch/x86/kvm/mmu/ |
D | mmu.c | 768 if (sp->role.level > PG_LEVEL_4K) in account_shadowed() 796 if (sp->role.level > PG_LEVEL_4K) in unaccount_shadowed() 943 return &slot->arch.rmap[level - PG_LEVEL_4K][idx]; in __gfn_to_rmap() 1074 WARN_ON(sptep_to_sp(sptep)->role.level == PG_LEVEL_4K); in __drop_large_spte() 1229 PG_LEVEL_4K, slot); in kvm_mmu_write_protect_pt_masked() 1258 PG_LEVEL_4K, slot); in kvm_mmu_clear_dirty_pt_masked() 1295 for (i = PG_LEVEL_4K; i <= KVM_MAX_HUGEPAGE_LEVEL; ++i) { in kvm_mmu_slot_gfn_write_protect() 1482 for_each_slot_rmap_range(memslot, PG_LEVEL_4K, in kvm_handle_hva_range() 1896 WARN_ON(s->role.level != PG_LEVEL_4K); in kvm_sync_pages() 1925 if (level == PG_LEVEL_4K) in mmu_pages_next() [all …]
|
D | paging_tmpl.h | 79 #define gpte_to_gfn(pte) gpte_to_gfn_lvl((pte), PG_LEVEL_4K) 198 if (FNAME(is_rsvd_bits_set)(vcpu->arch.mmu, gpte, PG_LEVEL_4K)) in FNAME() 435 if (PTTYPE == 32 && walker->level > PG_LEVEL_4K && is_cpuid_PSE36()) in FNAME() 552 mmu_set_spte(vcpu, spte, pte_access, false, PG_LEVEL_4K, gfn, pfn, in FNAME() 575 if (level == PG_LEVEL_4K) { in FNAME() 600 if (sp->role.level > PG_LEVEL_4K) in FNAME() 836 max_level = PG_LEVEL_4K; in FNAME() 893 WARN_ON(sp->role.level != PG_LEVEL_4K); in FNAME() 1082 pte_access, PG_LEVEL_4K, in FNAME()
|
D | mmu_audit.c | 103 if (level != PG_LEVEL_4K) { in audit_mappings() 179 if (sp->role.level != PG_LEVEL_4K) in check_mappings_rmap() 203 rmap_head = __gfn_to_rmap(sp->gfn, PG_LEVEL_4K, slot); in audit_write_protection()
|
D | page_track.c | 64 index = gfn_to_index(gfn, slot->base_gfn, PG_LEVEL_4K); in update_gfn_track() 154 index = gfn_to_index(gfn, slot->base_gfn, PG_LEVEL_4K); in kvm_page_track_is_active()
|
D | tdp_iter.h | 56 for_each_tdp_pte_min_level(iter, root, root_level, PG_LEVEL_4K, start, end)
|
D | spte.c | 107 if (level > PG_LEVEL_4K && (pte_access & ACC_EXEC_MASK) && in make_spte() 120 if (level > PG_LEVEL_4K) in make_spte()
|
D | tdp_mmu.c | 221 if (level > PG_LEVEL_4K) in handle_changed_spte_dirty_log() 259 WARN_ON(level < PG_LEVEL_4K); in __handle_changed_spte() 798 if (iter.level != PG_LEVEL_4K) in set_tdp_spte() 966 if (iter.level > PG_LEVEL_4K || in clear_dirty_pt_masked()
|
D | spte.h | 194 if (level == PG_LEVEL_4K) in is_last_spte()
|
/kernel/linux/linux-5.10/arch/x86/xen/ |
D | p2m.c | 250 BUG_ON(!ptep || level != PG_LEVEL_4K); in xen_build_mfn_list_list() 448 BUG_ON(!ptep || level != PG_LEVEL_4K); in get_phys_to_machine() 542 BUG_ON(!ptep || level != PG_LEVEL_4K); in xen_alloc_p2m_entry() 667 BUG_ON(!ptep || level != PG_LEVEL_4K); in __set_phys_to_machine()
|
/kernel/linux/linux-5.10/arch/x86/mm/ |
D | mem_encrypt.c | 209 case PG_LEVEL_4K: in __set_clr_pte_enc() 275 if (level == PG_LEVEL_4K) { in early_set_memory_enc_dec()
|
D | pti.c | 283 if (!pte || WARN_ON(level != PG_LEVEL_4K) || pte_none(*pte)) in pti_setup_vsyscall()
|
D | kmmio.c | 168 case PG_LEVEL_4K: in clear_page_presence()
|
D | init_32.c | 380 update_page_count(PG_LEVEL_4K, pages_4k); in kernel_physical_mapping_init()
|
D | init_64.c | 491 update_page_count(PG_LEVEL_4K, pages); in phys_pte_init() 1027 update_page_count(PG_LEVEL_4K, -pages); in remove_pte_table()
|
/kernel/linux/linux-5.10/arch/x86/mm/pat/ |
D | cpa-test.c | 206 if (level != PG_LEVEL_4K) { in pageattr_test()
|
D | set_memory.c | 100 direct_pages_count[PG_LEVEL_4K] << 2); in arch_report_meminfo() 485 if (lookup_address(start, &level) && (level != PG_LEVEL_4K)) in protect_kernel_text_ro() 611 *level = PG_LEVEL_4K; in lookup_address_in_pgd() 1526 if (level == PG_LEVEL_4K) { in __change_page_attr()
|
/kernel/linux/linux-5.10/arch/x86/include/asm/ |
D | pgtable_types.h | 544 PG_LEVEL_4K, enumerator
|
D | kvm_host.h | 116 #define KVM_NR_PAGE_SIZES (KVM_MAX_HUGEPAGE_LEVEL - PG_LEVEL_4K + 1)
|
/kernel/linux/linux-5.10/arch/x86/kvm/vmx/ |
D | vmx.c | 7934 ept_lpage_level = PG_LEVEL_4K; in hardware_setup()
|
/kernel/linux/linux-5.10/arch/x86/kvm/ |
D | x86.c | 10783 PG_LEVEL_2M : PG_LEVEL_4K; in kvm_mmu_slot_apply_flags()
|