/arch/x86/kernel/ |
D | paravirt.c | 359 .mmu.flush_tlb_user = native_flush_tlb, 360 .mmu.flush_tlb_kernel = native_flush_tlb_global, 361 .mmu.flush_tlb_one_user = native_flush_tlb_one_user, 362 .mmu.flush_tlb_others = native_flush_tlb_others, 363 .mmu.tlb_remove_table = 366 .mmu.exit_mmap = paravirt_nop, 369 .mmu.read_cr2 = __PV_IS_CALLEE_SAVE(native_read_cr2), 370 .mmu.write_cr2 = native_write_cr2, 371 .mmu.read_cr3 = __native_read_cr3, 372 .mmu.write_cr3 = native_write_cr3, [all …]
|
D | paravirt_patch.c | 97 PATCH_CASE(mmu, read_cr2, xxl, insn_buff, len); in native_patch() 98 PATCH_CASE(mmu, read_cr3, xxl, insn_buff, len); in native_patch() 99 PATCH_CASE(mmu, write_cr3, xxl, insn_buff, len); in native_patch()
|
/arch/arc/mm/ |
D | tlb.c | 252 struct cpuinfo_arc_mmu *mmu = &cpuinfo_arc700[smp_processor_id()].mmu; in local_flush_tlb_all() local 255 int num_tlb = mmu->sets * mmu->ways; in local_flush_tlb_all() 728 struct cpuinfo_arc_mmu *mmu = &cpuinfo_arc700[smp_processor_id()].mmu; in read_decode_mmu_bcr() local 760 mmu->ver = (tmp >> 24); in read_decode_mmu_bcr() 763 if (mmu->ver <= 2) { in read_decode_mmu_bcr() 765 mmu->pg_sz_k = TO_KB(0x2000); in read_decode_mmu_bcr() 766 mmu->sets = 1 << mmu2->sets; in read_decode_mmu_bcr() 767 mmu->ways = 1 << mmu2->ways; in read_decode_mmu_bcr() 768 mmu->u_dtlb = mmu2->u_dtlb; in read_decode_mmu_bcr() 769 mmu->u_itlb = mmu2->u_itlb; in read_decode_mmu_bcr() [all …]
|
/arch/x86/include/asm/ |
D | paravirt.h | 52 PVOP_VCALL0(mmu.flush_tlb_user); in __flush_tlb() 57 PVOP_VCALL0(mmu.flush_tlb_kernel); in __flush_tlb_global() 62 PVOP_VCALL1(mmu.flush_tlb_one_user, addr); in __flush_tlb_one_user() 68 PVOP_VCALL2(mmu.flush_tlb_others, cpumask, info); in flush_tlb_others() 73 PVOP_VCALL2(mmu.tlb_remove_table, tlb, table); in paravirt_tlb_remove_table() 78 PVOP_VCALL1(mmu.exit_mmap, mm); in paravirt_arch_exit_mmap() 119 return PVOP_CALLEE0(unsigned long, mmu.read_cr2); in read_cr2() 124 PVOP_VCALL1(mmu.write_cr2, x); in write_cr2() 129 return PVOP_CALL0(unsigned long, mmu.read_cr3); in __read_cr3() 134 PVOP_VCALL1(mmu.write_cr3, x); in write_cr3() [all …]
|
/arch/powerpc/kvm/ |
D | book3s_32_mmu.c | 404 struct kvmppc_mmu *mmu = &vcpu->arch.mmu; in kvmppc_mmu_book3s_32_init() local 406 mmu->mtsrin = kvmppc_mmu_book3s_32_mtsrin; in kvmppc_mmu_book3s_32_init() 407 mmu->mfsrin = kvmppc_mmu_book3s_32_mfsrin; in kvmppc_mmu_book3s_32_init() 408 mmu->xlate = kvmppc_mmu_book3s_32_xlate; in kvmppc_mmu_book3s_32_init() 409 mmu->reset_msr = kvmppc_mmu_book3s_32_reset_msr; in kvmppc_mmu_book3s_32_init() 410 mmu->tlbie = kvmppc_mmu_book3s_32_tlbie; in kvmppc_mmu_book3s_32_init() 411 mmu->esid_to_vsid = kvmppc_mmu_book3s_32_esid_to_vsid; in kvmppc_mmu_book3s_32_init() 412 mmu->ea_to_vp = kvmppc_mmu_book3s_32_ea_to_vp; in kvmppc_mmu_book3s_32_init() 413 mmu->is_dcbz32 = kvmppc_mmu_book3s_32_is_dcbz32; in kvmppc_mmu_book3s_32_init() 415 mmu->slbmte = NULL; in kvmppc_mmu_book3s_32_init() [all …]
|
D | book3s_64_mmu.c | 668 struct kvmppc_mmu *mmu = &vcpu->arch.mmu; in kvmppc_mmu_book3s_64_init() local 670 mmu->mfsrin = NULL; in kvmppc_mmu_book3s_64_init() 671 mmu->mtsrin = kvmppc_mmu_book3s_64_mtsrin; in kvmppc_mmu_book3s_64_init() 672 mmu->slbmte = kvmppc_mmu_book3s_64_slbmte; in kvmppc_mmu_book3s_64_init() 673 mmu->slbmfee = kvmppc_mmu_book3s_64_slbmfee; in kvmppc_mmu_book3s_64_init() 674 mmu->slbmfev = kvmppc_mmu_book3s_64_slbmfev; in kvmppc_mmu_book3s_64_init() 675 mmu->slbfee = kvmppc_mmu_book3s_64_slbfee; in kvmppc_mmu_book3s_64_init() 676 mmu->slbie = kvmppc_mmu_book3s_64_slbie; in kvmppc_mmu_book3s_64_init() 677 mmu->slbia = kvmppc_mmu_book3s_64_slbia; in kvmppc_mmu_book3s_64_init() 678 mmu->xlate = kvmppc_mmu_book3s_64_xlate; in kvmppc_mmu_book3s_64_init() [all …]
|
D | book3s_emulate.c | 321 if (vcpu->arch.mmu.mfsrin) { in kvmppc_core_emulate_op_pr() 323 sr = vcpu->arch.mmu.mfsrin(vcpu, srnum); in kvmppc_core_emulate_op_pr() 333 if (vcpu->arch.mmu.mfsrin) { in kvmppc_core_emulate_op_pr() 335 sr = vcpu->arch.mmu.mfsrin(vcpu, srnum); in kvmppc_core_emulate_op_pr() 341 vcpu->arch.mmu.mtsrin(vcpu, in kvmppc_core_emulate_op_pr() 346 vcpu->arch.mmu.mtsrin(vcpu, in kvmppc_core_emulate_op_pr() 355 vcpu->arch.mmu.tlbie(vcpu, addr, large); in kvmppc_core_emulate_op_pr() 389 if (!vcpu->arch.mmu.slbmte) in kvmppc_core_emulate_op_pr() 392 vcpu->arch.mmu.slbmte(vcpu, in kvmppc_core_emulate_op_pr() 397 if (!vcpu->arch.mmu.slbie) in kvmppc_core_emulate_op_pr() [all …]
|
/arch/x86/kvm/ |
D | mmu.h | 78 if (likely(vcpu->arch.mmu->root_hpa != INVALID_PAGE)) in kvm_mmu_reload() 100 if (VALID_PAGE(vcpu->arch.mmu->root_hpa)) in kvm_mmu_load_cr3() 101 vcpu->arch.mmu->set_cr3(vcpu, vcpu->arch.mmu->root_hpa | in kvm_mmu_load_cr3() 156 static inline u8 permission_fault(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, in permission_fault() argument 179 bool fault = (mmu->permissions[index] >> pte_access) & 1; in permission_fault() 183 if (unlikely(mmu->pkru_mask)) { in permission_fault() 198 pkru_bits &= mmu->pkru_mask >> offset; in permission_fault()
|
D | paging_tmpl.h | 34 #define PT_HAVE_ACCESSED_DIRTY(mmu) true argument 53 #define PT_HAVE_ACCESSED_DIRTY(mmu) true argument 66 #define PT_HAVE_ACCESSED_DIRTY(mmu) ((mmu)->ept_ad) argument 105 static inline void FNAME(protect_clean_gpte)(struct kvm_mmu *mmu, unsigned *access, in FNAME() 111 if (!PT_HAVE_ACCESSED_DIRTY(mmu)) in FNAME() 132 static int FNAME(cmpxchg_gpte)(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, in FNAME() 175 if (is_rsvd_bits_set(vcpu->arch.mmu, gpte, PT_PAGE_TABLE_LEVEL)) in FNAME() 182 if (PT_HAVE_ACCESSED_DIRTY(vcpu->arch.mmu) && in FNAME() 218 struct kvm_mmu *mmu, in FNAME() 229 if (!PT_HAVE_ACCESSED_DIRTY(mmu)) in FNAME() [all …]
|
D | mmu.c | 372 return vcpu->arch.mmu == &vcpu->arch.guest_mmu; in kvm_vcpu_ad_need_write_protect() 2365 vcpu->arch.mmu->sync_page(vcpu, sp) == 0) { in __kvm_sync_page() 2570 role = vcpu->arch.mmu->mmu_role.base; in kvm_mmu_get_page() 2576 if (!vcpu->arch.mmu->direct_map in kvm_mmu_get_page() 2577 && vcpu->arch.mmu->root_level <= PT32_ROOT_LEVEL) { in kvm_mmu_get_page() 2651 iterator->level = vcpu->arch.mmu->shadow_root_level; in shadow_walk_init_using_root() 2654 vcpu->arch.mmu->root_level < PT64_ROOT_4LEVEL && in shadow_walk_init_using_root() 2655 !vcpu->arch.mmu->direct_map) in shadow_walk_init_using_root() 2663 BUG_ON(root != vcpu->arch.mmu->root_hpa); in shadow_walk_init_using_root() 2666 = vcpu->arch.mmu->pae_root[(addr >> 30) & 3]; in shadow_walk_init_using_root() [all …]
|
D | mmu_audit.c | 59 if (!VALID_PAGE(vcpu->arch.mmu->root_hpa)) in mmu_spte_walk() 62 if (vcpu->arch.mmu->root_level >= PT64_ROOT_4LEVEL) { in mmu_spte_walk() 63 hpa_t root = vcpu->arch.mmu->root_hpa; in mmu_spte_walk() 66 __mmu_spte_walk(vcpu, sp, fn, vcpu->arch.mmu->root_level); in mmu_spte_walk() 71 hpa_t root = vcpu->arch.mmu->pae_root[i]; in mmu_spte_walk() 122 "ent %llxn", vcpu->arch.mmu->root_level, pfn, in audit_mappings()
|
/arch/um/kernel/skas/ |
D | mmu.c | 142 struct mm_context *mmu = &mm->context; in destroy_context() local 150 if (mmu->id.u.pid < 2) { in destroy_context() 152 mmu->id.u.pid); in destroy_context() 155 os_kill_ptraced_process(mmu->id.u.pid, 1); in destroy_context() 157 free_page(mmu->id.stack); in destroy_context() 158 free_ldt(mmu); in destroy_context()
|
D | Makefile | 6 obj-y := clone.o mmu.o process.o syscall.o uaccess.o
|
/arch/m68k/kernel/ |
D | setup_mm.c | 401 const char *cpu, *mmu, *fpu; in show_cpuinfo() local 450 mmu = "68851"; in show_cpuinfo() 452 mmu = "68030"; in show_cpuinfo() 454 mmu = "68040"; in show_cpuinfo() 456 mmu = "68060"; in show_cpuinfo() 458 mmu = "Sun-3"; in show_cpuinfo() 460 mmu = "Apollo"; in show_cpuinfo() 462 mmu = "ColdFire"; in show_cpuinfo() 464 mmu = "unknown"; in show_cpuinfo() 474 cpu, mmu, fpu, in show_cpuinfo()
|
D | setup_no.c | 178 char *cpu, *mmu, *fpu; in show_cpuinfo() local 182 mmu = "none"; in show_cpuinfo() 192 cpu, mmu, fpu, in show_cpuinfo()
|
/arch/arm64/boot/dts/renesas/ |
D | r8a7795-es1.dtsi | 80 ipmmu_mp1: mmu@ec680000 { 88 ipmmu_sy: mmu@e7730000 { 96 /delete-node/ mmu@fd950000; 97 /delete-node/ mmu@fd960000; 98 /delete-node/ mmu@fd970000; 99 /delete-node/ mmu@febe0000; 100 /delete-node/ mmu@fe980000;
|
/arch/riscv/kernel/ |
D | cpu.c | 132 const char *compat, *isa, *mmu; in c_show() local 138 if (!of_property_read_string(node, "mmu-type", &mmu)) in c_show() 139 print_mmu(m, mmu); in c_show()
|
/arch/arm64/boot/dts/amd/ |
D | amd-seattle-xgbe-b.dtsi | 89 compatible = "arm,mmu-401"; 98 mmu-masters = <&xgmac0 105 compatible = "arm,mmu-401"; 114 mmu-masters = <&xgmac1
|
/arch/sh/mm/ |
D | Makefile | 19 mmu-y := nommu.o extable_32.o 20 mmu-$(CONFIG_MMU) := extable_$(BITS).o fault.o ioremap.o kmap.o \ 23 obj-y += $(mmu-y)
|
/arch/sparc/kernel/ |
D | wuf.S | 266 LEON_PI(lda [%g0] ASI_LEON_MMUREGS, %l5) ! read mmu-ctrl reg 267 SUN_PI_(lda [%g0] ASI_M_MMUREGS, %l5) ! read mmu-ctrl reg 282 LEON_PI(lda [%g0] ASI_LEON_MMUREGS, %twin_tmp1) ! load mmu-ctrl again 283 SUN_PI_(lda [%g0] ASI_M_MMUREGS, %twin_tmp1) ! load mmu-ctrl again
|
/arch/arm/lib/ |
D | Makefile | 19 mmu-y := clear_user.o copy_page.o getuser.o putuser.o \ 31 lib-$(CONFIG_MMU) += $(mmu-y)
|
/arch/x86/hyperv/ |
D | Makefile | 2 obj-y := hv_init.o mmu.o nested.o
|
/arch/powerpc/mm/book3s32/ |
D | Makefile | 9 obj-y += mmu.o hash_low.o mmu_context.o tlb.o
|
/arch/xtensa/mm/ |
D | Makefile | 7 obj-$(CONFIG_MMU) += cache.o fault.o ioremap.o mmu.o tlb.o
|
/arch/unicore32/mm/ |
D | Makefile | 6 obj-y := extable.o fault.o init.o pgd.o mmu.o
|