Home
last modified time | relevance | path

Searched refs:mmu (Results 1 – 25 of 116) sorted by relevance

12345

/arch/x86/kernel/
Dparavirt.c359 .mmu.flush_tlb_user = native_flush_tlb,
360 .mmu.flush_tlb_kernel = native_flush_tlb_global,
361 .mmu.flush_tlb_one_user = native_flush_tlb_one_user,
362 .mmu.flush_tlb_others = native_flush_tlb_others,
363 .mmu.tlb_remove_table =
366 .mmu.exit_mmap = paravirt_nop,
369 .mmu.read_cr2 = __PV_IS_CALLEE_SAVE(native_read_cr2),
370 .mmu.write_cr2 = native_write_cr2,
371 .mmu.read_cr3 = __native_read_cr3,
372 .mmu.write_cr3 = native_write_cr3,
[all …]
Dparavirt_patch.c97 PATCH_CASE(mmu, read_cr2, xxl, insn_buff, len); in native_patch()
98 PATCH_CASE(mmu, read_cr3, xxl, insn_buff, len); in native_patch()
99 PATCH_CASE(mmu, write_cr3, xxl, insn_buff, len); in native_patch()
/arch/arc/mm/
Dtlb.c252 struct cpuinfo_arc_mmu *mmu = &cpuinfo_arc700[smp_processor_id()].mmu; in local_flush_tlb_all() local
255 int num_tlb = mmu->sets * mmu->ways; in local_flush_tlb_all()
728 struct cpuinfo_arc_mmu *mmu = &cpuinfo_arc700[smp_processor_id()].mmu; in read_decode_mmu_bcr() local
760 mmu->ver = (tmp >> 24); in read_decode_mmu_bcr()
763 if (mmu->ver <= 2) { in read_decode_mmu_bcr()
765 mmu->pg_sz_k = TO_KB(0x2000); in read_decode_mmu_bcr()
766 mmu->sets = 1 << mmu2->sets; in read_decode_mmu_bcr()
767 mmu->ways = 1 << mmu2->ways; in read_decode_mmu_bcr()
768 mmu->u_dtlb = mmu2->u_dtlb; in read_decode_mmu_bcr()
769 mmu->u_itlb = mmu2->u_itlb; in read_decode_mmu_bcr()
[all …]
/arch/x86/include/asm/
Dparavirt.h52 PVOP_VCALL0(mmu.flush_tlb_user); in __flush_tlb()
57 PVOP_VCALL0(mmu.flush_tlb_kernel); in __flush_tlb_global()
62 PVOP_VCALL1(mmu.flush_tlb_one_user, addr); in __flush_tlb_one_user()
68 PVOP_VCALL2(mmu.flush_tlb_others, cpumask, info); in flush_tlb_others()
73 PVOP_VCALL2(mmu.tlb_remove_table, tlb, table); in paravirt_tlb_remove_table()
78 PVOP_VCALL1(mmu.exit_mmap, mm); in paravirt_arch_exit_mmap()
119 return PVOP_CALLEE0(unsigned long, mmu.read_cr2); in read_cr2()
124 PVOP_VCALL1(mmu.write_cr2, x); in write_cr2()
129 return PVOP_CALL0(unsigned long, mmu.read_cr3); in __read_cr3()
134 PVOP_VCALL1(mmu.write_cr3, x); in write_cr3()
[all …]
/arch/powerpc/kvm/
Dbook3s_32_mmu.c404 struct kvmppc_mmu *mmu = &vcpu->arch.mmu; in kvmppc_mmu_book3s_32_init() local
406 mmu->mtsrin = kvmppc_mmu_book3s_32_mtsrin; in kvmppc_mmu_book3s_32_init()
407 mmu->mfsrin = kvmppc_mmu_book3s_32_mfsrin; in kvmppc_mmu_book3s_32_init()
408 mmu->xlate = kvmppc_mmu_book3s_32_xlate; in kvmppc_mmu_book3s_32_init()
409 mmu->reset_msr = kvmppc_mmu_book3s_32_reset_msr; in kvmppc_mmu_book3s_32_init()
410 mmu->tlbie = kvmppc_mmu_book3s_32_tlbie; in kvmppc_mmu_book3s_32_init()
411 mmu->esid_to_vsid = kvmppc_mmu_book3s_32_esid_to_vsid; in kvmppc_mmu_book3s_32_init()
412 mmu->ea_to_vp = kvmppc_mmu_book3s_32_ea_to_vp; in kvmppc_mmu_book3s_32_init()
413 mmu->is_dcbz32 = kvmppc_mmu_book3s_32_is_dcbz32; in kvmppc_mmu_book3s_32_init()
415 mmu->slbmte = NULL; in kvmppc_mmu_book3s_32_init()
[all …]
Dbook3s_64_mmu.c668 struct kvmppc_mmu *mmu = &vcpu->arch.mmu; in kvmppc_mmu_book3s_64_init() local
670 mmu->mfsrin = NULL; in kvmppc_mmu_book3s_64_init()
671 mmu->mtsrin = kvmppc_mmu_book3s_64_mtsrin; in kvmppc_mmu_book3s_64_init()
672 mmu->slbmte = kvmppc_mmu_book3s_64_slbmte; in kvmppc_mmu_book3s_64_init()
673 mmu->slbmfee = kvmppc_mmu_book3s_64_slbmfee; in kvmppc_mmu_book3s_64_init()
674 mmu->slbmfev = kvmppc_mmu_book3s_64_slbmfev; in kvmppc_mmu_book3s_64_init()
675 mmu->slbfee = kvmppc_mmu_book3s_64_slbfee; in kvmppc_mmu_book3s_64_init()
676 mmu->slbie = kvmppc_mmu_book3s_64_slbie; in kvmppc_mmu_book3s_64_init()
677 mmu->slbia = kvmppc_mmu_book3s_64_slbia; in kvmppc_mmu_book3s_64_init()
678 mmu->xlate = kvmppc_mmu_book3s_64_xlate; in kvmppc_mmu_book3s_64_init()
[all …]
Dbook3s_emulate.c321 if (vcpu->arch.mmu.mfsrin) { in kvmppc_core_emulate_op_pr()
323 sr = vcpu->arch.mmu.mfsrin(vcpu, srnum); in kvmppc_core_emulate_op_pr()
333 if (vcpu->arch.mmu.mfsrin) { in kvmppc_core_emulate_op_pr()
335 sr = vcpu->arch.mmu.mfsrin(vcpu, srnum); in kvmppc_core_emulate_op_pr()
341 vcpu->arch.mmu.mtsrin(vcpu, in kvmppc_core_emulate_op_pr()
346 vcpu->arch.mmu.mtsrin(vcpu, in kvmppc_core_emulate_op_pr()
355 vcpu->arch.mmu.tlbie(vcpu, addr, large); in kvmppc_core_emulate_op_pr()
389 if (!vcpu->arch.mmu.slbmte) in kvmppc_core_emulate_op_pr()
392 vcpu->arch.mmu.slbmte(vcpu, in kvmppc_core_emulate_op_pr()
397 if (!vcpu->arch.mmu.slbie) in kvmppc_core_emulate_op_pr()
[all …]
/arch/x86/kvm/
Dmmu.h78 if (likely(vcpu->arch.mmu->root_hpa != INVALID_PAGE)) in kvm_mmu_reload()
100 if (VALID_PAGE(vcpu->arch.mmu->root_hpa)) in kvm_mmu_load_cr3()
101 vcpu->arch.mmu->set_cr3(vcpu, vcpu->arch.mmu->root_hpa | in kvm_mmu_load_cr3()
156 static inline u8 permission_fault(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, in permission_fault() argument
179 bool fault = (mmu->permissions[index] >> pte_access) & 1; in permission_fault()
183 if (unlikely(mmu->pkru_mask)) { in permission_fault()
198 pkru_bits &= mmu->pkru_mask >> offset; in permission_fault()
Dpaging_tmpl.h34 #define PT_HAVE_ACCESSED_DIRTY(mmu) true argument
53 #define PT_HAVE_ACCESSED_DIRTY(mmu) true argument
66 #define PT_HAVE_ACCESSED_DIRTY(mmu) ((mmu)->ept_ad) argument
105 static inline void FNAME(protect_clean_gpte)(struct kvm_mmu *mmu, unsigned *access, in FNAME()
111 if (!PT_HAVE_ACCESSED_DIRTY(mmu)) in FNAME()
132 static int FNAME(cmpxchg_gpte)(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, in FNAME()
175 if (is_rsvd_bits_set(vcpu->arch.mmu, gpte, PT_PAGE_TABLE_LEVEL)) in FNAME()
182 if (PT_HAVE_ACCESSED_DIRTY(vcpu->arch.mmu) && in FNAME()
218 struct kvm_mmu *mmu, in FNAME()
229 if (!PT_HAVE_ACCESSED_DIRTY(mmu)) in FNAME()
[all …]
Dmmu.c372 return vcpu->arch.mmu == &vcpu->arch.guest_mmu; in kvm_vcpu_ad_need_write_protect()
2365 vcpu->arch.mmu->sync_page(vcpu, sp) == 0) { in __kvm_sync_page()
2570 role = vcpu->arch.mmu->mmu_role.base; in kvm_mmu_get_page()
2576 if (!vcpu->arch.mmu->direct_map in kvm_mmu_get_page()
2577 && vcpu->arch.mmu->root_level <= PT32_ROOT_LEVEL) { in kvm_mmu_get_page()
2651 iterator->level = vcpu->arch.mmu->shadow_root_level; in shadow_walk_init_using_root()
2654 vcpu->arch.mmu->root_level < PT64_ROOT_4LEVEL && in shadow_walk_init_using_root()
2655 !vcpu->arch.mmu->direct_map) in shadow_walk_init_using_root()
2663 BUG_ON(root != vcpu->arch.mmu->root_hpa); in shadow_walk_init_using_root()
2666 = vcpu->arch.mmu->pae_root[(addr >> 30) & 3]; in shadow_walk_init_using_root()
[all …]
Dmmu_audit.c59 if (!VALID_PAGE(vcpu->arch.mmu->root_hpa)) in mmu_spte_walk()
62 if (vcpu->arch.mmu->root_level >= PT64_ROOT_4LEVEL) { in mmu_spte_walk()
63 hpa_t root = vcpu->arch.mmu->root_hpa; in mmu_spte_walk()
66 __mmu_spte_walk(vcpu, sp, fn, vcpu->arch.mmu->root_level); in mmu_spte_walk()
71 hpa_t root = vcpu->arch.mmu->pae_root[i]; in mmu_spte_walk()
122 "ent %llxn", vcpu->arch.mmu->root_level, pfn, in audit_mappings()
/arch/um/kernel/skas/
Dmmu.c142 struct mm_context *mmu = &mm->context; in destroy_context() local
150 if (mmu->id.u.pid < 2) { in destroy_context()
152 mmu->id.u.pid); in destroy_context()
155 os_kill_ptraced_process(mmu->id.u.pid, 1); in destroy_context()
157 free_page(mmu->id.stack); in destroy_context()
158 free_ldt(mmu); in destroy_context()
DMakefile6 obj-y := clone.o mmu.o process.o syscall.o uaccess.o
/arch/m68k/kernel/
Dsetup_mm.c401 const char *cpu, *mmu, *fpu; in show_cpuinfo() local
450 mmu = "68851"; in show_cpuinfo()
452 mmu = "68030"; in show_cpuinfo()
454 mmu = "68040"; in show_cpuinfo()
456 mmu = "68060"; in show_cpuinfo()
458 mmu = "Sun-3"; in show_cpuinfo()
460 mmu = "Apollo"; in show_cpuinfo()
462 mmu = "ColdFire"; in show_cpuinfo()
464 mmu = "unknown"; in show_cpuinfo()
474 cpu, mmu, fpu, in show_cpuinfo()
Dsetup_no.c178 char *cpu, *mmu, *fpu; in show_cpuinfo() local
182 mmu = "none"; in show_cpuinfo()
192 cpu, mmu, fpu, in show_cpuinfo()
/arch/arm64/boot/dts/renesas/
Dr8a7795-es1.dtsi80 ipmmu_mp1: mmu@ec680000 {
88 ipmmu_sy: mmu@e7730000 {
96 /delete-node/ mmu@fd950000;
97 /delete-node/ mmu@fd960000;
98 /delete-node/ mmu@fd970000;
99 /delete-node/ mmu@febe0000;
100 /delete-node/ mmu@fe980000;
/arch/riscv/kernel/
Dcpu.c132 const char *compat, *isa, *mmu; in c_show() local
138 if (!of_property_read_string(node, "mmu-type", &mmu)) in c_show()
139 print_mmu(m, mmu); in c_show()
/arch/arm64/boot/dts/amd/
Damd-seattle-xgbe-b.dtsi89 compatible = "arm,mmu-401";
98 mmu-masters = <&xgmac0
105 compatible = "arm,mmu-401";
114 mmu-masters = <&xgmac1
/arch/sh/mm/
DMakefile19 mmu-y := nommu.o extable_32.o
20 mmu-$(CONFIG_MMU) := extable_$(BITS).o fault.o ioremap.o kmap.o \
23 obj-y += $(mmu-y)
/arch/sparc/kernel/
Dwuf.S266 LEON_PI(lda [%g0] ASI_LEON_MMUREGS, %l5) ! read mmu-ctrl reg
267 SUN_PI_(lda [%g0] ASI_M_MMUREGS, %l5) ! read mmu-ctrl reg
282 LEON_PI(lda [%g0] ASI_LEON_MMUREGS, %twin_tmp1) ! load mmu-ctrl again
283 SUN_PI_(lda [%g0] ASI_M_MMUREGS, %twin_tmp1) ! load mmu-ctrl again
/arch/arm/lib/
DMakefile19 mmu-y := clear_user.o copy_page.o getuser.o putuser.o \
31 lib-$(CONFIG_MMU) += $(mmu-y)
/arch/x86/hyperv/
DMakefile2 obj-y := hv_init.o mmu.o nested.o
/arch/powerpc/mm/book3s32/
DMakefile9 obj-y += mmu.o hash_low.o mmu_context.o tlb.o
/arch/xtensa/mm/
DMakefile7 obj-$(CONFIG_MMU) += cache.o fault.o ioremap.o mmu.o tlb.o
/arch/unicore32/mm/
DMakefile6 obj-y := extable.o fault.o init.o pgd.o mmu.o

12345