Searched refs:pv_mmu_ops (Results 1 – 9 of 9) sorted by relevance
/arch/x86/include/asm/ |
D | paravirt.h | 71 return PVOP_CALL0(unsigned long, pv_mmu_ops.read_cr2); in read_cr2() 76 PVOP_VCALL1(pv_mmu_ops.write_cr2, x); in write_cr2() 81 return PVOP_CALL0(unsigned long, pv_mmu_ops.read_cr3); in read_cr3() 86 PVOP_VCALL1(pv_mmu_ops.write_cr3, x); in write_cr3() 306 PVOP_VCALL2(pv_mmu_ops.activate_mm, prev, next); in paravirt_activate_mm() 312 PVOP_VCALL2(pv_mmu_ops.dup_mmap, oldmm, mm); in paravirt_arch_dup_mmap() 317 PVOP_VCALL1(pv_mmu_ops.exit_mmap, mm); in paravirt_arch_exit_mmap() 322 PVOP_VCALL0(pv_mmu_ops.flush_tlb_user); in __flush_tlb() 326 PVOP_VCALL0(pv_mmu_ops.flush_tlb_kernel); in __flush_tlb_global() 330 PVOP_VCALL1(pv_mmu_ops.flush_tlb_single, addr); in __flush_tlb_single() [all …]
|
D | paravirt_types.h | 222 struct pv_mmu_ops { struct 354 struct pv_mmu_ops pv_mmu_ops; member 364 extern struct pv_mmu_ops pv_mmu_ops;
|
/arch/x86/kernel/ |
D | paravirt_patch_32.c | 8 DEF_NATIVE(pv_mmu_ops, read_cr2, "mov %cr2, %eax"); 9 DEF_NATIVE(pv_mmu_ops, write_cr3, "mov %eax, %cr3"); 10 DEF_NATIVE(pv_mmu_ops, read_cr3, "mov %cr3, %eax"); 48 PATCH_SITE(pv_mmu_ops, read_cr2); in native_patch() 49 PATCH_SITE(pv_mmu_ops, read_cr3); in native_patch() 50 PATCH_SITE(pv_mmu_ops, write_cr3); in native_patch()
|
D | paravirt_patch_64.c | 9 DEF_NATIVE(pv_mmu_ops, read_cr2, "movq %cr2, %rax"); 10 DEF_NATIVE(pv_mmu_ops, read_cr3, "movq %cr3, %rax"); 11 DEF_NATIVE(pv_mmu_ops, write_cr3, "movq %rdi, %cr3"); 59 PATCH_SITE(pv_mmu_ops, read_cr2); in native_patch() 60 PATCH_SITE(pv_mmu_ops, read_cr3); in native_patch() 61 PATCH_SITE(pv_mmu_ops, write_cr3); in native_patch()
|
D | paravirt.c | 143 .pv_mmu_ops = pv_mmu_ops, in get_call_destination() 419 struct pv_mmu_ops pv_mmu_ops = { variable 493 EXPORT_SYMBOL (pv_mmu_ops);
|
D | asm-offsets.c | 69 OFFSET(PV_MMU_read_cr2, pv_mmu_ops, read_cr2); in common()
|
/arch/x86/lguest/ |
D | boot.c | 1457 pv_mmu_ops.write_cr3 = lguest_write_cr3; in lguest_init() 1458 pv_mmu_ops.flush_tlb_user = lguest_flush_tlb_user; in lguest_init() 1459 pv_mmu_ops.flush_tlb_single = lguest_flush_tlb_single; in lguest_init() 1460 pv_mmu_ops.flush_tlb_kernel = lguest_flush_tlb_kernel; in lguest_init() 1461 pv_mmu_ops.set_pte = lguest_set_pte; in lguest_init() 1462 pv_mmu_ops.set_pte_at = lguest_set_pte_at; in lguest_init() 1463 pv_mmu_ops.set_pmd = lguest_set_pmd; in lguest_init() 1465 pv_mmu_ops.set_pte_atomic = lguest_set_pte_atomic; in lguest_init() 1466 pv_mmu_ops.pte_clear = lguest_pte_clear; in lguest_init() 1467 pv_mmu_ops.pmd_clear = lguest_pmd_clear; in lguest_init() [all …]
|
/arch/x86/xen/ |
D | mmu.c | 2237 pv_mmu_ops.write_cr3 = &xen_write_cr3; in xen_write_cr3_init() 2390 pv_mmu_ops.set_pte = xen_set_pte; in xen_post_allocator_init() 2391 pv_mmu_ops.set_pmd = xen_set_pmd; in xen_post_allocator_init() 2392 pv_mmu_ops.set_pud = xen_set_pud; in xen_post_allocator_init() 2394 pv_mmu_ops.set_pgd = xen_set_pgd; in xen_post_allocator_init() 2399 pv_mmu_ops.alloc_pte = xen_alloc_pte; in xen_post_allocator_init() 2400 pv_mmu_ops.alloc_pmd = xen_alloc_pmd; in xen_post_allocator_init() 2401 pv_mmu_ops.release_pte = xen_release_pte; in xen_post_allocator_init() 2402 pv_mmu_ops.release_pmd = xen_release_pmd; in xen_post_allocator_init() 2404 pv_mmu_ops.alloc_pud = xen_alloc_pud; in xen_post_allocator_init() [all …]
|
D | enlighten.c | 1145 pv_mmu_ops.read_cr2 = xen_read_cr2_direct; in xen_setup_vcpu_info_placement() 1606 pv_mmu_ops.ptep_modify_prot_start = xen_ptep_modify_prot_start; in xen_start_kernel() 1607 pv_mmu_ops.ptep_modify_prot_commit = xen_ptep_modify_prot_commit; in xen_start_kernel()
|