Home
last modified time | relevance | path

Searched refs:mmu_lock (Results 1 – 21 of 21) sorted by relevance

/arch/x86/kvm/mmu/
Dpage_track.c188 write_lock(&kvm->mmu_lock); in kvm_page_track_register_notifier()
190 write_unlock(&kvm->mmu_lock); in kvm_page_track_register_notifier()
206 write_lock(&kvm->mmu_lock); in kvm_page_track_unregister_notifier()
208 write_unlock(&kvm->mmu_lock); in kvm_page_track_unregister_notifier()
Dtdp_mmu.c36 lockdep_assert_held_read(&kvm->mmu_lock); in kvm_lockdep_assert_mmu_lock_held()
38 lockdep_assert_held_write(&kvm->mmu_lock); in kvm_lockdep_assert_mmu_lock_held()
169 lockdep_is_held_type(&kvm->mmu_lock, 0) || \
212 lockdep_assert_held_write(&kvm->mmu_lock); in kvm_tdp_mmu_get_vcpu_root_hpa()
303 lockdep_assert_held_write(&kvm->mmu_lock); in tdp_mmu_unlink_page()
523 lockdep_assert_held_read(&kvm->mmu_lock); in tdp_mmu_set_spte_atomic_no_dirty_log()
635 lockdep_assert_held_write(&kvm->mmu_lock); in __tdp_mmu_set_spte()
717 if (need_resched() || rwlock_needbreak(&kvm->mmu_lock)) { in tdp_mmu_iter_cond_resched()
724 cond_resched_rwlock_read(&kvm->mmu_lock); in tdp_mmu_iter_cond_resched()
726 cond_resched_rwlock_write(&kvm->mmu_lock); in tdp_mmu_iter_cond_resched()
[all …]
Dmmu.c2047 if (need_resched() || rwlock_needbreak(&vcpu->kvm->mmu_lock)) { in mmu_sync_children()
2054 cond_resched_rwlock_write(&vcpu->kvm->mmu_lock); in mmu_sync_children()
2360 lockdep_assert_held_write(&kvm->mmu_lock); in __kvm_mmu_prepare_zap_page()
2521 write_lock(&kvm->mmu_lock); in kvm_mmu_change_mmu_pages()
2532 write_unlock(&kvm->mmu_lock); in kvm_mmu_change_mmu_pages()
2543 write_lock(&kvm->mmu_lock); in kvm_mmu_unprotect_page()
2551 write_unlock(&kvm->mmu_lock); in kvm_mmu_unprotect_page()
3339 write_lock(&kvm->mmu_lock); in kvm_mmu_free_roots()
3365 write_unlock(&kvm->mmu_lock); in kvm_mmu_free_roots()
3427 write_lock(&vcpu->kvm->mmu_lock); in mmu_alloc_direct_roots()
[all …]
Dtdp_mmu.h39 lockdep_assert_held_write(&kvm->mmu_lock); in kvm_tdp_mmu_zap_sp()
Dpaging_tmpl.h927 write_lock(&vcpu->kvm->mmu_lock); in FNAME()
940 write_unlock(&vcpu->kvm->mmu_lock); in FNAME()
978 write_lock(&vcpu->kvm->mmu_lock); in FNAME()
1013 write_unlock(&vcpu->kvm->mmu_lock); in FNAME()
/arch/powerpc/kvm/
Dbook3s_hv_nested.c723 spin_lock(&kvm->mmu_lock); in kvmhv_remove_nested()
734 spin_unlock(&kvm->mmu_lock); in kvmhv_remove_nested()
753 spin_lock(&kvm->mmu_lock); in kvmhv_release_all_nested()
765 spin_unlock(&kvm->mmu_lock); in kvmhv_release_all_nested()
782 spin_lock(&kvm->mmu_lock); in kvmhv_flush_nested()
784 spin_unlock(&kvm->mmu_lock); in kvmhv_flush_nested()
800 spin_lock(&kvm->mmu_lock); in kvmhv_get_nested()
804 spin_unlock(&kvm->mmu_lock); in kvmhv_get_nested()
812 spin_lock(&kvm->mmu_lock); in kvmhv_get_nested()
825 spin_unlock(&kvm->mmu_lock); in kvmhv_get_nested()
[all …]
Dbook3s_mmu_hpte.c63 spin_lock(&vcpu3s->mmu_lock); in kvmppc_mmu_hpte_cache_map()
92 spin_unlock(&vcpu3s->mmu_lock); in kvmppc_mmu_hpte_cache_map()
110 spin_lock(&vcpu3s->mmu_lock); in invalidate_pte()
114 spin_unlock(&vcpu3s->mmu_lock); in invalidate_pte()
127 spin_unlock(&vcpu3s->mmu_lock); in invalidate_pte()
369 spin_lock_init(&vcpu3s->mmu_lock); in kvmppc_mmu_hpte_init()
Dbook3s_64_mmu_radix.c635 spin_lock(&kvm->mmu_lock); in kvmppc_create_pte()
771 spin_unlock(&kvm->mmu_lock); in kvmppc_create_pte()
859 spin_lock(&kvm->mmu_lock); in kvmppc_book3s_instantiate_page()
864 spin_unlock(&kvm->mmu_lock); in kvmppc_book3s_instantiate_page()
994 spin_lock(&kvm->mmu_lock); in kvmppc_book3s_radix_page_fault()
998 spin_unlock(&kvm->mmu_lock); in kvmppc_book3s_radix_page_fault()
1104 spin_lock(&kvm->mmu_lock); in kvm_radix_test_clear_dirty()
1116 spin_unlock(&kvm->mmu_lock); in kvm_radix_test_clear_dirty()
1131 spin_unlock(&kvm->mmu_lock); in kvm_radix_test_clear_dirty()
1176 spin_lock(&kvm->mmu_lock); in kvmppc_radix_flush_memslot()
[all …]
Dbook3s_hv_rm_mmu.c245 arch_spin_lock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_enter()
260 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_enter()
274 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_enter()
935 arch_spin_lock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_page_init_zero()
947 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_page_init_zero()
963 arch_spin_lock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_page_init_copy()
978 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_page_init_copy()
Dbook3s_64_mmu_host.c153 spin_lock(&kvm->mmu_lock); in kvmppc_mmu_map_page()
205 spin_unlock(&kvm->mmu_lock); in kvmppc_mmu_map_page()
Dbook3s_64_vio_hv.c551 arch_spin_lock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_rm_h_put_tce_indirect()
592 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_rm_h_put_tce_indirect()
Dbook3s_64_mmu_hv.c609 spin_lock(&kvm->mmu_lock); in kvmppc_book3s_hv_page_fault()
614 spin_unlock(&kvm->mmu_lock); in kvmppc_book3s_hv_page_fault()
743 spin_lock(&kvm->mmu_lock); in kvmppc_rmap_reset()
750 spin_unlock(&kvm->mmu_lock); in kvmppc_rmap_reset()
1383 spin_lock(&kvm->mmu_lock); in resize_hpt_pivot()
1390 spin_unlock(&kvm->mmu_lock); in resize_hpt_pivot()
De500_mmu_host.c462 spin_lock(&kvm->mmu_lock); in kvmppc_e500_shadow_map()
502 spin_unlock(&kvm->mmu_lock); in kvmppc_e500_shadow_map()
Dbook3s_hv.c5083 spin_lock(&kvm->mmu_lock); in kvmppc_switch_mmu_to_hpt()
5085 spin_unlock(&kvm->mmu_lock); in kvmppc_switch_mmu_to_hpt()
5105 spin_lock(&kvm->mmu_lock); in kvmppc_switch_mmu_to_radix()
5107 spin_unlock(&kvm->mmu_lock); in kvmppc_switch_mmu_to_radix()
/arch/arm64/kvm/
Dmmu.c61 cond_resched_lock(&kvm->mmu_lock); in stage2_apply_range()
182 assert_spin_locked(&kvm->mmu_lock); in __unmap_stage2_range()
205 cond_resched_lock(&kvm->mmu_lock); in pkvm_stage2_flush()
232 spin_lock(&kvm->mmu_lock); in stage2_flush_vm()
242 spin_unlock(&kvm->mmu_lock); in stage2_flush_vm()
791 spin_lock(&kvm->mmu_lock); in stage2_unmap_vm()
797 spin_unlock(&kvm->mmu_lock); in stage2_unmap_vm()
810 spin_lock(&kvm->mmu_lock); in kvm_free_stage2_pgd()
817 spin_unlock(&kvm->mmu_lock); in kvm_free_stage2_pgd()
885 spin_lock(&kvm->mmu_lock); in kvm_phys_addr_ioremap()
[all …]
/arch/x86/kvm/
Ddebugfs.c110 write_lock(&kvm->mmu_lock); in kvm_mmu_rmaps_stat_show()
130 write_unlock(&kvm->mmu_lock); in kvm_mmu_rmaps_stat_show()
Dx86.c7842 write_lock(&vcpu->kvm->mmu_lock); in reexecute_instruction()
7844 write_unlock(&vcpu->kvm->mmu_lock); in reexecute_instruction()
/arch/mips/kvm/
Dmmu.c522 spin_lock(&kvm->mmu_lock); in _kvm_mips_map_page_fast()
557 spin_unlock(&kvm->mmu_lock); in _kvm_mips_map_page_fast()
639 spin_lock(&kvm->mmu_lock); in kvm_mips_map_page()
647 spin_unlock(&kvm->mmu_lock); in kvm_mips_map_page()
676 spin_unlock(&kvm->mmu_lock); in kvm_mips_map_page()
Dmips.c227 spin_lock(&kvm->mmu_lock); in kvm_arch_flush_shadow_memslot()
232 spin_unlock(&kvm->mmu_lock); in kvm_arch_flush_shadow_memslot()
267 spin_lock(&kvm->mmu_lock); in kvm_arch_commit_memory_region()
273 spin_unlock(&kvm->mmu_lock); in kvm_arch_commit_memory_region()
/arch/powerpc/include/asm/
Dkvm_book3s_64.h661 VM_WARN(!spin_is_locked(&kvm->mmu_lock), in find_kvm_secondary_pte()
673 VM_WARN(!spin_is_locked(&kvm->mmu_lock), in find_kvm_host_pte()
Dkvm_book3s.h134 spinlock_t mmu_lock; member