Searched refs:mmu_lock (Results 1 – 21 of 21) sorted by relevance
188 write_lock(&kvm->mmu_lock); in kvm_page_track_register_notifier()190 write_unlock(&kvm->mmu_lock); in kvm_page_track_register_notifier()206 write_lock(&kvm->mmu_lock); in kvm_page_track_unregister_notifier()208 write_unlock(&kvm->mmu_lock); in kvm_page_track_unregister_notifier()
36 lockdep_assert_held_read(&kvm->mmu_lock); in kvm_lockdep_assert_mmu_lock_held()38 lockdep_assert_held_write(&kvm->mmu_lock); in kvm_lockdep_assert_mmu_lock_held()169 lockdep_is_held_type(&kvm->mmu_lock, 0) || \212 lockdep_assert_held_write(&kvm->mmu_lock); in kvm_tdp_mmu_get_vcpu_root_hpa()303 lockdep_assert_held_write(&kvm->mmu_lock); in tdp_mmu_unlink_page()523 lockdep_assert_held_read(&kvm->mmu_lock); in tdp_mmu_set_spte_atomic_no_dirty_log()635 lockdep_assert_held_write(&kvm->mmu_lock); in __tdp_mmu_set_spte()717 if (need_resched() || rwlock_needbreak(&kvm->mmu_lock)) { in tdp_mmu_iter_cond_resched()724 cond_resched_rwlock_read(&kvm->mmu_lock); in tdp_mmu_iter_cond_resched()726 cond_resched_rwlock_write(&kvm->mmu_lock); in tdp_mmu_iter_cond_resched()[all …]
2047 if (need_resched() || rwlock_needbreak(&vcpu->kvm->mmu_lock)) { in mmu_sync_children()2054 cond_resched_rwlock_write(&vcpu->kvm->mmu_lock); in mmu_sync_children()2360 lockdep_assert_held_write(&kvm->mmu_lock); in __kvm_mmu_prepare_zap_page()2521 write_lock(&kvm->mmu_lock); in kvm_mmu_change_mmu_pages()2532 write_unlock(&kvm->mmu_lock); in kvm_mmu_change_mmu_pages()2543 write_lock(&kvm->mmu_lock); in kvm_mmu_unprotect_page()2551 write_unlock(&kvm->mmu_lock); in kvm_mmu_unprotect_page()3339 write_lock(&kvm->mmu_lock); in kvm_mmu_free_roots()3365 write_unlock(&kvm->mmu_lock); in kvm_mmu_free_roots()3427 write_lock(&vcpu->kvm->mmu_lock); in mmu_alloc_direct_roots()[all …]
39 lockdep_assert_held_write(&kvm->mmu_lock); in kvm_tdp_mmu_zap_sp()
927 write_lock(&vcpu->kvm->mmu_lock); in FNAME()940 write_unlock(&vcpu->kvm->mmu_lock); in FNAME()978 write_lock(&vcpu->kvm->mmu_lock); in FNAME()1013 write_unlock(&vcpu->kvm->mmu_lock); in FNAME()
723 spin_lock(&kvm->mmu_lock); in kvmhv_remove_nested()734 spin_unlock(&kvm->mmu_lock); in kvmhv_remove_nested()753 spin_lock(&kvm->mmu_lock); in kvmhv_release_all_nested()765 spin_unlock(&kvm->mmu_lock); in kvmhv_release_all_nested()782 spin_lock(&kvm->mmu_lock); in kvmhv_flush_nested()784 spin_unlock(&kvm->mmu_lock); in kvmhv_flush_nested()800 spin_lock(&kvm->mmu_lock); in kvmhv_get_nested()804 spin_unlock(&kvm->mmu_lock); in kvmhv_get_nested()812 spin_lock(&kvm->mmu_lock); in kvmhv_get_nested()825 spin_unlock(&kvm->mmu_lock); in kvmhv_get_nested()[all …]
63 spin_lock(&vcpu3s->mmu_lock); in kvmppc_mmu_hpte_cache_map()92 spin_unlock(&vcpu3s->mmu_lock); in kvmppc_mmu_hpte_cache_map()110 spin_lock(&vcpu3s->mmu_lock); in invalidate_pte()114 spin_unlock(&vcpu3s->mmu_lock); in invalidate_pte()127 spin_unlock(&vcpu3s->mmu_lock); in invalidate_pte()369 spin_lock_init(&vcpu3s->mmu_lock); in kvmppc_mmu_hpte_init()
635 spin_lock(&kvm->mmu_lock); in kvmppc_create_pte()771 spin_unlock(&kvm->mmu_lock); in kvmppc_create_pte()859 spin_lock(&kvm->mmu_lock); in kvmppc_book3s_instantiate_page()864 spin_unlock(&kvm->mmu_lock); in kvmppc_book3s_instantiate_page()994 spin_lock(&kvm->mmu_lock); in kvmppc_book3s_radix_page_fault()998 spin_unlock(&kvm->mmu_lock); in kvmppc_book3s_radix_page_fault()1104 spin_lock(&kvm->mmu_lock); in kvm_radix_test_clear_dirty()1116 spin_unlock(&kvm->mmu_lock); in kvm_radix_test_clear_dirty()1131 spin_unlock(&kvm->mmu_lock); in kvm_radix_test_clear_dirty()1176 spin_lock(&kvm->mmu_lock); in kvmppc_radix_flush_memslot()[all …]
245 arch_spin_lock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_enter()260 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_enter()274 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_enter()935 arch_spin_lock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_page_init_zero()947 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_page_init_zero()963 arch_spin_lock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_page_init_copy()978 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_page_init_copy()
153 spin_lock(&kvm->mmu_lock); in kvmppc_mmu_map_page()205 spin_unlock(&kvm->mmu_lock); in kvmppc_mmu_map_page()
551 arch_spin_lock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_rm_h_put_tce_indirect()592 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_rm_h_put_tce_indirect()
609 spin_lock(&kvm->mmu_lock); in kvmppc_book3s_hv_page_fault()614 spin_unlock(&kvm->mmu_lock); in kvmppc_book3s_hv_page_fault()743 spin_lock(&kvm->mmu_lock); in kvmppc_rmap_reset()750 spin_unlock(&kvm->mmu_lock); in kvmppc_rmap_reset()1383 spin_lock(&kvm->mmu_lock); in resize_hpt_pivot()1390 spin_unlock(&kvm->mmu_lock); in resize_hpt_pivot()
462 spin_lock(&kvm->mmu_lock); in kvmppc_e500_shadow_map()502 spin_unlock(&kvm->mmu_lock); in kvmppc_e500_shadow_map()
5083 spin_lock(&kvm->mmu_lock); in kvmppc_switch_mmu_to_hpt()5085 spin_unlock(&kvm->mmu_lock); in kvmppc_switch_mmu_to_hpt()5105 spin_lock(&kvm->mmu_lock); in kvmppc_switch_mmu_to_radix()5107 spin_unlock(&kvm->mmu_lock); in kvmppc_switch_mmu_to_radix()
61 cond_resched_lock(&kvm->mmu_lock); in stage2_apply_range()182 assert_spin_locked(&kvm->mmu_lock); in __unmap_stage2_range()205 cond_resched_lock(&kvm->mmu_lock); in pkvm_stage2_flush()232 spin_lock(&kvm->mmu_lock); in stage2_flush_vm()242 spin_unlock(&kvm->mmu_lock); in stage2_flush_vm()791 spin_lock(&kvm->mmu_lock); in stage2_unmap_vm()797 spin_unlock(&kvm->mmu_lock); in stage2_unmap_vm()810 spin_lock(&kvm->mmu_lock); in kvm_free_stage2_pgd()817 spin_unlock(&kvm->mmu_lock); in kvm_free_stage2_pgd()885 spin_lock(&kvm->mmu_lock); in kvm_phys_addr_ioremap()[all …]
110 write_lock(&kvm->mmu_lock); in kvm_mmu_rmaps_stat_show()130 write_unlock(&kvm->mmu_lock); in kvm_mmu_rmaps_stat_show()
7842 write_lock(&vcpu->kvm->mmu_lock); in reexecute_instruction()7844 write_unlock(&vcpu->kvm->mmu_lock); in reexecute_instruction()
522 spin_lock(&kvm->mmu_lock); in _kvm_mips_map_page_fast()557 spin_unlock(&kvm->mmu_lock); in _kvm_mips_map_page_fast()639 spin_lock(&kvm->mmu_lock); in kvm_mips_map_page()647 spin_unlock(&kvm->mmu_lock); in kvm_mips_map_page()676 spin_unlock(&kvm->mmu_lock); in kvm_mips_map_page()
227 spin_lock(&kvm->mmu_lock); in kvm_arch_flush_shadow_memslot()232 spin_unlock(&kvm->mmu_lock); in kvm_arch_flush_shadow_memslot()267 spin_lock(&kvm->mmu_lock); in kvm_arch_commit_memory_region()273 spin_unlock(&kvm->mmu_lock); in kvm_arch_commit_memory_region()
661 VM_WARN(!spin_is_locked(&kvm->mmu_lock), in find_kvm_secondary_pte()673 VM_WARN(!spin_is_locked(&kvm->mmu_lock), in find_kvm_host_pte()
134 spinlock_t mmu_lock; member