• Home
  • Raw
  • Download

Lines Matching refs:memslot

1459 	struct kvm_memory_slot *memslot;  in kvm_handle_hva_range()  local
1466 kvm_for_each_memslot(memslot, slots) { in kvm_handle_hva_range()
1470 hva_start = max(start, memslot->userspace_addr); in kvm_handle_hva_range()
1471 hva_end = min(end, memslot->userspace_addr + in kvm_handle_hva_range()
1472 (memslot->npages << PAGE_SHIFT)); in kvm_handle_hva_range()
1479 gfn_start = hva_to_gfn_memslot(hva_start, memslot); in kvm_handle_hva_range()
1480 gfn_end = hva_to_gfn_memslot(hva_end + PAGE_SIZE - 1, memslot); in kvm_handle_hva_range()
1482 for_each_slot_rmap_range(memslot, PG_LEVEL_4K, in kvm_handle_hva_range()
1486 ret |= handler(kvm, iterator.rmap, memslot, in kvm_handle_hva_range()
5235 slot_handle_level_range(struct kvm *kvm, struct kvm_memory_slot *memslot, in slot_handle_level_range() argument
5242 for_each_slot_rmap_range(memslot, start_level, end_level, start_gfn, in slot_handle_level_range()
5268 slot_handle_level(struct kvm *kvm, struct kvm_memory_slot *memslot, in slot_handle_level() argument
5272 return slot_handle_level_range(kvm, memslot, fn, start_level, in slot_handle_level()
5273 end_level, memslot->base_gfn, in slot_handle_level()
5274 memslot->base_gfn + memslot->npages - 1, in slot_handle_level()
5279 slot_handle_all_level(struct kvm *kvm, struct kvm_memory_slot *memslot, in slot_handle_all_level() argument
5282 return slot_handle_level(kvm, memslot, fn, PG_LEVEL_4K, in slot_handle_all_level()
5287 slot_handle_large_level(struct kvm *kvm, struct kvm_memory_slot *memslot, in slot_handle_large_level() argument
5290 return slot_handle_level(kvm, memslot, fn, PG_LEVEL_4K + 1, in slot_handle_large_level()
5295 slot_handle_leaf(struct kvm *kvm, struct kvm_memory_slot *memslot, in slot_handle_leaf() argument
5298 return slot_handle_level(kvm, memslot, fn, PG_LEVEL_4K, in slot_handle_leaf()
5506 struct kvm_memory_slot *memslot; in kvm_zap_gfn_range() local
5513 kvm_for_each_memslot(memslot, slots) { in kvm_zap_gfn_range()
5516 start = max(gfn_start, memslot->base_gfn); in kvm_zap_gfn_range()
5517 end = min(gfn_end, memslot->base_gfn + memslot->npages); in kvm_zap_gfn_range()
5521 slot_handle_level_range(kvm, memslot, kvm_zap_rmapp, in kvm_zap_gfn_range()
5544 struct kvm_memory_slot *memslot, in kvm_mmu_slot_remove_write_access() argument
5550 flush = slot_handle_level(kvm, memslot, slot_rmap_write_protect, in kvm_mmu_slot_remove_write_access()
5553 flush |= kvm_tdp_mmu_wrprot_slot(kvm, memslot, PG_LEVEL_4K); in kvm_mmu_slot_remove_write_access()
5568 kvm_arch_flush_remote_tlbs_memslot(kvm, memslot); in kvm_mmu_slot_remove_write_access()
5611 const struct kvm_memory_slot *memslot) in kvm_mmu_zap_collapsible_sptes() argument
5615 slot_handle_leaf(kvm, (struct kvm_memory_slot *)memslot, in kvm_mmu_zap_collapsible_sptes()
5619 kvm_tdp_mmu_zap_collapsible_sptes(kvm, memslot); in kvm_mmu_zap_collapsible_sptes()
5624 struct kvm_memory_slot *memslot) in kvm_arch_flush_remote_tlbs_memslot() argument
5634 kvm_flush_remote_tlbs_with_address(kvm, memslot->base_gfn, in kvm_arch_flush_remote_tlbs_memslot()
5635 memslot->npages); in kvm_arch_flush_remote_tlbs_memslot()
5639 struct kvm_memory_slot *memslot) in kvm_mmu_slot_leaf_clear_dirty() argument
5644 flush = slot_handle_leaf(kvm, memslot, __rmap_clear_dirty, false); in kvm_mmu_slot_leaf_clear_dirty()
5646 flush |= kvm_tdp_mmu_clear_dirty_slot(kvm, memslot); in kvm_mmu_slot_leaf_clear_dirty()
5656 kvm_arch_flush_remote_tlbs_memslot(kvm, memslot); in kvm_mmu_slot_leaf_clear_dirty()
5661 struct kvm_memory_slot *memslot) in kvm_mmu_slot_largepage_remove_write_access() argument
5666 flush = slot_handle_large_level(kvm, memslot, slot_rmap_write_protect, in kvm_mmu_slot_largepage_remove_write_access()
5669 flush |= kvm_tdp_mmu_wrprot_slot(kvm, memslot, PG_LEVEL_2M); in kvm_mmu_slot_largepage_remove_write_access()
5673 kvm_arch_flush_remote_tlbs_memslot(kvm, memslot); in kvm_mmu_slot_largepage_remove_write_access()
5678 struct kvm_memory_slot *memslot) in kvm_mmu_slot_set_dirty() argument
5683 flush = slot_handle_all_level(kvm, memslot, __rmap_set_dirty, false); in kvm_mmu_slot_set_dirty()
5685 flush |= kvm_tdp_mmu_slot_set_dirty(kvm, memslot); in kvm_mmu_slot_set_dirty()
5689 kvm_arch_flush_remote_tlbs_memslot(kvm, memslot); in kvm_mmu_slot_set_dirty()
5952 struct kvm_memory_slot *memslot; in kvm_mmu_calculate_default_mmu_pages() local
5958 kvm_for_each_memslot(memslot, slots) in kvm_mmu_calculate_default_mmu_pages()
5959 nr_pages += memslot->npages; in kvm_mmu_calculate_default_mmu_pages()