Lines Matching refs:mmu
532 pgt->mmu = NULL; in kvm_pgtable_hyp_init()
574 struct kvm_s2_mmu *mmu; member
630 void kvm_tlb_flush_vmid_range(struct kvm_s2_mmu *mmu, in kvm_tlb_flush_vmid_range() argument
636 kvm_call_hyp(__kvm_tlb_flush_vmid, mmu); in kvm_tlb_flush_vmid_range()
643 kvm_call_hyp(__kvm_tlb_flush_vmid_range, mmu, addr, inval_pages); in kvm_tlb_flush_vmid_range()
773 struct kvm_s2_mmu *mmu) in stage2_try_break_pte() argument
799 kvm_tlb_flush_vmid_range(mmu, addr, size); in stage2_try_break_pte()
801 kvm_call_hyp(__kvm_tlb_flush_vmid_ipa, mmu, in stage2_try_break_pte()
839 struct kvm_s2_mmu *mmu) in stage2_unmap_clear_pte() argument
846 kvm_call_hyp(__kvm_tlb_flush_vmid_ipa, mmu, ctx->addr, in stage2_unmap_clear_pte()
849 kvm_call_hyp(__kvm_tlb_flush_vmid_ipa, mmu, ctx->addr, in stage2_unmap_clear_pte()
856 struct kvm_s2_mmu *mmu, in stage2_unmap_put_pte() argument
864 stage2_unmap_clear_pte(ctx, mmu); in stage2_unmap_put_pte()
918 struct kvm_pgtable *pgt = data->mmu->pgt; in stage2_map_walker_try_leaf()
961 if (!stage2_try_break_pte(ctx, data->mmu)) in stage2_map_walker_try_leaf()
1040 struct kvm_pgtable *pgt = data->mmu->pgt; in stage2_map_walk_leaf()
1065 if (!stage2_try_break_pte(ctx, data->mmu)) { in stage2_map_walk_leaf()
1125 WARN_ON(!(data->mmu->pgt->flags & KVM_PGTABLE_S2_IDMAP)); in stage2_coalesce_walk_table_post()
1135 WARN_ON(!stage2_try_break_pte(ctx, data->mmu)); in stage2_coalesce_walk_table_post()
1182 .mmu = pgt->mmu, in kvm_pgtable_stage2_map()
1217 .mmu = pgt->mmu, in kvm_pgtable_stage2_annotate()
1240 struct kvm_s2_mmu *mmu = pgt->mmu; in stage2_unmap_walker() local
1269 stage2_unmap_put_pte(ctx, mmu, mm_ops); in stage2_unmap_walker()
1271 stage2_unmap_clear_pte(ctx, mmu); in stage2_unmap_walker()
1307 kvm_tlb_flush_vmid_range(pgt->mmu, addr, size); in kvm_pgtable_stage2_unmap()
1333 kvm_tlb_flush_vmid_range(data->mmu, addr, size); in stage2_reclaim_leaf_walker()
1346 .mmu = pgt->mmu, in kvm_pgtable_stage2_reclaim_leaves()
1532 kvm_call_hyp(__kvm_tlb_flush_vmid_ipa_nsh, pgt->mmu, addr, level); in __kvm_pgtable_stage2_relax_perms()
1572 .mmu = pgt->mmu, in kvm_pgtable_stage2_create_unlinked()
1646 struct kvm_s2_mmu *mmu; in stage2_split_walker() local
1685 mmu = container_of(mc, struct kvm_s2_mmu, split_page_cache); in stage2_split_walker()
1689 childp = kvm_pgtable_stage2_create_unlinked(mmu->pgt, phys, in stage2_split_walker()
1694 if (!stage2_try_break_pte(ctx, mmu)) { in stage2_split_walker()
1724 int __kvm_pgtable_stage2_init(struct kvm_pgtable *pgt, struct kvm_s2_mmu *mmu, in __kvm_pgtable_stage2_init() argument
1730 u64 vtcr = mmu->arch->vtcr; in __kvm_pgtable_stage2_init()
1743 pgt->mmu = mmu; in __kvm_pgtable_stage2_init()