Home
last modified time | relevance | path

Searched refs:kvm (Results 1 – 25 of 165) sorted by relevance

1234567

/arch/powerpc/include/asm/
Dkvm_book3s_uvmem.h9 int kvmppc_uvmem_slot_init(struct kvm *kvm, const struct kvm_memory_slot *slot);
10 void kvmppc_uvmem_slot_free(struct kvm *kvm,
12 unsigned long kvmppc_h_svm_page_in(struct kvm *kvm,
16 unsigned long kvmppc_h_svm_page_out(struct kvm *kvm,
20 unsigned long kvmppc_h_svm_init_start(struct kvm *kvm);
21 unsigned long kvmppc_h_svm_init_done(struct kvm *kvm);
22 int kvmppc_send_page_to_uv(struct kvm *kvm, unsigned long gfn);
23 unsigned long kvmppc_h_svm_init_abort(struct kvm *kvm);
25 struct kvm *kvm, bool skip_page_out);
26 int kvmppc_uvmem_memslot_create(struct kvm *kvm,
[all …]
Dkvm_ppc.h160 extern void kvmppc_set_hpt(struct kvm *kvm, struct kvm_hpt_info *info);
161 extern long kvmppc_alloc_reset_hpt(struct kvm *kvm, int order);
163 extern void kvmppc_rmap_reset(struct kvm *kvm);
164 extern long kvmppc_prepare_vrma(struct kvm *kvm,
169 extern long kvm_spapr_tce_attach_iommu_group(struct kvm *kvm, int tablefd,
171 extern void kvm_spapr_tce_release_iommu_group(struct kvm *kvm,
173 extern int kvmppc_switch_mmu_to_hpt(struct kvm *kvm);
174 extern int kvmppc_switch_mmu_to_radix(struct kvm *kvm);
175 extern void kvmppc_setup_partition_table(struct kvm *kvm);
177 extern long kvm_vm_ioctl_create_spapr_tce(struct kvm *kvm,
[all …]
Dkvm_book3s.h86 struct kvm *kvm; member
160 extern long kvmppc_hv_find_lock_hpte(struct kvm *kvm, gva_t eaddr,
174 extern int kvmppc_book3s_hcall_implemented(struct kvm *kvm, unsigned long hc);
193 extern void kvmppc_radix_tlbie_page(struct kvm *kvm, unsigned long addr,
195 extern void kvmppc_unmap_pte(struct kvm *kvm, pte_t *pte, unsigned long gpa,
199 extern bool kvmppc_hv_handle_set_rc(struct kvm *kvm, bool nested,
207 extern int kvmppc_init_vm_radix(struct kvm *kvm);
208 extern void kvmppc_free_radix(struct kvm *kvm);
209 extern void kvmppc_free_pgtable_radix(struct kvm *kvm, pgd_t *pgd,
213 extern int kvm_unmap_radix(struct kvm *kvm, struct kvm_memory_slot *memslot,
[all …]
/arch/x86/kvm/mmu/
Dtdp_mmu.h8 void kvm_mmu_init_tdp_mmu(struct kvm *kvm);
9 void kvm_mmu_uninit_tdp_mmu(struct kvm *kvm);
11 bool is_tdp_mmu_root(struct kvm *kvm, hpa_t root);
13 void kvm_tdp_mmu_free_root(struct kvm *kvm, struct kvm_mmu_page *root);
15 bool __kvm_tdp_mmu_zap_gfn_range(struct kvm *kvm, gfn_t start, gfn_t end,
17 static inline bool kvm_tdp_mmu_zap_gfn_range(struct kvm *kvm, gfn_t start, in kvm_tdp_mmu_zap_gfn_range() argument
20 return __kvm_tdp_mmu_zap_gfn_range(kvm, start, end, true); in kvm_tdp_mmu_zap_gfn_range()
22 static inline bool kvm_tdp_mmu_zap_sp(struct kvm *kvm, struct kvm_mmu_page *sp) in kvm_tdp_mmu_zap_sp() argument
30 return __kvm_tdp_mmu_zap_gfn_range(kvm, sp->gfn, end, false); in kvm_tdp_mmu_zap_sp()
32 void kvm_tdp_mmu_zap_all(struct kvm *kvm);
[all …]
Dtdp_mmu.c25 void kvm_mmu_init_tdp_mmu(struct kvm *kvm) in kvm_mmu_init_tdp_mmu() argument
31 kvm->arch.tdp_mmu_enabled = true; in kvm_mmu_init_tdp_mmu()
33 INIT_LIST_HEAD(&kvm->arch.tdp_mmu_roots); in kvm_mmu_init_tdp_mmu()
34 INIT_LIST_HEAD(&kvm->arch.tdp_mmu_pages); in kvm_mmu_init_tdp_mmu()
37 void kvm_mmu_uninit_tdp_mmu(struct kvm *kvm) in kvm_mmu_uninit_tdp_mmu() argument
39 if (!kvm->arch.tdp_mmu_enabled) in kvm_mmu_uninit_tdp_mmu()
42 WARN_ON(!list_empty(&kvm->arch.tdp_mmu_roots)); in kvm_mmu_uninit_tdp_mmu()
45 static void tdp_mmu_put_root(struct kvm *kvm, struct kvm_mmu_page *root) in tdp_mmu_put_root() argument
47 if (kvm_mmu_put_root(kvm, root)) in tdp_mmu_put_root()
48 kvm_tdp_mmu_free_root(kvm, root); in tdp_mmu_put_root()
[all …]
Dpage_track.c86 void kvm_slot_page_track_add_page(struct kvm *kvm, in kvm_slot_page_track_add_page() argument
103 if (kvm_mmu_slot_gfn_write_protect(kvm, slot, gfn)) in kvm_slot_page_track_add_page()
104 kvm_flush_remote_tlbs(kvm); in kvm_slot_page_track_add_page()
121 void kvm_slot_page_track_remove_page(struct kvm *kvm, in kvm_slot_page_track_remove_page() argument
158 void kvm_page_track_cleanup(struct kvm *kvm) in kvm_page_track_cleanup() argument
162 head = &kvm->arch.track_notifier_head; in kvm_page_track_cleanup()
166 int kvm_page_track_init(struct kvm *kvm) in kvm_page_track_init() argument
170 head = &kvm->arch.track_notifier_head; in kvm_page_track_init()
180 kvm_page_track_register_notifier(struct kvm *kvm, in kvm_page_track_register_notifier() argument
185 head = &kvm->arch.track_notifier_head; in kvm_page_track_register_notifier()
[all …]
Dmmu_audit.c28 #define audit_printk(kvm, fmt, args...) \ argument
30 fmt, audit_point_name[kvm->arch.audit_point], ##args)
83 typedef void (*sp_handler) (struct kvm *kvm, struct kvm_mmu_page *sp);
85 static void walk_all_active_sps(struct kvm *kvm, sp_handler fn) in walk_all_active_sps() argument
89 list_for_each_entry(sp, &kvm->arch.active_mmu_pages, link) in walk_all_active_sps()
90 fn(kvm, sp); in walk_all_active_sps()
104 audit_printk(vcpu->kvm, "unsync sp: %p " in audit_mappings()
121 audit_printk(vcpu->kvm, "levels %d pfn %llx hpa %llx " in audit_mappings()
126 static void inspect_spte_has_rmap(struct kvm *kvm, u64 *sptep) in inspect_spte_has_rmap() argument
138 slots = kvm_memslots_for_spte_role(kvm, rev_sp->role); in inspect_spte_has_rmap()
[all …]
/arch/powerpc/kvm/
Dbook3s_hv_uvmem.c231 struct kvm *kvm; member
246 int kvmppc_uvmem_slot_init(struct kvm *kvm, const struct kvm_memory_slot *slot) in kvmppc_uvmem_slot_init() argument
261 mutex_lock(&kvm->arch.uvmem_lock); in kvmppc_uvmem_slot_init()
262 list_add(&p->list, &kvm->arch.uvmem_pfns); in kvmppc_uvmem_slot_init()
263 mutex_unlock(&kvm->arch.uvmem_lock); in kvmppc_uvmem_slot_init()
271 void kvmppc_uvmem_slot_free(struct kvm *kvm, const struct kvm_memory_slot *slot) in kvmppc_uvmem_slot_free() argument
275 mutex_lock(&kvm->arch.uvmem_lock); in kvmppc_uvmem_slot_free()
276 list_for_each_entry_safe(p, next, &kvm->arch.uvmem_pfns, list) { in kvmppc_uvmem_slot_free()
284 mutex_unlock(&kvm->arch.uvmem_lock); in kvmppc_uvmem_slot_free()
287 static void kvmppc_mark_gfn(unsigned long gfn, struct kvm *kvm, in kvmppc_mark_gfn() argument
[all …]
DMakefile6 ccflags-y := -Ivirt/kvm -Iarch/powerpc/kvm
7 KVM := ../../../virt/kvm
19 kvm-e500-objs := \
29 kvm-objs-$(CONFIG_KVM_E500V2) := $(kvm-e500-objs)
31 kvm-e500mc-objs := \
41 kvm-objs-$(CONFIG_KVM_E500MC) := $(kvm-e500mc-objs)
43 kvm-book3s_64-builtin-objs-$(CONFIG_SPAPR_TCE_IOMMU) := \
46 kvm-pr-y := \
59 kvm-book3s_64-builtin-objs-$(CONFIG_KVM_BOOK3S_64_HANDLER) += \
63 kvm-book3s_64-builtin-objs-$(CONFIG_KVM_BOOK3S_64_HANDLER) += \
[all …]
Dbook3s_64_mmu_radix.c90 int lpid = vcpu->kvm->arch.lpid; in kvmhv_copy_tofrom_guest_radix()
134 struct kvm *kvm = vcpu->kvm; in kvmppc_mmu_walk_radix_tree() local
166 vcpu->srcu_idx = srcu_read_lock(&kvm->srcu); in kvmppc_mmu_walk_radix_tree()
167 ret = kvm_read_guest(kvm, addr, &rpte, sizeof(rpte)); in kvmppc_mmu_walk_radix_tree()
168 srcu_read_unlock(&kvm->srcu, vcpu->srcu_idx); in kvmppc_mmu_walk_radix_tree()
229 struct kvm *kvm = vcpu->kvm; in kvmppc_mmu_radix_translate_table() local
244 vcpu->srcu_idx = srcu_read_lock(&kvm->srcu); in kvmppc_mmu_radix_translate_table()
245 ret = kvm_read_guest(kvm, ptbl, &entry, sizeof(entry)); in kvmppc_mmu_radix_translate_table()
246 srcu_read_unlock(&kvm->srcu, vcpu->srcu_idx); in kvmppc_mmu_radix_translate_table()
276 vcpu->kvm->arch.process_table, pid, &pte); in kvmppc_mmu_radix_xlate()
[all …]
Dbook3s_64_mmu_hv.c45 static long kvmppc_virtmode_do_h_enter(struct kvm *kvm, unsigned long flags,
51 struct kvm *kvm; member
116 void kvmppc_set_hpt(struct kvm *kvm, struct kvm_hpt_info *info) in kvmppc_set_hpt() argument
118 atomic64_set(&kvm->arch.mmio_update, 0); in kvmppc_set_hpt()
119 kvm->arch.hpt = *info; in kvmppc_set_hpt()
120 kvm->arch.sdr1 = __pa(info->virt) | (info->order - 18); in kvmppc_set_hpt()
123 info->virt, (long)info->order, kvm->arch.lpid); in kvmppc_set_hpt()
126 long kvmppc_alloc_reset_hpt(struct kvm *kvm, int order) in kvmppc_alloc_reset_hpt() argument
131 mutex_lock(&kvm->arch.mmu_setup_lock); in kvmppc_alloc_reset_hpt()
132 if (kvm->arch.mmu_ready) { in kvmppc_alloc_reset_hpt()
[all …]
Dbook3s_hv_nested.c232 if (vcpu->kvm->arch.l1_ptcr == 0) in kvmhv_enter_nested_guest()
241 vcpu->srcu_idx = srcu_read_lock(&vcpu->kvm->srcu); in kvmhv_enter_nested_guest()
246 srcu_read_unlock(&vcpu->kvm->srcu, vcpu->srcu_idx); in kvmhv_enter_nested_guest()
278 l2 = kvmhv_get_nested(vcpu->kvm, l2_hv.lpid, true); in kvmhv_enter_nested_guest()
348 vcpu->srcu_idx = srcu_read_lock(&vcpu->kvm->srcu); in kvmhv_enter_nested_guest()
353 srcu_read_unlock(&vcpu->kvm->srcu, vcpu->srcu_idx); in kvmhv_enter_nested_guest()
454 void kvmhv_vm_nested_init(struct kvm *kvm) in kvmhv_vm_nested_init() argument
456 kvm->arch.max_nested_lpid = -1; in kvmhv_vm_nested_init()
466 struct kvm *kvm = vcpu->kvm; in kvmhv_set_partition_table() local
471 srcu_idx = srcu_read_lock(&kvm->srcu); in kvmhv_set_partition_table()
[all …]
Dbook3s_rtas.c34 rc = kvmppc_xive_set_xive(vcpu->kvm, irq, server, priority); in kvm_rtas_set_xive()
36 rc = kvmppc_xics_set_xive(vcpu->kvm, irq, server, priority); in kvm_rtas_set_xive()
57 rc = kvmppc_xive_get_xive(vcpu->kvm, irq, &server, &priority); in kvm_rtas_get_xive()
59 rc = kvmppc_xics_get_xive(vcpu->kvm, irq, &server, &priority); in kvm_rtas_get_xive()
84 rc = kvmppc_xive_int_off(vcpu->kvm, irq); in kvm_rtas_int_off()
86 rc = kvmppc_xics_int_off(vcpu->kvm, irq); in kvm_rtas_int_off()
106 rc = kvmppc_xive_int_on(vcpu->kvm, irq); in kvm_rtas_int_on()
108 rc = kvmppc_xics_int_on(vcpu->kvm, irq); in kvm_rtas_int_on()
142 static int rtas_token_undefine(struct kvm *kvm, char *name) in rtas_token_undefine() argument
146 lockdep_assert_held(&kvm->arch.rtas_token_lock); in rtas_token_undefine()
[all …]
/arch/x86/kvm/
Dirq_comm.c31 struct kvm *kvm, int irq_source_id, int level, in kvm_set_pic_irq() argument
34 struct kvm_pic *pic = kvm->arch.vpic; in kvm_set_pic_irq()
39 struct kvm *kvm, int irq_source_id, int level, in kvm_set_ioapic_irq() argument
42 struct kvm_ioapic *ioapic = kvm->arch.vioapic; in kvm_set_ioapic_irq()
47 int kvm_irq_delivery_to_apic(struct kvm *kvm, struct kvm_lapic *src, in kvm_irq_delivery_to_apic() argument
55 if (kvm_irq_delivery_to_apic_fast(kvm, src, irq, &r, dest_map)) in kvm_irq_delivery_to_apic()
66 kvm_for_each_vcpu(i, vcpu, kvm) { in kvm_irq_delivery_to_apic()
95 lowest = kvm_get_vcpu(kvm, idx); in kvm_irq_delivery_to_apic()
104 void kvm_set_msi_irq(struct kvm *kvm, struct kvm_kernel_irq_routing_entry *e, in kvm_set_msi_irq() argument
107 trace_kvm_msi_set_irq(e->msi.address_lo | (kvm->arch.x2apic_format ? in kvm_set_msi_irq()
[all …]
Dirq.h25 struct kvm;
53 struct kvm *kvm; member
63 int kvm_pic_init(struct kvm *kvm);
64 void kvm_pic_destroy(struct kvm *kvm);
65 int kvm_pic_read_irq(struct kvm *kvm);
68 static inline int irqchip_split(struct kvm *kvm) in irqchip_split() argument
70 int mode = kvm->arch.irqchip_mode; in irqchip_split()
77 static inline int irqchip_kernel(struct kvm *kvm) in irqchip_kernel() argument
79 int mode = kvm->arch.irqchip_mode; in irqchip_kernel()
86 static inline int pic_in_kernel(struct kvm *kvm) in pic_in_kernel() argument
[all …]
/arch/arm64/kvm/vgic/
Dvgic-init.c52 void kvm_vgic_early_init(struct kvm *kvm) in kvm_vgic_early_init() argument
54 struct vgic_dist *dist = &kvm->arch.vgic; in kvm_vgic_early_init()
71 int kvm_vgic_create(struct kvm *kvm, u32 type) in kvm_vgic_create() argument
76 if (irqchip_in_kernel(kvm)) in kvm_vgic_create()
90 if (!lock_all_vcpus(kvm)) in kvm_vgic_create()
93 kvm_for_each_vcpu(i, vcpu, kvm) { in kvm_vgic_create()
100 kvm->arch.max_vcpus = VGIC_V2_MAX_CPUS; in kvm_vgic_create()
102 kvm->arch.max_vcpus = VGIC_V3_MAX_CPUS; in kvm_vgic_create()
104 if (atomic_read(&kvm->online_vcpus) > kvm->arch.max_vcpus) { in kvm_vgic_create()
109 kvm->arch.vgic.in_kernel = true; in kvm_vgic_create()
[all …]
Dvgic.h162 struct vgic_irq *vgic_get_irq(struct kvm *kvm, struct kvm_vcpu *vcpu,
164 void __vgic_put_lpi_locked(struct kvm *kvm, struct vgic_irq *irq);
165 void vgic_put_irq(struct kvm *kvm, struct vgic_irq *irq);
169 bool vgic_queue_irq_unlock(struct kvm *kvm, struct vgic_irq *irq,
171 void vgic_kick_vcpus(struct kvm *kvm);
173 int vgic_check_ioaddr(struct kvm *kvm, phys_addr_t *ioaddr,
190 int vgic_v2_map_resources(struct kvm *kvm);
191 int vgic_register_dist_iodev(struct kvm *kvm, gpa_t dist_base_address,
219 int vgic_v3_map_resources(struct kvm *kvm);
220 int vgic_v3_lpi_sync_pending_status(struct kvm *kvm, struct vgic_irq *irq);
[all …]
Dvgic-debug.c55 static void iter_init(struct kvm *kvm, struct vgic_state_iter *iter, in iter_init() argument
58 int nr_cpus = atomic_read(&kvm->online_vcpus); in iter_init()
63 iter->nr_spis = kvm->arch.vgic.nr_spis; in iter_init()
64 if (kvm->arch.vgic.vgic_model == KVM_DEV_TYPE_ARM_VGIC_V3) { in iter_init()
65 iter->nr_lpis = vgic_copy_lpi_list(kvm, NULL, &iter->lpi_array); in iter_init()
85 struct kvm *kvm = (struct kvm *)s->private; in vgic_debug_start() local
88 mutex_lock(&kvm->lock); in vgic_debug_start()
89 iter = kvm->arch.vgic.iter; in vgic_debug_start()
101 iter_init(kvm, iter, *pos); in vgic_debug_start()
102 kvm->arch.vgic.iter = iter; in vgic_debug_start()
[all …]
/arch/s390/kvm/
Dkvm-s390.c245 static int sca_switch_to_extended(struct kvm *kvm);
278 struct kvm *kvm; in kvm_clock_sync() local
283 list_for_each_entry(kvm, &vm_list, vm_list) { in kvm_clock_sync()
284 kvm_for_each_vcpu(i, vcpu, kvm) { in kvm_clock_sync()
287 kvm->arch.epoch = vcpu->arch.sie_block->epoch; in kvm_clock_sync()
288 kvm->arch.epdx = vcpu->arch.sie_block->epdx; in kvm_clock_sync()
515 int kvm_vm_ioctl_check_extension(struct kvm *kvm, long ext) in kvm_vm_ioctl_check_extension() argument
553 if (hpage && !kvm_is_ucontrol(kvm)) in kvm_vm_ioctl_check_extension()
592 void kvm_arch_sync_dirty_log(struct kvm *kvm, struct kvm_memory_slot *memslot) in kvm_arch_sync_dirty_log() argument
597 struct gmap *gmap = kvm->arch.gmap; in kvm_arch_sync_dirty_log()
[all …]
Dpv.c26 KVM_UV_EVENT(vcpu->kvm, 3, "PROTVIRT DESTROY VCPU %d: rc %x rrc %x", in kvm_s390_pv_destroy_cpu()
68 uvcb.guest_handle = kvm_s390_pv_get_handle(vcpu->kvm); in kvm_s390_pv_create_cpu()
84 KVM_UV_EVENT(vcpu->kvm, 3, in kvm_s390_pv_create_cpu()
99 vcpu->arch.sie_block->pv_handle_config = kvm_s390_pv_get_handle(vcpu->kvm); in kvm_s390_pv_create_cpu()
106 static void kvm_s390_pv_dealloc_vm(struct kvm *kvm) in kvm_s390_pv_dealloc_vm() argument
108 vfree(kvm->arch.pv.stor_var); in kvm_s390_pv_dealloc_vm()
109 free_pages(kvm->arch.pv.stor_base, in kvm_s390_pv_dealloc_vm()
111 memset(&kvm->arch.pv, 0, sizeof(kvm->arch.pv)); in kvm_s390_pv_dealloc_vm()
114 static int kvm_s390_pv_alloc_vm(struct kvm *kvm) in kvm_s390_pv_alloc_vm() argument
121 kvm->arch.pv.stor_var = NULL; in kvm_s390_pv_alloc_vm()
[all …]
Dkvm-s390.h54 debug_sprintf_event(d_vcpu->kvm->arch.dbf, d_loglevel, \
82 return test_bit(kvm_vcpu_get_idx(vcpu), vcpu->kvm->arch.idle_mask); in is_vcpu_idle()
85 static inline int kvm_is_ucontrol(struct kvm *kvm) in kvm_is_ucontrol() argument
88 if (kvm->arch.gmap) in kvm_is_ucontrol()
182 static inline int test_kvm_facility(struct kvm *kvm, unsigned long nr) in test_kvm_facility() argument
184 return __test_facility(nr, kvm->arch.model.fac_mask) && in test_kvm_facility()
185 __test_facility(nr, kvm->arch.model.fac_list); in test_kvm_facility()
199 static inline int test_kvm_cpu_feat(struct kvm *kvm, unsigned long nr) in test_kvm_cpu_feat() argument
202 return test_bit_inv(nr, kvm->arch.cpu_feat); in test_kvm_cpu_feat()
206 static inline int kvm_s390_user_cpu_state_ctrl(struct kvm *kvm) in kvm_s390_user_cpu_state_ctrl() argument
[all …]
/arch/arm64/include/asm/
Dstage2_pgtable.h31 #define kvm_stage2_levels(kvm) VTCR_EL2_LVLS(kvm->arch.vtcr) argument
34 #define stage2_pgdir_shift(kvm) pt_levels_pgdir_shift(kvm_stage2_levels(kvm)) argument
35 #define stage2_pgdir_size(kvm) (1ULL << stage2_pgdir_shift(kvm)) argument
36 #define stage2_pgdir_mask(kvm) ~(stage2_pgdir_size(kvm) - 1) argument
43 #define kvm_mmu_cache_min_pages(kvm) (kvm_stage2_levels(kvm) - 1) argument
46 stage2_pgd_addr_end(struct kvm *kvm, phys_addr_t addr, phys_addr_t end) in stage2_pgd_addr_end() argument
48 phys_addr_t boundary = (addr + stage2_pgdir_size(kvm)) & stage2_pgdir_mask(kvm); in stage2_pgd_addr_end()
/arch/x86/include/asm/
Dkvm_page_track.h45 void (*track_flush_slot)(struct kvm *kvm, struct kvm_memory_slot *slot,
49 int kvm_page_track_init(struct kvm *kvm);
50 void kvm_page_track_cleanup(struct kvm *kvm);
56 void kvm_slot_page_track_add_page(struct kvm *kvm,
59 void kvm_slot_page_track_remove_page(struct kvm *kvm,
66 kvm_page_track_register_notifier(struct kvm *kvm,
69 kvm_page_track_unregister_notifier(struct kvm *kvm,
73 void kvm_page_track_flush_slot(struct kvm *kvm, struct kvm_memory_slot *slot);
/arch/arm64/kvm/
Dmmu.c42 static int stage2_apply_range(struct kvm *kvm, phys_addr_t addr, in stage2_apply_range() argument
51 struct kvm_pgtable *pgt = kvm->arch.mmu.pgt; in stage2_apply_range()
55 next = stage2_pgd_addr_end(kvm, addr, end); in stage2_apply_range()
61 cond_resched_lock(&kvm->mmu_lock); in stage2_apply_range()
67 #define stage2_apply_range_resched(kvm, addr, end, fn) \ argument
68 stage2_apply_range(kvm, addr, end, fn, true)
81 void kvm_flush_remote_tlbs(struct kvm *kvm) in kvm_flush_remote_tlbs() argument
83 kvm_call_hyp(__kvm_tlb_flush_vmid, &kvm->arch.mmu); in kvm_flush_remote_tlbs()
168 struct kvm *kvm = kvm_s2_mmu_to_kvm(mmu); in __unmap_stage2_range() local
171 assert_spin_locked(&kvm->mmu_lock); in __unmap_stage2_range()
[all …]
/arch/x86/kvm/svm/
Dsev.c108 static int sev_get_asid(struct kvm *kvm) in sev_get_asid() argument
110 struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info; in sev_get_asid()
150 static void sev_unbind_asid(struct kvm *kvm, unsigned int handle) in sev_unbind_asid() argument
174 static int sev_guest_init(struct kvm *kvm, struct kvm_sev_cmd *argp) in sev_guest_init() argument
176 struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info; in sev_guest_init()
179 if (kvm->created_vcpus) in sev_guest_init()
205 static int sev_bind_asid(struct kvm *kvm, unsigned int handle, int *error) in sev_bind_asid() argument
208 int asid = sev_get_asid(kvm); in sev_bind_asid()
239 static int sev_issue_cmd(struct kvm *kvm, int id, void *data, int *error) in sev_issue_cmd() argument
241 struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info; in sev_issue_cmd()
[all …]

1234567