Home
last modified time | relevance | path

Searched refs:vmx_pages (Results 1 – 9 of 9) sorted by relevance

/tools/testing/selftests/kvm/x86_64/
Devmcs_test.c32 void l1_guest_code(struct vmx_pages *vmx_pages) in l1_guest_code() argument
37 enable_vp_assist(vmx_pages->vp_assist_gpa, vmx_pages->vp_assist); in l1_guest_code()
39 GUEST_ASSERT(vmx_pages->vmcs_gpa); in l1_guest_code()
40 GUEST_ASSERT(prepare_for_vmx_operation(vmx_pages)); in l1_guest_code()
42 GUEST_ASSERT(load_vmcs(vmx_pages)); in l1_guest_code()
43 GUEST_ASSERT(vmptrstz() == vmx_pages->enlightened_vmcs_gpa); in l1_guest_code()
46 GUEST_ASSERT(vmptrstz() == vmx_pages->enlightened_vmcs_gpa); in l1_guest_code()
48 prepare_vmcs(vmx_pages, l2_guest_code, in l1_guest_code()
52 GUEST_ASSERT(vmptrstz() == vmx_pages->enlightened_vmcs_gpa); in l1_guest_code()
59 GUEST_ASSERT(vmptrstz() == vmx_pages->enlightened_vmcs_gpa); in l1_guest_code()
[all …]
Dstate_test.c77 static void vmx_l1_guest_code(struct vmx_pages *vmx_pages) in vmx_l1_guest_code() argument
81 GUEST_ASSERT(vmx_pages->vmcs_gpa); in vmx_l1_guest_code()
82 GUEST_ASSERT(prepare_for_vmx_operation(vmx_pages)); in vmx_l1_guest_code()
84 GUEST_ASSERT(load_vmcs(vmx_pages)); in vmx_l1_guest_code()
85 GUEST_ASSERT(vmptrstz() == vmx_pages->vmcs_gpa); in vmx_l1_guest_code()
88 GUEST_ASSERT(vmptrstz() == vmx_pages->vmcs_gpa); in vmx_l1_guest_code()
90 prepare_vmcs(vmx_pages, vmx_l2_guest_code, in vmx_l1_guest_code()
94 GUEST_ASSERT(vmptrstz() == vmx_pages->vmcs_gpa); in vmx_l1_guest_code()
96 GUEST_ASSERT(vmptrstz() == vmx_pages->vmcs_gpa); in vmx_l1_guest_code()
114 vmwrite(VMCS_LINK_POINTER, vmx_pages->shadow_vmcs_gpa); in vmx_l1_guest_code()
[all …]
Dvmx_close_while_nested_test.c37 static void l1_guest_code(struct vmx_pages *vmx_pages) in l1_guest_code() argument
42 GUEST_ASSERT(prepare_for_vmx_operation(vmx_pages)); in l1_guest_code()
43 GUEST_ASSERT(load_vmcs(vmx_pages)); in l1_guest_code()
46 prepare_vmcs(vmx_pages, l2_guest_code, in l1_guest_code()
Dvmx_preemption_timer_test.c69 void l1_guest_code(struct vmx_pages *vmx_pages) in l1_guest_code() argument
77 GUEST_ASSERT(vmx_pages->vmcs_gpa); in l1_guest_code()
78 GUEST_ASSERT(prepare_for_vmx_operation(vmx_pages)); in l1_guest_code()
79 GUEST_ASSERT(load_vmcs(vmx_pages)); in l1_guest_code()
80 GUEST_ASSERT(vmptrstz() == vmx_pages->vmcs_gpa); in l1_guest_code()
82 prepare_vmcs(vmx_pages, l2_guest_code, in l1_guest_code()
147 void guest_code(struct vmx_pages *vmx_pages) in guest_code() argument
149 if (vmx_pages) in guest_code()
150 l1_guest_code(vmx_pages); in guest_code()
Dvmx_apic_access_test.c42 static void l1_guest_code(struct vmx_pages *vmx_pages, unsigned long high_gpa) in l1_guest_code() argument
48 GUEST_ASSERT(prepare_for_vmx_operation(vmx_pages)); in l1_guest_code()
49 GUEST_ASSERT(load_vmcs(vmx_pages)); in l1_guest_code()
52 prepare_vmcs(vmx_pages, l2_guest_code, in l1_guest_code()
60 vmwrite(APIC_ACCESS_ADDR, vmx_pages->apic_access_gpa); in l1_guest_code()
84 struct vmx_pages *vmx; in main()
Dvmx_tsc_adjust_test.c83 static void l1_guest_code(struct vmx_pages *vmx_pages) in l1_guest_code() argument
94 GUEST_ASSERT(prepare_for_vmx_operation(vmx_pages)); in l1_guest_code()
95 GUEST_ASSERT(load_vmcs(vmx_pages)); in l1_guest_code()
98 prepare_vmcs(vmx_pages, l2_guest_code, in l1_guest_code()
Dvmx_dirty_log_test.c50 void l1_guest_code(struct vmx_pages *vmx) in l1_guest_code()
72 struct vmx_pages *vmx; in main()
/tools/testing/selftests/kvm/include/x86_64/
Dvmx.h540 struct vmx_pages { struct
605 struct vmx_pages *vcpu_alloc_vmx(struct kvm_vm *vm, vm_vaddr_t *p_vmx_gva);
606 bool prepare_for_vmx_operation(struct vmx_pages *vmx);
607 void prepare_vmcs(struct vmx_pages *vmx, void *guest_rip, void *guest_rsp);
608 bool load_vmcs(struct vmx_pages *vmx);
613 void nested_pg_map(struct vmx_pages *vmx, struct kvm_vm *vm,
615 void nested_map(struct vmx_pages *vmx, struct kvm_vm *vm,
618 void nested_map_memslot(struct vmx_pages *vmx, struct kvm_vm *vm,
620 void prepare_eptp(struct vmx_pages *vmx, struct kvm_vm *vm,
622 void prepare_virtualize_apic_accesses(struct vmx_pages *vmx, struct kvm_vm *vm,
/tools/testing/selftests/kvm/lib/x86_64/
Dvmx.c77 struct vmx_pages *
81 struct vmx_pages *vmx = addr_gva2hva(vm, vmx_gva); in vcpu_alloc_vmx()
133 bool prepare_for_vmx_operation(struct vmx_pages *vmx) in prepare_for_vmx_operation()
177 bool load_vmcs(struct vmx_pages *vmx) in load_vmcs()
206 static inline void init_vmcs_control_fields(struct vmx_pages *vmx) in init_vmcs_control_fields()
375 void prepare_vmcs(struct vmx_pages *vmx, void *guest_rip, void *guest_rsp) in prepare_vmcs()
397 void nested_pg_map(struct vmx_pages *vmx, struct kvm_vm *vm, in nested_pg_map()
496 void nested_map(struct vmx_pages *vmx, struct kvm_vm *vm, in nested_map()
516 void nested_map_memslot(struct vmx_pages *vmx, struct kvm_vm *vm, in nested_map_memslot()
538 void prepare_eptp(struct vmx_pages *vmx, struct kvm_vm *vm, in prepare_eptp()
[all …]