Home
last modified time | relevance | path

Searched refs:vmx (Results 1 – 14 of 14) sorted by relevance

/tools/testing/selftests/kvm/lib/x86_64/
Dvmx.c81 struct vmx_pages *vmx = addr_gva2hva(vm, vmx_gva); in vcpu_alloc_vmx() local
84 vmx->vmxon = (void *)vm_vaddr_alloc(vm, getpagesize(), 0x10000, 0, 0); in vcpu_alloc_vmx()
85 vmx->vmxon_hva = addr_gva2hva(vm, (uintptr_t)vmx->vmxon); in vcpu_alloc_vmx()
86 vmx->vmxon_gpa = addr_gva2gpa(vm, (uintptr_t)vmx->vmxon); in vcpu_alloc_vmx()
89 vmx->vmcs = (void *)vm_vaddr_alloc(vm, getpagesize(), 0x10000, 0, 0); in vcpu_alloc_vmx()
90 vmx->vmcs_hva = addr_gva2hva(vm, (uintptr_t)vmx->vmcs); in vcpu_alloc_vmx()
91 vmx->vmcs_gpa = addr_gva2gpa(vm, (uintptr_t)vmx->vmcs); in vcpu_alloc_vmx()
94 vmx->msr = (void *)vm_vaddr_alloc(vm, getpagesize(), 0x10000, 0, 0); in vcpu_alloc_vmx()
95 vmx->msr_hva = addr_gva2hva(vm, (uintptr_t)vmx->msr); in vcpu_alloc_vmx()
96 vmx->msr_gpa = addr_gva2gpa(vm, (uintptr_t)vmx->msr); in vcpu_alloc_vmx()
[all …]
/tools/testing/selftests/kvm/x86_64/
Dvmx_set_nested_state_test.c83 state->hdr.vmx.vmxon_pa = 0x1000; in set_default_vmx_state()
84 state->hdr.vmx.vmcs12_pa = 0x2000; in set_default_vmx_state()
85 state->hdr.vmx.smm.flags = 0; in set_default_vmx_state()
114 state->hdr.vmx.vmxon_pa = -1ull; in test_vmx_nested_state()
117 state->hdr.vmx.vmcs12_pa = -1ull; in test_vmx_nested_state()
134 state->hdr.vmx.vmxon_pa = -1ull; in test_vmx_nested_state()
135 state->hdr.vmx.vmcs12_pa = -1ull; in test_vmx_nested_state()
146 state->hdr.vmx.smm.flags = 1; in test_vmx_nested_state()
151 state->hdr.vmx.flags = ~0; in test_vmx_nested_state()
156 state->hdr.vmx.vmxon_pa = -1ull; in test_vmx_nested_state()
[all …]
Dvmx_dirty_log_test.c50 void l1_guest_code(struct vmx_pages *vmx) in l1_guest_code() argument
55 GUEST_ASSERT(vmx->vmcs_gpa); in l1_guest_code()
56 GUEST_ASSERT(prepare_for_vmx_operation(vmx)); in l1_guest_code()
57 GUEST_ASSERT(load_vmcs(vmx)); in l1_guest_code()
59 prepare_vmcs(vmx, l2_guest_code, in l1_guest_code()
72 struct vmx_pages *vmx; in main() local
86 vmx = vcpu_alloc_vmx(vm, &vmx_pages_gva); in main()
110 prepare_eptp(vmx, vm, 0); in main()
111 nested_map_memslot(vmx, vm, 0, 0); in main()
112 nested_map(vmx, vm, NESTED_TEST_MEM1, GUEST_TEST_MEM, 4096, 0); in main()
[all …]
Dvmx_apic_access_test.c84 struct vmx_pages *vmx; in main() local
99 vmx = vcpu_alloc_vmx(vm, &vmx_pages_gva); in main()
100 prepare_virtualize_apic_accesses(vmx, vm, 0); in main()
/tools/testing/selftests/powerpc/ptrace/
Dptrace-vsx.h31 int validate_vmx(unsigned long vmx[][2], unsigned long *load) in validate_vmx()
37 if ((vmx[i][0] != load[64 + 2 * i]) || in validate_vmx()
38 (vmx[i][1] != load[65 + 2 * i])) { in validate_vmx()
40 i, vmx[i][0], 64 + 2 * i, in validate_vmx()
43 i, vmx[i][1], 65 + 2 * i, in validate_vmx()
51 if ((vmx[i][0] != load[65 + 2 * i]) || in validate_vmx()
52 (vmx[i][1] != load[64 + 2 * i])) { in validate_vmx()
54 i, vmx[i][0], 65 + 2 * i, in validate_vmx()
57 i, vmx[i][1], 64 + 2 * i, in validate_vmx()
109 unsigned long vmx[][2]) in load_vsx_vmx()
[all …]
Dptrace-tm-vsx.c87 unsigned long vmx[VMX_MAX + 2][2]; in trace_tm_vsx() local
92 FAIL_IF(show_vmx(child, vmx)); in trace_tm_vsx()
93 FAIL_IF(validate_vmx(vmx, fp_load)); in trace_tm_vsx()
96 FAIL_IF(show_vmx_ckpt(child, vmx)); in trace_tm_vsx()
97 FAIL_IF(validate_vmx(vmx, fp_load_ckpt)); in trace_tm_vsx()
99 memset(vmx, 0, sizeof(vmx)); in trace_tm_vsx()
101 load_vsx_vmx(fp_load_ckpt_new, vsx, vmx); in trace_tm_vsx()
104 FAIL_IF(write_vmx_ckpt(child, vmx)); in trace_tm_vsx()
Dptrace-tm-spd-vsx.c99 unsigned long vmx[VMX_MAX + 2][2]; in trace_tm_spd_vsx() local
104 FAIL_IF(show_vmx(child, vmx)); in trace_tm_spd_vsx()
105 FAIL_IF(validate_vmx(vmx, fp_load)); in trace_tm_spd_vsx()
108 FAIL_IF(show_vmx_ckpt(child, vmx)); in trace_tm_spd_vsx()
109 FAIL_IF(validate_vmx(vmx, fp_load_ckpt)); in trace_tm_spd_vsx()
112 memset(vmx, 0, sizeof(vmx)); in trace_tm_spd_vsx()
114 load_vsx_vmx(fp_load_ckpt_new, vsx, vmx); in trace_tm_spd_vsx()
117 FAIL_IF(write_vmx_ckpt(child, vmx)); in trace_tm_spd_vsx()
Dptrace-vsx.c40 unsigned long vmx[VMX_MAX + 2][2]; in trace_vsx() local
45 FAIL_IF(show_vmx(child, vmx)); in trace_vsx()
46 FAIL_IF(validate_vmx(vmx, fp_load)); in trace_vsx()
49 memset(vmx, 0, sizeof(vmx)); in trace_vsx()
50 load_vsx_vmx(fp_load_new, vsx, vmx); in trace_vsx()
53 FAIL_IF(write_vmx(child, vmx)); in trace_vsx()
Dptrace.h529 int show_vmx(pid_t child, unsigned long vmx[][2]) in show_vmx()
533 ret = ptrace(PTRACE_GETVRREGS, child, 0, vmx); in show_vmx()
541 int show_vmx_ckpt(pid_t child, unsigned long vmx[][2]) in show_vmx_ckpt()
554 memcpy(vmx, regs, sizeof(regs)); in show_vmx_ckpt()
559 int write_vmx(pid_t child, unsigned long vmx[][2]) in write_vmx()
563 ret = ptrace(PTRACE_SETVRREGS, child, 0, vmx); in write_vmx()
571 int write_vmx_ckpt(pid_t child, unsigned long vmx[][2]) in write_vmx_ckpt()
577 memcpy(regs, vmx, sizeof(regs)); in write_vmx_ckpt()
/tools/testing/selftests/kvm/include/x86_64/
Dvmx.h606 bool prepare_for_vmx_operation(struct vmx_pages *vmx);
607 void prepare_vmcs(struct vmx_pages *vmx, void *guest_rip, void *guest_rsp);
608 bool load_vmcs(struct vmx_pages *vmx);
613 void nested_pg_map(struct vmx_pages *vmx, struct kvm_vm *vm,
615 void nested_map(struct vmx_pages *vmx, struct kvm_vm *vm,
618 void nested_map_memslot(struct vmx_pages *vmx, struct kvm_vm *vm,
620 void prepare_eptp(struct vmx_pages *vmx, struct kvm_vm *vm,
622 void prepare_virtualize_apic_accesses(struct vmx_pages *vmx, struct kvm_vm *vm,
/tools/testing/selftests/powerpc/tm/
DMakefile3 tm-signal-context-chk-vmx tm-signal-context-chk-vsx
6 tm-vmxcopy tm-fork tm-tar tm-tmspr tm-vmx-unavail tm-unavailable tm-trap \
22 $(OUTPUT)/tm-vmx-unavail: CFLAGS += -pthread -m64
D.gitignore13 tm-signal-context-chk-vmx
18 tm-vmx-unavail
/tools/arch/x86/include/uapi/asm/
Dkvm.h458 struct kvm_vmx_nested_state_hdr vmx; member
471 struct kvm_vmx_nested_state_data vmx[0]; member
/tools/testing/selftests/kvm/
DMakefile37 LIBKVM_x86_64 = lib/x86_64/processor.c lib/x86_64/vmx.c lib/x86_64/svm.c lib/x86_64/ucall.c lib/x86…