1 // SPDX-License-Identifier: GPL-2.0
2 /*
3 * Tests for MSR_IA32_TSC and MSR_IA32_TSC_ADJUST.
4 *
5 * Copyright (C) 2020, Red Hat, Inc.
6 */
7 #include <stdio.h>
8 #include <string.h>
9 #include "kvm_util.h"
10 #include "processor.h"
11
12 #define VCPU_ID 0
13
14 #define UNITY (1ull << 30)
15 #define HOST_ADJUST (UNITY * 64)
16 #define GUEST_STEP (UNITY * 4)
17 #define ROUND(x) ((x + UNITY / 2) & -UNITY)
18 #define rounded_rdmsr(x) ROUND(rdmsr(x))
19 #define rounded_host_rdmsr(x) ROUND(vcpu_get_msr(vm, 0, x))
20
21 #define GUEST_ASSERT_EQ(a, b) do { \
22 __typeof(a) _a = (a); \
23 __typeof(b) _b = (b); \
24 if (_a != _b) \
25 ucall(UCALL_ABORT, 4, \
26 "Failed guest assert: " \
27 #a " == " #b, __LINE__, _a, _b); \
28 } while(0)
29
guest_code(void)30 static void guest_code(void)
31 {
32 u64 val = 0;
33
34 GUEST_ASSERT_EQ(rounded_rdmsr(MSR_IA32_TSC), val);
35 GUEST_ASSERT_EQ(rounded_rdmsr(MSR_IA32_TSC_ADJUST), val);
36
37 /* Guest: writes to MSR_IA32_TSC affect both MSRs. */
38 val = 1ull * GUEST_STEP;
39 wrmsr(MSR_IA32_TSC, val);
40 GUEST_ASSERT_EQ(rounded_rdmsr(MSR_IA32_TSC), val);
41 GUEST_ASSERT_EQ(rounded_rdmsr(MSR_IA32_TSC_ADJUST), val);
42
43 /* Guest: writes to MSR_IA32_TSC_ADJUST affect both MSRs. */
44 GUEST_SYNC(2);
45 val = 2ull * GUEST_STEP;
46 wrmsr(MSR_IA32_TSC_ADJUST, val);
47 GUEST_ASSERT_EQ(rounded_rdmsr(MSR_IA32_TSC), val);
48 GUEST_ASSERT_EQ(rounded_rdmsr(MSR_IA32_TSC_ADJUST), val);
49
50 /* Host: setting the TSC offset. */
51 GUEST_SYNC(3);
52 GUEST_ASSERT_EQ(rounded_rdmsr(MSR_IA32_TSC), HOST_ADJUST + val);
53 GUEST_ASSERT_EQ(rounded_rdmsr(MSR_IA32_TSC_ADJUST), val);
54
55 /*
56 * Guest: writes to MSR_IA32_TSC_ADJUST do not destroy the
57 * host-side offset and affect both MSRs.
58 */
59 GUEST_SYNC(4);
60 val = 3ull * GUEST_STEP;
61 wrmsr(MSR_IA32_TSC_ADJUST, val);
62 GUEST_ASSERT_EQ(rounded_rdmsr(MSR_IA32_TSC), HOST_ADJUST + val);
63 GUEST_ASSERT_EQ(rounded_rdmsr(MSR_IA32_TSC_ADJUST), val);
64
65 /*
66 * Guest: writes to MSR_IA32_TSC affect both MSRs, so the host-side
67 * offset is now visible in MSR_IA32_TSC_ADJUST.
68 */
69 GUEST_SYNC(5);
70 val = 4ull * GUEST_STEP;
71 wrmsr(MSR_IA32_TSC, val);
72 GUEST_ASSERT_EQ(rounded_rdmsr(MSR_IA32_TSC), val);
73 GUEST_ASSERT_EQ(rounded_rdmsr(MSR_IA32_TSC_ADJUST), val - HOST_ADJUST);
74
75 GUEST_DONE();
76 }
77
run_vcpu(struct kvm_vm * vm,uint32_t vcpuid,int stage)78 static void run_vcpu(struct kvm_vm *vm, uint32_t vcpuid, int stage)
79 {
80 struct ucall uc;
81
82 vcpu_args_set(vm, vcpuid, 1, vcpuid);
83
84 vcpu_ioctl(vm, vcpuid, KVM_RUN, NULL);
85
86 switch (get_ucall(vm, vcpuid, &uc)) {
87 case UCALL_SYNC:
88 TEST_ASSERT(!strcmp((const char *)uc.args[0], "hello") &&
89 uc.args[1] == stage + 1, "Stage %d: Unexpected register values vmexit, got %lx",
90 stage + 1, (ulong)uc.args[1]);
91 return;
92 case UCALL_DONE:
93 return;
94 case UCALL_ABORT:
95 TEST_ASSERT(false, "%s at %s:%ld\n" \
96 "\tvalues: %#lx, %#lx", (const char *)uc.args[0],
97 __FILE__, uc.args[1], uc.args[2], uc.args[3]);
98 default:
99 TEST_ASSERT(false, "Unexpected exit: %s",
100 exit_reason_str(vcpu_state(vm, vcpuid)->exit_reason));
101 }
102 }
103
main(void)104 int main(void)
105 {
106 struct kvm_vm *vm;
107 uint64_t val;
108
109 vm = vm_create_default(VCPU_ID, 0, guest_code);
110 vcpu_set_cpuid(vm, VCPU_ID, kvm_get_supported_cpuid());
111
112 val = 0;
113 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC), val);
114 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC_ADJUST), val);
115
116 /* Guest: writes to MSR_IA32_TSC affect both MSRs. */
117 run_vcpu(vm, VCPU_ID, 1);
118 val = 1ull * GUEST_STEP;
119 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC), val);
120 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC_ADJUST), val);
121
122 /* Guest: writes to MSR_IA32_TSC_ADJUST affect both MSRs. */
123 run_vcpu(vm, VCPU_ID, 2);
124 val = 2ull * GUEST_STEP;
125 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC), val);
126 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC_ADJUST), val);
127
128 /*
129 * Host: writes to MSR_IA32_TSC set the host-side offset
130 * and therefore do not change MSR_IA32_TSC_ADJUST.
131 */
132 vcpu_set_msr(vm, 0, MSR_IA32_TSC, HOST_ADJUST + val);
133 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC), HOST_ADJUST + val);
134 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC_ADJUST), val);
135 run_vcpu(vm, VCPU_ID, 3);
136
137 /* Host: writes to MSR_IA32_TSC_ADJUST do not modify the TSC. */
138 vcpu_set_msr(vm, 0, MSR_IA32_TSC_ADJUST, UNITY * 123456);
139 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC), HOST_ADJUST + val);
140 ASSERT_EQ(vcpu_get_msr(vm, 0, MSR_IA32_TSC_ADJUST), UNITY * 123456);
141
142 /* Restore previous value. */
143 vcpu_set_msr(vm, 0, MSR_IA32_TSC_ADJUST, val);
144 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC), HOST_ADJUST + val);
145 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC_ADJUST), val);
146
147 /*
148 * Guest: writes to MSR_IA32_TSC_ADJUST do not destroy the
149 * host-side offset and affect both MSRs.
150 */
151 run_vcpu(vm, VCPU_ID, 4);
152 val = 3ull * GUEST_STEP;
153 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC), HOST_ADJUST + val);
154 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC_ADJUST), val);
155
156 /*
157 * Guest: writes to MSR_IA32_TSC affect both MSRs, so the host-side
158 * offset is now visible in MSR_IA32_TSC_ADJUST.
159 */
160 run_vcpu(vm, VCPU_ID, 5);
161 val = 4ull * GUEST_STEP;
162 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC), val);
163 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC_ADJUST), val - HOST_ADJUST);
164
165 kvm_vm_free(vm);
166
167 return 0;
168 }
169