1 // SPDX-License-Identifier: GPL-2.0
2 /*
3 * Tests for MSR_IA32_TSC and MSR_IA32_TSC_ADJUST.
4 *
5 * Copyright (C) 2020, Red Hat, Inc.
6 */
7 #include <stdio.h>
8 #include <string.h>
9 #include "kvm_util.h"
10 #include "processor.h"
11
12 #define UNITY (1ull << 30)
13 #define HOST_ADJUST (UNITY * 64)
14 #define GUEST_STEP (UNITY * 4)
15 #define ROUND(x) ((x + UNITY / 2) & -UNITY)
16 #define rounded_rdmsr(x) ROUND(rdmsr(x))
17 #define rounded_host_rdmsr(x) ROUND(vcpu_get_msr(vcpu, x))
18
guest_code(void)19 static void guest_code(void)
20 {
21 u64 val = 0;
22
23 GUEST_ASSERT_EQ(rounded_rdmsr(MSR_IA32_TSC), val);
24 GUEST_ASSERT_EQ(rounded_rdmsr(MSR_IA32_TSC_ADJUST), val);
25
26 /* Guest: writes to MSR_IA32_TSC affect both MSRs. */
27 val = 1ull * GUEST_STEP;
28 wrmsr(MSR_IA32_TSC, val);
29 GUEST_ASSERT_EQ(rounded_rdmsr(MSR_IA32_TSC), val);
30 GUEST_ASSERT_EQ(rounded_rdmsr(MSR_IA32_TSC_ADJUST), val);
31
32 /* Guest: writes to MSR_IA32_TSC_ADJUST affect both MSRs. */
33 GUEST_SYNC(2);
34 val = 2ull * GUEST_STEP;
35 wrmsr(MSR_IA32_TSC_ADJUST, val);
36 GUEST_ASSERT_EQ(rounded_rdmsr(MSR_IA32_TSC), val);
37 GUEST_ASSERT_EQ(rounded_rdmsr(MSR_IA32_TSC_ADJUST), val);
38
39 /* Host: setting the TSC offset. */
40 GUEST_SYNC(3);
41 GUEST_ASSERT_EQ(rounded_rdmsr(MSR_IA32_TSC), HOST_ADJUST + val);
42 GUEST_ASSERT_EQ(rounded_rdmsr(MSR_IA32_TSC_ADJUST), val);
43
44 /*
45 * Guest: writes to MSR_IA32_TSC_ADJUST do not destroy the
46 * host-side offset and affect both MSRs.
47 */
48 GUEST_SYNC(4);
49 val = 3ull * GUEST_STEP;
50 wrmsr(MSR_IA32_TSC_ADJUST, val);
51 GUEST_ASSERT_EQ(rounded_rdmsr(MSR_IA32_TSC), HOST_ADJUST + val);
52 GUEST_ASSERT_EQ(rounded_rdmsr(MSR_IA32_TSC_ADJUST), val);
53
54 /*
55 * Guest: writes to MSR_IA32_TSC affect both MSRs, so the host-side
56 * offset is now visible in MSR_IA32_TSC_ADJUST.
57 */
58 GUEST_SYNC(5);
59 val = 4ull * GUEST_STEP;
60 wrmsr(MSR_IA32_TSC, val);
61 GUEST_ASSERT_EQ(rounded_rdmsr(MSR_IA32_TSC), val);
62 GUEST_ASSERT_EQ(rounded_rdmsr(MSR_IA32_TSC_ADJUST), val - HOST_ADJUST);
63
64 GUEST_DONE();
65 }
66
run_vcpu(struct kvm_vcpu * vcpu,int stage)67 static void run_vcpu(struct kvm_vcpu *vcpu, int stage)
68 {
69 struct ucall uc;
70
71 vcpu_run(vcpu);
72
73 switch (get_ucall(vcpu, &uc)) {
74 case UCALL_SYNC:
75 if (!strcmp((const char *)uc.args[0], "hello") &&
76 uc.args[1] == stage + 1)
77 ksft_test_result_pass("stage %d passed\n", stage + 1);
78 else
79 ksft_test_result_fail(
80 "stage %d: Unexpected register values vmexit, got %lx",
81 stage + 1, (ulong)uc.args[1]);
82 return;
83 case UCALL_DONE:
84 ksft_test_result_pass("stage %d passed\n", stage + 1);
85 return;
86 case UCALL_ABORT:
87 REPORT_GUEST_ASSERT_2(uc, "values: %#lx, %#lx");
88 default:
89 TEST_ASSERT(false, "Unexpected exit: %s",
90 exit_reason_str(vcpu->run->exit_reason));
91 }
92 }
93
main(void)94 int main(void)
95 {
96 struct kvm_vcpu *vcpu;
97 struct kvm_vm *vm;
98 uint64_t val;
99
100 ksft_print_header();
101 ksft_set_plan(5);
102
103 vm = vm_create_with_one_vcpu(&vcpu, guest_code);
104
105 val = 0;
106 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC), val);
107 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC_ADJUST), val);
108
109 /* Guest: writes to MSR_IA32_TSC affect both MSRs. */
110 run_vcpu(vcpu, 1);
111 val = 1ull * GUEST_STEP;
112 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC), val);
113 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC_ADJUST), val);
114
115 /* Guest: writes to MSR_IA32_TSC_ADJUST affect both MSRs. */
116 run_vcpu(vcpu, 2);
117 val = 2ull * GUEST_STEP;
118 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC), val);
119 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC_ADJUST), val);
120
121 /*
122 * Host: writes to MSR_IA32_TSC set the host-side offset
123 * and therefore do not change MSR_IA32_TSC_ADJUST.
124 */
125 vcpu_set_msr(vcpu, MSR_IA32_TSC, HOST_ADJUST + val);
126 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC), HOST_ADJUST + val);
127 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC_ADJUST), val);
128 run_vcpu(vcpu, 3);
129
130 /* Host: writes to MSR_IA32_TSC_ADJUST do not modify the TSC. */
131 vcpu_set_msr(vcpu, MSR_IA32_TSC_ADJUST, UNITY * 123456);
132 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC), HOST_ADJUST + val);
133 ASSERT_EQ(vcpu_get_msr(vcpu, MSR_IA32_TSC_ADJUST), UNITY * 123456);
134
135 /* Restore previous value. */
136 vcpu_set_msr(vcpu, MSR_IA32_TSC_ADJUST, val);
137 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC), HOST_ADJUST + val);
138 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC_ADJUST), val);
139
140 /*
141 * Guest: writes to MSR_IA32_TSC_ADJUST do not destroy the
142 * host-side offset and affect both MSRs.
143 */
144 run_vcpu(vcpu, 4);
145 val = 3ull * GUEST_STEP;
146 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC), HOST_ADJUST + val);
147 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC_ADJUST), val);
148
149 /*
150 * Guest: writes to MSR_IA32_TSC affect both MSRs, so the host-side
151 * offset is now visible in MSR_IA32_TSC_ADJUST.
152 */
153 run_vcpu(vcpu, 5);
154 val = 4ull * GUEST_STEP;
155 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC), val);
156 ASSERT_EQ(rounded_host_rdmsr(MSR_IA32_TSC_ADJUST), val - HOST_ADJUST);
157
158 kvm_vm_free(vm);
159
160 ksft_finished(); /* Print results and exit() accordingly */
161 }
162