Merge tag 'hid-for-linus-2024020101' of git://git.kernel.org/pub/scm/linux/kernel...
[sfrench/cifs-2.6.git] / tools / testing / selftests / kvm / x86_64 / hyperv_ipi.c
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Hyper-V HvCallSendSyntheticClusterIpi{,Ex} tests
4  *
5  * Copyright (C) 2022, Red Hat, Inc.
6  *
7  */
8
9 #define _GNU_SOURCE /* for program_invocation_short_name */
10 #include <pthread.h>
11 #include <inttypes.h>
12
13 #include "kvm_util.h"
14 #include "hyperv.h"
15 #include "test_util.h"
16 #include "vmx.h"
17
18 #define RECEIVER_VCPU_ID_1 2
19 #define RECEIVER_VCPU_ID_2 65
20
21 #define IPI_VECTOR       0xfe
22
23 static volatile uint64_t ipis_rcvd[RECEIVER_VCPU_ID_2 + 1];
24
25 struct hv_vpset {
26         u64 format;
27         u64 valid_bank_mask;
28         u64 bank_contents[2];
29 };
30
31 enum HV_GENERIC_SET_FORMAT {
32         HV_GENERIC_SET_SPARSE_4K,
33         HV_GENERIC_SET_ALL,
34 };
35
36 /* HvCallSendSyntheticClusterIpi hypercall */
37 struct hv_send_ipi {
38         u32 vector;
39         u32 reserved;
40         u64 cpu_mask;
41 };
42
43 /* HvCallSendSyntheticClusterIpiEx hypercall */
44 struct hv_send_ipi_ex {
45         u32 vector;
46         u32 reserved;
47         struct hv_vpset vp_set;
48 };
49
50 static inline void hv_init(vm_vaddr_t pgs_gpa)
51 {
52         wrmsr(HV_X64_MSR_GUEST_OS_ID, HYPERV_LINUX_OS_ID);
53         wrmsr(HV_X64_MSR_HYPERCALL, pgs_gpa);
54 }
55
56 static void receiver_code(void *hcall_page, vm_vaddr_t pgs_gpa)
57 {
58         u32 vcpu_id;
59
60         x2apic_enable();
61         hv_init(pgs_gpa);
62
63         vcpu_id = rdmsr(HV_X64_MSR_VP_INDEX);
64
65         /* Signal sender vCPU we're ready */
66         ipis_rcvd[vcpu_id] = (u64)-1;
67
68         for (;;)
69                 asm volatile("sti; hlt; cli");
70 }
71
72 static void guest_ipi_handler(struct ex_regs *regs)
73 {
74         u32 vcpu_id = rdmsr(HV_X64_MSR_VP_INDEX);
75
76         ipis_rcvd[vcpu_id]++;
77         wrmsr(HV_X64_MSR_EOI, 1);
78 }
79
80 static inline void nop_loop(void)
81 {
82         int i;
83
84         for (i = 0; i < 100000000; i++)
85                 asm volatile("nop");
86 }
87
88 static void sender_guest_code(void *hcall_page, vm_vaddr_t pgs_gpa)
89 {
90         struct hv_send_ipi *ipi = (struct hv_send_ipi *)hcall_page;
91         struct hv_send_ipi_ex *ipi_ex = (struct hv_send_ipi_ex *)hcall_page;
92         int stage = 1, ipis_expected[2] = {0};
93
94         hv_init(pgs_gpa);
95         GUEST_SYNC(stage++);
96
97         /* Wait for receiver vCPUs to come up */
98         while (!ipis_rcvd[RECEIVER_VCPU_ID_1] || !ipis_rcvd[RECEIVER_VCPU_ID_2])
99                 nop_loop();
100         ipis_rcvd[RECEIVER_VCPU_ID_1] = ipis_rcvd[RECEIVER_VCPU_ID_2] = 0;
101
102         /* 'Slow' HvCallSendSyntheticClusterIpi to RECEIVER_VCPU_ID_1 */
103         ipi->vector = IPI_VECTOR;
104         ipi->cpu_mask = 1 << RECEIVER_VCPU_ID_1;
105         hyperv_hypercall(HVCALL_SEND_IPI, pgs_gpa, pgs_gpa + 4096);
106         nop_loop();
107         GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_1] == ++ipis_expected[0]);
108         GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_2] == ipis_expected[1]);
109         GUEST_SYNC(stage++);
110         /* 'Fast' HvCallSendSyntheticClusterIpi to RECEIVER_VCPU_ID_1 */
111         hyperv_hypercall(HVCALL_SEND_IPI | HV_HYPERCALL_FAST_BIT,
112                          IPI_VECTOR, 1 << RECEIVER_VCPU_ID_1);
113         nop_loop();
114         GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_1] == ++ipis_expected[0]);
115         GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_2] == ipis_expected[1]);
116         GUEST_SYNC(stage++);
117
118         /* 'Slow' HvCallSendSyntheticClusterIpiEx to RECEIVER_VCPU_ID_1 */
119         memset(hcall_page, 0, 4096);
120         ipi_ex->vector = IPI_VECTOR;
121         ipi_ex->vp_set.format = HV_GENERIC_SET_SPARSE_4K;
122         ipi_ex->vp_set.valid_bank_mask = 1 << 0;
123         ipi_ex->vp_set.bank_contents[0] = BIT(RECEIVER_VCPU_ID_1);
124         hyperv_hypercall(HVCALL_SEND_IPI_EX | (1 << HV_HYPERCALL_VARHEAD_OFFSET),
125                          pgs_gpa, pgs_gpa + 4096);
126         nop_loop();
127         GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_1] == ++ipis_expected[0]);
128         GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_2] == ipis_expected[1]);
129         GUEST_SYNC(stage++);
130         /* 'XMM Fast' HvCallSendSyntheticClusterIpiEx to RECEIVER_VCPU_ID_1 */
131         hyperv_write_xmm_input(&ipi_ex->vp_set.valid_bank_mask, 1);
132         hyperv_hypercall(HVCALL_SEND_IPI_EX | HV_HYPERCALL_FAST_BIT |
133                          (1 << HV_HYPERCALL_VARHEAD_OFFSET),
134                          IPI_VECTOR, HV_GENERIC_SET_SPARSE_4K);
135         nop_loop();
136         GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_1] == ++ipis_expected[0]);
137         GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_2] == ipis_expected[1]);
138         GUEST_SYNC(stage++);
139
140         /* 'Slow' HvCallSendSyntheticClusterIpiEx to RECEIVER_VCPU_ID_2 */
141         memset(hcall_page, 0, 4096);
142         ipi_ex->vector = IPI_VECTOR;
143         ipi_ex->vp_set.format = HV_GENERIC_SET_SPARSE_4K;
144         ipi_ex->vp_set.valid_bank_mask = 1 << 1;
145         ipi_ex->vp_set.bank_contents[0] = BIT(RECEIVER_VCPU_ID_2 - 64);
146         hyperv_hypercall(HVCALL_SEND_IPI_EX | (1 << HV_HYPERCALL_VARHEAD_OFFSET),
147                          pgs_gpa, pgs_gpa + 4096);
148         nop_loop();
149         GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_1] == ipis_expected[0]);
150         GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_2] == ++ipis_expected[1]);
151         GUEST_SYNC(stage++);
152         /* 'XMM Fast' HvCallSendSyntheticClusterIpiEx to RECEIVER_VCPU_ID_2 */
153         hyperv_write_xmm_input(&ipi_ex->vp_set.valid_bank_mask, 1);
154         hyperv_hypercall(HVCALL_SEND_IPI_EX | HV_HYPERCALL_FAST_BIT |
155                          (1 << HV_HYPERCALL_VARHEAD_OFFSET),
156                          IPI_VECTOR, HV_GENERIC_SET_SPARSE_4K);
157         nop_loop();
158         GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_1] == ipis_expected[0]);
159         GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_2] == ++ipis_expected[1]);
160         GUEST_SYNC(stage++);
161
162         /* 'Slow' HvCallSendSyntheticClusterIpiEx to both RECEIVER_VCPU_ID_{1,2} */
163         memset(hcall_page, 0, 4096);
164         ipi_ex->vector = IPI_VECTOR;
165         ipi_ex->vp_set.format = HV_GENERIC_SET_SPARSE_4K;
166         ipi_ex->vp_set.valid_bank_mask = 1 << 1 | 1;
167         ipi_ex->vp_set.bank_contents[0] = BIT(RECEIVER_VCPU_ID_1);
168         ipi_ex->vp_set.bank_contents[1] = BIT(RECEIVER_VCPU_ID_2 - 64);
169         hyperv_hypercall(HVCALL_SEND_IPI_EX | (2 << HV_HYPERCALL_VARHEAD_OFFSET),
170                          pgs_gpa, pgs_gpa + 4096);
171         nop_loop();
172         GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_1] == ++ipis_expected[0]);
173         GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_2] == ++ipis_expected[1]);
174         GUEST_SYNC(stage++);
175         /* 'XMM Fast' HvCallSendSyntheticClusterIpiEx to both RECEIVER_VCPU_ID_{1, 2} */
176         hyperv_write_xmm_input(&ipi_ex->vp_set.valid_bank_mask, 2);
177         hyperv_hypercall(HVCALL_SEND_IPI_EX | HV_HYPERCALL_FAST_BIT |
178                          (2 << HV_HYPERCALL_VARHEAD_OFFSET),
179                          IPI_VECTOR, HV_GENERIC_SET_SPARSE_4K);
180         nop_loop();
181         GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_1] == ++ipis_expected[0]);
182         GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_2] == ++ipis_expected[1]);
183         GUEST_SYNC(stage++);
184
185         /* 'Slow' HvCallSendSyntheticClusterIpiEx to HV_GENERIC_SET_ALL */
186         memset(hcall_page, 0, 4096);
187         ipi_ex->vector = IPI_VECTOR;
188         ipi_ex->vp_set.format = HV_GENERIC_SET_ALL;
189         hyperv_hypercall(HVCALL_SEND_IPI_EX, pgs_gpa, pgs_gpa + 4096);
190         nop_loop();
191         GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_1] == ++ipis_expected[0]);
192         GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_2] == ++ipis_expected[1]);
193         GUEST_SYNC(stage++);
194         /*
195          * 'XMM Fast' HvCallSendSyntheticClusterIpiEx to HV_GENERIC_SET_ALL.
196          */
197         ipi_ex->vp_set.valid_bank_mask = 0;
198         hyperv_write_xmm_input(&ipi_ex->vp_set.valid_bank_mask, 2);
199         hyperv_hypercall(HVCALL_SEND_IPI_EX | HV_HYPERCALL_FAST_BIT,
200                          IPI_VECTOR, HV_GENERIC_SET_ALL);
201         nop_loop();
202         GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_1] == ++ipis_expected[0]);
203         GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_2] == ++ipis_expected[1]);
204         GUEST_SYNC(stage++);
205
206         GUEST_DONE();
207 }
208
209 static void *vcpu_thread(void *arg)
210 {
211         struct kvm_vcpu *vcpu = (struct kvm_vcpu *)arg;
212         int old, r;
213
214         r = pthread_setcanceltype(PTHREAD_CANCEL_ASYNCHRONOUS, &old);
215         TEST_ASSERT(!r, "pthread_setcanceltype failed on vcpu_id=%u with errno=%d",
216                     vcpu->id, r);
217
218         vcpu_run(vcpu);
219
220         TEST_FAIL("vCPU %u exited unexpectedly", vcpu->id);
221
222         return NULL;
223 }
224
225 static void cancel_join_vcpu_thread(pthread_t thread, struct kvm_vcpu *vcpu)
226 {
227         void *retval;
228         int r;
229
230         r = pthread_cancel(thread);
231         TEST_ASSERT(!r, "pthread_cancel on vcpu_id=%d failed with errno=%d",
232                     vcpu->id, r);
233
234         r = pthread_join(thread, &retval);
235         TEST_ASSERT(!r, "pthread_join on vcpu_id=%d failed with errno=%d",
236                     vcpu->id, r);
237         TEST_ASSERT(retval == PTHREAD_CANCELED,
238                     "expected retval=%p, got %p", PTHREAD_CANCELED,
239                     retval);
240 }
241
242 int main(int argc, char *argv[])
243 {
244         struct kvm_vm *vm;
245         struct kvm_vcpu *vcpu[3];
246         vm_vaddr_t hcall_page;
247         pthread_t threads[2];
248         int stage = 1, r;
249         struct ucall uc;
250
251         TEST_REQUIRE(kvm_has_cap(KVM_CAP_HYPERV_SEND_IPI));
252
253         vm = vm_create_with_one_vcpu(&vcpu[0], sender_guest_code);
254
255         /* Hypercall input/output */
256         hcall_page = vm_vaddr_alloc_pages(vm, 2);
257         memset(addr_gva2hva(vm, hcall_page), 0x0, 2 * getpagesize());
258
259         vm_init_descriptor_tables(vm);
260
261         vcpu[1] = vm_vcpu_add(vm, RECEIVER_VCPU_ID_1, receiver_code);
262         vcpu_init_descriptor_tables(vcpu[1]);
263         vcpu_args_set(vcpu[1], 2, hcall_page, addr_gva2gpa(vm, hcall_page));
264         vcpu_set_msr(vcpu[1], HV_X64_MSR_VP_INDEX, RECEIVER_VCPU_ID_1);
265         vcpu_set_hv_cpuid(vcpu[1]);
266
267         vcpu[2] = vm_vcpu_add(vm, RECEIVER_VCPU_ID_2, receiver_code);
268         vcpu_init_descriptor_tables(vcpu[2]);
269         vcpu_args_set(vcpu[2], 2, hcall_page, addr_gva2gpa(vm, hcall_page));
270         vcpu_set_msr(vcpu[2], HV_X64_MSR_VP_INDEX, RECEIVER_VCPU_ID_2);
271         vcpu_set_hv_cpuid(vcpu[2]);
272
273         vm_install_exception_handler(vm, IPI_VECTOR, guest_ipi_handler);
274
275         vcpu_args_set(vcpu[0], 2, hcall_page, addr_gva2gpa(vm, hcall_page));
276         vcpu_set_hv_cpuid(vcpu[0]);
277
278         r = pthread_create(&threads[0], NULL, vcpu_thread, vcpu[1]);
279         TEST_ASSERT(!r, "pthread_create failed errno=%d", r);
280
281         r = pthread_create(&threads[1], NULL, vcpu_thread, vcpu[2]);
282         TEST_ASSERT(!r, "pthread_create failed errno=%d", errno);
283
284         while (true) {
285                 vcpu_run(vcpu[0]);
286
287                 TEST_ASSERT_KVM_EXIT_REASON(vcpu[0], KVM_EXIT_IO);
288
289                 switch (get_ucall(vcpu[0], &uc)) {
290                 case UCALL_SYNC:
291                         TEST_ASSERT(uc.args[1] == stage,
292                                     "Unexpected stage: %ld (%d expected)\n",
293                                     uc.args[1], stage);
294                         break;
295                 case UCALL_DONE:
296                         goto done;
297                 case UCALL_ABORT:
298                         REPORT_GUEST_ASSERT(uc);
299                         /* NOT REACHED */
300                 default:
301                         TEST_FAIL("Unknown ucall %lu", uc.cmd);
302                 }
303
304                 stage++;
305         }
306
307 done:
308         cancel_join_vcpu_thread(threads[0], vcpu[1]);
309         cancel_join_vcpu_thread(threads[1], vcpu[2]);
310         kvm_vm_free(vm);
311
312         return r;
313 }