5da884e11337a6d420e3dc71456b469057300d1c
[sfrench/cifs-2.6.git] / arch / arm64 / kvm / hypercalls.c
1 // SPDX-License-Identifier: GPL-2.0
2 // Copyright (C) 2019 Arm Ltd.
3
4 #include <linux/arm-smccc.h>
5 #include <linux/kvm_host.h>
6
7 #include <asm/kvm_emulate.h>
8
9 #include <kvm/arm_hypercalls.h>
10 #include <kvm/arm_psci.h>
11
12 #define KVM_ARM_SMCCC_STD_FEATURES                              \
13         GENMASK(KVM_REG_ARM_STD_BMAP_BIT_COUNT - 1, 0)
14 #define KVM_ARM_SMCCC_STD_HYP_FEATURES                          \
15         GENMASK(KVM_REG_ARM_STD_HYP_BMAP_BIT_COUNT - 1, 0)
16 #define KVM_ARM_SMCCC_VENDOR_HYP_FEATURES                       \
17         GENMASK(KVM_REG_ARM_VENDOR_HYP_BMAP_BIT_COUNT - 1, 0)
18
19 static void kvm_ptp_get_time(struct kvm_vcpu *vcpu, u64 *val)
20 {
21         struct system_time_snapshot systime_snapshot;
22         u64 cycles = ~0UL;
23         u32 feature;
24
25         /*
26          * system time and counter value must captured at the same
27          * time to keep consistency and precision.
28          */
29         ktime_get_snapshot(&systime_snapshot);
30
31         /*
32          * This is only valid if the current clocksource is the
33          * architected counter, as this is the only one the guest
34          * can see.
35          */
36         if (systime_snapshot.cs_id != CSID_ARM_ARCH_COUNTER)
37                 return;
38
39         /*
40          * The guest selects one of the two reference counters
41          * (virtual or physical) with the first argument of the SMCCC
42          * call. In case the identifier is not supported, error out.
43          */
44         feature = smccc_get_arg1(vcpu);
45         switch (feature) {
46         case KVM_PTP_VIRT_COUNTER:
47                 cycles = systime_snapshot.cycles - vcpu->kvm->arch.timer_data.voffset;
48                 break;
49         case KVM_PTP_PHYS_COUNTER:
50                 cycles = systime_snapshot.cycles;
51                 break;
52         default:
53                 return;
54         }
55
56         /*
57          * This relies on the top bit of val[0] never being set for
58          * valid values of system time, because that is *really* far
59          * in the future (about 292 years from 1970, and at that stage
60          * nobody will give a damn about it).
61          */
62         val[0] = upper_32_bits(systime_snapshot.real);
63         val[1] = lower_32_bits(systime_snapshot.real);
64         val[2] = upper_32_bits(cycles);
65         val[3] = lower_32_bits(cycles);
66 }
67
68 static bool kvm_hvc_call_default_allowed(u32 func_id)
69 {
70         switch (func_id) {
71         /*
72          * List of function-ids that are not gated with the bitmapped
73          * feature firmware registers, and are to be allowed for
74          * servicing the call by default.
75          */
76         case ARM_SMCCC_VERSION_FUNC_ID:
77         case ARM_SMCCC_ARCH_FEATURES_FUNC_ID:
78                 return true;
79         default:
80                 /* PSCI 0.2 and up is in the 0:0x1f range */
81                 if (ARM_SMCCC_OWNER_NUM(func_id) == ARM_SMCCC_OWNER_STANDARD &&
82                     ARM_SMCCC_FUNC_NUM(func_id) <= 0x1f)
83                         return true;
84
85                 /*
86                  * KVM's PSCI 0.1 doesn't comply with SMCCC, and has
87                  * its own function-id base and range
88                  */
89                 if (func_id >= KVM_PSCI_FN(0) && func_id <= KVM_PSCI_FN(3))
90                         return true;
91
92                 return false;
93         }
94 }
95
96 static bool kvm_hvc_call_allowed(struct kvm_vcpu *vcpu, u32 func_id)
97 {
98         struct kvm_smccc_features *smccc_feat = &vcpu->kvm->arch.smccc_feat;
99
100         switch (func_id) {
101         case ARM_SMCCC_TRNG_VERSION:
102         case ARM_SMCCC_TRNG_FEATURES:
103         case ARM_SMCCC_TRNG_GET_UUID:
104         case ARM_SMCCC_TRNG_RND32:
105         case ARM_SMCCC_TRNG_RND64:
106                 return test_bit(KVM_REG_ARM_STD_BIT_TRNG_V1_0,
107                                 &smccc_feat->std_bmap);
108         case ARM_SMCCC_HV_PV_TIME_FEATURES:
109         case ARM_SMCCC_HV_PV_TIME_ST:
110                 return test_bit(KVM_REG_ARM_STD_HYP_BIT_PV_TIME,
111                                 &smccc_feat->std_hyp_bmap);
112         case ARM_SMCCC_VENDOR_HYP_KVM_FEATURES_FUNC_ID:
113         case ARM_SMCCC_VENDOR_HYP_CALL_UID_FUNC_ID:
114                 return test_bit(KVM_REG_ARM_VENDOR_HYP_BIT_FUNC_FEAT,
115                                 &smccc_feat->vendor_hyp_bmap);
116         case ARM_SMCCC_VENDOR_HYP_KVM_PTP_FUNC_ID:
117                 return test_bit(KVM_REG_ARM_VENDOR_HYP_BIT_PTP,
118                                 &smccc_feat->vendor_hyp_bmap);
119         default:
120                 return kvm_hvc_call_default_allowed(func_id);
121         }
122 }
123
124 int kvm_hvc_call_handler(struct kvm_vcpu *vcpu)
125 {
126         struct kvm_smccc_features *smccc_feat = &vcpu->kvm->arch.smccc_feat;
127         u32 func_id = smccc_get_function(vcpu);
128         u64 val[4] = {SMCCC_RET_NOT_SUPPORTED};
129         u32 feature;
130         gpa_t gpa;
131
132         if (!kvm_hvc_call_allowed(vcpu, func_id))
133                 goto out;
134
135         switch (func_id) {
136         case ARM_SMCCC_VERSION_FUNC_ID:
137                 val[0] = ARM_SMCCC_VERSION_1_1;
138                 break;
139         case ARM_SMCCC_ARCH_FEATURES_FUNC_ID:
140                 feature = smccc_get_arg1(vcpu);
141                 switch (feature) {
142                 case ARM_SMCCC_ARCH_WORKAROUND_1:
143                         switch (arm64_get_spectre_v2_state()) {
144                         case SPECTRE_VULNERABLE:
145                                 break;
146                         case SPECTRE_MITIGATED:
147                                 val[0] = SMCCC_RET_SUCCESS;
148                                 break;
149                         case SPECTRE_UNAFFECTED:
150                                 val[0] = SMCCC_ARCH_WORKAROUND_RET_UNAFFECTED;
151                                 break;
152                         }
153                         break;
154                 case ARM_SMCCC_ARCH_WORKAROUND_2:
155                         switch (arm64_get_spectre_v4_state()) {
156                         case SPECTRE_VULNERABLE:
157                                 break;
158                         case SPECTRE_MITIGATED:
159                                 /*
160                                  * SSBS everywhere: Indicate no firmware
161                                  * support, as the SSBS support will be
162                                  * indicated to the guest and the default is
163                                  * safe.
164                                  *
165                                  * Otherwise, expose a permanent mitigation
166                                  * to the guest, and hide SSBS so that the
167                                  * guest stays protected.
168                                  */
169                                 if (cpus_have_final_cap(ARM64_SSBS))
170                                         break;
171                                 fallthrough;
172                         case SPECTRE_UNAFFECTED:
173                                 val[0] = SMCCC_RET_NOT_REQUIRED;
174                                 break;
175                         }
176                         break;
177                 case ARM_SMCCC_ARCH_WORKAROUND_3:
178                         switch (arm64_get_spectre_bhb_state()) {
179                         case SPECTRE_VULNERABLE:
180                                 break;
181                         case SPECTRE_MITIGATED:
182                                 val[0] = SMCCC_RET_SUCCESS;
183                                 break;
184                         case SPECTRE_UNAFFECTED:
185                                 val[0] = SMCCC_ARCH_WORKAROUND_RET_UNAFFECTED;
186                                 break;
187                         }
188                         break;
189                 case ARM_SMCCC_HV_PV_TIME_FEATURES:
190                         if (test_bit(KVM_REG_ARM_STD_HYP_BIT_PV_TIME,
191                                      &smccc_feat->std_hyp_bmap))
192                                 val[0] = SMCCC_RET_SUCCESS;
193                         break;
194                 }
195                 break;
196         case ARM_SMCCC_HV_PV_TIME_FEATURES:
197                 val[0] = kvm_hypercall_pv_features(vcpu);
198                 break;
199         case ARM_SMCCC_HV_PV_TIME_ST:
200                 gpa = kvm_init_stolen_time(vcpu);
201                 if (gpa != INVALID_GPA)
202                         val[0] = gpa;
203                 break;
204         case ARM_SMCCC_VENDOR_HYP_CALL_UID_FUNC_ID:
205                 val[0] = ARM_SMCCC_VENDOR_HYP_UID_KVM_REG_0;
206                 val[1] = ARM_SMCCC_VENDOR_HYP_UID_KVM_REG_1;
207                 val[2] = ARM_SMCCC_VENDOR_HYP_UID_KVM_REG_2;
208                 val[3] = ARM_SMCCC_VENDOR_HYP_UID_KVM_REG_3;
209                 break;
210         case ARM_SMCCC_VENDOR_HYP_KVM_FEATURES_FUNC_ID:
211                 val[0] = smccc_feat->vendor_hyp_bmap;
212                 break;
213         case ARM_SMCCC_VENDOR_HYP_KVM_PTP_FUNC_ID:
214                 kvm_ptp_get_time(vcpu, val);
215                 break;
216         case ARM_SMCCC_TRNG_VERSION:
217         case ARM_SMCCC_TRNG_FEATURES:
218         case ARM_SMCCC_TRNG_GET_UUID:
219         case ARM_SMCCC_TRNG_RND32:
220         case ARM_SMCCC_TRNG_RND64:
221                 return kvm_trng_call(vcpu);
222         default:
223                 return kvm_psci_call(vcpu);
224         }
225
226 out:
227         smccc_set_retval(vcpu, val[0], val[1], val[2], val[3]);
228         return 1;
229 }
230
231 static const u64 kvm_arm_fw_reg_ids[] = {
232         KVM_REG_ARM_PSCI_VERSION,
233         KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1,
234         KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2,
235         KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3,
236         KVM_REG_ARM_STD_BMAP,
237         KVM_REG_ARM_STD_HYP_BMAP,
238         KVM_REG_ARM_VENDOR_HYP_BMAP,
239 };
240
241 void kvm_arm_init_hypercalls(struct kvm *kvm)
242 {
243         struct kvm_smccc_features *smccc_feat = &kvm->arch.smccc_feat;
244
245         smccc_feat->std_bmap = KVM_ARM_SMCCC_STD_FEATURES;
246         smccc_feat->std_hyp_bmap = KVM_ARM_SMCCC_STD_HYP_FEATURES;
247         smccc_feat->vendor_hyp_bmap = KVM_ARM_SMCCC_VENDOR_HYP_FEATURES;
248 }
249
250 int kvm_arm_get_fw_num_regs(struct kvm_vcpu *vcpu)
251 {
252         return ARRAY_SIZE(kvm_arm_fw_reg_ids);
253 }
254
255 int kvm_arm_copy_fw_reg_indices(struct kvm_vcpu *vcpu, u64 __user *uindices)
256 {
257         int i;
258
259         for (i = 0; i < ARRAY_SIZE(kvm_arm_fw_reg_ids); i++) {
260                 if (put_user(kvm_arm_fw_reg_ids[i], uindices++))
261                         return -EFAULT;
262         }
263
264         return 0;
265 }
266
267 #define KVM_REG_FEATURE_LEVEL_MASK      GENMASK(3, 0)
268
269 /*
270  * Convert the workaround level into an easy-to-compare number, where higher
271  * values mean better protection.
272  */
273 static int get_kernel_wa_level(u64 regid)
274 {
275         switch (regid) {
276         case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1:
277                 switch (arm64_get_spectre_v2_state()) {
278                 case SPECTRE_VULNERABLE:
279                         return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1_NOT_AVAIL;
280                 case SPECTRE_MITIGATED:
281                         return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1_AVAIL;
282                 case SPECTRE_UNAFFECTED:
283                         return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1_NOT_REQUIRED;
284                 }
285                 return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1_NOT_AVAIL;
286         case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2:
287                 switch (arm64_get_spectre_v4_state()) {
288                 case SPECTRE_MITIGATED:
289                         /*
290                          * As for the hypercall discovery, we pretend we
291                          * don't have any FW mitigation if SSBS is there at
292                          * all times.
293                          */
294                         if (cpus_have_final_cap(ARM64_SSBS))
295                                 return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_AVAIL;
296                         fallthrough;
297                 case SPECTRE_UNAFFECTED:
298                         return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_REQUIRED;
299                 case SPECTRE_VULNERABLE:
300                         return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_AVAIL;
301                 }
302                 break;
303         case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3:
304                 switch (arm64_get_spectre_bhb_state()) {
305                 case SPECTRE_VULNERABLE:
306                         return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3_NOT_AVAIL;
307                 case SPECTRE_MITIGATED:
308                         return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3_AVAIL;
309                 case SPECTRE_UNAFFECTED:
310                         return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3_NOT_REQUIRED;
311                 }
312                 return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3_NOT_AVAIL;
313         }
314
315         return -EINVAL;
316 }
317
318 int kvm_arm_get_fw_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
319 {
320         struct kvm_smccc_features *smccc_feat = &vcpu->kvm->arch.smccc_feat;
321         void __user *uaddr = (void __user *)(long)reg->addr;
322         u64 val;
323
324         switch (reg->id) {
325         case KVM_REG_ARM_PSCI_VERSION:
326                 val = kvm_psci_version(vcpu);
327                 break;
328         case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1:
329         case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2:
330         case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3:
331                 val = get_kernel_wa_level(reg->id) & KVM_REG_FEATURE_LEVEL_MASK;
332                 break;
333         case KVM_REG_ARM_STD_BMAP:
334                 val = READ_ONCE(smccc_feat->std_bmap);
335                 break;
336         case KVM_REG_ARM_STD_HYP_BMAP:
337                 val = READ_ONCE(smccc_feat->std_hyp_bmap);
338                 break;
339         case KVM_REG_ARM_VENDOR_HYP_BMAP:
340                 val = READ_ONCE(smccc_feat->vendor_hyp_bmap);
341                 break;
342         default:
343                 return -ENOENT;
344         }
345
346         if (copy_to_user(uaddr, &val, KVM_REG_SIZE(reg->id)))
347                 return -EFAULT;
348
349         return 0;
350 }
351
352 static int kvm_arm_set_fw_reg_bmap(struct kvm_vcpu *vcpu, u64 reg_id, u64 val)
353 {
354         int ret = 0;
355         struct kvm *kvm = vcpu->kvm;
356         struct kvm_smccc_features *smccc_feat = &kvm->arch.smccc_feat;
357         unsigned long *fw_reg_bmap, fw_reg_features;
358
359         switch (reg_id) {
360         case KVM_REG_ARM_STD_BMAP:
361                 fw_reg_bmap = &smccc_feat->std_bmap;
362                 fw_reg_features = KVM_ARM_SMCCC_STD_FEATURES;
363                 break;
364         case KVM_REG_ARM_STD_HYP_BMAP:
365                 fw_reg_bmap = &smccc_feat->std_hyp_bmap;
366                 fw_reg_features = KVM_ARM_SMCCC_STD_HYP_FEATURES;
367                 break;
368         case KVM_REG_ARM_VENDOR_HYP_BMAP:
369                 fw_reg_bmap = &smccc_feat->vendor_hyp_bmap;
370                 fw_reg_features = KVM_ARM_SMCCC_VENDOR_HYP_FEATURES;
371                 break;
372         default:
373                 return -ENOENT;
374         }
375
376         /* Check for unsupported bit */
377         if (val & ~fw_reg_features)
378                 return -EINVAL;
379
380         mutex_lock(&kvm->lock);
381
382         if (test_bit(KVM_ARCH_FLAG_HAS_RAN_ONCE, &kvm->arch.flags) &&
383             val != *fw_reg_bmap) {
384                 ret = -EBUSY;
385                 goto out;
386         }
387
388         WRITE_ONCE(*fw_reg_bmap, val);
389 out:
390         mutex_unlock(&kvm->lock);
391         return ret;
392 }
393
394 int kvm_arm_set_fw_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
395 {
396         void __user *uaddr = (void __user *)(long)reg->addr;
397         u64 val;
398         int wa_level;
399
400         if (copy_from_user(&val, uaddr, KVM_REG_SIZE(reg->id)))
401                 return -EFAULT;
402
403         switch (reg->id) {
404         case KVM_REG_ARM_PSCI_VERSION:
405         {
406                 bool wants_02;
407
408                 wants_02 = test_bit(KVM_ARM_VCPU_PSCI_0_2, vcpu->arch.features);
409
410                 switch (val) {
411                 case KVM_ARM_PSCI_0_1:
412                         if (wants_02)
413                                 return -EINVAL;
414                         vcpu->kvm->arch.psci_version = val;
415                         return 0;
416                 case KVM_ARM_PSCI_0_2:
417                 case KVM_ARM_PSCI_1_0:
418                 case KVM_ARM_PSCI_1_1:
419                         if (!wants_02)
420                                 return -EINVAL;
421                         vcpu->kvm->arch.psci_version = val;
422                         return 0;
423                 }
424                 break;
425         }
426
427         case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1:
428         case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3:
429                 if (val & ~KVM_REG_FEATURE_LEVEL_MASK)
430                         return -EINVAL;
431
432                 if (get_kernel_wa_level(reg->id) < val)
433                         return -EINVAL;
434
435                 return 0;
436
437         case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2:
438                 if (val & ~(KVM_REG_FEATURE_LEVEL_MASK |
439                             KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_ENABLED))
440                         return -EINVAL;
441
442                 /* The enabled bit must not be set unless the level is AVAIL. */
443                 if ((val & KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_ENABLED) &&
444                     (val & KVM_REG_FEATURE_LEVEL_MASK) != KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_AVAIL)
445                         return -EINVAL;
446
447                 /*
448                  * Map all the possible incoming states to the only two we
449                  * really want to deal with.
450                  */
451                 switch (val & KVM_REG_FEATURE_LEVEL_MASK) {
452                 case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_AVAIL:
453                 case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_UNKNOWN:
454                         wa_level = KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_AVAIL;
455                         break;
456                 case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_AVAIL:
457                 case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_REQUIRED:
458                         wa_level = KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_REQUIRED;
459                         break;
460                 default:
461                         return -EINVAL;
462                 }
463
464                 /*
465                  * We can deal with NOT_AVAIL on NOT_REQUIRED, but not the
466                  * other way around.
467                  */
468                 if (get_kernel_wa_level(reg->id) < wa_level)
469                         return -EINVAL;
470
471                 return 0;
472         case KVM_REG_ARM_STD_BMAP:
473         case KVM_REG_ARM_STD_HYP_BMAP:
474         case KVM_REG_ARM_VENDOR_HYP_BMAP:
475                 return kvm_arm_set_fw_reg_bmap(vcpu, reg->id, val);
476         default:
477                 return -ENOENT;
478         }
479
480         return -EINVAL;
481 }