Linux 6.7-rc7
[linux-modified.git] / arch / arm64 / kvm / hypercalls.c
1 // SPDX-License-Identifier: GPL-2.0
2 // Copyright (C) 2019 Arm Ltd.
3
4 #include <linux/arm-smccc.h>
5 #include <linux/kvm_host.h>
6
7 #include <asm/kvm_emulate.h>
8
9 #include <kvm/arm_hypercalls.h>
10 #include <kvm/arm_psci.h>
11
12 #define KVM_ARM_SMCCC_STD_FEATURES                              \
13         GENMASK(KVM_REG_ARM_STD_BMAP_BIT_COUNT - 1, 0)
14 #define KVM_ARM_SMCCC_STD_HYP_FEATURES                          \
15         GENMASK(KVM_REG_ARM_STD_HYP_BMAP_BIT_COUNT - 1, 0)
16 #define KVM_ARM_SMCCC_VENDOR_HYP_FEATURES                       \
17         GENMASK(KVM_REG_ARM_VENDOR_HYP_BMAP_BIT_COUNT - 1, 0)
18
19 static void kvm_ptp_get_time(struct kvm_vcpu *vcpu, u64 *val)
20 {
21         struct system_time_snapshot systime_snapshot;
22         u64 cycles = ~0UL;
23         u32 feature;
24
25         /*
26          * system time and counter value must captured at the same
27          * time to keep consistency and precision.
28          */
29         ktime_get_snapshot(&systime_snapshot);
30
31         /*
32          * This is only valid if the current clocksource is the
33          * architected counter, as this is the only one the guest
34          * can see.
35          */
36         if (systime_snapshot.cs_id != CSID_ARM_ARCH_COUNTER)
37                 return;
38
39         /*
40          * The guest selects one of the two reference counters
41          * (virtual or physical) with the first argument of the SMCCC
42          * call. In case the identifier is not supported, error out.
43          */
44         feature = smccc_get_arg1(vcpu);
45         switch (feature) {
46         case KVM_PTP_VIRT_COUNTER:
47                 cycles = systime_snapshot.cycles - vcpu->kvm->arch.timer_data.voffset;
48                 break;
49         case KVM_PTP_PHYS_COUNTER:
50                 cycles = systime_snapshot.cycles - vcpu->kvm->arch.timer_data.poffset;
51                 break;
52         default:
53                 return;
54         }
55
56         /*
57          * This relies on the top bit of val[0] never being set for
58          * valid values of system time, because that is *really* far
59          * in the future (about 292 years from 1970, and at that stage
60          * nobody will give a damn about it).
61          */
62         val[0] = upper_32_bits(systime_snapshot.real);
63         val[1] = lower_32_bits(systime_snapshot.real);
64         val[2] = upper_32_bits(cycles);
65         val[3] = lower_32_bits(cycles);
66 }
67
68 static bool kvm_smccc_default_allowed(u32 func_id)
69 {
70         switch (func_id) {
71         /*
72          * List of function-ids that are not gated with the bitmapped
73          * feature firmware registers, and are to be allowed for
74          * servicing the call by default.
75          */
76         case ARM_SMCCC_VERSION_FUNC_ID:
77         case ARM_SMCCC_ARCH_FEATURES_FUNC_ID:
78                 return true;
79         default:
80                 /* PSCI 0.2 and up is in the 0:0x1f range */
81                 if (ARM_SMCCC_OWNER_NUM(func_id) == ARM_SMCCC_OWNER_STANDARD &&
82                     ARM_SMCCC_FUNC_NUM(func_id) <= 0x1f)
83                         return true;
84
85                 /*
86                  * KVM's PSCI 0.1 doesn't comply with SMCCC, and has
87                  * its own function-id base and range
88                  */
89                 if (func_id >= KVM_PSCI_FN(0) && func_id <= KVM_PSCI_FN(3))
90                         return true;
91
92                 return false;
93         }
94 }
95
96 static bool kvm_smccc_test_fw_bmap(struct kvm_vcpu *vcpu, u32 func_id)
97 {
98         struct kvm_smccc_features *smccc_feat = &vcpu->kvm->arch.smccc_feat;
99
100         switch (func_id) {
101         case ARM_SMCCC_TRNG_VERSION:
102         case ARM_SMCCC_TRNG_FEATURES:
103         case ARM_SMCCC_TRNG_GET_UUID:
104         case ARM_SMCCC_TRNG_RND32:
105         case ARM_SMCCC_TRNG_RND64:
106                 return test_bit(KVM_REG_ARM_STD_BIT_TRNG_V1_0,
107                                 &smccc_feat->std_bmap);
108         case ARM_SMCCC_HV_PV_TIME_FEATURES:
109         case ARM_SMCCC_HV_PV_TIME_ST:
110                 return test_bit(KVM_REG_ARM_STD_HYP_BIT_PV_TIME,
111                                 &smccc_feat->std_hyp_bmap);
112         case ARM_SMCCC_VENDOR_HYP_KVM_FEATURES_FUNC_ID:
113         case ARM_SMCCC_VENDOR_HYP_CALL_UID_FUNC_ID:
114                 return test_bit(KVM_REG_ARM_VENDOR_HYP_BIT_FUNC_FEAT,
115                                 &smccc_feat->vendor_hyp_bmap);
116         case ARM_SMCCC_VENDOR_HYP_KVM_PTP_FUNC_ID:
117                 return test_bit(KVM_REG_ARM_VENDOR_HYP_BIT_PTP,
118                                 &smccc_feat->vendor_hyp_bmap);
119         default:
120                 return false;
121         }
122 }
123
124 #define SMC32_ARCH_RANGE_BEGIN  ARM_SMCCC_VERSION_FUNC_ID
125 #define SMC32_ARCH_RANGE_END    ARM_SMCCC_CALL_VAL(ARM_SMCCC_FAST_CALL,         \
126                                                    ARM_SMCCC_SMC_32,            \
127                                                    0, ARM_SMCCC_FUNC_MASK)
128
129 #define SMC64_ARCH_RANGE_BEGIN  ARM_SMCCC_CALL_VAL(ARM_SMCCC_FAST_CALL,         \
130                                                    ARM_SMCCC_SMC_64,            \
131                                                    0, 0)
132 #define SMC64_ARCH_RANGE_END    ARM_SMCCC_CALL_VAL(ARM_SMCCC_FAST_CALL,         \
133                                                    ARM_SMCCC_SMC_64,            \
134                                                    0, ARM_SMCCC_FUNC_MASK)
135
136 static int kvm_smccc_filter_insert_reserved(struct kvm *kvm)
137 {
138         int r;
139
140         /*
141          * Prevent userspace from handling any SMCCC calls in the architecture
142          * range, avoiding the risk of misrepresenting Spectre mitigation status
143          * to the guest.
144          */
145         r = mtree_insert_range(&kvm->arch.smccc_filter,
146                                SMC32_ARCH_RANGE_BEGIN, SMC32_ARCH_RANGE_END,
147                                xa_mk_value(KVM_SMCCC_FILTER_HANDLE),
148                                GFP_KERNEL_ACCOUNT);
149         if (r)
150                 goto out_destroy;
151
152         r = mtree_insert_range(&kvm->arch.smccc_filter,
153                                SMC64_ARCH_RANGE_BEGIN, SMC64_ARCH_RANGE_END,
154                                xa_mk_value(KVM_SMCCC_FILTER_HANDLE),
155                                GFP_KERNEL_ACCOUNT);
156         if (r)
157                 goto out_destroy;
158
159         return 0;
160 out_destroy:
161         mtree_destroy(&kvm->arch.smccc_filter);
162         return r;
163 }
164
165 static bool kvm_smccc_filter_configured(struct kvm *kvm)
166 {
167         return !mtree_empty(&kvm->arch.smccc_filter);
168 }
169
170 static int kvm_smccc_set_filter(struct kvm *kvm, struct kvm_smccc_filter __user *uaddr)
171 {
172         const void *zero_page = page_to_virt(ZERO_PAGE(0));
173         struct kvm_smccc_filter filter;
174         u32 start, end;
175         int r;
176
177         if (copy_from_user(&filter, uaddr, sizeof(filter)))
178                 return -EFAULT;
179
180         if (memcmp(filter.pad, zero_page, sizeof(filter.pad)))
181                 return -EINVAL;
182
183         start = filter.base;
184         end = start + filter.nr_functions - 1;
185
186         if (end < start || filter.action >= NR_SMCCC_FILTER_ACTIONS)
187                 return -EINVAL;
188
189         mutex_lock(&kvm->arch.config_lock);
190
191         if (kvm_vm_has_ran_once(kvm)) {
192                 r = -EBUSY;
193                 goto out_unlock;
194         }
195
196         if (!kvm_smccc_filter_configured(kvm)) {
197                 r = kvm_smccc_filter_insert_reserved(kvm);
198                 if (WARN_ON_ONCE(r))
199                         goto out_unlock;
200         }
201
202         r = mtree_insert_range(&kvm->arch.smccc_filter, start, end,
203                                xa_mk_value(filter.action), GFP_KERNEL_ACCOUNT);
204 out_unlock:
205         mutex_unlock(&kvm->arch.config_lock);
206         return r;
207 }
208
209 static u8 kvm_smccc_filter_get_action(struct kvm *kvm, u32 func_id)
210 {
211         unsigned long idx = func_id;
212         void *val;
213
214         if (!kvm_smccc_filter_configured(kvm))
215                 return KVM_SMCCC_FILTER_HANDLE;
216
217         /*
218          * But where's the error handling, you say?
219          *
220          * mt_find() returns NULL if no entry was found, which just so happens
221          * to match KVM_SMCCC_FILTER_HANDLE.
222          */
223         val = mt_find(&kvm->arch.smccc_filter, &idx, idx);
224         return xa_to_value(val);
225 }
226
227 static u8 kvm_smccc_get_action(struct kvm_vcpu *vcpu, u32 func_id)
228 {
229         /*
230          * Intervening actions in the SMCCC filter take precedence over the
231          * pseudo-firmware register bitmaps.
232          */
233         u8 action = kvm_smccc_filter_get_action(vcpu->kvm, func_id);
234         if (action != KVM_SMCCC_FILTER_HANDLE)
235                 return action;
236
237         if (kvm_smccc_test_fw_bmap(vcpu, func_id) ||
238             kvm_smccc_default_allowed(func_id))
239                 return KVM_SMCCC_FILTER_HANDLE;
240
241         return KVM_SMCCC_FILTER_DENY;
242 }
243
244 static void kvm_prepare_hypercall_exit(struct kvm_vcpu *vcpu, u32 func_id)
245 {
246         u8 ec = ESR_ELx_EC(kvm_vcpu_get_esr(vcpu));
247         struct kvm_run *run = vcpu->run;
248         u64 flags = 0;
249
250         if (ec == ESR_ELx_EC_SMC32 || ec == ESR_ELx_EC_SMC64)
251                 flags |= KVM_HYPERCALL_EXIT_SMC;
252
253         if (!kvm_vcpu_trap_il_is32bit(vcpu))
254                 flags |= KVM_HYPERCALL_EXIT_16BIT;
255
256         run->exit_reason = KVM_EXIT_HYPERCALL;
257         run->hypercall = (typeof(run->hypercall)) {
258                 .nr     = func_id,
259                 .flags  = flags,
260         };
261 }
262
263 int kvm_smccc_call_handler(struct kvm_vcpu *vcpu)
264 {
265         struct kvm_smccc_features *smccc_feat = &vcpu->kvm->arch.smccc_feat;
266         u32 func_id = smccc_get_function(vcpu);
267         u64 val[4] = {SMCCC_RET_NOT_SUPPORTED};
268         u32 feature;
269         u8 action;
270         gpa_t gpa;
271
272         action = kvm_smccc_get_action(vcpu, func_id);
273         switch (action) {
274         case KVM_SMCCC_FILTER_HANDLE:
275                 break;
276         case KVM_SMCCC_FILTER_DENY:
277                 goto out;
278         case KVM_SMCCC_FILTER_FWD_TO_USER:
279                 kvm_prepare_hypercall_exit(vcpu, func_id);
280                 return 0;
281         default:
282                 WARN_RATELIMIT(1, "Unhandled SMCCC filter action: %d\n", action);
283                 goto out;
284         }
285
286         switch (func_id) {
287         case ARM_SMCCC_VERSION_FUNC_ID:
288                 val[0] = ARM_SMCCC_VERSION_1_1;
289                 break;
290         case ARM_SMCCC_ARCH_FEATURES_FUNC_ID:
291                 feature = smccc_get_arg1(vcpu);
292                 switch (feature) {
293                 case ARM_SMCCC_ARCH_WORKAROUND_1:
294                         switch (arm64_get_spectre_v2_state()) {
295                         case SPECTRE_VULNERABLE:
296                                 break;
297                         case SPECTRE_MITIGATED:
298                                 val[0] = SMCCC_RET_SUCCESS;
299                                 break;
300                         case SPECTRE_UNAFFECTED:
301                                 val[0] = SMCCC_ARCH_WORKAROUND_RET_UNAFFECTED;
302                                 break;
303                         }
304                         break;
305                 case ARM_SMCCC_ARCH_WORKAROUND_2:
306                         switch (arm64_get_spectre_v4_state()) {
307                         case SPECTRE_VULNERABLE:
308                                 break;
309                         case SPECTRE_MITIGATED:
310                                 /*
311                                  * SSBS everywhere: Indicate no firmware
312                                  * support, as the SSBS support will be
313                                  * indicated to the guest and the default is
314                                  * safe.
315                                  *
316                                  * Otherwise, expose a permanent mitigation
317                                  * to the guest, and hide SSBS so that the
318                                  * guest stays protected.
319                                  */
320                                 if (cpus_have_final_cap(ARM64_SSBS))
321                                         break;
322                                 fallthrough;
323                         case SPECTRE_UNAFFECTED:
324                                 val[0] = SMCCC_RET_NOT_REQUIRED;
325                                 break;
326                         }
327                         break;
328                 case ARM_SMCCC_ARCH_WORKAROUND_3:
329                         switch (arm64_get_spectre_bhb_state()) {
330                         case SPECTRE_VULNERABLE:
331                                 break;
332                         case SPECTRE_MITIGATED:
333                                 val[0] = SMCCC_RET_SUCCESS;
334                                 break;
335                         case SPECTRE_UNAFFECTED:
336                                 val[0] = SMCCC_ARCH_WORKAROUND_RET_UNAFFECTED;
337                                 break;
338                         }
339                         break;
340                 case ARM_SMCCC_HV_PV_TIME_FEATURES:
341                         if (test_bit(KVM_REG_ARM_STD_HYP_BIT_PV_TIME,
342                                      &smccc_feat->std_hyp_bmap))
343                                 val[0] = SMCCC_RET_SUCCESS;
344                         break;
345                 }
346                 break;
347         case ARM_SMCCC_HV_PV_TIME_FEATURES:
348                 val[0] = kvm_hypercall_pv_features(vcpu);
349                 break;
350         case ARM_SMCCC_HV_PV_TIME_ST:
351                 gpa = kvm_init_stolen_time(vcpu);
352                 if (gpa != INVALID_GPA)
353                         val[0] = gpa;
354                 break;
355         case ARM_SMCCC_VENDOR_HYP_CALL_UID_FUNC_ID:
356                 val[0] = ARM_SMCCC_VENDOR_HYP_UID_KVM_REG_0;
357                 val[1] = ARM_SMCCC_VENDOR_HYP_UID_KVM_REG_1;
358                 val[2] = ARM_SMCCC_VENDOR_HYP_UID_KVM_REG_2;
359                 val[3] = ARM_SMCCC_VENDOR_HYP_UID_KVM_REG_3;
360                 break;
361         case ARM_SMCCC_VENDOR_HYP_KVM_FEATURES_FUNC_ID:
362                 val[0] = smccc_feat->vendor_hyp_bmap;
363                 break;
364         case ARM_SMCCC_VENDOR_HYP_KVM_PTP_FUNC_ID:
365                 kvm_ptp_get_time(vcpu, val);
366                 break;
367         case ARM_SMCCC_TRNG_VERSION:
368         case ARM_SMCCC_TRNG_FEATURES:
369         case ARM_SMCCC_TRNG_GET_UUID:
370         case ARM_SMCCC_TRNG_RND32:
371         case ARM_SMCCC_TRNG_RND64:
372                 return kvm_trng_call(vcpu);
373         default:
374                 return kvm_psci_call(vcpu);
375         }
376
377 out:
378         smccc_set_retval(vcpu, val[0], val[1], val[2], val[3]);
379         return 1;
380 }
381
382 static const u64 kvm_arm_fw_reg_ids[] = {
383         KVM_REG_ARM_PSCI_VERSION,
384         KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1,
385         KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2,
386         KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3,
387         KVM_REG_ARM_STD_BMAP,
388         KVM_REG_ARM_STD_HYP_BMAP,
389         KVM_REG_ARM_VENDOR_HYP_BMAP,
390 };
391
392 void kvm_arm_init_hypercalls(struct kvm *kvm)
393 {
394         struct kvm_smccc_features *smccc_feat = &kvm->arch.smccc_feat;
395
396         smccc_feat->std_bmap = KVM_ARM_SMCCC_STD_FEATURES;
397         smccc_feat->std_hyp_bmap = KVM_ARM_SMCCC_STD_HYP_FEATURES;
398         smccc_feat->vendor_hyp_bmap = KVM_ARM_SMCCC_VENDOR_HYP_FEATURES;
399
400         mt_init(&kvm->arch.smccc_filter);
401 }
402
403 void kvm_arm_teardown_hypercalls(struct kvm *kvm)
404 {
405         mtree_destroy(&kvm->arch.smccc_filter);
406 }
407
408 int kvm_arm_get_fw_num_regs(struct kvm_vcpu *vcpu)
409 {
410         return ARRAY_SIZE(kvm_arm_fw_reg_ids);
411 }
412
413 int kvm_arm_copy_fw_reg_indices(struct kvm_vcpu *vcpu, u64 __user *uindices)
414 {
415         int i;
416
417         for (i = 0; i < ARRAY_SIZE(kvm_arm_fw_reg_ids); i++) {
418                 if (put_user(kvm_arm_fw_reg_ids[i], uindices++))
419                         return -EFAULT;
420         }
421
422         return 0;
423 }
424
425 #define KVM_REG_FEATURE_LEVEL_MASK      GENMASK(3, 0)
426
427 /*
428  * Convert the workaround level into an easy-to-compare number, where higher
429  * values mean better protection.
430  */
431 static int get_kernel_wa_level(u64 regid)
432 {
433         switch (regid) {
434         case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1:
435                 switch (arm64_get_spectre_v2_state()) {
436                 case SPECTRE_VULNERABLE:
437                         return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1_NOT_AVAIL;
438                 case SPECTRE_MITIGATED:
439                         return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1_AVAIL;
440                 case SPECTRE_UNAFFECTED:
441                         return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1_NOT_REQUIRED;
442                 }
443                 return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1_NOT_AVAIL;
444         case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2:
445                 switch (arm64_get_spectre_v4_state()) {
446                 case SPECTRE_MITIGATED:
447                         /*
448                          * As for the hypercall discovery, we pretend we
449                          * don't have any FW mitigation if SSBS is there at
450                          * all times.
451                          */
452                         if (cpus_have_final_cap(ARM64_SSBS))
453                                 return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_AVAIL;
454                         fallthrough;
455                 case SPECTRE_UNAFFECTED:
456                         return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_REQUIRED;
457                 case SPECTRE_VULNERABLE:
458                         return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_AVAIL;
459                 }
460                 break;
461         case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3:
462                 switch (arm64_get_spectre_bhb_state()) {
463                 case SPECTRE_VULNERABLE:
464                         return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3_NOT_AVAIL;
465                 case SPECTRE_MITIGATED:
466                         return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3_AVAIL;
467                 case SPECTRE_UNAFFECTED:
468                         return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3_NOT_REQUIRED;
469                 }
470                 return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3_NOT_AVAIL;
471         }
472
473         return -EINVAL;
474 }
475
476 int kvm_arm_get_fw_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
477 {
478         struct kvm_smccc_features *smccc_feat = &vcpu->kvm->arch.smccc_feat;
479         void __user *uaddr = (void __user *)(long)reg->addr;
480         u64 val;
481
482         switch (reg->id) {
483         case KVM_REG_ARM_PSCI_VERSION:
484                 val = kvm_psci_version(vcpu);
485                 break;
486         case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1:
487         case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2:
488         case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3:
489                 val = get_kernel_wa_level(reg->id) & KVM_REG_FEATURE_LEVEL_MASK;
490                 break;
491         case KVM_REG_ARM_STD_BMAP:
492                 val = READ_ONCE(smccc_feat->std_bmap);
493                 break;
494         case KVM_REG_ARM_STD_HYP_BMAP:
495                 val = READ_ONCE(smccc_feat->std_hyp_bmap);
496                 break;
497         case KVM_REG_ARM_VENDOR_HYP_BMAP:
498                 val = READ_ONCE(smccc_feat->vendor_hyp_bmap);
499                 break;
500         default:
501                 return -ENOENT;
502         }
503
504         if (copy_to_user(uaddr, &val, KVM_REG_SIZE(reg->id)))
505                 return -EFAULT;
506
507         return 0;
508 }
509
510 static int kvm_arm_set_fw_reg_bmap(struct kvm_vcpu *vcpu, u64 reg_id, u64 val)
511 {
512         int ret = 0;
513         struct kvm *kvm = vcpu->kvm;
514         struct kvm_smccc_features *smccc_feat = &kvm->arch.smccc_feat;
515         unsigned long *fw_reg_bmap, fw_reg_features;
516
517         switch (reg_id) {
518         case KVM_REG_ARM_STD_BMAP:
519                 fw_reg_bmap = &smccc_feat->std_bmap;
520                 fw_reg_features = KVM_ARM_SMCCC_STD_FEATURES;
521                 break;
522         case KVM_REG_ARM_STD_HYP_BMAP:
523                 fw_reg_bmap = &smccc_feat->std_hyp_bmap;
524                 fw_reg_features = KVM_ARM_SMCCC_STD_HYP_FEATURES;
525                 break;
526         case KVM_REG_ARM_VENDOR_HYP_BMAP:
527                 fw_reg_bmap = &smccc_feat->vendor_hyp_bmap;
528                 fw_reg_features = KVM_ARM_SMCCC_VENDOR_HYP_FEATURES;
529                 break;
530         default:
531                 return -ENOENT;
532         }
533
534         /* Check for unsupported bit */
535         if (val & ~fw_reg_features)
536                 return -EINVAL;
537
538         mutex_lock(&kvm->arch.config_lock);
539
540         if (kvm_vm_has_ran_once(kvm) && val != *fw_reg_bmap) {
541                 ret = -EBUSY;
542                 goto out;
543         }
544
545         WRITE_ONCE(*fw_reg_bmap, val);
546 out:
547         mutex_unlock(&kvm->arch.config_lock);
548         return ret;
549 }
550
551 int kvm_arm_set_fw_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
552 {
553         void __user *uaddr = (void __user *)(long)reg->addr;
554         u64 val;
555         int wa_level;
556
557         if (KVM_REG_SIZE(reg->id) != sizeof(val))
558                 return -ENOENT;
559         if (copy_from_user(&val, uaddr, KVM_REG_SIZE(reg->id)))
560                 return -EFAULT;
561
562         switch (reg->id) {
563         case KVM_REG_ARM_PSCI_VERSION:
564         {
565                 bool wants_02;
566
567                 wants_02 = vcpu_has_feature(vcpu, KVM_ARM_VCPU_PSCI_0_2);
568
569                 switch (val) {
570                 case KVM_ARM_PSCI_0_1:
571                         if (wants_02)
572                                 return -EINVAL;
573                         vcpu->kvm->arch.psci_version = val;
574                         return 0;
575                 case KVM_ARM_PSCI_0_2:
576                 case KVM_ARM_PSCI_1_0:
577                 case KVM_ARM_PSCI_1_1:
578                         if (!wants_02)
579                                 return -EINVAL;
580                         vcpu->kvm->arch.psci_version = val;
581                         return 0;
582                 }
583                 break;
584         }
585
586         case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1:
587         case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3:
588                 if (val & ~KVM_REG_FEATURE_LEVEL_MASK)
589                         return -EINVAL;
590
591                 if (get_kernel_wa_level(reg->id) < val)
592                         return -EINVAL;
593
594                 return 0;
595
596         case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2:
597                 if (val & ~(KVM_REG_FEATURE_LEVEL_MASK |
598                             KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_ENABLED))
599                         return -EINVAL;
600
601                 /* The enabled bit must not be set unless the level is AVAIL. */
602                 if ((val & KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_ENABLED) &&
603                     (val & KVM_REG_FEATURE_LEVEL_MASK) != KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_AVAIL)
604                         return -EINVAL;
605
606                 /*
607                  * Map all the possible incoming states to the only two we
608                  * really want to deal with.
609                  */
610                 switch (val & KVM_REG_FEATURE_LEVEL_MASK) {
611                 case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_AVAIL:
612                 case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_UNKNOWN:
613                         wa_level = KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_AVAIL;
614                         break;
615                 case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_AVAIL:
616                 case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_REQUIRED:
617                         wa_level = KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_REQUIRED;
618                         break;
619                 default:
620                         return -EINVAL;
621                 }
622
623                 /*
624                  * We can deal with NOT_AVAIL on NOT_REQUIRED, but not the
625                  * other way around.
626                  */
627                 if (get_kernel_wa_level(reg->id) < wa_level)
628                         return -EINVAL;
629
630                 return 0;
631         case KVM_REG_ARM_STD_BMAP:
632         case KVM_REG_ARM_STD_HYP_BMAP:
633         case KVM_REG_ARM_VENDOR_HYP_BMAP:
634                 return kvm_arm_set_fw_reg_bmap(vcpu, reg->id, val);
635         default:
636                 return -ENOENT;
637         }
638
639         return -EINVAL;
640 }
641
642 int kvm_vm_smccc_has_attr(struct kvm *kvm, struct kvm_device_attr *attr)
643 {
644         switch (attr->attr) {
645         case KVM_ARM_VM_SMCCC_FILTER:
646                 return 0;
647         default:
648                 return -ENXIO;
649         }
650 }
651
652 int kvm_vm_smccc_set_attr(struct kvm *kvm, struct kvm_device_attr *attr)
653 {
654         void __user *uaddr = (void __user *)attr->addr;
655
656         switch (attr->attr) {
657         case KVM_ARM_VM_SMCCC_FILTER:
658                 return kvm_smccc_set_filter(kvm, uaddr);
659         default:
660                 return -ENXIO;
661         }
662 }