xref: /linux/arch/arm64/kvm/hypercalls.c (revision 1634b7adcc5bef645b3666fdd564e5952a9e24e0)
1 // SPDX-License-Identifier: GPL-2.0
2 // Copyright (C) 2019 Arm Ltd.
3 
4 #include <linux/arm-smccc.h>
5 #include <linux/kvm_host.h>
6 
7 #include <asm/kvm_emulate.h>
8 
9 #include <kvm/arm_hypercalls.h>
10 #include <kvm/arm_psci.h>
11 
12 #define KVM_ARM_SMCCC_STD_FEATURES				\
13 	GENMASK(KVM_REG_ARM_STD_BMAP_BIT_COUNT - 1, 0)
14 #define KVM_ARM_SMCCC_STD_HYP_FEATURES				\
15 	GENMASK(KVM_REG_ARM_STD_HYP_BMAP_BIT_COUNT - 1, 0)
16 #define KVM_ARM_SMCCC_VENDOR_HYP_FEATURES			\
17 	GENMASK(KVM_REG_ARM_VENDOR_HYP_BMAP_BIT_COUNT - 1, 0)
18 
19 static void kvm_ptp_get_time(struct kvm_vcpu *vcpu, u64 *val)
20 {
21 	struct system_time_snapshot systime_snapshot;
22 	u64 cycles = ~0UL;
23 	u32 feature;
24 
25 	/*
26 	 * system time and counter value must captured at the same
27 	 * time to keep consistency and precision.
28 	 */
29 	ktime_get_snapshot(&systime_snapshot);
30 
31 	/*
32 	 * This is only valid if the current clocksource is the
33 	 * architected counter, as this is the only one the guest
34 	 * can see.
35 	 */
36 	if (systime_snapshot.cs_id != CSID_ARM_ARCH_COUNTER)
37 		return;
38 
39 	/*
40 	 * The guest selects one of the two reference counters
41 	 * (virtual or physical) with the first argument of the SMCCC
42 	 * call. In case the identifier is not supported, error out.
43 	 */
44 	feature = smccc_get_arg1(vcpu);
45 	switch (feature) {
46 	case KVM_PTP_VIRT_COUNTER:
47 		cycles = systime_snapshot.cycles - vcpu_read_sys_reg(vcpu, CNTVOFF_EL2);
48 		break;
49 	case KVM_PTP_PHYS_COUNTER:
50 		cycles = systime_snapshot.cycles;
51 		break;
52 	default:
53 		return;
54 	}
55 
56 	/*
57 	 * This relies on the top bit of val[0] never being set for
58 	 * valid values of system time, because that is *really* far
59 	 * in the future (about 292 years from 1970, and at that stage
60 	 * nobody will give a damn about it).
61 	 */
62 	val[0] = upper_32_bits(systime_snapshot.real);
63 	val[1] = lower_32_bits(systime_snapshot.real);
64 	val[2] = upper_32_bits(cycles);
65 	val[3] = lower_32_bits(cycles);
66 }
67 
68 static bool kvm_hvc_call_default_allowed(u32 func_id)
69 {
70 	switch (func_id) {
71 	/*
72 	 * List of function-ids that are not gated with the bitmapped
73 	 * feature firmware registers, and are to be allowed for
74 	 * servicing the call by default.
75 	 */
76 	case ARM_SMCCC_VERSION_FUNC_ID:
77 	case ARM_SMCCC_ARCH_FEATURES_FUNC_ID:
78 		return true;
79 	default:
80 		/* PSCI 0.2 and up is in the 0:0x1f range */
81 		if (ARM_SMCCC_OWNER_NUM(func_id) == ARM_SMCCC_OWNER_STANDARD &&
82 		    ARM_SMCCC_FUNC_NUM(func_id) <= 0x1f)
83 			return true;
84 
85 		/*
86 		 * KVM's PSCI 0.1 doesn't comply with SMCCC, and has
87 		 * its own function-id base and range
88 		 */
89 		if (func_id >= KVM_PSCI_FN(0) && func_id <= KVM_PSCI_FN(3))
90 			return true;
91 
92 		return false;
93 	}
94 }
95 
96 static bool kvm_hvc_call_allowed(struct kvm_vcpu *vcpu, u32 func_id)
97 {
98 	struct kvm_smccc_features *smccc_feat = &vcpu->kvm->arch.smccc_feat;
99 
100 	switch (func_id) {
101 	case ARM_SMCCC_TRNG_VERSION:
102 	case ARM_SMCCC_TRNG_FEATURES:
103 	case ARM_SMCCC_TRNG_GET_UUID:
104 	case ARM_SMCCC_TRNG_RND32:
105 	case ARM_SMCCC_TRNG_RND64:
106 		return test_bit(KVM_REG_ARM_STD_BIT_TRNG_V1_0,
107 				&smccc_feat->std_bmap);
108 	case ARM_SMCCC_HV_PV_TIME_FEATURES:
109 	case ARM_SMCCC_HV_PV_TIME_ST:
110 		return test_bit(KVM_REG_ARM_STD_HYP_BIT_PV_TIME,
111 				&smccc_feat->std_hyp_bmap);
112 	case ARM_SMCCC_VENDOR_HYP_KVM_FEATURES_FUNC_ID:
113 	case ARM_SMCCC_VENDOR_HYP_CALL_UID_FUNC_ID:
114 		return test_bit(KVM_REG_ARM_VENDOR_HYP_BIT_FUNC_FEAT,
115 				&smccc_feat->vendor_hyp_bmap);
116 	case ARM_SMCCC_VENDOR_HYP_KVM_PTP_FUNC_ID:
117 		return test_bit(KVM_REG_ARM_VENDOR_HYP_BIT_PTP,
118 				&smccc_feat->vendor_hyp_bmap);
119 	default:
120 		return kvm_hvc_call_default_allowed(func_id);
121 	}
122 }
123 
124 int kvm_hvc_call_handler(struct kvm_vcpu *vcpu)
125 {
126 	struct kvm_smccc_features *smccc_feat = &vcpu->kvm->arch.smccc_feat;
127 	u32 func_id = smccc_get_function(vcpu);
128 	u64 val[4] = {SMCCC_RET_NOT_SUPPORTED};
129 	u32 feature;
130 	gpa_t gpa;
131 
132 	if (!kvm_hvc_call_allowed(vcpu, func_id))
133 		goto out;
134 
135 	switch (func_id) {
136 	case ARM_SMCCC_VERSION_FUNC_ID:
137 		val[0] = ARM_SMCCC_VERSION_1_1;
138 		break;
139 	case ARM_SMCCC_ARCH_FEATURES_FUNC_ID:
140 		feature = smccc_get_arg1(vcpu);
141 		switch (feature) {
142 		case ARM_SMCCC_ARCH_WORKAROUND_1:
143 			switch (arm64_get_spectre_v2_state()) {
144 			case SPECTRE_VULNERABLE:
145 				break;
146 			case SPECTRE_MITIGATED:
147 				val[0] = SMCCC_RET_SUCCESS;
148 				break;
149 			case SPECTRE_UNAFFECTED:
150 				val[0] = SMCCC_ARCH_WORKAROUND_RET_UNAFFECTED;
151 				break;
152 			}
153 			break;
154 		case ARM_SMCCC_ARCH_WORKAROUND_2:
155 			switch (arm64_get_spectre_v4_state()) {
156 			case SPECTRE_VULNERABLE:
157 				break;
158 			case SPECTRE_MITIGATED:
159 				/*
160 				 * SSBS everywhere: Indicate no firmware
161 				 * support, as the SSBS support will be
162 				 * indicated to the guest and the default is
163 				 * safe.
164 				 *
165 				 * Otherwise, expose a permanent mitigation
166 				 * to the guest, and hide SSBS so that the
167 				 * guest stays protected.
168 				 */
169 				if (cpus_have_final_cap(ARM64_SSBS))
170 					break;
171 				fallthrough;
172 			case SPECTRE_UNAFFECTED:
173 				val[0] = SMCCC_RET_NOT_REQUIRED;
174 				break;
175 			}
176 			break;
177 		case ARM_SMCCC_ARCH_WORKAROUND_3:
178 			switch (arm64_get_spectre_bhb_state()) {
179 			case SPECTRE_VULNERABLE:
180 				break;
181 			case SPECTRE_MITIGATED:
182 				val[0] = SMCCC_RET_SUCCESS;
183 				break;
184 			case SPECTRE_UNAFFECTED:
185 				val[0] = SMCCC_ARCH_WORKAROUND_RET_UNAFFECTED;
186 				break;
187 			}
188 			break;
189 		case ARM_SMCCC_HV_PV_TIME_FEATURES:
190 			if (test_bit(KVM_REG_ARM_STD_HYP_BIT_PV_TIME,
191 				     &smccc_feat->std_hyp_bmap))
192 				val[0] = SMCCC_RET_SUCCESS;
193 			break;
194 		}
195 		break;
196 	case ARM_SMCCC_HV_PV_TIME_FEATURES:
197 		val[0] = kvm_hypercall_pv_features(vcpu);
198 		break;
199 	case ARM_SMCCC_HV_PV_TIME_ST:
200 		gpa = kvm_init_stolen_time(vcpu);
201 		if (gpa != INVALID_GPA)
202 			val[0] = gpa;
203 		break;
204 	case ARM_SMCCC_VENDOR_HYP_CALL_UID_FUNC_ID:
205 		val[0] = ARM_SMCCC_VENDOR_HYP_UID_KVM_REG_0;
206 		val[1] = ARM_SMCCC_VENDOR_HYP_UID_KVM_REG_1;
207 		val[2] = ARM_SMCCC_VENDOR_HYP_UID_KVM_REG_2;
208 		val[3] = ARM_SMCCC_VENDOR_HYP_UID_KVM_REG_3;
209 		break;
210 	case ARM_SMCCC_VENDOR_HYP_KVM_FEATURES_FUNC_ID:
211 		val[0] = smccc_feat->vendor_hyp_bmap;
212 		break;
213 	case ARM_SMCCC_VENDOR_HYP_KVM_PTP_FUNC_ID:
214 		kvm_ptp_get_time(vcpu, val);
215 		break;
216 	case ARM_SMCCC_TRNG_VERSION:
217 	case ARM_SMCCC_TRNG_FEATURES:
218 	case ARM_SMCCC_TRNG_GET_UUID:
219 	case ARM_SMCCC_TRNG_RND32:
220 	case ARM_SMCCC_TRNG_RND64:
221 		return kvm_trng_call(vcpu);
222 	default:
223 		return kvm_psci_call(vcpu);
224 	}
225 
226 out:
227 	smccc_set_retval(vcpu, val[0], val[1], val[2], val[3]);
228 	return 1;
229 }
230 
231 static const u64 kvm_arm_fw_reg_ids[] = {
232 	KVM_REG_ARM_PSCI_VERSION,
233 	KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1,
234 	KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2,
235 	KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3,
236 	KVM_REG_ARM_STD_BMAP,
237 	KVM_REG_ARM_STD_HYP_BMAP,
238 	KVM_REG_ARM_VENDOR_HYP_BMAP,
239 };
240 
241 void kvm_arm_init_hypercalls(struct kvm *kvm)
242 {
243 	struct kvm_smccc_features *smccc_feat = &kvm->arch.smccc_feat;
244 
245 	smccc_feat->std_bmap = KVM_ARM_SMCCC_STD_FEATURES;
246 	smccc_feat->std_hyp_bmap = KVM_ARM_SMCCC_STD_HYP_FEATURES;
247 	smccc_feat->vendor_hyp_bmap = KVM_ARM_SMCCC_VENDOR_HYP_FEATURES;
248 }
249 
250 int kvm_arm_get_fw_num_regs(struct kvm_vcpu *vcpu)
251 {
252 	return ARRAY_SIZE(kvm_arm_fw_reg_ids);
253 }
254 
255 int kvm_arm_copy_fw_reg_indices(struct kvm_vcpu *vcpu, u64 __user *uindices)
256 {
257 	int i;
258 
259 	for (i = 0; i < ARRAY_SIZE(kvm_arm_fw_reg_ids); i++) {
260 		if (put_user(kvm_arm_fw_reg_ids[i], uindices++))
261 			return -EFAULT;
262 	}
263 
264 	return 0;
265 }
266 
267 #define KVM_REG_FEATURE_LEVEL_MASK	GENMASK(3, 0)
268 
269 /*
270  * Convert the workaround level into an easy-to-compare number, where higher
271  * values mean better protection.
272  */
273 static int get_kernel_wa_level(u64 regid)
274 {
275 	switch (regid) {
276 	case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1:
277 		switch (arm64_get_spectre_v2_state()) {
278 		case SPECTRE_VULNERABLE:
279 			return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1_NOT_AVAIL;
280 		case SPECTRE_MITIGATED:
281 			return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1_AVAIL;
282 		case SPECTRE_UNAFFECTED:
283 			return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1_NOT_REQUIRED;
284 		}
285 		return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1_NOT_AVAIL;
286 	case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2:
287 		switch (arm64_get_spectre_v4_state()) {
288 		case SPECTRE_MITIGATED:
289 			/*
290 			 * As for the hypercall discovery, we pretend we
291 			 * don't have any FW mitigation if SSBS is there at
292 			 * all times.
293 			 */
294 			if (cpus_have_final_cap(ARM64_SSBS))
295 				return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_AVAIL;
296 			fallthrough;
297 		case SPECTRE_UNAFFECTED:
298 			return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_REQUIRED;
299 		case SPECTRE_VULNERABLE:
300 			return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_AVAIL;
301 		}
302 		break;
303 	case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3:
304 		switch (arm64_get_spectre_bhb_state()) {
305 		case SPECTRE_VULNERABLE:
306 			return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3_NOT_AVAIL;
307 		case SPECTRE_MITIGATED:
308 			return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3_AVAIL;
309 		case SPECTRE_UNAFFECTED:
310 			return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3_NOT_REQUIRED;
311 		}
312 		return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3_NOT_AVAIL;
313 	}
314 
315 	return -EINVAL;
316 }
317 
318 int kvm_arm_get_fw_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
319 {
320 	struct kvm_smccc_features *smccc_feat = &vcpu->kvm->arch.smccc_feat;
321 	void __user *uaddr = (void __user *)(long)reg->addr;
322 	u64 val;
323 
324 	switch (reg->id) {
325 	case KVM_REG_ARM_PSCI_VERSION:
326 		val = kvm_psci_version(vcpu);
327 		break;
328 	case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1:
329 	case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2:
330 	case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3:
331 		val = get_kernel_wa_level(reg->id) & KVM_REG_FEATURE_LEVEL_MASK;
332 		break;
333 	case KVM_REG_ARM_STD_BMAP:
334 		val = READ_ONCE(smccc_feat->std_bmap);
335 		break;
336 	case KVM_REG_ARM_STD_HYP_BMAP:
337 		val = READ_ONCE(smccc_feat->std_hyp_bmap);
338 		break;
339 	case KVM_REG_ARM_VENDOR_HYP_BMAP:
340 		val = READ_ONCE(smccc_feat->vendor_hyp_bmap);
341 		break;
342 	default:
343 		return -ENOENT;
344 	}
345 
346 	if (copy_to_user(uaddr, &val, KVM_REG_SIZE(reg->id)))
347 		return -EFAULT;
348 
349 	return 0;
350 }
351 
352 static int kvm_arm_set_fw_reg_bmap(struct kvm_vcpu *vcpu, u64 reg_id, u64 val)
353 {
354 	int ret = 0;
355 	struct kvm *kvm = vcpu->kvm;
356 	struct kvm_smccc_features *smccc_feat = &kvm->arch.smccc_feat;
357 	unsigned long *fw_reg_bmap, fw_reg_features;
358 
359 	switch (reg_id) {
360 	case KVM_REG_ARM_STD_BMAP:
361 		fw_reg_bmap = &smccc_feat->std_bmap;
362 		fw_reg_features = KVM_ARM_SMCCC_STD_FEATURES;
363 		break;
364 	case KVM_REG_ARM_STD_HYP_BMAP:
365 		fw_reg_bmap = &smccc_feat->std_hyp_bmap;
366 		fw_reg_features = KVM_ARM_SMCCC_STD_HYP_FEATURES;
367 		break;
368 	case KVM_REG_ARM_VENDOR_HYP_BMAP:
369 		fw_reg_bmap = &smccc_feat->vendor_hyp_bmap;
370 		fw_reg_features = KVM_ARM_SMCCC_VENDOR_HYP_FEATURES;
371 		break;
372 	default:
373 		return -ENOENT;
374 	}
375 
376 	/* Check for unsupported bit */
377 	if (val & ~fw_reg_features)
378 		return -EINVAL;
379 
380 	mutex_lock(&kvm->lock);
381 
382 	if (test_bit(KVM_ARCH_FLAG_HAS_RAN_ONCE, &kvm->arch.flags) &&
383 	    val != *fw_reg_bmap) {
384 		ret = -EBUSY;
385 		goto out;
386 	}
387 
388 	WRITE_ONCE(*fw_reg_bmap, val);
389 out:
390 	mutex_unlock(&kvm->lock);
391 	return ret;
392 }
393 
394 int kvm_arm_set_fw_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
395 {
396 	void __user *uaddr = (void __user *)(long)reg->addr;
397 	u64 val;
398 	int wa_level;
399 
400 	if (copy_from_user(&val, uaddr, KVM_REG_SIZE(reg->id)))
401 		return -EFAULT;
402 
403 	switch (reg->id) {
404 	case KVM_REG_ARM_PSCI_VERSION:
405 	{
406 		bool wants_02;
407 
408 		wants_02 = test_bit(KVM_ARM_VCPU_PSCI_0_2, vcpu->arch.features);
409 
410 		switch (val) {
411 		case KVM_ARM_PSCI_0_1:
412 			if (wants_02)
413 				return -EINVAL;
414 			vcpu->kvm->arch.psci_version = val;
415 			return 0;
416 		case KVM_ARM_PSCI_0_2:
417 		case KVM_ARM_PSCI_1_0:
418 		case KVM_ARM_PSCI_1_1:
419 			if (!wants_02)
420 				return -EINVAL;
421 			vcpu->kvm->arch.psci_version = val;
422 			return 0;
423 		}
424 		break;
425 	}
426 
427 	case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1:
428 	case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3:
429 		if (val & ~KVM_REG_FEATURE_LEVEL_MASK)
430 			return -EINVAL;
431 
432 		if (get_kernel_wa_level(reg->id) < val)
433 			return -EINVAL;
434 
435 		return 0;
436 
437 	case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2:
438 		if (val & ~(KVM_REG_FEATURE_LEVEL_MASK |
439 			    KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_ENABLED))
440 			return -EINVAL;
441 
442 		/* The enabled bit must not be set unless the level is AVAIL. */
443 		if ((val & KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_ENABLED) &&
444 		    (val & KVM_REG_FEATURE_LEVEL_MASK) != KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_AVAIL)
445 			return -EINVAL;
446 
447 		/*
448 		 * Map all the possible incoming states to the only two we
449 		 * really want to deal with.
450 		 */
451 		switch (val & KVM_REG_FEATURE_LEVEL_MASK) {
452 		case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_AVAIL:
453 		case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_UNKNOWN:
454 			wa_level = KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_AVAIL;
455 			break;
456 		case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_AVAIL:
457 		case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_REQUIRED:
458 			wa_level = KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_REQUIRED;
459 			break;
460 		default:
461 			return -EINVAL;
462 		}
463 
464 		/*
465 		 * We can deal with NOT_AVAIL on NOT_REQUIRED, but not the
466 		 * other way around.
467 		 */
468 		if (get_kernel_wa_level(reg->id) < wa_level)
469 			return -EINVAL;
470 
471 		return 0;
472 	case KVM_REG_ARM_STD_BMAP:
473 	case KVM_REG_ARM_STD_HYP_BMAP:
474 	case KVM_REG_ARM_VENDOR_HYP_BMAP:
475 		return kvm_arm_set_fw_reg_bmap(vcpu, reg->id, val);
476 	default:
477 		return -ENOENT;
478 	}
479 
480 	return -EINVAL;
481 }
482