xref: /linux/arch/arm64/kvm/vgic/vgic-init.c (revision ae22a94997b8a03dcb3c922857c203246711f9d4)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * Copyright (C) 2015, 2016 ARM Ltd.
4  */
5 
6 #include <linux/uaccess.h>
7 #include <linux/interrupt.h>
8 #include <linux/cpu.h>
9 #include <linux/kvm_host.h>
10 #include <kvm/arm_vgic.h>
11 #include <asm/kvm_emulate.h>
12 #include <asm/kvm_mmu.h>
13 #include "vgic.h"
14 
15 /*
16  * Initialization rules: there are multiple stages to the vgic
17  * initialization, both for the distributor and the CPU interfaces.  The basic
18  * idea is that even though the VGIC is not functional or not requested from
19  * user space, the critical path of the run loop can still call VGIC functions
20  * that just won't do anything, without them having to check additional
21  * initialization flags to ensure they don't look at uninitialized data
22  * structures.
23  *
24  * Distributor:
25  *
26  * - kvm_vgic_early_init(): initialization of static data that doesn't
27  *   depend on any sizing information or emulation type. No allocation
28  *   is allowed there.
29  *
30  * - vgic_init(): allocation and initialization of the generic data
31  *   structures that depend on sizing information (number of CPUs,
32  *   number of interrupts). Also initializes the vcpu specific data
33  *   structures. Can be executed lazily for GICv2.
34  *
35  * CPU Interface:
36  *
37  * - kvm_vgic_vcpu_init(): initialization of static data that
38  *   doesn't depend on any sizing information or emulation type. No
39  *   allocation is allowed there.
40  */
41 
42 /* EARLY INIT */
43 
44 /**
45  * kvm_vgic_early_init() - Initialize static VGIC VCPU data structures
46  * @kvm: The VM whose VGIC districutor should be initialized
47  *
48  * Only do initialization of static structures that don't require any
49  * allocation or sizing information from userspace.  vgic_init() called
50  * kvm_vgic_dist_init() which takes care of the rest.
51  */
52 void kvm_vgic_early_init(struct kvm *kvm)
53 {
54 	struct vgic_dist *dist = &kvm->arch.vgic;
55 
56 	INIT_LIST_HEAD(&dist->lpi_translation_cache);
57 	raw_spin_lock_init(&dist->lpi_list_lock);
58 	xa_init_flags(&dist->lpi_xa, XA_FLAGS_LOCK_IRQ);
59 }
60 
61 /* CREATION */
62 
63 /**
64  * kvm_vgic_create: triggered by the instantiation of the VGIC device by
65  * user space, either through the legacy KVM_CREATE_IRQCHIP ioctl (v2 only)
66  * or through the generic KVM_CREATE_DEVICE API ioctl.
67  * irqchip_in_kernel() tells you if this function succeeded or not.
68  * @kvm: kvm struct pointer
69  * @type: KVM_DEV_TYPE_ARM_VGIC_V[23]
70  */
71 int kvm_vgic_create(struct kvm *kvm, u32 type)
72 {
73 	struct kvm_vcpu *vcpu;
74 	unsigned long i;
75 	int ret;
76 
77 	/*
78 	 * This function is also called by the KVM_CREATE_IRQCHIP handler,
79 	 * which had no chance yet to check the availability of the GICv2
80 	 * emulation. So check this here again. KVM_CREATE_DEVICE does
81 	 * the proper checks already.
82 	 */
83 	if (type == KVM_DEV_TYPE_ARM_VGIC_V2 &&
84 		!kvm_vgic_global_state.can_emulate_gicv2)
85 		return -ENODEV;
86 
87 	/* Must be held to avoid race with vCPU creation */
88 	lockdep_assert_held(&kvm->lock);
89 
90 	ret = -EBUSY;
91 	if (!lock_all_vcpus(kvm))
92 		return ret;
93 
94 	mutex_lock(&kvm->arch.config_lock);
95 
96 	if (irqchip_in_kernel(kvm)) {
97 		ret = -EEXIST;
98 		goto out_unlock;
99 	}
100 
101 	kvm_for_each_vcpu(i, vcpu, kvm) {
102 		if (vcpu_has_run_once(vcpu))
103 			goto out_unlock;
104 	}
105 	ret = 0;
106 
107 	if (type == KVM_DEV_TYPE_ARM_VGIC_V2)
108 		kvm->max_vcpus = VGIC_V2_MAX_CPUS;
109 	else
110 		kvm->max_vcpus = VGIC_V3_MAX_CPUS;
111 
112 	if (atomic_read(&kvm->online_vcpus) > kvm->max_vcpus) {
113 		ret = -E2BIG;
114 		goto out_unlock;
115 	}
116 
117 	kvm->arch.vgic.in_kernel = true;
118 	kvm->arch.vgic.vgic_model = type;
119 
120 	kvm->arch.vgic.vgic_dist_base = VGIC_ADDR_UNDEF;
121 
122 	if (type == KVM_DEV_TYPE_ARM_VGIC_V2)
123 		kvm->arch.vgic.vgic_cpu_base = VGIC_ADDR_UNDEF;
124 	else
125 		INIT_LIST_HEAD(&kvm->arch.vgic.rd_regions);
126 
127 out_unlock:
128 	mutex_unlock(&kvm->arch.config_lock);
129 	unlock_all_vcpus(kvm);
130 	return ret;
131 }
132 
133 /* INIT/DESTROY */
134 
135 /**
136  * kvm_vgic_dist_init: initialize the dist data structures
137  * @kvm: kvm struct pointer
138  * @nr_spis: number of spis, frozen by caller
139  */
140 static int kvm_vgic_dist_init(struct kvm *kvm, unsigned int nr_spis)
141 {
142 	struct vgic_dist *dist = &kvm->arch.vgic;
143 	struct kvm_vcpu *vcpu0 = kvm_get_vcpu(kvm, 0);
144 	int i;
145 
146 	dist->spis = kcalloc(nr_spis, sizeof(struct vgic_irq), GFP_KERNEL_ACCOUNT);
147 	if (!dist->spis)
148 		return  -ENOMEM;
149 
150 	/*
151 	 * In the following code we do not take the irq struct lock since
152 	 * no other action on irq structs can happen while the VGIC is
153 	 * not initialized yet:
154 	 * If someone wants to inject an interrupt or does a MMIO access, we
155 	 * require prior initialization in case of a virtual GICv3 or trigger
156 	 * initialization when using a virtual GICv2.
157 	 */
158 	for (i = 0; i < nr_spis; i++) {
159 		struct vgic_irq *irq = &dist->spis[i];
160 
161 		irq->intid = i + VGIC_NR_PRIVATE_IRQS;
162 		INIT_LIST_HEAD(&irq->ap_list);
163 		raw_spin_lock_init(&irq->irq_lock);
164 		irq->vcpu = NULL;
165 		irq->target_vcpu = vcpu0;
166 		kref_init(&irq->refcount);
167 		switch (dist->vgic_model) {
168 		case KVM_DEV_TYPE_ARM_VGIC_V2:
169 			irq->targets = 0;
170 			irq->group = 0;
171 			break;
172 		case KVM_DEV_TYPE_ARM_VGIC_V3:
173 			irq->mpidr = 0;
174 			irq->group = 1;
175 			break;
176 		default:
177 			kfree(dist->spis);
178 			dist->spis = NULL;
179 			return -EINVAL;
180 		}
181 	}
182 	return 0;
183 }
184 
185 /**
186  * kvm_vgic_vcpu_init() - Initialize static VGIC VCPU data
187  * structures and register VCPU-specific KVM iodevs
188  *
189  * @vcpu: pointer to the VCPU being created and initialized
190  *
191  * Only do initialization, but do not actually enable the
192  * VGIC CPU interface
193  */
194 int kvm_vgic_vcpu_init(struct kvm_vcpu *vcpu)
195 {
196 	struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
197 	struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
198 	int ret = 0;
199 	int i;
200 
201 	vgic_cpu->rd_iodev.base_addr = VGIC_ADDR_UNDEF;
202 
203 	INIT_LIST_HEAD(&vgic_cpu->ap_list_head);
204 	raw_spin_lock_init(&vgic_cpu->ap_list_lock);
205 	atomic_set(&vgic_cpu->vgic_v3.its_vpe.vlpi_count, 0);
206 
207 	/*
208 	 * Enable and configure all SGIs to be edge-triggered and
209 	 * configure all PPIs as level-triggered.
210 	 */
211 	for (i = 0; i < VGIC_NR_PRIVATE_IRQS; i++) {
212 		struct vgic_irq *irq = &vgic_cpu->private_irqs[i];
213 
214 		INIT_LIST_HEAD(&irq->ap_list);
215 		raw_spin_lock_init(&irq->irq_lock);
216 		irq->intid = i;
217 		irq->vcpu = NULL;
218 		irq->target_vcpu = vcpu;
219 		kref_init(&irq->refcount);
220 		if (vgic_irq_is_sgi(i)) {
221 			/* SGIs */
222 			irq->enabled = 1;
223 			irq->config = VGIC_CONFIG_EDGE;
224 		} else {
225 			/* PPIs */
226 			irq->config = VGIC_CONFIG_LEVEL;
227 		}
228 	}
229 
230 	if (!irqchip_in_kernel(vcpu->kvm))
231 		return 0;
232 
233 	/*
234 	 * If we are creating a VCPU with a GICv3 we must also register the
235 	 * KVM io device for the redistributor that belongs to this VCPU.
236 	 */
237 	if (dist->vgic_model == KVM_DEV_TYPE_ARM_VGIC_V3) {
238 		mutex_lock(&vcpu->kvm->slots_lock);
239 		ret = vgic_register_redist_iodev(vcpu);
240 		mutex_unlock(&vcpu->kvm->slots_lock);
241 	}
242 	return ret;
243 }
244 
245 static void kvm_vgic_vcpu_enable(struct kvm_vcpu *vcpu)
246 {
247 	if (kvm_vgic_global_state.type == VGIC_V2)
248 		vgic_v2_enable(vcpu);
249 	else
250 		vgic_v3_enable(vcpu);
251 }
252 
253 /*
254  * vgic_init: allocates and initializes dist and vcpu data structures
255  * depending on two dimensioning parameters:
256  * - the number of spis
257  * - the number of vcpus
258  * The function is generally called when nr_spis has been explicitly set
259  * by the guest through the KVM DEVICE API. If not nr_spis is set to 256.
260  * vgic_initialized() returns true when this function has succeeded.
261  */
262 int vgic_init(struct kvm *kvm)
263 {
264 	struct vgic_dist *dist = &kvm->arch.vgic;
265 	struct kvm_vcpu *vcpu;
266 	int ret = 0, i;
267 	unsigned long idx;
268 
269 	lockdep_assert_held(&kvm->arch.config_lock);
270 
271 	if (vgic_initialized(kvm))
272 		return 0;
273 
274 	/* Are we also in the middle of creating a VCPU? */
275 	if (kvm->created_vcpus != atomic_read(&kvm->online_vcpus))
276 		return -EBUSY;
277 
278 	/* freeze the number of spis */
279 	if (!dist->nr_spis)
280 		dist->nr_spis = VGIC_NR_IRQS_LEGACY - VGIC_NR_PRIVATE_IRQS;
281 
282 	ret = kvm_vgic_dist_init(kvm, dist->nr_spis);
283 	if (ret)
284 		goto out;
285 
286 	/* Initialize groups on CPUs created before the VGIC type was known */
287 	kvm_for_each_vcpu(idx, vcpu, kvm) {
288 		struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
289 
290 		for (i = 0; i < VGIC_NR_PRIVATE_IRQS; i++) {
291 			struct vgic_irq *irq = &vgic_cpu->private_irqs[i];
292 			switch (dist->vgic_model) {
293 			case KVM_DEV_TYPE_ARM_VGIC_V3:
294 				irq->group = 1;
295 				irq->mpidr = kvm_vcpu_get_mpidr_aff(vcpu);
296 				break;
297 			case KVM_DEV_TYPE_ARM_VGIC_V2:
298 				irq->group = 0;
299 				irq->targets = 1U << idx;
300 				break;
301 			default:
302 				ret = -EINVAL;
303 				goto out;
304 			}
305 		}
306 	}
307 
308 	if (vgic_has_its(kvm))
309 		vgic_lpi_translation_cache_init(kvm);
310 
311 	/*
312 	 * If we have GICv4.1 enabled, unconditionally request enable the
313 	 * v4 support so that we get HW-accelerated vSGIs. Otherwise, only
314 	 * enable it if we present a virtual ITS to the guest.
315 	 */
316 	if (vgic_supports_direct_msis(kvm)) {
317 		ret = vgic_v4_init(kvm);
318 		if (ret)
319 			goto out;
320 	}
321 
322 	kvm_for_each_vcpu(idx, vcpu, kvm)
323 		kvm_vgic_vcpu_enable(vcpu);
324 
325 	ret = kvm_vgic_setup_default_irq_routing(kvm);
326 	if (ret)
327 		goto out;
328 
329 	vgic_debug_init(kvm);
330 
331 	/*
332 	 * If userspace didn't set the GIC implementation revision,
333 	 * default to the latest and greatest. You know want it.
334 	 */
335 	if (!dist->implementation_rev)
336 		dist->implementation_rev = KVM_VGIC_IMP_REV_LATEST;
337 	dist->initialized = true;
338 
339 out:
340 	return ret;
341 }
342 
343 static void kvm_vgic_dist_destroy(struct kvm *kvm)
344 {
345 	struct vgic_dist *dist = &kvm->arch.vgic;
346 	struct vgic_redist_region *rdreg, *next;
347 
348 	dist->ready = false;
349 	dist->initialized = false;
350 
351 	kfree(dist->spis);
352 	dist->spis = NULL;
353 	dist->nr_spis = 0;
354 	dist->vgic_dist_base = VGIC_ADDR_UNDEF;
355 
356 	if (dist->vgic_model == KVM_DEV_TYPE_ARM_VGIC_V3) {
357 		list_for_each_entry_safe(rdreg, next, &dist->rd_regions, list)
358 			vgic_v3_free_redist_region(rdreg);
359 		INIT_LIST_HEAD(&dist->rd_regions);
360 	} else {
361 		dist->vgic_cpu_base = VGIC_ADDR_UNDEF;
362 	}
363 
364 	if (vgic_has_its(kvm))
365 		vgic_lpi_translation_cache_destroy(kvm);
366 
367 	if (vgic_supports_direct_msis(kvm))
368 		vgic_v4_teardown(kvm);
369 
370 	xa_destroy(&dist->lpi_xa);
371 }
372 
373 static void __kvm_vgic_vcpu_destroy(struct kvm_vcpu *vcpu)
374 {
375 	struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu;
376 
377 	/*
378 	 * Retire all pending LPIs on this vcpu anyway as we're
379 	 * going to destroy it.
380 	 */
381 	vgic_flush_pending_lpis(vcpu);
382 
383 	INIT_LIST_HEAD(&vgic_cpu->ap_list_head);
384 	if (vcpu->kvm->arch.vgic.vgic_model == KVM_DEV_TYPE_ARM_VGIC_V3) {
385 		vgic_unregister_redist_iodev(vcpu);
386 		vgic_cpu->rd_iodev.base_addr = VGIC_ADDR_UNDEF;
387 	}
388 }
389 
390 void kvm_vgic_vcpu_destroy(struct kvm_vcpu *vcpu)
391 {
392 	struct kvm *kvm = vcpu->kvm;
393 
394 	mutex_lock(&kvm->slots_lock);
395 	__kvm_vgic_vcpu_destroy(vcpu);
396 	mutex_unlock(&kvm->slots_lock);
397 }
398 
399 void kvm_vgic_destroy(struct kvm *kvm)
400 {
401 	struct kvm_vcpu *vcpu;
402 	unsigned long i;
403 
404 	mutex_lock(&kvm->slots_lock);
405 
406 	vgic_debug_destroy(kvm);
407 
408 	kvm_for_each_vcpu(i, vcpu, kvm)
409 		__kvm_vgic_vcpu_destroy(vcpu);
410 
411 	mutex_lock(&kvm->arch.config_lock);
412 
413 	kvm_vgic_dist_destroy(kvm);
414 
415 	mutex_unlock(&kvm->arch.config_lock);
416 	mutex_unlock(&kvm->slots_lock);
417 }
418 
419 /**
420  * vgic_lazy_init: Lazy init is only allowed if the GIC exposed to the guest
421  * is a GICv2. A GICv3 must be explicitly initialized by userspace using the
422  * KVM_DEV_ARM_VGIC_GRP_CTRL KVM_DEVICE group.
423  * @kvm: kvm struct pointer
424  */
425 int vgic_lazy_init(struct kvm *kvm)
426 {
427 	int ret = 0;
428 
429 	if (unlikely(!vgic_initialized(kvm))) {
430 		/*
431 		 * We only provide the automatic initialization of the VGIC
432 		 * for the legacy case of a GICv2. Any other type must
433 		 * be explicitly initialized once setup with the respective
434 		 * KVM device call.
435 		 */
436 		if (kvm->arch.vgic.vgic_model != KVM_DEV_TYPE_ARM_VGIC_V2)
437 			return -EBUSY;
438 
439 		mutex_lock(&kvm->arch.config_lock);
440 		ret = vgic_init(kvm);
441 		mutex_unlock(&kvm->arch.config_lock);
442 	}
443 
444 	return ret;
445 }
446 
447 /* RESOURCE MAPPING */
448 
449 /**
450  * kvm_vgic_map_resources - map the MMIO regions
451  * @kvm: kvm struct pointer
452  *
453  * Map the MMIO regions depending on the VGIC model exposed to the guest
454  * called on the first VCPU run.
455  * Also map the virtual CPU interface into the VM.
456  * v2 calls vgic_init() if not already done.
457  * v3 and derivatives return an error if the VGIC is not initialized.
458  * vgic_ready() returns true if this function has succeeded.
459  */
460 int kvm_vgic_map_resources(struct kvm *kvm)
461 {
462 	struct vgic_dist *dist = &kvm->arch.vgic;
463 	enum vgic_type type;
464 	gpa_t dist_base;
465 	int ret = 0;
466 
467 	if (likely(vgic_ready(kvm)))
468 		return 0;
469 
470 	mutex_lock(&kvm->slots_lock);
471 	mutex_lock(&kvm->arch.config_lock);
472 	if (vgic_ready(kvm))
473 		goto out;
474 
475 	if (!irqchip_in_kernel(kvm))
476 		goto out;
477 
478 	if (dist->vgic_model == KVM_DEV_TYPE_ARM_VGIC_V2) {
479 		ret = vgic_v2_map_resources(kvm);
480 		type = VGIC_V2;
481 	} else {
482 		ret = vgic_v3_map_resources(kvm);
483 		type = VGIC_V3;
484 	}
485 
486 	if (ret)
487 		goto out;
488 
489 	dist->ready = true;
490 	dist_base = dist->vgic_dist_base;
491 	mutex_unlock(&kvm->arch.config_lock);
492 
493 	ret = vgic_register_dist_iodev(kvm, dist_base, type);
494 	if (ret)
495 		kvm_err("Unable to register VGIC dist MMIO regions\n");
496 
497 	goto out_slots;
498 out:
499 	mutex_unlock(&kvm->arch.config_lock);
500 out_slots:
501 	mutex_unlock(&kvm->slots_lock);
502 
503 	if (ret)
504 		kvm_vgic_destroy(kvm);
505 
506 	return ret;
507 }
508 
509 /* GENERIC PROBE */
510 
511 void kvm_vgic_cpu_up(void)
512 {
513 	enable_percpu_irq(kvm_vgic_global_state.maint_irq, 0);
514 }
515 
516 
517 void kvm_vgic_cpu_down(void)
518 {
519 	disable_percpu_irq(kvm_vgic_global_state.maint_irq);
520 }
521 
522 static irqreturn_t vgic_maintenance_handler(int irq, void *data)
523 {
524 	/*
525 	 * We cannot rely on the vgic maintenance interrupt to be
526 	 * delivered synchronously. This means we can only use it to
527 	 * exit the VM, and we perform the handling of EOIed
528 	 * interrupts on the exit path (see vgic_fold_lr_state).
529 	 */
530 	return IRQ_HANDLED;
531 }
532 
533 static struct gic_kvm_info *gic_kvm_info;
534 
535 void __init vgic_set_kvm_info(const struct gic_kvm_info *info)
536 {
537 	BUG_ON(gic_kvm_info != NULL);
538 	gic_kvm_info = kmalloc(sizeof(*info), GFP_KERNEL);
539 	if (gic_kvm_info)
540 		*gic_kvm_info = *info;
541 }
542 
543 /**
544  * kvm_vgic_init_cpu_hardware - initialize the GIC VE hardware
545  *
546  * For a specific CPU, initialize the GIC VE hardware.
547  */
548 void kvm_vgic_init_cpu_hardware(void)
549 {
550 	BUG_ON(preemptible());
551 
552 	/*
553 	 * We want to make sure the list registers start out clear so that we
554 	 * only have the program the used registers.
555 	 */
556 	if (kvm_vgic_global_state.type == VGIC_V2)
557 		vgic_v2_init_lrs();
558 	else
559 		kvm_call_hyp(__vgic_v3_init_lrs);
560 }
561 
562 /**
563  * kvm_vgic_hyp_init: populates the kvm_vgic_global_state variable
564  * according to the host GIC model. Accordingly calls either
565  * vgic_v2/v3_probe which registers the KVM_DEVICE that can be
566  * instantiated by a guest later on .
567  */
568 int kvm_vgic_hyp_init(void)
569 {
570 	bool has_mask;
571 	int ret;
572 
573 	if (!gic_kvm_info)
574 		return -ENODEV;
575 
576 	has_mask = !gic_kvm_info->no_maint_irq_mask;
577 
578 	if (has_mask && !gic_kvm_info->maint_irq) {
579 		kvm_err("No vgic maintenance irq\n");
580 		return -ENXIO;
581 	}
582 
583 	/*
584 	 * If we get one of these oddball non-GICs, taint the kernel,
585 	 * as we have no idea of how they *really* behave.
586 	 */
587 	if (gic_kvm_info->no_hw_deactivation) {
588 		kvm_info("Non-architectural vgic, tainting kernel\n");
589 		add_taint(TAINT_CPU_OUT_OF_SPEC, LOCKDEP_STILL_OK);
590 		kvm_vgic_global_state.no_hw_deactivation = true;
591 	}
592 
593 	switch (gic_kvm_info->type) {
594 	case GIC_V2:
595 		ret = vgic_v2_probe(gic_kvm_info);
596 		break;
597 	case GIC_V3:
598 		ret = vgic_v3_probe(gic_kvm_info);
599 		if (!ret) {
600 			static_branch_enable(&kvm_vgic_global_state.gicv3_cpuif);
601 			kvm_info("GIC system register CPU interface enabled\n");
602 		}
603 		break;
604 	default:
605 		ret = -ENODEV;
606 	}
607 
608 	kvm_vgic_global_state.maint_irq = gic_kvm_info->maint_irq;
609 
610 	kfree(gic_kvm_info);
611 	gic_kvm_info = NULL;
612 
613 	if (ret)
614 		return ret;
615 
616 	if (!has_mask && !kvm_vgic_global_state.maint_irq)
617 		return 0;
618 
619 	ret = request_percpu_irq(kvm_vgic_global_state.maint_irq,
620 				 vgic_maintenance_handler,
621 				 "vgic", kvm_get_running_vcpus());
622 	if (ret) {
623 		kvm_err("Cannot register interrupt %d\n",
624 			kvm_vgic_global_state.maint_irq);
625 		return ret;
626 	}
627 
628 	kvm_info("vgic interrupt IRQ%d\n", kvm_vgic_global_state.maint_irq);
629 	return 0;
630 }
631