xref: /linux/arch/riscv/kvm/aia.c (revision b8e85e6f3a09fc56b0ff574887798962ef8a8f80)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Copyright (C) 2021 Western Digital Corporation or its affiliates.
4  * Copyright (C) 2022 Ventana Micro Systems Inc.
5  *
6  * Authors:
7  *	Anup Patel <apatel@ventanamicro.com>
8  */
9 
10 #include <linux/kernel.h>
11 #include <linux/bitops.h>
12 #include <linux/irq.h>
13 #include <linux/irqdomain.h>
14 #include <linux/kvm_host.h>
15 #include <linux/percpu.h>
16 #include <linux/spinlock.h>
17 #include <asm/cpufeature.h>
18 #include <asm/kvm_aia_imsic.h>
19 
20 struct aia_hgei_control {
21 	raw_spinlock_t lock;
22 	unsigned long free_bitmap;
23 	struct kvm_vcpu *owners[BITS_PER_LONG];
24 };
25 static DEFINE_PER_CPU(struct aia_hgei_control, aia_hgei);
26 static int hgei_parent_irq;
27 
28 unsigned int kvm_riscv_aia_nr_hgei;
29 unsigned int kvm_riscv_aia_max_ids;
30 DEFINE_STATIC_KEY_FALSE(kvm_riscv_aia_available);
31 
32 static int aia_find_hgei(struct kvm_vcpu *owner)
33 {
34 	int i, hgei;
35 	unsigned long flags;
36 	struct aia_hgei_control *hgctrl = get_cpu_ptr(&aia_hgei);
37 
38 	raw_spin_lock_irqsave(&hgctrl->lock, flags);
39 
40 	hgei = -1;
41 	for (i = 1; i <= kvm_riscv_aia_nr_hgei; i++) {
42 		if (hgctrl->owners[i] == owner) {
43 			hgei = i;
44 			break;
45 		}
46 	}
47 
48 	raw_spin_unlock_irqrestore(&hgctrl->lock, flags);
49 
50 	put_cpu_ptr(&aia_hgei);
51 	return hgei;
52 }
53 
54 static void aia_set_hvictl(bool ext_irq_pending)
55 {
56 	unsigned long hvictl;
57 
58 	/*
59 	 * HVICTL.IID == 9 and HVICTL.IPRIO == 0 represents
60 	 * no interrupt in HVICTL.
61 	 */
62 
63 	hvictl = (IRQ_S_EXT << HVICTL_IID_SHIFT) & HVICTL_IID;
64 	hvictl |= ext_irq_pending;
65 	csr_write(CSR_HVICTL, hvictl);
66 }
67 
68 #ifdef CONFIG_32BIT
69 void kvm_riscv_vcpu_aia_flush_interrupts(struct kvm_vcpu *vcpu)
70 {
71 	struct kvm_vcpu_aia_csr *csr = &vcpu->arch.aia_context.guest_csr;
72 	unsigned long mask, val;
73 
74 	if (!kvm_riscv_aia_available())
75 		return;
76 
77 	if (READ_ONCE(vcpu->arch.irqs_pending_mask[1])) {
78 		mask = xchg_acquire(&vcpu->arch.irqs_pending_mask[1], 0);
79 		val = READ_ONCE(vcpu->arch.irqs_pending[1]) & mask;
80 
81 		csr->hviph &= ~mask;
82 		csr->hviph |= val;
83 	}
84 }
85 
86 void kvm_riscv_vcpu_aia_sync_interrupts(struct kvm_vcpu *vcpu)
87 {
88 	struct kvm_vcpu_aia_csr *csr = &vcpu->arch.aia_context.guest_csr;
89 
90 	if (kvm_riscv_aia_available())
91 		csr->vsieh = csr_read(CSR_VSIEH);
92 }
93 #endif
94 
95 bool kvm_riscv_vcpu_aia_has_interrupts(struct kvm_vcpu *vcpu, u64 mask)
96 {
97 	int hgei;
98 	unsigned long seip;
99 
100 	if (!kvm_riscv_aia_available())
101 		return false;
102 
103 #ifdef CONFIG_32BIT
104 	if (READ_ONCE(vcpu->arch.irqs_pending[1]) &
105 	    (vcpu->arch.aia_context.guest_csr.vsieh & upper_32_bits(mask)))
106 		return true;
107 #endif
108 
109 	seip = vcpu->arch.guest_csr.vsie;
110 	seip &= (unsigned long)mask;
111 	seip &= BIT(IRQ_S_EXT);
112 
113 	if (!kvm_riscv_aia_initialized(vcpu->kvm) || !seip)
114 		return false;
115 
116 	hgei = aia_find_hgei(vcpu);
117 	if (hgei > 0)
118 		return !!(csr_read(CSR_HGEIP) & BIT(hgei));
119 
120 	return false;
121 }
122 
123 void kvm_riscv_vcpu_aia_update_hvip(struct kvm_vcpu *vcpu)
124 {
125 	struct kvm_vcpu_csr *csr = &vcpu->arch.guest_csr;
126 
127 	if (!kvm_riscv_aia_available())
128 		return;
129 
130 #ifdef CONFIG_32BIT
131 	csr_write(CSR_HVIPH, vcpu->arch.aia_context.guest_csr.hviph);
132 #endif
133 	aia_set_hvictl(!!(csr->hvip & BIT(IRQ_VS_EXT)));
134 }
135 
136 void kvm_riscv_vcpu_aia_load(struct kvm_vcpu *vcpu, int cpu)
137 {
138 	struct kvm_vcpu_aia_csr *csr = &vcpu->arch.aia_context.guest_csr;
139 
140 	if (!kvm_riscv_aia_available())
141 		return;
142 
143 	csr_write(CSR_VSISELECT, csr->vsiselect);
144 	csr_write(CSR_HVIPRIO1, csr->hviprio1);
145 	csr_write(CSR_HVIPRIO2, csr->hviprio2);
146 #ifdef CONFIG_32BIT
147 	csr_write(CSR_VSIEH, csr->vsieh);
148 	csr_write(CSR_HVIPH, csr->hviph);
149 	csr_write(CSR_HVIPRIO1H, csr->hviprio1h);
150 	csr_write(CSR_HVIPRIO2H, csr->hviprio2h);
151 #endif
152 }
153 
154 void kvm_riscv_vcpu_aia_put(struct kvm_vcpu *vcpu)
155 {
156 	struct kvm_vcpu_aia_csr *csr = &vcpu->arch.aia_context.guest_csr;
157 
158 	if (!kvm_riscv_aia_available())
159 		return;
160 
161 	csr->vsiselect = csr_read(CSR_VSISELECT);
162 	csr->hviprio1 = csr_read(CSR_HVIPRIO1);
163 	csr->hviprio2 = csr_read(CSR_HVIPRIO2);
164 #ifdef CONFIG_32BIT
165 	csr->vsieh = csr_read(CSR_VSIEH);
166 	csr->hviph = csr_read(CSR_HVIPH);
167 	csr->hviprio1h = csr_read(CSR_HVIPRIO1H);
168 	csr->hviprio2h = csr_read(CSR_HVIPRIO2H);
169 #endif
170 }
171 
172 int kvm_riscv_vcpu_aia_get_csr(struct kvm_vcpu *vcpu,
173 			       unsigned long reg_num,
174 			       unsigned long *out_val)
175 {
176 	struct kvm_vcpu_aia_csr *csr = &vcpu->arch.aia_context.guest_csr;
177 
178 	if (reg_num >= sizeof(struct kvm_riscv_aia_csr) / sizeof(unsigned long))
179 		return -ENOENT;
180 
181 	*out_val = 0;
182 	if (kvm_riscv_aia_available())
183 		*out_val = ((unsigned long *)csr)[reg_num];
184 
185 	return 0;
186 }
187 
188 int kvm_riscv_vcpu_aia_set_csr(struct kvm_vcpu *vcpu,
189 			       unsigned long reg_num,
190 			       unsigned long val)
191 {
192 	struct kvm_vcpu_aia_csr *csr = &vcpu->arch.aia_context.guest_csr;
193 
194 	if (reg_num >= sizeof(struct kvm_riscv_aia_csr) / sizeof(unsigned long))
195 		return -ENOENT;
196 
197 	if (kvm_riscv_aia_available()) {
198 		((unsigned long *)csr)[reg_num] = val;
199 
200 #ifdef CONFIG_32BIT
201 		if (reg_num == KVM_REG_RISCV_CSR_AIA_REG(siph))
202 			WRITE_ONCE(vcpu->arch.irqs_pending_mask[1], 0);
203 #endif
204 	}
205 
206 	return 0;
207 }
208 
209 int kvm_riscv_vcpu_aia_rmw_topei(struct kvm_vcpu *vcpu,
210 				 unsigned int csr_num,
211 				 unsigned long *val,
212 				 unsigned long new_val,
213 				 unsigned long wr_mask)
214 {
215 	/* If AIA not available then redirect trap */
216 	if (!kvm_riscv_aia_available())
217 		return KVM_INSN_ILLEGAL_TRAP;
218 
219 	/* If AIA not initialized then forward to user space */
220 	if (!kvm_riscv_aia_initialized(vcpu->kvm))
221 		return KVM_INSN_EXIT_TO_USER_SPACE;
222 
223 	return kvm_riscv_vcpu_aia_imsic_rmw(vcpu, KVM_RISCV_AIA_IMSIC_TOPEI,
224 					    val, new_val, wr_mask);
225 }
226 
227 /*
228  * External IRQ priority always read-only zero. This means default
229  * priority order  is always preferred for external IRQs unless
230  * HVICTL.IID == 9 and HVICTL.IPRIO != 0
231  */
232 static int aia_irq2bitpos[] = {
233 0,     8,   -1,   -1,   16,   24,   -1,   -1, /* 0 - 7 */
234 32,   -1,   -1,   -1,   -1,   40,   48,   56, /* 8 - 15 */
235 64,   72,   80,   88,   96,  104,  112,  120, /* 16 - 23 */
236 -1,   -1,   -1,   -1,   -1,   -1,   -1,   -1, /* 24 - 31 */
237 -1,   -1,   -1,   -1,   -1,   -1,   -1,   -1, /* 32 - 39 */
238 -1,   -1,   -1,   -1,   -1,   -1,   -1,   -1, /* 40 - 47 */
239 -1,   -1,   -1,   -1,   -1,   -1,   -1,   -1, /* 48 - 55 */
240 -1,   -1,   -1,   -1,   -1,   -1,   -1,   -1, /* 56 - 63 */
241 };
242 
243 static u8 aia_get_iprio8(struct kvm_vcpu *vcpu, unsigned int irq)
244 {
245 	unsigned long hviprio;
246 	int bitpos = aia_irq2bitpos[irq];
247 
248 	if (bitpos < 0)
249 		return 0;
250 
251 	switch (bitpos / BITS_PER_LONG) {
252 	case 0:
253 		hviprio = csr_read(CSR_HVIPRIO1);
254 		break;
255 	case 1:
256 #ifndef CONFIG_32BIT
257 		hviprio = csr_read(CSR_HVIPRIO2);
258 		break;
259 #else
260 		hviprio = csr_read(CSR_HVIPRIO1H);
261 		break;
262 	case 2:
263 		hviprio = csr_read(CSR_HVIPRIO2);
264 		break;
265 	case 3:
266 		hviprio = csr_read(CSR_HVIPRIO2H);
267 		break;
268 #endif
269 	default:
270 		return 0;
271 	}
272 
273 	return (hviprio >> (bitpos % BITS_PER_LONG)) & TOPI_IPRIO_MASK;
274 }
275 
276 static void aia_set_iprio8(struct kvm_vcpu *vcpu, unsigned int irq, u8 prio)
277 {
278 	unsigned long hviprio;
279 	int bitpos = aia_irq2bitpos[irq];
280 
281 	if (bitpos < 0)
282 		return;
283 
284 	switch (bitpos / BITS_PER_LONG) {
285 	case 0:
286 		hviprio = csr_read(CSR_HVIPRIO1);
287 		break;
288 	case 1:
289 #ifndef CONFIG_32BIT
290 		hviprio = csr_read(CSR_HVIPRIO2);
291 		break;
292 #else
293 		hviprio = csr_read(CSR_HVIPRIO1H);
294 		break;
295 	case 2:
296 		hviprio = csr_read(CSR_HVIPRIO2);
297 		break;
298 	case 3:
299 		hviprio = csr_read(CSR_HVIPRIO2H);
300 		break;
301 #endif
302 	default:
303 		return;
304 	}
305 
306 	hviprio &= ~(TOPI_IPRIO_MASK << (bitpos % BITS_PER_LONG));
307 	hviprio |= (unsigned long)prio << (bitpos % BITS_PER_LONG);
308 
309 	switch (bitpos / BITS_PER_LONG) {
310 	case 0:
311 		csr_write(CSR_HVIPRIO1, hviprio);
312 		break;
313 	case 1:
314 #ifndef CONFIG_32BIT
315 		csr_write(CSR_HVIPRIO2, hviprio);
316 		break;
317 #else
318 		csr_write(CSR_HVIPRIO1H, hviprio);
319 		break;
320 	case 2:
321 		csr_write(CSR_HVIPRIO2, hviprio);
322 		break;
323 	case 3:
324 		csr_write(CSR_HVIPRIO2H, hviprio);
325 		break;
326 #endif
327 	default:
328 		return;
329 	}
330 }
331 
332 static int aia_rmw_iprio(struct kvm_vcpu *vcpu, unsigned int isel,
333 			 unsigned long *val, unsigned long new_val,
334 			 unsigned long wr_mask)
335 {
336 	int i, first_irq, nirqs;
337 	unsigned long old_val;
338 	u8 prio;
339 
340 #ifndef CONFIG_32BIT
341 	if (isel & 0x1)
342 		return KVM_INSN_ILLEGAL_TRAP;
343 #endif
344 
345 	nirqs = 4 * (BITS_PER_LONG / 32);
346 	first_irq = (isel - ISELECT_IPRIO0) * 4;
347 
348 	old_val = 0;
349 	for (i = 0; i < nirqs; i++) {
350 		prio = aia_get_iprio8(vcpu, first_irq + i);
351 		old_val |= (unsigned long)prio << (TOPI_IPRIO_BITS * i);
352 	}
353 
354 	if (val)
355 		*val = old_val;
356 
357 	if (wr_mask) {
358 		new_val = (old_val & ~wr_mask) | (new_val & wr_mask);
359 		for (i = 0; i < nirqs; i++) {
360 			prio = (new_val >> (TOPI_IPRIO_BITS * i)) &
361 				TOPI_IPRIO_MASK;
362 			aia_set_iprio8(vcpu, first_irq + i, prio);
363 		}
364 	}
365 
366 	return KVM_INSN_CONTINUE_NEXT_SEPC;
367 }
368 
369 int kvm_riscv_vcpu_aia_rmw_ireg(struct kvm_vcpu *vcpu, unsigned int csr_num,
370 				unsigned long *val, unsigned long new_val,
371 				unsigned long wr_mask)
372 {
373 	unsigned int isel;
374 
375 	/* If AIA not available then redirect trap */
376 	if (!kvm_riscv_aia_available())
377 		return KVM_INSN_ILLEGAL_TRAP;
378 
379 	/* First try to emulate in kernel space */
380 	isel = csr_read(CSR_VSISELECT) & ISELECT_MASK;
381 	if (isel >= ISELECT_IPRIO0 && isel <= ISELECT_IPRIO15)
382 		return aia_rmw_iprio(vcpu, isel, val, new_val, wr_mask);
383 	else if (isel >= IMSIC_FIRST && isel <= IMSIC_LAST &&
384 		 kvm_riscv_aia_initialized(vcpu->kvm))
385 		return kvm_riscv_vcpu_aia_imsic_rmw(vcpu, isel, val, new_val,
386 						    wr_mask);
387 
388 	/* We can't handle it here so redirect to user space */
389 	return KVM_INSN_EXIT_TO_USER_SPACE;
390 }
391 
392 int kvm_riscv_aia_alloc_hgei(int cpu, struct kvm_vcpu *owner,
393 			     void __iomem **hgei_va, phys_addr_t *hgei_pa)
394 {
395 	int ret = -ENOENT;
396 	unsigned long flags;
397 	struct aia_hgei_control *hgctrl = per_cpu_ptr(&aia_hgei, cpu);
398 
399 	if (!kvm_riscv_aia_available() || !hgctrl)
400 		return -ENODEV;
401 
402 	raw_spin_lock_irqsave(&hgctrl->lock, flags);
403 
404 	if (hgctrl->free_bitmap) {
405 		ret = __ffs(hgctrl->free_bitmap);
406 		hgctrl->free_bitmap &= ~BIT(ret);
407 		hgctrl->owners[ret] = owner;
408 	}
409 
410 	raw_spin_unlock_irqrestore(&hgctrl->lock, flags);
411 
412 	/* TODO: To be updated later by AIA IMSIC HW guest file support */
413 	if (hgei_va)
414 		*hgei_va = NULL;
415 	if (hgei_pa)
416 		*hgei_pa = 0;
417 
418 	return ret;
419 }
420 
421 void kvm_riscv_aia_free_hgei(int cpu, int hgei)
422 {
423 	unsigned long flags;
424 	struct aia_hgei_control *hgctrl = per_cpu_ptr(&aia_hgei, cpu);
425 
426 	if (!kvm_riscv_aia_available() || !hgctrl)
427 		return;
428 
429 	raw_spin_lock_irqsave(&hgctrl->lock, flags);
430 
431 	if (hgei > 0 && hgei <= kvm_riscv_aia_nr_hgei) {
432 		if (!(hgctrl->free_bitmap & BIT(hgei))) {
433 			hgctrl->free_bitmap |= BIT(hgei);
434 			hgctrl->owners[hgei] = NULL;
435 		}
436 	}
437 
438 	raw_spin_unlock_irqrestore(&hgctrl->lock, flags);
439 }
440 
441 void kvm_riscv_aia_wakeon_hgei(struct kvm_vcpu *owner, bool enable)
442 {
443 	int hgei;
444 
445 	if (!kvm_riscv_aia_available())
446 		return;
447 
448 	hgei = aia_find_hgei(owner);
449 	if (hgei > 0) {
450 		if (enable)
451 			csr_set(CSR_HGEIE, BIT(hgei));
452 		else
453 			csr_clear(CSR_HGEIE, BIT(hgei));
454 	}
455 }
456 
457 static irqreturn_t hgei_interrupt(int irq, void *dev_id)
458 {
459 	int i;
460 	unsigned long hgei_mask, flags;
461 	struct aia_hgei_control *hgctrl = get_cpu_ptr(&aia_hgei);
462 
463 	hgei_mask = csr_read(CSR_HGEIP) & csr_read(CSR_HGEIE);
464 	csr_clear(CSR_HGEIE, hgei_mask);
465 
466 	raw_spin_lock_irqsave(&hgctrl->lock, flags);
467 
468 	for_each_set_bit(i, &hgei_mask, BITS_PER_LONG) {
469 		if (hgctrl->owners[i])
470 			kvm_vcpu_kick(hgctrl->owners[i]);
471 	}
472 
473 	raw_spin_unlock_irqrestore(&hgctrl->lock, flags);
474 
475 	put_cpu_ptr(&aia_hgei);
476 	return IRQ_HANDLED;
477 }
478 
479 static int aia_hgei_init(void)
480 {
481 	int cpu, rc;
482 	struct irq_domain *domain;
483 	struct aia_hgei_control *hgctrl;
484 
485 	/* Initialize per-CPU guest external interrupt line management */
486 	for_each_possible_cpu(cpu) {
487 		hgctrl = per_cpu_ptr(&aia_hgei, cpu);
488 		raw_spin_lock_init(&hgctrl->lock);
489 		if (kvm_riscv_aia_nr_hgei) {
490 			hgctrl->free_bitmap =
491 				BIT(kvm_riscv_aia_nr_hgei + 1) - 1;
492 			hgctrl->free_bitmap &= ~BIT(0);
493 		} else
494 			hgctrl->free_bitmap = 0;
495 	}
496 
497 	/* Find INTC irq domain */
498 	domain = irq_find_matching_fwnode(riscv_get_intc_hwnode(),
499 					  DOMAIN_BUS_ANY);
500 	if (!domain) {
501 		kvm_err("unable to find INTC domain\n");
502 		return -ENOENT;
503 	}
504 
505 	/* Map per-CPU SGEI interrupt from INTC domain */
506 	hgei_parent_irq = irq_create_mapping(domain, IRQ_S_GEXT);
507 	if (!hgei_parent_irq) {
508 		kvm_err("unable to map SGEI IRQ\n");
509 		return -ENOMEM;
510 	}
511 
512 	/* Request per-CPU SGEI interrupt */
513 	rc = request_percpu_irq(hgei_parent_irq, hgei_interrupt,
514 				"riscv-kvm", &aia_hgei);
515 	if (rc) {
516 		kvm_err("failed to request SGEI IRQ\n");
517 		return rc;
518 	}
519 
520 	return 0;
521 }
522 
523 static void aia_hgei_exit(void)
524 {
525 	/* Free per-CPU SGEI interrupt */
526 	free_percpu_irq(hgei_parent_irq, &aia_hgei);
527 }
528 
529 void kvm_riscv_aia_enable(void)
530 {
531 	if (!kvm_riscv_aia_available())
532 		return;
533 
534 	aia_set_hvictl(false);
535 	csr_write(CSR_HVIPRIO1, 0x0);
536 	csr_write(CSR_HVIPRIO2, 0x0);
537 #ifdef CONFIG_32BIT
538 	csr_write(CSR_HVIPH, 0x0);
539 	csr_write(CSR_HIDELEGH, 0x0);
540 	csr_write(CSR_HVIPRIO1H, 0x0);
541 	csr_write(CSR_HVIPRIO2H, 0x0);
542 #endif
543 
544 	/* Enable per-CPU SGEI interrupt */
545 	enable_percpu_irq(hgei_parent_irq,
546 			  irq_get_trigger_type(hgei_parent_irq));
547 	csr_set(CSR_HIE, BIT(IRQ_S_GEXT));
548 }
549 
550 void kvm_riscv_aia_disable(void)
551 {
552 	int i;
553 	unsigned long flags;
554 	struct kvm_vcpu *vcpu;
555 	struct aia_hgei_control *hgctrl;
556 
557 	if (!kvm_riscv_aia_available())
558 		return;
559 	hgctrl = get_cpu_ptr(&aia_hgei);
560 
561 	/* Disable per-CPU SGEI interrupt */
562 	csr_clear(CSR_HIE, BIT(IRQ_S_GEXT));
563 	disable_percpu_irq(hgei_parent_irq);
564 
565 	aia_set_hvictl(false);
566 
567 	raw_spin_lock_irqsave(&hgctrl->lock, flags);
568 
569 	for (i = 0; i <= kvm_riscv_aia_nr_hgei; i++) {
570 		vcpu = hgctrl->owners[i];
571 		if (!vcpu)
572 			continue;
573 
574 		/*
575 		 * We release hgctrl->lock before notifying IMSIC
576 		 * so that we don't have lock ordering issues.
577 		 */
578 		raw_spin_unlock_irqrestore(&hgctrl->lock, flags);
579 
580 		/* Notify IMSIC */
581 		kvm_riscv_vcpu_aia_imsic_release(vcpu);
582 
583 		/*
584 		 * Wakeup VCPU if it was blocked so that it can
585 		 * run on other HARTs
586 		 */
587 		if (csr_read(CSR_HGEIE) & BIT(i)) {
588 			csr_clear(CSR_HGEIE, BIT(i));
589 			kvm_vcpu_kick(vcpu);
590 		}
591 
592 		raw_spin_lock_irqsave(&hgctrl->lock, flags);
593 	}
594 
595 	raw_spin_unlock_irqrestore(&hgctrl->lock, flags);
596 
597 	put_cpu_ptr(&aia_hgei);
598 }
599 
600 int kvm_riscv_aia_init(void)
601 {
602 	int rc;
603 
604 	if (!riscv_isa_extension_available(NULL, SxAIA))
605 		return -ENODEV;
606 
607 	/* Figure-out number of bits in HGEIE */
608 	csr_write(CSR_HGEIE, -1UL);
609 	kvm_riscv_aia_nr_hgei = fls_long(csr_read(CSR_HGEIE));
610 	csr_write(CSR_HGEIE, 0);
611 	if (kvm_riscv_aia_nr_hgei)
612 		kvm_riscv_aia_nr_hgei--;
613 
614 	/*
615 	 * Number of usable HGEI lines should be minimum of per-HART
616 	 * IMSIC guest files and number of bits in HGEIE
617 	 *
618 	 * TODO: To be updated later by AIA IMSIC HW guest file support
619 	 */
620 	kvm_riscv_aia_nr_hgei = 0;
621 
622 	/*
623 	 * Find number of guest MSI IDs
624 	 *
625 	 * TODO: To be updated later by AIA IMSIC HW guest file support
626 	 */
627 	kvm_riscv_aia_max_ids = IMSIC_MAX_ID;
628 
629 	/* Initialize guest external interrupt line management */
630 	rc = aia_hgei_init();
631 	if (rc)
632 		return rc;
633 
634 	/* Register device operations */
635 	rc = kvm_register_device_ops(&kvm_riscv_aia_device_ops,
636 				     KVM_DEV_TYPE_RISCV_AIA);
637 	if (rc) {
638 		aia_hgei_exit();
639 		return rc;
640 	}
641 
642 	/* Enable KVM AIA support */
643 	static_branch_enable(&kvm_riscv_aia_available);
644 
645 	return 0;
646 }
647 
648 void kvm_riscv_aia_exit(void)
649 {
650 	if (!kvm_riscv_aia_available())
651 		return;
652 
653 	/* Unregister device operations */
654 	kvm_unregister_device_ops(KVM_DEV_TYPE_RISCV_AIA);
655 
656 	/* Cleanup the HGEI state */
657 	aia_hgei_exit();
658 }
659