xref: /linux/arch/riscv/kvm/vm.c (revision e467705a9fb37f51595aa6deaca085ccb4005454)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Copyright (C) 2019 Western Digital Corporation or its affiliates.
4  *
5  * Authors:
6  *     Anup Patel <anup.patel@wdc.com>
7  */
8 
9 #include <linux/errno.h>
10 #include <linux/err.h>
11 #include <linux/module.h>
12 #include <linux/uaccess.h>
13 #include <linux/kvm_host.h>
14 
15 const struct _kvm_stats_desc kvm_vm_stats_desc[] = {
16 	KVM_GENERIC_VM_STATS()
17 };
18 static_assert(ARRAY_SIZE(kvm_vm_stats_desc) ==
19 		sizeof(struct kvm_vm_stat) / sizeof(u64));
20 
21 const struct kvm_stats_header kvm_vm_stats_header = {
22 	.name_size = KVM_STATS_NAME_SIZE,
23 	.num_desc = ARRAY_SIZE(kvm_vm_stats_desc),
24 	.id_offset =  sizeof(struct kvm_stats_header),
25 	.desc_offset = sizeof(struct kvm_stats_header) + KVM_STATS_NAME_SIZE,
26 	.data_offset = sizeof(struct kvm_stats_header) + KVM_STATS_NAME_SIZE +
27 		       sizeof(kvm_vm_stats_desc),
28 };
29 
30 int kvm_arch_init_vm(struct kvm *kvm, unsigned long type)
31 {
32 	int r;
33 
34 	r = kvm_riscv_gstage_alloc_pgd(kvm);
35 	if (r)
36 		return r;
37 
38 	r = kvm_riscv_gstage_vmid_init(kvm);
39 	if (r) {
40 		kvm_riscv_gstage_free_pgd(kvm);
41 		return r;
42 	}
43 
44 	kvm_riscv_aia_init_vm(kvm);
45 
46 	kvm_riscv_guest_timer_init(kvm);
47 
48 	return 0;
49 }
50 
51 void kvm_arch_destroy_vm(struct kvm *kvm)
52 {
53 	kvm_destroy_vcpus(kvm);
54 
55 	kvm_riscv_aia_destroy_vm(kvm);
56 }
57 
58 int kvm_vm_ioctl_irq_line(struct kvm *kvm, struct kvm_irq_level *irql,
59 			  bool line_status)
60 {
61 	if (!irqchip_in_kernel(kvm))
62 		return -ENXIO;
63 
64 	return kvm_riscv_aia_inject_irq(kvm, irql->irq, irql->level);
65 }
66 
67 int kvm_set_msi(struct kvm_kernel_irq_routing_entry *e,
68 		struct kvm *kvm, int irq_source_id,
69 		int level, bool line_status)
70 {
71 	struct kvm_msi msi;
72 
73 	if (!level)
74 		return -1;
75 
76 	msi.address_lo = e->msi.address_lo;
77 	msi.address_hi = e->msi.address_hi;
78 	msi.data = e->msi.data;
79 	msi.flags = e->msi.flags;
80 	msi.devid = e->msi.devid;
81 
82 	return kvm_riscv_aia_inject_msi(kvm, &msi);
83 }
84 
85 static int kvm_riscv_set_irq(struct kvm_kernel_irq_routing_entry *e,
86 			     struct kvm *kvm, int irq_source_id,
87 			     int level, bool line_status)
88 {
89 	return kvm_riscv_aia_inject_irq(kvm, e->irqchip.pin, level);
90 }
91 
92 int kvm_riscv_setup_default_irq_routing(struct kvm *kvm, u32 lines)
93 {
94 	struct kvm_irq_routing_entry *ents;
95 	int i, rc;
96 
97 	ents = kcalloc(lines, sizeof(*ents), GFP_KERNEL);
98 	if (!ents)
99 		return -ENOMEM;
100 
101 	for (i = 0; i < lines; i++) {
102 		ents[i].gsi = i;
103 		ents[i].type = KVM_IRQ_ROUTING_IRQCHIP;
104 		ents[i].u.irqchip.irqchip = 0;
105 		ents[i].u.irqchip.pin = i;
106 	}
107 	rc = kvm_set_irq_routing(kvm, ents, lines, 0);
108 	kfree(ents);
109 
110 	return rc;
111 }
112 
113 bool kvm_arch_can_set_irq_routing(struct kvm *kvm)
114 {
115 	return irqchip_in_kernel(kvm);
116 }
117 
118 int kvm_set_routing_entry(struct kvm *kvm,
119 			  struct kvm_kernel_irq_routing_entry *e,
120 			  const struct kvm_irq_routing_entry *ue)
121 {
122 	int r = -EINVAL;
123 
124 	switch (ue->type) {
125 	case KVM_IRQ_ROUTING_IRQCHIP:
126 		e->set = kvm_riscv_set_irq;
127 		e->irqchip.irqchip = ue->u.irqchip.irqchip;
128 		e->irqchip.pin = ue->u.irqchip.pin;
129 		if ((e->irqchip.pin >= KVM_IRQCHIP_NUM_PINS) ||
130 		    (e->irqchip.irqchip >= KVM_NR_IRQCHIPS))
131 			goto out;
132 		break;
133 	case KVM_IRQ_ROUTING_MSI:
134 		e->set = kvm_set_msi;
135 		e->msi.address_lo = ue->u.msi.address_lo;
136 		e->msi.address_hi = ue->u.msi.address_hi;
137 		e->msi.data = ue->u.msi.data;
138 		e->msi.flags = ue->flags;
139 		e->msi.devid = ue->u.msi.devid;
140 		break;
141 	default:
142 		goto out;
143 	}
144 	r = 0;
145 out:
146 	return r;
147 }
148 
149 int kvm_arch_set_irq_inatomic(struct kvm_kernel_irq_routing_entry *e,
150 			      struct kvm *kvm, int irq_source_id, int level,
151 			      bool line_status)
152 {
153 	if (!level)
154 		return -EWOULDBLOCK;
155 
156 	switch (e->type) {
157 	case KVM_IRQ_ROUTING_MSI:
158 		return kvm_set_msi(e, kvm, irq_source_id, level, line_status);
159 
160 	case KVM_IRQ_ROUTING_IRQCHIP:
161 		return kvm_riscv_set_irq(e, kvm, irq_source_id,
162 					 level, line_status);
163 	}
164 
165 	return -EWOULDBLOCK;
166 }
167 
168 bool kvm_arch_irqchip_in_kernel(struct kvm *kvm)
169 {
170 	return irqchip_in_kernel(kvm);
171 }
172 
173 int kvm_vm_ioctl_check_extension(struct kvm *kvm, long ext)
174 {
175 	int r;
176 
177 	switch (ext) {
178 	case KVM_CAP_IRQCHIP:
179 		r = kvm_riscv_aia_available();
180 		break;
181 	case KVM_CAP_IOEVENTFD:
182 	case KVM_CAP_USER_MEMORY:
183 	case KVM_CAP_SYNC_MMU:
184 	case KVM_CAP_DESTROY_MEMORY_REGION_WORKS:
185 	case KVM_CAP_ONE_REG:
186 	case KVM_CAP_READONLY_MEM:
187 	case KVM_CAP_MP_STATE:
188 	case KVM_CAP_IMMEDIATE_EXIT:
189 	case KVM_CAP_SET_GUEST_DEBUG:
190 		r = 1;
191 		break;
192 	case KVM_CAP_NR_VCPUS:
193 		r = min_t(unsigned int, num_online_cpus(), KVM_MAX_VCPUS);
194 		break;
195 	case KVM_CAP_MAX_VCPUS:
196 		r = KVM_MAX_VCPUS;
197 		break;
198 	case KVM_CAP_NR_MEMSLOTS:
199 		r = KVM_USER_MEM_SLOTS;
200 		break;
201 	case KVM_CAP_VM_GPA_BITS:
202 		r = kvm_riscv_gstage_gpa_bits();
203 		break;
204 	default:
205 		r = 0;
206 		break;
207 	}
208 
209 	return r;
210 }
211 
212 int kvm_arch_vm_ioctl(struct file *filp, unsigned int ioctl, unsigned long arg)
213 {
214 	return -EINVAL;
215 }
216