xref: /linux/arch/um/drivers/vfio_user.c (revision cfc4ca8986bb1f6182da6cd7bb57f228590b4643)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Copyright (C) 2025 Ant Group
4  * Author: Tiwei Bie <tiwei.btw@antgroup.com>
5  */
6 #include <errno.h>
7 #include <fcntl.h>
8 #include <unistd.h>
9 #include <stdio.h>
10 #include <stdint.h>
11 #include <stdlib.h>
12 #include <string.h>
13 #include <sys/ioctl.h>
14 #include <sys/eventfd.h>
15 #include <linux/limits.h>
16 #include <linux/vfio.h>
17 #include <linux/pci_regs.h>
18 #include <as-layout.h>
19 #include <um_malloc.h>
20 
21 #include "vfio_user.h"
22 
uml_vfio_user_open_container(void)23 int uml_vfio_user_open_container(void)
24 {
25 	int r, fd;
26 
27 	fd = open("/dev/vfio/vfio", O_RDWR);
28 	if (fd < 0)
29 		return -errno;
30 
31 	r = ioctl(fd, VFIO_GET_API_VERSION);
32 	if (r != VFIO_API_VERSION) {
33 		r = r < 0 ? -errno : -EINVAL;
34 		goto error;
35 	}
36 
37 	r = ioctl(fd, VFIO_CHECK_EXTENSION, VFIO_TYPE1_IOMMU);
38 	if (r <= 0) {
39 		r = r < 0 ? -errno : -EINVAL;
40 		goto error;
41 	}
42 
43 	return fd;
44 
45 error:
46 	close(fd);
47 	return r;
48 }
49 
uml_vfio_user_setup_iommu(int container)50 int uml_vfio_user_setup_iommu(int container)
51 {
52 	/*
53 	 * This is a bit tricky. See the big comment in
54 	 * vhost_user_set_mem_table() in virtio_uml.c.
55 	 */
56 	unsigned long reserved = uml_reserved - uml_physmem;
57 	struct vfio_iommu_type1_dma_map dma_map = {
58 		.argsz = sizeof(dma_map),
59 		.flags = VFIO_DMA_MAP_FLAG_READ | VFIO_DMA_MAP_FLAG_WRITE,
60 		.vaddr = uml_reserved,
61 		.iova = reserved,
62 		.size = physmem_size - reserved,
63 	};
64 
65 	if (ioctl(container, VFIO_SET_IOMMU, VFIO_TYPE1_IOMMU) < 0)
66 		return -errno;
67 
68 	if (ioctl(container, VFIO_IOMMU_MAP_DMA, &dma_map) < 0)
69 		return -errno;
70 
71 	return 0;
72 }
73 
uml_vfio_user_get_group_id(const char * device)74 int uml_vfio_user_get_group_id(const char *device)
75 {
76 	char *path, *buf, *end;
77 	const char *name;
78 	int r;
79 
80 	path = uml_kmalloc(PATH_MAX, UM_GFP_KERNEL);
81 	if (!path)
82 		return -ENOMEM;
83 
84 	sprintf(path, "/sys/bus/pci/devices/%s/iommu_group", device);
85 
86 	buf = uml_kmalloc(PATH_MAX + 1, UM_GFP_KERNEL);
87 	if (!buf) {
88 		r = -ENOMEM;
89 		goto free_path;
90 	}
91 
92 	r = readlink(path, buf, PATH_MAX);
93 	if (r < 0) {
94 		r = -errno;
95 		goto free_buf;
96 	}
97 	buf[r] = '\0';
98 
99 	name = basename(buf);
100 
101 	r = strtoul(name, &end, 10);
102 	if (*end != '\0' || end == name) {
103 		r = -EINVAL;
104 		goto free_buf;
105 	}
106 
107 free_buf:
108 	kfree(buf);
109 free_path:
110 	kfree(path);
111 	return r;
112 }
113 
uml_vfio_user_open_group(int group_id)114 int uml_vfio_user_open_group(int group_id)
115 {
116 	char *path;
117 	int fd;
118 
119 	path = uml_kmalloc(PATH_MAX, UM_GFP_KERNEL);
120 	if (!path)
121 		return -ENOMEM;
122 
123 	sprintf(path, "/dev/vfio/%d", group_id);
124 
125 	fd = open(path, O_RDWR);
126 	if (fd < 0) {
127 		fd = -errno;
128 		goto out;
129 	}
130 
131 out:
132 	kfree(path);
133 	return fd;
134 }
135 
uml_vfio_user_set_container(int container,int group)136 int uml_vfio_user_set_container(int container, int group)
137 {
138 	if (ioctl(group, VFIO_GROUP_SET_CONTAINER, &container) < 0)
139 		return -errno;
140 	return 0;
141 }
142 
uml_vfio_user_unset_container(int container,int group)143 int uml_vfio_user_unset_container(int container, int group)
144 {
145 	if (ioctl(group, VFIO_GROUP_UNSET_CONTAINER, &container) < 0)
146 		return -errno;
147 	return 0;
148 }
149 
vfio_set_irqs(int device,int start,int count,int * irqfd)150 static int vfio_set_irqs(int device, int start, int count, int *irqfd)
151 {
152 	struct vfio_irq_set *irq_set;
153 	int argsz = sizeof(*irq_set) + sizeof(*irqfd) * count;
154 	int err = 0;
155 
156 	irq_set = uml_kmalloc(argsz, UM_GFP_KERNEL);
157 	if (!irq_set)
158 		return -ENOMEM;
159 
160 	irq_set->argsz = argsz;
161 	irq_set->flags = VFIO_IRQ_SET_DATA_EVENTFD | VFIO_IRQ_SET_ACTION_TRIGGER;
162 	irq_set->index = VFIO_PCI_MSIX_IRQ_INDEX;
163 	irq_set->start = start;
164 	irq_set->count = count;
165 	memcpy(irq_set->data, irqfd, sizeof(*irqfd) * count);
166 
167 	if (ioctl(device, VFIO_DEVICE_SET_IRQS, irq_set) < 0) {
168 		err = -errno;
169 		goto out;
170 	}
171 
172 out:
173 	kfree(irq_set);
174 	return err;
175 }
176 
uml_vfio_user_setup_device(struct uml_vfio_user_device * dev,int group,const char * device)177 int uml_vfio_user_setup_device(struct uml_vfio_user_device *dev,
178 			       int group, const char *device)
179 {
180 	struct vfio_device_info device_info = { .argsz = sizeof(device_info) };
181 	struct vfio_irq_info irq_info = { .argsz = sizeof(irq_info) };
182 	int err, i;
183 
184 	dev->device = ioctl(group, VFIO_GROUP_GET_DEVICE_FD, device);
185 	if (dev->device < 0)
186 		return -errno;
187 
188 	if (ioctl(dev->device, VFIO_DEVICE_GET_INFO, &device_info) < 0) {
189 		err = -errno;
190 		goto close_device;
191 	}
192 
193 	dev->num_regions = device_info.num_regions;
194 	if (dev->num_regions > VFIO_PCI_CONFIG_REGION_INDEX + 1)
195 		dev->num_regions = VFIO_PCI_CONFIG_REGION_INDEX + 1;
196 
197 	dev->region = uml_kmalloc(sizeof(*dev->region) * dev->num_regions,
198 				  UM_GFP_KERNEL);
199 	if (!dev->region) {
200 		err = -ENOMEM;
201 		goto close_device;
202 	}
203 
204 	for (i = 0; i < dev->num_regions; i++) {
205 		struct vfio_region_info region = {
206 			.argsz = sizeof(region),
207 			.index = i,
208 		};
209 		if (ioctl(dev->device, VFIO_DEVICE_GET_REGION_INFO, &region) < 0) {
210 			err = -errno;
211 			goto free_region;
212 		}
213 		dev->region[i].size = region.size;
214 		dev->region[i].offset = region.offset;
215 	}
216 
217 	/* Only MSI-X is supported currently. */
218 	irq_info.index = VFIO_PCI_MSIX_IRQ_INDEX;
219 	if (ioctl(dev->device, VFIO_DEVICE_GET_IRQ_INFO, &irq_info) < 0) {
220 		err = -errno;
221 		goto free_region;
222 	}
223 
224 	dev->irq_count = irq_info.count;
225 
226 	dev->irqfd = uml_kmalloc(sizeof(int) * dev->irq_count, UM_GFP_KERNEL);
227 	if (!dev->irqfd) {
228 		err = -ENOMEM;
229 		goto free_region;
230 	}
231 
232 	memset(dev->irqfd, -1, sizeof(int) * dev->irq_count);
233 
234 	err = vfio_set_irqs(dev->device, 0, dev->irq_count, dev->irqfd);
235 	if (err)
236 		goto free_irqfd;
237 
238 	return 0;
239 
240 free_irqfd:
241 	kfree(dev->irqfd);
242 free_region:
243 	kfree(dev->region);
244 close_device:
245 	close(dev->device);
246 	return err;
247 }
248 
uml_vfio_user_teardown_device(struct uml_vfio_user_device * dev)249 void uml_vfio_user_teardown_device(struct uml_vfio_user_device *dev)
250 {
251 	kfree(dev->irqfd);
252 	kfree(dev->region);
253 	close(dev->device);
254 }
255 
uml_vfio_user_activate_irq(struct uml_vfio_user_device * dev,int index)256 int uml_vfio_user_activate_irq(struct uml_vfio_user_device *dev, int index)
257 {
258 	int irqfd;
259 
260 	irqfd = eventfd(0, EFD_NONBLOCK | EFD_CLOEXEC);
261 	if (irqfd < 0)
262 		return -errno;
263 
264 	dev->irqfd[index] = irqfd;
265 	return irqfd;
266 }
267 
uml_vfio_user_deactivate_irq(struct uml_vfio_user_device * dev,int index)268 void uml_vfio_user_deactivate_irq(struct uml_vfio_user_device *dev, int index)
269 {
270 	close(dev->irqfd[index]);
271 	dev->irqfd[index] = -1;
272 }
273 
uml_vfio_user_update_irqs(struct uml_vfio_user_device * dev)274 int uml_vfio_user_update_irqs(struct uml_vfio_user_device *dev)
275 {
276 	return vfio_set_irqs(dev->device, 0, dev->irq_count, dev->irqfd);
277 }
278 
vfio_region_read(struct uml_vfio_user_device * dev,unsigned int index,uint64_t offset,void * buf,uint64_t size)279 static int vfio_region_read(struct uml_vfio_user_device *dev, unsigned int index,
280 			    uint64_t offset, void *buf, uint64_t size)
281 {
282 	if (index >= dev->num_regions || offset + size > dev->region[index].size)
283 		return -EINVAL;
284 
285 	if (pread(dev->device, buf, size, dev->region[index].offset + offset) < 0)
286 		return -errno;
287 
288 	return 0;
289 }
290 
vfio_region_write(struct uml_vfio_user_device * dev,unsigned int index,uint64_t offset,const void * buf,uint64_t size)291 static int vfio_region_write(struct uml_vfio_user_device *dev, unsigned int index,
292 			     uint64_t offset, const void *buf, uint64_t size)
293 {
294 	if (index >= dev->num_regions || offset + size > dev->region[index].size)
295 		return -EINVAL;
296 
297 	if (pwrite(dev->device, buf, size, dev->region[index].offset + offset) < 0)
298 		return -errno;
299 
300 	return 0;
301 }
302 
uml_vfio_user_cfgspace_read(struct uml_vfio_user_device * dev,unsigned int offset,void * buf,int size)303 int uml_vfio_user_cfgspace_read(struct uml_vfio_user_device *dev,
304 				unsigned int offset, void *buf, int size)
305 {
306 	return vfio_region_read(dev, VFIO_PCI_CONFIG_REGION_INDEX,
307 				offset, buf, size);
308 }
309 
uml_vfio_user_cfgspace_write(struct uml_vfio_user_device * dev,unsigned int offset,const void * buf,int size)310 int uml_vfio_user_cfgspace_write(struct uml_vfio_user_device *dev,
311 				 unsigned int offset, const void *buf, int size)
312 {
313 	return vfio_region_write(dev, VFIO_PCI_CONFIG_REGION_INDEX,
314 				 offset, buf, size);
315 }
316 
uml_vfio_user_bar_read(struct uml_vfio_user_device * dev,int bar,unsigned int offset,void * buf,int size)317 int uml_vfio_user_bar_read(struct uml_vfio_user_device *dev, int bar,
318 			   unsigned int offset, void *buf, int size)
319 {
320 	return vfio_region_read(dev, bar, offset, buf, size);
321 }
322 
uml_vfio_user_bar_write(struct uml_vfio_user_device * dev,int bar,unsigned int offset,const void * buf,int size)323 int uml_vfio_user_bar_write(struct uml_vfio_user_device *dev, int bar,
324 			    unsigned int offset, const void *buf, int size)
325 {
326 	return vfio_region_write(dev, bar, offset, buf, size);
327 }
328