xref: /linux/drivers/vfio/pci/pds/cmds.c (revision 409c38d4f156740bf3165fd6ceae4fa6425eebf4)
1 // SPDX-License-Identifier: GPL-2.0
2 /* Copyright(c) 2023 Advanced Micro Devices, Inc. */
3 
4 #include <linux/io.h>
5 #include <linux/types.h>
6 #include <linux/delay.h>
7 
8 #include <linux/pds/pds_common.h>
9 #include <linux/pds/pds_core_if.h>
10 #include <linux/pds/pds_adminq.h>
11 
12 #include "vfio_dev.h"
13 #include "cmds.h"
14 
15 #define SUSPEND_TIMEOUT_S		5
16 #define SUSPEND_CHECK_INTERVAL_MS	1
17 
18 static int pds_vfio_client_adminq_cmd(struct pds_vfio_pci_device *pds_vfio,
19 				      union pds_core_adminq_cmd *req,
20 				      union pds_core_adminq_comp *resp,
21 				      bool fast_poll)
22 {
23 	struct pci_dev *pdev = pds_vfio_to_pci_dev(pds_vfio);
24 	union pds_core_adminq_cmd cmd = {};
25 	struct pdsc *pdsc;
26 	int err;
27 
28 	/* Wrap the client request */
29 	cmd.client_request.opcode = PDS_AQ_CMD_CLIENT_CMD;
30 	cmd.client_request.client_id = cpu_to_le16(pds_vfio->client_id);
31 	memcpy(cmd.client_request.client_cmd, req,
32 	       sizeof(cmd.client_request.client_cmd));
33 
34 	pdsc = pdsc_get_pf_struct(pdev);
35 	if (IS_ERR(pdsc))
36 		return PTR_ERR(pdsc);
37 
38 	err = pdsc_adminq_post(pdsc, &cmd, resp, fast_poll);
39 	if (err && err != -EAGAIN)
40 		dev_err(pds_vfio_to_dev(pds_vfio),
41 			"client admin cmd failed: %pe\n", ERR_PTR(err));
42 
43 	return err;
44 }
45 
46 int pds_vfio_register_client_cmd(struct pds_vfio_pci_device *pds_vfio)
47 {
48 	struct pci_dev *pdev = pds_vfio_to_pci_dev(pds_vfio);
49 	char devname[PDS_DEVNAME_LEN];
50 	struct pdsc *pdsc;
51 	int ci;
52 
53 	snprintf(devname, sizeof(devname), "%s.%d-%u", PDS_VFIO_LM_DEV_NAME,
54 		 pci_domain_nr(pdev->bus),
55 		 PCI_DEVID(pdev->bus->number, pdev->devfn));
56 
57 	pdsc = pdsc_get_pf_struct(pdev);
58 	if (IS_ERR(pdsc))
59 		return PTR_ERR(pdsc);
60 
61 	ci = pds_client_register(pdsc, devname);
62 	if (ci < 0)
63 		return ci;
64 
65 	pds_vfio->client_id = ci;
66 
67 	return 0;
68 }
69 
70 void pds_vfio_unregister_client_cmd(struct pds_vfio_pci_device *pds_vfio)
71 {
72 	struct pci_dev *pdev = pds_vfio_to_pci_dev(pds_vfio);
73 	struct pdsc *pdsc;
74 	int err;
75 
76 	pdsc = pdsc_get_pf_struct(pdev);
77 	if (IS_ERR(pdsc))
78 		return;
79 
80 	err = pds_client_unregister(pdsc, pds_vfio->client_id);
81 	if (err)
82 		dev_err(&pdev->dev, "unregister from DSC failed: %pe\n",
83 			ERR_PTR(err));
84 
85 	pds_vfio->client_id = 0;
86 }
87 
88 static int
89 pds_vfio_suspend_wait_device_cmd(struct pds_vfio_pci_device *pds_vfio, u8 type)
90 {
91 	union pds_core_adminq_cmd cmd = {
92 		.lm_suspend_status = {
93 			.opcode = PDS_LM_CMD_SUSPEND_STATUS,
94 			.vf_id = cpu_to_le16(pds_vfio->vf_id),
95 			.type = type,
96 		},
97 	};
98 	struct device *dev = pds_vfio_to_dev(pds_vfio);
99 	union pds_core_adminq_comp comp = {};
100 	unsigned long time_limit;
101 	unsigned long time_start;
102 	unsigned long time_done;
103 	int err;
104 
105 	time_start = jiffies;
106 	time_limit = time_start + HZ * SUSPEND_TIMEOUT_S;
107 	do {
108 		err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, &comp, true);
109 		if (err != -EAGAIN)
110 			break;
111 
112 		msleep(SUSPEND_CHECK_INTERVAL_MS);
113 	} while (time_before(jiffies, time_limit));
114 
115 	time_done = jiffies;
116 	dev_dbg(dev, "%s: vf%u: Suspend comp received in %d msecs\n", __func__,
117 		pds_vfio->vf_id, jiffies_to_msecs(time_done - time_start));
118 
119 	/* Check the results */
120 	if (time_after_eq(time_done, time_limit)) {
121 		dev_err(dev, "%s: vf%u: Suspend comp timeout\n", __func__,
122 			pds_vfio->vf_id);
123 		err = -ETIMEDOUT;
124 	}
125 
126 	return err;
127 }
128 
129 int pds_vfio_suspend_device_cmd(struct pds_vfio_pci_device *pds_vfio, u8 type)
130 {
131 	union pds_core_adminq_cmd cmd = {
132 		.lm_suspend = {
133 			.opcode = PDS_LM_CMD_SUSPEND,
134 			.vf_id = cpu_to_le16(pds_vfio->vf_id),
135 			.type = type,
136 		},
137 	};
138 	struct device *dev = pds_vfio_to_dev(pds_vfio);
139 	union pds_core_adminq_comp comp = {};
140 	int err;
141 
142 	dev_dbg(dev, "vf%u: Suspend device\n", pds_vfio->vf_id);
143 
144 	/*
145 	 * The initial suspend request to the firmware starts the device suspend
146 	 * operation and the firmware returns success if it's started
147 	 * successfully.
148 	 */
149 	err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, &comp, true);
150 	if (err) {
151 		dev_err(dev, "vf%u: Suspend failed: %pe\n", pds_vfio->vf_id,
152 			ERR_PTR(err));
153 		return err;
154 	}
155 
156 	/*
157 	 * The subsequent suspend status request(s) check if the firmware has
158 	 * completed the device suspend process.
159 	 */
160 	return pds_vfio_suspend_wait_device_cmd(pds_vfio, type);
161 }
162 
163 int pds_vfio_resume_device_cmd(struct pds_vfio_pci_device *pds_vfio, u8 type)
164 {
165 	union pds_core_adminq_cmd cmd = {
166 		.lm_resume = {
167 			.opcode = PDS_LM_CMD_RESUME,
168 			.vf_id = cpu_to_le16(pds_vfio->vf_id),
169 			.type = type,
170 		},
171 	};
172 	struct device *dev = pds_vfio_to_dev(pds_vfio);
173 	union pds_core_adminq_comp comp = {};
174 
175 	dev_dbg(dev, "vf%u: Resume device\n", pds_vfio->vf_id);
176 
177 	return pds_vfio_client_adminq_cmd(pds_vfio, &cmd, &comp, true);
178 }
179 
180 int pds_vfio_get_lm_state_size_cmd(struct pds_vfio_pci_device *pds_vfio, u64 *size)
181 {
182 	union pds_core_adminq_cmd cmd = {
183 		.lm_state_size = {
184 			.opcode = PDS_LM_CMD_STATE_SIZE,
185 			.vf_id = cpu_to_le16(pds_vfio->vf_id),
186 		},
187 	};
188 	struct device *dev = pds_vfio_to_dev(pds_vfio);
189 	union pds_core_adminq_comp comp = {};
190 	int err;
191 
192 	dev_dbg(dev, "vf%u: Get migration status\n", pds_vfio->vf_id);
193 
194 	err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, &comp, false);
195 	if (err)
196 		return err;
197 
198 	*size = le64_to_cpu(comp.lm_state_size.size);
199 	return 0;
200 }
201 
202 static int pds_vfio_dma_map_lm_file(struct device *dev,
203 				    enum dma_data_direction dir,
204 				    struct pds_vfio_lm_file *lm_file)
205 {
206 	struct pds_lm_sg_elem *sgl, *sge;
207 	struct scatterlist *sg;
208 	dma_addr_t sgl_addr;
209 	size_t sgl_size;
210 	int err;
211 	int i;
212 
213 	if (!lm_file)
214 		return -EINVAL;
215 
216 	/* dma map file pages */
217 	err = dma_map_sgtable(dev, &lm_file->sg_table, dir, 0);
218 	if (err)
219 		return err;
220 
221 	lm_file->num_sge = lm_file->sg_table.nents;
222 
223 	/* alloc sgl */
224 	sgl_size = lm_file->num_sge * sizeof(struct pds_lm_sg_elem);
225 	sgl = kzalloc(sgl_size, GFP_KERNEL);
226 	if (!sgl) {
227 		err = -ENOMEM;
228 		goto out_unmap_sgtable;
229 	}
230 
231 	/* fill sgl */
232 	sge = sgl;
233 	for_each_sgtable_dma_sg(&lm_file->sg_table, sg, i) {
234 		sge->addr = cpu_to_le64(sg_dma_address(sg));
235 		sge->len = cpu_to_le32(sg_dma_len(sg));
236 		dev_dbg(dev, "addr = %llx, len = %u\n", sge->addr, sge->len);
237 		sge++;
238 	}
239 
240 	sgl_addr = dma_map_single(dev, sgl, sgl_size, DMA_TO_DEVICE);
241 	if (dma_mapping_error(dev, sgl_addr)) {
242 		err = -EIO;
243 		goto out_free_sgl;
244 	}
245 
246 	lm_file->sgl = sgl;
247 	lm_file->sgl_addr = sgl_addr;
248 
249 	return 0;
250 
251 out_free_sgl:
252 	kfree(sgl);
253 out_unmap_sgtable:
254 	lm_file->num_sge = 0;
255 	dma_unmap_sgtable(dev, &lm_file->sg_table, dir, 0);
256 	return err;
257 }
258 
259 static void pds_vfio_dma_unmap_lm_file(struct device *dev,
260 				       enum dma_data_direction dir,
261 				       struct pds_vfio_lm_file *lm_file)
262 {
263 	if (!lm_file)
264 		return;
265 
266 	/* free sgl */
267 	if (lm_file->sgl) {
268 		dma_unmap_single(dev, lm_file->sgl_addr,
269 				 lm_file->num_sge * sizeof(*lm_file->sgl),
270 				 DMA_TO_DEVICE);
271 		kfree(lm_file->sgl);
272 		lm_file->sgl = NULL;
273 		lm_file->sgl_addr = DMA_MAPPING_ERROR;
274 		lm_file->num_sge = 0;
275 	}
276 
277 	/* dma unmap file pages */
278 	dma_unmap_sgtable(dev, &lm_file->sg_table, dir, 0);
279 }
280 
281 int pds_vfio_get_lm_state_cmd(struct pds_vfio_pci_device *pds_vfio)
282 {
283 	union pds_core_adminq_cmd cmd = {
284 		.lm_save = {
285 			.opcode = PDS_LM_CMD_SAVE,
286 			.vf_id = cpu_to_le16(pds_vfio->vf_id),
287 		},
288 	};
289 	struct pci_dev *pdev = pds_vfio_to_pci_dev(pds_vfio);
290 	struct device *pdsc_dev = &pci_physfn(pdev)->dev;
291 	union pds_core_adminq_comp comp = {};
292 	struct pds_vfio_lm_file *lm_file;
293 	int err;
294 
295 	dev_dbg(&pdev->dev, "vf%u: Get migration state\n", pds_vfio->vf_id);
296 
297 	lm_file = pds_vfio->save_file;
298 
299 	err = pds_vfio_dma_map_lm_file(pdsc_dev, DMA_FROM_DEVICE, lm_file);
300 	if (err) {
301 		dev_err(&pdev->dev, "failed to map save migration file: %pe\n",
302 			ERR_PTR(err));
303 		return err;
304 	}
305 
306 	cmd.lm_save.sgl_addr = cpu_to_le64(lm_file->sgl_addr);
307 	cmd.lm_save.num_sge = cpu_to_le32(lm_file->num_sge);
308 
309 	err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, &comp, false);
310 	if (err)
311 		dev_err(&pdev->dev, "failed to get migration state: %pe\n",
312 			ERR_PTR(err));
313 
314 	pds_vfio_dma_unmap_lm_file(pdsc_dev, DMA_FROM_DEVICE, lm_file);
315 
316 	return err;
317 }
318 
319 int pds_vfio_set_lm_state_cmd(struct pds_vfio_pci_device *pds_vfio)
320 {
321 	union pds_core_adminq_cmd cmd = {
322 		.lm_restore = {
323 			.opcode = PDS_LM_CMD_RESTORE,
324 			.vf_id = cpu_to_le16(pds_vfio->vf_id),
325 		},
326 	};
327 	struct pci_dev *pdev = pds_vfio_to_pci_dev(pds_vfio);
328 	struct device *pdsc_dev = &pci_physfn(pdev)->dev;
329 	union pds_core_adminq_comp comp = {};
330 	struct pds_vfio_lm_file *lm_file;
331 	int err;
332 
333 	dev_dbg(&pdev->dev, "vf%u: Set migration state\n", pds_vfio->vf_id);
334 
335 	lm_file = pds_vfio->restore_file;
336 
337 	err = pds_vfio_dma_map_lm_file(pdsc_dev, DMA_TO_DEVICE, lm_file);
338 	if (err) {
339 		dev_err(&pdev->dev,
340 			"failed to map restore migration file: %pe\n",
341 			ERR_PTR(err));
342 		return err;
343 	}
344 
345 	cmd.lm_restore.sgl_addr = cpu_to_le64(lm_file->sgl_addr);
346 	cmd.lm_restore.num_sge = cpu_to_le32(lm_file->num_sge);
347 
348 	err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, &comp, false);
349 	if (err)
350 		dev_err(&pdev->dev, "failed to set migration state: %pe\n",
351 			ERR_PTR(err));
352 
353 	pds_vfio_dma_unmap_lm_file(pdsc_dev, DMA_TO_DEVICE, lm_file);
354 
355 	return err;
356 }
357 
358 void pds_vfio_send_host_vf_lm_status_cmd(struct pds_vfio_pci_device *pds_vfio,
359 					 enum pds_lm_host_vf_status vf_status)
360 {
361 	union pds_core_adminq_cmd cmd = {
362 		.lm_host_vf_status = {
363 			.opcode = PDS_LM_CMD_HOST_VF_STATUS,
364 			.vf_id = cpu_to_le16(pds_vfio->vf_id),
365 			.status = vf_status,
366 		},
367 	};
368 	struct device *dev = pds_vfio_to_dev(pds_vfio);
369 	union pds_core_adminq_comp comp = {};
370 	int err;
371 
372 	dev_dbg(dev, "vf%u: Set host VF LM status: %u", pds_vfio->vf_id,
373 		vf_status);
374 	if (vf_status != PDS_LM_STA_IN_PROGRESS &&
375 	    vf_status != PDS_LM_STA_NONE) {
376 		dev_warn(dev, "Invalid host VF migration status, %d\n",
377 			 vf_status);
378 		return;
379 	}
380 
381 	err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, &comp, false);
382 	if (err)
383 		dev_warn(dev, "failed to send host VF migration status: %pe\n",
384 			 ERR_PTR(err));
385 }
386 
387 int pds_vfio_dirty_status_cmd(struct pds_vfio_pci_device *pds_vfio,
388 			      u64 regions_dma, u8 *max_regions, u8 *num_regions)
389 {
390 	union pds_core_adminq_cmd cmd = {
391 		.lm_dirty_status = {
392 			.opcode = PDS_LM_CMD_DIRTY_STATUS,
393 			.vf_id = cpu_to_le16(pds_vfio->vf_id),
394 		},
395 	};
396 	struct device *dev = pds_vfio_to_dev(pds_vfio);
397 	union pds_core_adminq_comp comp = {};
398 	int err;
399 
400 	dev_dbg(dev, "vf%u: Dirty status\n", pds_vfio->vf_id);
401 
402 	cmd.lm_dirty_status.regions_dma = cpu_to_le64(regions_dma);
403 	cmd.lm_dirty_status.max_regions = *max_regions;
404 
405 	err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, &comp, false);
406 	if (err) {
407 		dev_err(dev, "failed to get dirty status: %pe\n", ERR_PTR(err));
408 		return err;
409 	}
410 
411 	/* only support seq_ack approach for now */
412 	if (!(le32_to_cpu(comp.lm_dirty_status.bmp_type_mask) &
413 	      BIT(PDS_LM_DIRTY_BMP_TYPE_SEQ_ACK))) {
414 		dev_err(dev, "Dirty bitmap tracking SEQ_ACK not supported\n");
415 		return -EOPNOTSUPP;
416 	}
417 
418 	*num_regions = comp.lm_dirty_status.num_regions;
419 	*max_regions = comp.lm_dirty_status.max_regions;
420 
421 	dev_dbg(dev,
422 		"Page Tracking Status command successful, max_regions: %d, num_regions: %d, bmp_type: %s\n",
423 		*max_regions, *num_regions, "PDS_LM_DIRTY_BMP_TYPE_SEQ_ACK");
424 
425 	return 0;
426 }
427 
428 int pds_vfio_dirty_enable_cmd(struct pds_vfio_pci_device *pds_vfio,
429 			      u64 regions_dma, u8 num_regions)
430 {
431 	union pds_core_adminq_cmd cmd = {
432 		.lm_dirty_enable = {
433 			.opcode = PDS_LM_CMD_DIRTY_ENABLE,
434 			.vf_id = cpu_to_le16(pds_vfio->vf_id),
435 			.regions_dma = cpu_to_le64(regions_dma),
436 			.bmp_type = PDS_LM_DIRTY_BMP_TYPE_SEQ_ACK,
437 			.num_regions = num_regions,
438 		},
439 	};
440 	struct device *dev = pds_vfio_to_dev(pds_vfio);
441 	union pds_core_adminq_comp comp = {};
442 	int err;
443 
444 	err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, &comp, false);
445 	if (err) {
446 		dev_err(dev, "failed dirty tracking enable: %pe\n",
447 			ERR_PTR(err));
448 		return err;
449 	}
450 
451 	return 0;
452 }
453 
454 int pds_vfio_dirty_disable_cmd(struct pds_vfio_pci_device *pds_vfio)
455 {
456 	union pds_core_adminq_cmd cmd = {
457 		.lm_dirty_disable = {
458 			.opcode = PDS_LM_CMD_DIRTY_DISABLE,
459 			.vf_id = cpu_to_le16(pds_vfio->vf_id),
460 		},
461 	};
462 	struct device *dev = pds_vfio_to_dev(pds_vfio);
463 	union pds_core_adminq_comp comp = {};
464 	int err;
465 
466 	err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, &comp, false);
467 	if (err || comp.lm_dirty_status.num_regions != 0) {
468 		/* in case num_regions is still non-zero after disable */
469 		err = err ? err : -EIO;
470 		dev_err(dev,
471 			"failed dirty tracking disable: %pe, num_regions %d\n",
472 			ERR_PTR(err), comp.lm_dirty_status.num_regions);
473 		return err;
474 	}
475 
476 	return 0;
477 }
478 
479 int pds_vfio_dirty_seq_ack_cmd(struct pds_vfio_pci_device *pds_vfio,
480 			       u64 sgl_dma, u16 num_sge, u32 offset,
481 			       u32 total_len, bool read_seq)
482 {
483 	const char *cmd_type_str = read_seq ? "read_seq" : "write_ack";
484 	union pds_core_adminq_cmd cmd = {
485 		.lm_dirty_seq_ack = {
486 			.vf_id = cpu_to_le16(pds_vfio->vf_id),
487 			.len_bytes = cpu_to_le32(total_len),
488 			.off_bytes = cpu_to_le32(offset),
489 			.sgl_addr = cpu_to_le64(sgl_dma),
490 			.num_sge = cpu_to_le16(num_sge),
491 		},
492 	};
493 	struct device *dev = pds_vfio_to_dev(pds_vfio);
494 	union pds_core_adminq_comp comp = {};
495 	int err;
496 
497 	if (read_seq)
498 		cmd.lm_dirty_seq_ack.opcode = PDS_LM_CMD_DIRTY_READ_SEQ;
499 	else
500 		cmd.lm_dirty_seq_ack.opcode = PDS_LM_CMD_DIRTY_WRITE_ACK;
501 
502 	err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, &comp, false);
503 	if (err) {
504 		dev_err(dev, "failed cmd Page Tracking %s: %pe\n", cmd_type_str,
505 			ERR_PTR(err));
506 		return err;
507 	}
508 
509 	return 0;
510 }
511