xref: /linux/drivers/net/ethernet/intel/idpf/idpf_main.c (revision 260f6f4fda93c8485c8037865c941b42b9cba5d2)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /* Copyright (C) 2023 Intel Corporation */
3 
4 #include "idpf.h"
5 #include "idpf_devids.h"
6 #include "idpf_virtchnl.h"
7 
8 #define DRV_SUMMARY	"Intel(R) Infrastructure Data Path Function Linux Driver"
9 
10 MODULE_DESCRIPTION(DRV_SUMMARY);
11 MODULE_IMPORT_NS("LIBETH");
12 MODULE_LICENSE("GPL");
13 
14 /**
15  * idpf_remove - Device removal routine
16  * @pdev: PCI device information struct
17  */
18 static void idpf_remove(struct pci_dev *pdev)
19 {
20 	struct idpf_adapter *adapter = pci_get_drvdata(pdev);
21 	int i;
22 
23 	set_bit(IDPF_REMOVE_IN_PROG, adapter->flags);
24 
25 	/* Wait until vc_event_task is done to consider if any hard reset is
26 	 * in progress else we may go ahead and release the resources but the
27 	 * thread doing the hard reset might continue the init path and
28 	 * end up in bad state.
29 	 */
30 	cancel_delayed_work_sync(&adapter->vc_event_task);
31 	if (adapter->num_vfs)
32 		idpf_sriov_configure(pdev, 0);
33 
34 	idpf_vc_core_deinit(adapter);
35 
36 	/* Be a good citizen and leave the device clean on exit */
37 	adapter->dev_ops.reg_ops.trigger_reset(adapter, IDPF_HR_FUNC_RESET);
38 	idpf_deinit_dflt_mbx(adapter);
39 
40 	if (!adapter->netdevs)
41 		goto destroy_wqs;
42 
43 	/* There are some cases where it's possible to still have netdevs
44 	 * registered with the stack at this point, e.g. if the driver detected
45 	 * a HW reset and rmmod is called before it fully recovers. Unregister
46 	 * any stale netdevs here.
47 	 */
48 	for (i = 0; i < adapter->max_vports; i++) {
49 		if (!adapter->netdevs[i])
50 			continue;
51 		if (adapter->netdevs[i]->reg_state != NETREG_UNINITIALIZED)
52 			unregister_netdev(adapter->netdevs[i]);
53 		free_netdev(adapter->netdevs[i]);
54 		adapter->netdevs[i] = NULL;
55 	}
56 
57 destroy_wqs:
58 	destroy_workqueue(adapter->init_wq);
59 	destroy_workqueue(adapter->serv_wq);
60 	destroy_workqueue(adapter->mbx_wq);
61 	destroy_workqueue(adapter->stats_wq);
62 	destroy_workqueue(adapter->vc_event_wq);
63 
64 	for (i = 0; i < adapter->max_vports; i++) {
65 		kfree(adapter->vport_config[i]->user_config.q_coalesce);
66 		kfree(adapter->vport_config[i]);
67 		adapter->vport_config[i] = NULL;
68 	}
69 	kfree(adapter->vport_config);
70 	adapter->vport_config = NULL;
71 	kfree(adapter->netdevs);
72 	adapter->netdevs = NULL;
73 	kfree(adapter->vcxn_mngr);
74 	adapter->vcxn_mngr = NULL;
75 
76 	mutex_destroy(&adapter->vport_ctrl_lock);
77 	mutex_destroy(&adapter->vector_lock);
78 	mutex_destroy(&adapter->queue_lock);
79 	mutex_destroy(&adapter->vc_buf_lock);
80 
81 	pci_set_drvdata(pdev, NULL);
82 	kfree(adapter);
83 }
84 
85 /**
86  * idpf_shutdown - PCI callback for shutting down device
87  * @pdev: PCI device information struct
88  */
89 static void idpf_shutdown(struct pci_dev *pdev)
90 {
91 	struct idpf_adapter *adapter = pci_get_drvdata(pdev);
92 
93 	cancel_delayed_work_sync(&adapter->serv_task);
94 	cancel_delayed_work_sync(&adapter->vc_event_task);
95 	idpf_vc_core_deinit(adapter);
96 	idpf_deinit_dflt_mbx(adapter);
97 
98 	if (system_state == SYSTEM_POWER_OFF)
99 		pci_set_power_state(pdev, PCI_D3hot);
100 }
101 
102 /**
103  * idpf_cfg_hw - Initialize HW struct
104  * @adapter: adapter to setup hw struct for
105  *
106  * Returns 0 on success, negative on failure
107  */
108 static int idpf_cfg_hw(struct idpf_adapter *adapter)
109 {
110 	resource_size_t res_start, mbx_start, rstat_start;
111 	struct pci_dev *pdev = adapter->pdev;
112 	struct idpf_hw *hw = &adapter->hw;
113 	struct device *dev = &pdev->dev;
114 	long len;
115 
116 	res_start = pci_resource_start(pdev, 0);
117 
118 	/* Map mailbox space for virtchnl communication */
119 	mbx_start = res_start + adapter->dev_ops.static_reg_info[0].start;
120 	len = resource_size(&adapter->dev_ops.static_reg_info[0]);
121 	hw->mbx.vaddr = devm_ioremap(dev, mbx_start, len);
122 	if (!hw->mbx.vaddr) {
123 		pci_err(pdev, "failed to allocate BAR0 mbx region\n");
124 
125 		return -ENOMEM;
126 	}
127 	hw->mbx.addr_start = adapter->dev_ops.static_reg_info[0].start;
128 	hw->mbx.addr_len = len;
129 
130 	/* Map rstat space for resets */
131 	rstat_start = res_start + adapter->dev_ops.static_reg_info[1].start;
132 	len = resource_size(&adapter->dev_ops.static_reg_info[1]);
133 	hw->rstat.vaddr = devm_ioremap(dev, rstat_start, len);
134 	if (!hw->rstat.vaddr) {
135 		pci_err(pdev, "failed to allocate BAR0 rstat region\n");
136 
137 		return -ENOMEM;
138 	}
139 	hw->rstat.addr_start = adapter->dev_ops.static_reg_info[1].start;
140 	hw->rstat.addr_len = len;
141 
142 	hw->back = adapter;
143 
144 	return 0;
145 }
146 
147 /**
148  * idpf_probe - Device initialization routine
149  * @pdev: PCI device information struct
150  * @ent: entry in idpf_pci_tbl
151  *
152  * Returns 0 on success, negative on failure
153  */
154 static int idpf_probe(struct pci_dev *pdev, const struct pci_device_id *ent)
155 {
156 	struct device *dev = &pdev->dev;
157 	struct idpf_adapter *adapter;
158 	int err;
159 
160 	adapter = kzalloc(sizeof(*adapter), GFP_KERNEL);
161 	if (!adapter)
162 		return -ENOMEM;
163 
164 	adapter->req_tx_splitq = true;
165 	adapter->req_rx_splitq = true;
166 
167 	switch (ent->device) {
168 	case IDPF_DEV_ID_PF:
169 		idpf_dev_ops_init(adapter);
170 		break;
171 	case IDPF_DEV_ID_VF:
172 		idpf_vf_dev_ops_init(adapter);
173 		adapter->crc_enable = true;
174 		break;
175 	default:
176 		err = -ENODEV;
177 		dev_err(&pdev->dev, "Unexpected dev ID 0x%x in idpf probe\n",
178 			ent->device);
179 		goto err_free;
180 	}
181 
182 	adapter->pdev = pdev;
183 	err = pcim_enable_device(pdev);
184 	if (err)
185 		goto err_free;
186 
187 	err = pcim_request_region(pdev, 0, pci_name(pdev));
188 	if (err) {
189 		pci_err(pdev, "pcim_request_region failed %pe\n", ERR_PTR(err));
190 
191 		goto err_free;
192 	}
193 
194 	err = pci_enable_ptm(pdev, NULL);
195 	if (err)
196 		pci_dbg(pdev, "PCIe PTM is not supported by PCIe bus/controller\n");
197 
198 	/* set up for high or low dma */
199 	err = dma_set_mask_and_coherent(dev, DMA_BIT_MASK(64));
200 	if (err) {
201 		pci_err(pdev, "DMA configuration failed: %pe\n", ERR_PTR(err));
202 
203 		goto err_free;
204 	}
205 
206 	pci_set_master(pdev);
207 	pci_set_drvdata(pdev, adapter);
208 
209 	adapter->init_wq = alloc_workqueue("%s-%s-init",
210 					   WQ_UNBOUND | WQ_MEM_RECLAIM, 0,
211 					   dev_driver_string(dev),
212 					   dev_name(dev));
213 	if (!adapter->init_wq) {
214 		dev_err(dev, "Failed to allocate init workqueue\n");
215 		err = -ENOMEM;
216 		goto err_free;
217 	}
218 
219 	adapter->serv_wq = alloc_workqueue("%s-%s-service",
220 					   WQ_UNBOUND | WQ_MEM_RECLAIM, 0,
221 					   dev_driver_string(dev),
222 					   dev_name(dev));
223 	if (!adapter->serv_wq) {
224 		dev_err(dev, "Failed to allocate service workqueue\n");
225 		err = -ENOMEM;
226 		goto err_serv_wq_alloc;
227 	}
228 
229 	adapter->mbx_wq = alloc_workqueue("%s-%s-mbx", WQ_UNBOUND | WQ_HIGHPRI,
230 					  0, dev_driver_string(dev),
231 					  dev_name(dev));
232 	if (!adapter->mbx_wq) {
233 		dev_err(dev, "Failed to allocate mailbox workqueue\n");
234 		err = -ENOMEM;
235 		goto err_mbx_wq_alloc;
236 	}
237 
238 	adapter->stats_wq = alloc_workqueue("%s-%s-stats",
239 					    WQ_UNBOUND | WQ_MEM_RECLAIM, 0,
240 					    dev_driver_string(dev),
241 					    dev_name(dev));
242 	if (!adapter->stats_wq) {
243 		dev_err(dev, "Failed to allocate workqueue\n");
244 		err = -ENOMEM;
245 		goto err_stats_wq_alloc;
246 	}
247 
248 	adapter->vc_event_wq = alloc_workqueue("%s-%s-vc_event",
249 					       WQ_UNBOUND | WQ_MEM_RECLAIM, 0,
250 					       dev_driver_string(dev),
251 					       dev_name(dev));
252 	if (!adapter->vc_event_wq) {
253 		dev_err(dev, "Failed to allocate virtchnl event workqueue\n");
254 		err = -ENOMEM;
255 		goto err_vc_event_wq_alloc;
256 	}
257 
258 	/* setup msglvl */
259 	adapter->msg_enable = netif_msg_init(-1, IDPF_AVAIL_NETIF_M);
260 
261 	err = idpf_cfg_hw(adapter);
262 	if (err) {
263 		dev_err(dev, "Failed to configure HW structure for adapter: %d\n",
264 			err);
265 		goto err_cfg_hw;
266 	}
267 
268 	mutex_init(&adapter->vport_ctrl_lock);
269 	mutex_init(&adapter->vector_lock);
270 	mutex_init(&adapter->queue_lock);
271 	mutex_init(&adapter->vc_buf_lock);
272 
273 	INIT_DELAYED_WORK(&adapter->init_task, idpf_init_task);
274 	INIT_DELAYED_WORK(&adapter->serv_task, idpf_service_task);
275 	INIT_DELAYED_WORK(&adapter->mbx_task, idpf_mbx_task);
276 	INIT_DELAYED_WORK(&adapter->stats_task, idpf_statistics_task);
277 	INIT_DELAYED_WORK(&adapter->vc_event_task, idpf_vc_event_task);
278 
279 	adapter->dev_ops.reg_ops.reset_reg_init(adapter);
280 	set_bit(IDPF_HR_DRV_LOAD, adapter->flags);
281 	queue_delayed_work(adapter->vc_event_wq, &adapter->vc_event_task,
282 			   msecs_to_jiffies(10 * (pdev->devfn & 0x07)));
283 
284 	return 0;
285 
286 err_cfg_hw:
287 	destroy_workqueue(adapter->vc_event_wq);
288 err_vc_event_wq_alloc:
289 	destroy_workqueue(adapter->stats_wq);
290 err_stats_wq_alloc:
291 	destroy_workqueue(adapter->mbx_wq);
292 err_mbx_wq_alloc:
293 	destroy_workqueue(adapter->serv_wq);
294 err_serv_wq_alloc:
295 	destroy_workqueue(adapter->init_wq);
296 err_free:
297 	kfree(adapter);
298 	return err;
299 }
300 
301 /* idpf_pci_tbl - PCI Dev idpf ID Table
302  */
303 static const struct pci_device_id idpf_pci_tbl[] = {
304 	{ PCI_VDEVICE(INTEL, IDPF_DEV_ID_PF)},
305 	{ PCI_VDEVICE(INTEL, IDPF_DEV_ID_VF)},
306 	{ /* Sentinel */ }
307 };
308 MODULE_DEVICE_TABLE(pci, idpf_pci_tbl);
309 
310 static struct pci_driver idpf_driver = {
311 	.name			= KBUILD_MODNAME,
312 	.id_table		= idpf_pci_tbl,
313 	.probe			= idpf_probe,
314 	.sriov_configure	= idpf_sriov_configure,
315 	.remove			= idpf_remove,
316 	.shutdown		= idpf_shutdown,
317 };
318 module_pci_driver(idpf_driver);
319