1 // SPDX-License-Identifier: GPL-2.0-only 2 /* Copyright (C) 2023 Intel Corporation */ 3 4 #include "idpf.h" 5 #include "idpf_devids.h" 6 #include "idpf_virtchnl.h" 7 8 #define DRV_SUMMARY "Intel(R) Infrastructure Data Path Function Linux Driver" 9 10 MODULE_DESCRIPTION(DRV_SUMMARY); 11 MODULE_IMPORT_NS("LIBETH"); 12 MODULE_LICENSE("GPL"); 13 14 /** 15 * idpf_remove - Device removal routine 16 * @pdev: PCI device information struct 17 */ 18 static void idpf_remove(struct pci_dev *pdev) 19 { 20 struct idpf_adapter *adapter = pci_get_drvdata(pdev); 21 int i; 22 23 set_bit(IDPF_REMOVE_IN_PROG, adapter->flags); 24 25 /* Wait until vc_event_task is done to consider if any hard reset is 26 * in progress else we may go ahead and release the resources but the 27 * thread doing the hard reset might continue the init path and 28 * end up in bad state. 29 */ 30 cancel_delayed_work_sync(&adapter->vc_event_task); 31 if (adapter->num_vfs) 32 idpf_sriov_configure(pdev, 0); 33 34 idpf_vc_core_deinit(adapter); 35 36 /* Be a good citizen and leave the device clean on exit */ 37 adapter->dev_ops.reg_ops.trigger_reset(adapter, IDPF_HR_FUNC_RESET); 38 idpf_deinit_dflt_mbx(adapter); 39 40 if (!adapter->netdevs) 41 goto destroy_wqs; 42 43 /* There are some cases where it's possible to still have netdevs 44 * registered with the stack at this point, e.g. if the driver detected 45 * a HW reset and rmmod is called before it fully recovers. Unregister 46 * any stale netdevs here. 47 */ 48 for (i = 0; i < adapter->max_vports; i++) { 49 if (!adapter->netdevs[i]) 50 continue; 51 if (adapter->netdevs[i]->reg_state != NETREG_UNINITIALIZED) 52 unregister_netdev(adapter->netdevs[i]); 53 free_netdev(adapter->netdevs[i]); 54 adapter->netdevs[i] = NULL; 55 } 56 57 destroy_wqs: 58 destroy_workqueue(adapter->init_wq); 59 destroy_workqueue(adapter->serv_wq); 60 destroy_workqueue(adapter->mbx_wq); 61 destroy_workqueue(adapter->stats_wq); 62 destroy_workqueue(adapter->vc_event_wq); 63 64 for (i = 0; i < adapter->max_vports; i++) { 65 kfree(adapter->vport_config[i]->user_config.q_coalesce); 66 kfree(adapter->vport_config[i]); 67 adapter->vport_config[i] = NULL; 68 } 69 kfree(adapter->vport_config); 70 adapter->vport_config = NULL; 71 kfree(adapter->netdevs); 72 adapter->netdevs = NULL; 73 kfree(adapter->vcxn_mngr); 74 adapter->vcxn_mngr = NULL; 75 76 mutex_destroy(&adapter->vport_ctrl_lock); 77 mutex_destroy(&adapter->vector_lock); 78 mutex_destroy(&adapter->queue_lock); 79 mutex_destroy(&adapter->vc_buf_lock); 80 81 pci_set_drvdata(pdev, NULL); 82 kfree(adapter); 83 } 84 85 /** 86 * idpf_shutdown - PCI callback for shutting down device 87 * @pdev: PCI device information struct 88 */ 89 static void idpf_shutdown(struct pci_dev *pdev) 90 { 91 struct idpf_adapter *adapter = pci_get_drvdata(pdev); 92 93 cancel_delayed_work_sync(&adapter->serv_task); 94 cancel_delayed_work_sync(&adapter->vc_event_task); 95 idpf_vc_core_deinit(adapter); 96 idpf_deinit_dflt_mbx(adapter); 97 98 if (system_state == SYSTEM_POWER_OFF) 99 pci_set_power_state(pdev, PCI_D3hot); 100 } 101 102 /** 103 * idpf_cfg_hw - Initialize HW struct 104 * @adapter: adapter to setup hw struct for 105 * 106 * Returns 0 on success, negative on failure 107 */ 108 static int idpf_cfg_hw(struct idpf_adapter *adapter) 109 { 110 resource_size_t res_start, mbx_start, rstat_start; 111 struct pci_dev *pdev = adapter->pdev; 112 struct idpf_hw *hw = &adapter->hw; 113 struct device *dev = &pdev->dev; 114 long len; 115 116 res_start = pci_resource_start(pdev, 0); 117 118 /* Map mailbox space for virtchnl communication */ 119 mbx_start = res_start + adapter->dev_ops.static_reg_info[0].start; 120 len = resource_size(&adapter->dev_ops.static_reg_info[0]); 121 hw->mbx.vaddr = devm_ioremap(dev, mbx_start, len); 122 if (!hw->mbx.vaddr) { 123 pci_err(pdev, "failed to allocate BAR0 mbx region\n"); 124 125 return -ENOMEM; 126 } 127 hw->mbx.addr_start = adapter->dev_ops.static_reg_info[0].start; 128 hw->mbx.addr_len = len; 129 130 /* Map rstat space for resets */ 131 rstat_start = res_start + adapter->dev_ops.static_reg_info[1].start; 132 len = resource_size(&adapter->dev_ops.static_reg_info[1]); 133 hw->rstat.vaddr = devm_ioremap(dev, rstat_start, len); 134 if (!hw->rstat.vaddr) { 135 pci_err(pdev, "failed to allocate BAR0 rstat region\n"); 136 137 return -ENOMEM; 138 } 139 hw->rstat.addr_start = adapter->dev_ops.static_reg_info[1].start; 140 hw->rstat.addr_len = len; 141 142 hw->back = adapter; 143 144 return 0; 145 } 146 147 /** 148 * idpf_probe - Device initialization routine 149 * @pdev: PCI device information struct 150 * @ent: entry in idpf_pci_tbl 151 * 152 * Returns 0 on success, negative on failure 153 */ 154 static int idpf_probe(struct pci_dev *pdev, const struct pci_device_id *ent) 155 { 156 struct device *dev = &pdev->dev; 157 struct idpf_adapter *adapter; 158 int err; 159 160 adapter = kzalloc(sizeof(*adapter), GFP_KERNEL); 161 if (!adapter) 162 return -ENOMEM; 163 164 adapter->req_tx_splitq = true; 165 adapter->req_rx_splitq = true; 166 167 switch (ent->device) { 168 case IDPF_DEV_ID_PF: 169 idpf_dev_ops_init(adapter); 170 break; 171 case IDPF_DEV_ID_VF: 172 idpf_vf_dev_ops_init(adapter); 173 adapter->crc_enable = true; 174 break; 175 default: 176 err = -ENODEV; 177 dev_err(&pdev->dev, "Unexpected dev ID 0x%x in idpf probe\n", 178 ent->device); 179 goto err_free; 180 } 181 182 adapter->pdev = pdev; 183 err = pcim_enable_device(pdev); 184 if (err) 185 goto err_free; 186 187 err = pcim_request_region(pdev, 0, pci_name(pdev)); 188 if (err) { 189 pci_err(pdev, "pcim_request_region failed %pe\n", ERR_PTR(err)); 190 191 goto err_free; 192 } 193 194 err = pci_enable_ptm(pdev, NULL); 195 if (err) 196 pci_dbg(pdev, "PCIe PTM is not supported by PCIe bus/controller\n"); 197 198 /* set up for high or low dma */ 199 err = dma_set_mask_and_coherent(dev, DMA_BIT_MASK(64)); 200 if (err) { 201 pci_err(pdev, "DMA configuration failed: %pe\n", ERR_PTR(err)); 202 203 goto err_free; 204 } 205 206 pci_set_master(pdev); 207 pci_set_drvdata(pdev, adapter); 208 209 adapter->init_wq = alloc_workqueue("%s-%s-init", 210 WQ_UNBOUND | WQ_MEM_RECLAIM, 0, 211 dev_driver_string(dev), 212 dev_name(dev)); 213 if (!adapter->init_wq) { 214 dev_err(dev, "Failed to allocate init workqueue\n"); 215 err = -ENOMEM; 216 goto err_free; 217 } 218 219 adapter->serv_wq = alloc_workqueue("%s-%s-service", 220 WQ_UNBOUND | WQ_MEM_RECLAIM, 0, 221 dev_driver_string(dev), 222 dev_name(dev)); 223 if (!adapter->serv_wq) { 224 dev_err(dev, "Failed to allocate service workqueue\n"); 225 err = -ENOMEM; 226 goto err_serv_wq_alloc; 227 } 228 229 adapter->mbx_wq = alloc_workqueue("%s-%s-mbx", WQ_UNBOUND | WQ_HIGHPRI, 230 0, dev_driver_string(dev), 231 dev_name(dev)); 232 if (!adapter->mbx_wq) { 233 dev_err(dev, "Failed to allocate mailbox workqueue\n"); 234 err = -ENOMEM; 235 goto err_mbx_wq_alloc; 236 } 237 238 adapter->stats_wq = alloc_workqueue("%s-%s-stats", 239 WQ_UNBOUND | WQ_MEM_RECLAIM, 0, 240 dev_driver_string(dev), 241 dev_name(dev)); 242 if (!adapter->stats_wq) { 243 dev_err(dev, "Failed to allocate workqueue\n"); 244 err = -ENOMEM; 245 goto err_stats_wq_alloc; 246 } 247 248 adapter->vc_event_wq = alloc_workqueue("%s-%s-vc_event", 249 WQ_UNBOUND | WQ_MEM_RECLAIM, 0, 250 dev_driver_string(dev), 251 dev_name(dev)); 252 if (!adapter->vc_event_wq) { 253 dev_err(dev, "Failed to allocate virtchnl event workqueue\n"); 254 err = -ENOMEM; 255 goto err_vc_event_wq_alloc; 256 } 257 258 /* setup msglvl */ 259 adapter->msg_enable = netif_msg_init(-1, IDPF_AVAIL_NETIF_M); 260 261 err = idpf_cfg_hw(adapter); 262 if (err) { 263 dev_err(dev, "Failed to configure HW structure for adapter: %d\n", 264 err); 265 goto err_cfg_hw; 266 } 267 268 mutex_init(&adapter->vport_ctrl_lock); 269 mutex_init(&adapter->vector_lock); 270 mutex_init(&adapter->queue_lock); 271 mutex_init(&adapter->vc_buf_lock); 272 273 INIT_DELAYED_WORK(&adapter->init_task, idpf_init_task); 274 INIT_DELAYED_WORK(&adapter->serv_task, idpf_service_task); 275 INIT_DELAYED_WORK(&adapter->mbx_task, idpf_mbx_task); 276 INIT_DELAYED_WORK(&adapter->stats_task, idpf_statistics_task); 277 INIT_DELAYED_WORK(&adapter->vc_event_task, idpf_vc_event_task); 278 279 adapter->dev_ops.reg_ops.reset_reg_init(adapter); 280 set_bit(IDPF_HR_DRV_LOAD, adapter->flags); 281 queue_delayed_work(adapter->vc_event_wq, &adapter->vc_event_task, 282 msecs_to_jiffies(10 * (pdev->devfn & 0x07))); 283 284 return 0; 285 286 err_cfg_hw: 287 destroy_workqueue(adapter->vc_event_wq); 288 err_vc_event_wq_alloc: 289 destroy_workqueue(adapter->stats_wq); 290 err_stats_wq_alloc: 291 destroy_workqueue(adapter->mbx_wq); 292 err_mbx_wq_alloc: 293 destroy_workqueue(adapter->serv_wq); 294 err_serv_wq_alloc: 295 destroy_workqueue(adapter->init_wq); 296 err_free: 297 kfree(adapter); 298 return err; 299 } 300 301 /* idpf_pci_tbl - PCI Dev idpf ID Table 302 */ 303 static const struct pci_device_id idpf_pci_tbl[] = { 304 { PCI_VDEVICE(INTEL, IDPF_DEV_ID_PF)}, 305 { PCI_VDEVICE(INTEL, IDPF_DEV_ID_VF)}, 306 { /* Sentinel */ } 307 }; 308 MODULE_DEVICE_TABLE(pci, idpf_pci_tbl); 309 310 static struct pci_driver idpf_driver = { 311 .name = KBUILD_MODNAME, 312 .id_table = idpf_pci_tbl, 313 .probe = idpf_probe, 314 .sriov_configure = idpf_sriov_configure, 315 .remove = idpf_remove, 316 .shutdown = idpf_shutdown, 317 }; 318 module_pci_driver(idpf_driver); 319