1 /****************************************************************************** 2 3 Copyright (c) 2013-2018, Intel Corporation 4 All rights reserved. 5 6 Redistribution and use in source and binary forms, with or without 7 modification, are permitted provided that the following conditions are met: 8 9 1. Redistributions of source code must retain the above copyright notice, 10 this list of conditions and the following disclaimer. 11 12 2. Redistributions in binary form must reproduce the above copyright 13 notice, this list of conditions and the following disclaimer in the 14 documentation and/or other materials provided with the distribution. 15 16 3. Neither the name of the Intel Corporation nor the names of its 17 contributors may be used to endorse or promote products derived from 18 this software without specific prior written permission. 19 20 THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" 21 AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 22 IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 23 ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE 24 LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 25 CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 26 SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 27 INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 28 CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 29 ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE 30 POSSIBILITY OF SUCH DAMAGE. 31 32 ******************************************************************************/ 33 /*$FreeBSD$*/ 34 35 #include "ixl_pf_iov.h" 36 37 /* Private functions */ 38 static void ixl_vf_map_vsi_queue(struct i40e_hw *hw, struct ixl_vf *vf, int qnum, uint32_t val); 39 static void ixl_vf_disable_queue_intr(struct i40e_hw *hw, uint32_t vfint_reg); 40 static void ixl_vf_unregister_intr(struct i40e_hw *hw, uint32_t vpint_reg); 41 42 static int ixl_vc_opcode_level(uint16_t opcode); 43 44 static int ixl_vf_mac_valid(struct ixl_vf *vf, const uint8_t *addr); 45 46 static int ixl_vf_alloc_vsi(struct ixl_pf *pf, struct ixl_vf *vf); 47 static int ixl_vf_setup_vsi(struct ixl_pf *pf, struct ixl_vf *vf); 48 static void ixl_vf_map_queues(struct ixl_pf *pf, struct ixl_vf *vf); 49 static void ixl_vf_vsi_release(struct ixl_pf *pf, struct ixl_vsi *vsi); 50 static void ixl_vf_release_resources(struct ixl_pf *pf, struct ixl_vf *vf); 51 static int ixl_flush_pcie(struct ixl_pf *pf, struct ixl_vf *vf); 52 static void ixl_reset_vf(struct ixl_pf *pf, struct ixl_vf *vf); 53 static void ixl_reinit_vf(struct ixl_pf *pf, struct ixl_vf *vf); 54 static void ixl_send_vf_msg(struct ixl_pf *pf, struct ixl_vf *vf, uint16_t op, enum i40e_status_code status, void *msg, uint16_t len); 55 static void ixl_send_vf_ack(struct ixl_pf *pf, struct ixl_vf *vf, uint16_t op); 56 static void ixl_send_vf_nack_msg(struct ixl_pf *pf, struct ixl_vf *vf, uint16_t op, enum i40e_status_code status, const char *file, int line); 57 static void ixl_vf_version_msg(struct ixl_pf *pf, struct ixl_vf *vf, void *msg, uint16_t msg_size); 58 static void ixl_vf_reset_msg(struct ixl_pf *pf, struct ixl_vf *vf, void *msg, uint16_t msg_size); 59 static void ixl_vf_get_resources_msg(struct ixl_pf *pf, struct ixl_vf *vf, void *msg, uint16_t msg_size); 60 static int ixl_vf_config_tx_queue(struct ixl_pf *pf, struct ixl_vf *vf, struct virtchnl_txq_info *info); 61 static int ixl_vf_config_rx_queue(struct ixl_pf *pf, struct ixl_vf *vf, struct virtchnl_rxq_info *info); 62 static void ixl_vf_config_vsi_msg(struct ixl_pf *pf, struct ixl_vf *vf, void *msg, uint16_t msg_size); 63 static void ixl_vf_set_qctl(struct ixl_pf *pf, const struct virtchnl_vector_map *vector, enum i40e_queue_type cur_type, uint16_t cur_queue, 64 enum i40e_queue_type *last_type, uint16_t *last_queue); 65 static void ixl_vf_config_vector(struct ixl_pf *pf, struct ixl_vf *vf, const struct virtchnl_vector_map *vector); 66 static void ixl_vf_config_irq_msg(struct ixl_pf *pf, struct ixl_vf *vf, void *msg, uint16_t msg_size); 67 static void ixl_vf_enable_queues_msg(struct ixl_pf *pf, struct ixl_vf *vf, void *msg, uint16_t msg_size); 68 static void ixl_vf_disable_queues_msg(struct ixl_pf *pf, struct ixl_vf *vf, void *msg, uint16_t msg_size); 69 static void ixl_vf_add_mac_msg(struct ixl_pf *pf, struct ixl_vf *vf, void *msg, uint16_t msg_size); 70 static void ixl_vf_del_mac_msg(struct ixl_pf *pf, struct ixl_vf *vf, void *msg, uint16_t msg_size); 71 static enum i40e_status_code ixl_vf_enable_vlan_strip(struct ixl_pf *pf, struct ixl_vf *vf); 72 static void ixl_vf_add_vlan_msg(struct ixl_pf *pf, struct ixl_vf *vf, void *msg, uint16_t msg_size); 73 static void ixl_vf_del_vlan_msg(struct ixl_pf *pf, struct ixl_vf *vf, void *msg, uint16_t msg_size); 74 static void ixl_vf_config_promisc_msg(struct ixl_pf *pf, struct ixl_vf *vf, void *msg, uint16_t msg_size); 75 static void ixl_vf_get_stats_msg(struct ixl_pf *pf, struct ixl_vf *vf, void *msg, uint16_t msg_size); 76 static int ixl_vf_reserve_queues(struct ixl_pf *pf, struct ixl_vf *vf, int num_queues); 77 static int ixl_config_pf_vsi_loopback(struct ixl_pf *pf, bool enable); 78 79 static int ixl_adminq_err_to_errno(enum i40e_admin_queue_err err); 80 81 /* 82 * TODO: Move pieces of this into iflib and call the rest in a handler? 83 * 84 * e.g. ixl_if_iov_set_schema 85 * 86 * It's odd to do pci_iov_detach() there while doing pci_iov_attach() 87 * in the driver. 88 */ 89 void 90 ixl_initialize_sriov(struct ixl_pf *pf) 91 { 92 device_t dev = pf->dev; 93 struct i40e_hw *hw = &pf->hw; 94 nvlist_t *pf_schema, *vf_schema; 95 int iov_error; 96 97 pf_schema = pci_iov_schema_alloc_node(); 98 vf_schema = pci_iov_schema_alloc_node(); 99 pci_iov_schema_add_unicast_mac(vf_schema, "mac-addr", 0, NULL); 100 pci_iov_schema_add_bool(vf_schema, "mac-anti-spoof", 101 IOV_SCHEMA_HASDEFAULT, TRUE); 102 pci_iov_schema_add_bool(vf_schema, "allow-set-mac", 103 IOV_SCHEMA_HASDEFAULT, FALSE); 104 pci_iov_schema_add_bool(vf_schema, "allow-promisc", 105 IOV_SCHEMA_HASDEFAULT, FALSE); 106 pci_iov_schema_add_uint16(vf_schema, "num-queues", 107 IOV_SCHEMA_HASDEFAULT, 108 max(1, min(hw->func_caps.num_msix_vectors_vf - 1, IAVF_MAX_QUEUES))); 109 110 iov_error = pci_iov_attach(dev, pf_schema, vf_schema); 111 if (iov_error != 0) { 112 device_printf(dev, 113 "Failed to initialize SR-IOV (error=%d)\n", 114 iov_error); 115 } else 116 device_printf(dev, "SR-IOV ready\n"); 117 118 pf->vc_debug_lvl = 1; 119 } 120 121 /* 122 * Allocate the VSI for a VF. 123 */ 124 static int 125 ixl_vf_alloc_vsi(struct ixl_pf *pf, struct ixl_vf *vf) 126 { 127 device_t dev; 128 struct i40e_hw *hw; 129 struct ixl_vsi *vsi; 130 struct i40e_vsi_context vsi_ctx; 131 int i; 132 enum i40e_status_code code; 133 134 hw = &pf->hw; 135 vsi = &pf->vsi; 136 dev = pf->dev; 137 138 vsi_ctx.pf_num = hw->pf_id; 139 vsi_ctx.uplink_seid = pf->veb_seid; 140 vsi_ctx.connection_type = IXL_VSI_DATA_PORT; 141 vsi_ctx.vf_num = hw->func_caps.vf_base_id + vf->vf_num; 142 vsi_ctx.flags = I40E_AQ_VSI_TYPE_VF; 143 144 bzero(&vsi_ctx.info, sizeof(vsi_ctx.info)); 145 146 vsi_ctx.info.valid_sections = htole16(I40E_AQ_VSI_PROP_SWITCH_VALID); 147 if (pf->enable_vf_loopback) 148 vsi_ctx.info.switch_id = 149 htole16(I40E_AQ_VSI_SW_ID_FLAG_ALLOW_LB); 150 151 vsi_ctx.info.valid_sections |= htole16(I40E_AQ_VSI_PROP_SECURITY_VALID); 152 vsi_ctx.info.sec_flags = 0; 153 if (vf->vf_flags & VF_FLAG_MAC_ANTI_SPOOF) 154 vsi_ctx.info.sec_flags |= I40E_AQ_VSI_SEC_FLAG_ENABLE_MAC_CHK; 155 156 vsi_ctx.info.valid_sections |= htole16(I40E_AQ_VSI_PROP_VLAN_VALID); 157 vsi_ctx.info.port_vlan_flags = I40E_AQ_VSI_PVLAN_MODE_ALL | 158 I40E_AQ_VSI_PVLAN_EMOD_NOTHING; 159 160 vsi_ctx.info.valid_sections |= 161 htole16(I40E_AQ_VSI_PROP_QUEUE_MAP_VALID); 162 vsi_ctx.info.mapping_flags = htole16(I40E_AQ_VSI_QUE_MAP_NONCONTIG); 163 164 /* XXX: Only scattered allocation is supported for VFs right now */ 165 for (i = 0; i < vf->qtag.num_active; i++) 166 vsi_ctx.info.queue_mapping[i] = vf->qtag.qidx[i]; 167 for (; i < nitems(vsi_ctx.info.queue_mapping); i++) 168 vsi_ctx.info.queue_mapping[i] = htole16(I40E_AQ_VSI_QUEUE_MASK); 169 170 vsi_ctx.info.tc_mapping[0] = htole16( 171 (0 << I40E_AQ_VSI_TC_QUE_OFFSET_SHIFT) | 172 ((fls(vf->qtag.num_allocated) - 1) << I40E_AQ_VSI_TC_QUE_NUMBER_SHIFT)); 173 174 code = i40e_aq_add_vsi(hw, &vsi_ctx, NULL); 175 if (code != I40E_SUCCESS) 176 return (ixl_adminq_err_to_errno(hw->aq.asq_last_status)); 177 vf->vsi.seid = vsi_ctx.seid; 178 vf->vsi.vsi_num = vsi_ctx.vsi_number; 179 vf->vsi.num_rx_queues = vf->qtag.num_active; 180 vf->vsi.num_tx_queues = vf->qtag.num_active; 181 182 code = i40e_aq_get_vsi_params(hw, &vsi_ctx, NULL); 183 if (code != I40E_SUCCESS) 184 return (ixl_adminq_err_to_errno(hw->aq.asq_last_status)); 185 186 code = i40e_aq_config_vsi_bw_limit(hw, vf->vsi.seid, 0, 0, NULL); 187 if (code != I40E_SUCCESS) { 188 device_printf(dev, "Failed to disable BW limit: %d\n", 189 ixl_adminq_err_to_errno(hw->aq.asq_last_status)); 190 return (ixl_adminq_err_to_errno(hw->aq.asq_last_status)); 191 } 192 193 memcpy(&vf->vsi.info, &vsi_ctx.info, sizeof(vf->vsi.info)); 194 return (0); 195 } 196 197 static int 198 ixl_vf_setup_vsi(struct ixl_pf *pf, struct ixl_vf *vf) 199 { 200 struct i40e_hw *hw; 201 int error; 202 203 hw = &pf->hw; 204 vf->vsi.flags |= IXL_FLAGS_IS_VF; 205 206 error = ixl_vf_alloc_vsi(pf, vf); 207 if (error != 0) 208 return (error); 209 210 vf->vsi.dev = pf->dev; 211 212 ixl_init_filters(&vf->vsi); 213 /* Let VF receive broadcast Ethernet frames */ 214 error = i40e_aq_set_vsi_broadcast(hw, vf->vsi.seid, TRUE, NULL); 215 if (error) 216 device_printf(pf->dev, "Error configuring VF VSI for broadcast promiscuous\n"); 217 /* Re-add VF's MAC/VLAN filters to its VSI */ 218 ixl_reconfigure_filters(&vf->vsi); 219 220 return (0); 221 } 222 223 static void 224 ixl_vf_map_vsi_queue(struct i40e_hw *hw, struct ixl_vf *vf, int qnum, 225 uint32_t val) 226 { 227 uint32_t qtable; 228 int index, shift; 229 230 /* 231 * Two queues are mapped in a single register, so we have to do some 232 * gymnastics to convert the queue number into a register index and 233 * shift. 234 */ 235 index = qnum / 2; 236 shift = (qnum % 2) * I40E_VSILAN_QTABLE_QINDEX_1_SHIFT; 237 238 qtable = i40e_read_rx_ctl(hw, I40E_VSILAN_QTABLE(index, vf->vsi.vsi_num)); 239 qtable &= ~(I40E_VSILAN_QTABLE_QINDEX_0_MASK << shift); 240 qtable |= val << shift; 241 i40e_write_rx_ctl(hw, I40E_VSILAN_QTABLE(index, vf->vsi.vsi_num), qtable); 242 } 243 244 static void 245 ixl_vf_map_queues(struct ixl_pf *pf, struct ixl_vf *vf) 246 { 247 struct i40e_hw *hw; 248 uint32_t qtable; 249 int i; 250 251 hw = &pf->hw; 252 253 /* 254 * Contiguous mappings aren't actually supported by the hardware, 255 * so we have to use non-contiguous mappings. 256 */ 257 i40e_write_rx_ctl(hw, I40E_VSILAN_QBASE(vf->vsi.vsi_num), 258 I40E_VSILAN_QBASE_VSIQTABLE_ENA_MASK); 259 260 /* Enable LAN traffic on this VF */ 261 wr32(hw, I40E_VPLAN_MAPENA(vf->vf_num), 262 I40E_VPLAN_MAPENA_TXRX_ENA_MASK); 263 264 /* Program index of each VF queue into PF queue space 265 * (This is only needed if QTABLE is enabled) */ 266 for (i = 0; i < vf->vsi.num_tx_queues; i++) { 267 qtable = ixl_pf_qidx_from_vsi_qidx(&vf->qtag, i) << 268 I40E_VPLAN_QTABLE_QINDEX_SHIFT; 269 270 wr32(hw, I40E_VPLAN_QTABLE(i, vf->vf_num), qtable); 271 } 272 for (; i < IXL_MAX_VSI_QUEUES; i++) 273 wr32(hw, I40E_VPLAN_QTABLE(i, vf->vf_num), 274 I40E_VPLAN_QTABLE_QINDEX_MASK); 275 276 /* Map queues allocated to VF to its VSI; 277 * This mapping matches the VF-wide mapping since the VF 278 * is only given a single VSI */ 279 for (i = 0; i < vf->vsi.num_tx_queues; i++) 280 ixl_vf_map_vsi_queue(hw, vf, i, 281 ixl_pf_qidx_from_vsi_qidx(&vf->qtag, i)); 282 283 /* Set rest of VSI queues as unused. */ 284 for (; i < IXL_MAX_VSI_QUEUES; i++) 285 ixl_vf_map_vsi_queue(hw, vf, i, 286 I40E_VSILAN_QTABLE_QINDEX_0_MASK); 287 288 ixl_flush(hw); 289 } 290 291 static void 292 ixl_vf_vsi_release(struct ixl_pf *pf, struct ixl_vsi *vsi) 293 { 294 struct i40e_hw *hw; 295 296 hw = &pf->hw; 297 298 if (vsi->seid == 0) 299 return; 300 301 i40e_aq_delete_element(hw, vsi->seid, NULL); 302 } 303 304 static void 305 ixl_vf_disable_queue_intr(struct i40e_hw *hw, uint32_t vfint_reg) 306 { 307 308 wr32(hw, vfint_reg, I40E_VFINT_DYN_CTLN_CLEARPBA_MASK); 309 ixl_flush(hw); 310 } 311 312 static void 313 ixl_vf_unregister_intr(struct i40e_hw *hw, uint32_t vpint_reg) 314 { 315 316 wr32(hw, vpint_reg, I40E_VPINT_LNKLSTN_FIRSTQ_TYPE_MASK | 317 I40E_VPINT_LNKLSTN_FIRSTQ_INDX_MASK); 318 ixl_flush(hw); 319 } 320 321 static void 322 ixl_vf_release_resources(struct ixl_pf *pf, struct ixl_vf *vf) 323 { 324 struct i40e_hw *hw; 325 uint32_t vfint_reg, vpint_reg; 326 int i; 327 328 hw = &pf->hw; 329 330 ixl_vf_vsi_release(pf, &vf->vsi); 331 332 /* Index 0 has a special register. */ 333 ixl_vf_disable_queue_intr(hw, I40E_VFINT_DYN_CTL0(vf->vf_num)); 334 335 for (i = 1; i < hw->func_caps.num_msix_vectors_vf; i++) { 336 vfint_reg = IXL_VFINT_DYN_CTLN_REG(hw, i , vf->vf_num); 337 ixl_vf_disable_queue_intr(hw, vfint_reg); 338 } 339 340 /* Index 0 has a special register. */ 341 ixl_vf_unregister_intr(hw, I40E_VPINT_LNKLST0(vf->vf_num)); 342 343 for (i = 1; i < hw->func_caps.num_msix_vectors_vf; i++) { 344 vpint_reg = IXL_VPINT_LNKLSTN_REG(hw, i, vf->vf_num); 345 ixl_vf_unregister_intr(hw, vpint_reg); 346 } 347 348 vf->vsi.num_tx_queues = 0; 349 vf->vsi.num_rx_queues = 0; 350 } 351 352 static int 353 ixl_flush_pcie(struct ixl_pf *pf, struct ixl_vf *vf) 354 { 355 struct i40e_hw *hw; 356 int i; 357 uint16_t global_vf_num; 358 uint32_t ciad; 359 360 hw = &pf->hw; 361 global_vf_num = hw->func_caps.vf_base_id + vf->vf_num; 362 363 wr32(hw, I40E_PF_PCI_CIAA, IXL_PF_PCI_CIAA_VF_DEVICE_STATUS | 364 (global_vf_num << I40E_PF_PCI_CIAA_VF_NUM_SHIFT)); 365 for (i = 0; i < IXL_VF_RESET_TIMEOUT; i++) { 366 ciad = rd32(hw, I40E_PF_PCI_CIAD); 367 if ((ciad & IXL_PF_PCI_CIAD_VF_TRANS_PENDING_MASK) == 0) 368 return (0); 369 DELAY(1); 370 } 371 372 return (ETIMEDOUT); 373 } 374 375 static void 376 ixl_reset_vf(struct ixl_pf *pf, struct ixl_vf *vf) 377 { 378 struct i40e_hw *hw; 379 uint32_t vfrtrig; 380 381 hw = &pf->hw; 382 383 ixl_dbg_iov(pf, "Resetting VF-%d\n", vf->vf_num); 384 385 vfrtrig = rd32(hw, I40E_VPGEN_VFRTRIG(vf->vf_num)); 386 vfrtrig |= I40E_VPGEN_VFRTRIG_VFSWR_MASK; 387 wr32(hw, I40E_VPGEN_VFRTRIG(vf->vf_num), vfrtrig); 388 ixl_flush(hw); 389 390 ixl_reinit_vf(pf, vf); 391 392 ixl_dbg_iov(pf, "Resetting VF-%d done.\n", vf->vf_num); 393 } 394 395 static void 396 ixl_reinit_vf(struct ixl_pf *pf, struct ixl_vf *vf) 397 { 398 struct i40e_hw *hw; 399 uint32_t vfrstat, vfrtrig; 400 int i, error; 401 402 hw = &pf->hw; 403 404 error = ixl_flush_pcie(pf, vf); 405 if (error != 0) 406 device_printf(pf->dev, 407 "Timed out waiting for PCIe activity to stop on VF-%d\n", 408 vf->vf_num); 409 410 for (i = 0; i < IXL_VF_RESET_TIMEOUT; i++) { 411 DELAY(10); 412 413 vfrstat = rd32(hw, I40E_VPGEN_VFRSTAT(vf->vf_num)); 414 if (vfrstat & I40E_VPGEN_VFRSTAT_VFRD_MASK) 415 break; 416 } 417 418 if (i == IXL_VF_RESET_TIMEOUT) 419 device_printf(pf->dev, "VF %d failed to reset\n", vf->vf_num); 420 421 wr32(hw, I40E_VFGEN_RSTAT1(vf->vf_num), VIRTCHNL_VFR_COMPLETED); 422 423 vfrtrig = rd32(hw, I40E_VPGEN_VFRTRIG(vf->vf_num)); 424 vfrtrig &= ~I40E_VPGEN_VFRTRIG_VFSWR_MASK; 425 wr32(hw, I40E_VPGEN_VFRTRIG(vf->vf_num), vfrtrig); 426 427 if (vf->vsi.seid != 0) 428 ixl_disable_rings(pf, &vf->vsi, &vf->qtag); 429 ixl_pf_qmgr_clear_queue_flags(&vf->qtag); 430 431 ixl_vf_release_resources(pf, vf); 432 ixl_vf_setup_vsi(pf, vf); 433 ixl_vf_map_queues(pf, vf); 434 435 wr32(hw, I40E_VFGEN_RSTAT1(vf->vf_num), VIRTCHNL_VFR_VFACTIVE); 436 ixl_flush(hw); 437 } 438 439 static int 440 ixl_vc_opcode_level(uint16_t opcode) 441 { 442 switch (opcode) { 443 case VIRTCHNL_OP_GET_STATS: 444 return (10); 445 default: 446 return (5); 447 } 448 } 449 450 static void 451 ixl_send_vf_msg(struct ixl_pf *pf, struct ixl_vf *vf, uint16_t op, 452 enum i40e_status_code status, void *msg, uint16_t len) 453 { 454 struct i40e_hw *hw; 455 int global_vf_id; 456 457 hw = &pf->hw; 458 global_vf_id = hw->func_caps.vf_base_id + vf->vf_num; 459 460 I40E_VC_DEBUG(pf, ixl_vc_opcode_level(op), 461 "Sending msg (op=%s[%d], status=%d) to VF-%d\n", 462 ixl_vc_opcode_str(op), op, status, vf->vf_num); 463 464 i40e_aq_send_msg_to_vf(hw, global_vf_id, op, status, msg, len, NULL); 465 } 466 467 static void 468 ixl_send_vf_ack(struct ixl_pf *pf, struct ixl_vf *vf, uint16_t op) 469 { 470 471 ixl_send_vf_msg(pf, vf, op, I40E_SUCCESS, NULL, 0); 472 } 473 474 static void 475 ixl_send_vf_nack_msg(struct ixl_pf *pf, struct ixl_vf *vf, uint16_t op, 476 enum i40e_status_code status, const char *file, int line) 477 { 478 479 I40E_VC_DEBUG(pf, 1, 480 "Sending NACK (op=%s[%d], err=%s[%d]) to VF-%d from %s:%d\n", 481 ixl_vc_opcode_str(op), op, i40e_stat_str(&pf->hw, status), 482 status, vf->vf_num, file, line); 483 ixl_send_vf_msg(pf, vf, op, status, NULL, 0); 484 } 485 486 static void 487 ixl_vf_version_msg(struct ixl_pf *pf, struct ixl_vf *vf, void *msg, 488 uint16_t msg_size) 489 { 490 struct virtchnl_version_info *recv_vf_version; 491 device_t dev = pf->dev; 492 493 recv_vf_version = (struct virtchnl_version_info *)msg; 494 495 /* VFs running the 1.0 API expect to get 1.0 back */ 496 if (VF_IS_V10(recv_vf_version)) { 497 vf->version.major = 1; 498 vf->version.minor = VIRTCHNL_VERSION_MINOR_NO_VF_CAPS; 499 } else { 500 vf->version.major = VIRTCHNL_VERSION_MAJOR; 501 vf->version.minor = VIRTCHNL_VERSION_MINOR; 502 503 if ((recv_vf_version->major != VIRTCHNL_VERSION_MAJOR) || 504 (recv_vf_version->minor != VIRTCHNL_VERSION_MINOR)) 505 device_printf(dev, 506 "%s: VF-%d requested version (%d.%d) differs from PF version (%d.%d)\n", 507 __func__, vf->vf_num, 508 recv_vf_version->major, recv_vf_version->minor, 509 VIRTCHNL_VERSION_MAJOR, VIRTCHNL_VERSION_MINOR); 510 } 511 512 ixl_send_vf_msg(pf, vf, VIRTCHNL_OP_VERSION, I40E_SUCCESS, 513 &vf->version, sizeof(vf->version)); 514 } 515 516 static void 517 ixl_vf_reset_msg(struct ixl_pf *pf, struct ixl_vf *vf, void *msg, 518 uint16_t msg_size) 519 { 520 ixl_reset_vf(pf, vf); 521 522 /* No response to a reset message. */ 523 } 524 525 static void 526 ixl_vf_get_resources_msg(struct ixl_pf *pf, struct ixl_vf *vf, void *msg, 527 uint16_t msg_size) 528 { 529 struct virtchnl_vf_resource reply; 530 531 bzero(&reply, sizeof(reply)); 532 533 if (vf->version.minor == VIRTCHNL_VERSION_MINOR_NO_VF_CAPS) 534 reply.vf_cap_flags = VIRTCHNL_VF_OFFLOAD_L2 | 535 VIRTCHNL_VF_OFFLOAD_RSS_REG | 536 VIRTCHNL_VF_OFFLOAD_VLAN; 537 else 538 /* Force VF RSS setup by PF in 1.1+ VFs */ 539 reply.vf_cap_flags = *(u32 *)msg & ( 540 VIRTCHNL_VF_OFFLOAD_L2 | 541 VIRTCHNL_VF_OFFLOAD_RSS_PF | 542 VIRTCHNL_VF_OFFLOAD_VLAN); 543 544 reply.num_vsis = 1; 545 reply.num_queue_pairs = vf->vsi.num_tx_queues; 546 reply.max_vectors = pf->hw.func_caps.num_msix_vectors_vf; 547 reply.rss_key_size = 52; 548 reply.rss_lut_size = 64; 549 reply.vsi_res[0].vsi_id = vf->vsi.vsi_num; 550 reply.vsi_res[0].vsi_type = VIRTCHNL_VSI_SRIOV; 551 reply.vsi_res[0].num_queue_pairs = vf->vsi.num_tx_queues; 552 memcpy(reply.vsi_res[0].default_mac_addr, vf->mac, ETHER_ADDR_LEN); 553 554 ixl_send_vf_msg(pf, vf, VIRTCHNL_OP_GET_VF_RESOURCES, 555 I40E_SUCCESS, &reply, sizeof(reply)); 556 } 557 558 static int 559 ixl_vf_config_tx_queue(struct ixl_pf *pf, struct ixl_vf *vf, 560 struct virtchnl_txq_info *info) 561 { 562 struct i40e_hw *hw; 563 struct i40e_hmc_obj_txq txq; 564 uint16_t global_queue_num, global_vf_num; 565 enum i40e_status_code status; 566 uint32_t qtx_ctl; 567 568 hw = &pf->hw; 569 global_queue_num = ixl_pf_qidx_from_vsi_qidx(&vf->qtag, info->queue_id); 570 global_vf_num = hw->func_caps.vf_base_id + vf->vf_num; 571 bzero(&txq, sizeof(txq)); 572 573 DDPRINTF(pf->dev, "VF %d: PF TX queue %d / VF TX queue %d (Global VF %d)\n", 574 vf->vf_num, global_queue_num, info->queue_id, global_vf_num); 575 576 status = i40e_clear_lan_tx_queue_context(hw, global_queue_num); 577 if (status != I40E_SUCCESS) 578 return (EINVAL); 579 580 txq.base = info->dma_ring_addr / IXL_TX_CTX_BASE_UNITS; 581 582 txq.head_wb_ena = info->headwb_enabled; 583 txq.head_wb_addr = info->dma_headwb_addr; 584 txq.qlen = info->ring_len; 585 txq.rdylist = le16_to_cpu(vf->vsi.info.qs_handle[0]); 586 txq.rdylist_act = 0; 587 588 status = i40e_set_lan_tx_queue_context(hw, global_queue_num, &txq); 589 if (status != I40E_SUCCESS) 590 return (EINVAL); 591 592 qtx_ctl = I40E_QTX_CTL_VF_QUEUE | 593 (hw->pf_id << I40E_QTX_CTL_PF_INDX_SHIFT) | 594 (global_vf_num << I40E_QTX_CTL_VFVM_INDX_SHIFT); 595 wr32(hw, I40E_QTX_CTL(global_queue_num), qtx_ctl); 596 ixl_flush(hw); 597 598 ixl_pf_qmgr_mark_queue_configured(&vf->qtag, info->queue_id, true); 599 600 return (0); 601 } 602 603 static int 604 ixl_vf_config_rx_queue(struct ixl_pf *pf, struct ixl_vf *vf, 605 struct virtchnl_rxq_info *info) 606 { 607 struct i40e_hw *hw; 608 struct i40e_hmc_obj_rxq rxq; 609 uint16_t global_queue_num; 610 enum i40e_status_code status; 611 612 hw = &pf->hw; 613 global_queue_num = ixl_pf_qidx_from_vsi_qidx(&vf->qtag, info->queue_id); 614 bzero(&rxq, sizeof(rxq)); 615 616 DDPRINTF(pf->dev, "VF %d: PF RX queue %d / VF RX queue %d\n", 617 vf->vf_num, global_queue_num, info->queue_id); 618 619 if (info->databuffer_size > IXL_VF_MAX_BUFFER) 620 return (EINVAL); 621 622 if (info->max_pkt_size > IXL_VF_MAX_FRAME || 623 info->max_pkt_size < ETHER_MIN_LEN) 624 return (EINVAL); 625 626 if (info->splithdr_enabled) { 627 if (info->hdr_size > IXL_VF_MAX_HDR_BUFFER) 628 return (EINVAL); 629 630 rxq.hsplit_0 = info->rx_split_pos & 631 (I40E_HMC_OBJ_RX_HSPLIT_0_SPLIT_L2 | 632 I40E_HMC_OBJ_RX_HSPLIT_0_SPLIT_IP | 633 I40E_HMC_OBJ_RX_HSPLIT_0_SPLIT_TCP_UDP | 634 I40E_HMC_OBJ_RX_HSPLIT_0_SPLIT_SCTP); 635 rxq.hbuff = info->hdr_size >> I40E_RXQ_CTX_HBUFF_SHIFT; 636 637 rxq.dtype = 2; 638 } 639 640 status = i40e_clear_lan_rx_queue_context(hw, global_queue_num); 641 if (status != I40E_SUCCESS) 642 return (EINVAL); 643 644 rxq.base = info->dma_ring_addr / IXL_RX_CTX_BASE_UNITS; 645 rxq.qlen = info->ring_len; 646 647 rxq.dbuff = info->databuffer_size >> I40E_RXQ_CTX_DBUFF_SHIFT; 648 649 rxq.dsize = 1; 650 rxq.crcstrip = 1; 651 rxq.l2tsel = 1; 652 653 rxq.rxmax = info->max_pkt_size; 654 rxq.tphrdesc_ena = 1; 655 rxq.tphwdesc_ena = 1; 656 rxq.tphdata_ena = 1; 657 rxq.tphhead_ena = 1; 658 rxq.lrxqthresh = 1; 659 rxq.prefena = 1; 660 661 status = i40e_set_lan_rx_queue_context(hw, global_queue_num, &rxq); 662 if (status != I40E_SUCCESS) 663 return (EINVAL); 664 665 ixl_pf_qmgr_mark_queue_configured(&vf->qtag, info->queue_id, false); 666 667 return (0); 668 } 669 670 static void 671 ixl_vf_config_vsi_msg(struct ixl_pf *pf, struct ixl_vf *vf, void *msg, 672 uint16_t msg_size) 673 { 674 struct virtchnl_vsi_queue_config_info *info; 675 struct virtchnl_queue_pair_info *pair; 676 int i; 677 678 info = msg; 679 if (info->num_queue_pairs == 0 || info->num_queue_pairs > vf->vsi.num_tx_queues) { 680 device_printf(pf->dev, "VF %d: invalid # of qpairs (msg has %d, VSI has %d)\n", 681 vf->vf_num, info->num_queue_pairs, vf->vsi.num_tx_queues); 682 i40e_send_vf_nack(pf, vf, VIRTCHNL_OP_CONFIG_VSI_QUEUES, 683 I40E_ERR_PARAM); 684 return; 685 } 686 687 if (info->vsi_id != vf->vsi.vsi_num) { 688 device_printf(pf->dev, "VF %d: VSI id in recvd message (%d) does not match expected id (%d)\n", 689 vf->vf_num, info->vsi_id, vf->vsi.vsi_num); 690 i40e_send_vf_nack(pf, vf, VIRTCHNL_OP_CONFIG_VSI_QUEUES, 691 I40E_ERR_PARAM); 692 return; 693 } 694 695 for (i = 0; i < info->num_queue_pairs; i++) { 696 pair = &info->qpair[i]; 697 698 if (pair->txq.vsi_id != vf->vsi.vsi_num || 699 pair->rxq.vsi_id != vf->vsi.vsi_num || 700 pair->txq.queue_id != pair->rxq.queue_id || 701 pair->txq.queue_id >= vf->vsi.num_tx_queues) { 702 703 i40e_send_vf_nack(pf, vf, 704 VIRTCHNL_OP_CONFIG_VSI_QUEUES, I40E_ERR_PARAM); 705 return; 706 } 707 708 if (ixl_vf_config_tx_queue(pf, vf, &pair->txq) != 0) { 709 i40e_send_vf_nack(pf, vf, 710 VIRTCHNL_OP_CONFIG_VSI_QUEUES, I40E_ERR_PARAM); 711 return; 712 } 713 714 if (ixl_vf_config_rx_queue(pf, vf, &pair->rxq) != 0) { 715 i40e_send_vf_nack(pf, vf, 716 VIRTCHNL_OP_CONFIG_VSI_QUEUES, I40E_ERR_PARAM); 717 return; 718 } 719 } 720 721 ixl_send_vf_ack(pf, vf, VIRTCHNL_OP_CONFIG_VSI_QUEUES); 722 } 723 724 static void 725 ixl_vf_set_qctl(struct ixl_pf *pf, 726 const struct virtchnl_vector_map *vector, 727 enum i40e_queue_type cur_type, uint16_t cur_queue, 728 enum i40e_queue_type *last_type, uint16_t *last_queue) 729 { 730 uint32_t offset, qctl; 731 uint16_t itr_indx; 732 733 if (cur_type == I40E_QUEUE_TYPE_RX) { 734 offset = I40E_QINT_RQCTL(cur_queue); 735 itr_indx = vector->rxitr_idx; 736 } else { 737 offset = I40E_QINT_TQCTL(cur_queue); 738 itr_indx = vector->txitr_idx; 739 } 740 741 qctl = htole32((vector->vector_id << I40E_QINT_RQCTL_MSIX_INDX_SHIFT) | 742 (*last_type << I40E_QINT_RQCTL_NEXTQ_TYPE_SHIFT) | 743 (*last_queue << I40E_QINT_RQCTL_NEXTQ_INDX_SHIFT) | 744 I40E_QINT_RQCTL_CAUSE_ENA_MASK | 745 (itr_indx << I40E_QINT_RQCTL_ITR_INDX_SHIFT)); 746 747 wr32(&pf->hw, offset, qctl); 748 749 *last_type = cur_type; 750 *last_queue = cur_queue; 751 } 752 753 static void 754 ixl_vf_config_vector(struct ixl_pf *pf, struct ixl_vf *vf, 755 const struct virtchnl_vector_map *vector) 756 { 757 struct i40e_hw *hw; 758 u_int qindex; 759 enum i40e_queue_type type, last_type; 760 uint32_t lnklst_reg; 761 uint16_t rxq_map, txq_map, cur_queue, last_queue; 762 763 hw = &pf->hw; 764 765 rxq_map = vector->rxq_map; 766 txq_map = vector->txq_map; 767 768 last_queue = IXL_END_OF_INTR_LNKLST; 769 last_type = I40E_QUEUE_TYPE_RX; 770 771 /* 772 * The datasheet says to optimize performance, RX queues and TX queues 773 * should be interleaved in the interrupt linked list, so we process 774 * both at once here. 775 */ 776 while ((rxq_map != 0) || (txq_map != 0)) { 777 if (txq_map != 0) { 778 qindex = ffs(txq_map) - 1; 779 type = I40E_QUEUE_TYPE_TX; 780 cur_queue = ixl_pf_qidx_from_vsi_qidx(&vf->qtag, qindex); 781 ixl_vf_set_qctl(pf, vector, type, cur_queue, 782 &last_type, &last_queue); 783 txq_map &= ~(1 << qindex); 784 } 785 786 if (rxq_map != 0) { 787 qindex = ffs(rxq_map) - 1; 788 type = I40E_QUEUE_TYPE_RX; 789 cur_queue = ixl_pf_qidx_from_vsi_qidx(&vf->qtag, qindex); 790 ixl_vf_set_qctl(pf, vector, type, cur_queue, 791 &last_type, &last_queue); 792 rxq_map &= ~(1 << qindex); 793 } 794 } 795 796 if (vector->vector_id == 0) 797 lnklst_reg = I40E_VPINT_LNKLST0(vf->vf_num); 798 else 799 lnklst_reg = IXL_VPINT_LNKLSTN_REG(hw, vector->vector_id, 800 vf->vf_num); 801 wr32(hw, lnklst_reg, 802 (last_queue << I40E_VPINT_LNKLST0_FIRSTQ_INDX_SHIFT) | 803 (last_type << I40E_VPINT_LNKLST0_FIRSTQ_TYPE_SHIFT)); 804 805 ixl_flush(hw); 806 } 807 808 static void 809 ixl_vf_config_irq_msg(struct ixl_pf *pf, struct ixl_vf *vf, void *msg, 810 uint16_t msg_size) 811 { 812 struct virtchnl_irq_map_info *map; 813 struct virtchnl_vector_map *vector; 814 struct i40e_hw *hw; 815 int i, largest_txq, largest_rxq; 816 817 hw = &pf->hw; 818 map = msg; 819 820 for (i = 0; i < map->num_vectors; i++) { 821 vector = &map->vecmap[i]; 822 823 if ((vector->vector_id >= hw->func_caps.num_msix_vectors_vf) || 824 vector->vsi_id != vf->vsi.vsi_num) { 825 i40e_send_vf_nack(pf, vf, 826 VIRTCHNL_OP_CONFIG_IRQ_MAP, I40E_ERR_PARAM); 827 return; 828 } 829 830 if (vector->rxq_map != 0) { 831 largest_rxq = fls(vector->rxq_map) - 1; 832 if (largest_rxq >= vf->vsi.num_rx_queues) { 833 i40e_send_vf_nack(pf, vf, 834 VIRTCHNL_OP_CONFIG_IRQ_MAP, 835 I40E_ERR_PARAM); 836 return; 837 } 838 } 839 840 if (vector->txq_map != 0) { 841 largest_txq = fls(vector->txq_map) - 1; 842 if (largest_txq >= vf->vsi.num_tx_queues) { 843 i40e_send_vf_nack(pf, vf, 844 VIRTCHNL_OP_CONFIG_IRQ_MAP, 845 I40E_ERR_PARAM); 846 return; 847 } 848 } 849 850 if (vector->rxitr_idx > IXL_MAX_ITR_IDX || 851 vector->txitr_idx > IXL_MAX_ITR_IDX) { 852 i40e_send_vf_nack(pf, vf, 853 VIRTCHNL_OP_CONFIG_IRQ_MAP, 854 I40E_ERR_PARAM); 855 return; 856 } 857 858 ixl_vf_config_vector(pf, vf, vector); 859 } 860 861 ixl_send_vf_ack(pf, vf, VIRTCHNL_OP_CONFIG_IRQ_MAP); 862 } 863 864 static void 865 ixl_vf_enable_queues_msg(struct ixl_pf *pf, struct ixl_vf *vf, void *msg, 866 uint16_t msg_size) 867 { 868 struct virtchnl_queue_select *select; 869 int error = 0; 870 871 select = msg; 872 873 if (select->vsi_id != vf->vsi.vsi_num || 874 select->rx_queues == 0 || select->tx_queues == 0) { 875 i40e_send_vf_nack(pf, vf, VIRTCHNL_OP_ENABLE_QUEUES, 876 I40E_ERR_PARAM); 877 return; 878 } 879 880 /* Enable TX rings selected by the VF */ 881 for (int i = 0; i < 32; i++) { 882 if ((1 << i) & select->tx_queues) { 883 /* Warn if queue is out of VF allocation range */ 884 if (i >= vf->vsi.num_tx_queues) { 885 device_printf(pf->dev, "VF %d: TX ring %d is outside of VF VSI allocation!\n", 886 vf->vf_num, i); 887 break; 888 } 889 /* Skip this queue if it hasn't been configured */ 890 if (!ixl_pf_qmgr_is_queue_configured(&vf->qtag, i, true)) 891 continue; 892 /* Warn if this queue is already marked as enabled */ 893 if (ixl_pf_qmgr_is_queue_enabled(&vf->qtag, i, true)) 894 ixl_dbg_iov(pf, "VF %d: TX ring %d is already enabled!\n", 895 vf->vf_num, i); 896 897 error = ixl_enable_tx_ring(pf, &vf->qtag, i); 898 if (error) 899 break; 900 else 901 ixl_pf_qmgr_mark_queue_enabled(&vf->qtag, i, true); 902 } 903 } 904 905 /* Enable RX rings selected by the VF */ 906 for (int i = 0; i < 32; i++) { 907 if ((1 << i) & select->rx_queues) { 908 /* Warn if queue is out of VF allocation range */ 909 if (i >= vf->vsi.num_rx_queues) { 910 device_printf(pf->dev, "VF %d: RX ring %d is outside of VF VSI allocation!\n", 911 vf->vf_num, i); 912 break; 913 } 914 /* Skip this queue if it hasn't been configured */ 915 if (!ixl_pf_qmgr_is_queue_configured(&vf->qtag, i, false)) 916 continue; 917 /* Warn if this queue is already marked as enabled */ 918 if (ixl_pf_qmgr_is_queue_enabled(&vf->qtag, i, false)) 919 ixl_dbg_iov(pf, "VF %d: RX ring %d is already enabled!\n", 920 vf->vf_num, i); 921 error = ixl_enable_rx_ring(pf, &vf->qtag, i); 922 if (error) 923 break; 924 else 925 ixl_pf_qmgr_mark_queue_enabled(&vf->qtag, i, false); 926 } 927 } 928 929 if (error) { 930 i40e_send_vf_nack(pf, vf, VIRTCHNL_OP_ENABLE_QUEUES, 931 I40E_ERR_TIMEOUT); 932 return; 933 } 934 935 ixl_send_vf_ack(pf, vf, VIRTCHNL_OP_ENABLE_QUEUES); 936 } 937 938 static void 939 ixl_vf_disable_queues_msg(struct ixl_pf *pf, struct ixl_vf *vf, 940 void *msg, uint16_t msg_size) 941 { 942 struct virtchnl_queue_select *select; 943 int error = 0; 944 945 select = msg; 946 947 if (select->vsi_id != vf->vsi.vsi_num || 948 select->rx_queues == 0 || select->tx_queues == 0) { 949 i40e_send_vf_nack(pf, vf, VIRTCHNL_OP_DISABLE_QUEUES, 950 I40E_ERR_PARAM); 951 return; 952 } 953 954 /* Disable TX rings selected by the VF */ 955 for (int i = 0; i < 32; i++) { 956 if ((1 << i) & select->tx_queues) { 957 /* Warn if queue is out of VF allocation range */ 958 if (i >= vf->vsi.num_tx_queues) { 959 device_printf(pf->dev, "VF %d: TX ring %d is outside of VF VSI allocation!\n", 960 vf->vf_num, i); 961 break; 962 } 963 /* Skip this queue if it hasn't been configured */ 964 if (!ixl_pf_qmgr_is_queue_configured(&vf->qtag, i, true)) 965 continue; 966 /* Warn if this queue is already marked as disabled */ 967 if (!ixl_pf_qmgr_is_queue_enabled(&vf->qtag, i, true)) { 968 ixl_dbg_iov(pf, "VF %d: TX ring %d is already disabled!\n", 969 vf->vf_num, i); 970 continue; 971 } 972 error = ixl_disable_tx_ring(pf, &vf->qtag, i); 973 if (error) 974 break; 975 else 976 ixl_pf_qmgr_mark_queue_disabled(&vf->qtag, i, true); 977 } 978 } 979 980 /* Enable RX rings selected by the VF */ 981 for (int i = 0; i < 32; i++) { 982 if ((1 << i) & select->rx_queues) { 983 /* Warn if queue is out of VF allocation range */ 984 if (i >= vf->vsi.num_rx_queues) { 985 device_printf(pf->dev, "VF %d: RX ring %d is outside of VF VSI allocation!\n", 986 vf->vf_num, i); 987 break; 988 } 989 /* Skip this queue if it hasn't been configured */ 990 if (!ixl_pf_qmgr_is_queue_configured(&vf->qtag, i, false)) 991 continue; 992 /* Warn if this queue is already marked as disabled */ 993 if (!ixl_pf_qmgr_is_queue_enabled(&vf->qtag, i, false)) { 994 ixl_dbg_iov(pf, "VF %d: RX ring %d is already disabled!\n", 995 vf->vf_num, i); 996 continue; 997 } 998 error = ixl_disable_rx_ring(pf, &vf->qtag, i); 999 if (error) 1000 break; 1001 else 1002 ixl_pf_qmgr_mark_queue_disabled(&vf->qtag, i, false); 1003 } 1004 } 1005 1006 if (error) { 1007 i40e_send_vf_nack(pf, vf, VIRTCHNL_OP_DISABLE_QUEUES, 1008 I40E_ERR_TIMEOUT); 1009 return; 1010 } 1011 1012 ixl_send_vf_ack(pf, vf, VIRTCHNL_OP_DISABLE_QUEUES); 1013 } 1014 1015 static int 1016 ixl_vf_mac_valid(struct ixl_vf *vf, const uint8_t *addr) 1017 { 1018 1019 if (ETHER_IS_ZERO(addr) || ETHER_IS_BROADCAST(addr)) 1020 return (EINVAL); 1021 1022 /* 1023 * If the VF is not allowed to change its MAC address, don't let it 1024 * set a MAC filter for an address that is not a multicast address and 1025 * is not its assigned MAC. 1026 */ 1027 if (!(vf->vf_flags & VF_FLAG_SET_MAC_CAP) && 1028 !(ETHER_IS_MULTICAST(addr) || !ixl_ether_is_equal(addr, vf->mac))) 1029 return (EPERM); 1030 1031 return (0); 1032 } 1033 1034 static void 1035 ixl_vf_add_mac_msg(struct ixl_pf *pf, struct ixl_vf *vf, void *msg, 1036 uint16_t msg_size) 1037 { 1038 struct virtchnl_ether_addr_list *addr_list; 1039 struct virtchnl_ether_addr *addr; 1040 struct ixl_vsi *vsi; 1041 int i; 1042 1043 vsi = &vf->vsi; 1044 addr_list = msg; 1045 1046 if (addr_list->vsi_id != vsi->vsi_num) { 1047 i40e_send_vf_nack(pf, vf, VIRTCHNL_OP_ADD_ETH_ADDR, 1048 I40E_ERR_PARAM); 1049 return; 1050 } 1051 1052 for (i = 0; i < addr_list->num_elements; i++) { 1053 if (ixl_vf_mac_valid(vf, addr_list->list[i].addr) != 0) { 1054 i40e_send_vf_nack(pf, vf, 1055 VIRTCHNL_OP_ADD_ETH_ADDR, I40E_ERR_PARAM); 1056 return; 1057 } 1058 } 1059 1060 for (i = 0; i < addr_list->num_elements; i++) { 1061 addr = &addr_list->list[i]; 1062 ixl_add_filter(vsi, addr->addr, IXL_VLAN_ANY); 1063 } 1064 1065 ixl_send_vf_ack(pf, vf, VIRTCHNL_OP_ADD_ETH_ADDR); 1066 } 1067 1068 static void 1069 ixl_vf_del_mac_msg(struct ixl_pf *pf, struct ixl_vf *vf, void *msg, 1070 uint16_t msg_size) 1071 { 1072 struct virtchnl_ether_addr_list *addr_list; 1073 struct virtchnl_ether_addr *addr; 1074 struct ixl_vsi *vsi; 1075 int i; 1076 1077 vsi = &vf->vsi; 1078 addr_list = msg; 1079 1080 if (addr_list->vsi_id != vsi->vsi_num) { 1081 i40e_send_vf_nack(pf, vf, VIRTCHNL_OP_DEL_ETH_ADDR, 1082 I40E_ERR_PARAM); 1083 return; 1084 } 1085 1086 for (i = 0; i < addr_list->num_elements; i++) { 1087 addr = &addr_list->list[i]; 1088 if (ETHER_IS_ZERO(addr->addr) || ETHER_IS_BROADCAST(addr->addr)) { 1089 i40e_send_vf_nack(pf, vf, 1090 VIRTCHNL_OP_DEL_ETH_ADDR, I40E_ERR_PARAM); 1091 return; 1092 } 1093 } 1094 1095 for (i = 0; i < addr_list->num_elements; i++) { 1096 addr = &addr_list->list[i]; 1097 ixl_del_filter(&vf->vsi, addr->addr, IXL_VLAN_ANY); 1098 } 1099 1100 ixl_send_vf_ack(pf, vf, VIRTCHNL_OP_DEL_ETH_ADDR); 1101 } 1102 1103 static enum i40e_status_code 1104 ixl_vf_enable_vlan_strip(struct ixl_pf *pf, struct ixl_vf *vf) 1105 { 1106 struct i40e_vsi_context vsi_ctx; 1107 1108 vsi_ctx.seid = vf->vsi.seid; 1109 1110 bzero(&vsi_ctx.info, sizeof(vsi_ctx.info)); 1111 vsi_ctx.info.valid_sections = htole16(I40E_AQ_VSI_PROP_VLAN_VALID); 1112 vsi_ctx.info.port_vlan_flags = I40E_AQ_VSI_PVLAN_MODE_ALL | 1113 I40E_AQ_VSI_PVLAN_EMOD_STR_BOTH; 1114 return (i40e_aq_update_vsi_params(&pf->hw, &vsi_ctx, NULL)); 1115 } 1116 1117 static void 1118 ixl_vf_add_vlan_msg(struct ixl_pf *pf, struct ixl_vf *vf, void *msg, 1119 uint16_t msg_size) 1120 { 1121 struct virtchnl_vlan_filter_list *filter_list; 1122 enum i40e_status_code code; 1123 int i; 1124 1125 filter_list = msg; 1126 1127 if (filter_list->vsi_id != vf->vsi.vsi_num) { 1128 i40e_send_vf_nack(pf, vf, VIRTCHNL_OP_ADD_VLAN, 1129 I40E_ERR_PARAM); 1130 return; 1131 } 1132 1133 if (!(vf->vf_flags & VF_FLAG_VLAN_CAP)) { 1134 i40e_send_vf_nack(pf, vf, VIRTCHNL_OP_ADD_VLAN, 1135 I40E_ERR_PARAM); 1136 return; 1137 } 1138 1139 for (i = 0; i < filter_list->num_elements; i++) { 1140 if (filter_list->vlan_id[i] > EVL_VLID_MASK) { 1141 i40e_send_vf_nack(pf, vf, VIRTCHNL_OP_ADD_VLAN, 1142 I40E_ERR_PARAM); 1143 return; 1144 } 1145 } 1146 1147 code = ixl_vf_enable_vlan_strip(pf, vf); 1148 if (code != I40E_SUCCESS) { 1149 i40e_send_vf_nack(pf, vf, VIRTCHNL_OP_ADD_VLAN, 1150 I40E_ERR_PARAM); 1151 } 1152 1153 for (i = 0; i < filter_list->num_elements; i++) 1154 ixl_add_filter(&vf->vsi, vf->mac, filter_list->vlan_id[i]); 1155 1156 ixl_send_vf_ack(pf, vf, VIRTCHNL_OP_ADD_VLAN); 1157 } 1158 1159 static void 1160 ixl_vf_del_vlan_msg(struct ixl_pf *pf, struct ixl_vf *vf, void *msg, 1161 uint16_t msg_size) 1162 { 1163 struct virtchnl_vlan_filter_list *filter_list; 1164 int i; 1165 1166 filter_list = msg; 1167 1168 if (filter_list->vsi_id != vf->vsi.vsi_num) { 1169 i40e_send_vf_nack(pf, vf, VIRTCHNL_OP_DEL_VLAN, 1170 I40E_ERR_PARAM); 1171 return; 1172 } 1173 1174 for (i = 0; i < filter_list->num_elements; i++) { 1175 if (filter_list->vlan_id[i] > EVL_VLID_MASK) { 1176 i40e_send_vf_nack(pf, vf, VIRTCHNL_OP_ADD_VLAN, 1177 I40E_ERR_PARAM); 1178 return; 1179 } 1180 } 1181 1182 if (!(vf->vf_flags & VF_FLAG_VLAN_CAP)) { 1183 i40e_send_vf_nack(pf, vf, VIRTCHNL_OP_ADD_VLAN, 1184 I40E_ERR_PARAM); 1185 return; 1186 } 1187 1188 for (i = 0; i < filter_list->num_elements; i++) 1189 ixl_del_filter(&vf->vsi, vf->mac, filter_list->vlan_id[i]); 1190 1191 ixl_send_vf_ack(pf, vf, VIRTCHNL_OP_DEL_VLAN); 1192 } 1193 1194 static void 1195 ixl_vf_config_promisc_msg(struct ixl_pf *pf, struct ixl_vf *vf, 1196 void *msg, uint16_t msg_size) 1197 { 1198 struct virtchnl_promisc_info *info; 1199 struct i40e_hw *hw = &pf->hw; 1200 enum i40e_status_code code; 1201 1202 if (!(vf->vf_flags & VF_FLAG_PROMISC_CAP)) { 1203 /* 1204 * Do the same thing as the Linux PF driver -- lie to the VF 1205 */ 1206 ixl_send_vf_ack(pf, vf, 1207 VIRTCHNL_OP_CONFIG_PROMISCUOUS_MODE); 1208 return; 1209 } 1210 1211 info = msg; 1212 if (info->vsi_id != vf->vsi.vsi_num) { 1213 i40e_send_vf_nack(pf, vf, 1214 VIRTCHNL_OP_CONFIG_PROMISCUOUS_MODE, I40E_ERR_PARAM); 1215 return; 1216 } 1217 1218 code = i40e_aq_set_vsi_unicast_promiscuous(hw, vf->vsi.seid, 1219 info->flags & FLAG_VF_UNICAST_PROMISC, NULL, TRUE); 1220 if (code != I40E_SUCCESS) { 1221 device_printf(pf->dev, "i40e_aq_set_vsi_unicast_promiscuous (seid %d) failed: status %s," 1222 " error %s\n", vf->vsi.seid, i40e_stat_str(hw, code), 1223 i40e_aq_str(hw, hw->aq.asq_last_status)); 1224 i40e_send_vf_nack(pf, vf, 1225 VIRTCHNL_OP_CONFIG_PROMISCUOUS_MODE, I40E_ERR_PARAM); 1226 return; 1227 } 1228 1229 code = i40e_aq_set_vsi_multicast_promiscuous(hw, vf->vsi.seid, 1230 info->flags & FLAG_VF_MULTICAST_PROMISC, NULL); 1231 if (code != I40E_SUCCESS) { 1232 device_printf(pf->dev, "i40e_aq_set_vsi_multicast_promiscuous (seid %d) failed: status %s," 1233 " error %s\n", vf->vsi.seid, i40e_stat_str(hw, code), 1234 i40e_aq_str(hw, hw->aq.asq_last_status)); 1235 i40e_send_vf_nack(pf, vf, 1236 VIRTCHNL_OP_CONFIG_PROMISCUOUS_MODE, I40E_ERR_PARAM); 1237 return; 1238 } 1239 1240 ixl_send_vf_ack(pf, vf, VIRTCHNL_OP_CONFIG_PROMISCUOUS_MODE); 1241 } 1242 1243 static void 1244 ixl_vf_get_stats_msg(struct ixl_pf *pf, struct ixl_vf *vf, void *msg, 1245 uint16_t msg_size) 1246 { 1247 struct virtchnl_queue_select *queue; 1248 1249 queue = msg; 1250 if (queue->vsi_id != vf->vsi.vsi_num) { 1251 i40e_send_vf_nack(pf, vf, VIRTCHNL_OP_GET_STATS, 1252 I40E_ERR_PARAM); 1253 return; 1254 } 1255 1256 ixl_update_eth_stats(&vf->vsi); 1257 1258 ixl_send_vf_msg(pf, vf, VIRTCHNL_OP_GET_STATS, 1259 I40E_SUCCESS, &vf->vsi.eth_stats, sizeof(vf->vsi.eth_stats)); 1260 } 1261 1262 static void 1263 ixl_vf_config_rss_key_msg(struct ixl_pf *pf, struct ixl_vf *vf, void *msg, 1264 uint16_t msg_size) 1265 { 1266 struct i40e_hw *hw; 1267 struct virtchnl_rss_key *key; 1268 struct i40e_aqc_get_set_rss_key_data key_data; 1269 enum i40e_status_code status; 1270 1271 hw = &pf->hw; 1272 1273 key = msg; 1274 1275 if (key->key_len > 52) { 1276 device_printf(pf->dev, "VF %d: Key size in msg (%d) is greater than max key size (%d)\n", 1277 vf->vf_num, key->key_len, 52); 1278 i40e_send_vf_nack(pf, vf, VIRTCHNL_OP_CONFIG_RSS_KEY, 1279 I40E_ERR_PARAM); 1280 return; 1281 } 1282 1283 if (key->vsi_id != vf->vsi.vsi_num) { 1284 device_printf(pf->dev, "VF %d: VSI id in recvd message (%d) does not match expected id (%d)\n", 1285 vf->vf_num, key->vsi_id, vf->vsi.vsi_num); 1286 i40e_send_vf_nack(pf, vf, VIRTCHNL_OP_CONFIG_RSS_KEY, 1287 I40E_ERR_PARAM); 1288 return; 1289 } 1290 1291 /* Fill out hash using MAC-dependent method */ 1292 if (hw->mac.type == I40E_MAC_X722) { 1293 bzero(&key_data, sizeof(key_data)); 1294 if (key->key_len <= 40) 1295 bcopy(key->key, key_data.standard_rss_key, key->key_len); 1296 else { 1297 bcopy(key->key, key_data.standard_rss_key, 40); 1298 bcopy(&key->key[40], key_data.extended_hash_key, key->key_len - 40); 1299 } 1300 status = i40e_aq_set_rss_key(hw, vf->vsi.vsi_num, &key_data); 1301 if (status) { 1302 device_printf(pf->dev, "i40e_aq_set_rss_key status %s, error %s\n", 1303 i40e_stat_str(hw, status), i40e_aq_str(hw, hw->aq.asq_last_status)); 1304 i40e_send_vf_nack(pf, vf, VIRTCHNL_OP_CONFIG_RSS_KEY, 1305 I40E_ERR_ADMIN_QUEUE_ERROR); 1306 return; 1307 } 1308 } else { 1309 for (int i = 0; i < (key->key_len / 4); i++) 1310 i40e_write_rx_ctl(hw, I40E_VFQF_HKEY1(i, vf->vf_num), ((u32 *)key->key)[i]); 1311 } 1312 1313 DDPRINTF(pf->dev, "VF %d: Programmed key starting with 0x%x ok!", 1314 vf->vf_num, key->key[0]); 1315 1316 ixl_send_vf_ack(pf, vf, VIRTCHNL_OP_CONFIG_RSS_KEY); 1317 } 1318 1319 static void 1320 ixl_vf_config_rss_lut_msg(struct ixl_pf *pf, struct ixl_vf *vf, void *msg, 1321 uint16_t msg_size) 1322 { 1323 struct i40e_hw *hw; 1324 struct virtchnl_rss_lut *lut; 1325 enum i40e_status_code status; 1326 1327 hw = &pf->hw; 1328 1329 lut = msg; 1330 1331 if (lut->lut_entries > 64) { 1332 device_printf(pf->dev, "VF %d: # of LUT entries in msg (%d) is greater than max (%d)\n", 1333 vf->vf_num, lut->lut_entries, 64); 1334 i40e_send_vf_nack(pf, vf, VIRTCHNL_OP_CONFIG_RSS_LUT, 1335 I40E_ERR_PARAM); 1336 return; 1337 } 1338 1339 if (lut->vsi_id != vf->vsi.vsi_num) { 1340 device_printf(pf->dev, "VF %d: VSI id in recvd message (%d) does not match expected id (%d)\n", 1341 vf->vf_num, lut->vsi_id, vf->vsi.vsi_num); 1342 i40e_send_vf_nack(pf, vf, VIRTCHNL_OP_CONFIG_RSS_LUT, 1343 I40E_ERR_PARAM); 1344 return; 1345 } 1346 1347 /* Fill out LUT using MAC-dependent method */ 1348 if (hw->mac.type == I40E_MAC_X722) { 1349 status = i40e_aq_set_rss_lut(hw, vf->vsi.vsi_num, false, lut->lut, lut->lut_entries); 1350 if (status) { 1351 device_printf(pf->dev, "i40e_aq_set_rss_lut status %s, error %s\n", 1352 i40e_stat_str(hw, status), i40e_aq_str(hw, hw->aq.asq_last_status)); 1353 i40e_send_vf_nack(pf, vf, VIRTCHNL_OP_CONFIG_RSS_LUT, 1354 I40E_ERR_ADMIN_QUEUE_ERROR); 1355 return; 1356 } 1357 } else { 1358 for (int i = 0; i < (lut->lut_entries / 4); i++) 1359 i40e_write_rx_ctl(hw, I40E_VFQF_HLUT1(i, vf->vf_num), ((u32 *)lut->lut)[i]); 1360 } 1361 1362 DDPRINTF(pf->dev, "VF %d: Programmed LUT starting with 0x%x and length %d ok!", 1363 vf->vf_num, lut->lut[0], lut->lut_entries); 1364 1365 ixl_send_vf_ack(pf, vf, VIRTCHNL_OP_CONFIG_RSS_LUT); 1366 } 1367 1368 static void 1369 ixl_vf_set_rss_hena_msg(struct ixl_pf *pf, struct ixl_vf *vf, void *msg, 1370 uint16_t msg_size) 1371 { 1372 struct i40e_hw *hw; 1373 struct virtchnl_rss_hena *hena; 1374 1375 hw = &pf->hw; 1376 hena = msg; 1377 1378 /* Set HENA */ 1379 i40e_write_rx_ctl(hw, I40E_VFQF_HENA1(0, vf->vf_num), (u32)hena->hena); 1380 i40e_write_rx_ctl(hw, I40E_VFQF_HENA1(1, vf->vf_num), (u32)(hena->hena >> 32)); 1381 1382 DDPRINTF(pf->dev, "VF %d: Programmed HENA with 0x%016lx", 1383 vf->vf_num, hena->hena); 1384 1385 ixl_send_vf_ack(pf, vf, VIRTCHNL_OP_SET_RSS_HENA); 1386 } 1387 1388 static void 1389 ixl_notify_vf_link_state(struct ixl_pf *pf, struct ixl_vf *vf) 1390 { 1391 struct virtchnl_pf_event event; 1392 struct i40e_hw *hw; 1393 1394 hw = &pf->hw; 1395 event.event = VIRTCHNL_EVENT_LINK_CHANGE; 1396 event.severity = PF_EVENT_SEVERITY_INFO; 1397 event.event_data.link_event.link_status = pf->vsi.link_active; 1398 event.event_data.link_event.link_speed = 1399 i40e_virtchnl_link_speed(hw->phy.link_info.link_speed); 1400 1401 ixl_send_vf_msg(pf, vf, VIRTCHNL_OP_EVENT, I40E_SUCCESS, &event, 1402 sizeof(event)); 1403 } 1404 1405 void 1406 ixl_broadcast_link_state(struct ixl_pf *pf) 1407 { 1408 int i; 1409 1410 for (i = 0; i < pf->num_vfs; i++) 1411 ixl_notify_vf_link_state(pf, &pf->vfs[i]); 1412 } 1413 1414 void 1415 ixl_handle_vf_msg(struct ixl_pf *pf, struct i40e_arq_event_info *event) 1416 { 1417 device_t dev = pf->dev; 1418 struct ixl_vf *vf; 1419 uint16_t vf_num, msg_size; 1420 uint32_t opcode; 1421 void *msg; 1422 int err; 1423 1424 vf_num = le16toh(event->desc.retval) - pf->hw.func_caps.vf_base_id; 1425 opcode = le32toh(event->desc.cookie_high); 1426 1427 if (vf_num >= pf->num_vfs) { 1428 device_printf(pf->dev, "Got msg from illegal VF: %d\n", vf_num); 1429 return; 1430 } 1431 1432 vf = &pf->vfs[vf_num]; 1433 msg = event->msg_buf; 1434 msg_size = event->msg_len; 1435 1436 I40E_VC_DEBUG(pf, ixl_vc_opcode_level(opcode), 1437 "Got msg %s(%d) from%sVF-%d of size %d\n", 1438 ixl_vc_opcode_str(opcode), opcode, 1439 (vf->vf_flags & VF_FLAG_ENABLED) ? " " : " disabled ", 1440 vf_num, msg_size); 1441 1442 /* Perform basic checks on the msg */ 1443 err = virtchnl_vc_validate_vf_msg(&vf->version, opcode, msg, msg_size); 1444 if (err) { 1445 device_printf(dev, "%s: Received invalid msg from VF-%d: opcode %d, len %d, error %d\n", 1446 __func__, vf->vf_num, opcode, msg_size, err); 1447 i40e_send_vf_nack(pf, vf, opcode, I40E_ERR_PARAM); 1448 return; 1449 } 1450 1451 /* This must be a stray msg from a previously destroyed VF. */ 1452 if (!(vf->vf_flags & VF_FLAG_ENABLED)) 1453 return; 1454 1455 switch (opcode) { 1456 case VIRTCHNL_OP_VERSION: 1457 ixl_vf_version_msg(pf, vf, msg, msg_size); 1458 break; 1459 case VIRTCHNL_OP_RESET_VF: 1460 ixl_vf_reset_msg(pf, vf, msg, msg_size); 1461 break; 1462 case VIRTCHNL_OP_GET_VF_RESOURCES: 1463 ixl_vf_get_resources_msg(pf, vf, msg, msg_size); 1464 /* Notify VF of link state after it obtains queues, as this is 1465 * the last thing it will do as part of initialization 1466 */ 1467 ixl_notify_vf_link_state(pf, vf); 1468 break; 1469 case VIRTCHNL_OP_CONFIG_VSI_QUEUES: 1470 ixl_vf_config_vsi_msg(pf, vf, msg, msg_size); 1471 break; 1472 case VIRTCHNL_OP_CONFIG_IRQ_MAP: 1473 ixl_vf_config_irq_msg(pf, vf, msg, msg_size); 1474 break; 1475 case VIRTCHNL_OP_ENABLE_QUEUES: 1476 ixl_vf_enable_queues_msg(pf, vf, msg, msg_size); 1477 /* Notify VF of link state after it obtains queues, as this is 1478 * the last thing it will do as part of initialization 1479 */ 1480 ixl_notify_vf_link_state(pf, vf); 1481 break; 1482 case VIRTCHNL_OP_DISABLE_QUEUES: 1483 ixl_vf_disable_queues_msg(pf, vf, msg, msg_size); 1484 break; 1485 case VIRTCHNL_OP_ADD_ETH_ADDR: 1486 ixl_vf_add_mac_msg(pf, vf, msg, msg_size); 1487 break; 1488 case VIRTCHNL_OP_DEL_ETH_ADDR: 1489 ixl_vf_del_mac_msg(pf, vf, msg, msg_size); 1490 break; 1491 case VIRTCHNL_OP_ADD_VLAN: 1492 ixl_vf_add_vlan_msg(pf, vf, msg, msg_size); 1493 break; 1494 case VIRTCHNL_OP_DEL_VLAN: 1495 ixl_vf_del_vlan_msg(pf, vf, msg, msg_size); 1496 break; 1497 case VIRTCHNL_OP_CONFIG_PROMISCUOUS_MODE: 1498 ixl_vf_config_promisc_msg(pf, vf, msg, msg_size); 1499 break; 1500 case VIRTCHNL_OP_GET_STATS: 1501 ixl_vf_get_stats_msg(pf, vf, msg, msg_size); 1502 break; 1503 case VIRTCHNL_OP_CONFIG_RSS_KEY: 1504 ixl_vf_config_rss_key_msg(pf, vf, msg, msg_size); 1505 break; 1506 case VIRTCHNL_OP_CONFIG_RSS_LUT: 1507 ixl_vf_config_rss_lut_msg(pf, vf, msg, msg_size); 1508 break; 1509 case VIRTCHNL_OP_SET_RSS_HENA: 1510 ixl_vf_set_rss_hena_msg(pf, vf, msg, msg_size); 1511 break; 1512 1513 /* These two opcodes have been superseded by CONFIG_VSI_QUEUES. */ 1514 case VIRTCHNL_OP_CONFIG_TX_QUEUE: 1515 case VIRTCHNL_OP_CONFIG_RX_QUEUE: 1516 default: 1517 i40e_send_vf_nack(pf, vf, opcode, I40E_ERR_NOT_IMPLEMENTED); 1518 break; 1519 } 1520 } 1521 1522 /* Handle any VFs that have reset themselves via a Function Level Reset(FLR). */ 1523 void 1524 ixl_handle_vflr(struct ixl_pf *pf) 1525 { 1526 struct ixl_vf *vf; 1527 struct i40e_hw *hw; 1528 uint16_t global_vf_num; 1529 uint32_t vflrstat_index, vflrstat_mask, vflrstat, icr0; 1530 int i; 1531 1532 hw = &pf->hw; 1533 1534 ixl_dbg_iov(pf, "%s: begin\n", __func__); 1535 1536 /* Re-enable VFLR interrupt cause so driver doesn't miss a 1537 * reset interrupt for another VF */ 1538 icr0 = rd32(hw, I40E_PFINT_ICR0_ENA); 1539 icr0 |= I40E_PFINT_ICR0_ENA_VFLR_MASK; 1540 wr32(hw, I40E_PFINT_ICR0_ENA, icr0); 1541 ixl_flush(hw); 1542 1543 for (i = 0; i < pf->num_vfs; i++) { 1544 global_vf_num = hw->func_caps.vf_base_id + i; 1545 1546 vf = &pf->vfs[i]; 1547 if (!(vf->vf_flags & VF_FLAG_ENABLED)) 1548 continue; 1549 1550 vflrstat_index = IXL_GLGEN_VFLRSTAT_INDEX(global_vf_num); 1551 vflrstat_mask = IXL_GLGEN_VFLRSTAT_MASK(global_vf_num); 1552 vflrstat = rd32(hw, I40E_GLGEN_VFLRSTAT(vflrstat_index)); 1553 if (vflrstat & vflrstat_mask) { 1554 wr32(hw, I40E_GLGEN_VFLRSTAT(vflrstat_index), 1555 vflrstat_mask); 1556 1557 ixl_dbg_iov(pf, "Reinitializing VF-%d\n", i); 1558 ixl_reinit_vf(pf, vf); 1559 ixl_dbg_iov(pf, "Reinitializing VF-%d done\n", i); 1560 } 1561 } 1562 1563 } 1564 1565 static int 1566 ixl_adminq_err_to_errno(enum i40e_admin_queue_err err) 1567 { 1568 1569 switch (err) { 1570 case I40E_AQ_RC_EPERM: 1571 return (EPERM); 1572 case I40E_AQ_RC_ENOENT: 1573 return (ENOENT); 1574 case I40E_AQ_RC_ESRCH: 1575 return (ESRCH); 1576 case I40E_AQ_RC_EINTR: 1577 return (EINTR); 1578 case I40E_AQ_RC_EIO: 1579 return (EIO); 1580 case I40E_AQ_RC_ENXIO: 1581 return (ENXIO); 1582 case I40E_AQ_RC_E2BIG: 1583 return (E2BIG); 1584 case I40E_AQ_RC_EAGAIN: 1585 return (EAGAIN); 1586 case I40E_AQ_RC_ENOMEM: 1587 return (ENOMEM); 1588 case I40E_AQ_RC_EACCES: 1589 return (EACCES); 1590 case I40E_AQ_RC_EFAULT: 1591 return (EFAULT); 1592 case I40E_AQ_RC_EBUSY: 1593 return (EBUSY); 1594 case I40E_AQ_RC_EEXIST: 1595 return (EEXIST); 1596 case I40E_AQ_RC_EINVAL: 1597 return (EINVAL); 1598 case I40E_AQ_RC_ENOTTY: 1599 return (ENOTTY); 1600 case I40E_AQ_RC_ENOSPC: 1601 return (ENOSPC); 1602 case I40E_AQ_RC_ENOSYS: 1603 return (ENOSYS); 1604 case I40E_AQ_RC_ERANGE: 1605 return (ERANGE); 1606 case I40E_AQ_RC_EFLUSHED: 1607 return (EINVAL); /* No exact equivalent in errno.h */ 1608 case I40E_AQ_RC_BAD_ADDR: 1609 return (EFAULT); 1610 case I40E_AQ_RC_EMODE: 1611 return (EPERM); 1612 case I40E_AQ_RC_EFBIG: 1613 return (EFBIG); 1614 default: 1615 return (EINVAL); 1616 } 1617 } 1618 1619 static int 1620 ixl_config_pf_vsi_loopback(struct ixl_pf *pf, bool enable) 1621 { 1622 struct i40e_hw *hw = &pf->hw; 1623 device_t dev = pf->dev; 1624 struct ixl_vsi *vsi = &pf->vsi; 1625 struct i40e_vsi_context ctxt; 1626 int error; 1627 1628 memset(&ctxt, 0, sizeof(ctxt)); 1629 1630 ctxt.seid = vsi->seid; 1631 if (pf->veb_seid != 0) 1632 ctxt.uplink_seid = pf->veb_seid; 1633 ctxt.pf_num = hw->pf_id; 1634 ctxt.connection_type = IXL_VSI_DATA_PORT; 1635 1636 ctxt.info.valid_sections = htole16(I40E_AQ_VSI_PROP_SWITCH_VALID); 1637 ctxt.info.switch_id = (enable) ? 1638 htole16(I40E_AQ_VSI_SW_ID_FLAG_ALLOW_LB) : 0; 1639 1640 /* error is set to 0 on success */ 1641 error = i40e_aq_update_vsi_params(hw, &ctxt, NULL); 1642 if (error) { 1643 device_printf(dev, "i40e_aq_update_vsi_params() failed, error %d," 1644 " aq_error %d\n", error, hw->aq.asq_last_status); 1645 } 1646 1647 return (error); 1648 } 1649 1650 int 1651 ixl_if_iov_init(if_ctx_t ctx, uint16_t num_vfs, const nvlist_t *params) 1652 { 1653 struct ixl_pf *pf = iflib_get_softc(ctx); 1654 device_t dev = iflib_get_dev(ctx); 1655 struct i40e_hw *hw; 1656 struct ixl_vsi *pf_vsi; 1657 enum i40e_status_code ret; 1658 int error; 1659 1660 hw = &pf->hw; 1661 pf_vsi = &pf->vsi; 1662 1663 pf->vfs = malloc(sizeof(struct ixl_vf) * num_vfs, M_IXL, M_NOWAIT | 1664 M_ZERO); 1665 if (pf->vfs == NULL) { 1666 error = ENOMEM; 1667 goto fail; 1668 } 1669 1670 /* 1671 * Add the VEB and ... 1672 * - do nothing: VEPA mode 1673 * - enable loopback mode on connected VSIs: VEB mode 1674 */ 1675 ret = i40e_aq_add_veb(hw, pf_vsi->uplink_seid, pf_vsi->seid, 1676 1, FALSE, &pf->veb_seid, FALSE, NULL); 1677 if (ret != I40E_SUCCESS) { 1678 error = hw->aq.asq_last_status; 1679 device_printf(dev, "i40e_aq_add_veb failed; status %s error %s", 1680 i40e_stat_str(hw, ret), i40e_aq_str(hw, error)); 1681 goto fail; 1682 } 1683 if (pf->enable_vf_loopback) 1684 ixl_config_pf_vsi_loopback(pf, true); 1685 1686 /* 1687 * Adding a VEB brings back the default MAC filter(s). Remove them, 1688 * and let the driver add the proper filters back. 1689 */ 1690 ixl_del_default_hw_filters(pf_vsi); 1691 ixl_reconfigure_filters(pf_vsi); 1692 1693 pf->num_vfs = num_vfs; 1694 return (0); 1695 1696 fail: 1697 free(pf->vfs, M_IXL); 1698 pf->vfs = NULL; 1699 return (error); 1700 } 1701 1702 void 1703 ixl_if_iov_uninit(if_ctx_t ctx) 1704 { 1705 struct ixl_pf *pf = iflib_get_softc(ctx); 1706 struct i40e_hw *hw; 1707 struct ixl_vsi *vsi; 1708 struct ifnet *ifp; 1709 struct ixl_vf *vfs; 1710 int i, num_vfs; 1711 1712 hw = &pf->hw; 1713 vsi = &pf->vsi; 1714 ifp = vsi->ifp; 1715 1716 for (i = 0; i < pf->num_vfs; i++) { 1717 if (pf->vfs[i].vsi.seid != 0) 1718 i40e_aq_delete_element(hw, pf->vfs[i].vsi.seid, NULL); 1719 ixl_pf_qmgr_release(&pf->qmgr, &pf->vfs[i].qtag); 1720 ixl_free_filters(&pf->vfs[i].vsi.ftl); 1721 ixl_dbg_iov(pf, "VF %d: %d released\n", 1722 i, pf->vfs[i].qtag.num_allocated); 1723 ixl_dbg_iov(pf, "Unallocated total: %d\n", ixl_pf_qmgr_get_num_free(&pf->qmgr)); 1724 } 1725 1726 if (pf->veb_seid != 0) { 1727 i40e_aq_delete_element(hw, pf->veb_seid, NULL); 1728 pf->veb_seid = 0; 1729 } 1730 /* Reset PF VSI loopback mode */ 1731 if (pf->enable_vf_loopback) 1732 ixl_config_pf_vsi_loopback(pf, false); 1733 1734 vfs = pf->vfs; 1735 num_vfs = pf->num_vfs; 1736 1737 pf->vfs = NULL; 1738 pf->num_vfs = 0; 1739 1740 /* sysctl_ctx_free might sleep, but this func is called w/ an sx lock */ 1741 for (i = 0; i < num_vfs; i++) 1742 sysctl_ctx_free(&vfs[i].vsi.sysctl_ctx); 1743 free(vfs, M_IXL); 1744 } 1745 1746 static int 1747 ixl_vf_reserve_queues(struct ixl_pf *pf, struct ixl_vf *vf, int num_queues) 1748 { 1749 device_t dev = pf->dev; 1750 int error; 1751 1752 /* Validate, and clamp value if invalid */ 1753 if (num_queues < 1 || num_queues > 16) 1754 device_printf(dev, "Invalid num-queues (%d) for VF %d\n", 1755 num_queues, vf->vf_num); 1756 if (num_queues < 1) { 1757 device_printf(dev, "Setting VF %d num-queues to 1\n", vf->vf_num); 1758 num_queues = 1; 1759 } else if (num_queues > IAVF_MAX_QUEUES) { 1760 device_printf(dev, "Setting VF %d num-queues to %d\n", vf->vf_num, IAVF_MAX_QUEUES); 1761 num_queues = IAVF_MAX_QUEUES; 1762 } 1763 error = ixl_pf_qmgr_alloc_scattered(&pf->qmgr, num_queues, &vf->qtag); 1764 if (error) { 1765 device_printf(dev, "Error allocating %d queues for VF %d's VSI\n", 1766 num_queues, vf->vf_num); 1767 return (ENOSPC); 1768 } 1769 1770 ixl_dbg_iov(pf, "VF %d: %d allocated, %d active\n", 1771 vf->vf_num, vf->qtag.num_allocated, vf->qtag.num_active); 1772 ixl_dbg_iov(pf, "Unallocated total: %d\n", ixl_pf_qmgr_get_num_free(&pf->qmgr)); 1773 1774 return (0); 1775 } 1776 1777 int 1778 ixl_if_iov_vf_add(if_ctx_t ctx, uint16_t vfnum, const nvlist_t *params) 1779 { 1780 struct ixl_pf *pf = iflib_get_softc(ctx); 1781 char sysctl_name[IXL_QUEUE_NAME_LEN]; 1782 struct ixl_vf *vf; 1783 const void *mac; 1784 size_t size; 1785 int error; 1786 int vf_num_queues; 1787 1788 vf = &pf->vfs[vfnum]; 1789 vf->vf_num = vfnum; 1790 vf->vsi.back = pf; 1791 vf->vf_flags = VF_FLAG_ENABLED; 1792 1793 /* Reserve queue allocation from PF */ 1794 vf_num_queues = nvlist_get_number(params, "num-queues"); 1795 error = ixl_vf_reserve_queues(pf, vf, vf_num_queues); 1796 if (error != 0) 1797 goto out; 1798 1799 error = ixl_vf_setup_vsi(pf, vf); 1800 if (error != 0) 1801 goto out; 1802 1803 if (nvlist_exists_binary(params, "mac-addr")) { 1804 mac = nvlist_get_binary(params, "mac-addr", &size); 1805 bcopy(mac, vf->mac, ETHER_ADDR_LEN); 1806 1807 if (nvlist_get_bool(params, "allow-set-mac")) 1808 vf->vf_flags |= VF_FLAG_SET_MAC_CAP; 1809 } else 1810 /* 1811 * If the administrator has not specified a MAC address then 1812 * we must allow the VF to choose one. 1813 */ 1814 vf->vf_flags |= VF_FLAG_SET_MAC_CAP; 1815 1816 if (nvlist_get_bool(params, "mac-anti-spoof")) 1817 vf->vf_flags |= VF_FLAG_MAC_ANTI_SPOOF; 1818 1819 if (nvlist_get_bool(params, "allow-promisc")) 1820 vf->vf_flags |= VF_FLAG_PROMISC_CAP; 1821 1822 vf->vf_flags |= VF_FLAG_VLAN_CAP; 1823 1824 /* VF needs to be reset before it can be used */ 1825 ixl_reset_vf(pf, vf); 1826 out: 1827 if (error == 0) { 1828 snprintf(sysctl_name, sizeof(sysctl_name), "vf%d", vfnum); 1829 ixl_vsi_add_sysctls(&vf->vsi, sysctl_name, false); 1830 } 1831 1832 return (error); 1833 } 1834 1835