Lines Matching full:pf

37 ixl_configure_tx_itr(struct ixl_pf *pf)  in ixl_configure_tx_itr()  argument
39 struct i40e_hw *hw = &pf->hw; in ixl_configure_tx_itr()
40 struct ixl_vsi *vsi = &pf->vsi; in ixl_configure_tx_itr()
43 vsi->tx_itr_setting = pf->tx_itr; in ixl_configure_tx_itr()
56 ixl_configure_rx_itr(struct ixl_pf *pf) in ixl_configure_rx_itr() argument
58 struct i40e_hw *hw = &pf->hw; in ixl_configure_rx_itr()
59 struct ixl_vsi *vsi = &pf->vsi; in ixl_configure_rx_itr()
62 vsi->rx_itr_setting = pf->rx_itr; in ixl_configure_rx_itr()
77 struct ixl_pf *pf = arg; in ixl_intr() local
78 struct i40e_hw *hw = &pf->hw; in ixl_intr()
79 struct ixl_vsi *vsi = &pf->vsi; in ixl_intr()
135 struct ixl_pf *pf = arg; in ixl_msix_adminq() local
136 struct i40e_hw *hw = &pf->hw; in ixl_msix_adminq()
137 device_t dev = pf->dev; in ixl_msix_adminq()
143 ++pf->admin_irq; in ixl_msix_adminq()
160 ixl_set_state(&pf->state, IXL_STATE_MDD_PENDING); in ixl_msix_adminq()
187 ixl_set_state(&pf->state, IXL_STATE_RESETTING); in ixl_msix_adminq()
193 * mask out these three causes, then request a PF reset in ixl_msix_adminq()
204 ixl_set_state(&pf->state, in ixl_msix_adminq()
223 iflib_iov_intr_deferred(pf->vsi.ctx); in ixl_msix_adminq()
242 ixl_configure_queue_intr_msix(struct ixl_pf *pf) in ixl_configure_queue_intr_msix() argument
244 struct i40e_hw *hw = &pf->hw; in ixl_configure_queue_intr_msix()
245 struct ixl_vsi *vsi = &pf->vsi; in ixl_configure_queue_intr_msix()
280 ixl_configure_legacy(struct ixl_pf *pf) in ixl_configure_legacy() argument
282 struct i40e_hw *hw = &pf->hw; in ixl_configure_legacy()
283 struct ixl_vsi *vsi = &pf->vsi; in ixl_configure_legacy()
320 ixl_free_pci_resources(struct ixl_pf *pf) in ixl_free_pci_resources() argument
322 struct ixl_vsi *vsi = &pf->vsi; in ixl_free_pci_resources()
338 if (pf->pci_mem != NULL) in ixl_free_pci_resources()
340 rman_get_rid(pf->pci_mem), pf->pci_mem); in ixl_free_pci_resources()
349 ixl_setup_interface(device_t dev, struct ixl_pf *pf) in ixl_setup_interface() argument
351 struct ixl_vsi *vsi = &pf->vsi; in ixl_setup_interface()
353 struct i40e_hw *hw = &pf->hw; in ixl_setup_interface()
364 if (IXL_PF_IN_RECOVERY_MODE(pf)) in ixl_setup_interface()
383 pf->supported_speeds = abilities.link_speed; in ixl_setup_interface()
384 if_setbaudrate(ifp, ixl_max_aq_speed_to_value(pf->supported_speeds)); in ixl_setup_interface()
401 ixl_link_event(struct ixl_pf *pf, struct i40e_arq_event_info *e) in ixl_link_event() argument
403 struct i40e_hw *hw = &pf->hw; in ixl_link_event()
404 device_t dev = iflib_get_dev(pf->vsi.ctx); in ixl_link_event()
414 i40e_get_link_status(hw, &pf->link_up); in ixl_link_event()
418 (pf->advertised_speed) && in ixl_link_event()
419 (if_getflags(pf->vsi.ifp) & IFF_UP) && in ixl_link_event()
438 struct ixl_pf *pf = vsi->back; in ixl_initialize_vsi() local
450 if (pf->veb_seid != 0) in ixl_initialize_vsi()
451 ctxt.uplink_seid = pf->veb_seid; in ixl_initialize_vsi()
459 ixl_dbg(pf, IXL_DBG_SWITCH_INFO, in ixl_initialize_vsi()
481 tc_queues = fls(pf->qtag.num_allocated) - 1; in ixl_initialize_vsi()
482 ctxt.info.tc_mapping[0] = ((pf->qtag.first_qidx << I40E_AQ_VSI_TC_QUE_OFFSET_SHIFT) in ixl_initialize_vsi()
497 if (ixl_enable_iwarp && pf->iw_enabled) { in ixl_initialize_vsi()
556 /* Associate the ring with this PF */ in ixl_initialize_vsi()
623 struct ixl_pf *pf = (struct ixl_pf *)vsi->back; in ixl_set_queue_rx_itr() local
634 if (pf->dynamic_rx_itr) { in ixl_set_queue_rx_itr()
675 vsi->rx_itr_setting = pf->rx_itr; in ixl_set_queue_rx_itr()
696 struct ixl_pf *pf = (struct ixl_pf *)vsi->back; in ixl_set_queue_tx_itr() local
708 if (pf->dynamic_tx_itr) { in ixl_set_queue_tx_itr()
749 vsi->tx_itr_setting = pf->tx_itr; in ixl_set_queue_tx_itr()
809 ixl_add_hw_stats(struct ixl_pf *pf) in ixl_add_hw_stats() argument
811 struct ixl_vsi *vsi = &pf->vsi; in ixl_add_hw_stats()
813 struct i40e_hw_port_stats *pf_stats = &pf->stats; in ixl_add_hw_stats()
821 CTLFLAG_RD, &pf->admin_irq, in ixl_add_hw_stats()
825 ixl_vsi_add_sysctls(vsi, "pf", true); in ixl_add_hw_stats()
831 ixl_set_rss_hlut(struct ixl_pf *pf) in ixl_set_rss_hlut() argument
833 struct i40e_hw *hw = &pf->hw; in ixl_set_rss_hlut()
834 struct ixl_vsi *vsi = &pf->vsi; in ixl_set_rss_hlut()
841 lut_entry_width = pf->hw.func_caps.rss_table_entry_width; in ixl_set_rss_hlut()
845 for (i = 0; i < pf->hw.func_caps.rss_table_size; i++) { in ixl_set_rss_hlut()
867 for (i = 0; i < pf->hw.func_caps.rss_table_size >> 2; i++) in ixl_set_rss_hlut()
873 /* For PF VSI only */
877 struct ixl_pf *pf = vsi->back; in ixl_enable_rings() local
881 error = ixl_enable_tx_ring(pf, &pf->qtag, i); in ixl_enable_rings()
884 error = ixl_enable_rx_ring(pf, &pf->qtag, i); in ixl_enable_rings()
890 ixl_disable_rings(struct ixl_pf *pf, struct ixl_vsi *vsi, struct ixl_pf_qtag *qtag) in ixl_disable_rings() argument
895 error = ixl_disable_tx_ring(pf, qtag, i); in ixl_disable_rings()
898 error = ixl_disable_rx_ring(pf, qtag, i); in ixl_disable_rings()
927 ixl_prepare_for_reset(struct ixl_pf *pf, bool is_up) in ixl_prepare_for_reset() argument
929 struct i40e_hw *hw = &pf->hw; in ixl_prepare_for_reset()
930 device_t dev = pf->dev; in ixl_prepare_for_reset()
934 ixl_if_stop(pf->vsi.ctx); in ixl_prepare_for_reset()
936 ixl_shutdown_hmc(pf); in ixl_prepare_for_reset()
945 ixl_pf_qmgr_release(&pf->qmgr, &pf->qtag); in ixl_prepare_for_reset()
950 ixl_rebuild_hw_structs_after_reset(struct ixl_pf *pf, bool is_up) in ixl_rebuild_hw_structs_after_reset() argument
952 struct i40e_hw *hw = &pf->hw; in ixl_rebuild_hw_structs_after_reset()
953 struct ixl_vsi *vsi = &pf->vsi; in ixl_rebuild_hw_structs_after_reset()
954 device_t dev = pf->dev; in ixl_rebuild_hw_structs_after_reset()
968 if (IXL_PF_IN_RECOVERY_MODE(pf)) { in ixl_rebuild_hw_structs_after_reset()
971 ixl_configure_intr0_msix(pf); in ixl_rebuild_hw_structs_after_reset()
980 error = ixl_get_hw_capabilities(pf); in ixl_rebuild_hw_structs_after_reset()
986 error = ixl_setup_hmc(pf); in ixl_rebuild_hw_structs_after_reset()
990 /* reserve a contiguous allocation for the PF's VSI */ in ixl_rebuild_hw_structs_after_reset()
991 error = ixl_pf_qmgr_alloc_contiguous(&pf->qmgr, vsi->num_tx_queues, &pf->qtag); in ixl_rebuild_hw_structs_after_reset()
993 device_printf(dev, "Failed to reserve queues for PF LAN VSI, error %d\n", in ixl_rebuild_hw_structs_after_reset()
997 error = ixl_switch_config(pf); in ixl_rebuild_hw_structs_after_reset()
1027 i40e_aq_set_vsi_broadcast(&pf->hw, vsi->seid, TRUE, NULL); in ixl_rebuild_hw_structs_after_reset()
1030 ixl_attach_get_link_status(pf); in ixl_rebuild_hw_structs_after_reset()
1037 ixl_set_state(&pf->state, in ixl_rebuild_hw_structs_after_reset()
1040 ixl_clear_state(&pf->state, in ixl_rebuild_hw_structs_after_reset()
1047 ixl_configure_intr0_msix(pf); in ixl_rebuild_hw_structs_after_reset()
1074 struct ixl_pf *pf = (struct ixl_pf *)arg1; in ixl_sysctl_set_flowcntl() local
1075 struct i40e_hw *hw = &pf->hw; in ixl_sysctl_set_flowcntl()
1076 device_t dev = pf->dev; in ixl_sysctl_set_flowcntl()
1082 requested_fc = pf->fc; in ixl_sysctl_set_flowcntl()
1101 pf->fc = requested_fc; in ixl_sysctl_set_flowcntl()