xref: /freebsd/sys/dev/ixl/ixl_pf_iflib.c (revision 71625ec9ad2a9bc8c09784fbd23b759830e0ee5f)
1 /******************************************************************************
2 
3   Copyright (c) 2013-2020, Intel Corporation
4   All rights reserved.
5 
6   Redistribution and use in source and binary forms, with or without
7   modification, are permitted provided that the following conditions are met:
8 
9    1. Redistributions of source code must retain the above copyright notice,
10       this list of conditions and the following disclaimer.
11 
12    2. Redistributions in binary form must reproduce the above copyright
13       notice, this list of conditions and the following disclaimer in the
14       documentation and/or other materials provided with the distribution.
15 
16    3. Neither the name of the Intel Corporation nor the names of its
17       contributors may be used to endorse or promote products derived from
18       this software without specific prior written permission.
19 
20   THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
21   AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
22   IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
23   ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
24   LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
25   CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
26   SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
27   INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
28   CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
29   ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
30   POSSIBILITY OF SUCH DAMAGE.
31 
32 ******************************************************************************/
33 
34 #include "ixl_pf.h"
35 
36 void
ixl_configure_tx_itr(struct ixl_pf * pf)37 ixl_configure_tx_itr(struct ixl_pf *pf)
38 {
39 	struct i40e_hw		*hw = &pf->hw;
40 	struct ixl_vsi		*vsi = &pf->vsi;
41 	struct ixl_tx_queue	*que = vsi->tx_queues;
42 
43 	vsi->tx_itr_setting = pf->tx_itr;
44 
45 	for (int i = 0; i < vsi->num_tx_queues; i++, que++) {
46 		struct tx_ring	*txr = &que->txr;
47 
48 		wr32(hw, I40E_PFINT_ITRN(IXL_TX_ITR, i),
49 		    vsi->tx_itr_setting);
50 		txr->itr = vsi->tx_itr_setting;
51 		txr->latency = IXL_AVE_LATENCY;
52 	}
53 }
54 
55 void
ixl_configure_rx_itr(struct ixl_pf * pf)56 ixl_configure_rx_itr(struct ixl_pf *pf)
57 {
58 	struct i40e_hw		*hw = &pf->hw;
59 	struct ixl_vsi		*vsi = &pf->vsi;
60 	struct ixl_rx_queue	*que = vsi->rx_queues;
61 
62 	vsi->rx_itr_setting = pf->rx_itr;
63 
64 	for (int i = 0; i < vsi->num_rx_queues; i++, que++) {
65 		struct rx_ring 	*rxr = &que->rxr;
66 
67 		wr32(hw, I40E_PFINT_ITRN(IXL_RX_ITR, i),
68 		    vsi->rx_itr_setting);
69 		rxr->itr = vsi->rx_itr_setting;
70 		rxr->latency = IXL_AVE_LATENCY;
71 	}
72 }
73 
74 int
ixl_intr(void * arg)75 ixl_intr(void *arg)
76 {
77 	struct ixl_pf		*pf = arg;
78 	struct i40e_hw		*hw =  &pf->hw;
79 	struct ixl_vsi		*vsi = &pf->vsi;
80 	struct ixl_rx_queue	*que = vsi->rx_queues;
81         u32			icr0;
82 
83 	++que->irqs;
84 
85 	/* Clear PBA at start of ISR if using legacy interrupts */
86 	if (vsi->shared->isc_intr == IFLIB_INTR_LEGACY)
87 		wr32(hw, I40E_PFINT_DYN_CTL0,
88 		    I40E_PFINT_DYN_CTLN_CLEARPBA_MASK |
89 		    (IXL_ITR_NONE << I40E_PFINT_DYN_CTLN_ITR_INDX_SHIFT));
90 
91 	icr0 = rd32(hw, I40E_PFINT_ICR0);
92 
93 
94 #ifdef PCI_IOV
95 	if (icr0 & I40E_PFINT_ICR0_VFLR_MASK)
96 		iflib_iov_intr_deferred(vsi->ctx);
97 #endif
98 
99 	if (icr0 & I40E_PFINT_ICR0_ADMINQ_MASK)
100 		iflib_admin_intr_deferred(vsi->ctx);
101 
102 	ixl_enable_intr0(hw);
103 
104 	if (icr0 & I40E_PFINT_ICR0_QUEUE_0_MASK)
105 		return (FILTER_SCHEDULE_THREAD);
106 	else
107 		return (FILTER_HANDLED);
108 }
109 
110 /*********************************************************************
111  *
112  *  MSI-X VSI Interrupt Service routine
113  *
114  **********************************************************************/
115 int
ixl_msix_que(void * arg)116 ixl_msix_que(void *arg)
117 {
118 	struct ixl_rx_queue *rx_que = arg;
119 
120 	++rx_que->irqs;
121 
122 	ixl_set_queue_rx_itr(rx_que);
123 
124 	return (FILTER_SCHEDULE_THREAD);
125 }
126 
127 /*********************************************************************
128  *
129  *  MSI-X Admin Queue Interrupt Service routine
130  *
131  **********************************************************************/
132 int
ixl_msix_adminq(void * arg)133 ixl_msix_adminq(void *arg)
134 {
135 	struct ixl_pf	*pf = arg;
136 	struct i40e_hw	*hw = &pf->hw;
137 	device_t	dev = pf->dev;
138 	u32		reg, mask, rstat_reg;
139 	bool		do_task = FALSE;
140 
141 	DDPRINTF(dev, "begin");
142 
143 	++pf->admin_irq;
144 
145 	reg = rd32(hw, I40E_PFINT_ICR0);
146 	/*
147 	 * For masking off interrupt causes that need to be handled before
148 	 * they can be re-enabled
149 	 */
150 	mask = rd32(hw, I40E_PFINT_ICR0_ENA);
151 
152 	/* Check on the cause */
153 	if (reg & I40E_PFINT_ICR0_ADMINQ_MASK) {
154 		mask &= ~I40E_PFINT_ICR0_ENA_ADMINQ_MASK;
155 		do_task = TRUE;
156 	}
157 
158 	if (reg & I40E_PFINT_ICR0_MAL_DETECT_MASK) {
159 		mask &= ~I40E_PFINT_ICR0_ENA_MAL_DETECT_MASK;
160 		ixl_set_state(&pf->state, IXL_STATE_MDD_PENDING);
161 		do_task = TRUE;
162 	}
163 
164 	if (reg & I40E_PFINT_ICR0_GRST_MASK) {
165 		const char *reset_type;
166 		mask &= ~I40E_PFINT_ICR0_ENA_GRST_MASK;
167 		rstat_reg = rd32(hw, I40E_GLGEN_RSTAT);
168 		rstat_reg = (rstat_reg & I40E_GLGEN_RSTAT_RESET_TYPE_MASK)
169 		    >> I40E_GLGEN_RSTAT_RESET_TYPE_SHIFT;
170 		switch (rstat_reg) {
171 		/* These others might be handled similarly to an EMPR reset */
172 		case I40E_RESET_CORER:
173 			reset_type = "CORER";
174 			break;
175 		case I40E_RESET_GLOBR:
176 			reset_type = "GLOBR";
177 			break;
178 		case I40E_RESET_EMPR:
179 			reset_type = "EMPR";
180 			break;
181 		default:
182 			reset_type = "POR";
183 			break;
184 		}
185 		device_printf(dev, "Reset Requested! (%s)\n", reset_type);
186 		/* overload admin queue task to check reset progress */
187 		ixl_set_state(&pf->state, IXL_STATE_RESETTING);
188 		do_task = TRUE;
189 	}
190 
191 	/*
192 	 * PE / PCI / ECC exceptions are all handled in the same way:
193 	 * mask out these three causes, then request a PF reset
194 	 */
195 	if (reg & I40E_PFINT_ICR0_ECC_ERR_MASK)
196 		device_printf(dev, "ECC Error detected!\n");
197 	if (reg & I40E_PFINT_ICR0_PCI_EXCEPTION_MASK)
198 		device_printf(dev, "PCI Exception detected!\n");
199 	if (reg & I40E_PFINT_ICR0_PE_CRITERR_MASK)
200 		device_printf(dev, "Critical Protocol Engine Error detected!\n");
201 	/* Checks against the conditions above */
202 	if (reg & IXL_ICR0_CRIT_ERR_MASK) {
203 		mask &= ~IXL_ICR0_CRIT_ERR_MASK;
204 		ixl_set_state(&pf->state,
205 		    IXL_STATE_PF_RESET_REQ | IXL_STATE_PF_CRIT_ERR);
206 		do_task = TRUE;
207 	}
208 
209 	if (reg & I40E_PFINT_ICR0_HMC_ERR_MASK) {
210 		reg = rd32(hw, I40E_PFHMC_ERRORINFO);
211 		if (reg & I40E_PFHMC_ERRORINFO_ERROR_DETECTED_MASK) {
212 			device_printf(dev, "HMC Error detected!\n");
213 			device_printf(dev, "INFO 0x%08x\n", reg);
214 			reg = rd32(hw, I40E_PFHMC_ERRORDATA);
215 			device_printf(dev, "DATA 0x%08x\n", reg);
216 			wr32(hw, I40E_PFHMC_ERRORINFO, 0);
217 		}
218 	}
219 
220 #ifdef PCI_IOV
221 	if (reg & I40E_PFINT_ICR0_VFLR_MASK) {
222 		mask &= ~I40E_PFINT_ICR0_ENA_VFLR_MASK;
223 		iflib_iov_intr_deferred(pf->vsi.ctx);
224 	}
225 #endif
226 
227 	wr32(hw, I40E_PFINT_ICR0_ENA, mask);
228 	ixl_enable_intr0(hw);
229 
230 	if (do_task)
231 		return (FILTER_SCHEDULE_THREAD);
232 	else
233 		return (FILTER_HANDLED);
234 }
235 
236 /*
237  * Configure queue interrupt cause registers in hardware.
238  *
239  * Linked list for each vector LNKLSTN(i) -> RQCTL(i) -> TQCTL(i) -> EOL
240  */
241 void
ixl_configure_queue_intr_msix(struct ixl_pf * pf)242 ixl_configure_queue_intr_msix(struct ixl_pf *pf)
243 {
244 	struct i40e_hw *hw = &pf->hw;
245 	struct ixl_vsi *vsi = &pf->vsi;
246 	u32		reg;
247 	u16		vector = 1;
248 
249 	for (int i = 0; i < max(vsi->num_rx_queues, vsi->num_tx_queues); i++, vector++) {
250 		/* Make sure interrupt is disabled */
251 		wr32(hw, I40E_PFINT_DYN_CTLN(i), 0);
252 		/* Set linked list head to point to corresponding RX queue
253 		 * e.g. vector 1 (LNKLSTN register 0) points to queue pair 0's RX queue */
254 		reg = ((i << I40E_PFINT_LNKLSTN_FIRSTQ_INDX_SHIFT)
255 		        & I40E_PFINT_LNKLSTN_FIRSTQ_INDX_MASK) |
256 		    ((I40E_QUEUE_TYPE_RX << I40E_PFINT_LNKLSTN_FIRSTQ_TYPE_SHIFT)
257 		        & I40E_PFINT_LNKLSTN_FIRSTQ_TYPE_MASK);
258 		wr32(hw, I40E_PFINT_LNKLSTN(i), reg);
259 
260 		reg = I40E_QINT_RQCTL_CAUSE_ENA_MASK |
261 		(IXL_RX_ITR << I40E_QINT_RQCTL_ITR_INDX_SHIFT) |
262 		(vector << I40E_QINT_RQCTL_MSIX_INDX_SHIFT) |
263 		(i << I40E_QINT_RQCTL_NEXTQ_INDX_SHIFT) |
264 		(I40E_QUEUE_TYPE_TX << I40E_QINT_RQCTL_NEXTQ_TYPE_SHIFT);
265 		wr32(hw, I40E_QINT_RQCTL(i), reg);
266 
267 		reg = I40E_QINT_TQCTL_CAUSE_ENA_MASK |
268 		(IXL_TX_ITR << I40E_QINT_TQCTL_ITR_INDX_SHIFT) |
269 		(vector << I40E_QINT_TQCTL_MSIX_INDX_SHIFT) |
270 		(IXL_QUEUE_EOL << I40E_QINT_TQCTL_NEXTQ_INDX_SHIFT) |
271 		(I40E_QUEUE_TYPE_RX << I40E_QINT_TQCTL_NEXTQ_TYPE_SHIFT);
272 		wr32(hw, I40E_QINT_TQCTL(i), reg);
273 	}
274 }
275 
276 /*
277  * Configure for single interrupt vector operation
278  */
279 void
ixl_configure_legacy(struct ixl_pf * pf)280 ixl_configure_legacy(struct ixl_pf *pf)
281 {
282 	struct i40e_hw	*hw = &pf->hw;
283 	struct ixl_vsi	*vsi = &pf->vsi;
284 	u32 reg;
285 
286 	vsi->rx_queues[0].rxr.itr = vsi->rx_itr_setting;
287 
288 	/* Setup "other" causes */
289 	reg = I40E_PFINT_ICR0_ENA_ECC_ERR_MASK
290 	    | I40E_PFINT_ICR0_ENA_MAL_DETECT_MASK
291 	    | I40E_PFINT_ICR0_ENA_GRST_MASK
292 	    | I40E_PFINT_ICR0_ENA_PCI_EXCEPTION_MASK
293 	    | I40E_PFINT_ICR0_ENA_HMC_ERR_MASK
294 	    | I40E_PFINT_ICR0_ENA_PE_CRITERR_MASK
295 	    | I40E_PFINT_ICR0_ENA_VFLR_MASK
296 	    | I40E_PFINT_ICR0_ENA_ADMINQ_MASK
297 	    ;
298 	wr32(hw, I40E_PFINT_ICR0_ENA, reg);
299 
300 	/* No ITR for non-queue interrupts */
301 	wr32(hw, I40E_PFINT_STAT_CTL0,
302 	    IXL_ITR_NONE << I40E_PFINT_STAT_CTL0_OTHER_ITR_INDX_SHIFT);
303 
304 	/* FIRSTQ_INDX = 0, FIRSTQ_TYPE = 0 (rx) */
305 	wr32(hw, I40E_PFINT_LNKLST0, 0);
306 
307 	/* Associate the queue pair to the vector and enable the q int */
308 	reg = I40E_QINT_RQCTL_CAUSE_ENA_MASK
309 	    | (IXL_RX_ITR << I40E_QINT_RQCTL_ITR_INDX_SHIFT)
310 	    | (I40E_QUEUE_TYPE_TX << I40E_QINT_RQCTL_NEXTQ_TYPE_SHIFT);
311 	wr32(hw, I40E_QINT_RQCTL(0), reg);
312 
313 	reg = I40E_QINT_TQCTL_CAUSE_ENA_MASK
314 	    | (IXL_TX_ITR << I40E_QINT_TQCTL_ITR_INDX_SHIFT)
315 	    | (IXL_QUEUE_EOL << I40E_QINT_TQCTL_NEXTQ_INDX_SHIFT);
316 	wr32(hw, I40E_QINT_TQCTL(0), reg);
317 }
318 
319 void
ixl_free_pci_resources(struct ixl_pf * pf)320 ixl_free_pci_resources(struct ixl_pf *pf)
321 {
322 	struct ixl_vsi		*vsi = &pf->vsi;
323 	device_t		dev = iflib_get_dev(vsi->ctx);
324 	struct ixl_rx_queue	*rx_que = vsi->rx_queues;
325 
326 	/* We may get here before stations are set up */
327 	if (rx_que == NULL)
328 		goto early;
329 
330 	/*
331 	**  Release all MSI-X VSI resources:
332 	*/
333 	iflib_irq_free(vsi->ctx, &vsi->irq);
334 
335 	for (int i = 0; i < vsi->num_rx_queues; i++, rx_que++)
336 		iflib_irq_free(vsi->ctx, &rx_que->que_irq);
337 early:
338 	if (pf->pci_mem != NULL)
339 		bus_release_resource(dev, SYS_RES_MEMORY,
340 		    rman_get_rid(pf->pci_mem), pf->pci_mem);
341 }
342 
343 /*********************************************************************
344  *
345  *  Setup networking device structure and register an interface.
346  *
347  **********************************************************************/
348 int
ixl_setup_interface(device_t dev,struct ixl_pf * pf)349 ixl_setup_interface(device_t dev, struct ixl_pf *pf)
350 {
351 	struct ixl_vsi *vsi = &pf->vsi;
352 	if_ctx_t ctx = vsi->ctx;
353 	struct i40e_hw *hw = &pf->hw;
354 	if_t ifp = iflib_get_ifp(ctx);
355 	struct i40e_aq_get_phy_abilities_resp abilities;
356 	enum i40e_status_code aq_error = 0;
357 
358 	INIT_DBG_DEV(dev, "begin");
359 
360 	vsi->shared->isc_max_frame_size =
361 	    if_getmtu(ifp) + ETHER_HDR_LEN + ETHER_CRC_LEN
362 	    + ETHER_VLAN_ENCAP_LEN;
363 
364 	if (IXL_PF_IN_RECOVERY_MODE(pf))
365 		goto only_auto;
366 
367 	aq_error = i40e_aq_get_phy_capabilities(hw,
368 	    FALSE, TRUE, &abilities, NULL);
369 	/* May need delay to detect fiber correctly */
370 	if (aq_error == I40E_ERR_UNKNOWN_PHY) {
371 		i40e_msec_delay(200);
372 		aq_error = i40e_aq_get_phy_capabilities(hw, FALSE,
373 		    TRUE, &abilities, NULL);
374 	}
375 	if (aq_error) {
376 		if (aq_error == I40E_ERR_UNKNOWN_PHY)
377 			device_printf(dev, "Unknown PHY type detected!\n");
378 		else
379 			device_printf(dev,
380 			    "Error getting supported media types, err %d,"
381 			    " AQ error %d\n", aq_error, hw->aq.asq_last_status);
382 	} else {
383 		pf->supported_speeds = abilities.link_speed;
384 		if_setbaudrate(ifp, ixl_max_aq_speed_to_value(pf->supported_speeds));
385 
386 		ixl_add_ifmedia(vsi->media, hw->phy.phy_types);
387 	}
388 
389 only_auto:
390 	/* Use autoselect media by default */
391 	ifmedia_add(vsi->media, IFM_ETHER | IFM_AUTO, 0, NULL);
392 	ifmedia_set(vsi->media, IFM_ETHER | IFM_AUTO);
393 
394 	return (0);
395 }
396 
397 /*
398 ** Run when the Admin Queue gets a link state change interrupt.
399 */
400 void
ixl_link_event(struct ixl_pf * pf,struct i40e_arq_event_info * e)401 ixl_link_event(struct ixl_pf *pf, struct i40e_arq_event_info *e)
402 {
403 	struct i40e_hw *hw = &pf->hw;
404 	device_t dev = iflib_get_dev(pf->vsi.ctx);
405 	struct i40e_link_status *link_info = &hw->phy.link_info;
406 
407 	/* Driver needs to re-enable delivering of link status events
408 	 * by FW after each event reception. Call i40e_get_link_status
409 	 * to do that. To not lose information about link state changes,
410 	 * which happened between receiving an event and the call,
411 	 * do not rely on status from event but use most recent
412 	 * status information retrieved by the call. */
413 	hw->phy.get_link_info = TRUE;
414 	i40e_get_link_status(hw, &pf->link_up);
415 
416 	/* Print out message if an unqualified module is found */
417 	if ((link_info->link_info & I40E_AQ_MEDIA_AVAILABLE) &&
418 	    (pf->advertised_speed) &&
419 	    (if_getflags(pf->vsi.ifp) & IFF_UP) &&
420 	    (!(link_info->an_info & I40E_AQ_QUALIFIED_MODULE)) &&
421 	    (!(link_info->link_info & I40E_AQ_LINK_UP)))
422 		device_printf(dev, "Link failed because "
423 		    "an unqualified module was detected!\n");
424 
425 	/* OS link info is updated elsewhere */
426 }
427 
428 /*********************************************************************
429  *
430  *  Initialize the VSI:  this handles contexts, which means things
431  *  			 like the number of descriptors, buffer size,
432  *			 plus we init the rings thru this function.
433  *
434  **********************************************************************/
435 int
ixl_initialize_vsi(struct ixl_vsi * vsi)436 ixl_initialize_vsi(struct ixl_vsi *vsi)
437 {
438 	struct ixl_pf *pf = vsi->back;
439 	if_softc_ctx_t		scctx = iflib_get_softc_ctx(vsi->ctx);
440 	struct ixl_tx_queue	*tx_que = vsi->tx_queues;
441 	struct ixl_rx_queue	*rx_que = vsi->rx_queues;
442 	device_t		dev = iflib_get_dev(vsi->ctx);
443 	struct i40e_hw		*hw = vsi->hw;
444 	struct i40e_vsi_context	ctxt;
445 	int 			tc_queues;
446 	int			err = 0;
447 
448 	memset(&ctxt, 0, sizeof(ctxt));
449 	ctxt.seid = vsi->seid;
450 	if (pf->veb_seid != 0)
451 		ctxt.uplink_seid = pf->veb_seid;
452 	ctxt.pf_num = hw->pf_id;
453 	err = i40e_aq_get_vsi_params(hw, &ctxt, NULL);
454 	if (err) {
455 		device_printf(dev, "i40e_aq_get_vsi_params() failed, error %d"
456 		    " aq_error %d\n", err, hw->aq.asq_last_status);
457 		return (err);
458 	}
459 	ixl_dbg(pf, IXL_DBG_SWITCH_INFO,
460 	    "get_vsi_params: seid: %d, uplinkseid: %d, vsi_number: %d, "
461 	    "vsis_allocated: %d, vsis_unallocated: %d, flags: 0x%x, "
462 	    "pfnum: %d, vfnum: %d, stat idx: %d, enabled: %d\n", ctxt.seid,
463 	    ctxt.uplink_seid, ctxt.vsi_number,
464 	    ctxt.vsis_allocated, ctxt.vsis_unallocated,
465 	    ctxt.flags, ctxt.pf_num, ctxt.vf_num,
466 	    ctxt.info.stat_counter_idx, ctxt.info.up_enable_bits);
467 	/*
468 	** Set the queue and traffic class bits
469 	**  - when multiple traffic classes are supported
470 	**    this will need to be more robust.
471 	*/
472 	ctxt.info.valid_sections = I40E_AQ_VSI_PROP_QUEUE_MAP_VALID;
473 	ctxt.info.mapping_flags |= I40E_AQ_VSI_QUE_MAP_CONTIG;
474 	/* In contig mode, que_mapping[0] is first queue index used by this VSI */
475 	ctxt.info.queue_mapping[0] = 0;
476 	/*
477 	 * This VSI will only use traffic class 0; start traffic class 0's
478 	 * queue allocation at queue 0, and assign it 2^tc_queues queues (though
479 	 * the driver may not use all of them).
480 	 */
481 	tc_queues = fls(pf->qtag.num_allocated) - 1;
482 	ctxt.info.tc_mapping[0] = ((pf->qtag.first_qidx << I40E_AQ_VSI_TC_QUE_OFFSET_SHIFT)
483 	    & I40E_AQ_VSI_TC_QUE_OFFSET_MASK) |
484 	    ((tc_queues << I40E_AQ_VSI_TC_QUE_NUMBER_SHIFT)
485 	    & I40E_AQ_VSI_TC_QUE_NUMBER_MASK);
486 
487 	/* Set VLAN receive stripping mode */
488 	ctxt.info.valid_sections |= I40E_AQ_VSI_PROP_VLAN_VALID;
489 	ctxt.info.port_vlan_flags = I40E_AQ_VSI_PVLAN_MODE_ALL;
490 	if (if_getcapenable(vsi->ifp) & IFCAP_VLAN_HWTAGGING)
491 		ctxt.info.port_vlan_flags |= I40E_AQ_VSI_PVLAN_EMOD_STR_BOTH;
492 	else
493 		ctxt.info.port_vlan_flags |= I40E_AQ_VSI_PVLAN_EMOD_NOTHING;
494 
495 #ifdef IXL_IW
496 	/* Set TCP Enable for iWARP capable VSI */
497 	if (ixl_enable_iwarp && pf->iw_enabled) {
498 		ctxt.info.valid_sections |=
499 		    htole16(I40E_AQ_VSI_PROP_QUEUE_OPT_VALID);
500 		ctxt.info.queueing_opt_flags |= I40E_AQ_VSI_QUE_OPT_TCP_ENA;
501 	}
502 #endif
503 	/* Save VSI number and info for use later */
504 	vsi->vsi_num = ctxt.vsi_number;
505 	bcopy(&ctxt.info, &vsi->info, sizeof(vsi->info));
506 
507 	ctxt.flags = htole16(I40E_AQ_VSI_TYPE_PF);
508 
509 	err = i40e_aq_update_vsi_params(hw, &ctxt, NULL);
510 	if (err) {
511 		device_printf(dev, "i40e_aq_update_vsi_params() failed, error %d,"
512 		    " aq_error %d\n", err, hw->aq.asq_last_status);
513 		return (err);
514 	}
515 
516 	for (int i = 0; i < vsi->num_tx_queues; i++, tx_que++) {
517 		struct tx_ring		*txr = &tx_que->txr;
518 		struct i40e_hmc_obj_txq tctx;
519 		u32			txctl;
520 
521 		/* Setup the HMC TX Context  */
522 		bzero(&tctx, sizeof(tctx));
523 		tctx.new_context = 1;
524 		tctx.base = (txr->tx_paddr/IXL_TX_CTX_BASE_UNITS);
525 		tctx.qlen = scctx->isc_ntxd[0];
526 		tctx.fc_ena = 0;	/* Disable FCoE */
527 		/*
528 		 * This value needs to pulled from the VSI that this queue
529 		 * is assigned to. Index into array is traffic class.
530 		 */
531 		tctx.rdylist = vsi->info.qs_handle[0];
532 		/*
533 		 * Set these to enable Head Writeback
534 		 * - Address is last entry in TX ring (reserved for HWB index)
535 		 * Leave these as 0 for Descriptor Writeback
536 		 */
537 		if (vsi->enable_head_writeback) {
538 			tctx.head_wb_ena = 1;
539 			tctx.head_wb_addr = txr->tx_paddr +
540 			    (scctx->isc_ntxd[0] * sizeof(struct i40e_tx_desc));
541 		} else {
542 			tctx.head_wb_ena = 0;
543 			tctx.head_wb_addr = 0;
544 		}
545 		tctx.rdylist_act = 0;
546 		err = i40e_clear_lan_tx_queue_context(hw, i);
547 		if (err) {
548 			device_printf(dev, "Unable to clear TX context\n");
549 			break;
550 		}
551 		err = i40e_set_lan_tx_queue_context(hw, i, &tctx);
552 		if (err) {
553 			device_printf(dev, "Unable to set TX context\n");
554 			break;
555 		}
556 		/* Associate the ring with this PF */
557 		txctl = I40E_QTX_CTL_PF_QUEUE;
558 		txctl |= ((hw->pf_id << I40E_QTX_CTL_PF_INDX_SHIFT) &
559 		    I40E_QTX_CTL_PF_INDX_MASK);
560 		wr32(hw, I40E_QTX_CTL(i), txctl);
561 		ixl_flush(hw);
562 
563 		/* Do ring (re)init */
564 		ixl_init_tx_ring(vsi, tx_que);
565 	}
566 	for (int i = 0; i < vsi->num_rx_queues; i++, rx_que++) {
567 		struct rx_ring 		*rxr = &rx_que->rxr;
568 		struct i40e_hmc_obj_rxq rctx;
569 
570 		/* Next setup the HMC RX Context  */
571 		rxr->mbuf_sz = iflib_get_rx_mbuf_sz(vsi->ctx);
572 
573 		u16 max_rxmax = rxr->mbuf_sz * hw->func_caps.rx_buf_chain_len;
574 
575 		/* Set up an RX context for the HMC */
576 		memset(&rctx, 0, sizeof(struct i40e_hmc_obj_rxq));
577 		rctx.dbuff = rxr->mbuf_sz >> I40E_RXQ_CTX_DBUFF_SHIFT;
578 		/* ignore header split for now */
579 		rctx.hbuff = 0 >> I40E_RXQ_CTX_HBUFF_SHIFT;
580 		rctx.rxmax = (scctx->isc_max_frame_size < max_rxmax) ?
581 		    scctx->isc_max_frame_size : max_rxmax;
582 		rctx.dtype = 0;
583 		rctx.dsize = 1;		/* do 32byte descriptors */
584 		rctx.hsplit_0 = 0;	/* no header split */
585 		rctx.base = (rxr->rx_paddr/IXL_RX_CTX_BASE_UNITS);
586 		rctx.qlen = scctx->isc_nrxd[0];
587 		rctx.tphrdesc_ena = 1;
588 		rctx.tphwdesc_ena = 1;
589 		rctx.tphdata_ena = 0;	/* Header Split related */
590 		rctx.tphhead_ena = 0;	/* Header Split related */
591 		rctx.lrxqthresh = 1;	/* Interrupt at <64 desc avail */
592 		rctx.crcstrip = 1;
593 		rctx.l2tsel = 1;
594 		rctx.showiv = 1;	/* Strip inner VLAN header */
595 		rctx.fc_ena = 0;	/* Disable FCoE */
596 		rctx.prefena = 1;	/* Prefetch descriptors */
597 
598 		err = i40e_clear_lan_rx_queue_context(hw, i);
599 		if (err) {
600 			device_printf(dev,
601 			    "Unable to clear RX context %d\n", i);
602 			break;
603 		}
604 		err = i40e_set_lan_rx_queue_context(hw, i, &rctx);
605 		if (err) {
606 			device_printf(dev, "Unable to set RX context %d\n", i);
607 			break;
608 		}
609 		wr32(vsi->hw, I40E_QRX_TAIL(i), 0);
610 	}
611 	return (err);
612 }
613 
614 
615 /*
616 ** Provide a update to the queue RX
617 ** interrupt moderation value.
618 */
619 void
ixl_set_queue_rx_itr(struct ixl_rx_queue * que)620 ixl_set_queue_rx_itr(struct ixl_rx_queue *que)
621 {
622 	struct ixl_vsi	*vsi = que->vsi;
623 	struct ixl_pf	*pf = (struct ixl_pf *)vsi->back;
624 	struct i40e_hw	*hw = vsi->hw;
625 	struct rx_ring	*rxr = &que->rxr;
626 	u16		rx_itr;
627 	u16		rx_latency = 0;
628 	int		rx_bytes;
629 
630 	/* Idle, do nothing */
631 	if (rxr->bytes == 0)
632 		return;
633 
634 	if (pf->dynamic_rx_itr) {
635 		rx_bytes = rxr->bytes/rxr->itr;
636 		rx_itr = rxr->itr;
637 
638 		/* Adjust latency range */
639 		switch (rxr->latency) {
640 		case IXL_LOW_LATENCY:
641 			if (rx_bytes > 10) {
642 				rx_latency = IXL_AVE_LATENCY;
643 				rx_itr = IXL_ITR_20K;
644 			}
645 			break;
646 		case IXL_AVE_LATENCY:
647 			if (rx_bytes > 20) {
648 				rx_latency = IXL_BULK_LATENCY;
649 				rx_itr = IXL_ITR_8K;
650 			} else if (rx_bytes <= 10) {
651 				rx_latency = IXL_LOW_LATENCY;
652 				rx_itr = IXL_ITR_100K;
653 			}
654 			break;
655 		case IXL_BULK_LATENCY:
656 			if (rx_bytes <= 20) {
657 				rx_latency = IXL_AVE_LATENCY;
658 				rx_itr = IXL_ITR_20K;
659 			}
660 			break;
661 		}
662 
663 		rxr->latency = rx_latency;
664 
665 		if (rx_itr != rxr->itr) {
666 			/* do an exponential smoothing */
667 			rx_itr = (10 * rx_itr * rxr->itr) /
668 			    ((9 * rx_itr) + rxr->itr);
669 			rxr->itr = min(rx_itr, IXL_MAX_ITR);
670 			wr32(hw, I40E_PFINT_ITRN(IXL_RX_ITR,
671 			    rxr->me), rxr->itr);
672 		}
673 	} else { /* We may have have toggled to non-dynamic */
674 		if (vsi->rx_itr_setting & IXL_ITR_DYNAMIC)
675 			vsi->rx_itr_setting = pf->rx_itr;
676 		/* Update the hardware if needed */
677 		if (rxr->itr != vsi->rx_itr_setting) {
678 			rxr->itr = vsi->rx_itr_setting;
679 			wr32(hw, I40E_PFINT_ITRN(IXL_RX_ITR,
680 			    rxr->me), rxr->itr);
681 		}
682 	}
683 	rxr->bytes = 0;
684 	rxr->packets = 0;
685 }
686 
687 
688 /*
689 ** Provide a update to the queue TX
690 ** interrupt moderation value.
691 */
692 void
ixl_set_queue_tx_itr(struct ixl_tx_queue * que)693 ixl_set_queue_tx_itr(struct ixl_tx_queue *que)
694 {
695 	struct ixl_vsi	*vsi = que->vsi;
696 	struct ixl_pf	*pf = (struct ixl_pf *)vsi->back;
697 	struct i40e_hw	*hw = vsi->hw;
698 	struct tx_ring	*txr = &que->txr;
699 	u16		tx_itr;
700 	u16		tx_latency = 0;
701 	int		tx_bytes;
702 
703 
704 	/* Idle, do nothing */
705 	if (txr->bytes == 0)
706 		return;
707 
708 	if (pf->dynamic_tx_itr) {
709 		tx_bytes = txr->bytes/txr->itr;
710 		tx_itr = txr->itr;
711 
712 		switch (txr->latency) {
713 		case IXL_LOW_LATENCY:
714 			if (tx_bytes > 10) {
715 				tx_latency = IXL_AVE_LATENCY;
716 				tx_itr = IXL_ITR_20K;
717 			}
718 			break;
719 		case IXL_AVE_LATENCY:
720 			if (tx_bytes > 20) {
721 				tx_latency = IXL_BULK_LATENCY;
722 				tx_itr = IXL_ITR_8K;
723 			} else if (tx_bytes <= 10) {
724 				tx_latency = IXL_LOW_LATENCY;
725 				tx_itr = IXL_ITR_100K;
726 			}
727 			break;
728 		case IXL_BULK_LATENCY:
729 			if (tx_bytes <= 20) {
730 				tx_latency = IXL_AVE_LATENCY;
731 				tx_itr = IXL_ITR_20K;
732 			}
733 			break;
734 		}
735 
736 		txr->latency = tx_latency;
737 
738 		if (tx_itr != txr->itr) {
739 			/* do an exponential smoothing */
740 			tx_itr = (10 * tx_itr * txr->itr) /
741 			    ((9 * tx_itr) + txr->itr);
742 			txr->itr = min(tx_itr, IXL_MAX_ITR);
743 			wr32(hw, I40E_PFINT_ITRN(IXL_TX_ITR,
744 			    txr->me), txr->itr);
745 		}
746 
747 	} else { /* We may have have toggled to non-dynamic */
748 		if (vsi->tx_itr_setting & IXL_ITR_DYNAMIC)
749 			vsi->tx_itr_setting = pf->tx_itr;
750 		/* Update the hardware if needed */
751 		if (txr->itr != vsi->tx_itr_setting) {
752 			txr->itr = vsi->tx_itr_setting;
753 			wr32(hw, I40E_PFINT_ITRN(IXL_TX_ITR,
754 			    txr->me), txr->itr);
755 		}
756 	}
757 	txr->bytes = 0;
758 	txr->packets = 0;
759 	return;
760 }
761 
762 #ifdef IXL_DEBUG
763 /**
764  * ixl_sysctl_qtx_tail_handler
765  * Retrieves I40E_QTX_TAIL value from hardware
766  * for a sysctl.
767  */
768 int
ixl_sysctl_qtx_tail_handler(SYSCTL_HANDLER_ARGS)769 ixl_sysctl_qtx_tail_handler(SYSCTL_HANDLER_ARGS)
770 {
771 	struct ixl_tx_queue *tx_que;
772 	int error;
773 	u32 val;
774 
775 	tx_que = ((struct ixl_tx_queue *)oidp->oid_arg1);
776 	if (!tx_que) return 0;
777 
778 	val = rd32(tx_que->vsi->hw, tx_que->txr.tail);
779 	error = sysctl_handle_int(oidp, &val, 0, req);
780 	if (error || !req->newptr)
781 		return error;
782 	return (0);
783 }
784 
785 /**
786  * ixl_sysctl_qrx_tail_handler
787  * Retrieves I40E_QRX_TAIL value from hardware
788  * for a sysctl.
789  */
790 int
ixl_sysctl_qrx_tail_handler(SYSCTL_HANDLER_ARGS)791 ixl_sysctl_qrx_tail_handler(SYSCTL_HANDLER_ARGS)
792 {
793 	struct ixl_rx_queue *rx_que;
794 	int error;
795 	u32 val;
796 
797 	rx_que = ((struct ixl_rx_queue *)oidp->oid_arg1);
798 	if (!rx_que) return 0;
799 
800 	val = rd32(rx_que->vsi->hw, rx_que->rxr.tail);
801 	error = sysctl_handle_int(oidp, &val, 0, req);
802 	if (error || !req->newptr)
803 		return error;
804 	return (0);
805 }
806 #endif
807 
808 void
ixl_add_hw_stats(struct ixl_pf * pf)809 ixl_add_hw_stats(struct ixl_pf *pf)
810 {
811 	struct ixl_vsi *vsi = &pf->vsi;
812 	device_t dev = iflib_get_dev(vsi->ctx);
813 	struct i40e_hw_port_stats *pf_stats = &pf->stats;
814 
815 	struct sysctl_ctx_list *ctx = device_get_sysctl_ctx(dev);
816 	struct sysctl_oid *tree = device_get_sysctl_tree(dev);
817 	struct sysctl_oid_list *child = SYSCTL_CHILDREN(tree);
818 
819 	/* Driver statistics */
820 	SYSCTL_ADD_UQUAD(ctx, child, OID_AUTO, "admin_irq",
821 			CTLFLAG_RD, &pf->admin_irq,
822 			"Admin Queue IRQs received");
823 
824 	sysctl_ctx_init(&vsi->sysctl_ctx);
825 	ixl_vsi_add_sysctls(vsi, "pf", true);
826 
827 	ixl_add_sysctls_mac_stats(ctx, child, pf_stats);
828 }
829 
830 void
ixl_set_rss_hlut(struct ixl_pf * pf)831 ixl_set_rss_hlut(struct ixl_pf *pf)
832 {
833 	struct i40e_hw	*hw = &pf->hw;
834 	struct ixl_vsi *vsi = &pf->vsi;
835 	device_t	dev = iflib_get_dev(vsi->ctx);
836 	int		i, que_id;
837 	int		lut_entry_width;
838 	u32		lut = 0;
839 	enum i40e_status_code status;
840 
841 	lut_entry_width = pf->hw.func_caps.rss_table_entry_width;
842 
843 	/* Populate the LUT with max no. of queues in round robin fashion */
844 	u8 hlut_buf[512];
845 	for (i = 0; i < pf->hw.func_caps.rss_table_size; i++) {
846 #ifdef RSS
847 		/*
848 		 * Fetch the RSS bucket id for the given indirection entry.
849 		 * Cap it at the number of configured buckets (which is
850 		 * num_queues.)
851 		 */
852 		que_id = rss_get_indirection_to_bucket(i);
853 		que_id = que_id % vsi->num_rx_queues;
854 #else
855 		que_id = i % vsi->num_rx_queues;
856 #endif
857 		lut = (que_id & ((0x1 << lut_entry_width) - 1));
858 		hlut_buf[i] = lut;
859 	}
860 
861 	if (hw->mac.type == I40E_MAC_X722) {
862 		status = i40e_aq_set_rss_lut(hw, vsi->vsi_num, TRUE, hlut_buf, sizeof(hlut_buf));
863 		if (status)
864 			device_printf(dev, "i40e_aq_set_rss_lut status %s, error %s\n",
865 			    i40e_stat_str(hw, status), i40e_aq_str(hw, hw->aq.asq_last_status));
866 	} else {
867 		for (i = 0; i < pf->hw.func_caps.rss_table_size >> 2; i++)
868 			wr32(hw, I40E_PFQF_HLUT(i), ((u32 *)hlut_buf)[i]);
869 		ixl_flush(hw);
870 	}
871 }
872 
873 /* For PF VSI only */
874 int
ixl_enable_rings(struct ixl_vsi * vsi)875 ixl_enable_rings(struct ixl_vsi *vsi)
876 {
877 	struct ixl_pf	*pf = vsi->back;
878 	int		error = 0;
879 
880 	for (int i = 0; i < vsi->num_tx_queues; i++)
881 		error = ixl_enable_tx_ring(pf, &pf->qtag, i);
882 
883 	for (int i = 0; i < vsi->num_rx_queues; i++)
884 		error = ixl_enable_rx_ring(pf, &pf->qtag, i);
885 
886 	return (error);
887 }
888 
889 int
ixl_disable_rings(struct ixl_pf * pf,struct ixl_vsi * vsi,struct ixl_pf_qtag * qtag)890 ixl_disable_rings(struct ixl_pf *pf, struct ixl_vsi *vsi, struct ixl_pf_qtag *qtag)
891 {
892 	int error = 0;
893 
894 	for (int i = 0; i < vsi->num_tx_queues; i++)
895 		error = ixl_disable_tx_ring(pf, qtag, i);
896 
897 	for (int i = 0; i < vsi->num_rx_queues; i++)
898 		error = ixl_disable_rx_ring(pf, qtag, i);
899 
900 	return (error);
901 }
902 
903 void
ixl_enable_intr(struct ixl_vsi * vsi)904 ixl_enable_intr(struct ixl_vsi *vsi)
905 {
906 	struct i40e_hw		*hw = vsi->hw;
907 	struct ixl_rx_queue	*que = vsi->rx_queues;
908 
909 	if (vsi->shared->isc_intr == IFLIB_INTR_MSIX) {
910 		for (int i = 0; i < vsi->num_rx_queues; i++, que++)
911 			ixl_enable_queue(hw, que->rxr.me);
912 	} else
913 		ixl_enable_intr0(hw);
914 }
915 
916 void
ixl_disable_rings_intr(struct ixl_vsi * vsi)917 ixl_disable_rings_intr(struct ixl_vsi *vsi)
918 {
919 	struct i40e_hw		*hw = vsi->hw;
920 	struct ixl_rx_queue	*que = vsi->rx_queues;
921 
922 	for (int i = 0; i < vsi->num_rx_queues; i++, que++)
923 		ixl_disable_queue(hw, que->rxr.me);
924 }
925 
926 int
ixl_prepare_for_reset(struct ixl_pf * pf,bool is_up)927 ixl_prepare_for_reset(struct ixl_pf *pf, bool is_up)
928 {
929 	struct i40e_hw *hw = &pf->hw;
930 	device_t dev = pf->dev;
931 	int error = 0;
932 
933 	if (is_up)
934 		ixl_if_stop(pf->vsi.ctx);
935 
936 	ixl_shutdown_hmc(pf);
937 
938 	ixl_disable_intr0(hw);
939 
940 	error = i40e_shutdown_adminq(hw);
941 	if (error)
942 		device_printf(dev,
943 		    "Shutdown Admin queue failed with code %d\n", error);
944 
945 	ixl_pf_qmgr_release(&pf->qmgr, &pf->qtag);
946 	return (error);
947 }
948 
949 int
ixl_rebuild_hw_structs_after_reset(struct ixl_pf * pf,bool is_up)950 ixl_rebuild_hw_structs_after_reset(struct ixl_pf *pf, bool is_up)
951 {
952 	struct i40e_hw *hw = &pf->hw;
953 	struct ixl_vsi *vsi = &pf->vsi;
954 	device_t dev = pf->dev;
955 	enum i40e_get_fw_lldp_status_resp lldp_status;
956 	int error = 0;
957 
958 	device_printf(dev, "Rebuilding driver state...\n");
959 
960 	/* Setup */
961 	error = i40e_init_adminq(hw);
962 	if (error != 0 && error != I40E_ERR_FIRMWARE_API_VERSION) {
963 		device_printf(dev, "Unable to initialize Admin Queue, error %d\n",
964 		    error);
965 		goto ixl_rebuild_hw_structs_after_reset_err;
966 	}
967 
968 	if (IXL_PF_IN_RECOVERY_MODE(pf)) {
969 		/* Keep admin queue interrupts active while driver is loaded */
970 		if (vsi->shared->isc_intr == IFLIB_INTR_MSIX) {
971 			ixl_configure_intr0_msix(pf);
972 			ixl_enable_intr0(hw);
973 		}
974 
975 		return (0);
976 	}
977 
978 	i40e_clear_pxe_mode(hw);
979 
980 	error = ixl_get_hw_capabilities(pf);
981 	if (error) {
982 		device_printf(dev, "ixl_get_hw_capabilities failed: %d\n", error);
983 		goto ixl_rebuild_hw_structs_after_reset_err;
984 	}
985 
986 	error = ixl_setup_hmc(pf);
987 	if (error)
988 		goto ixl_rebuild_hw_structs_after_reset_err;
989 
990 	/* reserve a contiguous allocation for the PF's VSI */
991 	error = ixl_pf_qmgr_alloc_contiguous(&pf->qmgr, vsi->num_tx_queues, &pf->qtag);
992 	if (error) {
993 		device_printf(dev, "Failed to reserve queues for PF LAN VSI, error %d\n",
994 		    error);
995 	}
996 
997 	error = ixl_switch_config(pf);
998 	if (error) {
999 		device_printf(dev, "ixl_rebuild_hw_structs_after_reset: ixl_switch_config() failed: %d\n",
1000 		     error);
1001 		error = EIO;
1002 		goto ixl_rebuild_hw_structs_after_reset_err;
1003 	}
1004 
1005 	error = i40e_aq_set_phy_int_mask(hw, IXL_DEFAULT_PHY_INT_MASK,
1006 	    NULL);
1007 	if (error) {
1008 		device_printf(dev, "init: i40e_aq_set_phy_mask() failed: err %d,"
1009 		    " aq_err %d\n", error, hw->aq.asq_last_status);
1010 		error = EIO;
1011 		goto ixl_rebuild_hw_structs_after_reset_err;
1012 	}
1013 
1014 	u8 set_fc_err_mask;
1015 	error = i40e_set_fc(hw, &set_fc_err_mask, true);
1016 	if (error) {
1017 		device_printf(dev, "init: setting link flow control failed; retcode %d,"
1018 		    " fc_err_mask 0x%02x\n", error, set_fc_err_mask);
1019 		error = EIO;
1020 		goto ixl_rebuild_hw_structs_after_reset_err;
1021 	}
1022 
1023 	/* Remove default filters reinstalled by FW on reset */
1024 	ixl_del_default_hw_filters(vsi);
1025 
1026 	/* Receive broadcast Ethernet frames */
1027 	i40e_aq_set_vsi_broadcast(&pf->hw, vsi->seid, TRUE, NULL);
1028 
1029 	/* Determine link state */
1030 	ixl_attach_get_link_status(pf);
1031 
1032 	i40e_aq_set_dcb_parameters(hw, TRUE, NULL);
1033 
1034 	/* Query device FW LLDP status */
1035 	if (i40e_get_fw_lldp_status(hw, &lldp_status) == I40E_SUCCESS) {
1036 		if (lldp_status == I40E_GET_FW_LLDP_STATUS_DISABLED) {
1037 			ixl_set_state(&pf->state,
1038 			    IXL_STATE_FW_LLDP_DISABLED);
1039 		} else {
1040 			ixl_clear_state(&pf->state,
1041 			    IXL_STATE_FW_LLDP_DISABLED);
1042 		}
1043 	}
1044 
1045 	/* Keep admin queue interrupts active while driver is loaded */
1046 	if (vsi->shared->isc_intr == IFLIB_INTR_MSIX) {
1047 		ixl_configure_intr0_msix(pf);
1048 		ixl_enable_intr0(hw);
1049 	}
1050 
1051 	if (is_up) {
1052 		iflib_request_reset(vsi->ctx);
1053 		iflib_admin_intr_deferred(vsi->ctx);
1054 	}
1055 
1056 	device_printf(dev, "Rebuilding driver state done.\n");
1057 	return (0);
1058 
1059 ixl_rebuild_hw_structs_after_reset_err:
1060 	device_printf(dev, "Reload the driver to recover\n");
1061 	return (error);
1062 }
1063 
1064 /*
1065 ** Set flow control using sysctl:
1066 ** 	0 - off
1067 **	1 - rx pause
1068 **	2 - tx pause
1069 **	3 - full
1070 */
1071 int
ixl_sysctl_set_flowcntl(SYSCTL_HANDLER_ARGS)1072 ixl_sysctl_set_flowcntl(SYSCTL_HANDLER_ARGS)
1073 {
1074 	struct ixl_pf *pf = (struct ixl_pf *)arg1;
1075 	struct i40e_hw *hw = &pf->hw;
1076 	device_t dev = pf->dev;
1077 	int requested_fc, error = 0;
1078 	enum i40e_status_code aq_error = 0;
1079 	u8 fc_aq_err = 0;
1080 
1081 	/* Get request */
1082 	requested_fc = pf->fc;
1083 	error = sysctl_handle_int(oidp, &requested_fc, 0, req);
1084 	if ((error) || (req->newptr == NULL))
1085 		return (error);
1086 	if (requested_fc < 0 || requested_fc > 3) {
1087 		device_printf(dev,
1088 		    "Invalid fc mode; valid modes are 0 through 3\n");
1089 		return (EINVAL);
1090 	}
1091 
1092 	/* Set fc ability for port */
1093 	hw->fc.requested_mode = requested_fc;
1094 	aq_error = i40e_set_fc(hw, &fc_aq_err, TRUE);
1095 	if (aq_error) {
1096 		device_printf(dev,
1097 		    "%s: Error setting Flow Control mode %d; fc_err %#x\n",
1098 		    __func__, aq_error, fc_aq_err);
1099 		return (EIO);
1100 	}
1101 	pf->fc = requested_fc;
1102 
1103 	return (0);
1104 }
1105