1 // SPDX-License-Identifier: GPL-2.0-only
2 /* Copyright (C) 2023 Intel Corporation */
3
4 #include "idpf.h"
5 #include "idpf_ptp.h"
6 #include "idpf_virtchnl.h"
7 #include "xdp.h"
8 #include "xsk.h"
9
10 #define idpf_tx_buf_next(buf) (*(u32 *)&(buf)->priv)
11 LIBETH_SQE_CHECK_PRIV(u32);
12
13 /**
14 * idpf_chk_linearize - Check if skb exceeds max descriptors per packet
15 * @skb: send buffer
16 * @max_bufs: maximum scatter gather buffers for single packet
17 * @count: number of buffers this packet needs
18 *
19 * Make sure we don't exceed maximum scatter gather buffers for a single
20 * packet.
21 * TSO case has been handled earlier from idpf_features_check().
22 */
idpf_chk_linearize(const struct sk_buff * skb,unsigned int max_bufs,unsigned int count)23 static bool idpf_chk_linearize(const struct sk_buff *skb,
24 unsigned int max_bufs,
25 unsigned int count)
26 {
27 if (likely(count <= max_bufs))
28 return false;
29
30 if (skb_is_gso(skb))
31 return false;
32
33 return true;
34 }
35
36 /**
37 * idpf_tx_timeout - Respond to a Tx Hang
38 * @netdev: network interface device structure
39 * @txqueue: TX queue
40 */
idpf_tx_timeout(struct net_device * netdev,unsigned int txqueue)41 void idpf_tx_timeout(struct net_device *netdev, unsigned int txqueue)
42 {
43 struct idpf_adapter *adapter = idpf_netdev_to_adapter(netdev);
44
45 adapter->tx_timeout_count++;
46
47 netdev_err(netdev, "Detected Tx timeout: Count %d, Queue %d\n",
48 adapter->tx_timeout_count, txqueue);
49 if (!idpf_is_reset_in_prog(adapter)) {
50 set_bit(IDPF_HR_FUNC_RESET, adapter->flags);
51 queue_delayed_work(adapter->vc_event_wq,
52 &adapter->vc_event_task,
53 msecs_to_jiffies(10));
54 }
55 }
56
idpf_tx_buf_clean(struct idpf_tx_queue * txq)57 static void idpf_tx_buf_clean(struct idpf_tx_queue *txq)
58 {
59 struct libeth_sq_napi_stats ss = { };
60 struct xdp_frame_bulk bq;
61 struct libeth_cq_pp cp = {
62 .dev = txq->dev,
63 .bq = &bq,
64 .ss = &ss,
65 };
66
67 xdp_frame_bulk_init(&bq);
68
69 /* Free all the Tx buffer sk_buffs */
70 for (u32 i = 0; i < txq->buf_pool_size; i++)
71 libeth_tx_complete_any(&txq->tx_buf[i], &cp);
72
73 xdp_flush_frame_bulk(&bq);
74 }
75
76 /**
77 * idpf_tx_buf_rel_all - Free any empty Tx buffers
78 * @txq: queue to be cleaned
79 */
idpf_tx_buf_rel_all(struct idpf_tx_queue * txq)80 static void idpf_tx_buf_rel_all(struct idpf_tx_queue *txq)
81 {
82 /* Buffers already cleared, nothing to do */
83 if (!txq->tx_buf)
84 return;
85
86 if (idpf_queue_has(XSK, txq))
87 idpf_xsksq_clean(txq);
88 else
89 idpf_tx_buf_clean(txq);
90
91 kfree(txq->tx_buf);
92 txq->tx_buf = NULL;
93 }
94
95 /**
96 * idpf_tx_desc_rel - Free Tx resources per queue
97 * @txq: Tx descriptor ring for a specific queue
98 *
99 * Free all transmit software resources
100 */
idpf_tx_desc_rel(struct idpf_tx_queue * txq)101 static void idpf_tx_desc_rel(struct idpf_tx_queue *txq)
102 {
103 bool xdp = idpf_queue_has(XDP, txq);
104
105 if (xdp)
106 libeth_xdpsq_deinit_timer(txq->timer);
107
108 idpf_tx_buf_rel_all(txq);
109
110 if (!xdp)
111 netdev_tx_reset_subqueue(txq->netdev, txq->idx);
112
113 idpf_xsk_clear_queue(txq, VIRTCHNL2_QUEUE_TYPE_TX);
114
115 if (!txq->desc_ring)
116 return;
117
118 if (!xdp && txq->refillq)
119 kfree(txq->refillq->ring);
120
121 dmam_free_coherent(txq->dev, txq->size, txq->desc_ring, txq->dma);
122 txq->desc_ring = NULL;
123 txq->next_to_use = 0;
124 txq->next_to_clean = 0;
125 }
126
127 /**
128 * idpf_compl_desc_rel - Free completion resources per queue
129 * @complq: completion queue
130 *
131 * Free all completion software resources.
132 */
idpf_compl_desc_rel(struct idpf_compl_queue * complq)133 static void idpf_compl_desc_rel(struct idpf_compl_queue *complq)
134 {
135 idpf_xsk_clear_queue(complq, VIRTCHNL2_QUEUE_TYPE_TX_COMPLETION);
136
137 if (!complq->comp)
138 return;
139
140 dma_free_coherent(complq->netdev->dev.parent, complq->size,
141 complq->desc_ring, complq->dma);
142 complq->desc_ring = NULL;
143 complq->next_to_use = 0;
144 complq->next_to_clean = 0;
145 }
146
147 /**
148 * idpf_tx_desc_rel_all - Free Tx Resources for All Queues
149 * @vport: virtual port structure
150 *
151 * Free all transmit software resources
152 */
idpf_tx_desc_rel_all(struct idpf_vport * vport)153 static void idpf_tx_desc_rel_all(struct idpf_vport *vport)
154 {
155 int i, j;
156
157 if (!vport->txq_grps)
158 return;
159
160 for (i = 0; i < vport->num_txq_grp; i++) {
161 struct idpf_txq_group *txq_grp = &vport->txq_grps[i];
162
163 for (j = 0; j < txq_grp->num_txq; j++)
164 idpf_tx_desc_rel(txq_grp->txqs[j]);
165
166 if (idpf_is_queue_model_split(vport->txq_model))
167 idpf_compl_desc_rel(txq_grp->complq);
168 }
169 }
170
171 /**
172 * idpf_tx_buf_alloc_all - Allocate memory for all buffer resources
173 * @tx_q: queue for which the buffers are allocated
174 *
175 * Returns 0 on success, negative on failure
176 */
idpf_tx_buf_alloc_all(struct idpf_tx_queue * tx_q)177 static int idpf_tx_buf_alloc_all(struct idpf_tx_queue *tx_q)
178 {
179 /* Allocate book keeping buffers only. Buffers to be supplied to HW
180 * are allocated by kernel network stack and received as part of skb
181 */
182 if (idpf_queue_has(FLOW_SCH_EN, tx_q))
183 tx_q->buf_pool_size = U16_MAX;
184 else
185 tx_q->buf_pool_size = tx_q->desc_count;
186 tx_q->tx_buf = kcalloc(tx_q->buf_pool_size, sizeof(*tx_q->tx_buf),
187 GFP_KERNEL);
188 if (!tx_q->tx_buf)
189 return -ENOMEM;
190
191 return 0;
192 }
193
194 /**
195 * idpf_tx_desc_alloc - Allocate the Tx descriptors
196 * @vport: vport to allocate resources for
197 * @tx_q: the tx ring to set up
198 *
199 * Returns 0 on success, negative on failure
200 */
idpf_tx_desc_alloc(const struct idpf_vport * vport,struct idpf_tx_queue * tx_q)201 static int idpf_tx_desc_alloc(const struct idpf_vport *vport,
202 struct idpf_tx_queue *tx_q)
203 {
204 struct device *dev = tx_q->dev;
205 struct idpf_sw_queue *refillq;
206 int err;
207
208 err = idpf_tx_buf_alloc_all(tx_q);
209 if (err)
210 goto err_alloc;
211
212 tx_q->size = tx_q->desc_count * sizeof(*tx_q->base_tx);
213
214 /* Allocate descriptors also round up to nearest 4K */
215 tx_q->size = ALIGN(tx_q->size, 4096);
216 tx_q->desc_ring = dmam_alloc_coherent(dev, tx_q->size, &tx_q->dma,
217 GFP_KERNEL);
218 if (!tx_q->desc_ring) {
219 dev_err(dev, "Unable to allocate memory for the Tx descriptor ring, size=%d\n",
220 tx_q->size);
221 err = -ENOMEM;
222 goto err_alloc;
223 }
224
225 tx_q->next_to_use = 0;
226 tx_q->next_to_clean = 0;
227 idpf_queue_set(GEN_CHK, tx_q);
228
229 idpf_xsk_setup_queue(vport, tx_q, VIRTCHNL2_QUEUE_TYPE_TX);
230
231 if (!idpf_queue_has(FLOW_SCH_EN, tx_q))
232 return 0;
233
234 refillq = tx_q->refillq;
235 refillq->desc_count = tx_q->buf_pool_size;
236 refillq->ring = kcalloc(refillq->desc_count, sizeof(u32),
237 GFP_KERNEL);
238 if (!refillq->ring) {
239 err = -ENOMEM;
240 goto err_alloc;
241 }
242
243 for (unsigned int i = 0; i < refillq->desc_count; i++)
244 refillq->ring[i] =
245 FIELD_PREP(IDPF_RFL_BI_BUFID_M, i) |
246 FIELD_PREP(IDPF_RFL_BI_GEN_M,
247 idpf_queue_has(GEN_CHK, refillq));
248
249 /* Go ahead and flip the GEN bit since this counts as filling
250 * up the ring, i.e. we already ring wrapped.
251 */
252 idpf_queue_change(GEN_CHK, refillq);
253
254 tx_q->last_re = tx_q->desc_count - IDPF_TX_SPLITQ_RE_MIN_GAP;
255
256 return 0;
257
258 err_alloc:
259 idpf_tx_desc_rel(tx_q);
260
261 return err;
262 }
263
264 /**
265 * idpf_compl_desc_alloc - allocate completion descriptors
266 * @vport: vport to allocate resources for
267 * @complq: completion queue to set up
268 *
269 * Return: 0 on success, -errno on failure.
270 */
idpf_compl_desc_alloc(const struct idpf_vport * vport,struct idpf_compl_queue * complq)271 static int idpf_compl_desc_alloc(const struct idpf_vport *vport,
272 struct idpf_compl_queue *complq)
273 {
274 u32 desc_size;
275
276 desc_size = idpf_queue_has(FLOW_SCH_EN, complq) ?
277 sizeof(*complq->comp) : sizeof(*complq->comp_4b);
278 complq->size = array_size(complq->desc_count, desc_size);
279
280 complq->desc_ring = dma_alloc_coherent(complq->netdev->dev.parent,
281 complq->size, &complq->dma,
282 GFP_KERNEL);
283 if (!complq->desc_ring)
284 return -ENOMEM;
285
286 complq->next_to_use = 0;
287 complq->next_to_clean = 0;
288 idpf_queue_set(GEN_CHK, complq);
289
290 idpf_xsk_setup_queue(vport, complq,
291 VIRTCHNL2_QUEUE_TYPE_TX_COMPLETION);
292
293 return 0;
294 }
295
296 /**
297 * idpf_tx_desc_alloc_all - allocate all queues Tx resources
298 * @vport: virtual port private structure
299 *
300 * Returns 0 on success, negative on failure
301 */
idpf_tx_desc_alloc_all(struct idpf_vport * vport)302 static int idpf_tx_desc_alloc_all(struct idpf_vport *vport)
303 {
304 int err = 0;
305 int i, j;
306
307 /* Setup buffer queues. In single queue model buffer queues and
308 * completion queues will be same
309 */
310 for (i = 0; i < vport->num_txq_grp; i++) {
311 for (j = 0; j < vport->txq_grps[i].num_txq; j++) {
312 struct idpf_tx_queue *txq = vport->txq_grps[i].txqs[j];
313
314 err = idpf_tx_desc_alloc(vport, txq);
315 if (err) {
316 pci_err(vport->adapter->pdev,
317 "Allocation for Tx Queue %u failed\n",
318 i);
319 goto err_out;
320 }
321 }
322
323 if (!idpf_is_queue_model_split(vport->txq_model))
324 continue;
325
326 /* Setup completion queues */
327 err = idpf_compl_desc_alloc(vport, vport->txq_grps[i].complq);
328 if (err) {
329 pci_err(vport->adapter->pdev,
330 "Allocation for Tx Completion Queue %u failed\n",
331 i);
332 goto err_out;
333 }
334 }
335
336 err_out:
337 if (err)
338 idpf_tx_desc_rel_all(vport);
339
340 return err;
341 }
342
343 /**
344 * idpf_rx_page_rel - Release an rx buffer page
345 * @rx_buf: the buffer to free
346 */
idpf_rx_page_rel(struct libeth_fqe * rx_buf)347 static void idpf_rx_page_rel(struct libeth_fqe *rx_buf)
348 {
349 if (unlikely(!rx_buf->netmem))
350 return;
351
352 libeth_rx_recycle_slow(rx_buf->netmem);
353
354 rx_buf->netmem = 0;
355 rx_buf->offset = 0;
356 }
357
358 /**
359 * idpf_rx_hdr_buf_rel_all - Release header buffer memory
360 * @bufq: queue to use
361 */
idpf_rx_hdr_buf_rel_all(struct idpf_buf_queue * bufq)362 static void idpf_rx_hdr_buf_rel_all(struct idpf_buf_queue *bufq)
363 {
364 struct libeth_fq fq = {
365 .fqes = bufq->hdr_buf,
366 .pp = bufq->hdr_pp,
367 };
368
369 for (u32 i = 0; i < bufq->desc_count; i++)
370 idpf_rx_page_rel(&bufq->hdr_buf[i]);
371
372 libeth_rx_fq_destroy(&fq);
373 bufq->hdr_buf = NULL;
374 bufq->hdr_pp = NULL;
375 }
376
377 /**
378 * idpf_rx_buf_rel_bufq - Free all Rx buffer resources for a buffer queue
379 * @bufq: queue to be cleaned
380 */
idpf_rx_buf_rel_bufq(struct idpf_buf_queue * bufq)381 static void idpf_rx_buf_rel_bufq(struct idpf_buf_queue *bufq)
382 {
383 struct libeth_fq fq = {
384 .fqes = bufq->buf,
385 .pp = bufq->pp,
386 };
387
388 /* queue already cleared, nothing to do */
389 if (!bufq->buf)
390 return;
391
392 if (idpf_queue_has(XSK, bufq)) {
393 idpf_xskfq_rel(bufq);
394 return;
395 }
396
397 /* Free all the bufs allocated and given to hw on Rx queue */
398 for (u32 i = 0; i < bufq->desc_count; i++)
399 idpf_rx_page_rel(&bufq->buf[i]);
400
401 if (idpf_queue_has(HSPLIT_EN, bufq))
402 idpf_rx_hdr_buf_rel_all(bufq);
403
404 libeth_rx_fq_destroy(&fq);
405 bufq->buf = NULL;
406 bufq->pp = NULL;
407 }
408
409 /**
410 * idpf_rx_buf_rel_all - Free all Rx buffer resources for a receive queue
411 * @rxq: queue to be cleaned
412 */
idpf_rx_buf_rel_all(struct idpf_rx_queue * rxq)413 static void idpf_rx_buf_rel_all(struct idpf_rx_queue *rxq)
414 {
415 struct libeth_fq fq = {
416 .fqes = rxq->rx_buf,
417 .pp = rxq->pp,
418 };
419
420 if (!rxq->rx_buf)
421 return;
422
423 for (u32 i = 0; i < rxq->desc_count; i++)
424 idpf_rx_page_rel(&rxq->rx_buf[i]);
425
426 libeth_rx_fq_destroy(&fq);
427 rxq->rx_buf = NULL;
428 rxq->pp = NULL;
429 }
430
431 /**
432 * idpf_rx_desc_rel - Free a specific Rx q resources
433 * @rxq: queue to clean the resources from
434 * @dev: device to free DMA memory
435 * @model: single or split queue model
436 *
437 * Free a specific rx queue resources
438 */
idpf_rx_desc_rel(struct idpf_rx_queue * rxq,struct device * dev,u32 model)439 static void idpf_rx_desc_rel(struct idpf_rx_queue *rxq, struct device *dev,
440 u32 model)
441 {
442 if (!rxq)
443 return;
444
445 if (!idpf_queue_has(XSK, rxq))
446 libeth_xdp_return_stash(&rxq->xdp);
447
448 if (!idpf_is_queue_model_split(model))
449 idpf_rx_buf_rel_all(rxq);
450
451 idpf_xsk_clear_queue(rxq, VIRTCHNL2_QUEUE_TYPE_RX);
452
453 rxq->next_to_alloc = 0;
454 rxq->next_to_clean = 0;
455 rxq->next_to_use = 0;
456 if (!rxq->desc_ring)
457 return;
458
459 dmam_free_coherent(dev, rxq->size, rxq->desc_ring, rxq->dma);
460 rxq->desc_ring = NULL;
461 }
462
463 /**
464 * idpf_rx_desc_rel_bufq - free buffer queue resources
465 * @bufq: buffer queue to clean the resources from
466 * @dev: device to free DMA memory
467 */
idpf_rx_desc_rel_bufq(struct idpf_buf_queue * bufq,struct device * dev)468 static void idpf_rx_desc_rel_bufq(struct idpf_buf_queue *bufq,
469 struct device *dev)
470 {
471 if (!bufq)
472 return;
473
474 idpf_rx_buf_rel_bufq(bufq);
475 idpf_xsk_clear_queue(bufq, VIRTCHNL2_QUEUE_TYPE_RX_BUFFER);
476
477 bufq->next_to_alloc = 0;
478 bufq->next_to_clean = 0;
479 bufq->next_to_use = 0;
480
481 if (!bufq->split_buf)
482 return;
483
484 dma_free_coherent(dev, bufq->size, bufq->split_buf, bufq->dma);
485 bufq->split_buf = NULL;
486 }
487
488 /**
489 * idpf_rx_desc_rel_all - Free Rx Resources for All Queues
490 * @vport: virtual port structure
491 *
492 * Free all rx queues resources
493 */
idpf_rx_desc_rel_all(struct idpf_vport * vport)494 static void idpf_rx_desc_rel_all(struct idpf_vport *vport)
495 {
496 struct device *dev = &vport->adapter->pdev->dev;
497 struct idpf_rxq_group *rx_qgrp;
498 u16 num_rxq;
499 int i, j;
500
501 if (!vport->rxq_grps)
502 return;
503
504 for (i = 0; i < vport->num_rxq_grp; i++) {
505 rx_qgrp = &vport->rxq_grps[i];
506
507 if (!idpf_is_queue_model_split(vport->rxq_model)) {
508 for (j = 0; j < rx_qgrp->singleq.num_rxq; j++)
509 idpf_rx_desc_rel(rx_qgrp->singleq.rxqs[j], dev,
510 VIRTCHNL2_QUEUE_MODEL_SINGLE);
511 continue;
512 }
513
514 num_rxq = rx_qgrp->splitq.num_rxq_sets;
515 for (j = 0; j < num_rxq; j++)
516 idpf_rx_desc_rel(&rx_qgrp->splitq.rxq_sets[j]->rxq,
517 dev, VIRTCHNL2_QUEUE_MODEL_SPLIT);
518
519 if (!rx_qgrp->splitq.bufq_sets)
520 continue;
521
522 for (j = 0; j < vport->num_bufqs_per_qgrp; j++) {
523 struct idpf_bufq_set *bufq_set =
524 &rx_qgrp->splitq.bufq_sets[j];
525
526 idpf_rx_desc_rel_bufq(&bufq_set->bufq, dev);
527 }
528 }
529 }
530
531 /**
532 * idpf_rx_buf_hw_update - Store the new tail and head values
533 * @bufq: queue to bump
534 * @val: new head index
535 */
idpf_rx_buf_hw_update(struct idpf_buf_queue * bufq,u32 val)536 static void idpf_rx_buf_hw_update(struct idpf_buf_queue *bufq, u32 val)
537 {
538 bufq->next_to_use = val;
539
540 if (unlikely(!bufq->tail))
541 return;
542
543 /* writel has an implicit memory barrier */
544 writel(val, bufq->tail);
545 }
546
547 /**
548 * idpf_rx_hdr_buf_alloc_all - Allocate memory for header buffers
549 * @bufq: ring to use
550 *
551 * Returns 0 on success, negative on failure.
552 */
idpf_rx_hdr_buf_alloc_all(struct idpf_buf_queue * bufq)553 static int idpf_rx_hdr_buf_alloc_all(struct idpf_buf_queue *bufq)
554 {
555 struct libeth_fq fq = {
556 .count = bufq->desc_count,
557 .type = LIBETH_FQE_HDR,
558 .xdp = idpf_xdp_enabled(bufq->q_vector->vport),
559 .nid = idpf_q_vector_to_mem(bufq->q_vector),
560 };
561 int ret;
562
563 ret = libeth_rx_fq_create(&fq, &bufq->q_vector->napi);
564 if (ret)
565 return ret;
566
567 bufq->hdr_pp = fq.pp;
568 bufq->hdr_buf = fq.fqes;
569 bufq->hdr_truesize = fq.truesize;
570 bufq->rx_hbuf_size = fq.buf_len;
571
572 return 0;
573 }
574
575 /**
576 * idpf_post_buf_refill - Post buffer id to refill queue
577 * @refillq: refill queue to post to
578 * @buf_id: buffer id to post
579 */
idpf_post_buf_refill(struct idpf_sw_queue * refillq,u16 buf_id)580 static void idpf_post_buf_refill(struct idpf_sw_queue *refillq, u16 buf_id)
581 {
582 u32 nta = refillq->next_to_use;
583
584 /* store the buffer ID and the SW maintained GEN bit to the refillq */
585 refillq->ring[nta] =
586 FIELD_PREP(IDPF_RFL_BI_BUFID_M, buf_id) |
587 FIELD_PREP(IDPF_RFL_BI_GEN_M,
588 idpf_queue_has(GEN_CHK, refillq));
589
590 if (unlikely(++nta == refillq->desc_count)) {
591 nta = 0;
592 idpf_queue_change(GEN_CHK, refillq);
593 }
594
595 refillq->next_to_use = nta;
596 }
597
598 /**
599 * idpf_rx_post_buf_desc - Post buffer to bufq descriptor ring
600 * @bufq: buffer queue to post to
601 * @buf_id: buffer id to post
602 *
603 * Returns false if buffer could not be allocated, true otherwise.
604 */
idpf_rx_post_buf_desc(struct idpf_buf_queue * bufq,u16 buf_id)605 static bool idpf_rx_post_buf_desc(struct idpf_buf_queue *bufq, u16 buf_id)
606 {
607 struct virtchnl2_splitq_rx_buf_desc *splitq_rx_desc = NULL;
608 struct libeth_fq_fp fq = {
609 .count = bufq->desc_count,
610 };
611 u16 nta = bufq->next_to_alloc;
612 dma_addr_t addr;
613
614 splitq_rx_desc = &bufq->split_buf[nta];
615
616 if (idpf_queue_has(HSPLIT_EN, bufq)) {
617 fq.pp = bufq->hdr_pp;
618 fq.fqes = bufq->hdr_buf;
619 fq.truesize = bufq->hdr_truesize;
620
621 addr = libeth_rx_alloc(&fq, buf_id);
622 if (addr == DMA_MAPPING_ERROR)
623 return false;
624
625 splitq_rx_desc->hdr_addr = cpu_to_le64(addr);
626 }
627
628 fq.pp = bufq->pp;
629 fq.fqes = bufq->buf;
630 fq.truesize = bufq->truesize;
631
632 addr = libeth_rx_alloc(&fq, buf_id);
633 if (addr == DMA_MAPPING_ERROR)
634 return false;
635
636 splitq_rx_desc->pkt_addr = cpu_to_le64(addr);
637 splitq_rx_desc->qword0.buf_id = cpu_to_le16(buf_id);
638
639 nta++;
640 if (unlikely(nta == bufq->desc_count))
641 nta = 0;
642 bufq->next_to_alloc = nta;
643
644 return true;
645 }
646
647 /**
648 * idpf_rx_post_init_bufs - Post initial buffers to bufq
649 * @bufq: buffer queue to post working set to
650 * @working_set: number of buffers to put in working set
651 *
652 * Returns true if @working_set bufs were posted successfully, false otherwise.
653 */
idpf_rx_post_init_bufs(struct idpf_buf_queue * bufq,u16 working_set)654 static bool idpf_rx_post_init_bufs(struct idpf_buf_queue *bufq,
655 u16 working_set)
656 {
657 int i;
658
659 for (i = 0; i < working_set; i++) {
660 if (!idpf_rx_post_buf_desc(bufq, i))
661 return false;
662 }
663
664 idpf_rx_buf_hw_update(bufq, ALIGN_DOWN(bufq->next_to_alloc,
665 IDPF_RX_BUF_STRIDE));
666
667 return true;
668 }
669
670 /**
671 * idpf_rx_buf_alloc_singleq - Allocate memory for all buffer resources
672 * @rxq: queue for which the buffers are allocated
673 *
674 * Return: 0 on success, -ENOMEM on failure.
675 */
idpf_rx_buf_alloc_singleq(struct idpf_rx_queue * rxq)676 static int idpf_rx_buf_alloc_singleq(struct idpf_rx_queue *rxq)
677 {
678 if (idpf_rx_singleq_buf_hw_alloc_all(rxq, rxq->desc_count - 1))
679 goto err;
680
681 return 0;
682
683 err:
684 idpf_rx_buf_rel_all(rxq);
685
686 return -ENOMEM;
687 }
688
689 /**
690 * idpf_rx_bufs_init_singleq - Initialize page pool and allocate Rx bufs
691 * @rxq: buffer queue to create page pool for
692 *
693 * Return: 0 on success, -errno on failure.
694 */
idpf_rx_bufs_init_singleq(struct idpf_rx_queue * rxq)695 static int idpf_rx_bufs_init_singleq(struct idpf_rx_queue *rxq)
696 {
697 struct libeth_fq fq = {
698 .count = rxq->desc_count,
699 .type = LIBETH_FQE_MTU,
700 .nid = idpf_q_vector_to_mem(rxq->q_vector),
701 };
702 int ret;
703
704 ret = libeth_rx_fq_create(&fq, &rxq->q_vector->napi);
705 if (ret)
706 return ret;
707
708 rxq->pp = fq.pp;
709 rxq->rx_buf = fq.fqes;
710 rxq->truesize = fq.truesize;
711 rxq->rx_buf_size = fq.buf_len;
712
713 return idpf_rx_buf_alloc_singleq(rxq);
714 }
715
716 /**
717 * idpf_rx_buf_alloc_all - Allocate memory for all buffer resources
718 * @rxbufq: queue for which the buffers are allocated
719 *
720 * Returns 0 on success, negative on failure
721 */
idpf_rx_buf_alloc_all(struct idpf_buf_queue * rxbufq)722 static int idpf_rx_buf_alloc_all(struct idpf_buf_queue *rxbufq)
723 {
724 int err = 0;
725
726 if (idpf_queue_has(HSPLIT_EN, rxbufq)) {
727 err = idpf_rx_hdr_buf_alloc_all(rxbufq);
728 if (err)
729 goto rx_buf_alloc_all_out;
730 }
731
732 /* Allocate buffers to be given to HW. */
733 if (!idpf_rx_post_init_bufs(rxbufq, IDPF_RX_BUFQ_WORKING_SET(rxbufq)))
734 err = -ENOMEM;
735
736 rx_buf_alloc_all_out:
737 if (err)
738 idpf_rx_buf_rel_bufq(rxbufq);
739
740 return err;
741 }
742
743 /**
744 * idpf_rx_bufs_init - Initialize page pool, allocate rx bufs, and post to HW
745 * @bufq: buffer queue to create page pool for
746 * @type: type of Rx buffers to allocate
747 *
748 * Returns 0 on success, negative on failure
749 */
idpf_rx_bufs_init(struct idpf_buf_queue * bufq,enum libeth_fqe_type type)750 static int idpf_rx_bufs_init(struct idpf_buf_queue *bufq,
751 enum libeth_fqe_type type)
752 {
753 struct libeth_fq fq = {
754 .truesize = bufq->truesize,
755 .count = bufq->desc_count,
756 .type = type,
757 .hsplit = idpf_queue_has(HSPLIT_EN, bufq),
758 .xdp = idpf_xdp_enabled(bufq->q_vector->vport),
759 .nid = idpf_q_vector_to_mem(bufq->q_vector),
760 };
761 int ret;
762
763 if (idpf_queue_has(XSK, bufq))
764 return idpf_xskfq_init(bufq);
765
766 ret = libeth_rx_fq_create(&fq, &bufq->q_vector->napi);
767 if (ret)
768 return ret;
769
770 bufq->pp = fq.pp;
771 bufq->buf = fq.fqes;
772 bufq->truesize = fq.truesize;
773 bufq->rx_buf_size = fq.buf_len;
774
775 return idpf_rx_buf_alloc_all(bufq);
776 }
777
778 /**
779 * idpf_rx_bufs_init_all - Initialize all RX bufs
780 * @vport: virtual port struct
781 *
782 * Returns 0 on success, negative on failure
783 */
idpf_rx_bufs_init_all(struct idpf_vport * vport)784 int idpf_rx_bufs_init_all(struct idpf_vport *vport)
785 {
786 bool split = idpf_is_queue_model_split(vport->rxq_model);
787 int i, j, err;
788
789 idpf_xdp_copy_prog_to_rqs(vport, vport->xdp_prog);
790
791 for (i = 0; i < vport->num_rxq_grp; i++) {
792 struct idpf_rxq_group *rx_qgrp = &vport->rxq_grps[i];
793 u32 truesize = 0;
794
795 /* Allocate bufs for the rxq itself in singleq */
796 if (!split) {
797 int num_rxq = rx_qgrp->singleq.num_rxq;
798
799 for (j = 0; j < num_rxq; j++) {
800 struct idpf_rx_queue *q;
801
802 q = rx_qgrp->singleq.rxqs[j];
803 err = idpf_rx_bufs_init_singleq(q);
804 if (err)
805 return err;
806 }
807
808 continue;
809 }
810
811 /* Otherwise, allocate bufs for the buffer queues */
812 for (j = 0; j < vport->num_bufqs_per_qgrp; j++) {
813 enum libeth_fqe_type type;
814 struct idpf_buf_queue *q;
815
816 q = &rx_qgrp->splitq.bufq_sets[j].bufq;
817 q->truesize = truesize;
818
819 type = truesize ? LIBETH_FQE_SHORT : LIBETH_FQE_MTU;
820
821 err = idpf_rx_bufs_init(q, type);
822 if (err)
823 return err;
824
825 truesize = q->truesize >> 1;
826 }
827 }
828
829 return 0;
830 }
831
832 /**
833 * idpf_rx_desc_alloc - Allocate queue Rx resources
834 * @vport: vport to allocate resources for
835 * @rxq: Rx queue for which the resources are setup
836 *
837 * Returns 0 on success, negative on failure
838 */
idpf_rx_desc_alloc(const struct idpf_vport * vport,struct idpf_rx_queue * rxq)839 static int idpf_rx_desc_alloc(const struct idpf_vport *vport,
840 struct idpf_rx_queue *rxq)
841 {
842 struct device *dev = &vport->adapter->pdev->dev;
843
844 rxq->size = rxq->desc_count * sizeof(union virtchnl2_rx_desc);
845
846 /* Allocate descriptors and also round up to nearest 4K */
847 rxq->size = ALIGN(rxq->size, 4096);
848 rxq->desc_ring = dmam_alloc_coherent(dev, rxq->size,
849 &rxq->dma, GFP_KERNEL);
850 if (!rxq->desc_ring) {
851 dev_err(dev, "Unable to allocate memory for the Rx descriptor ring, size=%d\n",
852 rxq->size);
853 return -ENOMEM;
854 }
855
856 rxq->next_to_alloc = 0;
857 rxq->next_to_clean = 0;
858 rxq->next_to_use = 0;
859 idpf_queue_set(GEN_CHK, rxq);
860
861 idpf_xsk_setup_queue(vport, rxq, VIRTCHNL2_QUEUE_TYPE_RX);
862
863 return 0;
864 }
865
866 /**
867 * idpf_bufq_desc_alloc - Allocate buffer queue descriptor ring
868 * @vport: vport to allocate resources for
869 * @bufq: buffer queue for which the resources are set up
870 *
871 * Return: 0 on success, -ENOMEM on failure.
872 */
idpf_bufq_desc_alloc(const struct idpf_vport * vport,struct idpf_buf_queue * bufq)873 static int idpf_bufq_desc_alloc(const struct idpf_vport *vport,
874 struct idpf_buf_queue *bufq)
875 {
876 struct device *dev = &vport->adapter->pdev->dev;
877
878 bufq->size = array_size(bufq->desc_count, sizeof(*bufq->split_buf));
879
880 bufq->split_buf = dma_alloc_coherent(dev, bufq->size, &bufq->dma,
881 GFP_KERNEL);
882 if (!bufq->split_buf)
883 return -ENOMEM;
884
885 bufq->next_to_alloc = 0;
886 bufq->next_to_clean = 0;
887 bufq->next_to_use = 0;
888 idpf_queue_set(GEN_CHK, bufq);
889
890 idpf_xsk_setup_queue(vport, bufq, VIRTCHNL2_QUEUE_TYPE_RX_BUFFER);
891
892 return 0;
893 }
894
895 /**
896 * idpf_rx_desc_alloc_all - allocate all RX queues resources
897 * @vport: virtual port structure
898 *
899 * Returns 0 on success, negative on failure
900 */
idpf_rx_desc_alloc_all(struct idpf_vport * vport)901 static int idpf_rx_desc_alloc_all(struct idpf_vport *vport)
902 {
903 struct idpf_rxq_group *rx_qgrp;
904 int i, j, err;
905 u16 num_rxq;
906
907 for (i = 0; i < vport->num_rxq_grp; i++) {
908 rx_qgrp = &vport->rxq_grps[i];
909 if (idpf_is_queue_model_split(vport->rxq_model))
910 num_rxq = rx_qgrp->splitq.num_rxq_sets;
911 else
912 num_rxq = rx_qgrp->singleq.num_rxq;
913
914 for (j = 0; j < num_rxq; j++) {
915 struct idpf_rx_queue *q;
916
917 if (idpf_is_queue_model_split(vport->rxq_model))
918 q = &rx_qgrp->splitq.rxq_sets[j]->rxq;
919 else
920 q = rx_qgrp->singleq.rxqs[j];
921
922 err = idpf_rx_desc_alloc(vport, q);
923 if (err) {
924 pci_err(vport->adapter->pdev,
925 "Memory allocation for Rx Queue %u failed\n",
926 i);
927 goto err_out;
928 }
929 }
930
931 if (!idpf_is_queue_model_split(vport->rxq_model))
932 continue;
933
934 for (j = 0; j < vport->num_bufqs_per_qgrp; j++) {
935 struct idpf_buf_queue *q;
936
937 q = &rx_qgrp->splitq.bufq_sets[j].bufq;
938
939 err = idpf_bufq_desc_alloc(vport, q);
940 if (err) {
941 pci_err(vport->adapter->pdev,
942 "Memory allocation for Rx Buffer Queue %u failed\n",
943 i);
944 goto err_out;
945 }
946 }
947 }
948
949 return 0;
950
951 err_out:
952 idpf_rx_desc_rel_all(vport);
953
954 return err;
955 }
956
idpf_init_queue_set(const struct idpf_queue_set * qs)957 static int idpf_init_queue_set(const struct idpf_queue_set *qs)
958 {
959 const struct idpf_vport *vport = qs->vport;
960 bool splitq;
961 int err;
962
963 splitq = idpf_is_queue_model_split(vport->rxq_model);
964
965 for (u32 i = 0; i < qs->num; i++) {
966 const struct idpf_queue_ptr *q = &qs->qs[i];
967 struct idpf_buf_queue *bufq;
968
969 switch (q->type) {
970 case VIRTCHNL2_QUEUE_TYPE_RX:
971 err = idpf_rx_desc_alloc(vport, q->rxq);
972 if (err)
973 break;
974
975 err = idpf_xdp_rxq_info_init(q->rxq);
976 if (err)
977 break;
978
979 if (!splitq)
980 err = idpf_rx_bufs_init_singleq(q->rxq);
981
982 break;
983 case VIRTCHNL2_QUEUE_TYPE_RX_BUFFER:
984 bufq = q->bufq;
985
986 err = idpf_bufq_desc_alloc(vport, bufq);
987 if (err)
988 break;
989
990 for (u32 j = 0; j < bufq->q_vector->num_bufq; j++) {
991 struct idpf_buf_queue * const *bufqs;
992 enum libeth_fqe_type type;
993 u32 ts;
994
995 bufqs = bufq->q_vector->bufq;
996 if (bufqs[j] != bufq)
997 continue;
998
999 if (j) {
1000 type = LIBETH_FQE_SHORT;
1001 ts = bufqs[j - 1]->truesize >> 1;
1002 } else {
1003 type = LIBETH_FQE_MTU;
1004 ts = 0;
1005 }
1006
1007 bufq->truesize = ts;
1008
1009 err = idpf_rx_bufs_init(bufq, type);
1010 break;
1011 }
1012
1013 break;
1014 case VIRTCHNL2_QUEUE_TYPE_TX:
1015 err = idpf_tx_desc_alloc(vport, q->txq);
1016 break;
1017 case VIRTCHNL2_QUEUE_TYPE_TX_COMPLETION:
1018 err = idpf_compl_desc_alloc(vport, q->complq);
1019 break;
1020 default:
1021 continue;
1022 }
1023
1024 if (err)
1025 return err;
1026 }
1027
1028 return 0;
1029 }
1030
idpf_clean_queue_set(const struct idpf_queue_set * qs)1031 static void idpf_clean_queue_set(const struct idpf_queue_set *qs)
1032 {
1033 const struct idpf_vport *vport = qs->vport;
1034 struct device *dev = vport->netdev->dev.parent;
1035
1036 for (u32 i = 0; i < qs->num; i++) {
1037 const struct idpf_queue_ptr *q = &qs->qs[i];
1038
1039 switch (q->type) {
1040 case VIRTCHNL2_QUEUE_TYPE_RX:
1041 idpf_xdp_rxq_info_deinit(q->rxq, vport->rxq_model);
1042 idpf_rx_desc_rel(q->rxq, dev, vport->rxq_model);
1043 break;
1044 case VIRTCHNL2_QUEUE_TYPE_RX_BUFFER:
1045 idpf_rx_desc_rel_bufq(q->bufq, dev);
1046 break;
1047 case VIRTCHNL2_QUEUE_TYPE_TX:
1048 idpf_tx_desc_rel(q->txq);
1049
1050 if (idpf_queue_has(XDP, q->txq)) {
1051 q->txq->pending = 0;
1052 q->txq->xdp_tx = 0;
1053 } else {
1054 q->txq->txq_grp->num_completions_pending = 0;
1055 }
1056
1057 writel(q->txq->next_to_use, q->txq->tail);
1058 break;
1059 case VIRTCHNL2_QUEUE_TYPE_TX_COMPLETION:
1060 idpf_compl_desc_rel(q->complq);
1061 q->complq->num_completions = 0;
1062 break;
1063 default:
1064 break;
1065 }
1066 }
1067 }
1068
idpf_qvec_ena_irq(struct idpf_q_vector * qv)1069 static void idpf_qvec_ena_irq(struct idpf_q_vector *qv)
1070 {
1071 if (qv->num_txq) {
1072 u32 itr;
1073
1074 if (IDPF_ITR_IS_DYNAMIC(qv->tx_intr_mode))
1075 itr = qv->vport->tx_itr_profile[qv->tx_dim.profile_ix];
1076 else
1077 itr = qv->tx_itr_value;
1078
1079 idpf_vport_intr_write_itr(qv, itr, true);
1080 }
1081
1082 if (qv->num_rxq) {
1083 u32 itr;
1084
1085 if (IDPF_ITR_IS_DYNAMIC(qv->rx_intr_mode))
1086 itr = qv->vport->rx_itr_profile[qv->rx_dim.profile_ix];
1087 else
1088 itr = qv->rx_itr_value;
1089
1090 idpf_vport_intr_write_itr(qv, itr, false);
1091 }
1092
1093 if (qv->num_txq || qv->num_rxq)
1094 idpf_vport_intr_update_itr_ena_irq(qv);
1095 }
1096
1097 /**
1098 * idpf_vector_to_queue_set - create a queue set associated with the given
1099 * queue vector
1100 * @qv: queue vector corresponding to the queue pair
1101 *
1102 * Returns a pointer to a dynamically allocated array of pointers to all
1103 * queues associated with a given queue vector (@qv).
1104 * Please note that the caller is responsible to free the memory allocated
1105 * by this function using kfree().
1106 *
1107 * Return: &idpf_queue_set on success, %NULL in case of error.
1108 */
1109 static struct idpf_queue_set *
idpf_vector_to_queue_set(struct idpf_q_vector * qv)1110 idpf_vector_to_queue_set(struct idpf_q_vector *qv)
1111 {
1112 bool xdp = qv->vport->xdp_txq_offset && !qv->num_xsksq;
1113 struct idpf_vport *vport = qv->vport;
1114 struct idpf_queue_set *qs;
1115 u32 num;
1116
1117 num = qv->num_rxq + qv->num_bufq + qv->num_txq + qv->num_complq;
1118 num += xdp ? qv->num_rxq * 2 : qv->num_xsksq * 2;
1119 if (!num)
1120 return NULL;
1121
1122 qs = idpf_alloc_queue_set(vport, num);
1123 if (!qs)
1124 return NULL;
1125
1126 num = 0;
1127
1128 for (u32 i = 0; i < qv->num_bufq; i++) {
1129 qs->qs[num].type = VIRTCHNL2_QUEUE_TYPE_RX_BUFFER;
1130 qs->qs[num++].bufq = qv->bufq[i];
1131 }
1132
1133 for (u32 i = 0; i < qv->num_rxq; i++) {
1134 qs->qs[num].type = VIRTCHNL2_QUEUE_TYPE_RX;
1135 qs->qs[num++].rxq = qv->rx[i];
1136 }
1137
1138 for (u32 i = 0; i < qv->num_txq; i++) {
1139 qs->qs[num].type = VIRTCHNL2_QUEUE_TYPE_TX;
1140 qs->qs[num++].txq = qv->tx[i];
1141 }
1142
1143 for (u32 i = 0; i < qv->num_complq; i++) {
1144 qs->qs[num].type = VIRTCHNL2_QUEUE_TYPE_TX_COMPLETION;
1145 qs->qs[num++].complq = qv->complq[i];
1146 }
1147
1148 if (!vport->xdp_txq_offset)
1149 goto finalize;
1150
1151 if (xdp) {
1152 for (u32 i = 0; i < qv->num_rxq; i++) {
1153 u32 idx = vport->xdp_txq_offset + qv->rx[i]->idx;
1154
1155 qs->qs[num].type = VIRTCHNL2_QUEUE_TYPE_TX;
1156 qs->qs[num++].txq = vport->txqs[idx];
1157
1158 qs->qs[num].type = VIRTCHNL2_QUEUE_TYPE_TX_COMPLETION;
1159 qs->qs[num++].complq = vport->txqs[idx]->complq;
1160 }
1161 } else {
1162 for (u32 i = 0; i < qv->num_xsksq; i++) {
1163 qs->qs[num].type = VIRTCHNL2_QUEUE_TYPE_TX;
1164 qs->qs[num++].txq = qv->xsksq[i];
1165
1166 qs->qs[num].type = VIRTCHNL2_QUEUE_TYPE_TX_COMPLETION;
1167 qs->qs[num++].complq = qv->xsksq[i]->complq;
1168 }
1169 }
1170
1171 finalize:
1172 if (num != qs->num) {
1173 kfree(qs);
1174 return NULL;
1175 }
1176
1177 return qs;
1178 }
1179
idpf_qp_enable(const struct idpf_queue_set * qs,u32 qid)1180 static int idpf_qp_enable(const struct idpf_queue_set *qs, u32 qid)
1181 {
1182 struct idpf_vport *vport = qs->vport;
1183 struct idpf_q_vector *q_vector;
1184 int err;
1185
1186 q_vector = idpf_find_rxq_vec(vport, qid);
1187
1188 err = idpf_init_queue_set(qs);
1189 if (err) {
1190 netdev_err(vport->netdev, "Could not initialize queues in pair %u: %pe\n",
1191 qid, ERR_PTR(err));
1192 return err;
1193 }
1194
1195 if (!vport->xdp_txq_offset)
1196 goto config;
1197
1198 q_vector->xsksq = kcalloc(DIV_ROUND_UP(vport->num_rxq_grp,
1199 vport->num_q_vectors),
1200 sizeof(*q_vector->xsksq), GFP_KERNEL);
1201 if (!q_vector->xsksq)
1202 return -ENOMEM;
1203
1204 for (u32 i = 0; i < qs->num; i++) {
1205 const struct idpf_queue_ptr *q = &qs->qs[i];
1206
1207 if (q->type != VIRTCHNL2_QUEUE_TYPE_TX)
1208 continue;
1209
1210 if (!idpf_queue_has(XSK, q->txq))
1211 continue;
1212
1213 idpf_xsk_init_wakeup(q_vector);
1214
1215 q->txq->q_vector = q_vector;
1216 q_vector->xsksq[q_vector->num_xsksq++] = q->txq;
1217 }
1218
1219 config:
1220 err = idpf_send_config_queue_set_msg(qs);
1221 if (err) {
1222 netdev_err(vport->netdev, "Could not configure queues in pair %u: %pe\n",
1223 qid, ERR_PTR(err));
1224 return err;
1225 }
1226
1227 err = idpf_send_enable_queue_set_msg(qs);
1228 if (err) {
1229 netdev_err(vport->netdev, "Could not enable queues in pair %u: %pe\n",
1230 qid, ERR_PTR(err));
1231 return err;
1232 }
1233
1234 napi_enable(&q_vector->napi);
1235 idpf_qvec_ena_irq(q_vector);
1236
1237 netif_start_subqueue(vport->netdev, qid);
1238
1239 return 0;
1240 }
1241
idpf_qp_disable(const struct idpf_queue_set * qs,u32 qid)1242 static int idpf_qp_disable(const struct idpf_queue_set *qs, u32 qid)
1243 {
1244 struct idpf_vport *vport = qs->vport;
1245 struct idpf_q_vector *q_vector;
1246 int err;
1247
1248 q_vector = idpf_find_rxq_vec(vport, qid);
1249 netif_stop_subqueue(vport->netdev, qid);
1250
1251 writel(0, q_vector->intr_reg.dyn_ctl);
1252 napi_disable(&q_vector->napi);
1253
1254 err = idpf_send_disable_queue_set_msg(qs);
1255 if (err) {
1256 netdev_err(vport->netdev, "Could not disable queues in pair %u: %pe\n",
1257 qid, ERR_PTR(err));
1258 return err;
1259 }
1260
1261 idpf_clean_queue_set(qs);
1262
1263 kfree(q_vector->xsksq);
1264 q_vector->num_xsksq = 0;
1265
1266 return 0;
1267 }
1268
1269 /**
1270 * idpf_qp_switch - enable or disable queues associated with queue pair
1271 * @vport: vport to switch the pair for
1272 * @qid: index of the queue pair to switch
1273 * @en: whether to enable or disable the pair
1274 *
1275 * Return: 0 on success, -errno on failure.
1276 */
idpf_qp_switch(struct idpf_vport * vport,u32 qid,bool en)1277 int idpf_qp_switch(struct idpf_vport *vport, u32 qid, bool en)
1278 {
1279 struct idpf_q_vector *q_vector = idpf_find_rxq_vec(vport, qid);
1280 struct idpf_queue_set *qs __free(kfree) = NULL;
1281
1282 if (idpf_find_txq_vec(vport, qid) != q_vector)
1283 return -EINVAL;
1284
1285 qs = idpf_vector_to_queue_set(q_vector);
1286 if (!qs)
1287 return -ENOMEM;
1288
1289 return en ? idpf_qp_enable(qs, qid) : idpf_qp_disable(qs, qid);
1290 }
1291
1292 /**
1293 * idpf_txq_group_rel - Release all resources for txq groups
1294 * @vport: vport to release txq groups on
1295 */
idpf_txq_group_rel(struct idpf_vport * vport)1296 static void idpf_txq_group_rel(struct idpf_vport *vport)
1297 {
1298 bool split, flow_sch_en;
1299 int i, j;
1300
1301 if (!vport->txq_grps)
1302 return;
1303
1304 split = idpf_is_queue_model_split(vport->txq_model);
1305 flow_sch_en = !idpf_is_cap_ena(vport->adapter, IDPF_OTHER_CAPS,
1306 VIRTCHNL2_CAP_SPLITQ_QSCHED);
1307
1308 for (i = 0; i < vport->num_txq_grp; i++) {
1309 struct idpf_txq_group *txq_grp = &vport->txq_grps[i];
1310
1311 for (j = 0; j < txq_grp->num_txq; j++) {
1312 if (flow_sch_en) {
1313 kfree(txq_grp->txqs[j]->refillq);
1314 txq_grp->txqs[j]->refillq = NULL;
1315 }
1316
1317 kfree(txq_grp->txqs[j]);
1318 txq_grp->txqs[j] = NULL;
1319 }
1320
1321 if (!split)
1322 continue;
1323
1324 kfree(txq_grp->complq);
1325 txq_grp->complq = NULL;
1326 }
1327 kfree(vport->txq_grps);
1328 vport->txq_grps = NULL;
1329 }
1330
1331 /**
1332 * idpf_rxq_sw_queue_rel - Release software queue resources
1333 * @rx_qgrp: rx queue group with software queues
1334 */
idpf_rxq_sw_queue_rel(struct idpf_rxq_group * rx_qgrp)1335 static void idpf_rxq_sw_queue_rel(struct idpf_rxq_group *rx_qgrp)
1336 {
1337 int i, j;
1338
1339 for (i = 0; i < rx_qgrp->vport->num_bufqs_per_qgrp; i++) {
1340 struct idpf_bufq_set *bufq_set = &rx_qgrp->splitq.bufq_sets[i];
1341
1342 for (j = 0; j < bufq_set->num_refillqs; j++) {
1343 kfree(bufq_set->refillqs[j].ring);
1344 bufq_set->refillqs[j].ring = NULL;
1345 }
1346 kfree(bufq_set->refillqs);
1347 bufq_set->refillqs = NULL;
1348 }
1349 }
1350
1351 /**
1352 * idpf_rxq_group_rel - Release all resources for rxq groups
1353 * @vport: vport to release rxq groups on
1354 */
idpf_rxq_group_rel(struct idpf_vport * vport)1355 static void idpf_rxq_group_rel(struct idpf_vport *vport)
1356 {
1357 int i;
1358
1359 if (!vport->rxq_grps)
1360 return;
1361
1362 for (i = 0; i < vport->num_rxq_grp; i++) {
1363 struct idpf_rxq_group *rx_qgrp = &vport->rxq_grps[i];
1364 u16 num_rxq;
1365 int j;
1366
1367 if (idpf_is_queue_model_split(vport->rxq_model)) {
1368 num_rxq = rx_qgrp->splitq.num_rxq_sets;
1369 for (j = 0; j < num_rxq; j++) {
1370 kfree(rx_qgrp->splitq.rxq_sets[j]);
1371 rx_qgrp->splitq.rxq_sets[j] = NULL;
1372 }
1373
1374 idpf_rxq_sw_queue_rel(rx_qgrp);
1375 kfree(rx_qgrp->splitq.bufq_sets);
1376 rx_qgrp->splitq.bufq_sets = NULL;
1377 } else {
1378 num_rxq = rx_qgrp->singleq.num_rxq;
1379 for (j = 0; j < num_rxq; j++) {
1380 kfree(rx_qgrp->singleq.rxqs[j]);
1381 rx_qgrp->singleq.rxqs[j] = NULL;
1382 }
1383 }
1384 }
1385 kfree(vport->rxq_grps);
1386 vport->rxq_grps = NULL;
1387 }
1388
1389 /**
1390 * idpf_vport_queue_grp_rel_all - Release all queue groups
1391 * @vport: vport to release queue groups for
1392 */
idpf_vport_queue_grp_rel_all(struct idpf_vport * vport)1393 static void idpf_vport_queue_grp_rel_all(struct idpf_vport *vport)
1394 {
1395 idpf_txq_group_rel(vport);
1396 idpf_rxq_group_rel(vport);
1397 }
1398
1399 /**
1400 * idpf_vport_queues_rel - Free memory for all queues
1401 * @vport: virtual port
1402 *
1403 * Free the memory allocated for queues associated to a vport
1404 */
idpf_vport_queues_rel(struct idpf_vport * vport)1405 void idpf_vport_queues_rel(struct idpf_vport *vport)
1406 {
1407 idpf_xdp_copy_prog_to_rqs(vport, NULL);
1408
1409 idpf_tx_desc_rel_all(vport);
1410 idpf_rx_desc_rel_all(vport);
1411
1412 idpf_xdpsqs_put(vport);
1413 idpf_vport_queue_grp_rel_all(vport);
1414
1415 kfree(vport->txqs);
1416 vport->txqs = NULL;
1417 }
1418
1419 /**
1420 * idpf_vport_init_fast_path_txqs - Initialize fast path txq array
1421 * @vport: vport to init txqs on
1422 *
1423 * We get a queue index from skb->queue_mapping and we need a fast way to
1424 * dereference the queue from queue groups. This allows us to quickly pull a
1425 * txq based on a queue index.
1426 *
1427 * Returns 0 on success, negative on failure
1428 */
idpf_vport_init_fast_path_txqs(struct idpf_vport * vport)1429 static int idpf_vport_init_fast_path_txqs(struct idpf_vport *vport)
1430 {
1431 struct idpf_ptp_vport_tx_tstamp_caps *caps = vport->tx_tstamp_caps;
1432 struct work_struct *tstamp_task = &vport->tstamp_task;
1433 int i, j, k = 0;
1434
1435 vport->txqs = kcalloc(vport->num_txq, sizeof(*vport->txqs),
1436 GFP_KERNEL);
1437
1438 if (!vport->txqs)
1439 return -ENOMEM;
1440
1441 for (i = 0; i < vport->num_txq_grp; i++) {
1442 struct idpf_txq_group *tx_grp = &vport->txq_grps[i];
1443
1444 for (j = 0; j < tx_grp->num_txq; j++, k++) {
1445 vport->txqs[k] = tx_grp->txqs[j];
1446 vport->txqs[k]->idx = k;
1447
1448 if (!caps)
1449 continue;
1450
1451 vport->txqs[k]->cached_tstamp_caps = caps;
1452 vport->txqs[k]->tstamp_task = tstamp_task;
1453 }
1454 }
1455
1456 return 0;
1457 }
1458
1459 /**
1460 * idpf_vport_init_num_qs - Initialize number of queues
1461 * @vport: vport to initialize queues
1462 * @vport_msg: data to be filled into vport
1463 */
idpf_vport_init_num_qs(struct idpf_vport * vport,struct virtchnl2_create_vport * vport_msg)1464 void idpf_vport_init_num_qs(struct idpf_vport *vport,
1465 struct virtchnl2_create_vport *vport_msg)
1466 {
1467 struct idpf_vport_user_config_data *config_data;
1468 u16 idx = vport->idx;
1469
1470 config_data = &vport->adapter->vport_config[idx]->user_config;
1471 vport->num_txq = le16_to_cpu(vport_msg->num_tx_q);
1472 vport->num_rxq = le16_to_cpu(vport_msg->num_rx_q);
1473 /* number of txqs and rxqs in config data will be zeros only in the
1474 * driver load path and we dont update them there after
1475 */
1476 if (!config_data->num_req_tx_qs && !config_data->num_req_rx_qs) {
1477 config_data->num_req_tx_qs = le16_to_cpu(vport_msg->num_tx_q);
1478 config_data->num_req_rx_qs = le16_to_cpu(vport_msg->num_rx_q);
1479 }
1480
1481 if (idpf_is_queue_model_split(vport->txq_model))
1482 vport->num_complq = le16_to_cpu(vport_msg->num_tx_complq);
1483 if (idpf_is_queue_model_split(vport->rxq_model))
1484 vport->num_bufq = le16_to_cpu(vport_msg->num_rx_bufq);
1485
1486 vport->xdp_prog = config_data->xdp_prog;
1487 if (idpf_xdp_enabled(vport)) {
1488 vport->xdp_txq_offset = config_data->num_req_tx_qs;
1489 vport->num_xdp_txq = le16_to_cpu(vport_msg->num_tx_q) -
1490 vport->xdp_txq_offset;
1491 vport->xdpsq_share = libeth_xdpsq_shared(vport->num_xdp_txq);
1492 } else {
1493 vport->xdp_txq_offset = 0;
1494 vport->num_xdp_txq = 0;
1495 vport->xdpsq_share = false;
1496 }
1497
1498 /* Adjust number of buffer queues per Rx queue group. */
1499 if (!idpf_is_queue_model_split(vport->rxq_model)) {
1500 vport->num_bufqs_per_qgrp = 0;
1501
1502 return;
1503 }
1504
1505 vport->num_bufqs_per_qgrp = IDPF_MAX_BUFQS_PER_RXQ_GRP;
1506 }
1507
1508 /**
1509 * idpf_vport_calc_num_q_desc - Calculate number of queue groups
1510 * @vport: vport to calculate q groups for
1511 */
idpf_vport_calc_num_q_desc(struct idpf_vport * vport)1512 void idpf_vport_calc_num_q_desc(struct idpf_vport *vport)
1513 {
1514 struct idpf_vport_user_config_data *config_data;
1515 int num_bufqs = vport->num_bufqs_per_qgrp;
1516 u32 num_req_txq_desc, num_req_rxq_desc;
1517 u16 idx = vport->idx;
1518 int i;
1519
1520 config_data = &vport->adapter->vport_config[idx]->user_config;
1521 num_req_txq_desc = config_data->num_req_txq_desc;
1522 num_req_rxq_desc = config_data->num_req_rxq_desc;
1523
1524 vport->complq_desc_count = 0;
1525 if (num_req_txq_desc) {
1526 vport->txq_desc_count = num_req_txq_desc;
1527 if (idpf_is_queue_model_split(vport->txq_model)) {
1528 vport->complq_desc_count = num_req_txq_desc;
1529 if (vport->complq_desc_count < IDPF_MIN_TXQ_COMPLQ_DESC)
1530 vport->complq_desc_count =
1531 IDPF_MIN_TXQ_COMPLQ_DESC;
1532 }
1533 } else {
1534 vport->txq_desc_count = IDPF_DFLT_TX_Q_DESC_COUNT;
1535 if (idpf_is_queue_model_split(vport->txq_model))
1536 vport->complq_desc_count =
1537 IDPF_DFLT_TX_COMPLQ_DESC_COUNT;
1538 }
1539
1540 if (num_req_rxq_desc)
1541 vport->rxq_desc_count = num_req_rxq_desc;
1542 else
1543 vport->rxq_desc_count = IDPF_DFLT_RX_Q_DESC_COUNT;
1544
1545 for (i = 0; i < num_bufqs; i++) {
1546 if (!vport->bufq_desc_count[i])
1547 vport->bufq_desc_count[i] =
1548 IDPF_RX_BUFQ_DESC_COUNT(vport->rxq_desc_count,
1549 num_bufqs);
1550 }
1551 }
1552
1553 /**
1554 * idpf_vport_calc_total_qs - Calculate total number of queues
1555 * @adapter: private data struct
1556 * @vport_idx: vport idx to retrieve vport pointer
1557 * @vport_msg: message to fill with data
1558 * @max_q: vport max queue info
1559 *
1560 * Return 0 on success, error value on failure.
1561 */
idpf_vport_calc_total_qs(struct idpf_adapter * adapter,u16 vport_idx,struct virtchnl2_create_vport * vport_msg,struct idpf_vport_max_q * max_q)1562 int idpf_vport_calc_total_qs(struct idpf_adapter *adapter, u16 vport_idx,
1563 struct virtchnl2_create_vport *vport_msg,
1564 struct idpf_vport_max_q *max_q)
1565 {
1566 int dflt_splitq_txq_grps = 0, dflt_singleq_txqs = 0;
1567 int dflt_splitq_rxq_grps = 0, dflt_singleq_rxqs = 0;
1568 u16 num_req_tx_qs = 0, num_req_rx_qs = 0;
1569 struct idpf_vport_user_config_data *user;
1570 struct idpf_vport_config *vport_config;
1571 u16 num_txq_grps, num_rxq_grps;
1572 u32 num_qs, num_xdpsq;
1573
1574 vport_config = adapter->vport_config[vport_idx];
1575 if (vport_config) {
1576 num_req_tx_qs = vport_config->user_config.num_req_tx_qs;
1577 num_req_rx_qs = vport_config->user_config.num_req_rx_qs;
1578 } else {
1579 u32 num_cpus = netif_get_num_default_rss_queues();
1580
1581 dflt_splitq_txq_grps = min_t(int, max_q->max_txq, num_cpus);
1582 dflt_singleq_txqs = min_t(int, max_q->max_txq, num_cpus);
1583 dflt_splitq_rxq_grps = min_t(int, max_q->max_rxq, num_cpus);
1584 dflt_singleq_rxqs = min_t(int, max_q->max_rxq, num_cpus);
1585 }
1586
1587 if (idpf_is_queue_model_split(le16_to_cpu(vport_msg->txq_model))) {
1588 num_txq_grps = num_req_tx_qs ? num_req_tx_qs : dflt_splitq_txq_grps;
1589 vport_msg->num_tx_complq = cpu_to_le16(num_txq_grps *
1590 IDPF_COMPLQ_PER_GROUP);
1591 vport_msg->num_tx_q = cpu_to_le16(num_txq_grps *
1592 IDPF_DFLT_SPLITQ_TXQ_PER_GROUP);
1593 } else {
1594 num_txq_grps = IDPF_DFLT_SINGLEQ_TX_Q_GROUPS;
1595 num_qs = num_txq_grps * (num_req_tx_qs ? num_req_tx_qs :
1596 dflt_singleq_txqs);
1597 vport_msg->num_tx_q = cpu_to_le16(num_qs);
1598 vport_msg->num_tx_complq = 0;
1599 }
1600 if (idpf_is_queue_model_split(le16_to_cpu(vport_msg->rxq_model))) {
1601 num_rxq_grps = num_req_rx_qs ? num_req_rx_qs : dflt_splitq_rxq_grps;
1602 vport_msg->num_rx_bufq = cpu_to_le16(num_rxq_grps *
1603 IDPF_MAX_BUFQS_PER_RXQ_GRP);
1604 vport_msg->num_rx_q = cpu_to_le16(num_rxq_grps *
1605 IDPF_DFLT_SPLITQ_RXQ_PER_GROUP);
1606 } else {
1607 num_rxq_grps = IDPF_DFLT_SINGLEQ_RX_Q_GROUPS;
1608 num_qs = num_rxq_grps * (num_req_rx_qs ? num_req_rx_qs :
1609 dflt_singleq_rxqs);
1610 vport_msg->num_rx_q = cpu_to_le16(num_qs);
1611 vport_msg->num_rx_bufq = 0;
1612 }
1613
1614 if (!vport_config)
1615 return 0;
1616
1617 user = &vport_config->user_config;
1618 user->num_req_rx_qs = le16_to_cpu(vport_msg->num_rx_q);
1619 user->num_req_tx_qs = le16_to_cpu(vport_msg->num_tx_q);
1620
1621 if (vport_config->user_config.xdp_prog)
1622 num_xdpsq = libeth_xdpsq_num(user->num_req_rx_qs,
1623 user->num_req_tx_qs,
1624 vport_config->max_q.max_txq);
1625 else
1626 num_xdpsq = 0;
1627
1628 vport_msg->num_tx_q = cpu_to_le16(user->num_req_tx_qs + num_xdpsq);
1629 if (idpf_is_queue_model_split(le16_to_cpu(vport_msg->txq_model)))
1630 vport_msg->num_tx_complq = vport_msg->num_tx_q;
1631
1632 return 0;
1633 }
1634
1635 /**
1636 * idpf_vport_calc_num_q_groups - Calculate number of queue groups
1637 * @vport: vport to calculate q groups for
1638 */
idpf_vport_calc_num_q_groups(struct idpf_vport * vport)1639 void idpf_vport_calc_num_q_groups(struct idpf_vport *vport)
1640 {
1641 if (idpf_is_queue_model_split(vport->txq_model))
1642 vport->num_txq_grp = vport->num_txq;
1643 else
1644 vport->num_txq_grp = IDPF_DFLT_SINGLEQ_TX_Q_GROUPS;
1645
1646 if (idpf_is_queue_model_split(vport->rxq_model))
1647 vport->num_rxq_grp = vport->num_rxq;
1648 else
1649 vport->num_rxq_grp = IDPF_DFLT_SINGLEQ_RX_Q_GROUPS;
1650 }
1651
1652 /**
1653 * idpf_vport_calc_numq_per_grp - Calculate number of queues per group
1654 * @vport: vport to calculate queues for
1655 * @num_txq: return parameter for number of TX queues
1656 * @num_rxq: return parameter for number of RX queues
1657 */
idpf_vport_calc_numq_per_grp(struct idpf_vport * vport,u16 * num_txq,u16 * num_rxq)1658 static void idpf_vport_calc_numq_per_grp(struct idpf_vport *vport,
1659 u16 *num_txq, u16 *num_rxq)
1660 {
1661 if (idpf_is_queue_model_split(vport->txq_model))
1662 *num_txq = IDPF_DFLT_SPLITQ_TXQ_PER_GROUP;
1663 else
1664 *num_txq = vport->num_txq;
1665
1666 if (idpf_is_queue_model_split(vport->rxq_model))
1667 *num_rxq = IDPF_DFLT_SPLITQ_RXQ_PER_GROUP;
1668 else
1669 *num_rxq = vport->num_rxq;
1670 }
1671
1672 /**
1673 * idpf_rxq_set_descids - set the descids supported by this queue
1674 * @vport: virtual port data structure
1675 * @q: rx queue for which descids are set
1676 *
1677 */
idpf_rxq_set_descids(const struct idpf_vport * vport,struct idpf_rx_queue * q)1678 static void idpf_rxq_set_descids(const struct idpf_vport *vport,
1679 struct idpf_rx_queue *q)
1680 {
1681 if (idpf_is_queue_model_split(vport->rxq_model))
1682 return;
1683
1684 if (vport->base_rxd)
1685 q->rxdids = VIRTCHNL2_RXDID_1_32B_BASE_M;
1686 else
1687 q->rxdids = VIRTCHNL2_RXDID_2_FLEX_SQ_NIC_M;
1688 }
1689
1690 /**
1691 * idpf_txq_group_alloc - Allocate all txq group resources
1692 * @vport: vport to allocate txq groups for
1693 * @num_txq: number of txqs to allocate for each group
1694 *
1695 * Returns 0 on success, negative on failure
1696 */
idpf_txq_group_alloc(struct idpf_vport * vport,u16 num_txq)1697 static int idpf_txq_group_alloc(struct idpf_vport *vport, u16 num_txq)
1698 {
1699 bool split, flow_sch_en;
1700 int i;
1701
1702 vport->txq_grps = kcalloc(vport->num_txq_grp,
1703 sizeof(*vport->txq_grps), GFP_KERNEL);
1704 if (!vport->txq_grps)
1705 return -ENOMEM;
1706
1707 split = idpf_is_queue_model_split(vport->txq_model);
1708 flow_sch_en = !idpf_is_cap_ena(vport->adapter, IDPF_OTHER_CAPS,
1709 VIRTCHNL2_CAP_SPLITQ_QSCHED);
1710
1711 for (i = 0; i < vport->num_txq_grp; i++) {
1712 struct idpf_txq_group *tx_qgrp = &vport->txq_grps[i];
1713 struct idpf_adapter *adapter = vport->adapter;
1714 int j;
1715
1716 tx_qgrp->vport = vport;
1717 tx_qgrp->num_txq = num_txq;
1718
1719 for (j = 0; j < tx_qgrp->num_txq; j++) {
1720 tx_qgrp->txqs[j] = kzalloc(sizeof(*tx_qgrp->txqs[j]),
1721 GFP_KERNEL);
1722 if (!tx_qgrp->txqs[j])
1723 goto err_alloc;
1724 }
1725
1726 for (j = 0; j < tx_qgrp->num_txq; j++) {
1727 struct idpf_tx_queue *q = tx_qgrp->txqs[j];
1728
1729 q->dev = &adapter->pdev->dev;
1730 q->desc_count = vport->txq_desc_count;
1731 q->tx_max_bufs = idpf_get_max_tx_bufs(adapter);
1732 q->tx_min_pkt_len = idpf_get_min_tx_pkt_len(adapter);
1733 q->netdev = vport->netdev;
1734 q->txq_grp = tx_qgrp;
1735 q->rel_q_id = j;
1736
1737 if (!split) {
1738 q->clean_budget = vport->compln_clean_budget;
1739 idpf_queue_assign(CRC_EN, q,
1740 vport->crc_enable);
1741 }
1742
1743 if (!flow_sch_en)
1744 continue;
1745
1746 idpf_queue_set(FLOW_SCH_EN, q);
1747
1748 q->refillq = kzalloc(sizeof(*q->refillq), GFP_KERNEL);
1749 if (!q->refillq)
1750 goto err_alloc;
1751
1752 idpf_queue_set(GEN_CHK, q->refillq);
1753 idpf_queue_set(RFL_GEN_CHK, q->refillq);
1754 }
1755
1756 if (!split)
1757 continue;
1758
1759 tx_qgrp->complq = kcalloc(IDPF_COMPLQ_PER_GROUP,
1760 sizeof(*tx_qgrp->complq),
1761 GFP_KERNEL);
1762 if (!tx_qgrp->complq)
1763 goto err_alloc;
1764
1765 tx_qgrp->complq->desc_count = vport->complq_desc_count;
1766 tx_qgrp->complq->txq_grp = tx_qgrp;
1767 tx_qgrp->complq->netdev = vport->netdev;
1768 tx_qgrp->complq->clean_budget = vport->compln_clean_budget;
1769
1770 if (flow_sch_en)
1771 idpf_queue_set(FLOW_SCH_EN, tx_qgrp->complq);
1772 }
1773
1774 return 0;
1775
1776 err_alloc:
1777 idpf_txq_group_rel(vport);
1778
1779 return -ENOMEM;
1780 }
1781
1782 /**
1783 * idpf_rxq_group_alloc - Allocate all rxq group resources
1784 * @vport: vport to allocate rxq groups for
1785 * @num_rxq: number of rxqs to allocate for each group
1786 *
1787 * Returns 0 on success, negative on failure
1788 */
idpf_rxq_group_alloc(struct idpf_vport * vport,u16 num_rxq)1789 static int idpf_rxq_group_alloc(struct idpf_vport *vport, u16 num_rxq)
1790 {
1791 int i, k, err = 0;
1792 bool hs;
1793
1794 vport->rxq_grps = kcalloc(vport->num_rxq_grp,
1795 sizeof(struct idpf_rxq_group), GFP_KERNEL);
1796 if (!vport->rxq_grps)
1797 return -ENOMEM;
1798
1799 hs = idpf_vport_get_hsplit(vport) == ETHTOOL_TCP_DATA_SPLIT_ENABLED;
1800
1801 for (i = 0; i < vport->num_rxq_grp; i++) {
1802 struct idpf_rxq_group *rx_qgrp = &vport->rxq_grps[i];
1803 int j;
1804
1805 rx_qgrp->vport = vport;
1806 if (!idpf_is_queue_model_split(vport->rxq_model)) {
1807 rx_qgrp->singleq.num_rxq = num_rxq;
1808 for (j = 0; j < num_rxq; j++) {
1809 rx_qgrp->singleq.rxqs[j] =
1810 kzalloc(sizeof(*rx_qgrp->singleq.rxqs[j]),
1811 GFP_KERNEL);
1812 if (!rx_qgrp->singleq.rxqs[j]) {
1813 err = -ENOMEM;
1814 goto err_alloc;
1815 }
1816 }
1817 goto skip_splitq_rx_init;
1818 }
1819 rx_qgrp->splitq.num_rxq_sets = num_rxq;
1820
1821 for (j = 0; j < num_rxq; j++) {
1822 rx_qgrp->splitq.rxq_sets[j] =
1823 kzalloc(sizeof(struct idpf_rxq_set),
1824 GFP_KERNEL);
1825 if (!rx_qgrp->splitq.rxq_sets[j]) {
1826 err = -ENOMEM;
1827 goto err_alloc;
1828 }
1829 }
1830
1831 rx_qgrp->splitq.bufq_sets = kcalloc(vport->num_bufqs_per_qgrp,
1832 sizeof(struct idpf_bufq_set),
1833 GFP_KERNEL);
1834 if (!rx_qgrp->splitq.bufq_sets) {
1835 err = -ENOMEM;
1836 goto err_alloc;
1837 }
1838
1839 for (j = 0; j < vport->num_bufqs_per_qgrp; j++) {
1840 struct idpf_bufq_set *bufq_set =
1841 &rx_qgrp->splitq.bufq_sets[j];
1842 int swq_size = sizeof(struct idpf_sw_queue);
1843 struct idpf_buf_queue *q;
1844
1845 q = &rx_qgrp->splitq.bufq_sets[j].bufq;
1846 q->desc_count = vport->bufq_desc_count[j];
1847 q->rx_buffer_low_watermark = IDPF_LOW_WATERMARK;
1848
1849 idpf_queue_assign(HSPLIT_EN, q, hs);
1850
1851 bufq_set->num_refillqs = num_rxq;
1852 bufq_set->refillqs = kcalloc(num_rxq, swq_size,
1853 GFP_KERNEL);
1854 if (!bufq_set->refillqs) {
1855 err = -ENOMEM;
1856 goto err_alloc;
1857 }
1858 for (k = 0; k < bufq_set->num_refillqs; k++) {
1859 struct idpf_sw_queue *refillq =
1860 &bufq_set->refillqs[k];
1861
1862 refillq->desc_count =
1863 vport->bufq_desc_count[j];
1864 idpf_queue_set(GEN_CHK, refillq);
1865 idpf_queue_set(RFL_GEN_CHK, refillq);
1866 refillq->ring = kcalloc(refillq->desc_count,
1867 sizeof(*refillq->ring),
1868 GFP_KERNEL);
1869 if (!refillq->ring) {
1870 err = -ENOMEM;
1871 goto err_alloc;
1872 }
1873 }
1874 }
1875
1876 skip_splitq_rx_init:
1877 for (j = 0; j < num_rxq; j++) {
1878 struct idpf_rx_queue *q;
1879
1880 if (!idpf_is_queue_model_split(vport->rxq_model)) {
1881 q = rx_qgrp->singleq.rxqs[j];
1882 goto setup_rxq;
1883 }
1884 q = &rx_qgrp->splitq.rxq_sets[j]->rxq;
1885 rx_qgrp->splitq.rxq_sets[j]->refillq[0] =
1886 &rx_qgrp->splitq.bufq_sets[0].refillqs[j];
1887 if (vport->num_bufqs_per_qgrp > IDPF_SINGLE_BUFQ_PER_RXQ_GRP)
1888 rx_qgrp->splitq.rxq_sets[j]->refillq[1] =
1889 &rx_qgrp->splitq.bufq_sets[1].refillqs[j];
1890
1891 idpf_queue_assign(HSPLIT_EN, q, hs);
1892
1893 setup_rxq:
1894 q->desc_count = vport->rxq_desc_count;
1895 q->rx_ptype_lkup = vport->rx_ptype_lkup;
1896 q->bufq_sets = rx_qgrp->splitq.bufq_sets;
1897 q->idx = (i * num_rxq) + j;
1898 q->rx_buffer_low_watermark = IDPF_LOW_WATERMARK;
1899 q->rx_max_pkt_size = vport->netdev->mtu +
1900 LIBETH_RX_LL_LEN;
1901 idpf_rxq_set_descids(vport, q);
1902 }
1903 }
1904
1905 err_alloc:
1906 if (err)
1907 idpf_rxq_group_rel(vport);
1908
1909 return err;
1910 }
1911
1912 /**
1913 * idpf_vport_queue_grp_alloc_all - Allocate all queue groups/resources
1914 * @vport: vport with qgrps to allocate
1915 *
1916 * Returns 0 on success, negative on failure
1917 */
idpf_vport_queue_grp_alloc_all(struct idpf_vport * vport)1918 static int idpf_vport_queue_grp_alloc_all(struct idpf_vport *vport)
1919 {
1920 u16 num_txq, num_rxq;
1921 int err;
1922
1923 idpf_vport_calc_numq_per_grp(vport, &num_txq, &num_rxq);
1924
1925 err = idpf_txq_group_alloc(vport, num_txq);
1926 if (err)
1927 goto err_out;
1928
1929 err = idpf_rxq_group_alloc(vport, num_rxq);
1930 if (err)
1931 goto err_out;
1932
1933 return 0;
1934
1935 err_out:
1936 idpf_vport_queue_grp_rel_all(vport);
1937
1938 return err;
1939 }
1940
1941 /**
1942 * idpf_vport_queues_alloc - Allocate memory for all queues
1943 * @vport: virtual port
1944 *
1945 * Allocate memory for queues associated with a vport. Returns 0 on success,
1946 * negative on failure.
1947 */
idpf_vport_queues_alloc(struct idpf_vport * vport)1948 int idpf_vport_queues_alloc(struct idpf_vport *vport)
1949 {
1950 int err;
1951
1952 err = idpf_vport_queue_grp_alloc_all(vport);
1953 if (err)
1954 goto err_out;
1955
1956 err = idpf_vport_init_fast_path_txqs(vport);
1957 if (err)
1958 goto err_out;
1959
1960 err = idpf_xdpsqs_get(vport);
1961 if (err)
1962 goto err_out;
1963
1964 err = idpf_tx_desc_alloc_all(vport);
1965 if (err)
1966 goto err_out;
1967
1968 err = idpf_rx_desc_alloc_all(vport);
1969 if (err)
1970 goto err_out;
1971
1972 return 0;
1973
1974 err_out:
1975 idpf_vport_queues_rel(vport);
1976
1977 return err;
1978 }
1979
1980 /**
1981 * idpf_tx_read_tstamp - schedule a work to read Tx timestamp value
1982 * @txq: queue to read the timestamp from
1983 * @skb: socket buffer to provide Tx timestamp value
1984 *
1985 * Schedule a work to read Tx timestamp value generated once the packet is
1986 * transmitted.
1987 */
idpf_tx_read_tstamp(struct idpf_tx_queue * txq,struct sk_buff * skb)1988 static void idpf_tx_read_tstamp(struct idpf_tx_queue *txq, struct sk_buff *skb)
1989 {
1990 struct idpf_ptp_vport_tx_tstamp_caps *tx_tstamp_caps;
1991 struct idpf_ptp_tx_tstamp_status *tx_tstamp_status;
1992
1993 tx_tstamp_caps = txq->cached_tstamp_caps;
1994 spin_lock_bh(&tx_tstamp_caps->status_lock);
1995
1996 for (u32 i = 0; i < tx_tstamp_caps->num_entries; i++) {
1997 tx_tstamp_status = &tx_tstamp_caps->tx_tstamp_status[i];
1998 if (tx_tstamp_status->state != IDPF_PTP_FREE)
1999 continue;
2000
2001 tx_tstamp_status->skb = skb;
2002 tx_tstamp_status->state = IDPF_PTP_REQUEST;
2003
2004 /* Fetch timestamp from completion descriptor through
2005 * virtchnl msg to report to stack.
2006 */
2007 queue_work(system_unbound_wq, txq->tstamp_task);
2008 break;
2009 }
2010
2011 spin_unlock_bh(&tx_tstamp_caps->status_lock);
2012 }
2013
2014 #define idpf_tx_splitq_clean_bump_ntc(txq, ntc, desc, buf) \
2015 do { \
2016 if (unlikely(++(ntc) == (txq)->desc_count)) { \
2017 ntc = 0; \
2018 buf = (txq)->tx_buf; \
2019 desc = &(txq)->flex_tx[0]; \
2020 } else { \
2021 (buf)++; \
2022 (desc)++; \
2023 } \
2024 } while (0)
2025
2026 /**
2027 * idpf_tx_splitq_clean - Reclaim resources from buffer queue
2028 * @tx_q: Tx queue to clean
2029 * @end: queue index until which it should be cleaned
2030 * @napi_budget: Used to determine if we are in netpoll
2031 * @cleaned: pointer to stats struct to track cleaned packets/bytes
2032 * @descs_only: true if queue is using flow-based scheduling and should
2033 * not clean buffers at this time
2034 *
2035 * Cleans the queue descriptor ring. If the queue is using queue-based
2036 * scheduling, the buffers will be cleaned as well. If the queue is using
2037 * flow-based scheduling, only the descriptors are cleaned at this time.
2038 * Separate packet completion events will be reported on the completion queue,
2039 * and the buffers will be cleaned separately. The stats are not updated from
2040 * this function when using flow-based scheduling.
2041 */
idpf_tx_splitq_clean(struct idpf_tx_queue * tx_q,u16 end,int napi_budget,struct libeth_sq_napi_stats * cleaned,bool descs_only)2042 static void idpf_tx_splitq_clean(struct idpf_tx_queue *tx_q, u16 end,
2043 int napi_budget,
2044 struct libeth_sq_napi_stats *cleaned,
2045 bool descs_only)
2046 {
2047 union idpf_tx_flex_desc *next_pending_desc = NULL;
2048 union idpf_tx_flex_desc *tx_desc;
2049 u32 ntc = tx_q->next_to_clean;
2050 struct libeth_cq_pp cp = {
2051 .dev = tx_q->dev,
2052 .ss = cleaned,
2053 .napi = napi_budget,
2054 };
2055 struct idpf_tx_buf *tx_buf;
2056
2057 if (descs_only) {
2058 /* Bump ring index to mark as cleaned. */
2059 tx_q->next_to_clean = end;
2060 return;
2061 }
2062
2063 tx_desc = &tx_q->flex_tx[ntc];
2064 next_pending_desc = &tx_q->flex_tx[end];
2065 tx_buf = &tx_q->tx_buf[ntc];
2066
2067 while (tx_desc != next_pending_desc) {
2068 u32 eop_idx;
2069
2070 /* If this entry in the ring was used as a context descriptor,
2071 * it's corresponding entry in the buffer ring is reserved. We
2072 * can skip this descriptor since there is no buffer to clean.
2073 */
2074 if (tx_buf->type <= LIBETH_SQE_CTX)
2075 goto fetch_next_txq_desc;
2076
2077 if (unlikely(tx_buf->type != LIBETH_SQE_SKB))
2078 break;
2079
2080 eop_idx = tx_buf->rs_idx;
2081 libeth_tx_complete(tx_buf, &cp);
2082
2083 /* unmap remaining buffers */
2084 while (ntc != eop_idx) {
2085 idpf_tx_splitq_clean_bump_ntc(tx_q, ntc,
2086 tx_desc, tx_buf);
2087
2088 /* unmap any remaining paged data */
2089 libeth_tx_complete(tx_buf, &cp);
2090 }
2091
2092 fetch_next_txq_desc:
2093 idpf_tx_splitq_clean_bump_ntc(tx_q, ntc, tx_desc, tx_buf);
2094 }
2095
2096 tx_q->next_to_clean = ntc;
2097 }
2098
2099 /**
2100 * idpf_tx_clean_bufs - clean flow scheduling TX queue buffers
2101 * @txq: queue to clean
2102 * @buf_id: packet's starting buffer ID, from completion descriptor
2103 * @cleaned: pointer to stats struct to track cleaned packets/bytes
2104 * @budget: Used to determine if we are in netpoll
2105 *
2106 * Clean all buffers associated with the packet starting at buf_id. Returns the
2107 * byte/segment count for the cleaned packet.
2108 */
idpf_tx_clean_bufs(struct idpf_tx_queue * txq,u32 buf_id,struct libeth_sq_napi_stats * cleaned,int budget)2109 static void idpf_tx_clean_bufs(struct idpf_tx_queue *txq, u32 buf_id,
2110 struct libeth_sq_napi_stats *cleaned,
2111 int budget)
2112 {
2113 struct idpf_tx_buf *tx_buf = NULL;
2114 struct libeth_cq_pp cp = {
2115 .dev = txq->dev,
2116 .ss = cleaned,
2117 .napi = budget,
2118 };
2119
2120 tx_buf = &txq->tx_buf[buf_id];
2121 if (tx_buf->type == LIBETH_SQE_SKB) {
2122 if (skb_shinfo(tx_buf->skb)->tx_flags & SKBTX_IN_PROGRESS)
2123 idpf_tx_read_tstamp(txq, tx_buf->skb);
2124
2125 libeth_tx_complete(tx_buf, &cp);
2126 idpf_post_buf_refill(txq->refillq, buf_id);
2127 }
2128
2129 while (idpf_tx_buf_next(tx_buf) != IDPF_TXBUF_NULL) {
2130 buf_id = idpf_tx_buf_next(tx_buf);
2131
2132 tx_buf = &txq->tx_buf[buf_id];
2133 libeth_tx_complete(tx_buf, &cp);
2134 idpf_post_buf_refill(txq->refillq, buf_id);
2135 }
2136 }
2137
2138 /**
2139 * idpf_tx_handle_rs_completion - clean a single packet and all of its buffers
2140 * whether on the buffer ring or in the hash table
2141 * @txq: Tx ring to clean
2142 * @desc: pointer to completion queue descriptor to extract completion
2143 * information from
2144 * @cleaned: pointer to stats struct to track cleaned packets/bytes
2145 * @budget: Used to determine if we are in netpoll
2146 *
2147 * Returns bytes/packets cleaned
2148 */
idpf_tx_handle_rs_completion(struct idpf_tx_queue * txq,struct idpf_splitq_tx_compl_desc * desc,struct libeth_sq_napi_stats * cleaned,int budget)2149 static void idpf_tx_handle_rs_completion(struct idpf_tx_queue *txq,
2150 struct idpf_splitq_tx_compl_desc *desc,
2151 struct libeth_sq_napi_stats *cleaned,
2152 int budget)
2153 {
2154 /* RS completion contains queue head for queue based scheduling or
2155 * completion tag for flow based scheduling.
2156 */
2157 u16 rs_compl_val = le16_to_cpu(desc->common.q_head_compl_tag.q_head);
2158
2159 if (!idpf_queue_has(FLOW_SCH_EN, txq)) {
2160 idpf_tx_splitq_clean(txq, rs_compl_val, budget, cleaned, false);
2161 return;
2162 }
2163
2164 idpf_tx_clean_bufs(txq, rs_compl_val, cleaned, budget);
2165 }
2166
2167 /**
2168 * idpf_tx_clean_complq - Reclaim resources on completion queue
2169 * @complq: Tx ring to clean
2170 * @budget: Used to determine if we are in netpoll
2171 * @cleaned: returns number of packets cleaned
2172 *
2173 * Returns true if there's any budget left (e.g. the clean is finished)
2174 */
idpf_tx_clean_complq(struct idpf_compl_queue * complq,int budget,int * cleaned)2175 static bool idpf_tx_clean_complq(struct idpf_compl_queue *complq, int budget,
2176 int *cleaned)
2177 {
2178 struct idpf_splitq_tx_compl_desc *tx_desc;
2179 s16 ntc = complq->next_to_clean;
2180 struct idpf_netdev_priv *np;
2181 unsigned int complq_budget;
2182 bool complq_ok = true;
2183 int i;
2184
2185 complq_budget = complq->clean_budget;
2186 tx_desc = &complq->comp[ntc];
2187 ntc -= complq->desc_count;
2188
2189 do {
2190 struct libeth_sq_napi_stats cleaned_stats = { };
2191 struct idpf_tx_queue *tx_q;
2192 __le16 hw_head;
2193 int rel_tx_qid;
2194 u8 ctype; /* completion type */
2195 u16 gen;
2196
2197 /* if the descriptor isn't done, no work yet to do */
2198 gen = le16_get_bits(tx_desc->common.qid_comptype_gen,
2199 IDPF_TXD_COMPLQ_GEN_M);
2200 if (idpf_queue_has(GEN_CHK, complq) != gen)
2201 break;
2202
2203 /* Find necessary info of TX queue to clean buffers */
2204 rel_tx_qid = le16_get_bits(tx_desc->common.qid_comptype_gen,
2205 IDPF_TXD_COMPLQ_QID_M);
2206 if (rel_tx_qid >= complq->txq_grp->num_txq ||
2207 !complq->txq_grp->txqs[rel_tx_qid]) {
2208 netdev_err(complq->netdev, "TxQ not found\n");
2209 goto fetch_next_desc;
2210 }
2211 tx_q = complq->txq_grp->txqs[rel_tx_qid];
2212
2213 /* Determine completion type */
2214 ctype = le16_get_bits(tx_desc->common.qid_comptype_gen,
2215 IDPF_TXD_COMPLQ_COMPL_TYPE_M);
2216 switch (ctype) {
2217 case IDPF_TXD_COMPLT_RE:
2218 hw_head = tx_desc->common.q_head_compl_tag.q_head;
2219
2220 idpf_tx_splitq_clean(tx_q, le16_to_cpu(hw_head),
2221 budget, &cleaned_stats, true);
2222 break;
2223 case IDPF_TXD_COMPLT_RS:
2224 idpf_tx_handle_rs_completion(tx_q, tx_desc,
2225 &cleaned_stats, budget);
2226 break;
2227 default:
2228 netdev_err(tx_q->netdev,
2229 "Unknown TX completion type: %d\n", ctype);
2230 goto fetch_next_desc;
2231 }
2232
2233 u64_stats_update_begin(&tx_q->stats_sync);
2234 u64_stats_add(&tx_q->q_stats.packets, cleaned_stats.packets);
2235 u64_stats_add(&tx_q->q_stats.bytes, cleaned_stats.bytes);
2236 tx_q->cleaned_pkts += cleaned_stats.packets;
2237 tx_q->cleaned_bytes += cleaned_stats.bytes;
2238 complq->num_completions++;
2239 u64_stats_update_end(&tx_q->stats_sync);
2240
2241 fetch_next_desc:
2242 tx_desc++;
2243 ntc++;
2244 if (unlikely(!ntc)) {
2245 ntc -= complq->desc_count;
2246 tx_desc = &complq->comp[0];
2247 idpf_queue_change(GEN_CHK, complq);
2248 }
2249
2250 prefetch(tx_desc);
2251
2252 /* update budget accounting */
2253 complq_budget--;
2254 } while (likely(complq_budget));
2255
2256 /* Store the state of the complq to be used later in deciding if a
2257 * TXQ can be started again
2258 */
2259 if (unlikely(IDPF_TX_COMPLQ_PENDING(complq->txq_grp) >
2260 IDPF_TX_COMPLQ_OVERFLOW_THRESH(complq)))
2261 complq_ok = false;
2262
2263 np = netdev_priv(complq->netdev);
2264 for (i = 0; i < complq->txq_grp->num_txq; ++i) {
2265 struct idpf_tx_queue *tx_q = complq->txq_grp->txqs[i];
2266 struct netdev_queue *nq;
2267 bool dont_wake;
2268
2269 /* We didn't clean anything on this queue, move along */
2270 if (!tx_q->cleaned_bytes)
2271 continue;
2272
2273 *cleaned += tx_q->cleaned_pkts;
2274
2275 /* Update BQL */
2276 nq = netdev_get_tx_queue(tx_q->netdev, tx_q->idx);
2277
2278 dont_wake = !complq_ok || np->state != __IDPF_VPORT_UP ||
2279 !netif_carrier_ok(tx_q->netdev);
2280 /* Check if the TXQ needs to and can be restarted */
2281 __netif_txq_completed_wake(nq, tx_q->cleaned_pkts, tx_q->cleaned_bytes,
2282 IDPF_DESC_UNUSED(tx_q), IDPF_TX_WAKE_THRESH,
2283 dont_wake);
2284
2285 /* Reset cleaned stats for the next time this queue is
2286 * cleaned
2287 */
2288 tx_q->cleaned_bytes = 0;
2289 tx_q->cleaned_pkts = 0;
2290 }
2291
2292 ntc += complq->desc_count;
2293 complq->next_to_clean = ntc;
2294
2295 return !!complq_budget;
2296 }
2297
2298 /**
2299 * idpf_wait_for_sw_marker_completion - wait for SW marker of disabled Tx queue
2300 * @txq: disabled Tx queue
2301 *
2302 * When Tx queue is requested for disabling, the CP sends a special completion
2303 * descriptor called "SW marker", meaning the queue is ready to be destroyed.
2304 * If, for some reason, the marker is not received within 500 ms, break the
2305 * polling to not hang the driver.
2306 */
idpf_wait_for_sw_marker_completion(const struct idpf_tx_queue * txq)2307 void idpf_wait_for_sw_marker_completion(const struct idpf_tx_queue *txq)
2308 {
2309 struct idpf_compl_queue *complq;
2310 unsigned long timeout;
2311 bool flow, gen_flag;
2312 u32 ntc;
2313
2314 if (!idpf_queue_has(SW_MARKER, txq))
2315 return;
2316
2317 complq = idpf_queue_has(XDP, txq) ? txq->complq : txq->txq_grp->complq;
2318 ntc = complq->next_to_clean;
2319
2320 flow = idpf_queue_has(FLOW_SCH_EN, complq);
2321 gen_flag = idpf_queue_has(GEN_CHK, complq);
2322
2323 timeout = jiffies + msecs_to_jiffies(IDPF_WAIT_FOR_MARKER_TIMEO);
2324
2325 do {
2326 struct idpf_splitq_4b_tx_compl_desc *tx_desc;
2327 struct idpf_tx_queue *target;
2328 u32 ctype_gen, id;
2329
2330 tx_desc = flow ? &complq->comp[ntc].common :
2331 &complq->comp_4b[ntc];
2332 ctype_gen = le16_to_cpu(tx_desc->qid_comptype_gen);
2333
2334 if (!!(ctype_gen & IDPF_TXD_COMPLQ_GEN_M) != gen_flag) {
2335 usleep_range(500, 1000);
2336 continue;
2337 }
2338
2339 if (FIELD_GET(IDPF_TXD_COMPLQ_COMPL_TYPE_M, ctype_gen) !=
2340 IDPF_TXD_COMPLT_SW_MARKER)
2341 goto next;
2342
2343 id = FIELD_GET(IDPF_TXD_COMPLQ_QID_M, ctype_gen);
2344 target = complq->txq_grp->txqs[id];
2345
2346 idpf_queue_clear(SW_MARKER, target);
2347 if (target == txq)
2348 break;
2349
2350 next:
2351 if (unlikely(++ntc == complq->desc_count)) {
2352 ntc = 0;
2353 gen_flag = !gen_flag;
2354 }
2355 } while (time_before(jiffies, timeout));
2356
2357 idpf_queue_assign(GEN_CHK, complq, gen_flag);
2358 complq->next_to_clean = ntc;
2359 }
2360
2361 /**
2362 * idpf_tx_splitq_build_ctb - populate command tag and size for queue
2363 * based scheduling descriptors
2364 * @desc: descriptor to populate
2365 * @params: pointer to tx params struct
2366 * @td_cmd: command to be filled in desc
2367 * @size: size of buffer
2368 */
idpf_tx_splitq_build_ctb(union idpf_tx_flex_desc * desc,struct idpf_tx_splitq_params * params,u16 td_cmd,u16 size)2369 void idpf_tx_splitq_build_ctb(union idpf_tx_flex_desc *desc,
2370 struct idpf_tx_splitq_params *params,
2371 u16 td_cmd, u16 size)
2372 {
2373 desc->q.qw1.cmd_dtype =
2374 le16_encode_bits(params->dtype, IDPF_FLEX_TXD_QW1_DTYPE_M);
2375 desc->q.qw1.cmd_dtype |=
2376 le16_encode_bits(td_cmd, IDPF_FLEX_TXD_QW1_CMD_M);
2377 desc->q.qw1.buf_size = cpu_to_le16(size);
2378 desc->q.qw1.l2tags.l2tag1 = cpu_to_le16(params->td_tag);
2379 }
2380
2381 /**
2382 * idpf_tx_splitq_build_flow_desc - populate command tag and size for flow
2383 * scheduling descriptors
2384 * @desc: descriptor to populate
2385 * @params: pointer to tx params struct
2386 * @td_cmd: command to be filled in desc
2387 * @size: size of buffer
2388 */
idpf_tx_splitq_build_flow_desc(union idpf_tx_flex_desc * desc,struct idpf_tx_splitq_params * params,u16 td_cmd,u16 size)2389 void idpf_tx_splitq_build_flow_desc(union idpf_tx_flex_desc *desc,
2390 struct idpf_tx_splitq_params *params,
2391 u16 td_cmd, u16 size)
2392 {
2393 *(u32 *)&desc->flow.qw1.cmd_dtype = (u8)(params->dtype | td_cmd);
2394 desc->flow.qw1.rxr_bufsize = cpu_to_le16((u16)size);
2395 desc->flow.qw1.compl_tag = cpu_to_le16(params->compl_tag);
2396 }
2397
2398 /**
2399 * idpf_tx_splitq_has_room - check if enough Tx splitq resources are available
2400 * @tx_q: the queue to be checked
2401 * @descs_needed: number of descriptors required for this packet
2402 * @bufs_needed: number of Tx buffers required for this packet
2403 *
2404 * Return: 0 if no room available, 1 otherwise
2405 */
idpf_txq_has_room(struct idpf_tx_queue * tx_q,u32 descs_needed,u32 bufs_needed)2406 static int idpf_txq_has_room(struct idpf_tx_queue *tx_q, u32 descs_needed,
2407 u32 bufs_needed)
2408 {
2409 if (IDPF_DESC_UNUSED(tx_q) < descs_needed ||
2410 IDPF_TX_COMPLQ_PENDING(tx_q->txq_grp) >
2411 IDPF_TX_COMPLQ_OVERFLOW_THRESH(tx_q->txq_grp->complq) ||
2412 idpf_tx_splitq_get_free_bufs(tx_q->refillq) < bufs_needed)
2413 return 0;
2414 return 1;
2415 }
2416
2417 /**
2418 * idpf_tx_maybe_stop_splitq - 1st level check for Tx splitq stop conditions
2419 * @tx_q: the queue to be checked
2420 * @descs_needed: number of descriptors required for this packet
2421 * @bufs_needed: number of buffers needed for this packet
2422 *
2423 * Return: 0 if stop is not needed
2424 */
idpf_tx_maybe_stop_splitq(struct idpf_tx_queue * tx_q,u32 descs_needed,u32 bufs_needed)2425 static int idpf_tx_maybe_stop_splitq(struct idpf_tx_queue *tx_q,
2426 u32 descs_needed,
2427 u32 bufs_needed)
2428 {
2429 /* Since we have multiple resources to check for splitq, our
2430 * start,stop_thrs becomes a boolean check instead of a count
2431 * threshold.
2432 */
2433 if (netif_subqueue_maybe_stop(tx_q->netdev, tx_q->idx,
2434 idpf_txq_has_room(tx_q, descs_needed,
2435 bufs_needed),
2436 1, 1))
2437 return 0;
2438
2439 u64_stats_update_begin(&tx_q->stats_sync);
2440 u64_stats_inc(&tx_q->q_stats.q_busy);
2441 u64_stats_update_end(&tx_q->stats_sync);
2442
2443 return -EBUSY;
2444 }
2445
2446 /**
2447 * idpf_tx_buf_hw_update - Store the new tail value
2448 * @tx_q: queue to bump
2449 * @val: new tail index
2450 * @xmit_more: more skb's pending
2451 *
2452 * The naming here is special in that 'hw' signals that this function is about
2453 * to do a register write to update our queue status. We know this can only
2454 * mean tail here as HW should be owning head for TX.
2455 */
idpf_tx_buf_hw_update(struct idpf_tx_queue * tx_q,u32 val,bool xmit_more)2456 void idpf_tx_buf_hw_update(struct idpf_tx_queue *tx_q, u32 val,
2457 bool xmit_more)
2458 {
2459 struct netdev_queue *nq;
2460
2461 nq = netdev_get_tx_queue(tx_q->netdev, tx_q->idx);
2462 tx_q->next_to_use = val;
2463
2464 /* Force memory writes to complete before letting h/w
2465 * know there are new descriptors to fetch. (Only
2466 * applicable for weak-ordered memory model archs,
2467 * such as IA-64).
2468 */
2469 wmb();
2470
2471 /* notify HW of packet */
2472 if (netif_xmit_stopped(nq) || !xmit_more)
2473 writel(val, tx_q->tail);
2474 }
2475
2476 /**
2477 * idpf_tx_res_count_required - get number of Tx resources needed for this pkt
2478 * @txq: queue to send buffer on
2479 * @skb: send buffer
2480 * @bufs_needed: (output) number of buffers needed for this skb.
2481 *
2482 * Return: number of data descriptors and buffers needed for this skb.
2483 */
idpf_tx_res_count_required(struct idpf_tx_queue * txq,struct sk_buff * skb,u32 * bufs_needed)2484 unsigned int idpf_tx_res_count_required(struct idpf_tx_queue *txq,
2485 struct sk_buff *skb,
2486 u32 *bufs_needed)
2487 {
2488 const struct skb_shared_info *shinfo;
2489 unsigned int count = 0, i;
2490
2491 count += !!skb_headlen(skb);
2492
2493 if (!skb_is_nonlinear(skb))
2494 return count;
2495
2496 shinfo = skb_shinfo(skb);
2497 *bufs_needed += shinfo->nr_frags;
2498 for (i = 0; i < shinfo->nr_frags; i++) {
2499 unsigned int size;
2500
2501 size = skb_frag_size(&shinfo->frags[i]);
2502
2503 /* We only need to use the idpf_size_to_txd_count check if the
2504 * fragment is going to span multiple descriptors,
2505 * i.e. size >= 16K.
2506 */
2507 if (size >= SZ_16K)
2508 count += idpf_size_to_txd_count(size);
2509 else
2510 count++;
2511 }
2512
2513 if (idpf_chk_linearize(skb, txq->tx_max_bufs, count)) {
2514 if (__skb_linearize(skb))
2515 return 0;
2516
2517 count = idpf_size_to_txd_count(skb->len);
2518 u64_stats_update_begin(&txq->stats_sync);
2519 u64_stats_inc(&txq->q_stats.linearize);
2520 u64_stats_update_end(&txq->stats_sync);
2521 }
2522
2523 return count;
2524 }
2525
2526 /**
2527 * idpf_tx_splitq_bump_ntu - adjust NTU and generation
2528 * @txq: the tx ring to wrap
2529 * @ntu: ring index to bump
2530 */
idpf_tx_splitq_bump_ntu(struct idpf_tx_queue * txq,u16 ntu)2531 static unsigned int idpf_tx_splitq_bump_ntu(struct idpf_tx_queue *txq, u16 ntu)
2532 {
2533 ntu++;
2534
2535 if (ntu == txq->desc_count)
2536 ntu = 0;
2537
2538 return ntu;
2539 }
2540
2541 /**
2542 * idpf_tx_get_free_buf_id - get a free buffer ID from the refill queue
2543 * @refillq: refill queue to get buffer ID from
2544 * @buf_id: return buffer ID
2545 *
2546 * Return: true if a buffer ID was found, false if not
2547 */
idpf_tx_get_free_buf_id(struct idpf_sw_queue * refillq,u32 * buf_id)2548 static bool idpf_tx_get_free_buf_id(struct idpf_sw_queue *refillq,
2549 u32 *buf_id)
2550 {
2551 u32 ntc = refillq->next_to_clean;
2552 u32 refill_desc;
2553
2554 refill_desc = refillq->ring[ntc];
2555
2556 if (unlikely(idpf_queue_has(RFL_GEN_CHK, refillq) !=
2557 !!(refill_desc & IDPF_RFL_BI_GEN_M)))
2558 return false;
2559
2560 *buf_id = FIELD_GET(IDPF_RFL_BI_BUFID_M, refill_desc);
2561
2562 if (unlikely(++ntc == refillq->desc_count)) {
2563 idpf_queue_change(RFL_GEN_CHK, refillq);
2564 ntc = 0;
2565 }
2566
2567 refillq->next_to_clean = ntc;
2568
2569 return true;
2570 }
2571
2572 /**
2573 * idpf_tx_splitq_pkt_err_unmap - Unmap buffers and bump tail in case of error
2574 * @txq: Tx queue to unwind
2575 * @params: pointer to splitq params struct
2576 * @first: starting buffer for packet to unmap
2577 */
idpf_tx_splitq_pkt_err_unmap(struct idpf_tx_queue * txq,struct idpf_tx_splitq_params * params,struct idpf_tx_buf * first)2578 static void idpf_tx_splitq_pkt_err_unmap(struct idpf_tx_queue *txq,
2579 struct idpf_tx_splitq_params *params,
2580 struct idpf_tx_buf *first)
2581 {
2582 struct idpf_sw_queue *refillq = txq->refillq;
2583 struct libeth_sq_napi_stats ss = { };
2584 struct idpf_tx_buf *tx_buf = first;
2585 struct libeth_cq_pp cp = {
2586 .dev = txq->dev,
2587 .ss = &ss,
2588 };
2589
2590 u64_stats_update_begin(&txq->stats_sync);
2591 u64_stats_inc(&txq->q_stats.dma_map_errs);
2592 u64_stats_update_end(&txq->stats_sync);
2593
2594 libeth_tx_complete(tx_buf, &cp);
2595 while (idpf_tx_buf_next(tx_buf) != IDPF_TXBUF_NULL) {
2596 tx_buf = &txq->tx_buf[idpf_tx_buf_next(tx_buf)];
2597 libeth_tx_complete(tx_buf, &cp);
2598 }
2599
2600 /* Update tail in case netdev_xmit_more was previously true. */
2601 idpf_tx_buf_hw_update(txq, params->prev_ntu, false);
2602
2603 if (!refillq)
2604 return;
2605
2606 /* Restore refillq state to avoid leaking tags. */
2607 if (params->prev_refill_gen != idpf_queue_has(RFL_GEN_CHK, refillq))
2608 idpf_queue_change(RFL_GEN_CHK, refillq);
2609 refillq->next_to_clean = params->prev_refill_ntc;
2610 }
2611
2612 /**
2613 * idpf_tx_splitq_map - Build the Tx flex descriptor
2614 * @tx_q: queue to send buffer on
2615 * @params: pointer to splitq params struct
2616 * @first: first buffer info buffer to use
2617 *
2618 * This function loops over the skb data pointed to by *first
2619 * and gets a physical address for each memory location and programs
2620 * it and the length into the transmit flex descriptor.
2621 */
idpf_tx_splitq_map(struct idpf_tx_queue * tx_q,struct idpf_tx_splitq_params * params,struct idpf_tx_buf * first)2622 static void idpf_tx_splitq_map(struct idpf_tx_queue *tx_q,
2623 struct idpf_tx_splitq_params *params,
2624 struct idpf_tx_buf *first)
2625 {
2626 union idpf_tx_flex_desc *tx_desc;
2627 unsigned int data_len, size;
2628 struct idpf_tx_buf *tx_buf;
2629 u16 i = tx_q->next_to_use;
2630 struct netdev_queue *nq;
2631 struct sk_buff *skb;
2632 skb_frag_t *frag;
2633 u32 next_buf_id;
2634 u16 td_cmd = 0;
2635 dma_addr_t dma;
2636
2637 skb = first->skb;
2638
2639 td_cmd = params->offload.td_cmd;
2640
2641 data_len = skb->data_len;
2642 size = skb_headlen(skb);
2643
2644 tx_desc = &tx_q->flex_tx[i];
2645
2646 dma = dma_map_single(tx_q->dev, skb->data, size, DMA_TO_DEVICE);
2647
2648 tx_buf = first;
2649 first->nr_frags = 0;
2650
2651 for (frag = &skb_shinfo(skb)->frags[0];; frag++) {
2652 unsigned int max_data = IDPF_TX_MAX_DESC_DATA_ALIGNED;
2653
2654 if (unlikely(dma_mapping_error(tx_q->dev, dma))) {
2655 idpf_tx_buf_next(tx_buf) = IDPF_TXBUF_NULL;
2656 return idpf_tx_splitq_pkt_err_unmap(tx_q, params,
2657 first);
2658 }
2659
2660 first->nr_frags++;
2661 tx_buf->type = LIBETH_SQE_FRAG;
2662
2663 /* record length, and DMA address */
2664 dma_unmap_len_set(tx_buf, len, size);
2665 dma_unmap_addr_set(tx_buf, dma, dma);
2666
2667 /* buf_addr is in same location for both desc types */
2668 tx_desc->q.buf_addr = cpu_to_le64(dma);
2669
2670 /* The stack can send us fragments that are too large for a
2671 * single descriptor i.e. frag size > 16K-1. We will need to
2672 * split the fragment across multiple descriptors in this case.
2673 * To adhere to HW alignment restrictions, the fragment needs
2674 * to be split such that the first chunk ends on a 4K boundary
2675 * and all subsequent chunks start on a 4K boundary. We still
2676 * want to send as much data as possible though, so our
2677 * intermediate descriptor chunk size will be 12K.
2678 *
2679 * For example, consider a 32K fragment mapped to DMA addr 2600.
2680 * ------------------------------------------------------------
2681 * | frag_size = 32K |
2682 * ------------------------------------------------------------
2683 * |2600 |16384 |28672
2684 *
2685 * 3 descriptors will be used for this fragment. The HW expects
2686 * the descriptors to contain the following:
2687 * ------------------------------------------------------------
2688 * | size = 13784 | size = 12K | size = 6696 |
2689 * | dma = 2600 | dma = 16384 | dma = 28672 |
2690 * ------------------------------------------------------------
2691 *
2692 * We need to first adjust the max_data for the first chunk so
2693 * that it ends on a 4K boundary. By negating the value of the
2694 * DMA address and taking only the low order bits, we're
2695 * effectively calculating
2696 * 4K - (DMA addr lower order bits) =
2697 * bytes to next boundary.
2698 *
2699 * Add that to our base aligned max_data (12K) and we have
2700 * our first chunk size. In the example above,
2701 * 13784 = 12K + (4096-2600)
2702 *
2703 * After guaranteeing the first chunk ends on a 4K boundary, we
2704 * will give the intermediate descriptors 12K chunks and
2705 * whatever is left to the final descriptor. This ensures that
2706 * all descriptors used for the remaining chunks of the
2707 * fragment start on a 4K boundary and we use as few
2708 * descriptors as possible.
2709 */
2710 max_data += -dma & (IDPF_TX_MAX_READ_REQ_SIZE - 1);
2711 while (unlikely(size > IDPF_TX_MAX_DESC_DATA)) {
2712 idpf_tx_splitq_build_desc(tx_desc, params, td_cmd,
2713 max_data);
2714
2715 if (unlikely(++i == tx_q->desc_count)) {
2716 tx_desc = &tx_q->flex_tx[0];
2717 i = 0;
2718 } else {
2719 tx_desc++;
2720 }
2721
2722 /* Adjust the DMA offset and the remaining size of the
2723 * fragment. On the first iteration of this loop,
2724 * max_data will be >= 12K and <= 16K-1. On any
2725 * subsequent iteration of this loop, max_data will
2726 * always be 12K.
2727 */
2728 dma += max_data;
2729 size -= max_data;
2730
2731 /* Reset max_data since remaining chunks will be 12K
2732 * at most
2733 */
2734 max_data = IDPF_TX_MAX_DESC_DATA_ALIGNED;
2735
2736 /* buf_addr is in same location for both desc types */
2737 tx_desc->q.buf_addr = cpu_to_le64(dma);
2738 }
2739
2740 if (!data_len)
2741 break;
2742
2743 idpf_tx_splitq_build_desc(tx_desc, params, td_cmd, size);
2744
2745 if (unlikely(++i == tx_q->desc_count)) {
2746 tx_desc = &tx_q->flex_tx[0];
2747 i = 0;
2748 } else {
2749 tx_desc++;
2750 }
2751
2752 if (idpf_queue_has(FLOW_SCH_EN, tx_q)) {
2753 if (unlikely(!idpf_tx_get_free_buf_id(tx_q->refillq,
2754 &next_buf_id))) {
2755 idpf_tx_buf_next(tx_buf) = IDPF_TXBUF_NULL;
2756 return idpf_tx_splitq_pkt_err_unmap(tx_q, params,
2757 first);
2758 }
2759 } else {
2760 next_buf_id = i;
2761 }
2762 idpf_tx_buf_next(tx_buf) = next_buf_id;
2763 tx_buf = &tx_q->tx_buf[next_buf_id];
2764
2765 size = skb_frag_size(frag);
2766 data_len -= size;
2767
2768 dma = skb_frag_dma_map(tx_q->dev, frag, 0, size,
2769 DMA_TO_DEVICE);
2770 }
2771
2772 /* record SW timestamp if HW timestamp is not available */
2773 skb_tx_timestamp(skb);
2774
2775 first->type = LIBETH_SQE_SKB;
2776
2777 /* write last descriptor with RS and EOP bits */
2778 first->rs_idx = i;
2779 idpf_tx_buf_next(tx_buf) = IDPF_TXBUF_NULL;
2780 td_cmd |= params->eop_cmd;
2781 idpf_tx_splitq_build_desc(tx_desc, params, td_cmd, size);
2782 i = idpf_tx_splitq_bump_ntu(tx_q, i);
2783
2784 tx_q->txq_grp->num_completions_pending++;
2785
2786 /* record bytecount for BQL */
2787 nq = netdev_get_tx_queue(tx_q->netdev, tx_q->idx);
2788 netdev_tx_sent_queue(nq, first->bytes);
2789
2790 idpf_tx_buf_hw_update(tx_q, i, netdev_xmit_more());
2791 }
2792
2793 /**
2794 * idpf_tso - computes mss and TSO length to prepare for TSO
2795 * @skb: pointer to skb
2796 * @off: pointer to struct that holds offload parameters
2797 *
2798 * Returns error (negative) if TSO was requested but cannot be applied to the
2799 * given skb, 0 if TSO does not apply to the given skb, or 1 otherwise.
2800 */
idpf_tso(struct sk_buff * skb,struct idpf_tx_offload_params * off)2801 int idpf_tso(struct sk_buff *skb, struct idpf_tx_offload_params *off)
2802 {
2803 const struct skb_shared_info *shinfo;
2804 union {
2805 struct iphdr *v4;
2806 struct ipv6hdr *v6;
2807 unsigned char *hdr;
2808 } ip;
2809 union {
2810 struct tcphdr *tcp;
2811 struct udphdr *udp;
2812 unsigned char *hdr;
2813 } l4;
2814 u32 paylen, l4_start;
2815 int err;
2816
2817 if (!skb_is_gso(skb))
2818 return 0;
2819
2820 err = skb_cow_head(skb, 0);
2821 if (err < 0)
2822 return err;
2823
2824 shinfo = skb_shinfo(skb);
2825
2826 ip.hdr = skb_network_header(skb);
2827 l4.hdr = skb_transport_header(skb);
2828
2829 /* initialize outer IP header fields */
2830 if (ip.v4->version == 4) {
2831 ip.v4->tot_len = 0;
2832 ip.v4->check = 0;
2833 } else if (ip.v6->version == 6) {
2834 ip.v6->payload_len = 0;
2835 }
2836
2837 l4_start = skb_transport_offset(skb);
2838
2839 /* remove payload length from checksum */
2840 paylen = skb->len - l4_start;
2841
2842 switch (shinfo->gso_type & ~SKB_GSO_DODGY) {
2843 case SKB_GSO_TCPV4:
2844 case SKB_GSO_TCPV6:
2845 csum_replace_by_diff(&l4.tcp->check,
2846 (__force __wsum)htonl(paylen));
2847 off->tso_hdr_len = __tcp_hdrlen(l4.tcp) + l4_start;
2848 break;
2849 case SKB_GSO_UDP_L4:
2850 csum_replace_by_diff(&l4.udp->check,
2851 (__force __wsum)htonl(paylen));
2852 /* compute length of segmentation header */
2853 off->tso_hdr_len = sizeof(struct udphdr) + l4_start;
2854 l4.udp->len = htons(shinfo->gso_size + sizeof(struct udphdr));
2855 break;
2856 default:
2857 return -EINVAL;
2858 }
2859
2860 off->tso_len = skb->len - off->tso_hdr_len;
2861 off->mss = shinfo->gso_size;
2862 off->tso_segs = shinfo->gso_segs;
2863
2864 off->tx_flags |= IDPF_TX_FLAGS_TSO;
2865
2866 return 1;
2867 }
2868
2869
2870 /**
2871 * idpf_tx_splitq_get_ctx_desc - grab next desc and update buffer ring
2872 * @txq: queue to put context descriptor on
2873 *
2874 * Since the TX buffer rings mimics the descriptor ring, update the tx buffer
2875 * ring entry to reflect that this index is a context descriptor
2876 */
2877 static union idpf_flex_tx_ctx_desc *
idpf_tx_splitq_get_ctx_desc(struct idpf_tx_queue * txq)2878 idpf_tx_splitq_get_ctx_desc(struct idpf_tx_queue *txq)
2879 {
2880 union idpf_flex_tx_ctx_desc *desc;
2881 int i = txq->next_to_use;
2882
2883 /* grab the next descriptor */
2884 desc = &txq->flex_ctx[i];
2885 txq->next_to_use = idpf_tx_splitq_bump_ntu(txq, i);
2886
2887 return desc;
2888 }
2889
2890 /**
2891 * idpf_tx_drop_skb - free the SKB and bump tail if necessary
2892 * @tx_q: queue to send buffer on
2893 * @skb: pointer to skb
2894 */
idpf_tx_drop_skb(struct idpf_tx_queue * tx_q,struct sk_buff * skb)2895 netdev_tx_t idpf_tx_drop_skb(struct idpf_tx_queue *tx_q, struct sk_buff *skb)
2896 {
2897 u64_stats_update_begin(&tx_q->stats_sync);
2898 u64_stats_inc(&tx_q->q_stats.skb_drops);
2899 u64_stats_update_end(&tx_q->stats_sync);
2900
2901 idpf_tx_buf_hw_update(tx_q, tx_q->next_to_use, false);
2902
2903 dev_kfree_skb(skb);
2904
2905 return NETDEV_TX_OK;
2906 }
2907
2908 #if (IS_ENABLED(CONFIG_PTP_1588_CLOCK))
2909 /**
2910 * idpf_tx_tstamp - set up context descriptor for hardware timestamp
2911 * @tx_q: queue to send buffer on
2912 * @skb: pointer to the SKB we're sending
2913 * @off: pointer to the offload struct
2914 *
2915 * Return: Positive index number on success, negative otherwise.
2916 */
idpf_tx_tstamp(struct idpf_tx_queue * tx_q,struct sk_buff * skb,struct idpf_tx_offload_params * off)2917 static int idpf_tx_tstamp(struct idpf_tx_queue *tx_q, struct sk_buff *skb,
2918 struct idpf_tx_offload_params *off)
2919 {
2920 int err, idx;
2921
2922 /* only timestamp the outbound packet if the user has requested it */
2923 if (likely(!(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP)))
2924 return -1;
2925
2926 if (!idpf_ptp_get_txq_tstamp_capability(tx_q))
2927 return -1;
2928
2929 /* Tx timestamps cannot be sampled when doing TSO */
2930 if (off->tx_flags & IDPF_TX_FLAGS_TSO)
2931 return -1;
2932
2933 /* Grab an open timestamp slot */
2934 err = idpf_ptp_request_ts(tx_q, skb, &idx);
2935 if (err) {
2936 u64_stats_update_begin(&tx_q->stats_sync);
2937 u64_stats_inc(&tx_q->q_stats.tstamp_skipped);
2938 u64_stats_update_end(&tx_q->stats_sync);
2939
2940 return -1;
2941 }
2942
2943 off->tx_flags |= IDPF_TX_FLAGS_TSYN;
2944
2945 return idx;
2946 }
2947
2948 /**
2949 * idpf_tx_set_tstamp_desc - Set the Tx descriptor fields needed to generate
2950 * PHY Tx timestamp
2951 * @ctx_desc: Context descriptor
2952 * @idx: Index of the Tx timestamp latch
2953 */
idpf_tx_set_tstamp_desc(union idpf_flex_tx_ctx_desc * ctx_desc,u32 idx)2954 static void idpf_tx_set_tstamp_desc(union idpf_flex_tx_ctx_desc *ctx_desc,
2955 u32 idx)
2956 {
2957 ctx_desc->tsyn.qw1 = le64_encode_bits(IDPF_TX_DESC_DTYPE_CTX,
2958 IDPF_TX_CTX_DTYPE_M) |
2959 le64_encode_bits(IDPF_TX_CTX_DESC_TSYN,
2960 IDPF_TX_CTX_CMD_M) |
2961 le64_encode_bits(idx, IDPF_TX_CTX_TSYN_REG_M);
2962 }
2963 #else /* CONFIG_PTP_1588_CLOCK */
idpf_tx_tstamp(struct idpf_tx_queue * tx_q,struct sk_buff * skb,struct idpf_tx_offload_params * off)2964 static int idpf_tx_tstamp(struct idpf_tx_queue *tx_q, struct sk_buff *skb,
2965 struct idpf_tx_offload_params *off)
2966 {
2967 return -1;
2968 }
2969
idpf_tx_set_tstamp_desc(union idpf_flex_tx_ctx_desc * ctx_desc,u32 idx)2970 static void idpf_tx_set_tstamp_desc(union idpf_flex_tx_ctx_desc *ctx_desc,
2971 u32 idx)
2972 { }
2973 #endif /* CONFIG_PTP_1588_CLOCK */
2974
2975 /**
2976 * idpf_tx_splitq_need_re - check whether RE bit needs to be set
2977 * @tx_q: pointer to Tx queue
2978 *
2979 * Return: true if RE bit needs to be set, false otherwise
2980 */
idpf_tx_splitq_need_re(struct idpf_tx_queue * tx_q)2981 static bool idpf_tx_splitq_need_re(struct idpf_tx_queue *tx_q)
2982 {
2983 int gap = tx_q->next_to_use - tx_q->last_re;
2984
2985 gap += (gap < 0) ? tx_q->desc_count : 0;
2986
2987 return gap >= IDPF_TX_SPLITQ_RE_MIN_GAP;
2988 }
2989
2990 /**
2991 * idpf_tx_splitq_frame - Sends buffer on Tx ring using flex descriptors
2992 * @skb: send buffer
2993 * @tx_q: queue to send buffer on
2994 *
2995 * Returns NETDEV_TX_OK if sent, else an error code
2996 */
idpf_tx_splitq_frame(struct sk_buff * skb,struct idpf_tx_queue * tx_q)2997 static netdev_tx_t idpf_tx_splitq_frame(struct sk_buff *skb,
2998 struct idpf_tx_queue *tx_q)
2999 {
3000 struct idpf_tx_splitq_params tx_params = {
3001 .prev_ntu = tx_q->next_to_use,
3002 };
3003 union idpf_flex_tx_ctx_desc *ctx_desc;
3004 struct idpf_tx_buf *first;
3005 u32 count, buf_count = 1;
3006 int tso, idx;
3007 u32 buf_id;
3008
3009 count = idpf_tx_res_count_required(tx_q, skb, &buf_count);
3010 if (unlikely(!count))
3011 return idpf_tx_drop_skb(tx_q, skb);
3012
3013 tso = idpf_tso(skb, &tx_params.offload);
3014 if (unlikely(tso < 0))
3015 return idpf_tx_drop_skb(tx_q, skb);
3016
3017 /* Check for splitq specific TX resources */
3018 count += (IDPF_TX_DESCS_PER_CACHE_LINE + tso);
3019 if (idpf_tx_maybe_stop_splitq(tx_q, count, buf_count)) {
3020 idpf_tx_buf_hw_update(tx_q, tx_q->next_to_use, false);
3021
3022 return NETDEV_TX_BUSY;
3023 }
3024
3025 if (tso) {
3026 /* If tso is needed, set up context desc */
3027 ctx_desc = idpf_tx_splitq_get_ctx_desc(tx_q);
3028
3029 ctx_desc->tso.qw1.cmd_dtype =
3030 cpu_to_le16(IDPF_TX_DESC_DTYPE_FLEX_TSO_CTX |
3031 IDPF_TX_FLEX_CTX_DESC_CMD_TSO);
3032 ctx_desc->tso.qw0.flex_tlen =
3033 cpu_to_le32(tx_params.offload.tso_len &
3034 IDPF_TXD_FLEX_CTX_TLEN_M);
3035 ctx_desc->tso.qw0.mss_rt =
3036 cpu_to_le16(tx_params.offload.mss &
3037 IDPF_TXD_FLEX_CTX_MSS_RT_M);
3038 ctx_desc->tso.qw0.hdr_len = tx_params.offload.tso_hdr_len;
3039
3040 u64_stats_update_begin(&tx_q->stats_sync);
3041 u64_stats_inc(&tx_q->q_stats.lso_pkts);
3042 u64_stats_update_end(&tx_q->stats_sync);
3043 }
3044
3045 idx = idpf_tx_tstamp(tx_q, skb, &tx_params.offload);
3046 if (idx != -1) {
3047 ctx_desc = idpf_tx_splitq_get_ctx_desc(tx_q);
3048 idpf_tx_set_tstamp_desc(ctx_desc, idx);
3049 }
3050
3051 if (idpf_queue_has(FLOW_SCH_EN, tx_q)) {
3052 struct idpf_sw_queue *refillq = tx_q->refillq;
3053
3054 /* Save refillq state in case of a packet rollback. Otherwise,
3055 * the tags will be leaked since they will be popped from the
3056 * refillq but never reposted during cleaning.
3057 */
3058 tx_params.prev_refill_gen =
3059 idpf_queue_has(RFL_GEN_CHK, refillq);
3060 tx_params.prev_refill_ntc = refillq->next_to_clean;
3061
3062 if (unlikely(!idpf_tx_get_free_buf_id(tx_q->refillq,
3063 &buf_id))) {
3064 if (tx_params.prev_refill_gen !=
3065 idpf_queue_has(RFL_GEN_CHK, refillq))
3066 idpf_queue_change(RFL_GEN_CHK, refillq);
3067 refillq->next_to_clean = tx_params.prev_refill_ntc;
3068
3069 tx_q->next_to_use = tx_params.prev_ntu;
3070 return idpf_tx_drop_skb(tx_q, skb);
3071 }
3072 tx_params.compl_tag = buf_id;
3073
3074 tx_params.dtype = IDPF_TX_DESC_DTYPE_FLEX_FLOW_SCHE;
3075 tx_params.eop_cmd = IDPF_TXD_FLEX_FLOW_CMD_EOP;
3076 /* Set the RE bit to periodically "clean" the descriptor ring.
3077 * MIN_GAP is set to MIN_RING size to ensure it will be set at
3078 * least once each time around the ring.
3079 */
3080 if (idpf_tx_splitq_need_re(tx_q)) {
3081 tx_params.eop_cmd |= IDPF_TXD_FLEX_FLOW_CMD_RE;
3082 tx_q->txq_grp->num_completions_pending++;
3083 tx_q->last_re = tx_q->next_to_use;
3084 }
3085
3086 if (skb->ip_summed == CHECKSUM_PARTIAL)
3087 tx_params.offload.td_cmd |= IDPF_TXD_FLEX_FLOW_CMD_CS_EN;
3088
3089 } else {
3090 buf_id = tx_q->next_to_use;
3091
3092 tx_params.dtype = IDPF_TX_DESC_DTYPE_FLEX_L2TAG1_L2TAG2;
3093 tx_params.eop_cmd = IDPF_TXD_LAST_DESC_CMD;
3094
3095 if (skb->ip_summed == CHECKSUM_PARTIAL)
3096 tx_params.offload.td_cmd |= IDPF_TX_FLEX_DESC_CMD_CS_EN;
3097 }
3098
3099 first = &tx_q->tx_buf[buf_id];
3100 first->skb = skb;
3101
3102 if (tso) {
3103 first->packets = tx_params.offload.tso_segs;
3104 first->bytes = skb->len +
3105 ((first->packets - 1) * tx_params.offload.tso_hdr_len);
3106 } else {
3107 first->packets = 1;
3108 first->bytes = max_t(unsigned int, skb->len, ETH_ZLEN);
3109 }
3110
3111 idpf_tx_splitq_map(tx_q, &tx_params, first);
3112
3113 return NETDEV_TX_OK;
3114 }
3115
3116 /**
3117 * idpf_tx_start - Selects the right Tx queue to send buffer
3118 * @skb: send buffer
3119 * @netdev: network interface device structure
3120 *
3121 * Returns NETDEV_TX_OK if sent, else an error code
3122 */
idpf_tx_start(struct sk_buff * skb,struct net_device * netdev)3123 netdev_tx_t idpf_tx_start(struct sk_buff *skb, struct net_device *netdev)
3124 {
3125 const struct idpf_vport *vport = idpf_netdev_to_vport(netdev);
3126 struct idpf_tx_queue *tx_q;
3127
3128 if (unlikely(skb_get_queue_mapping(skb) >=
3129 vport->num_txq - vport->num_xdp_txq)) {
3130 dev_kfree_skb_any(skb);
3131
3132 return NETDEV_TX_OK;
3133 }
3134
3135 tx_q = vport->txqs[skb_get_queue_mapping(skb)];
3136
3137 /* hardware can't handle really short frames, hardware padding works
3138 * beyond this point
3139 */
3140 if (skb_put_padto(skb, tx_q->tx_min_pkt_len)) {
3141 idpf_tx_buf_hw_update(tx_q, tx_q->next_to_use, false);
3142
3143 return NETDEV_TX_OK;
3144 }
3145
3146 if (idpf_is_queue_model_split(vport->txq_model))
3147 return idpf_tx_splitq_frame(skb, tx_q);
3148 else
3149 return idpf_tx_singleq_frame(skb, tx_q);
3150 }
3151
3152 /**
3153 * idpf_rx_hash - set the hash value in the skb
3154 * @rxq: Rx descriptor ring packet is being transacted on
3155 * @skb: pointer to current skb being populated
3156 * @rx_desc: Receive descriptor
3157 * @decoded: Decoded Rx packet type related fields
3158 */
3159 static void
idpf_rx_hash(const struct idpf_rx_queue * rxq,struct sk_buff * skb,const struct virtchnl2_rx_flex_desc_adv_nic_3 * rx_desc,struct libeth_rx_pt decoded)3160 idpf_rx_hash(const struct idpf_rx_queue *rxq, struct sk_buff *skb,
3161 const struct virtchnl2_rx_flex_desc_adv_nic_3 *rx_desc,
3162 struct libeth_rx_pt decoded)
3163 {
3164 u32 hash;
3165
3166 if (!libeth_rx_pt_has_hash(rxq->xdp_rxq.dev, decoded))
3167 return;
3168
3169 hash = le16_to_cpu(rx_desc->hash1) |
3170 (rx_desc->ff2_mirrid_hash2.hash2 << 16) |
3171 (rx_desc->hash3 << 24);
3172
3173 libeth_rx_pt_set_hash(skb, hash, decoded);
3174 }
3175
3176 /**
3177 * idpf_rx_csum - Indicate in skb if checksum is good
3178 * @rxq: Rx descriptor ring packet is being transacted on
3179 * @skb: pointer to current skb being populated
3180 * @csum_bits: checksum fields extracted from the descriptor
3181 * @decoded: Decoded Rx packet type related fields
3182 *
3183 * skb->protocol must be set before this function is called
3184 */
idpf_rx_csum(struct idpf_rx_queue * rxq,struct sk_buff * skb,struct libeth_rx_csum csum_bits,struct libeth_rx_pt decoded)3185 static void idpf_rx_csum(struct idpf_rx_queue *rxq, struct sk_buff *skb,
3186 struct libeth_rx_csum csum_bits,
3187 struct libeth_rx_pt decoded)
3188 {
3189 bool ipv4, ipv6;
3190
3191 /* check if Rx checksum is enabled */
3192 if (!libeth_rx_pt_has_checksum(rxq->xdp_rxq.dev, decoded))
3193 return;
3194
3195 /* check if HW has decoded the packet and checksum */
3196 if (unlikely(!csum_bits.l3l4p))
3197 return;
3198
3199 ipv4 = libeth_rx_pt_get_ip_ver(decoded) == LIBETH_RX_PT_OUTER_IPV4;
3200 ipv6 = libeth_rx_pt_get_ip_ver(decoded) == LIBETH_RX_PT_OUTER_IPV6;
3201
3202 if (unlikely(ipv4 && (csum_bits.ipe || csum_bits.eipe)))
3203 goto checksum_fail;
3204
3205 if (unlikely(ipv6 && csum_bits.ipv6exadd))
3206 return;
3207
3208 /* check for L4 errors and handle packets that were not able to be
3209 * checksummed
3210 */
3211 if (unlikely(csum_bits.l4e))
3212 goto checksum_fail;
3213
3214 if (!csum_bits.raw_csum_valid ||
3215 decoded.inner_prot == LIBETH_RX_PT_INNER_SCTP) {
3216 skb->ip_summed = CHECKSUM_UNNECESSARY;
3217 return;
3218 }
3219
3220 skb->csum = csum_unfold((__force __sum16)~swab16(csum_bits.raw_csum));
3221 skb->ip_summed = CHECKSUM_COMPLETE;
3222
3223 return;
3224
3225 checksum_fail:
3226 u64_stats_update_begin(&rxq->stats_sync);
3227 u64_stats_inc(&rxq->q_stats.hw_csum_err);
3228 u64_stats_update_end(&rxq->stats_sync);
3229 }
3230
3231 /**
3232 * idpf_rx_splitq_extract_csum_bits - Extract checksum bits from descriptor
3233 * @rx_desc: receive descriptor
3234 *
3235 * Return: parsed checksum status.
3236 **/
3237 static struct libeth_rx_csum
idpf_rx_splitq_extract_csum_bits(const struct virtchnl2_rx_flex_desc_adv_nic_3 * rx_desc)3238 idpf_rx_splitq_extract_csum_bits(const struct virtchnl2_rx_flex_desc_adv_nic_3 *rx_desc)
3239 {
3240 struct libeth_rx_csum csum = { };
3241 u8 qword0, qword1;
3242
3243 qword0 = rx_desc->status_err0_qw0;
3244 qword1 = rx_desc->status_err0_qw1;
3245
3246 csum.ipe = FIELD_GET(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_IPE_M,
3247 qword1);
3248 csum.eipe = FIELD_GET(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_EIPE_M,
3249 qword1);
3250 csum.l4e = FIELD_GET(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_L4E_M,
3251 qword1);
3252 csum.l3l4p = FIELD_GET(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_L3L4P_M,
3253 qword1);
3254 csum.ipv6exadd = FIELD_GET(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_IPV6EXADD_M,
3255 qword0);
3256 csum.raw_csum_valid =
3257 !le16_get_bits(rx_desc->ptype_err_fflags0,
3258 VIRTCHNL2_RX_FLEX_DESC_ADV_RAW_CSUM_INV_M);
3259 csum.raw_csum = le16_to_cpu(rx_desc->misc.raw_cs);
3260
3261 return csum;
3262 }
3263
3264 /**
3265 * idpf_rx_rsc - Set the RSC fields in the skb
3266 * @rxq : Rx descriptor ring packet is being transacted on
3267 * @skb : pointer to current skb being populated
3268 * @rx_desc: Receive descriptor
3269 * @decoded: Decoded Rx packet type related fields
3270 *
3271 * Return 0 on success and error code on failure
3272 *
3273 * Populate the skb fields with the total number of RSC segments, RSC payload
3274 * length and packet type.
3275 */
idpf_rx_rsc(struct idpf_rx_queue * rxq,struct sk_buff * skb,const struct virtchnl2_rx_flex_desc_adv_nic_3 * rx_desc,struct libeth_rx_pt decoded)3276 static int idpf_rx_rsc(struct idpf_rx_queue *rxq, struct sk_buff *skb,
3277 const struct virtchnl2_rx_flex_desc_adv_nic_3 *rx_desc,
3278 struct libeth_rx_pt decoded)
3279 {
3280 u16 rsc_segments, rsc_seg_len;
3281 bool ipv4, ipv6;
3282 int len;
3283
3284 if (unlikely(libeth_rx_pt_get_ip_ver(decoded) ==
3285 LIBETH_RX_PT_OUTER_L2))
3286 return -EINVAL;
3287
3288 rsc_seg_len = le16_to_cpu(rx_desc->misc.rscseglen);
3289 if (unlikely(!rsc_seg_len))
3290 return -EINVAL;
3291
3292 ipv4 = libeth_rx_pt_get_ip_ver(decoded) == LIBETH_RX_PT_OUTER_IPV4;
3293 ipv6 = libeth_rx_pt_get_ip_ver(decoded) == LIBETH_RX_PT_OUTER_IPV6;
3294
3295 if (unlikely(!(ipv4 ^ ipv6)))
3296 return -EINVAL;
3297
3298 rsc_segments = DIV_ROUND_UP(skb->data_len, rsc_seg_len);
3299
3300 NAPI_GRO_CB(skb)->count = rsc_segments;
3301 skb_shinfo(skb)->gso_size = rsc_seg_len;
3302
3303 skb_reset_network_header(skb);
3304
3305 if (ipv4) {
3306 struct iphdr *ipv4h = ip_hdr(skb);
3307
3308 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV4;
3309
3310 /* Reset and set transport header offset in skb */
3311 skb_set_transport_header(skb, sizeof(struct iphdr));
3312 len = skb->len - skb_transport_offset(skb);
3313
3314 /* Compute the TCP pseudo header checksum*/
3315 tcp_hdr(skb)->check =
3316 ~tcp_v4_check(len, ipv4h->saddr, ipv4h->daddr, 0);
3317 } else {
3318 struct ipv6hdr *ipv6h = ipv6_hdr(skb);
3319
3320 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV6;
3321 skb_set_transport_header(skb, sizeof(struct ipv6hdr));
3322 len = skb->len - skb_transport_offset(skb);
3323 tcp_hdr(skb)->check =
3324 ~tcp_v6_check(len, &ipv6h->saddr, &ipv6h->daddr, 0);
3325 }
3326
3327 tcp_gro_complete(skb);
3328
3329 u64_stats_update_begin(&rxq->stats_sync);
3330 u64_stats_inc(&rxq->q_stats.rsc_pkts);
3331 u64_stats_update_end(&rxq->stats_sync);
3332
3333 return 0;
3334 }
3335
3336 /**
3337 * idpf_rx_hwtstamp - check for an RX timestamp and pass up the stack
3338 * @rxq: pointer to the rx queue that receives the timestamp
3339 * @rx_desc: pointer to rx descriptor containing timestamp
3340 * @skb: skb to put timestamp in
3341 */
3342 static void
idpf_rx_hwtstamp(const struct idpf_rx_queue * rxq,const struct virtchnl2_rx_flex_desc_adv_nic_3 * rx_desc,struct sk_buff * skb)3343 idpf_rx_hwtstamp(const struct idpf_rx_queue *rxq,
3344 const struct virtchnl2_rx_flex_desc_adv_nic_3 *rx_desc,
3345 struct sk_buff *skb)
3346 {
3347 u64 cached_time, ts_ns;
3348 u32 ts_high;
3349
3350 if (!(rx_desc->ts_low & VIRTCHNL2_RX_FLEX_TSTAMP_VALID))
3351 return;
3352
3353 cached_time = READ_ONCE(rxq->cached_phc_time);
3354
3355 ts_high = le32_to_cpu(rx_desc->ts_high);
3356 ts_ns = idpf_ptp_tstamp_extend_32b_to_64b(cached_time, ts_high);
3357
3358 *skb_hwtstamps(skb) = (struct skb_shared_hwtstamps) {
3359 .hwtstamp = ns_to_ktime(ts_ns),
3360 };
3361 }
3362
3363 /**
3364 * __idpf_rx_process_skb_fields - Populate skb header fields from Rx descriptor
3365 * @rxq: Rx descriptor ring packet is being transacted on
3366 * @skb: pointer to current skb being populated
3367 * @rx_desc: Receive descriptor
3368 *
3369 * This function checks the ring, descriptor, and packet information in
3370 * order to populate the hash, checksum, protocol, and
3371 * other fields within the skb.
3372 */
3373 static int
__idpf_rx_process_skb_fields(struct idpf_rx_queue * rxq,struct sk_buff * skb,const struct virtchnl2_rx_flex_desc_adv_nic_3 * rx_desc)3374 __idpf_rx_process_skb_fields(struct idpf_rx_queue *rxq, struct sk_buff *skb,
3375 const struct virtchnl2_rx_flex_desc_adv_nic_3 *rx_desc)
3376 {
3377 struct libeth_rx_csum csum_bits;
3378 struct libeth_rx_pt decoded;
3379 u16 rx_ptype;
3380
3381 rx_ptype = le16_get_bits(rx_desc->ptype_err_fflags0,
3382 VIRTCHNL2_RX_FLEX_DESC_ADV_PTYPE_M);
3383 decoded = rxq->rx_ptype_lkup[rx_ptype];
3384
3385 /* process RSS/hash */
3386 idpf_rx_hash(rxq, skb, rx_desc, decoded);
3387
3388 if (idpf_queue_has(PTP, rxq))
3389 idpf_rx_hwtstamp(rxq, rx_desc, skb);
3390
3391 if (le16_get_bits(rx_desc->hdrlen_flags,
3392 VIRTCHNL2_RX_FLEX_DESC_ADV_RSC_M))
3393 return idpf_rx_rsc(rxq, skb, rx_desc, decoded);
3394
3395 csum_bits = idpf_rx_splitq_extract_csum_bits(rx_desc);
3396 idpf_rx_csum(rxq, skb, csum_bits, decoded);
3397
3398 return 0;
3399 }
3400
idpf_rx_process_skb_fields(struct sk_buff * skb,const struct libeth_xdp_buff * xdp,struct libeth_rq_napi_stats * rs)3401 bool idpf_rx_process_skb_fields(struct sk_buff *skb,
3402 const struct libeth_xdp_buff *xdp,
3403 struct libeth_rq_napi_stats *rs)
3404 {
3405 struct idpf_rx_queue *rxq;
3406
3407 rxq = libeth_xdp_buff_to_rq(xdp, typeof(*rxq), xdp_rxq);
3408
3409 return !__idpf_rx_process_skb_fields(rxq, skb, xdp->desc);
3410 }
3411
3412 LIBETH_XDP_DEFINE_START();
3413 LIBETH_XDP_DEFINE_RUN(static idpf_xdp_run_pass, idpf_xdp_run_prog,
3414 idpf_xdp_tx_flush_bulk, idpf_rx_process_skb_fields);
3415 LIBETH_XDP_DEFINE_FINALIZE(static idpf_xdp_finalize_rx, idpf_xdp_tx_flush_bulk,
3416 idpf_xdp_tx_finalize);
3417 LIBETH_XDP_DEFINE_END();
3418
3419 /**
3420 * idpf_rx_hsplit_wa - handle header buffer overflows and split errors
3421 * @hdr: Rx buffer for the headers
3422 * @buf: Rx buffer for the payload
3423 * @data_len: number of bytes received to the payload buffer
3424 *
3425 * When a header buffer overflow occurs or the HW was unable do parse the
3426 * packet type to perform header split, the whole frame gets placed to the
3427 * payload buffer. We can't build a valid skb around a payload buffer when
3428 * the header split is active since it doesn't reserve any head- or tailroom.
3429 * In that case, copy either the whole frame when it's short or just the
3430 * Ethernet header to the header buffer to be able to build an skb and adjust
3431 * the data offset in the payload buffer, IOW emulate the header split.
3432 *
3433 * Return: number of bytes copied to the header buffer.
3434 */
idpf_rx_hsplit_wa(const struct libeth_fqe * hdr,struct libeth_fqe * buf,u32 data_len)3435 static u32 idpf_rx_hsplit_wa(const struct libeth_fqe *hdr,
3436 struct libeth_fqe *buf, u32 data_len)
3437 {
3438 u32 copy = data_len <= L1_CACHE_BYTES ? data_len : ETH_HLEN;
3439 struct page *hdr_page, *buf_page;
3440 const void *src;
3441 void *dst;
3442
3443 if (unlikely(netmem_is_net_iov(buf->netmem)) ||
3444 !libeth_rx_sync_for_cpu(buf, copy))
3445 return 0;
3446
3447 hdr_page = __netmem_to_page(hdr->netmem);
3448 buf_page = __netmem_to_page(buf->netmem);
3449 dst = page_address(hdr_page) + hdr->offset +
3450 pp_page_to_nmdesc(hdr_page)->pp->p.offset;
3451 src = page_address(buf_page) + buf->offset +
3452 pp_page_to_nmdesc(buf_page)->pp->p.offset;
3453
3454 memcpy(dst, src, LARGEST_ALIGN(copy));
3455 buf->offset += copy;
3456
3457 return copy;
3458 }
3459
3460 /**
3461 * idpf_rx_splitq_test_staterr - tests bits in Rx descriptor
3462 * status and error fields
3463 * @stat_err_field: field from descriptor to test bits in
3464 * @stat_err_bits: value to mask
3465 *
3466 */
idpf_rx_splitq_test_staterr(const u8 stat_err_field,const u8 stat_err_bits)3467 static bool idpf_rx_splitq_test_staterr(const u8 stat_err_field,
3468 const u8 stat_err_bits)
3469 {
3470 return !!(stat_err_field & stat_err_bits);
3471 }
3472
3473 /**
3474 * idpf_rx_splitq_is_eop - process handling of EOP buffers
3475 * @rx_desc: Rx descriptor for current buffer
3476 *
3477 * If the buffer is an EOP buffer, this function exits returning true,
3478 * otherwise return false indicating that this is in fact a non-EOP buffer.
3479 */
idpf_rx_splitq_is_eop(struct virtchnl2_rx_flex_desc_adv_nic_3 * rx_desc)3480 static bool idpf_rx_splitq_is_eop(struct virtchnl2_rx_flex_desc_adv_nic_3 *rx_desc)
3481 {
3482 /* if we are the last buffer then there is nothing else to do */
3483 return likely(idpf_rx_splitq_test_staterr(rx_desc->status_err0_qw1,
3484 IDPF_RXD_EOF_SPLITQ));
3485 }
3486
3487 /**
3488 * idpf_rx_splitq_clean - Clean completed descriptors from Rx queue
3489 * @rxq: Rx descriptor queue to retrieve receive buffer queue
3490 * @budget: Total limit on number of packets to process
3491 *
3492 * This function provides a "bounce buffer" approach to Rx interrupt
3493 * processing. The advantage to this is that on systems that have
3494 * expensive overhead for IOMMU access this provides a means of avoiding
3495 * it by maintaining the mapping of the page to the system.
3496 *
3497 * Returns amount of work completed
3498 */
idpf_rx_splitq_clean(struct idpf_rx_queue * rxq,int budget)3499 static int idpf_rx_splitq_clean(struct idpf_rx_queue *rxq, int budget)
3500 {
3501 struct idpf_buf_queue *rx_bufq = NULL;
3502 struct libeth_rq_napi_stats rs = { };
3503 u16 ntc = rxq->next_to_clean;
3504 LIBETH_XDP_ONSTACK_BUFF(xdp);
3505 LIBETH_XDP_ONSTACK_BULK(bq);
3506
3507 libeth_xdp_tx_init_bulk(&bq, rxq->xdp_prog, rxq->xdp_rxq.dev,
3508 rxq->xdpsqs, rxq->num_xdp_txq);
3509 libeth_xdp_init_buff(xdp, &rxq->xdp, &rxq->xdp_rxq);
3510
3511 /* Process Rx packets bounded by budget */
3512 while (likely(rs.packets < budget)) {
3513 struct virtchnl2_rx_flex_desc_adv_nic_3 *rx_desc;
3514 struct libeth_fqe *hdr, *rx_buf = NULL;
3515 struct idpf_sw_queue *refillq = NULL;
3516 struct idpf_rxq_set *rxq_set = NULL;
3517 unsigned int pkt_len = 0;
3518 unsigned int hdr_len = 0;
3519 u16 gen_id, buf_id = 0;
3520 int bufq_id;
3521 u8 rxdid;
3522
3523 /* get the Rx desc from Rx queue based on 'next_to_clean' */
3524 rx_desc = &rxq->rx[ntc].flex_adv_nic_3_wb;
3525
3526 /* if the descriptor isn't done, no work yet to do */
3527 gen_id = le16_get_bits(rx_desc->pktlen_gen_bufq_id,
3528 VIRTCHNL2_RX_FLEX_DESC_ADV_GEN_M);
3529 if (idpf_queue_has(GEN_CHK, rxq) != gen_id)
3530 break;
3531
3532 dma_rmb();
3533
3534 rxdid = FIELD_GET(VIRTCHNL2_RX_FLEX_DESC_ADV_RXDID_M,
3535 rx_desc->rxdid_ucast);
3536 if (rxdid != VIRTCHNL2_RXDID_2_FLEX_SPLITQ) {
3537 IDPF_RX_BUMP_NTC(rxq, ntc);
3538 u64_stats_update_begin(&rxq->stats_sync);
3539 u64_stats_inc(&rxq->q_stats.bad_descs);
3540 u64_stats_update_end(&rxq->stats_sync);
3541 continue;
3542 }
3543
3544 pkt_len = le16_get_bits(rx_desc->pktlen_gen_bufq_id,
3545 VIRTCHNL2_RX_FLEX_DESC_ADV_LEN_PBUF_M);
3546
3547 bufq_id = le16_get_bits(rx_desc->pktlen_gen_bufq_id,
3548 VIRTCHNL2_RX_FLEX_DESC_ADV_BUFQ_ID_M);
3549
3550 rxq_set = container_of(rxq, struct idpf_rxq_set, rxq);
3551 refillq = rxq_set->refillq[bufq_id];
3552
3553 /* retrieve buffer from the rxq */
3554 rx_bufq = &rxq->bufq_sets[bufq_id].bufq;
3555
3556 buf_id = le16_to_cpu(rx_desc->buf_id);
3557
3558 rx_buf = &rx_bufq->buf[buf_id];
3559
3560 if (!rx_bufq->hdr_pp)
3561 goto payload;
3562
3563 #define __HBO_BIT VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_HBO_M
3564 #define __HDR_LEN_MASK VIRTCHNL2_RX_FLEX_DESC_ADV_LEN_HDR_M
3565 if (likely(!(rx_desc->status_err0_qw1 & __HBO_BIT)))
3566 /* If a header buffer overflow, occurs, i.e. header is
3567 * too large to fit in the header split buffer, HW will
3568 * put the entire packet, including headers, in the
3569 * data/payload buffer.
3570 */
3571 hdr_len = le16_get_bits(rx_desc->hdrlen_flags,
3572 __HDR_LEN_MASK);
3573 #undef __HDR_LEN_MASK
3574 #undef __HBO_BIT
3575
3576 hdr = &rx_bufq->hdr_buf[buf_id];
3577
3578 if (unlikely(!hdr_len && !xdp->data)) {
3579 hdr_len = idpf_rx_hsplit_wa(hdr, rx_buf, pkt_len);
3580 /* If failed, drop both buffers by setting len to 0 */
3581 pkt_len -= hdr_len ? : pkt_len;
3582
3583 u64_stats_update_begin(&rxq->stats_sync);
3584 u64_stats_inc(&rxq->q_stats.hsplit_buf_ovf);
3585 u64_stats_update_end(&rxq->stats_sync);
3586 }
3587
3588 if (libeth_xdp_process_buff(xdp, hdr, hdr_len))
3589 rs.hsplit++;
3590
3591 hdr->netmem = 0;
3592
3593 payload:
3594 libeth_xdp_process_buff(xdp, rx_buf, pkt_len);
3595 rx_buf->netmem = 0;
3596
3597 idpf_post_buf_refill(refillq, buf_id);
3598 IDPF_RX_BUMP_NTC(rxq, ntc);
3599
3600 /* skip if it is non EOP desc */
3601 if (!idpf_rx_splitq_is_eop(rx_desc) || unlikely(!xdp->data))
3602 continue;
3603
3604 idpf_xdp_run_pass(xdp, &bq, rxq->napi, &rs, rx_desc);
3605 }
3606
3607 idpf_xdp_finalize_rx(&bq);
3608
3609 rxq->next_to_clean = ntc;
3610 libeth_xdp_save_buff(&rxq->xdp, xdp);
3611
3612 u64_stats_update_begin(&rxq->stats_sync);
3613 u64_stats_add(&rxq->q_stats.packets, rs.packets);
3614 u64_stats_add(&rxq->q_stats.bytes, rs.bytes);
3615 u64_stats_add(&rxq->q_stats.hsplit_pkts, rs.hsplit);
3616 u64_stats_update_end(&rxq->stats_sync);
3617
3618 return rs.packets;
3619 }
3620
3621 /**
3622 * idpf_rx_update_bufq_desc - Update buffer queue descriptor
3623 * @bufq: Pointer to the buffer queue
3624 * @buf_id: buffer ID
3625 * @buf_desc: Buffer queue descriptor
3626 *
3627 * Return 0 on success and negative on failure.
3628 */
idpf_rx_update_bufq_desc(struct idpf_buf_queue * bufq,u32 buf_id,struct virtchnl2_splitq_rx_buf_desc * buf_desc)3629 static int idpf_rx_update_bufq_desc(struct idpf_buf_queue *bufq, u32 buf_id,
3630 struct virtchnl2_splitq_rx_buf_desc *buf_desc)
3631 {
3632 struct libeth_fq_fp fq = {
3633 .pp = bufq->pp,
3634 .fqes = bufq->buf,
3635 .truesize = bufq->truesize,
3636 .count = bufq->desc_count,
3637 };
3638 dma_addr_t addr;
3639
3640 addr = libeth_rx_alloc(&fq, buf_id);
3641 if (addr == DMA_MAPPING_ERROR)
3642 return -ENOMEM;
3643
3644 buf_desc->pkt_addr = cpu_to_le64(addr);
3645 buf_desc->qword0.buf_id = cpu_to_le16(buf_id);
3646
3647 if (!idpf_queue_has(HSPLIT_EN, bufq))
3648 return 0;
3649
3650 fq.pp = bufq->hdr_pp;
3651 fq.fqes = bufq->hdr_buf;
3652 fq.truesize = bufq->hdr_truesize;
3653
3654 addr = libeth_rx_alloc(&fq, buf_id);
3655 if (addr == DMA_MAPPING_ERROR)
3656 return -ENOMEM;
3657
3658 buf_desc->hdr_addr = cpu_to_le64(addr);
3659
3660 return 0;
3661 }
3662
3663 /**
3664 * idpf_rx_clean_refillq - Clean refill queue buffers
3665 * @bufq: buffer queue to post buffers back to
3666 * @refillq: refill queue to clean
3667 *
3668 * This function takes care of the buffer refill management
3669 */
idpf_rx_clean_refillq(struct idpf_buf_queue * bufq,struct idpf_sw_queue * refillq)3670 static void idpf_rx_clean_refillq(struct idpf_buf_queue *bufq,
3671 struct idpf_sw_queue *refillq)
3672 {
3673 struct virtchnl2_splitq_rx_buf_desc *buf_desc;
3674 u16 bufq_nta = bufq->next_to_alloc;
3675 u16 ntc = refillq->next_to_clean;
3676 int cleaned = 0;
3677
3678 buf_desc = &bufq->split_buf[bufq_nta];
3679
3680 /* make sure we stop at ring wrap in the unlikely case ring is full */
3681 while (likely(cleaned < refillq->desc_count)) {
3682 u32 buf_id, refill_desc = refillq->ring[ntc];
3683 bool failure;
3684
3685 if (idpf_queue_has(RFL_GEN_CHK, refillq) !=
3686 !!(refill_desc & IDPF_RFL_BI_GEN_M))
3687 break;
3688
3689 buf_id = FIELD_GET(IDPF_RFL_BI_BUFID_M, refill_desc);
3690 failure = idpf_rx_update_bufq_desc(bufq, buf_id, buf_desc);
3691 if (failure)
3692 break;
3693
3694 if (unlikely(++ntc == refillq->desc_count)) {
3695 idpf_queue_change(RFL_GEN_CHK, refillq);
3696 ntc = 0;
3697 }
3698
3699 if (unlikely(++bufq_nta == bufq->desc_count)) {
3700 buf_desc = &bufq->split_buf[0];
3701 bufq_nta = 0;
3702 } else {
3703 buf_desc++;
3704 }
3705
3706 cleaned++;
3707 }
3708
3709 if (!cleaned)
3710 return;
3711
3712 /* We want to limit how many transactions on the bus we trigger with
3713 * tail writes so we only do it in strides. It's also important we
3714 * align the write to a multiple of 8 as required by HW.
3715 */
3716 if (((bufq->next_to_use <= bufq_nta ? 0 : bufq->desc_count) +
3717 bufq_nta - bufq->next_to_use) >= IDPF_RX_BUF_POST_STRIDE)
3718 idpf_rx_buf_hw_update(bufq, ALIGN_DOWN(bufq_nta,
3719 IDPF_RX_BUF_POST_STRIDE));
3720
3721 /* update next to alloc since we have filled the ring */
3722 refillq->next_to_clean = ntc;
3723 bufq->next_to_alloc = bufq_nta;
3724 }
3725
3726 /**
3727 * idpf_rx_clean_refillq_all - Clean all refill queues
3728 * @bufq: buffer queue with refill queues
3729 * @nid: ID of the closest NUMA node with memory
3730 *
3731 * Iterates through all refill queues assigned to the buffer queue assigned to
3732 * this vector. Returns true if clean is complete within budget, false
3733 * otherwise.
3734 */
idpf_rx_clean_refillq_all(struct idpf_buf_queue * bufq,int nid)3735 static void idpf_rx_clean_refillq_all(struct idpf_buf_queue *bufq, int nid)
3736 {
3737 struct idpf_bufq_set *bufq_set;
3738 int i;
3739
3740 page_pool_nid_changed(bufq->pp, nid);
3741 if (bufq->hdr_pp)
3742 page_pool_nid_changed(bufq->hdr_pp, nid);
3743
3744 bufq_set = container_of(bufq, struct idpf_bufq_set, bufq);
3745 for (i = 0; i < bufq_set->num_refillqs; i++)
3746 idpf_rx_clean_refillq(bufq, &bufq_set->refillqs[i]);
3747 }
3748
3749 /**
3750 * idpf_vport_intr_clean_queues - MSIX mode Interrupt Handler
3751 * @irq: interrupt number
3752 * @data: pointer to a q_vector
3753 *
3754 */
idpf_vport_intr_clean_queues(int __always_unused irq,void * data)3755 static irqreturn_t idpf_vport_intr_clean_queues(int __always_unused irq,
3756 void *data)
3757 {
3758 struct idpf_q_vector *q_vector = (struct idpf_q_vector *)data;
3759
3760 q_vector->total_events++;
3761 napi_schedule_irqoff(&q_vector->napi);
3762
3763 return IRQ_HANDLED;
3764 }
3765
3766 /**
3767 * idpf_vport_intr_napi_del_all - Unregister napi for all q_vectors in vport
3768 * @vport: virtual port structure
3769 *
3770 */
idpf_vport_intr_napi_del_all(struct idpf_vport * vport)3771 static void idpf_vport_intr_napi_del_all(struct idpf_vport *vport)
3772 {
3773 u16 v_idx;
3774
3775 for (v_idx = 0; v_idx < vport->num_q_vectors; v_idx++)
3776 netif_napi_del(&vport->q_vectors[v_idx].napi);
3777 }
3778
3779 /**
3780 * idpf_vport_intr_napi_dis_all - Disable NAPI for all q_vectors in the vport
3781 * @vport: main vport structure
3782 */
idpf_vport_intr_napi_dis_all(struct idpf_vport * vport)3783 static void idpf_vport_intr_napi_dis_all(struct idpf_vport *vport)
3784 {
3785 int v_idx;
3786
3787 for (v_idx = 0; v_idx < vport->num_q_vectors; v_idx++)
3788 napi_disable(&vport->q_vectors[v_idx].napi);
3789 }
3790
3791 /**
3792 * idpf_vport_intr_rel - Free memory allocated for interrupt vectors
3793 * @vport: virtual port
3794 *
3795 * Free the memory allocated for interrupt vectors associated to a vport
3796 */
idpf_vport_intr_rel(struct idpf_vport * vport)3797 void idpf_vport_intr_rel(struct idpf_vport *vport)
3798 {
3799 for (u32 v_idx = 0; v_idx < vport->num_q_vectors; v_idx++) {
3800 struct idpf_q_vector *q_vector = &vport->q_vectors[v_idx];
3801
3802 kfree(q_vector->xsksq);
3803 q_vector->xsksq = NULL;
3804 kfree(q_vector->complq);
3805 q_vector->complq = NULL;
3806 kfree(q_vector->bufq);
3807 q_vector->bufq = NULL;
3808 kfree(q_vector->tx);
3809 q_vector->tx = NULL;
3810 kfree(q_vector->rx);
3811 q_vector->rx = NULL;
3812 }
3813
3814 kfree(vport->q_vectors);
3815 vport->q_vectors = NULL;
3816 }
3817
idpf_q_vector_set_napi(struct idpf_q_vector * q_vector,bool link)3818 static void idpf_q_vector_set_napi(struct idpf_q_vector *q_vector, bool link)
3819 {
3820 struct napi_struct *napi = link ? &q_vector->napi : NULL;
3821 struct net_device *dev = q_vector->vport->netdev;
3822
3823 for (u32 i = 0; i < q_vector->num_rxq; i++)
3824 netif_queue_set_napi(dev, q_vector->rx[i]->idx,
3825 NETDEV_QUEUE_TYPE_RX, napi);
3826
3827 for (u32 i = 0; i < q_vector->num_txq; i++)
3828 netif_queue_set_napi(dev, q_vector->tx[i]->idx,
3829 NETDEV_QUEUE_TYPE_TX, napi);
3830 }
3831
3832 /**
3833 * idpf_vport_intr_rel_irq - Free the IRQ association with the OS
3834 * @vport: main vport structure
3835 */
idpf_vport_intr_rel_irq(struct idpf_vport * vport)3836 static void idpf_vport_intr_rel_irq(struct idpf_vport *vport)
3837 {
3838 struct idpf_adapter *adapter = vport->adapter;
3839 int vector;
3840
3841 for (vector = 0; vector < vport->num_q_vectors; vector++) {
3842 struct idpf_q_vector *q_vector = &vport->q_vectors[vector];
3843 int irq_num, vidx;
3844
3845 /* free only the irqs that were actually requested */
3846 if (!q_vector)
3847 continue;
3848
3849 vidx = vport->q_vector_idxs[vector];
3850 irq_num = adapter->msix_entries[vidx].vector;
3851
3852 idpf_q_vector_set_napi(q_vector, false);
3853 kfree(free_irq(irq_num, q_vector));
3854 }
3855 }
3856
3857 /**
3858 * idpf_vport_intr_dis_irq_all - Disable all interrupt
3859 * @vport: main vport structure
3860 */
idpf_vport_intr_dis_irq_all(struct idpf_vport * vport)3861 static void idpf_vport_intr_dis_irq_all(struct idpf_vport *vport)
3862 {
3863 struct idpf_q_vector *q_vector = vport->q_vectors;
3864 int q_idx;
3865
3866 writel(0, vport->noirq_dyn_ctl);
3867
3868 for (q_idx = 0; q_idx < vport->num_q_vectors; q_idx++)
3869 writel(0, q_vector[q_idx].intr_reg.dyn_ctl);
3870 }
3871
3872 /**
3873 * idpf_vport_intr_buildreg_itr - Enable default interrupt generation settings
3874 * @q_vector: pointer to q_vector
3875 */
idpf_vport_intr_buildreg_itr(struct idpf_q_vector * q_vector)3876 static u32 idpf_vport_intr_buildreg_itr(struct idpf_q_vector *q_vector)
3877 {
3878 u32 itr_val = q_vector->intr_reg.dyn_ctl_intena_m;
3879 int type = IDPF_NO_ITR_UPDATE_IDX;
3880 u16 itr = 0;
3881
3882 if (q_vector->wb_on_itr) {
3883 /*
3884 * Trigger a software interrupt when exiting wb_on_itr, to make
3885 * sure we catch any pending write backs that might have been
3886 * missed due to interrupt state transition.
3887 */
3888 itr_val |= q_vector->intr_reg.dyn_ctl_swint_trig_m |
3889 q_vector->intr_reg.dyn_ctl_sw_itridx_ena_m;
3890 type = IDPF_SW_ITR_UPDATE_IDX;
3891 itr = IDPF_ITR_20K;
3892 }
3893
3894 itr &= IDPF_ITR_MASK;
3895 /* Don't clear PBA because that can cause lost interrupts that
3896 * came in while we were cleaning/polling
3897 */
3898 itr_val |= (type << q_vector->intr_reg.dyn_ctl_itridx_s) |
3899 (itr << (q_vector->intr_reg.dyn_ctl_intrvl_s - 1));
3900
3901 return itr_val;
3902 }
3903
3904 /**
3905 * idpf_update_dim_sample - Update dim sample with packets and bytes
3906 * @q_vector: the vector associated with the interrupt
3907 * @dim_sample: dim sample to update
3908 * @dim: dim instance structure
3909 * @packets: total packets
3910 * @bytes: total bytes
3911 *
3912 * Update the dim sample with the packets and bytes which are passed to this
3913 * function. Set the dim state appropriately if the dim settings gets stale.
3914 */
idpf_update_dim_sample(struct idpf_q_vector * q_vector,struct dim_sample * dim_sample,struct dim * dim,u64 packets,u64 bytes)3915 static void idpf_update_dim_sample(struct idpf_q_vector *q_vector,
3916 struct dim_sample *dim_sample,
3917 struct dim *dim, u64 packets, u64 bytes)
3918 {
3919 dim_update_sample(q_vector->total_events, packets, bytes, dim_sample);
3920 dim_sample->comp_ctr = 0;
3921
3922 /* if dim settings get stale, like when not updated for 1 second or
3923 * longer, force it to start again. This addresses the frequent case
3924 * of an idle queue being switched to by the scheduler.
3925 */
3926 if (ktime_ms_delta(dim_sample->time, dim->start_sample.time) >= HZ)
3927 dim->state = DIM_START_MEASURE;
3928 }
3929
3930 /**
3931 * idpf_net_dim - Update net DIM algorithm
3932 * @q_vector: the vector associated with the interrupt
3933 *
3934 * Create a DIM sample and notify net_dim() so that it can possibly decide
3935 * a new ITR value based on incoming packets, bytes, and interrupts.
3936 *
3937 * This function is a no-op if the queue is not configured to dynamic ITR.
3938 */
idpf_net_dim(struct idpf_q_vector * q_vector)3939 static void idpf_net_dim(struct idpf_q_vector *q_vector)
3940 {
3941 struct dim_sample dim_sample = { };
3942 u64 packets, bytes;
3943 u32 i;
3944
3945 if (!IDPF_ITR_IS_DYNAMIC(q_vector->tx_intr_mode))
3946 goto check_rx_itr;
3947
3948 for (i = 0, packets = 0, bytes = 0; i < q_vector->num_txq; i++) {
3949 struct idpf_tx_queue *txq = q_vector->tx[i];
3950 unsigned int start;
3951
3952 do {
3953 start = u64_stats_fetch_begin(&txq->stats_sync);
3954 packets += u64_stats_read(&txq->q_stats.packets);
3955 bytes += u64_stats_read(&txq->q_stats.bytes);
3956 } while (u64_stats_fetch_retry(&txq->stats_sync, start));
3957 }
3958
3959 idpf_update_dim_sample(q_vector, &dim_sample, &q_vector->tx_dim,
3960 packets, bytes);
3961 net_dim(&q_vector->tx_dim, &dim_sample);
3962
3963 check_rx_itr:
3964 if (!IDPF_ITR_IS_DYNAMIC(q_vector->rx_intr_mode))
3965 return;
3966
3967 for (i = 0, packets = 0, bytes = 0; i < q_vector->num_rxq; i++) {
3968 struct idpf_rx_queue *rxq = q_vector->rx[i];
3969 unsigned int start;
3970
3971 do {
3972 start = u64_stats_fetch_begin(&rxq->stats_sync);
3973 packets += u64_stats_read(&rxq->q_stats.packets);
3974 bytes += u64_stats_read(&rxq->q_stats.bytes);
3975 } while (u64_stats_fetch_retry(&rxq->stats_sync, start));
3976 }
3977
3978 idpf_update_dim_sample(q_vector, &dim_sample, &q_vector->rx_dim,
3979 packets, bytes);
3980 net_dim(&q_vector->rx_dim, &dim_sample);
3981 }
3982
3983 /**
3984 * idpf_vport_intr_update_itr_ena_irq - Update itr and re-enable MSIX interrupt
3985 * @q_vector: q_vector for which itr is being updated and interrupt enabled
3986 *
3987 * Update the net_dim() algorithm and re-enable the interrupt associated with
3988 * this vector.
3989 */
idpf_vport_intr_update_itr_ena_irq(struct idpf_q_vector * q_vector)3990 void idpf_vport_intr_update_itr_ena_irq(struct idpf_q_vector *q_vector)
3991 {
3992 u32 intval;
3993
3994 /* net_dim() updates ITR out-of-band using a work item */
3995 idpf_net_dim(q_vector);
3996
3997 intval = idpf_vport_intr_buildreg_itr(q_vector);
3998 q_vector->wb_on_itr = false;
3999
4000 writel(intval, q_vector->intr_reg.dyn_ctl);
4001 }
4002
4003 /**
4004 * idpf_vport_intr_req_irq - get MSI-X vectors from the OS for the vport
4005 * @vport: main vport structure
4006 */
idpf_vport_intr_req_irq(struct idpf_vport * vport)4007 static int idpf_vport_intr_req_irq(struct idpf_vport *vport)
4008 {
4009 struct idpf_adapter *adapter = vport->adapter;
4010 const char *drv_name, *if_name, *vec_name;
4011 int vector, err, irq_num, vidx;
4012
4013 drv_name = dev_driver_string(&adapter->pdev->dev);
4014 if_name = netdev_name(vport->netdev);
4015
4016 for (vector = 0; vector < vport->num_q_vectors; vector++) {
4017 struct idpf_q_vector *q_vector = &vport->q_vectors[vector];
4018 char *name;
4019
4020 vidx = vport->q_vector_idxs[vector];
4021 irq_num = adapter->msix_entries[vidx].vector;
4022
4023 if (q_vector->num_rxq && q_vector->num_txq)
4024 vec_name = "TxRx";
4025 else if (q_vector->num_rxq)
4026 vec_name = "Rx";
4027 else if (q_vector->num_txq)
4028 vec_name = "Tx";
4029 else
4030 continue;
4031
4032 name = kasprintf(GFP_KERNEL, "%s-%s-%s-%d", drv_name, if_name,
4033 vec_name, vidx);
4034
4035 err = request_irq(irq_num, idpf_vport_intr_clean_queues, 0,
4036 name, q_vector);
4037 if (err) {
4038 netdev_err(vport->netdev,
4039 "Request_irq failed, error: %d\n", err);
4040 goto free_q_irqs;
4041 }
4042
4043 idpf_q_vector_set_napi(q_vector, true);
4044 }
4045
4046 return 0;
4047
4048 free_q_irqs:
4049 while (--vector >= 0) {
4050 vidx = vport->q_vector_idxs[vector];
4051 irq_num = adapter->msix_entries[vidx].vector;
4052 kfree(free_irq(irq_num, &vport->q_vectors[vector]));
4053 }
4054
4055 return err;
4056 }
4057
4058 /**
4059 * idpf_vport_intr_write_itr - Write ITR value to the ITR register
4060 * @q_vector: q_vector structure
4061 * @itr: Interrupt throttling rate
4062 * @tx: Tx or Rx ITR
4063 */
idpf_vport_intr_write_itr(struct idpf_q_vector * q_vector,u16 itr,bool tx)4064 void idpf_vport_intr_write_itr(struct idpf_q_vector *q_vector, u16 itr, bool tx)
4065 {
4066 struct idpf_intr_reg *intr_reg;
4067
4068 if (tx && !q_vector->tx)
4069 return;
4070 else if (!tx && !q_vector->rx)
4071 return;
4072
4073 intr_reg = &q_vector->intr_reg;
4074 writel(ITR_REG_ALIGN(itr) >> IDPF_ITR_GRAN_S,
4075 tx ? intr_reg->tx_itr : intr_reg->rx_itr);
4076 }
4077
4078 /**
4079 * idpf_vport_intr_ena_irq_all - Enable IRQ for the given vport
4080 * @vport: main vport structure
4081 */
idpf_vport_intr_ena_irq_all(struct idpf_vport * vport)4082 static void idpf_vport_intr_ena_irq_all(struct idpf_vport *vport)
4083 {
4084 bool dynamic;
4085 int q_idx;
4086 u16 itr;
4087
4088 for (q_idx = 0; q_idx < vport->num_q_vectors; q_idx++) {
4089 struct idpf_q_vector *qv = &vport->q_vectors[q_idx];
4090
4091 /* Set the initial ITR values */
4092 if (qv->num_txq) {
4093 dynamic = IDPF_ITR_IS_DYNAMIC(qv->tx_intr_mode);
4094 itr = vport->tx_itr_profile[qv->tx_dim.profile_ix];
4095 idpf_vport_intr_write_itr(qv, dynamic ?
4096 itr : qv->tx_itr_value,
4097 true);
4098 }
4099
4100 if (qv->num_rxq) {
4101 dynamic = IDPF_ITR_IS_DYNAMIC(qv->rx_intr_mode);
4102 itr = vport->rx_itr_profile[qv->rx_dim.profile_ix];
4103 idpf_vport_intr_write_itr(qv, dynamic ?
4104 itr : qv->rx_itr_value,
4105 false);
4106 }
4107
4108 if (qv->num_txq || qv->num_rxq)
4109 idpf_vport_intr_update_itr_ena_irq(qv);
4110 }
4111
4112 writel(vport->noirq_dyn_ctl_ena, vport->noirq_dyn_ctl);
4113 }
4114
4115 /**
4116 * idpf_vport_intr_deinit - Release all vector associations for the vport
4117 * @vport: main vport structure
4118 */
idpf_vport_intr_deinit(struct idpf_vport * vport)4119 void idpf_vport_intr_deinit(struct idpf_vport *vport)
4120 {
4121 idpf_vport_intr_dis_irq_all(vport);
4122 idpf_vport_intr_napi_dis_all(vport);
4123 idpf_vport_intr_napi_del_all(vport);
4124 idpf_vport_intr_rel_irq(vport);
4125 }
4126
4127 /**
4128 * idpf_tx_dim_work - Call back from the stack
4129 * @work: work queue structure
4130 */
idpf_tx_dim_work(struct work_struct * work)4131 static void idpf_tx_dim_work(struct work_struct *work)
4132 {
4133 struct idpf_q_vector *q_vector;
4134 struct idpf_vport *vport;
4135 struct dim *dim;
4136 u16 itr;
4137
4138 dim = container_of(work, struct dim, work);
4139 q_vector = container_of(dim, struct idpf_q_vector, tx_dim);
4140 vport = q_vector->vport;
4141
4142 if (dim->profile_ix >= ARRAY_SIZE(vport->tx_itr_profile))
4143 dim->profile_ix = ARRAY_SIZE(vport->tx_itr_profile) - 1;
4144
4145 /* look up the values in our local table */
4146 itr = vport->tx_itr_profile[dim->profile_ix];
4147
4148 idpf_vport_intr_write_itr(q_vector, itr, true);
4149
4150 dim->state = DIM_START_MEASURE;
4151 }
4152
4153 /**
4154 * idpf_rx_dim_work - Call back from the stack
4155 * @work: work queue structure
4156 */
idpf_rx_dim_work(struct work_struct * work)4157 static void idpf_rx_dim_work(struct work_struct *work)
4158 {
4159 struct idpf_q_vector *q_vector;
4160 struct idpf_vport *vport;
4161 struct dim *dim;
4162 u16 itr;
4163
4164 dim = container_of(work, struct dim, work);
4165 q_vector = container_of(dim, struct idpf_q_vector, rx_dim);
4166 vport = q_vector->vport;
4167
4168 if (dim->profile_ix >= ARRAY_SIZE(vport->rx_itr_profile))
4169 dim->profile_ix = ARRAY_SIZE(vport->rx_itr_profile) - 1;
4170
4171 /* look up the values in our local table */
4172 itr = vport->rx_itr_profile[dim->profile_ix];
4173
4174 idpf_vport_intr_write_itr(q_vector, itr, false);
4175
4176 dim->state = DIM_START_MEASURE;
4177 }
4178
4179 /**
4180 * idpf_init_dim - Set up dynamic interrupt moderation
4181 * @qv: q_vector structure
4182 */
idpf_init_dim(struct idpf_q_vector * qv)4183 static void idpf_init_dim(struct idpf_q_vector *qv)
4184 {
4185 INIT_WORK(&qv->tx_dim.work, idpf_tx_dim_work);
4186 qv->tx_dim.mode = DIM_CQ_PERIOD_MODE_START_FROM_EQE;
4187 qv->tx_dim.profile_ix = IDPF_DIM_DEFAULT_PROFILE_IX;
4188
4189 INIT_WORK(&qv->rx_dim.work, idpf_rx_dim_work);
4190 qv->rx_dim.mode = DIM_CQ_PERIOD_MODE_START_FROM_EQE;
4191 qv->rx_dim.profile_ix = IDPF_DIM_DEFAULT_PROFILE_IX;
4192 }
4193
4194 /**
4195 * idpf_vport_intr_napi_ena_all - Enable NAPI for all q_vectors in the vport
4196 * @vport: main vport structure
4197 */
idpf_vport_intr_napi_ena_all(struct idpf_vport * vport)4198 static void idpf_vport_intr_napi_ena_all(struct idpf_vport *vport)
4199 {
4200 int q_idx;
4201
4202 for (q_idx = 0; q_idx < vport->num_q_vectors; q_idx++) {
4203 struct idpf_q_vector *q_vector = &vport->q_vectors[q_idx];
4204
4205 idpf_init_dim(q_vector);
4206 napi_enable(&q_vector->napi);
4207 }
4208 }
4209
4210 /**
4211 * idpf_tx_splitq_clean_all- Clean completion queues
4212 * @q_vec: queue vector
4213 * @budget: Used to determine if we are in netpoll
4214 * @cleaned: returns number of packets cleaned
4215 *
4216 * Returns false if clean is not complete else returns true
4217 */
idpf_tx_splitq_clean_all(struct idpf_q_vector * q_vec,int budget,int * cleaned)4218 static bool idpf_tx_splitq_clean_all(struct idpf_q_vector *q_vec,
4219 int budget, int *cleaned)
4220 {
4221 u16 num_complq = q_vec->num_complq;
4222 bool clean_complete = true;
4223 int i, budget_per_q;
4224
4225 if (unlikely(!num_complq))
4226 return true;
4227
4228 budget_per_q = DIV_ROUND_UP(budget, num_complq);
4229
4230 for (i = 0; i < num_complq; i++)
4231 clean_complete &= idpf_tx_clean_complq(q_vec->complq[i],
4232 budget_per_q, cleaned);
4233
4234 return clean_complete;
4235 }
4236
4237 /**
4238 * idpf_rx_splitq_clean_all- Clean completion queues
4239 * @q_vec: queue vector
4240 * @budget: Used to determine if we are in netpoll
4241 * @cleaned: returns number of packets cleaned
4242 *
4243 * Returns false if clean is not complete else returns true
4244 */
idpf_rx_splitq_clean_all(struct idpf_q_vector * q_vec,int budget,int * cleaned)4245 static bool idpf_rx_splitq_clean_all(struct idpf_q_vector *q_vec, int budget,
4246 int *cleaned)
4247 {
4248 u16 num_rxq = q_vec->num_rxq;
4249 bool clean_complete = true;
4250 int pkts_cleaned = 0;
4251 int i, budget_per_q;
4252 int nid;
4253
4254 /* We attempt to distribute budget to each Rx queue fairly, but don't
4255 * allow the budget to go below 1 because that would exit polling early.
4256 */
4257 budget_per_q = num_rxq ? max(budget / num_rxq, 1) : 0;
4258 for (i = 0; i < num_rxq; i++) {
4259 struct idpf_rx_queue *rxq = q_vec->rx[i];
4260 int pkts_cleaned_per_q;
4261
4262 pkts_cleaned_per_q = idpf_queue_has(XSK, rxq) ?
4263 idpf_xskrq_poll(rxq, budget_per_q) :
4264 idpf_rx_splitq_clean(rxq, budget_per_q);
4265 /* if we clean as many as budgeted, we must not be done */
4266 if (pkts_cleaned_per_q >= budget_per_q)
4267 clean_complete = false;
4268 pkts_cleaned += pkts_cleaned_per_q;
4269 }
4270 *cleaned = pkts_cleaned;
4271
4272 nid = numa_mem_id();
4273
4274 for (i = 0; i < q_vec->num_bufq; i++) {
4275 if (!idpf_queue_has(XSK, q_vec->bufq[i]))
4276 idpf_rx_clean_refillq_all(q_vec->bufq[i], nid);
4277 }
4278
4279 return clean_complete;
4280 }
4281
4282 /**
4283 * idpf_vport_splitq_napi_poll - NAPI handler
4284 * @napi: struct from which you get q_vector
4285 * @budget: budget provided by stack
4286 */
idpf_vport_splitq_napi_poll(struct napi_struct * napi,int budget)4287 static int idpf_vport_splitq_napi_poll(struct napi_struct *napi, int budget)
4288 {
4289 struct idpf_q_vector *q_vector =
4290 container_of(napi, struct idpf_q_vector, napi);
4291 bool clean_complete = true;
4292 int work_done = 0;
4293
4294 /* Handle case where we are called by netpoll with a budget of 0 */
4295 if (unlikely(!budget)) {
4296 idpf_tx_splitq_clean_all(q_vector, budget, &work_done);
4297
4298 return 0;
4299 }
4300
4301 for (u32 i = 0; i < q_vector->num_xsksq; i++)
4302 clean_complete &= idpf_xsk_xmit(q_vector->xsksq[i]);
4303
4304 clean_complete &= idpf_tx_splitq_clean_all(q_vector, budget,
4305 &work_done);
4306 clean_complete &= idpf_rx_splitq_clean_all(q_vector, budget,
4307 &work_done);
4308
4309 /* If work not completed, return budget and polling will return */
4310 if (!clean_complete) {
4311 idpf_vport_intr_set_wb_on_itr(q_vector);
4312 return budget;
4313 }
4314
4315 work_done = min_t(int, work_done, budget - 1);
4316
4317 /* Exit the polling mode, but don't re-enable interrupts if stack might
4318 * poll us due to busy-polling
4319 */
4320 if (napi_complete_done(napi, work_done))
4321 idpf_vport_intr_update_itr_ena_irq(q_vector);
4322 else
4323 idpf_vport_intr_set_wb_on_itr(q_vector);
4324
4325 return work_done;
4326 }
4327
4328 /**
4329 * idpf_vport_intr_map_vector_to_qs - Map vectors to queues
4330 * @vport: virtual port
4331 *
4332 * Mapping for vectors to queues
4333 */
idpf_vport_intr_map_vector_to_qs(struct idpf_vport * vport)4334 static void idpf_vport_intr_map_vector_to_qs(struct idpf_vport *vport)
4335 {
4336 u16 num_txq_grp = vport->num_txq_grp - vport->num_xdp_txq;
4337 bool split = idpf_is_queue_model_split(vport->rxq_model);
4338 struct idpf_rxq_group *rx_qgrp;
4339 struct idpf_txq_group *tx_qgrp;
4340 u32 i, qv_idx, q_index;
4341
4342 for (i = 0, qv_idx = 0; i < vport->num_rxq_grp; i++) {
4343 u16 num_rxq;
4344
4345 if (qv_idx >= vport->num_q_vectors)
4346 qv_idx = 0;
4347
4348 rx_qgrp = &vport->rxq_grps[i];
4349 if (split)
4350 num_rxq = rx_qgrp->splitq.num_rxq_sets;
4351 else
4352 num_rxq = rx_qgrp->singleq.num_rxq;
4353
4354 for (u32 j = 0; j < num_rxq; j++) {
4355 struct idpf_rx_queue *q;
4356
4357 if (split)
4358 q = &rx_qgrp->splitq.rxq_sets[j]->rxq;
4359 else
4360 q = rx_qgrp->singleq.rxqs[j];
4361 q->q_vector = &vport->q_vectors[qv_idx];
4362 q_index = q->q_vector->num_rxq;
4363 q->q_vector->rx[q_index] = q;
4364 q->q_vector->num_rxq++;
4365
4366 if (split)
4367 q->napi = &q->q_vector->napi;
4368 }
4369
4370 if (split) {
4371 for (u32 j = 0; j < vport->num_bufqs_per_qgrp; j++) {
4372 struct idpf_buf_queue *bufq;
4373
4374 bufq = &rx_qgrp->splitq.bufq_sets[j].bufq;
4375 bufq->q_vector = &vport->q_vectors[qv_idx];
4376 q_index = bufq->q_vector->num_bufq;
4377 bufq->q_vector->bufq[q_index] = bufq;
4378 bufq->q_vector->num_bufq++;
4379 }
4380 }
4381
4382 qv_idx++;
4383 }
4384
4385 split = idpf_is_queue_model_split(vport->txq_model);
4386
4387 for (i = 0, qv_idx = 0; i < num_txq_grp; i++) {
4388 u16 num_txq;
4389
4390 if (qv_idx >= vport->num_q_vectors)
4391 qv_idx = 0;
4392
4393 tx_qgrp = &vport->txq_grps[i];
4394 num_txq = tx_qgrp->num_txq;
4395
4396 for (u32 j = 0; j < num_txq; j++) {
4397 struct idpf_tx_queue *q;
4398
4399 q = tx_qgrp->txqs[j];
4400 q->q_vector = &vport->q_vectors[qv_idx];
4401 q->q_vector->tx[q->q_vector->num_txq++] = q;
4402 }
4403
4404 if (split) {
4405 struct idpf_compl_queue *q = tx_qgrp->complq;
4406
4407 q->q_vector = &vport->q_vectors[qv_idx];
4408 q->q_vector->complq[q->q_vector->num_complq++] = q;
4409 }
4410
4411 qv_idx++;
4412 }
4413
4414 for (i = 0; i < vport->num_xdp_txq; i++) {
4415 struct idpf_tx_queue *xdpsq;
4416 struct idpf_q_vector *qv;
4417
4418 xdpsq = vport->txqs[vport->xdp_txq_offset + i];
4419 if (!idpf_queue_has(XSK, xdpsq))
4420 continue;
4421
4422 qv = idpf_find_rxq_vec(vport, i);
4423 idpf_xsk_init_wakeup(qv);
4424
4425 xdpsq->q_vector = qv;
4426 qv->xsksq[qv->num_xsksq++] = xdpsq;
4427 }
4428 }
4429
4430 /**
4431 * idpf_vport_intr_init_vec_idx - Initialize the vector indexes
4432 * @vport: virtual port
4433 *
4434 * Initialize vector indexes with values returened over mailbox
4435 */
idpf_vport_intr_init_vec_idx(struct idpf_vport * vport)4436 static int idpf_vport_intr_init_vec_idx(struct idpf_vport *vport)
4437 {
4438 struct idpf_adapter *adapter = vport->adapter;
4439 struct virtchnl2_alloc_vectors *ac;
4440 u16 *vecids, total_vecs;
4441 int i;
4442
4443 ac = adapter->req_vec_chunks;
4444 if (!ac) {
4445 for (i = 0; i < vport->num_q_vectors; i++)
4446 vport->q_vectors[i].v_idx = vport->q_vector_idxs[i];
4447
4448 vport->noirq_v_idx = vport->q_vector_idxs[i];
4449
4450 return 0;
4451 }
4452
4453 total_vecs = idpf_get_reserved_vecs(adapter);
4454 vecids = kcalloc(total_vecs, sizeof(u16), GFP_KERNEL);
4455 if (!vecids)
4456 return -ENOMEM;
4457
4458 idpf_get_vec_ids(adapter, vecids, total_vecs, &ac->vchunks);
4459
4460 for (i = 0; i < vport->num_q_vectors; i++)
4461 vport->q_vectors[i].v_idx = vecids[vport->q_vector_idxs[i]];
4462
4463 vport->noirq_v_idx = vecids[vport->q_vector_idxs[i]];
4464
4465 kfree(vecids);
4466
4467 return 0;
4468 }
4469
4470 /**
4471 * idpf_vport_intr_napi_add_all- Register napi handler for all qvectors
4472 * @vport: virtual port structure
4473 */
idpf_vport_intr_napi_add_all(struct idpf_vport * vport)4474 static void idpf_vport_intr_napi_add_all(struct idpf_vport *vport)
4475 {
4476 int (*napi_poll)(struct napi_struct *napi, int budget);
4477 u16 v_idx, qv_idx;
4478 int irq_num;
4479
4480 if (idpf_is_queue_model_split(vport->txq_model))
4481 napi_poll = idpf_vport_splitq_napi_poll;
4482 else
4483 napi_poll = idpf_vport_singleq_napi_poll;
4484
4485 for (v_idx = 0; v_idx < vport->num_q_vectors; v_idx++) {
4486 struct idpf_q_vector *q_vector = &vport->q_vectors[v_idx];
4487 qv_idx = vport->q_vector_idxs[v_idx];
4488 irq_num = vport->adapter->msix_entries[qv_idx].vector;
4489
4490 netif_napi_add_config(vport->netdev, &q_vector->napi,
4491 napi_poll, v_idx);
4492 netif_napi_set_irq(&q_vector->napi, irq_num);
4493 }
4494 }
4495
4496 /**
4497 * idpf_vport_intr_alloc - Allocate memory for interrupt vectors
4498 * @vport: virtual port
4499 *
4500 * We allocate one q_vector per queue interrupt. If allocation fails we
4501 * return -ENOMEM.
4502 */
idpf_vport_intr_alloc(struct idpf_vport * vport)4503 int idpf_vport_intr_alloc(struct idpf_vport *vport)
4504 {
4505 u16 txqs_per_vector, rxqs_per_vector, bufqs_per_vector;
4506 struct idpf_vport_user_config_data *user_config;
4507 struct idpf_q_vector *q_vector;
4508 struct idpf_q_coalesce *q_coal;
4509 u32 complqs_per_vector, v_idx;
4510 u16 idx = vport->idx;
4511
4512 user_config = &vport->adapter->vport_config[idx]->user_config;
4513 vport->q_vectors = kcalloc(vport->num_q_vectors,
4514 sizeof(struct idpf_q_vector), GFP_KERNEL);
4515 if (!vport->q_vectors)
4516 return -ENOMEM;
4517
4518 txqs_per_vector = DIV_ROUND_UP(vport->num_txq_grp,
4519 vport->num_q_vectors);
4520 rxqs_per_vector = DIV_ROUND_UP(vport->num_rxq_grp,
4521 vport->num_q_vectors);
4522 bufqs_per_vector = vport->num_bufqs_per_qgrp *
4523 DIV_ROUND_UP(vport->num_rxq_grp,
4524 vport->num_q_vectors);
4525 complqs_per_vector = DIV_ROUND_UP(vport->num_txq_grp,
4526 vport->num_q_vectors);
4527
4528 for (v_idx = 0; v_idx < vport->num_q_vectors; v_idx++) {
4529 q_vector = &vport->q_vectors[v_idx];
4530 q_coal = &user_config->q_coalesce[v_idx];
4531 q_vector->vport = vport;
4532
4533 q_vector->tx_itr_value = q_coal->tx_coalesce_usecs;
4534 q_vector->tx_intr_mode = q_coal->tx_intr_mode;
4535 q_vector->tx_itr_idx = VIRTCHNL2_ITR_IDX_1;
4536
4537 q_vector->rx_itr_value = q_coal->rx_coalesce_usecs;
4538 q_vector->rx_intr_mode = q_coal->rx_intr_mode;
4539 q_vector->rx_itr_idx = VIRTCHNL2_ITR_IDX_0;
4540
4541 q_vector->tx = kcalloc(txqs_per_vector, sizeof(*q_vector->tx),
4542 GFP_KERNEL);
4543 if (!q_vector->tx)
4544 goto error;
4545
4546 q_vector->rx = kcalloc(rxqs_per_vector, sizeof(*q_vector->rx),
4547 GFP_KERNEL);
4548 if (!q_vector->rx)
4549 goto error;
4550
4551 if (!idpf_is_queue_model_split(vport->rxq_model))
4552 continue;
4553
4554 q_vector->bufq = kcalloc(bufqs_per_vector,
4555 sizeof(*q_vector->bufq),
4556 GFP_KERNEL);
4557 if (!q_vector->bufq)
4558 goto error;
4559
4560 q_vector->complq = kcalloc(complqs_per_vector,
4561 sizeof(*q_vector->complq),
4562 GFP_KERNEL);
4563 if (!q_vector->complq)
4564 goto error;
4565
4566 if (!vport->xdp_txq_offset)
4567 continue;
4568
4569 q_vector->xsksq = kcalloc(rxqs_per_vector,
4570 sizeof(*q_vector->xsksq),
4571 GFP_KERNEL);
4572 if (!q_vector->xsksq)
4573 goto error;
4574 }
4575
4576 return 0;
4577
4578 error:
4579 idpf_vport_intr_rel(vport);
4580
4581 return -ENOMEM;
4582 }
4583
4584 /**
4585 * idpf_vport_intr_init - Setup all vectors for the given vport
4586 * @vport: virtual port
4587 *
4588 * Returns 0 on success or negative on failure
4589 */
idpf_vport_intr_init(struct idpf_vport * vport)4590 int idpf_vport_intr_init(struct idpf_vport *vport)
4591 {
4592 int err;
4593
4594 err = idpf_vport_intr_init_vec_idx(vport);
4595 if (err)
4596 return err;
4597
4598 idpf_vport_intr_map_vector_to_qs(vport);
4599 idpf_vport_intr_napi_add_all(vport);
4600
4601 err = vport->adapter->dev_ops.reg_ops.intr_reg_init(vport);
4602 if (err)
4603 goto unroll_vectors_alloc;
4604
4605 err = idpf_vport_intr_req_irq(vport);
4606 if (err)
4607 goto unroll_vectors_alloc;
4608
4609 return 0;
4610
4611 unroll_vectors_alloc:
4612 idpf_vport_intr_napi_del_all(vport);
4613
4614 return err;
4615 }
4616
idpf_vport_intr_ena(struct idpf_vport * vport)4617 void idpf_vport_intr_ena(struct idpf_vport *vport)
4618 {
4619 idpf_vport_intr_napi_ena_all(vport);
4620 idpf_vport_intr_ena_irq_all(vport);
4621 }
4622
4623 /**
4624 * idpf_config_rss - Send virtchnl messages to configure RSS
4625 * @vport: virtual port
4626 *
4627 * Return 0 on success, negative on failure
4628 */
idpf_config_rss(struct idpf_vport * vport)4629 int idpf_config_rss(struct idpf_vport *vport)
4630 {
4631 int err;
4632
4633 err = idpf_send_get_set_rss_key_msg(vport, false);
4634 if (err)
4635 return err;
4636
4637 return idpf_send_get_set_rss_lut_msg(vport, false);
4638 }
4639
4640 /**
4641 * idpf_fill_dflt_rss_lut - Fill the indirection table with the default values
4642 * @vport: virtual port structure
4643 */
idpf_fill_dflt_rss_lut(struct idpf_vport * vport)4644 static void idpf_fill_dflt_rss_lut(struct idpf_vport *vport)
4645 {
4646 struct idpf_adapter *adapter = vport->adapter;
4647 u16 num_active_rxq = vport->num_rxq;
4648 struct idpf_rss_data *rss_data;
4649 int i;
4650
4651 rss_data = &adapter->vport_config[vport->idx]->user_config.rss_data;
4652
4653 for (i = 0; i < rss_data->rss_lut_size; i++) {
4654 rss_data->rss_lut[i] = i % num_active_rxq;
4655 rss_data->cached_lut[i] = rss_data->rss_lut[i];
4656 }
4657 }
4658
4659 /**
4660 * idpf_init_rss - Allocate and initialize RSS resources
4661 * @vport: virtual port
4662 *
4663 * Return 0 on success, negative on failure
4664 */
idpf_init_rss(struct idpf_vport * vport)4665 int idpf_init_rss(struct idpf_vport *vport)
4666 {
4667 struct idpf_adapter *adapter = vport->adapter;
4668 struct idpf_rss_data *rss_data;
4669 u32 lut_size;
4670
4671 rss_data = &adapter->vport_config[vport->idx]->user_config.rss_data;
4672
4673 lut_size = rss_data->rss_lut_size * sizeof(u32);
4674 rss_data->rss_lut = kzalloc(lut_size, GFP_KERNEL);
4675 if (!rss_data->rss_lut)
4676 return -ENOMEM;
4677
4678 rss_data->cached_lut = kzalloc(lut_size, GFP_KERNEL);
4679 if (!rss_data->cached_lut) {
4680 kfree(rss_data->rss_lut);
4681 rss_data->rss_lut = NULL;
4682
4683 return -ENOMEM;
4684 }
4685
4686 /* Fill the default RSS lut values */
4687 idpf_fill_dflt_rss_lut(vport);
4688
4689 return idpf_config_rss(vport);
4690 }
4691
4692 /**
4693 * idpf_deinit_rss - Release RSS resources
4694 * @vport: virtual port
4695 */
idpf_deinit_rss(struct idpf_vport * vport)4696 void idpf_deinit_rss(struct idpf_vport *vport)
4697 {
4698 struct idpf_adapter *adapter = vport->adapter;
4699 struct idpf_rss_data *rss_data;
4700
4701 rss_data = &adapter->vport_config[vport->idx]->user_config.rss_data;
4702 kfree(rss_data->cached_lut);
4703 rss_data->cached_lut = NULL;
4704 kfree(rss_data->rss_lut);
4705 rss_data->rss_lut = NULL;
4706 }
4707