1 // SPDX-License-Identifier: GPL-2.0 2 /* Copyright(c) 2017 - 2019 Pensando Systems, Inc */ 3 4 #include <linux/ip.h> 5 #include <linux/ipv6.h> 6 #include <linux/if_vlan.h> 7 #include <net/ip6_checksum.h> 8 #include <net/netdev_queues.h> 9 10 #include "ionic.h" 11 #include "ionic_lif.h" 12 #include "ionic_txrx.h" 13 14 static dma_addr_t ionic_tx_map_single(struct ionic_queue *q, 15 void *data, size_t len); 16 17 static dma_addr_t ionic_tx_map_frag(struct ionic_queue *q, 18 const skb_frag_t *frag, 19 size_t offset, size_t len); 20 21 static void ionic_tx_desc_unmap_bufs(struct ionic_queue *q, 22 struct ionic_tx_desc_info *desc_info); 23 24 static void ionic_tx_clean(struct ionic_queue *q, 25 struct ionic_tx_desc_info *desc_info, 26 struct ionic_txq_comp *comp); 27 28 static inline void ionic_txq_post(struct ionic_queue *q, bool ring_dbell) 29 { 30 ionic_q_post(q, ring_dbell); 31 } 32 33 static inline void ionic_rxq_post(struct ionic_queue *q, bool ring_dbell) 34 { 35 ionic_q_post(q, ring_dbell); 36 } 37 38 bool ionic_txq_poke_doorbell(struct ionic_queue *q) 39 { 40 struct netdev_queue *netdev_txq; 41 unsigned long now, then, dif; 42 struct net_device *netdev; 43 44 netdev = q->lif->netdev; 45 netdev_txq = netdev_get_tx_queue(netdev, q->index); 46 47 HARD_TX_LOCK(netdev, netdev_txq, smp_processor_id()); 48 49 if (q->tail_idx == q->head_idx) { 50 HARD_TX_UNLOCK(netdev, netdev_txq); 51 return false; 52 } 53 54 now = READ_ONCE(jiffies); 55 then = q->dbell_jiffies; 56 dif = now - then; 57 58 if (dif > q->dbell_deadline) { 59 ionic_dbell_ring(q->lif->kern_dbpage, q->hw_type, 60 q->dbval | q->head_idx); 61 62 q->dbell_jiffies = now; 63 } 64 65 HARD_TX_UNLOCK(netdev, netdev_txq); 66 67 return true; 68 } 69 70 bool ionic_rxq_poke_doorbell(struct ionic_queue *q) 71 { 72 unsigned long now, then, dif; 73 74 /* no lock, called from rx napi or txrx napi, nothing else can fill */ 75 76 if (q->tail_idx == q->head_idx) 77 return false; 78 79 now = READ_ONCE(jiffies); 80 then = q->dbell_jiffies; 81 dif = now - then; 82 83 if (dif > q->dbell_deadline) { 84 ionic_dbell_ring(q->lif->kern_dbpage, q->hw_type, 85 q->dbval | q->head_idx); 86 87 q->dbell_jiffies = now; 88 89 dif = 2 * q->dbell_deadline; 90 if (dif > IONIC_RX_MAX_DOORBELL_DEADLINE) 91 dif = IONIC_RX_MAX_DOORBELL_DEADLINE; 92 93 q->dbell_deadline = dif; 94 } 95 96 return true; 97 } 98 99 static inline struct ionic_txq_sg_elem *ionic_tx_sg_elems(struct ionic_queue *q) 100 { 101 if (likely(q->sg_desc_size == sizeof(struct ionic_txq_sg_desc_v1))) 102 return q->txq_sgl_v1[q->head_idx].elems; 103 else 104 return q->txq_sgl[q->head_idx].elems; 105 } 106 107 static inline struct netdev_queue *q_to_ndq(struct net_device *netdev, 108 struct ionic_queue *q) 109 { 110 return netdev_get_tx_queue(netdev, q->index); 111 } 112 113 static void *ionic_rx_buf_va(struct ionic_buf_info *buf_info) 114 { 115 return page_address(buf_info->page) + buf_info->page_offset; 116 } 117 118 static dma_addr_t ionic_rx_buf_pa(struct ionic_buf_info *buf_info) 119 { 120 return buf_info->dma_addr + buf_info->page_offset; 121 } 122 123 static unsigned int ionic_rx_buf_size(struct ionic_buf_info *buf_info) 124 { 125 return min_t(u32, IONIC_MAX_BUF_LEN, IONIC_PAGE_SIZE - buf_info->page_offset); 126 } 127 128 static int ionic_rx_page_alloc(struct ionic_queue *q, 129 struct ionic_buf_info *buf_info) 130 { 131 struct device *dev = q->dev; 132 dma_addr_t dma_addr; 133 struct page *page; 134 135 page = alloc_pages(IONIC_PAGE_GFP_MASK, 0); 136 if (unlikely(!page)) { 137 net_err_ratelimited("%s: %s page alloc failed\n", 138 dev_name(dev), q->name); 139 q_to_rx_stats(q)->alloc_err++; 140 return -ENOMEM; 141 } 142 143 dma_addr = dma_map_page(dev, page, 0, 144 IONIC_PAGE_SIZE, DMA_FROM_DEVICE); 145 if (unlikely(dma_mapping_error(dev, dma_addr))) { 146 __free_pages(page, 0); 147 net_err_ratelimited("%s: %s dma map failed\n", 148 dev_name(dev), q->name); 149 q_to_rx_stats(q)->dma_map_err++; 150 return -EIO; 151 } 152 153 buf_info->dma_addr = dma_addr; 154 buf_info->page = page; 155 buf_info->page_offset = 0; 156 157 return 0; 158 } 159 160 static void ionic_rx_page_free(struct ionic_queue *q, 161 struct ionic_buf_info *buf_info) 162 { 163 struct device *dev = q->dev; 164 165 if (unlikely(!buf_info)) { 166 net_err_ratelimited("%s: %s invalid buf_info in free\n", 167 dev_name(dev), q->name); 168 return; 169 } 170 171 if (!buf_info->page) 172 return; 173 174 dma_unmap_page(dev, buf_info->dma_addr, IONIC_PAGE_SIZE, DMA_FROM_DEVICE); 175 __free_pages(buf_info->page, 0); 176 buf_info->page = NULL; 177 } 178 179 static bool ionic_rx_buf_recycle(struct ionic_queue *q, 180 struct ionic_buf_info *buf_info, u32 len) 181 { 182 u32 size; 183 184 /* don't re-use pages allocated in low-mem condition */ 185 if (page_is_pfmemalloc(buf_info->page)) 186 return false; 187 188 /* don't re-use buffers from non-local numa nodes */ 189 if (page_to_nid(buf_info->page) != numa_mem_id()) 190 return false; 191 192 size = ALIGN(len, q->xdp_rxq_info ? IONIC_PAGE_SIZE : IONIC_PAGE_SPLIT_SZ); 193 buf_info->page_offset += size; 194 if (buf_info->page_offset >= IONIC_PAGE_SIZE) 195 return false; 196 197 get_page(buf_info->page); 198 199 return true; 200 } 201 202 static void ionic_rx_add_skb_frag(struct ionic_queue *q, 203 struct sk_buff *skb, 204 struct ionic_buf_info *buf_info, 205 u32 off, u32 len, 206 bool synced) 207 { 208 if (!synced) 209 dma_sync_single_range_for_cpu(q->dev, ionic_rx_buf_pa(buf_info), 210 off, len, DMA_FROM_DEVICE); 211 212 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, 213 buf_info->page, buf_info->page_offset + off, 214 len, 215 IONIC_PAGE_SIZE); 216 217 if (!ionic_rx_buf_recycle(q, buf_info, len)) { 218 dma_unmap_page(q->dev, buf_info->dma_addr, 219 IONIC_PAGE_SIZE, DMA_FROM_DEVICE); 220 buf_info->page = NULL; 221 } 222 } 223 224 static struct sk_buff *ionic_rx_build_skb(struct ionic_queue *q, 225 struct ionic_rx_desc_info *desc_info, 226 unsigned int headroom, 227 unsigned int len, 228 unsigned int num_sg_elems, 229 bool synced) 230 { 231 struct ionic_buf_info *buf_info; 232 struct sk_buff *skb; 233 unsigned int i; 234 u16 frag_len; 235 236 buf_info = &desc_info->bufs[0]; 237 prefetchw(buf_info->page); 238 239 skb = napi_get_frags(&q_to_qcq(q)->napi); 240 if (unlikely(!skb)) { 241 net_warn_ratelimited("%s: SKB alloc failed on %s!\n", 242 dev_name(q->dev), q->name); 243 q_to_rx_stats(q)->alloc_err++; 244 return NULL; 245 } 246 247 if (headroom) 248 frag_len = min_t(u16, len, 249 IONIC_XDP_MAX_LINEAR_MTU + VLAN_ETH_HLEN); 250 else 251 frag_len = min_t(u16, len, ionic_rx_buf_size(buf_info)); 252 253 if (unlikely(!buf_info->page)) 254 goto err_bad_buf_page; 255 ionic_rx_add_skb_frag(q, skb, buf_info, headroom, frag_len, synced); 256 len -= frag_len; 257 buf_info++; 258 259 for (i = 0; i < num_sg_elems; i++, buf_info++) { 260 if (unlikely(!buf_info->page)) 261 goto err_bad_buf_page; 262 frag_len = min_t(u16, len, ionic_rx_buf_size(buf_info)); 263 ionic_rx_add_skb_frag(q, skb, buf_info, 0, frag_len, synced); 264 len -= frag_len; 265 } 266 267 return skb; 268 269 err_bad_buf_page: 270 dev_kfree_skb(skb); 271 return NULL; 272 } 273 274 static struct sk_buff *ionic_rx_copybreak(struct net_device *netdev, 275 struct ionic_queue *q, 276 struct ionic_rx_desc_info *desc_info, 277 unsigned int headroom, 278 unsigned int len, 279 bool synced) 280 { 281 struct ionic_buf_info *buf_info; 282 struct device *dev = q->dev; 283 struct sk_buff *skb; 284 285 buf_info = &desc_info->bufs[0]; 286 287 skb = napi_alloc_skb(&q_to_qcq(q)->napi, len); 288 if (unlikely(!skb)) { 289 net_warn_ratelimited("%s: SKB alloc failed on %s!\n", 290 dev_name(dev), q->name); 291 q_to_rx_stats(q)->alloc_err++; 292 return NULL; 293 } 294 295 if (unlikely(!buf_info->page)) { 296 dev_kfree_skb(skb); 297 return NULL; 298 } 299 300 if (!synced) 301 dma_sync_single_range_for_cpu(dev, ionic_rx_buf_pa(buf_info), 302 headroom, len, DMA_FROM_DEVICE); 303 skb_copy_to_linear_data(skb, ionic_rx_buf_va(buf_info) + headroom, len); 304 dma_sync_single_range_for_device(dev, ionic_rx_buf_pa(buf_info), 305 headroom, len, DMA_FROM_DEVICE); 306 307 skb_put(skb, len); 308 skb->protocol = eth_type_trans(skb, netdev); 309 310 return skb; 311 } 312 313 static void ionic_xdp_tx_desc_clean(struct ionic_queue *q, 314 struct ionic_tx_desc_info *desc_info) 315 { 316 unsigned int nbufs = desc_info->nbufs; 317 struct ionic_buf_info *buf_info; 318 struct device *dev = q->dev; 319 int i; 320 321 if (!nbufs) 322 return; 323 324 buf_info = desc_info->bufs; 325 dma_unmap_single(dev, buf_info->dma_addr, 326 buf_info->len, DMA_TO_DEVICE); 327 if (desc_info->act == XDP_TX) 328 __free_pages(buf_info->page, 0); 329 buf_info->page = NULL; 330 331 buf_info++; 332 for (i = 1; i < nbufs + 1 && buf_info->page; i++, buf_info++) { 333 dma_unmap_page(dev, buf_info->dma_addr, 334 buf_info->len, DMA_TO_DEVICE); 335 if (desc_info->act == XDP_TX) 336 __free_pages(buf_info->page, 0); 337 buf_info->page = NULL; 338 } 339 340 if (desc_info->act == XDP_REDIRECT) 341 xdp_return_frame(desc_info->xdpf); 342 343 desc_info->nbufs = 0; 344 desc_info->xdpf = NULL; 345 desc_info->act = 0; 346 } 347 348 static int ionic_xdp_post_frame(struct ionic_queue *q, struct xdp_frame *frame, 349 enum xdp_action act, struct page *page, int off, 350 bool ring_doorbell) 351 { 352 struct ionic_tx_desc_info *desc_info; 353 struct ionic_buf_info *buf_info; 354 struct ionic_tx_stats *stats; 355 struct ionic_txq_desc *desc; 356 size_t len = frame->len; 357 dma_addr_t dma_addr; 358 u64 cmd; 359 360 desc_info = &q->tx_info[q->head_idx]; 361 desc = &q->txq[q->head_idx]; 362 buf_info = desc_info->bufs; 363 stats = q_to_tx_stats(q); 364 365 dma_addr = ionic_tx_map_single(q, frame->data, len); 366 if (!dma_addr) 367 return -EIO; 368 buf_info->dma_addr = dma_addr; 369 buf_info->len = len; 370 buf_info->page = page; 371 buf_info->page_offset = off; 372 373 desc_info->nbufs = 1; 374 desc_info->xdpf = frame; 375 desc_info->act = act; 376 377 if (xdp_frame_has_frags(frame)) { 378 struct ionic_txq_sg_elem *elem; 379 struct skb_shared_info *sinfo; 380 struct ionic_buf_info *bi; 381 skb_frag_t *frag; 382 int i; 383 384 bi = &buf_info[1]; 385 sinfo = xdp_get_shared_info_from_frame(frame); 386 frag = sinfo->frags; 387 elem = ionic_tx_sg_elems(q); 388 for (i = 0; i < sinfo->nr_frags; i++, frag++, bi++) { 389 dma_addr = ionic_tx_map_frag(q, frag, 0, skb_frag_size(frag)); 390 if (!dma_addr) { 391 ionic_tx_desc_unmap_bufs(q, desc_info); 392 return -EIO; 393 } 394 bi->dma_addr = dma_addr; 395 bi->len = skb_frag_size(frag); 396 bi->page = skb_frag_page(frag); 397 398 elem->addr = cpu_to_le64(bi->dma_addr); 399 elem->len = cpu_to_le16(bi->len); 400 elem++; 401 402 desc_info->nbufs++; 403 } 404 } 405 406 cmd = encode_txq_desc_cmd(IONIC_TXQ_DESC_OPCODE_CSUM_NONE, 407 0, (desc_info->nbufs - 1), buf_info->dma_addr); 408 desc->cmd = cpu_to_le64(cmd); 409 desc->len = cpu_to_le16(len); 410 desc->csum_start = 0; 411 desc->csum_offset = 0; 412 413 stats->xdp_frames++; 414 stats->pkts++; 415 stats->bytes += len; 416 417 ionic_txq_post(q, ring_doorbell); 418 419 return 0; 420 } 421 422 int ionic_xdp_xmit(struct net_device *netdev, int n, 423 struct xdp_frame **xdp_frames, u32 flags) 424 { 425 struct ionic_lif *lif = netdev_priv(netdev); 426 struct ionic_queue *txq; 427 struct netdev_queue *nq; 428 int nxmit; 429 int space; 430 int cpu; 431 int qi; 432 433 if (unlikely(!test_bit(IONIC_LIF_F_UP, lif->state))) 434 return -ENETDOWN; 435 436 if (unlikely(flags & ~XDP_XMIT_FLAGS_MASK)) 437 return -EINVAL; 438 439 /* AdminQ is assumed on cpu 0, while we attempt to affinitize the 440 * TxRx queue pairs 0..n-1 on cpus 1..n. We try to keep with that 441 * affinitization here, but of course irqbalance and friends might 442 * have juggled things anyway, so we have to check for the 0 case. 443 */ 444 cpu = smp_processor_id(); 445 qi = cpu ? (cpu - 1) % lif->nxqs : cpu; 446 447 txq = &lif->txqcqs[qi]->q; 448 nq = netdev_get_tx_queue(netdev, txq->index); 449 __netif_tx_lock(nq, cpu); 450 txq_trans_cond_update(nq); 451 452 if (netif_tx_queue_stopped(nq) || 453 !netif_txq_maybe_stop(q_to_ndq(netdev, txq), 454 ionic_q_space_avail(txq), 455 1, 1)) { 456 __netif_tx_unlock(nq); 457 return -EIO; 458 } 459 460 space = min_t(int, n, ionic_q_space_avail(txq)); 461 for (nxmit = 0; nxmit < space ; nxmit++) { 462 if (ionic_xdp_post_frame(txq, xdp_frames[nxmit], 463 XDP_REDIRECT, 464 virt_to_page(xdp_frames[nxmit]->data), 465 0, false)) { 466 nxmit--; 467 break; 468 } 469 } 470 471 if (flags & XDP_XMIT_FLUSH) 472 ionic_dbell_ring(lif->kern_dbpage, txq->hw_type, 473 txq->dbval | txq->head_idx); 474 475 netif_txq_maybe_stop(q_to_ndq(netdev, txq), 476 ionic_q_space_avail(txq), 477 4, 4); 478 __netif_tx_unlock(nq); 479 480 return nxmit; 481 } 482 483 static bool ionic_run_xdp(struct ionic_rx_stats *stats, 484 struct net_device *netdev, 485 struct bpf_prog *xdp_prog, 486 struct ionic_queue *rxq, 487 struct ionic_buf_info *buf_info, 488 int len) 489 { 490 u32 xdp_action = XDP_ABORTED; 491 struct xdp_buff xdp_buf; 492 struct ionic_queue *txq; 493 struct netdev_queue *nq; 494 struct xdp_frame *xdpf; 495 int remain_len; 496 int frag_len; 497 int err = 0; 498 499 xdp_init_buff(&xdp_buf, IONIC_PAGE_SIZE, rxq->xdp_rxq_info); 500 frag_len = min_t(u16, len, IONIC_XDP_MAX_LINEAR_MTU + VLAN_ETH_HLEN); 501 xdp_prepare_buff(&xdp_buf, ionic_rx_buf_va(buf_info), 502 XDP_PACKET_HEADROOM, frag_len, false); 503 504 dma_sync_single_range_for_cpu(rxq->dev, ionic_rx_buf_pa(buf_info), 505 XDP_PACKET_HEADROOM, frag_len, 506 DMA_FROM_DEVICE); 507 508 prefetchw(&xdp_buf.data_hard_start); 509 510 /* We limit MTU size to one buffer if !xdp_has_frags, so 511 * if the recv len is bigger than one buffer 512 * then we know we have frag info to gather 513 */ 514 remain_len = len - frag_len; 515 if (remain_len) { 516 struct skb_shared_info *sinfo; 517 struct ionic_buf_info *bi; 518 skb_frag_t *frag; 519 520 bi = buf_info; 521 sinfo = xdp_get_shared_info_from_buff(&xdp_buf); 522 sinfo->nr_frags = 0; 523 sinfo->xdp_frags_size = 0; 524 xdp_buff_set_frags_flag(&xdp_buf); 525 526 do { 527 if (unlikely(sinfo->nr_frags >= MAX_SKB_FRAGS)) { 528 err = -ENOSPC; 529 goto out_xdp_abort; 530 } 531 532 frag = &sinfo->frags[sinfo->nr_frags]; 533 sinfo->nr_frags++; 534 bi++; 535 frag_len = min_t(u16, remain_len, ionic_rx_buf_size(bi)); 536 dma_sync_single_range_for_cpu(rxq->dev, ionic_rx_buf_pa(bi), 537 0, frag_len, DMA_FROM_DEVICE); 538 skb_frag_fill_page_desc(frag, bi->page, 0, frag_len); 539 sinfo->xdp_frags_size += frag_len; 540 remain_len -= frag_len; 541 542 if (page_is_pfmemalloc(bi->page)) 543 xdp_buff_set_frag_pfmemalloc(&xdp_buf); 544 } while (remain_len > 0); 545 } 546 547 xdp_action = bpf_prog_run_xdp(xdp_prog, &xdp_buf); 548 549 switch (xdp_action) { 550 case XDP_PASS: 551 stats->xdp_pass++; 552 return false; /* false = we didn't consume the packet */ 553 554 case XDP_DROP: 555 ionic_rx_page_free(rxq, buf_info); 556 stats->xdp_drop++; 557 break; 558 559 case XDP_TX: 560 xdpf = xdp_convert_buff_to_frame(&xdp_buf); 561 if (!xdpf) 562 goto out_xdp_abort; 563 564 txq = rxq->partner; 565 nq = netdev_get_tx_queue(netdev, txq->index); 566 __netif_tx_lock(nq, smp_processor_id()); 567 txq_trans_cond_update(nq); 568 569 if (netif_tx_queue_stopped(nq) || 570 !netif_txq_maybe_stop(q_to_ndq(netdev, txq), 571 ionic_q_space_avail(txq), 572 1, 1)) { 573 __netif_tx_unlock(nq); 574 goto out_xdp_abort; 575 } 576 577 dma_unmap_page(rxq->dev, buf_info->dma_addr, 578 IONIC_PAGE_SIZE, DMA_FROM_DEVICE); 579 580 err = ionic_xdp_post_frame(txq, xdpf, XDP_TX, 581 buf_info->page, 582 buf_info->page_offset, 583 true); 584 __netif_tx_unlock(nq); 585 if (unlikely(err)) { 586 netdev_dbg(netdev, "tx ionic_xdp_post_frame err %d\n", err); 587 goto out_xdp_abort; 588 } 589 buf_info->page = NULL; 590 stats->xdp_tx++; 591 592 /* the Tx completion will free the buffers */ 593 break; 594 595 case XDP_REDIRECT: 596 /* unmap the pages before handing them to a different device */ 597 dma_unmap_page(rxq->dev, buf_info->dma_addr, 598 IONIC_PAGE_SIZE, DMA_FROM_DEVICE); 599 600 err = xdp_do_redirect(netdev, &xdp_buf, xdp_prog); 601 if (unlikely(err)) { 602 netdev_dbg(netdev, "xdp_do_redirect err %d\n", err); 603 goto out_xdp_abort; 604 } 605 buf_info->page = NULL; 606 rxq->xdp_flush = true; 607 stats->xdp_redirect++; 608 break; 609 610 case XDP_ABORTED: 611 default: 612 goto out_xdp_abort; 613 } 614 615 return true; 616 617 out_xdp_abort: 618 trace_xdp_exception(netdev, xdp_prog, xdp_action); 619 ionic_rx_page_free(rxq, buf_info); 620 stats->xdp_aborted++; 621 622 return true; 623 } 624 625 static void ionic_rx_clean(struct ionic_queue *q, 626 struct ionic_rx_desc_info *desc_info, 627 struct ionic_rxq_comp *comp) 628 { 629 struct net_device *netdev = q->lif->netdev; 630 struct ionic_qcq *qcq = q_to_qcq(q); 631 struct ionic_rx_stats *stats; 632 struct bpf_prog *xdp_prog; 633 unsigned int headroom; 634 struct sk_buff *skb; 635 bool synced = false; 636 bool use_copybreak; 637 u16 len; 638 639 stats = q_to_rx_stats(q); 640 641 if (comp->status) { 642 stats->dropped++; 643 return; 644 } 645 646 len = le16_to_cpu(comp->len); 647 stats->pkts++; 648 stats->bytes += len; 649 650 xdp_prog = READ_ONCE(q->lif->xdp_prog); 651 if (xdp_prog) { 652 if (ionic_run_xdp(stats, netdev, xdp_prog, q, desc_info->bufs, len)) 653 return; 654 synced = true; 655 } 656 657 headroom = q->xdp_rxq_info ? XDP_PACKET_HEADROOM : 0; 658 use_copybreak = len <= q->lif->rx_copybreak; 659 if (use_copybreak) 660 skb = ionic_rx_copybreak(netdev, q, desc_info, 661 headroom, len, synced); 662 else 663 skb = ionic_rx_build_skb(q, desc_info, headroom, len, 664 comp->num_sg_elems, synced); 665 666 if (unlikely(!skb)) { 667 stats->dropped++; 668 return; 669 } 670 671 skb_record_rx_queue(skb, q->index); 672 673 if (likely(netdev->features & NETIF_F_RXHASH)) { 674 switch (comp->pkt_type_color & IONIC_RXQ_COMP_PKT_TYPE_MASK) { 675 case IONIC_PKT_TYPE_IPV4: 676 case IONIC_PKT_TYPE_IPV6: 677 skb_set_hash(skb, le32_to_cpu(comp->rss_hash), 678 PKT_HASH_TYPE_L3); 679 break; 680 case IONIC_PKT_TYPE_IPV4_TCP: 681 case IONIC_PKT_TYPE_IPV6_TCP: 682 case IONIC_PKT_TYPE_IPV4_UDP: 683 case IONIC_PKT_TYPE_IPV6_UDP: 684 skb_set_hash(skb, le32_to_cpu(comp->rss_hash), 685 PKT_HASH_TYPE_L4); 686 break; 687 } 688 } 689 690 if (likely(netdev->features & NETIF_F_RXCSUM) && 691 (comp->csum_flags & IONIC_RXQ_COMP_CSUM_F_CALC)) { 692 skb->ip_summed = CHECKSUM_COMPLETE; 693 skb->csum = (__force __wsum)le16_to_cpu(comp->csum); 694 stats->csum_complete++; 695 } else { 696 stats->csum_none++; 697 } 698 699 if (unlikely((comp->csum_flags & IONIC_RXQ_COMP_CSUM_F_TCP_BAD) || 700 (comp->csum_flags & IONIC_RXQ_COMP_CSUM_F_UDP_BAD) || 701 (comp->csum_flags & IONIC_RXQ_COMP_CSUM_F_IP_BAD))) 702 stats->csum_error++; 703 704 if (likely(netdev->features & NETIF_F_HW_VLAN_CTAG_RX) && 705 (comp->csum_flags & IONIC_RXQ_COMP_CSUM_F_VLAN)) { 706 __vlan_hwaccel_put_tag(skb, htons(ETH_P_8021Q), 707 le16_to_cpu(comp->vlan_tci)); 708 stats->vlan_stripped++; 709 } 710 711 if (unlikely(q->features & IONIC_RXQ_F_HWSTAMP)) { 712 __le64 *cq_desc_hwstamp; 713 u64 hwstamp; 714 715 cq_desc_hwstamp = 716 (void *)comp + 717 qcq->cq.desc_size - 718 sizeof(struct ionic_rxq_comp) - 719 IONIC_HWSTAMP_CQ_NEGOFFSET; 720 721 hwstamp = le64_to_cpu(*cq_desc_hwstamp); 722 723 if (hwstamp != IONIC_HWSTAMP_INVALID) { 724 skb_hwtstamps(skb)->hwtstamp = ionic_lif_phc_ktime(q->lif, hwstamp); 725 stats->hwstamp_valid++; 726 } else { 727 stats->hwstamp_invalid++; 728 } 729 } 730 731 if (use_copybreak) 732 napi_gro_receive(&qcq->napi, skb); 733 else 734 napi_gro_frags(&qcq->napi); 735 } 736 737 bool ionic_rx_service(struct ionic_cq *cq) 738 { 739 struct ionic_rx_desc_info *desc_info; 740 struct ionic_queue *q = cq->bound_q; 741 struct ionic_rxq_comp *comp; 742 743 comp = &((struct ionic_rxq_comp *)cq->base)[cq->tail_idx]; 744 745 if (!color_match(comp->pkt_type_color, cq->done_color)) 746 return false; 747 748 /* check for empty queue */ 749 if (q->tail_idx == q->head_idx) 750 return false; 751 752 if (q->tail_idx != le16_to_cpu(comp->comp_index)) 753 return false; 754 755 desc_info = &q->rx_info[q->tail_idx]; 756 q->tail_idx = (q->tail_idx + 1) & (q->num_descs - 1); 757 758 /* clean the related q entry, only one per qc completion */ 759 ionic_rx_clean(q, desc_info, comp); 760 761 return true; 762 } 763 764 static inline void ionic_write_cmb_desc(struct ionic_queue *q, 765 void *desc) 766 { 767 /* Since Rx and Tx descriptors are the same size, we can 768 * save an instruction or two and skip the qtype check. 769 */ 770 if (unlikely(q_to_qcq(q)->flags & IONIC_QCQ_F_CMB_RINGS)) 771 memcpy_toio(&q->cmb_txq[q->head_idx], desc, sizeof(q->cmb_txq[0])); 772 } 773 774 void ionic_rx_fill(struct ionic_queue *q) 775 { 776 struct net_device *netdev = q->lif->netdev; 777 struct ionic_rx_desc_info *desc_info; 778 struct ionic_rxq_sg_elem *sg_elem; 779 struct ionic_buf_info *buf_info; 780 unsigned int fill_threshold; 781 struct ionic_rxq_desc *desc; 782 unsigned int remain_len; 783 unsigned int frag_len; 784 unsigned int nfrags; 785 unsigned int n_fill; 786 unsigned int len; 787 unsigned int i; 788 unsigned int j; 789 790 n_fill = ionic_q_space_avail(q); 791 792 fill_threshold = min_t(unsigned int, IONIC_RX_FILL_THRESHOLD, 793 q->num_descs / IONIC_RX_FILL_DIV); 794 if (n_fill < fill_threshold) 795 return; 796 797 len = netdev->mtu + VLAN_ETH_HLEN; 798 799 for (i = n_fill; i; i--) { 800 unsigned int headroom; 801 unsigned int buf_len; 802 803 nfrags = 0; 804 remain_len = len; 805 desc = &q->rxq[q->head_idx]; 806 desc_info = &q->rx_info[q->head_idx]; 807 buf_info = &desc_info->bufs[0]; 808 809 if (!buf_info->page) { /* alloc a new buffer? */ 810 if (unlikely(ionic_rx_page_alloc(q, buf_info))) { 811 desc->addr = 0; 812 desc->len = 0; 813 return; 814 } 815 } 816 817 /* fill main descriptor - buf[0] 818 * XDP uses space in the first buffer, so account for 819 * head room, tail room, and ip header in the first frag size. 820 */ 821 headroom = q->xdp_rxq_info ? XDP_PACKET_HEADROOM : 0; 822 if (q->xdp_rxq_info) 823 buf_len = IONIC_XDP_MAX_LINEAR_MTU + VLAN_ETH_HLEN; 824 else 825 buf_len = ionic_rx_buf_size(buf_info); 826 frag_len = min_t(u16, len, buf_len); 827 828 desc->addr = cpu_to_le64(ionic_rx_buf_pa(buf_info) + headroom); 829 desc->len = cpu_to_le16(frag_len); 830 remain_len -= frag_len; 831 buf_info++; 832 nfrags++; 833 834 /* fill sg descriptors - buf[1..n] */ 835 sg_elem = q->rxq_sgl[q->head_idx].elems; 836 for (j = 0; remain_len > 0 && j < q->max_sg_elems; j++, sg_elem++) { 837 if (!buf_info->page) { /* alloc a new sg buffer? */ 838 if (unlikely(ionic_rx_page_alloc(q, buf_info))) { 839 sg_elem->addr = 0; 840 sg_elem->len = 0; 841 return; 842 } 843 } 844 845 sg_elem->addr = cpu_to_le64(ionic_rx_buf_pa(buf_info)); 846 frag_len = min_t(u16, remain_len, ionic_rx_buf_size(buf_info)); 847 sg_elem->len = cpu_to_le16(frag_len); 848 remain_len -= frag_len; 849 buf_info++; 850 nfrags++; 851 } 852 853 /* clear end sg element as a sentinel */ 854 if (j < q->max_sg_elems) 855 memset(sg_elem, 0, sizeof(*sg_elem)); 856 857 desc->opcode = (nfrags > 1) ? IONIC_RXQ_DESC_OPCODE_SG : 858 IONIC_RXQ_DESC_OPCODE_SIMPLE; 859 desc_info->nbufs = nfrags; 860 861 ionic_write_cmb_desc(q, desc); 862 863 ionic_rxq_post(q, false); 864 } 865 866 ionic_dbell_ring(q->lif->kern_dbpage, q->hw_type, 867 q->dbval | q->head_idx); 868 869 q->dbell_deadline = IONIC_RX_MIN_DOORBELL_DEADLINE; 870 q->dbell_jiffies = jiffies; 871 } 872 873 void ionic_rx_empty(struct ionic_queue *q) 874 { 875 struct ionic_rx_desc_info *desc_info; 876 struct ionic_buf_info *buf_info; 877 unsigned int i, j; 878 879 for (i = 0; i < q->num_descs; i++) { 880 desc_info = &q->rx_info[i]; 881 for (j = 0; j < ARRAY_SIZE(desc_info->bufs); j++) { 882 buf_info = &desc_info->bufs[j]; 883 if (buf_info->page) 884 ionic_rx_page_free(q, buf_info); 885 } 886 887 desc_info->nbufs = 0; 888 } 889 890 q->head_idx = 0; 891 q->tail_idx = 0; 892 } 893 894 static void ionic_dim_update(struct ionic_qcq *qcq, int napi_mode) 895 { 896 struct dim_sample dim_sample; 897 struct ionic_lif *lif; 898 unsigned int qi; 899 u64 pkts, bytes; 900 901 if (!qcq->intr.dim_coal_hw) 902 return; 903 904 lif = qcq->q.lif; 905 qi = qcq->cq.bound_q->index; 906 907 switch (napi_mode) { 908 case IONIC_LIF_F_TX_DIM_INTR: 909 pkts = lif->txqstats[qi].pkts; 910 bytes = lif->txqstats[qi].bytes; 911 break; 912 case IONIC_LIF_F_RX_DIM_INTR: 913 pkts = lif->rxqstats[qi].pkts; 914 bytes = lif->rxqstats[qi].bytes; 915 break; 916 default: 917 pkts = lif->txqstats[qi].pkts + lif->rxqstats[qi].pkts; 918 bytes = lif->txqstats[qi].bytes + lif->rxqstats[qi].bytes; 919 break; 920 } 921 922 dim_update_sample(qcq->cq.bound_intr->rearm_count, 923 pkts, bytes, &dim_sample); 924 925 net_dim(&qcq->dim, dim_sample); 926 } 927 928 int ionic_tx_napi(struct napi_struct *napi, int budget) 929 { 930 struct ionic_qcq *qcq = napi_to_qcq(napi); 931 struct ionic_cq *cq = napi_to_cq(napi); 932 u32 work_done = 0; 933 u32 flags = 0; 934 935 work_done = ionic_tx_cq_service(cq, budget); 936 937 if (unlikely(!budget)) 938 return budget; 939 940 if (work_done < budget && napi_complete_done(napi, work_done)) { 941 ionic_dim_update(qcq, IONIC_LIF_F_TX_DIM_INTR); 942 flags |= IONIC_INTR_CRED_UNMASK; 943 cq->bound_intr->rearm_count++; 944 } 945 946 if (work_done || flags) { 947 flags |= IONIC_INTR_CRED_RESET_COALESCE; 948 ionic_intr_credits(cq->idev->intr_ctrl, 949 cq->bound_intr->index, 950 work_done, flags); 951 } 952 953 if (!work_done && cq->bound_q->lif->doorbell_wa) 954 ionic_txq_poke_doorbell(&qcq->q); 955 956 return work_done; 957 } 958 959 static void ionic_xdp_do_flush(struct ionic_cq *cq) 960 { 961 if (cq->bound_q->xdp_flush) { 962 xdp_do_flush(); 963 cq->bound_q->xdp_flush = false; 964 } 965 } 966 967 int ionic_rx_napi(struct napi_struct *napi, int budget) 968 { 969 struct ionic_qcq *qcq = napi_to_qcq(napi); 970 struct ionic_cq *cq = napi_to_cq(napi); 971 u32 work_done = 0; 972 u32 flags = 0; 973 974 if (unlikely(!budget)) 975 return budget; 976 977 work_done = ionic_cq_service(cq, budget, 978 ionic_rx_service, NULL, NULL); 979 980 ionic_rx_fill(cq->bound_q); 981 982 ionic_xdp_do_flush(cq); 983 if (work_done < budget && napi_complete_done(napi, work_done)) { 984 ionic_dim_update(qcq, IONIC_LIF_F_RX_DIM_INTR); 985 flags |= IONIC_INTR_CRED_UNMASK; 986 cq->bound_intr->rearm_count++; 987 } 988 989 if (work_done || flags) { 990 flags |= IONIC_INTR_CRED_RESET_COALESCE; 991 ionic_intr_credits(cq->idev->intr_ctrl, 992 cq->bound_intr->index, 993 work_done, flags); 994 } 995 996 if (!work_done && cq->bound_q->lif->doorbell_wa) 997 ionic_rxq_poke_doorbell(&qcq->q); 998 999 return work_done; 1000 } 1001 1002 int ionic_txrx_napi(struct napi_struct *napi, int budget) 1003 { 1004 struct ionic_qcq *rxqcq = napi_to_qcq(napi); 1005 struct ionic_cq *rxcq = napi_to_cq(napi); 1006 unsigned int qi = rxcq->bound_q->index; 1007 struct ionic_qcq *txqcq; 1008 struct ionic_lif *lif; 1009 struct ionic_cq *txcq; 1010 u32 rx_work_done = 0; 1011 u32 tx_work_done = 0; 1012 u32 flags = 0; 1013 1014 lif = rxcq->bound_q->lif; 1015 txqcq = lif->txqcqs[qi]; 1016 txcq = &lif->txqcqs[qi]->cq; 1017 1018 tx_work_done = ionic_tx_cq_service(txcq, IONIC_TX_BUDGET_DEFAULT); 1019 1020 if (unlikely(!budget)) 1021 return budget; 1022 1023 rx_work_done = ionic_cq_service(rxcq, budget, 1024 ionic_rx_service, NULL, NULL); 1025 1026 ionic_rx_fill(rxcq->bound_q); 1027 1028 ionic_xdp_do_flush(rxcq); 1029 if (rx_work_done < budget && napi_complete_done(napi, rx_work_done)) { 1030 ionic_dim_update(rxqcq, 0); 1031 flags |= IONIC_INTR_CRED_UNMASK; 1032 rxcq->bound_intr->rearm_count++; 1033 } 1034 1035 if (rx_work_done || flags) { 1036 flags |= IONIC_INTR_CRED_RESET_COALESCE; 1037 ionic_intr_credits(rxcq->idev->intr_ctrl, rxcq->bound_intr->index, 1038 tx_work_done + rx_work_done, flags); 1039 } 1040 1041 if (lif->doorbell_wa) { 1042 if (!rx_work_done) 1043 ionic_rxq_poke_doorbell(&rxqcq->q); 1044 if (!tx_work_done) 1045 ionic_txq_poke_doorbell(&txqcq->q); 1046 } 1047 1048 return rx_work_done; 1049 } 1050 1051 static dma_addr_t ionic_tx_map_single(struct ionic_queue *q, 1052 void *data, size_t len) 1053 { 1054 struct device *dev = q->dev; 1055 dma_addr_t dma_addr; 1056 1057 dma_addr = dma_map_single(dev, data, len, DMA_TO_DEVICE); 1058 if (unlikely(dma_mapping_error(dev, dma_addr))) { 1059 net_warn_ratelimited("%s: DMA single map failed on %s!\n", 1060 dev_name(dev), q->name); 1061 q_to_tx_stats(q)->dma_map_err++; 1062 return 0; 1063 } 1064 return dma_addr; 1065 } 1066 1067 static dma_addr_t ionic_tx_map_frag(struct ionic_queue *q, 1068 const skb_frag_t *frag, 1069 size_t offset, size_t len) 1070 { 1071 struct device *dev = q->dev; 1072 dma_addr_t dma_addr; 1073 1074 dma_addr = skb_frag_dma_map(dev, frag, offset, len, DMA_TO_DEVICE); 1075 if (unlikely(dma_mapping_error(dev, dma_addr))) { 1076 net_warn_ratelimited("%s: DMA frag map failed on %s!\n", 1077 dev_name(dev), q->name); 1078 q_to_tx_stats(q)->dma_map_err++; 1079 return 0; 1080 } 1081 return dma_addr; 1082 } 1083 1084 static int ionic_tx_map_skb(struct ionic_queue *q, struct sk_buff *skb, 1085 struct ionic_tx_desc_info *desc_info) 1086 { 1087 struct ionic_buf_info *buf_info = desc_info->bufs; 1088 struct device *dev = q->dev; 1089 dma_addr_t dma_addr; 1090 unsigned int nfrags; 1091 skb_frag_t *frag; 1092 int frag_idx; 1093 1094 dma_addr = ionic_tx_map_single(q, skb->data, skb_headlen(skb)); 1095 if (!dma_addr) 1096 return -EIO; 1097 buf_info->dma_addr = dma_addr; 1098 buf_info->len = skb_headlen(skb); 1099 buf_info++; 1100 1101 frag = skb_shinfo(skb)->frags; 1102 nfrags = skb_shinfo(skb)->nr_frags; 1103 for (frag_idx = 0; frag_idx < nfrags; frag_idx++, frag++) { 1104 dma_addr = ionic_tx_map_frag(q, frag, 0, skb_frag_size(frag)); 1105 if (!dma_addr) 1106 goto dma_fail; 1107 buf_info->dma_addr = dma_addr; 1108 buf_info->len = skb_frag_size(frag); 1109 buf_info++; 1110 } 1111 1112 desc_info->nbufs = 1 + nfrags; 1113 1114 return 0; 1115 1116 dma_fail: 1117 /* unwind the frag mappings and the head mapping */ 1118 while (frag_idx > 0) { 1119 frag_idx--; 1120 buf_info--; 1121 dma_unmap_page(dev, buf_info->dma_addr, 1122 buf_info->len, DMA_TO_DEVICE); 1123 } 1124 dma_unmap_single(dev, desc_info->bufs[0].dma_addr, 1125 desc_info->bufs[0].len, DMA_TO_DEVICE); 1126 return -EIO; 1127 } 1128 1129 static void ionic_tx_desc_unmap_bufs(struct ionic_queue *q, 1130 struct ionic_tx_desc_info *desc_info) 1131 { 1132 struct ionic_buf_info *buf_info = desc_info->bufs; 1133 struct device *dev = q->dev; 1134 unsigned int i; 1135 1136 if (!desc_info->nbufs) 1137 return; 1138 1139 dma_unmap_single(dev, buf_info->dma_addr, 1140 buf_info->len, DMA_TO_DEVICE); 1141 buf_info++; 1142 for (i = 1; i < desc_info->nbufs; i++, buf_info++) 1143 dma_unmap_page(dev, buf_info->dma_addr, 1144 buf_info->len, DMA_TO_DEVICE); 1145 1146 desc_info->nbufs = 0; 1147 } 1148 1149 static void ionic_tx_clean(struct ionic_queue *q, 1150 struct ionic_tx_desc_info *desc_info, 1151 struct ionic_txq_comp *comp) 1152 { 1153 struct ionic_tx_stats *stats = q_to_tx_stats(q); 1154 struct ionic_qcq *qcq = q_to_qcq(q); 1155 struct sk_buff *skb; 1156 1157 if (desc_info->xdpf) { 1158 ionic_xdp_tx_desc_clean(q->partner, desc_info); 1159 stats->clean++; 1160 1161 if (unlikely(__netif_subqueue_stopped(q->lif->netdev, q->index))) 1162 netif_wake_subqueue(q->lif->netdev, q->index); 1163 1164 return; 1165 } 1166 1167 ionic_tx_desc_unmap_bufs(q, desc_info); 1168 1169 skb = desc_info->skb; 1170 if (!skb) 1171 return; 1172 1173 if (unlikely(ionic_txq_hwstamp_enabled(q))) { 1174 if (comp) { 1175 struct skb_shared_hwtstamps hwts = {}; 1176 __le64 *cq_desc_hwstamp; 1177 u64 hwstamp; 1178 1179 cq_desc_hwstamp = 1180 (void *)comp + 1181 qcq->cq.desc_size - 1182 sizeof(struct ionic_txq_comp) - 1183 IONIC_HWSTAMP_CQ_NEGOFFSET; 1184 1185 hwstamp = le64_to_cpu(*cq_desc_hwstamp); 1186 1187 if (hwstamp != IONIC_HWSTAMP_INVALID) { 1188 hwts.hwtstamp = ionic_lif_phc_ktime(q->lif, hwstamp); 1189 1190 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; 1191 skb_tstamp_tx(skb, &hwts); 1192 1193 stats->hwstamp_valid++; 1194 } else { 1195 stats->hwstamp_invalid++; 1196 } 1197 } 1198 } 1199 1200 desc_info->bytes = skb->len; 1201 stats->clean++; 1202 1203 napi_consume_skb(skb, 1); 1204 } 1205 1206 static bool ionic_tx_service(struct ionic_cq *cq, 1207 unsigned int *total_pkts, unsigned int *total_bytes) 1208 { 1209 struct ionic_tx_desc_info *desc_info; 1210 struct ionic_queue *q = cq->bound_q; 1211 struct ionic_txq_comp *comp; 1212 unsigned int bytes = 0; 1213 unsigned int pkts = 0; 1214 u16 index; 1215 1216 comp = &((struct ionic_txq_comp *)cq->base)[cq->tail_idx]; 1217 1218 if (!color_match(comp->color, cq->done_color)) 1219 return false; 1220 1221 /* clean the related q entries, there could be 1222 * several q entries completed for each cq completion 1223 */ 1224 do { 1225 desc_info = &q->tx_info[q->tail_idx]; 1226 desc_info->bytes = 0; 1227 index = q->tail_idx; 1228 q->tail_idx = (q->tail_idx + 1) & (q->num_descs - 1); 1229 ionic_tx_clean(q, desc_info, comp); 1230 if (desc_info->skb) { 1231 pkts++; 1232 bytes += desc_info->bytes; 1233 desc_info->skb = NULL; 1234 } 1235 } while (index != le16_to_cpu(comp->comp_index)); 1236 1237 (*total_pkts) += pkts; 1238 (*total_bytes) += bytes; 1239 1240 return true; 1241 } 1242 1243 unsigned int ionic_tx_cq_service(struct ionic_cq *cq, unsigned int work_to_do) 1244 { 1245 unsigned int work_done = 0; 1246 unsigned int bytes = 0; 1247 unsigned int pkts = 0; 1248 1249 if (work_to_do == 0) 1250 return 0; 1251 1252 while (ionic_tx_service(cq, &pkts, &bytes)) { 1253 if (cq->tail_idx == cq->num_descs - 1) 1254 cq->done_color = !cq->done_color; 1255 cq->tail_idx = (cq->tail_idx + 1) & (cq->num_descs - 1); 1256 1257 if (++work_done >= work_to_do) 1258 break; 1259 } 1260 1261 if (work_done) { 1262 struct ionic_queue *q = cq->bound_q; 1263 1264 if (likely(!ionic_txq_hwstamp_enabled(q))) 1265 netif_txq_completed_wake(q_to_ndq(q->lif->netdev, q), 1266 pkts, bytes, 1267 ionic_q_space_avail(q), 1268 IONIC_TSO_DESCS_NEEDED); 1269 } 1270 1271 return work_done; 1272 } 1273 1274 void ionic_tx_flush(struct ionic_cq *cq) 1275 { 1276 u32 work_done; 1277 1278 work_done = ionic_tx_cq_service(cq, cq->num_descs); 1279 if (work_done) 1280 ionic_intr_credits(cq->idev->intr_ctrl, cq->bound_intr->index, 1281 work_done, IONIC_INTR_CRED_RESET_COALESCE); 1282 } 1283 1284 void ionic_tx_empty(struct ionic_queue *q) 1285 { 1286 struct ionic_tx_desc_info *desc_info; 1287 int bytes = 0; 1288 int pkts = 0; 1289 1290 /* walk the not completed tx entries, if any */ 1291 while (q->head_idx != q->tail_idx) { 1292 desc_info = &q->tx_info[q->tail_idx]; 1293 desc_info->bytes = 0; 1294 q->tail_idx = (q->tail_idx + 1) & (q->num_descs - 1); 1295 ionic_tx_clean(q, desc_info, NULL); 1296 if (desc_info->skb) { 1297 pkts++; 1298 bytes += desc_info->bytes; 1299 desc_info->skb = NULL; 1300 } 1301 } 1302 1303 if (likely(!ionic_txq_hwstamp_enabled(q))) { 1304 struct netdev_queue *ndq = q_to_ndq(q->lif->netdev, q); 1305 1306 netdev_tx_completed_queue(ndq, pkts, bytes); 1307 netdev_tx_reset_queue(ndq); 1308 } 1309 } 1310 1311 static int ionic_tx_tcp_inner_pseudo_csum(struct sk_buff *skb) 1312 { 1313 int err; 1314 1315 err = skb_cow_head(skb, 0); 1316 if (unlikely(err)) 1317 return err; 1318 1319 if (skb->protocol == cpu_to_be16(ETH_P_IP)) { 1320 inner_ip_hdr(skb)->check = 0; 1321 inner_tcp_hdr(skb)->check = 1322 ~csum_tcpudp_magic(inner_ip_hdr(skb)->saddr, 1323 inner_ip_hdr(skb)->daddr, 1324 0, IPPROTO_TCP, 0); 1325 } else if (skb->protocol == cpu_to_be16(ETH_P_IPV6)) { 1326 inner_tcp_hdr(skb)->check = 1327 ~csum_ipv6_magic(&inner_ipv6_hdr(skb)->saddr, 1328 &inner_ipv6_hdr(skb)->daddr, 1329 0, IPPROTO_TCP, 0); 1330 } 1331 1332 return 0; 1333 } 1334 1335 static int ionic_tx_tcp_pseudo_csum(struct sk_buff *skb) 1336 { 1337 int err; 1338 1339 err = skb_cow_head(skb, 0); 1340 if (unlikely(err)) 1341 return err; 1342 1343 if (skb->protocol == cpu_to_be16(ETH_P_IP)) { 1344 ip_hdr(skb)->check = 0; 1345 tcp_hdr(skb)->check = 1346 ~csum_tcpudp_magic(ip_hdr(skb)->saddr, 1347 ip_hdr(skb)->daddr, 1348 0, IPPROTO_TCP, 0); 1349 } else if (skb->protocol == cpu_to_be16(ETH_P_IPV6)) { 1350 tcp_v6_gso_csum_prep(skb); 1351 } 1352 1353 return 0; 1354 } 1355 1356 static void ionic_tx_tso_post(struct net_device *netdev, struct ionic_queue *q, 1357 struct ionic_txq_desc *desc, 1358 struct sk_buff *skb, 1359 dma_addr_t addr, u8 nsge, u16 len, 1360 unsigned int hdrlen, unsigned int mss, 1361 bool outer_csum, 1362 u16 vlan_tci, bool has_vlan, 1363 bool start, bool done) 1364 { 1365 u8 flags = 0; 1366 u64 cmd; 1367 1368 flags |= has_vlan ? IONIC_TXQ_DESC_FLAG_VLAN : 0; 1369 flags |= outer_csum ? IONIC_TXQ_DESC_FLAG_ENCAP : 0; 1370 flags |= start ? IONIC_TXQ_DESC_FLAG_TSO_SOT : 0; 1371 flags |= done ? IONIC_TXQ_DESC_FLAG_TSO_EOT : 0; 1372 1373 cmd = encode_txq_desc_cmd(IONIC_TXQ_DESC_OPCODE_TSO, flags, nsge, addr); 1374 desc->cmd = cpu_to_le64(cmd); 1375 desc->len = cpu_to_le16(len); 1376 desc->vlan_tci = cpu_to_le16(vlan_tci); 1377 desc->hdr_len = cpu_to_le16(hdrlen); 1378 desc->mss = cpu_to_le16(mss); 1379 1380 ionic_write_cmb_desc(q, desc); 1381 1382 if (start) { 1383 skb_tx_timestamp(skb); 1384 if (likely(!ionic_txq_hwstamp_enabled(q))) 1385 netdev_tx_sent_queue(q_to_ndq(netdev, q), skb->len); 1386 ionic_txq_post(q, false); 1387 } else { 1388 ionic_txq_post(q, done); 1389 } 1390 } 1391 1392 static int ionic_tx_tso(struct net_device *netdev, struct ionic_queue *q, 1393 struct sk_buff *skb) 1394 { 1395 struct ionic_tx_stats *stats = q_to_tx_stats(q); 1396 struct ionic_tx_desc_info *desc_info; 1397 struct ionic_buf_info *buf_info; 1398 struct ionic_txq_sg_elem *elem; 1399 struct ionic_txq_desc *desc; 1400 unsigned int chunk_len; 1401 unsigned int frag_rem; 1402 unsigned int tso_rem; 1403 unsigned int seg_rem; 1404 dma_addr_t desc_addr; 1405 dma_addr_t frag_addr; 1406 unsigned int hdrlen; 1407 unsigned int len; 1408 unsigned int mss; 1409 bool start, done; 1410 bool outer_csum; 1411 bool has_vlan; 1412 u16 desc_len; 1413 u8 desc_nsge; 1414 u16 vlan_tci; 1415 bool encap; 1416 int err; 1417 1418 desc_info = &q->tx_info[q->head_idx]; 1419 1420 if (unlikely(ionic_tx_map_skb(q, skb, desc_info))) 1421 return -EIO; 1422 1423 len = skb->len; 1424 mss = skb_shinfo(skb)->gso_size; 1425 outer_csum = (skb_shinfo(skb)->gso_type & (SKB_GSO_GRE | 1426 SKB_GSO_GRE_CSUM | 1427 SKB_GSO_IPXIP4 | 1428 SKB_GSO_IPXIP6 | 1429 SKB_GSO_UDP_TUNNEL | 1430 SKB_GSO_UDP_TUNNEL_CSUM)); 1431 has_vlan = !!skb_vlan_tag_present(skb); 1432 vlan_tci = skb_vlan_tag_get(skb); 1433 encap = skb->encapsulation; 1434 1435 /* Preload inner-most TCP csum field with IP pseudo hdr 1436 * calculated with IP length set to zero. HW will later 1437 * add in length to each TCP segment resulting from the TSO. 1438 */ 1439 1440 if (encap) 1441 err = ionic_tx_tcp_inner_pseudo_csum(skb); 1442 else 1443 err = ionic_tx_tcp_pseudo_csum(skb); 1444 if (unlikely(err)) { 1445 /* clean up mapping from ionic_tx_map_skb */ 1446 ionic_tx_desc_unmap_bufs(q, desc_info); 1447 return err; 1448 } 1449 1450 if (encap) 1451 hdrlen = skb_inner_tcp_all_headers(skb); 1452 else 1453 hdrlen = skb_tcp_all_headers(skb); 1454 1455 desc_info->skb = skb; 1456 buf_info = desc_info->bufs; 1457 tso_rem = len; 1458 seg_rem = min(tso_rem, hdrlen + mss); 1459 1460 frag_addr = 0; 1461 frag_rem = 0; 1462 1463 start = true; 1464 1465 while (tso_rem > 0) { 1466 desc = NULL; 1467 elem = NULL; 1468 desc_addr = 0; 1469 desc_len = 0; 1470 desc_nsge = 0; 1471 /* use fragments until we have enough to post a single descriptor */ 1472 while (seg_rem > 0) { 1473 /* if the fragment is exhausted then move to the next one */ 1474 if (frag_rem == 0) { 1475 /* grab the next fragment */ 1476 frag_addr = buf_info->dma_addr; 1477 frag_rem = buf_info->len; 1478 buf_info++; 1479 } 1480 chunk_len = min(frag_rem, seg_rem); 1481 if (!desc) { 1482 /* fill main descriptor */ 1483 desc = &q->txq[q->head_idx]; 1484 elem = ionic_tx_sg_elems(q); 1485 desc_addr = frag_addr; 1486 desc_len = chunk_len; 1487 } else { 1488 /* fill sg descriptor */ 1489 elem->addr = cpu_to_le64(frag_addr); 1490 elem->len = cpu_to_le16(chunk_len); 1491 elem++; 1492 desc_nsge++; 1493 } 1494 frag_addr += chunk_len; 1495 frag_rem -= chunk_len; 1496 tso_rem -= chunk_len; 1497 seg_rem -= chunk_len; 1498 } 1499 seg_rem = min(tso_rem, mss); 1500 done = (tso_rem == 0); 1501 /* post descriptor */ 1502 ionic_tx_tso_post(netdev, q, desc, skb, desc_addr, desc_nsge, 1503 desc_len, hdrlen, mss, outer_csum, vlan_tci, 1504 has_vlan, start, done); 1505 start = false; 1506 /* Buffer information is stored with the first tso descriptor */ 1507 desc_info = &q->tx_info[q->head_idx]; 1508 desc_info->nbufs = 0; 1509 } 1510 1511 stats->pkts += DIV_ROUND_UP(len - hdrlen, mss); 1512 stats->bytes += len; 1513 stats->tso++; 1514 stats->tso_bytes = len; 1515 1516 return 0; 1517 } 1518 1519 static void ionic_tx_calc_csum(struct ionic_queue *q, struct sk_buff *skb, 1520 struct ionic_tx_desc_info *desc_info) 1521 { 1522 struct ionic_txq_desc *desc = &q->txq[q->head_idx]; 1523 struct ionic_buf_info *buf_info = desc_info->bufs; 1524 struct ionic_tx_stats *stats = q_to_tx_stats(q); 1525 bool has_vlan; 1526 u8 flags = 0; 1527 bool encap; 1528 u64 cmd; 1529 1530 has_vlan = !!skb_vlan_tag_present(skb); 1531 encap = skb->encapsulation; 1532 1533 flags |= has_vlan ? IONIC_TXQ_DESC_FLAG_VLAN : 0; 1534 flags |= encap ? IONIC_TXQ_DESC_FLAG_ENCAP : 0; 1535 1536 cmd = encode_txq_desc_cmd(IONIC_TXQ_DESC_OPCODE_CSUM_PARTIAL, 1537 flags, skb_shinfo(skb)->nr_frags, 1538 buf_info->dma_addr); 1539 desc->cmd = cpu_to_le64(cmd); 1540 desc->len = cpu_to_le16(buf_info->len); 1541 if (has_vlan) { 1542 desc->vlan_tci = cpu_to_le16(skb_vlan_tag_get(skb)); 1543 stats->vlan_inserted++; 1544 } else { 1545 desc->vlan_tci = 0; 1546 } 1547 desc->csum_start = cpu_to_le16(skb_checksum_start_offset(skb)); 1548 desc->csum_offset = cpu_to_le16(skb->csum_offset); 1549 1550 ionic_write_cmb_desc(q, desc); 1551 1552 if (skb_csum_is_sctp(skb)) 1553 stats->crc32_csum++; 1554 else 1555 stats->csum++; 1556 } 1557 1558 static void ionic_tx_calc_no_csum(struct ionic_queue *q, struct sk_buff *skb, 1559 struct ionic_tx_desc_info *desc_info) 1560 { 1561 struct ionic_txq_desc *desc = &q->txq[q->head_idx]; 1562 struct ionic_buf_info *buf_info = desc_info->bufs; 1563 struct ionic_tx_stats *stats = q_to_tx_stats(q); 1564 bool has_vlan; 1565 u8 flags = 0; 1566 bool encap; 1567 u64 cmd; 1568 1569 has_vlan = !!skb_vlan_tag_present(skb); 1570 encap = skb->encapsulation; 1571 1572 flags |= has_vlan ? IONIC_TXQ_DESC_FLAG_VLAN : 0; 1573 flags |= encap ? IONIC_TXQ_DESC_FLAG_ENCAP : 0; 1574 1575 cmd = encode_txq_desc_cmd(IONIC_TXQ_DESC_OPCODE_CSUM_NONE, 1576 flags, skb_shinfo(skb)->nr_frags, 1577 buf_info->dma_addr); 1578 desc->cmd = cpu_to_le64(cmd); 1579 desc->len = cpu_to_le16(buf_info->len); 1580 if (has_vlan) { 1581 desc->vlan_tci = cpu_to_le16(skb_vlan_tag_get(skb)); 1582 stats->vlan_inserted++; 1583 } else { 1584 desc->vlan_tci = 0; 1585 } 1586 desc->csum_start = 0; 1587 desc->csum_offset = 0; 1588 1589 ionic_write_cmb_desc(q, desc); 1590 1591 stats->csum_none++; 1592 } 1593 1594 static void ionic_tx_skb_frags(struct ionic_queue *q, struct sk_buff *skb, 1595 struct ionic_tx_desc_info *desc_info) 1596 { 1597 struct ionic_buf_info *buf_info = &desc_info->bufs[1]; 1598 struct ionic_tx_stats *stats = q_to_tx_stats(q); 1599 struct ionic_txq_sg_elem *elem; 1600 unsigned int i; 1601 1602 elem = ionic_tx_sg_elems(q); 1603 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++, buf_info++, elem++) { 1604 elem->addr = cpu_to_le64(buf_info->dma_addr); 1605 elem->len = cpu_to_le16(buf_info->len); 1606 } 1607 1608 stats->frags += skb_shinfo(skb)->nr_frags; 1609 } 1610 1611 static int ionic_tx(struct net_device *netdev, struct ionic_queue *q, 1612 struct sk_buff *skb) 1613 { 1614 struct ionic_tx_desc_info *desc_info = &q->tx_info[q->head_idx]; 1615 struct ionic_tx_stats *stats = q_to_tx_stats(q); 1616 bool ring_dbell = true; 1617 1618 if (unlikely(ionic_tx_map_skb(q, skb, desc_info))) 1619 return -EIO; 1620 1621 desc_info->skb = skb; 1622 1623 /* set up the initial descriptor */ 1624 if (skb->ip_summed == CHECKSUM_PARTIAL) 1625 ionic_tx_calc_csum(q, skb, desc_info); 1626 else 1627 ionic_tx_calc_no_csum(q, skb, desc_info); 1628 1629 /* add frags */ 1630 ionic_tx_skb_frags(q, skb, desc_info); 1631 1632 skb_tx_timestamp(skb); 1633 stats->pkts++; 1634 stats->bytes += skb->len; 1635 1636 if (likely(!ionic_txq_hwstamp_enabled(q))) { 1637 struct netdev_queue *ndq = q_to_ndq(netdev, q); 1638 1639 if (unlikely(!ionic_q_has_space(q, MAX_SKB_FRAGS + 1))) 1640 netif_tx_stop_queue(ndq); 1641 ring_dbell = __netdev_tx_sent_queue(ndq, skb->len, 1642 netdev_xmit_more()); 1643 } 1644 ionic_txq_post(q, ring_dbell); 1645 1646 return 0; 1647 } 1648 1649 static int ionic_tx_descs_needed(struct ionic_queue *q, struct sk_buff *skb) 1650 { 1651 int nr_frags = skb_shinfo(skb)->nr_frags; 1652 bool too_many_frags = false; 1653 skb_frag_t *frag; 1654 int desc_bufs; 1655 int chunk_len; 1656 int frag_rem; 1657 int tso_rem; 1658 int seg_rem; 1659 bool encap; 1660 int hdrlen; 1661 int ndescs; 1662 int err; 1663 1664 /* Each desc is mss long max, so a descriptor for each gso_seg */ 1665 if (skb_is_gso(skb)) { 1666 ndescs = skb_shinfo(skb)->gso_segs; 1667 if (!nr_frags) 1668 return ndescs; 1669 } else { 1670 ndescs = 1; 1671 if (!nr_frags) 1672 return ndescs; 1673 1674 if (unlikely(nr_frags > q->max_sg_elems)) { 1675 too_many_frags = true; 1676 goto linearize; 1677 } 1678 1679 return ndescs; 1680 } 1681 1682 /* We need to scan the skb to be sure that none of the MTU sized 1683 * packets in the TSO will require more sgs per descriptor than we 1684 * can support. We loop through the frags, add up the lengths for 1685 * a packet, and count the number of sgs used per packet. 1686 */ 1687 tso_rem = skb->len; 1688 frag = skb_shinfo(skb)->frags; 1689 encap = skb->encapsulation; 1690 1691 /* start with just hdr in first part of first descriptor */ 1692 if (encap) 1693 hdrlen = skb_inner_tcp_all_headers(skb); 1694 else 1695 hdrlen = skb_tcp_all_headers(skb); 1696 seg_rem = min_t(int, tso_rem, hdrlen + skb_shinfo(skb)->gso_size); 1697 frag_rem = hdrlen; 1698 1699 while (tso_rem > 0) { 1700 desc_bufs = 0; 1701 while (seg_rem > 0) { 1702 desc_bufs++; 1703 1704 /* We add the +1 because we can take buffers for one 1705 * more than we have SGs: one for the initial desc data 1706 * in addition to the SG segments that might follow. 1707 */ 1708 if (desc_bufs > q->max_sg_elems + 1) { 1709 too_many_frags = true; 1710 goto linearize; 1711 } 1712 1713 if (frag_rem == 0) { 1714 frag_rem = skb_frag_size(frag); 1715 frag++; 1716 } 1717 chunk_len = min(frag_rem, seg_rem); 1718 frag_rem -= chunk_len; 1719 tso_rem -= chunk_len; 1720 seg_rem -= chunk_len; 1721 } 1722 1723 seg_rem = min_t(int, tso_rem, skb_shinfo(skb)->gso_size); 1724 } 1725 1726 linearize: 1727 if (too_many_frags) { 1728 err = skb_linearize(skb); 1729 if (unlikely(err)) 1730 return err; 1731 q_to_tx_stats(q)->linearize++; 1732 } 1733 1734 return ndescs; 1735 } 1736 1737 static netdev_tx_t ionic_start_hwstamp_xmit(struct sk_buff *skb, 1738 struct net_device *netdev) 1739 { 1740 struct ionic_lif *lif = netdev_priv(netdev); 1741 struct ionic_queue *q; 1742 int err, ndescs; 1743 1744 /* Does not stop/start txq, because we post to a separate tx queue 1745 * for timestamping, and if a packet can't be posted immediately to 1746 * the timestamping queue, it is dropped. 1747 */ 1748 1749 q = &lif->hwstamp_txq->q; 1750 ndescs = ionic_tx_descs_needed(q, skb); 1751 if (unlikely(ndescs < 0)) 1752 goto err_out_drop; 1753 1754 if (unlikely(!ionic_q_has_space(q, ndescs))) 1755 goto err_out_drop; 1756 1757 skb_shinfo(skb)->tx_flags |= SKBTX_HW_TSTAMP; 1758 if (skb_is_gso(skb)) 1759 err = ionic_tx_tso(netdev, q, skb); 1760 else 1761 err = ionic_tx(netdev, q, skb); 1762 1763 if (unlikely(err)) 1764 goto err_out_drop; 1765 1766 return NETDEV_TX_OK; 1767 1768 err_out_drop: 1769 q->drop++; 1770 dev_kfree_skb(skb); 1771 return NETDEV_TX_OK; 1772 } 1773 1774 netdev_tx_t ionic_start_xmit(struct sk_buff *skb, struct net_device *netdev) 1775 { 1776 u16 queue_index = skb_get_queue_mapping(skb); 1777 struct ionic_lif *lif = netdev_priv(netdev); 1778 struct ionic_queue *q; 1779 int ndescs; 1780 int err; 1781 1782 if (unlikely(!test_bit(IONIC_LIF_F_UP, lif->state))) { 1783 dev_kfree_skb(skb); 1784 return NETDEV_TX_OK; 1785 } 1786 1787 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP)) 1788 if (lif->hwstamp_txq && lif->phc->ts_config_tx_mode) 1789 return ionic_start_hwstamp_xmit(skb, netdev); 1790 1791 if (unlikely(queue_index >= lif->nxqs)) 1792 queue_index = 0; 1793 q = &lif->txqcqs[queue_index]->q; 1794 1795 ndescs = ionic_tx_descs_needed(q, skb); 1796 if (ndescs < 0) 1797 goto err_out_drop; 1798 1799 if (!netif_txq_maybe_stop(q_to_ndq(netdev, q), 1800 ionic_q_space_avail(q), 1801 ndescs, ndescs)) 1802 return NETDEV_TX_BUSY; 1803 1804 if (skb_is_gso(skb)) 1805 err = ionic_tx_tso(netdev, q, skb); 1806 else 1807 err = ionic_tx(netdev, q, skb); 1808 1809 if (unlikely(err)) 1810 goto err_out_drop; 1811 1812 return NETDEV_TX_OK; 1813 1814 err_out_drop: 1815 q->drop++; 1816 dev_kfree_skb(skb); 1817 return NETDEV_TX_OK; 1818 } 1819