| /linux/drivers/net/wwan/t7xx/ |
| H A D | t7xx_hif_dpmaif_rx.c | 84 struct dpmaif_rx_queue *rxq = &dpmaif_ctrl->rxq[q_num]; in t7xx_dpmaif_update_bat_wr_idx() local 85 struct dpmaif_bat_request *bat_req = rxq->bat_req; in t7xx_dpmaif_update_bat_wr_idx() 88 if (!rxq->que_started) { in t7xx_dpmaif_update_bat_wr_idx() 89 dev_err(dpmaif_ctrl->dev, "RX queue %d has not been started\n", rxq->index); in t7xx_dpmaif_update_bat_wr_idx() 236 static int t7xx_dpmaifq_release_pit_entry(struct dpmaif_rx_queue *rxq, in t7xx_dpmaifq_release_pit_entry() argument 239 struct dpmaif_hw_info *hw_info = &rxq->dpmaif_ctrl->hw_info; in t7xx_dpmaifq_release_pit_entry() 243 if (!rxq->que_started) in t7xx_dpmaifq_release_pit_entry() 246 if (rel_entry_num >= rxq->pit_size_cnt) { in t7xx_dpmaifq_release_pit_entry() 247 dev_err(rxq->dpmaif_ctrl->dev, "Invalid PIT release index\n"); in t7xx_dpmaifq_release_pit_entry() 251 old_rel_idx = rxq->pit_release_rd_idx; in t7xx_dpmaifq_release_pit_entry() [all …]
|
| H A D | t7xx_hif_dpmaif.c | 227 rx_q = &dpmaif_ctrl->rxq[rx_idx]; in t7xx_dpmaif_rxtx_sw_allocs() 267 rx_q = &dpmaif_ctrl->rxq[i]; in t7xx_dpmaif_rxtx_sw_allocs() 294 rx_q = &dpmaif_ctrl->rxq[i]; in t7xx_dpmaif_sw_release() 303 struct dpmaif_rx_queue *rxq; in t7xx_dpmaif_start() local 314 rxq = &dpmaif_ctrl->rxq[i]; in t7xx_dpmaif_start() 315 rxq->que_started = true; in t7xx_dpmaif_start() 316 rxq->index = i; in t7xx_dpmaif_start() 317 rxq->budget = rxq->bat_req->bat_size_cnt - 1; in t7xx_dpmaif_start() 319 hw_init_para.pkt_bat_base_addr[i] = rxq->bat_req->bat_bus_addr; in t7xx_dpmaif_start() 320 hw_init_para.pkt_bat_size_cnt[i] = rxq->bat_req->bat_size_cnt; in t7xx_dpmaif_start() [all …]
|
| /linux/drivers/net/ethernet/huawei/hinic/ |
| H A D | hinic_rx.c | 53 static void hinic_rxq_clean_stats(struct hinic_rxq *rxq) in hinic_rxq_clean_stats() argument 55 struct hinic_rxq_stats *rxq_stats = &rxq->rxq_stats; in hinic_rxq_clean_stats() 71 void hinic_rxq_get_stats(struct hinic_rxq *rxq, struct hinic_rxq_stats *stats) in hinic_rxq_get_stats() argument 73 struct hinic_rxq_stats *rxq_stats = &rxq->rxq_stats; in hinic_rxq_get_stats() 91 static void rxq_stats_init(struct hinic_rxq *rxq) in rxq_stats_init() argument 93 struct hinic_rxq_stats *rxq_stats = &rxq->rxq_stats; in rxq_stats_init() 96 hinic_rxq_clean_stats(rxq); in rxq_stats_init() 99 static void rx_csum(struct hinic_rxq *rxq, u32 status, in rx_csum() argument 102 struct net_device *netdev = rxq->netdev; in rx_csum() 115 rxq->rxq_stats.csum_errors++; in rx_csum() [all …]
|
| /linux/drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/ |
| H A D | rx.c | 120 static int iwl_rxq_space(const struct iwl_rxq *rxq) in iwl_rxq_space() argument 123 WARN_ON(rxq->queue_size & (rxq->queue_size - 1)); in iwl_rxq_space() 131 return (rxq->read - rxq->write - 1) & (rxq->queue_size - 1); in iwl_rxq_space() 168 struct iwl_rxq *rxq) in iwl_pcie_rxq_inc_wr_ptr() argument 172 lockdep_assert_held(&rxq->lock); in iwl_pcie_rxq_inc_wr_ptr() 188 rxq->need_update = true; in iwl_pcie_rxq_inc_wr_ptr() 193 rxq->write_actual = round_down(rxq->write, 8); in iwl_pcie_rxq_inc_wr_ptr() 195 iwl_write32(trans, FH_RSCSR_CHNL0_WPTR, rxq->write_actual); in iwl_pcie_rxq_inc_wr_ptr() 197 iwl_write32(trans, HBUS_TARG_WRPTR, rxq->write_actual | in iwl_pcie_rxq_inc_wr_ptr() 198 HBUS_TARG_WRPTR_RX_Q(rxq->id)); in iwl_pcie_rxq_inc_wr_ptr() [all …]
|
| /linux/drivers/net/ethernet/huawei/hinic3/ |
| H A D | hinic3_rx.c | 53 static void hinic3_rxq_stats_init(struct hinic3_rxq *rxq) in hinic3_rxq_stats_init() argument 55 struct hinic3_rxq_stats *rxq_stats = &rxq->rxq_stats; in hinic3_rxq_stats_init() 66 struct hinic3_rxq *rxq; in hinic3_alloc_rxqs() local 74 rxq = &nic_dev->rxqs[q_id]; in hinic3_alloc_rxqs() 75 rxq->netdev = netdev; in hinic3_alloc_rxqs() 76 rxq->dev = &pdev->dev; in hinic3_alloc_rxqs() 77 rxq->q_id = q_id; in hinic3_alloc_rxqs() 78 rxq->buf_len = nic_dev->rx_buf_len; in hinic3_alloc_rxqs() 79 rxq->buf_len_shift = ilog2(nic_dev->rx_buf_len); in hinic3_alloc_rxqs() 80 rxq->q_depth = nic_dev->q_params.rq_depth; in hinic3_alloc_rxqs() [all …]
|
| H A D | hinic3_irq.c | 19 struct hinic3_rxq *rxq = irq_cfg->rxq; in hinic3_net_dim() local 26 dim_update_sample(irq_cfg->total_events, rxq->rxq_stats.packets, in hinic3_net_dim() 27 rxq->rxq_stats.bytes, &sample); in hinic3_net_dim() 28 net_dim(&rxq->dim, &sample); in hinic3_net_dim() 46 work_done = hinic3_rx_poll(irq_cfg->rxq, budget); in hinic3_poll() 184 struct hinic3_rxq *rxq; in hinic3_rx_dim_work() local 188 rxq = container_of(dim, struct hinic3_rxq, dim); in hinic3_rx_dim_work() 192 hinic3_update_queue_coal(rxq->netdev, rxq->q_id, in hinic3_rx_dim_work() 216 irq_cfg->rxq = &nic_dev->rxqs[q_id]; in hinic3_qps_irq_init() 231 INIT_WORK(&irq_cfg->rxq->dim.work, hinic3_rx_dim_work); in hinic3_qps_irq_init() [all …]
|
| /linux/drivers/infiniband/hw/hfi1/ |
| H A D | netdev_rx.c | 194 rx->rxq = kcalloc_node(rx->num_rx_q, sizeof(*rx->rxq), in hfi1_netdev_rxq_init() 197 if (!rx->rxq) { in hfi1_netdev_rxq_init() 203 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; in hfi1_netdev_rxq_init() local 205 rc = hfi1_netdev_allot_ctxt(rx, &rxq->rcd); in hfi1_netdev_rxq_init() 209 hfi1_rcd_get(rxq->rcd); in hfi1_netdev_rxq_init() 210 rxq->rx = rx; in hfi1_netdev_rxq_init() 211 rxq->rcd->napi = &rxq->napi; in hfi1_netdev_rxq_init() 213 i, rxq->rcd->ctxt); in hfi1_netdev_rxq_init() 218 set_bit(NAPI_STATE_NO_BUSY_POLL, &rxq->napi.state); in hfi1_netdev_rxq_init() 219 netif_napi_add(dev, &rxq->napi, hfi1_netdev_rx_napi); in hfi1_netdev_rxq_init() [all …]
|
| H A D | vnic_main.c | 292 static inline int hfi1_vnic_decap_skb(struct hfi1_vnic_rx_queue *rxq, in hfi1_vnic_decap_skb() argument 295 struct hfi1_vnic_vport_info *vinfo = rxq->vinfo; in hfi1_vnic_decap_skb() 303 vinfo->stats[rxq->idx].rx_oversize++; in hfi1_vnic_decap_skb() 305 vinfo->stats[rxq->idx].rx_runt++; in hfi1_vnic_decap_skb() 336 struct hfi1_vnic_rx_queue *rxq; in hfi1_vnic_bypass_rcv() local 370 rxq = &vinfo->rxq[q_idx]; in hfi1_vnic_bypass_rcv() 389 rc = hfi1_vnic_decap_skb(rxq, skb); in hfi1_vnic_bypass_rcv() 392 hfi1_vnic_update_rx_counters(vinfo, rxq->idx, skb, rc); in hfi1_vnic_bypass_rcv() 399 skb->protocol = eth_type_trans(skb, rxq->netdev); in hfi1_vnic_bypass_rcv() 401 napi_gro_receive(&rxq->napi, skb); in hfi1_vnic_bypass_rcv() [all …]
|
| /linux/drivers/net/ethernet/qlogic/qede/ |
| H A D | qede_fp.c | 28 int qede_alloc_rx_buffer(struct qede_rx_queue *rxq, bool allow_lazy) in qede_alloc_rx_buffer() argument 39 if (allow_lazy && likely(rxq->filled_buffers > 12)) { in qede_alloc_rx_buffer() 40 rxq->filled_buffers--; in qede_alloc_rx_buffer() 51 mapping = dma_map_page(rxq->dev, data, 0, in qede_alloc_rx_buffer() 52 PAGE_SIZE, rxq->data_direction); in qede_alloc_rx_buffer() 53 if (unlikely(dma_mapping_error(rxq->dev, mapping))) { in qede_alloc_rx_buffer() 58 sw_rx_data = &rxq->sw_rx_ring[rxq->sw_rx_prod & NUM_RX_BDS_MAX]; in qede_alloc_rx_buffer() 64 rx_bd = (struct eth_rx_bd *)qed_chain_produce(&rxq->rx_bd_ring); in qede_alloc_rx_buffer() 68 rxq->rx_headroom); in qede_alloc_rx_buffer() 70 rxq->sw_rx_prod++; in qede_alloc_rx_buffer() [all …]
|
| H A D | qede_main.c | 946 if (fp->rxq && xdp_rxq_info_is_reg(&fp->rxq->xdp_rxq)) in qede_free_fp_array() 947 xdp_rxq_info_unreg(&fp->rxq->xdp_rxq); in qede_free_fp_array() 948 kfree(fp->rxq); in qede_free_fp_array() 1014 fp->rxq = kzalloc_obj(*fp->rxq); in qede_alloc_fp_array() 1015 if (!fp->rxq) in qede_alloc_fp_array() 1506 struct qede_rx_queue *rxq) in qede_free_rx_buffers() argument 1510 for (i = rxq->sw_rx_cons; i != rxq->sw_rx_prod; i++) { in qede_free_rx_buffers() 1514 rx_buf = &rxq->sw_rx_ring[i & NUM_RX_BDS_MAX]; in qede_free_rx_buffers() 1518 rx_buf->mapping, PAGE_SIZE, rxq->data_direction); in qede_free_rx_buffers() 1525 static void qede_free_mem_rxq(struct qede_dev *edev, struct qede_rx_queue *rxq) in qede_free_mem_rxq() argument [all …]
|
| /linux/drivers/net/ethernet/intel/idpf/ |
| H A D | xdp.c | 11 int (*fn)(struct idpf_rx_queue *rxq, void *arg), in idpf_rxq_for_each() argument 33 q = &rx_qgrp->splitq.rxq_sets[j]->rxq; in idpf_rxq_for_each() 46 static int __idpf_xdp_rxq_info_init(struct idpf_rx_queue *rxq, void *arg) in __idpf_xdp_rxq_info_init() argument 48 const struct idpf_vport *vport = rxq->q_vector->vport; in __idpf_xdp_rxq_info_init() 54 if (idpf_queue_has(XSK, rxq)) in __idpf_xdp_rxq_info_init() 55 frag_size = rxq->bufq_sets[0].bufq.truesize; in __idpf_xdp_rxq_info_init() 57 err = __xdp_rxq_info_reg(&rxq->xdp_rxq, vport->netdev, rxq->idx, in __idpf_xdp_rxq_info_init() 58 rxq->q_vector->napi.napi_id, in __idpf_xdp_rxq_info_init() 66 if (idpf_queue_has(XSK, rxq)) { in __idpf_xdp_rxq_info_init() 67 err = xdp_rxq_info_reg_mem_model(&rxq->xdp_rxq, in __idpf_xdp_rxq_info_init() [all …]
|
| H A D | xsk.c | 13 struct idpf_rx_queue *rxq) in idpf_xsk_setup_rxq() argument 17 pool = xsk_get_pool_from_qid(vport->netdev, rxq->idx); in idpf_xsk_setup_rxq() 21 rxq->pool = pool; in idpf_xsk_setup_rxq() 23 idpf_queue_set(XSK, rxq); in idpf_xsk_setup_rxq() 38 qid = grp->splitq.rxq_sets[0]->rxq.idx; in idpf_xsk_setup_bufq() 128 struct idpf_rx_queue *rxq; in idpf_xsk_clear_queue() local 133 rxq = q; in idpf_xsk_clear_queue() 134 if (!idpf_queue_has_clear(XSK, rxq)) in idpf_xsk_clear_queue() 137 rxq->pool = NULL; in idpf_xsk_clear_queue() 467 int idpf_xskrq_poll(struct idpf_rx_queue *rxq, u32 budget) in idpf_xskrq_poll() argument [all …]
|
| H A D | idpf_txrx.c | 413 static void idpf_rx_buf_rel_all(struct idpf_rx_queue *rxq) in idpf_rx_buf_rel_all() argument 416 .fqes = rxq->rx_buf, in idpf_rx_buf_rel_all() 417 .pp = rxq->pp, in idpf_rx_buf_rel_all() 420 if (!rxq->rx_buf) in idpf_rx_buf_rel_all() 423 for (u32 i = 0; i < rxq->desc_count; i++) in idpf_rx_buf_rel_all() 424 idpf_rx_page_rel(&rxq->rx_buf[i]); in idpf_rx_buf_rel_all() 427 rxq->rx_buf = NULL; in idpf_rx_buf_rel_all() 428 rxq->pp = NULL; in idpf_rx_buf_rel_all() 439 static void idpf_rx_desc_rel(struct idpf_rx_queue *rxq, struct device *dev, in idpf_rx_desc_rel() argument 442 if (!rxq) in idpf_rx_desc_rel() [all …]
|
| /linux/drivers/net/ethernet/microsoft/mana/ |
| H A D | mana_en.c | 62 static void mana_put_rx_page(struct mana_rxq *rxq, struct page *page, in mana_put_rx_page() argument 66 page_pool_put_full_page(rxq->page_pool, page, false); in mana_put_rx_page() 702 static void *mana_get_rxbuf_pre(struct mana_rxq *rxq, dma_addr_t *da) in mana_get_rxbuf_pre() argument 704 struct net_device *ndev = rxq->ndev; in mana_get_rxbuf_pre() 716 if (mpc->rxbpre_datasize != rxq->datasize) { in mana_get_rxbuf_pre() 718 mpc->rxbpre_datasize, rxq->datasize); in mana_get_rxbuf_pre() 722 if (mpc->rxbpre_alloc_size != rxq->alloc_size) { in mana_get_rxbuf_pre() 724 mpc->rxbpre_alloc_size, rxq->alloc_size); in mana_get_rxbuf_pre() 728 if (mpc->rxbpre_headroom != rxq->headroom) { in mana_get_rxbuf_pre() 730 mpc->rxbpre_headroom, rxq->headroom); in mana_get_rxbuf_pre() [all …]
|
| /linux/drivers/net/ethernet/marvell/ |
| H A D | mvneta.c | 109 #define MVNETA_VLAN_PRIO_RXQ_MAP(prio, rxq) ((rxq) << ((prio) * 3)) argument 134 #define MVNETA_CPU_RXQ_ACCESS(rxq) BIT(rxq) argument 857 struct mvneta_rx_queue *rxq, in mvneta_rxq_non_occup_desc_add() argument 864 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), in mvneta_rxq_non_occup_desc_add() 870 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), in mvneta_rxq_non_occup_desc_add() 876 struct mvneta_rx_queue *rxq) in mvneta_rxq_busy_desc_num_get() argument 880 val = mvreg_read(pp, MVNETA_RXQ_STATUS_REG(rxq->id)); in mvneta_rxq_busy_desc_num_get() 888 struct mvneta_rx_queue *rxq, in mvneta_rxq_desc_num_update() argument 896 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), val); in mvneta_rxq_desc_num_update() 916 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), val); in mvneta_rxq_desc_num_update() [all …]
|
| H A D | mv643xx_eth.c | 399 struct rx_queue rxq[8]; member 441 static struct mv643xx_eth_private *rxq_to_mp(struct rx_queue *rxq) in rxq_to_mp() argument 443 return container_of(rxq, struct mv643xx_eth_private, rxq[rxq->index]); in rxq_to_mp() 451 static void rxq_enable(struct rx_queue *rxq) in rxq_enable() argument 453 struct mv643xx_eth_private *mp = rxq_to_mp(rxq); in rxq_enable() 454 wrlp(mp, RXQ_COMMAND, 1 << rxq->index); in rxq_enable() 457 static void rxq_disable(struct rx_queue *rxq) in rxq_disable() argument 459 struct mv643xx_eth_private *mp = rxq_to_mp(rxq); in rxq_disable() 460 u8 mask = 1 << rxq->index; in rxq_disable() 506 static int rxq_process(struct rx_queue *rxq, int budget) in rxq_process() argument [all …]
|
| /linux/drivers/net/ethernet/freescale/ |
| H A D | fec_main.c | 471 struct fec_enet_priv_rx_q *rxq) in fec_enet_create_page_pool() argument 477 .pool_size = rxq->bd.ring_size, in fec_enet_create_page_pool() 486 rxq->page_pool = page_pool_create(&pp_params); in fec_enet_create_page_pool() 487 if (IS_ERR(rxq->page_pool)) { in fec_enet_create_page_pool() 488 err = PTR_ERR(rxq->page_pool); in fec_enet_create_page_pool() 489 rxq->page_pool = NULL; in fec_enet_create_page_pool() 977 struct fec_enet_priv_rx_q *rxq; in fec_enet_bd_init() local 984 rxq = fep->rx_queue[q]; in fec_enet_bd_init() 985 bdp = rxq->bd.base; in fec_enet_bd_init() 987 for (i = 0; i < rxq->bd.ring_size; i++) { in fec_enet_bd_init() [all …]
|
| /linux/drivers/net/ethernet/atheros/alx/ |
| H A D | main.c | 73 struct alx_rx_queue *rxq = alx->qnapi[0]->rxq; in alx_refill_rx_ring() local 79 next = cur = rxq->write_idx; in alx_refill_rx_ring() 82 cur_buf = &rxq->bufs[cur]; in alx_refill_rx_ring() 84 while (!cur_buf->skb && next != rxq->read_idx) { in alx_refill_rx_ring() 85 struct alx_rfd *rfd = &rxq->rfd[cur]; in alx_refill_rx_ring() 127 cur_buf = &rxq->bufs[cur]; in alx_refill_rx_ring() 134 rxq->write_idx = cur; in alx_refill_rx_ring() 216 static int alx_clean_rx_irq(struct alx_rx_queue *rxq, int budget) in alx_clean_rx_irq() argument 225 alx = netdev_priv(rxq->netdev); in alx_clean_rx_irq() 228 rrd = &rxq->rrd[rxq->rrd_read_idx]; in alx_clean_rx_irq() [all …]
|
| /linux/drivers/net/wireless/intel/iwlegacy/ |
| H A D | 3945-mac.c | 929 struct il_rx_queue *rxq = &il->rxq; in il3945_rx_queue_restock() local 934 spin_lock_irqsave(&rxq->lock, flags); in il3945_rx_queue_restock() 935 while (il_rx_queue_space(rxq) > 0 && rxq->free_count) { in il3945_rx_queue_restock() 937 element = rxq->rx_free.next; in il3945_rx_queue_restock() 942 rxq->bd[rxq->write] = in il3945_rx_queue_restock() 944 rxq->queue[rxq->write] = rxb; in il3945_rx_queue_restock() 945 rxq->write = (rxq->write + 1) & RX_QUEUE_MASK; in il3945_rx_queue_restock() 946 rxq->free_count--; in il3945_rx_queue_restock() 948 spin_unlock_irqrestore(&rxq->lock, flags); in il3945_rx_queue_restock() 951 if (rxq->free_count <= RX_LOW_WATERMARK) in il3945_rx_queue_restock() [all …]
|
| /linux/tools/testing/selftests/bpf/ |
| H A D | xdp_hw_metadata.c | 73 int rxq; variable 444 static int verify_metadata(struct xsk *rx_xsk, int rxq, int server_fd, clockid_t clock_id) in verify_metadata() argument 447 struct pollfd fds[rxq + 1]; in verify_metadata() 455 for (i = 0; i < rxq; i++) { in verify_metadata() 461 fds[rxq].fd = server_fd; in verify_metadata() 462 fds[rxq].events = POLLIN; in verify_metadata() 463 fds[rxq].revents = 0; in verify_metadata() 468 for (i = 0; i < rxq; i++) { in verify_metadata() 474 ret = poll(fds, rxq + 1, 1000); in verify_metadata() 483 if (fds[rxq].revents) in verify_metadata() [all …]
|
| /linux/drivers/vhost/ |
| H A D | net.c | 131 struct vhost_net_buf rxq; member 154 static void *vhost_net_buf_get_ptr(struct vhost_net_buf *rxq) in vhost_net_buf_get_ptr() argument 156 if (rxq->tail != rxq->head) in vhost_net_buf_get_ptr() 157 return rxq->queue[rxq->head]; in vhost_net_buf_get_ptr() 162 static int vhost_net_buf_get_size(struct vhost_net_buf *rxq) in vhost_net_buf_get_size() argument 164 return rxq->tail - rxq->head; in vhost_net_buf_get_size() 167 static int vhost_net_buf_is_empty(struct vhost_net_buf *rxq) in vhost_net_buf_is_empty() argument 169 return rxq->tail == rxq->head; in vhost_net_buf_is_empty() 172 static void *vhost_net_buf_consume(struct vhost_net_buf *rxq) in vhost_net_buf_consume() argument 174 void *ret = vhost_net_buf_get_ptr(rxq); in vhost_net_buf_consume() [all …]
|
| /linux/drivers/bluetooth/ |
| H A D | btintel_pcie.c | 440 static void btintel_pcie_prepare_rx(struct rxq *rxq, u16 frbd_index) in btintel_pcie_prepare_rx() argument 446 buf = &rxq->bufs[frbd_index]; in btintel_pcie_prepare_rx() 448 frbd = &rxq->frbds[frbd_index]; in btintel_pcie_prepare_rx() 459 struct rxq *rxq = &data->rxq; in btintel_pcie_submit_rx() local 463 if (frbd_index > rxq->count) in btintel_pcie_submit_rx() 469 btintel_pcie_prepare_rx(rxq, frbd_index); in btintel_pcie_submit_rx() 471 frbd_index = (frbd_index + 1) % rxq->count; in btintel_pcie_submit_rx() 484 struct rxq *rxq = &data->rxq; in btintel_pcie_start_rx() local 490 for (i = 0; i < rxq->count - 3; i++) { in btintel_pcie_start_rx() 1389 struct rxq *rxq; in btintel_pcie_msix_rx_handle() local [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| H A D | en_arfs.c | 90 int rxq; member 436 arfs_rule->rxq, arfs_rule->flow_id, in arfs_may_expire_flow() 448 priv->channel_stats[arfs_rule->rxq]->rq.arfs_expired++; in arfs_may_expire_flow() 527 priv->channel_stats[arfs_rule->rxq]->rq.arfs_err++; in arfs_add_rule() 596 dest.tir_num = mlx5e_rx_res_get_tirn_direct(priv->rx_res, arfs_rule->rxq); in arfs_add_rule() 600 priv->channel_stats[arfs_rule->rxq]->rq.arfs_err++; in arfs_add_rule() 603 __func__, arfs_rule->filter_id, arfs_rule->rxq, in arfs_add_rule() 613 struct mlx5_flow_handle *rule, u16 rxq) in arfs_modify_rule_rq() argument 619 dst.tir_num = mlx5e_rx_res_get_tirn_direct(priv->rx_res, rxq); in arfs_modify_rule_rq() 622 priv->channel_stats[rxq]->rq.arfs_err++; in arfs_modify_rule_rq() [all …]
|
| /linux/drivers/net/ethernet/chelsio/cxgb4vf/ |
| H A D | cxgb4vf_main.c | 383 int rxq, msi, err; in request_msix_queue_irqs() local 397 for_each_ethrxq(s, rxq) { in request_msix_queue_irqs() 401 &s->ethrxq[rxq].rspq); in request_msix_queue_irqs() 409 while (--rxq >= 0) in request_msix_queue_irqs() 410 free_irq(adapter->msix_info[--msi].vec, &s->ethrxq[rxq].rspq); in request_msix_queue_irqs() 421 int rxq, msi; in free_msix_queue_irqs() local 425 for_each_ethrxq(s, rxq) in free_msix_queue_irqs() 427 &s->ethrxq[rxq].rspq); in free_msix_queue_irqs() 452 int rxq; in enable_rx() local 455 for_each_ethrxq(s, rxq) in enable_rx() [all …]
|
| H A D | sge.c | 1564 static void do_gro(struct sge_eth_rxq *rxq, const struct pkt_gl *gl, in do_gro() argument 1567 struct adapter *adapter = rxq->rspq.adapter; in do_gro() 1573 skb = napi_get_frags(&rxq->rspq.napi); in do_gro() 1576 rxq->stats.rx_drops++; in do_gro() 1585 skb_record_rx_queue(skb, rxq->rspq.idx); in do_gro() 1591 rxq->stats.vlan_ex++; in do_gro() 1593 ret = napi_gro_frags(&rxq->rspq.napi); in do_gro() 1596 rxq->stats.lro_pkts++; in do_gro() 1598 rxq->stats.lro_merged++; in do_gro() 1599 rxq->stats.pkts++; in do_gro() [all …]
|