Home
last modified time | relevance | path

Searched refs:rxq (Results 1 – 25 of 163) sorted by relevance

1234567

/linux/drivers/net/wwan/t7xx/
H A Dt7xx_hif_dpmaif_rx.c84 struct dpmaif_rx_queue *rxq = &dpmaif_ctrl->rxq[q_num]; in t7xx_dpmaif_update_bat_wr_idx() local
85 struct dpmaif_bat_request *bat_req = rxq->bat_req; in t7xx_dpmaif_update_bat_wr_idx()
88 if (!rxq->que_started) { in t7xx_dpmaif_update_bat_wr_idx()
89 dev_err(dpmaif_ctrl->dev, "RX queue %d has not been started\n", rxq->index); in t7xx_dpmaif_update_bat_wr_idx()
236 static int t7xx_dpmaifq_release_pit_entry(struct dpmaif_rx_queue *rxq, in t7xx_dpmaifq_release_pit_entry() argument
239 struct dpmaif_hw_info *hw_info = &rxq->dpmaif_ctrl->hw_info; in t7xx_dpmaifq_release_pit_entry()
243 if (!rxq->que_started) in t7xx_dpmaifq_release_pit_entry()
246 if (rel_entry_num >= rxq->pit_size_cnt) { in t7xx_dpmaifq_release_pit_entry()
247 dev_err(rxq->dpmaif_ctrl->dev, "Invalid PIT release index\n"); in t7xx_dpmaifq_release_pit_entry()
251 old_rel_idx = rxq->pit_release_rd_idx; in t7xx_dpmaifq_release_pit_entry()
[all …]
H A Dt7xx_hif_dpmaif.c227 rx_q = &dpmaif_ctrl->rxq[rx_idx]; in t7xx_dpmaif_rxtx_sw_allocs()
267 rx_q = &dpmaif_ctrl->rxq[i]; in t7xx_dpmaif_rxtx_sw_allocs()
294 rx_q = &dpmaif_ctrl->rxq[i]; in t7xx_dpmaif_sw_release()
303 struct dpmaif_rx_queue *rxq; in t7xx_dpmaif_start() local
314 rxq = &dpmaif_ctrl->rxq[i]; in t7xx_dpmaif_start()
315 rxq->que_started = true; in t7xx_dpmaif_start()
316 rxq->index = i; in t7xx_dpmaif_start()
317 rxq->budget = rxq->bat_req->bat_size_cnt - 1; in t7xx_dpmaif_start()
319 hw_init_para.pkt_bat_base_addr[i] = rxq->bat_req->bat_bus_addr; in t7xx_dpmaif_start()
320 hw_init_para.pkt_bat_size_cnt[i] = rxq->bat_req->bat_size_cnt; in t7xx_dpmaif_start()
[all …]
/linux/drivers/net/ethernet/huawei/hinic/
H A Dhinic_rx.c53 static void hinic_rxq_clean_stats(struct hinic_rxq *rxq) in hinic_rxq_clean_stats() argument
55 struct hinic_rxq_stats *rxq_stats = &rxq->rxq_stats; in hinic_rxq_clean_stats()
71 void hinic_rxq_get_stats(struct hinic_rxq *rxq, struct hinic_rxq_stats *stats) in hinic_rxq_get_stats() argument
73 struct hinic_rxq_stats *rxq_stats = &rxq->rxq_stats; in hinic_rxq_get_stats()
91 static void rxq_stats_init(struct hinic_rxq *rxq) in rxq_stats_init() argument
93 struct hinic_rxq_stats *rxq_stats = &rxq->rxq_stats; in rxq_stats_init()
96 hinic_rxq_clean_stats(rxq); in rxq_stats_init()
99 static void rx_csum(struct hinic_rxq *rxq, u32 status, in rx_csum() argument
102 struct net_device *netdev = rxq->netdev; in rx_csum()
115 rxq->rxq_stats.csum_errors++; in rx_csum()
[all …]
/linux/drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/
H A Drx.c120 static int iwl_rxq_space(const struct iwl_rxq *rxq) in iwl_rxq_space() argument
123 WARN_ON(rxq->queue_size & (rxq->queue_size - 1)); in iwl_rxq_space()
131 return (rxq->read - rxq->write - 1) & (rxq->queue_size - 1); in iwl_rxq_space()
168 struct iwl_rxq *rxq) in iwl_pcie_rxq_inc_wr_ptr() argument
172 lockdep_assert_held(&rxq->lock); in iwl_pcie_rxq_inc_wr_ptr()
188 rxq->need_update = true; in iwl_pcie_rxq_inc_wr_ptr()
193 rxq->write_actual = round_down(rxq->write, 8); in iwl_pcie_rxq_inc_wr_ptr()
195 iwl_write32(trans, FH_RSCSR_CHNL0_WPTR, rxq->write_actual); in iwl_pcie_rxq_inc_wr_ptr()
197 iwl_write32(trans, HBUS_TARG_WRPTR, rxq->write_actual | in iwl_pcie_rxq_inc_wr_ptr()
198 HBUS_TARG_WRPTR_RX_Q(rxq->id)); in iwl_pcie_rxq_inc_wr_ptr()
[all …]
/linux/drivers/net/ethernet/huawei/hinic3/
H A Dhinic3_rx.c53 static void hinic3_rxq_stats_init(struct hinic3_rxq *rxq) in hinic3_rxq_stats_init() argument
55 struct hinic3_rxq_stats *rxq_stats = &rxq->rxq_stats; in hinic3_rxq_stats_init()
66 struct hinic3_rxq *rxq; in hinic3_alloc_rxqs() local
74 rxq = &nic_dev->rxqs[q_id]; in hinic3_alloc_rxqs()
75 rxq->netdev = netdev; in hinic3_alloc_rxqs()
76 rxq->dev = &pdev->dev; in hinic3_alloc_rxqs()
77 rxq->q_id = q_id; in hinic3_alloc_rxqs()
78 rxq->buf_len = nic_dev->rx_buf_len; in hinic3_alloc_rxqs()
79 rxq->buf_len_shift = ilog2(nic_dev->rx_buf_len); in hinic3_alloc_rxqs()
80 rxq->q_depth = nic_dev->q_params.rq_depth; in hinic3_alloc_rxqs()
[all …]
H A Dhinic3_irq.c19 struct hinic3_rxq *rxq = irq_cfg->rxq; in hinic3_net_dim() local
26 dim_update_sample(irq_cfg->total_events, rxq->rxq_stats.packets, in hinic3_net_dim()
27 rxq->rxq_stats.bytes, &sample); in hinic3_net_dim()
28 net_dim(&rxq->dim, &sample); in hinic3_net_dim()
46 work_done = hinic3_rx_poll(irq_cfg->rxq, budget); in hinic3_poll()
184 struct hinic3_rxq *rxq; in hinic3_rx_dim_work() local
188 rxq = container_of(dim, struct hinic3_rxq, dim); in hinic3_rx_dim_work()
192 hinic3_update_queue_coal(rxq->netdev, rxq->q_id, in hinic3_rx_dim_work()
216 irq_cfg->rxq = &nic_dev->rxqs[q_id]; in hinic3_qps_irq_init()
231 INIT_WORK(&irq_cfg->rxq->dim.work, hinic3_rx_dim_work); in hinic3_qps_irq_init()
[all …]
/linux/drivers/infiniband/hw/hfi1/
H A Dnetdev_rx.c194 rx->rxq = kcalloc_node(rx->num_rx_q, sizeof(*rx->rxq), in hfi1_netdev_rxq_init()
197 if (!rx->rxq) { in hfi1_netdev_rxq_init()
203 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; in hfi1_netdev_rxq_init() local
205 rc = hfi1_netdev_allot_ctxt(rx, &rxq->rcd); in hfi1_netdev_rxq_init()
209 hfi1_rcd_get(rxq->rcd); in hfi1_netdev_rxq_init()
210 rxq->rx = rx; in hfi1_netdev_rxq_init()
211 rxq->rcd->napi = &rxq->napi; in hfi1_netdev_rxq_init()
213 i, rxq->rcd->ctxt); in hfi1_netdev_rxq_init()
218 set_bit(NAPI_STATE_NO_BUSY_POLL, &rxq->napi.state); in hfi1_netdev_rxq_init()
219 netif_napi_add(dev, &rxq->napi, hfi1_netdev_rx_napi); in hfi1_netdev_rxq_init()
[all …]
H A Dvnic_main.c292 static inline int hfi1_vnic_decap_skb(struct hfi1_vnic_rx_queue *rxq, in hfi1_vnic_decap_skb() argument
295 struct hfi1_vnic_vport_info *vinfo = rxq->vinfo; in hfi1_vnic_decap_skb()
303 vinfo->stats[rxq->idx].rx_oversize++; in hfi1_vnic_decap_skb()
305 vinfo->stats[rxq->idx].rx_runt++; in hfi1_vnic_decap_skb()
336 struct hfi1_vnic_rx_queue *rxq; in hfi1_vnic_bypass_rcv() local
370 rxq = &vinfo->rxq[q_idx]; in hfi1_vnic_bypass_rcv()
389 rc = hfi1_vnic_decap_skb(rxq, skb); in hfi1_vnic_bypass_rcv()
392 hfi1_vnic_update_rx_counters(vinfo, rxq->idx, skb, rc); in hfi1_vnic_bypass_rcv()
399 skb->protocol = eth_type_trans(skb, rxq->netdev); in hfi1_vnic_bypass_rcv()
401 napi_gro_receive(&rxq->napi, skb); in hfi1_vnic_bypass_rcv()
[all …]
/linux/drivers/net/ethernet/qlogic/qede/
H A Dqede_fp.c28 int qede_alloc_rx_buffer(struct qede_rx_queue *rxq, bool allow_lazy) in qede_alloc_rx_buffer() argument
39 if (allow_lazy && likely(rxq->filled_buffers > 12)) { in qede_alloc_rx_buffer()
40 rxq->filled_buffers--; in qede_alloc_rx_buffer()
51 mapping = dma_map_page(rxq->dev, data, 0, in qede_alloc_rx_buffer()
52 PAGE_SIZE, rxq->data_direction); in qede_alloc_rx_buffer()
53 if (unlikely(dma_mapping_error(rxq->dev, mapping))) { in qede_alloc_rx_buffer()
58 sw_rx_data = &rxq->sw_rx_ring[rxq->sw_rx_prod & NUM_RX_BDS_MAX]; in qede_alloc_rx_buffer()
64 rx_bd = (struct eth_rx_bd *)qed_chain_produce(&rxq->rx_bd_ring); in qede_alloc_rx_buffer()
68 rxq->rx_headroom); in qede_alloc_rx_buffer()
70 rxq->sw_rx_prod++; in qede_alloc_rx_buffer()
[all …]
H A Dqede_main.c946 if (fp->rxq && xdp_rxq_info_is_reg(&fp->rxq->xdp_rxq)) in qede_free_fp_array()
947 xdp_rxq_info_unreg(&fp->rxq->xdp_rxq); in qede_free_fp_array()
948 kfree(fp->rxq); in qede_free_fp_array()
1014 fp->rxq = kzalloc_obj(*fp->rxq); in qede_alloc_fp_array()
1015 if (!fp->rxq) in qede_alloc_fp_array()
1506 struct qede_rx_queue *rxq) in qede_free_rx_buffers() argument
1510 for (i = rxq->sw_rx_cons; i != rxq->sw_rx_prod; i++) { in qede_free_rx_buffers()
1514 rx_buf = &rxq->sw_rx_ring[i & NUM_RX_BDS_MAX]; in qede_free_rx_buffers()
1518 rx_buf->mapping, PAGE_SIZE, rxq->data_direction); in qede_free_rx_buffers()
1525 static void qede_free_mem_rxq(struct qede_dev *edev, struct qede_rx_queue *rxq) in qede_free_mem_rxq() argument
[all …]
/linux/drivers/net/ethernet/intel/idpf/
H A Dxdp.c11 int (*fn)(struct idpf_rx_queue *rxq, void *arg), in idpf_rxq_for_each() argument
33 q = &rx_qgrp->splitq.rxq_sets[j]->rxq; in idpf_rxq_for_each()
46 static int __idpf_xdp_rxq_info_init(struct idpf_rx_queue *rxq, void *arg) in __idpf_xdp_rxq_info_init() argument
48 const struct idpf_vport *vport = rxq->q_vector->vport; in __idpf_xdp_rxq_info_init()
54 if (idpf_queue_has(XSK, rxq)) in __idpf_xdp_rxq_info_init()
55 frag_size = rxq->bufq_sets[0].bufq.truesize; in __idpf_xdp_rxq_info_init()
57 err = __xdp_rxq_info_reg(&rxq->xdp_rxq, vport->netdev, rxq->idx, in __idpf_xdp_rxq_info_init()
58 rxq->q_vector->napi.napi_id, in __idpf_xdp_rxq_info_init()
66 if (idpf_queue_has(XSK, rxq)) { in __idpf_xdp_rxq_info_init()
67 err = xdp_rxq_info_reg_mem_model(&rxq->xdp_rxq, in __idpf_xdp_rxq_info_init()
[all …]
H A Dxsk.c13 struct idpf_rx_queue *rxq) in idpf_xsk_setup_rxq() argument
17 pool = xsk_get_pool_from_qid(vport->netdev, rxq->idx); in idpf_xsk_setup_rxq()
21 rxq->pool = pool; in idpf_xsk_setup_rxq()
23 idpf_queue_set(XSK, rxq); in idpf_xsk_setup_rxq()
38 qid = grp->splitq.rxq_sets[0]->rxq.idx; in idpf_xsk_setup_bufq()
128 struct idpf_rx_queue *rxq; in idpf_xsk_clear_queue() local
133 rxq = q; in idpf_xsk_clear_queue()
134 if (!idpf_queue_has_clear(XSK, rxq)) in idpf_xsk_clear_queue()
137 rxq->pool = NULL; in idpf_xsk_clear_queue()
467 int idpf_xskrq_poll(struct idpf_rx_queue *rxq, u32 budget) in idpf_xskrq_poll() argument
[all …]
H A Didpf_txrx.c413 static void idpf_rx_buf_rel_all(struct idpf_rx_queue *rxq) in idpf_rx_buf_rel_all() argument
416 .fqes = rxq->rx_buf, in idpf_rx_buf_rel_all()
417 .pp = rxq->pp, in idpf_rx_buf_rel_all()
420 if (!rxq->rx_buf) in idpf_rx_buf_rel_all()
423 for (u32 i = 0; i < rxq->desc_count; i++) in idpf_rx_buf_rel_all()
424 idpf_rx_page_rel(&rxq->rx_buf[i]); in idpf_rx_buf_rel_all()
427 rxq->rx_buf = NULL; in idpf_rx_buf_rel_all()
428 rxq->pp = NULL; in idpf_rx_buf_rel_all()
439 static void idpf_rx_desc_rel(struct idpf_rx_queue *rxq, struct device *dev, in idpf_rx_desc_rel() argument
442 if (!rxq) in idpf_rx_desc_rel()
[all …]
/linux/drivers/net/ethernet/microsoft/mana/
H A Dmana_en.c62 static void mana_put_rx_page(struct mana_rxq *rxq, struct page *page, in mana_put_rx_page() argument
66 page_pool_put_full_page(rxq->page_pool, page, false); in mana_put_rx_page()
702 static void *mana_get_rxbuf_pre(struct mana_rxq *rxq, dma_addr_t *da) in mana_get_rxbuf_pre() argument
704 struct net_device *ndev = rxq->ndev; in mana_get_rxbuf_pre()
716 if (mpc->rxbpre_datasize != rxq->datasize) { in mana_get_rxbuf_pre()
718 mpc->rxbpre_datasize, rxq->datasize); in mana_get_rxbuf_pre()
722 if (mpc->rxbpre_alloc_size != rxq->alloc_size) { in mana_get_rxbuf_pre()
724 mpc->rxbpre_alloc_size, rxq->alloc_size); in mana_get_rxbuf_pre()
728 if (mpc->rxbpre_headroom != rxq->headroom) { in mana_get_rxbuf_pre()
730 mpc->rxbpre_headroom, rxq->headroom); in mana_get_rxbuf_pre()
[all …]
/linux/drivers/net/ethernet/marvell/
H A Dmvneta.c109 #define MVNETA_VLAN_PRIO_RXQ_MAP(prio, rxq) ((rxq) << ((prio) * 3)) argument
134 #define MVNETA_CPU_RXQ_ACCESS(rxq) BIT(rxq) argument
857 struct mvneta_rx_queue *rxq, in mvneta_rxq_non_occup_desc_add() argument
864 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), in mvneta_rxq_non_occup_desc_add()
870 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), in mvneta_rxq_non_occup_desc_add()
876 struct mvneta_rx_queue *rxq) in mvneta_rxq_busy_desc_num_get() argument
880 val = mvreg_read(pp, MVNETA_RXQ_STATUS_REG(rxq->id)); in mvneta_rxq_busy_desc_num_get()
888 struct mvneta_rx_queue *rxq, in mvneta_rxq_desc_num_update() argument
896 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), val); in mvneta_rxq_desc_num_update()
916 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), val); in mvneta_rxq_desc_num_update()
[all …]
H A Dmv643xx_eth.c399 struct rx_queue rxq[8]; member
441 static struct mv643xx_eth_private *rxq_to_mp(struct rx_queue *rxq) in rxq_to_mp() argument
443 return container_of(rxq, struct mv643xx_eth_private, rxq[rxq->index]); in rxq_to_mp()
451 static void rxq_enable(struct rx_queue *rxq) in rxq_enable() argument
453 struct mv643xx_eth_private *mp = rxq_to_mp(rxq); in rxq_enable()
454 wrlp(mp, RXQ_COMMAND, 1 << rxq->index); in rxq_enable()
457 static void rxq_disable(struct rx_queue *rxq) in rxq_disable() argument
459 struct mv643xx_eth_private *mp = rxq_to_mp(rxq); in rxq_disable()
460 u8 mask = 1 << rxq->index; in rxq_disable()
506 static int rxq_process(struct rx_queue *rxq, int budget) in rxq_process() argument
[all …]
/linux/drivers/net/ethernet/freescale/
H A Dfec_main.c471 struct fec_enet_priv_rx_q *rxq) in fec_enet_create_page_pool() argument
477 .pool_size = rxq->bd.ring_size, in fec_enet_create_page_pool()
486 rxq->page_pool = page_pool_create(&pp_params); in fec_enet_create_page_pool()
487 if (IS_ERR(rxq->page_pool)) { in fec_enet_create_page_pool()
488 err = PTR_ERR(rxq->page_pool); in fec_enet_create_page_pool()
489 rxq->page_pool = NULL; in fec_enet_create_page_pool()
977 struct fec_enet_priv_rx_q *rxq; in fec_enet_bd_init() local
984 rxq = fep->rx_queue[q]; in fec_enet_bd_init()
985 bdp = rxq->bd.base; in fec_enet_bd_init()
987 for (i = 0; i < rxq->bd.ring_size; i++) { in fec_enet_bd_init()
[all …]
/linux/drivers/net/ethernet/atheros/alx/
H A Dmain.c73 struct alx_rx_queue *rxq = alx->qnapi[0]->rxq; in alx_refill_rx_ring() local
79 next = cur = rxq->write_idx; in alx_refill_rx_ring()
82 cur_buf = &rxq->bufs[cur]; in alx_refill_rx_ring()
84 while (!cur_buf->skb && next != rxq->read_idx) { in alx_refill_rx_ring()
85 struct alx_rfd *rfd = &rxq->rfd[cur]; in alx_refill_rx_ring()
127 cur_buf = &rxq->bufs[cur]; in alx_refill_rx_ring()
134 rxq->write_idx = cur; in alx_refill_rx_ring()
216 static int alx_clean_rx_irq(struct alx_rx_queue *rxq, int budget) in alx_clean_rx_irq() argument
225 alx = netdev_priv(rxq->netdev); in alx_clean_rx_irq()
228 rrd = &rxq->rrd[rxq->rrd_read_idx]; in alx_clean_rx_irq()
[all …]
/linux/drivers/net/wireless/intel/iwlegacy/
H A D3945-mac.c929 struct il_rx_queue *rxq = &il->rxq; in il3945_rx_queue_restock() local
934 spin_lock_irqsave(&rxq->lock, flags); in il3945_rx_queue_restock()
935 while (il_rx_queue_space(rxq) > 0 && rxq->free_count) { in il3945_rx_queue_restock()
937 element = rxq->rx_free.next; in il3945_rx_queue_restock()
942 rxq->bd[rxq->write] = in il3945_rx_queue_restock()
944 rxq->queue[rxq->write] = rxb; in il3945_rx_queue_restock()
945 rxq->write = (rxq->write + 1) & RX_QUEUE_MASK; in il3945_rx_queue_restock()
946 rxq->free_count--; in il3945_rx_queue_restock()
948 spin_unlock_irqrestore(&rxq->lock, flags); in il3945_rx_queue_restock()
951 if (rxq->free_count <= RX_LOW_WATERMARK) in il3945_rx_queue_restock()
[all …]
/linux/tools/testing/selftests/bpf/
H A Dxdp_hw_metadata.c73 int rxq; variable
444 static int verify_metadata(struct xsk *rx_xsk, int rxq, int server_fd, clockid_t clock_id) in verify_metadata() argument
447 struct pollfd fds[rxq + 1]; in verify_metadata()
455 for (i = 0; i < rxq; i++) { in verify_metadata()
461 fds[rxq].fd = server_fd; in verify_metadata()
462 fds[rxq].events = POLLIN; in verify_metadata()
463 fds[rxq].revents = 0; in verify_metadata()
468 for (i = 0; i < rxq; i++) { in verify_metadata()
474 ret = poll(fds, rxq + 1, 1000); in verify_metadata()
483 if (fds[rxq].revents) in verify_metadata()
[all …]
/linux/drivers/vhost/
H A Dnet.c131 struct vhost_net_buf rxq; member
154 static void *vhost_net_buf_get_ptr(struct vhost_net_buf *rxq) in vhost_net_buf_get_ptr() argument
156 if (rxq->tail != rxq->head) in vhost_net_buf_get_ptr()
157 return rxq->queue[rxq->head]; in vhost_net_buf_get_ptr()
162 static int vhost_net_buf_get_size(struct vhost_net_buf *rxq) in vhost_net_buf_get_size() argument
164 return rxq->tail - rxq->head; in vhost_net_buf_get_size()
167 static int vhost_net_buf_is_empty(struct vhost_net_buf *rxq) in vhost_net_buf_is_empty() argument
169 return rxq->tail == rxq->head; in vhost_net_buf_is_empty()
172 static void *vhost_net_buf_consume(struct vhost_net_buf *rxq) in vhost_net_buf_consume() argument
174 void *ret = vhost_net_buf_get_ptr(rxq); in vhost_net_buf_consume()
[all …]
/linux/drivers/bluetooth/
H A Dbtintel_pcie.c440 static void btintel_pcie_prepare_rx(struct rxq *rxq, u16 frbd_index) in btintel_pcie_prepare_rx() argument
446 buf = &rxq->bufs[frbd_index]; in btintel_pcie_prepare_rx()
448 frbd = &rxq->frbds[frbd_index]; in btintel_pcie_prepare_rx()
459 struct rxq *rxq = &data->rxq; in btintel_pcie_submit_rx() local
463 if (frbd_index > rxq->count) in btintel_pcie_submit_rx()
469 btintel_pcie_prepare_rx(rxq, frbd_index); in btintel_pcie_submit_rx()
471 frbd_index = (frbd_index + 1) % rxq->count; in btintel_pcie_submit_rx()
484 struct rxq *rxq = &data->rxq; in btintel_pcie_start_rx() local
490 for (i = 0; i < rxq->count - 3; i++) { in btintel_pcie_start_rx()
1389 struct rxq *rxq; in btintel_pcie_msix_rx_handle() local
[all …]
/linux/drivers/net/ethernet/mellanox/mlx5/core/
H A Den_arfs.c90 int rxq; member
436 arfs_rule->rxq, arfs_rule->flow_id, in arfs_may_expire_flow()
448 priv->channel_stats[arfs_rule->rxq]->rq.arfs_expired++; in arfs_may_expire_flow()
527 priv->channel_stats[arfs_rule->rxq]->rq.arfs_err++; in arfs_add_rule()
596 dest.tir_num = mlx5e_rx_res_get_tirn_direct(priv->rx_res, arfs_rule->rxq); in arfs_add_rule()
600 priv->channel_stats[arfs_rule->rxq]->rq.arfs_err++; in arfs_add_rule()
603 __func__, arfs_rule->filter_id, arfs_rule->rxq, in arfs_add_rule()
613 struct mlx5_flow_handle *rule, u16 rxq) in arfs_modify_rule_rq() argument
619 dst.tir_num = mlx5e_rx_res_get_tirn_direct(priv->rx_res, rxq); in arfs_modify_rule_rq()
622 priv->channel_stats[rxq]->rq.arfs_err++; in arfs_modify_rule_rq()
[all …]
/linux/drivers/net/ethernet/chelsio/cxgb4vf/
H A Dcxgb4vf_main.c383 int rxq, msi, err; in request_msix_queue_irqs() local
397 for_each_ethrxq(s, rxq) { in request_msix_queue_irqs()
401 &s->ethrxq[rxq].rspq); in request_msix_queue_irqs()
409 while (--rxq >= 0) in request_msix_queue_irqs()
410 free_irq(adapter->msix_info[--msi].vec, &s->ethrxq[rxq].rspq); in request_msix_queue_irqs()
421 int rxq, msi; in free_msix_queue_irqs() local
425 for_each_ethrxq(s, rxq) in free_msix_queue_irqs()
427 &s->ethrxq[rxq].rspq); in free_msix_queue_irqs()
452 int rxq; in enable_rx() local
455 for_each_ethrxq(s, rxq) in enable_rx()
[all …]
H A Dsge.c1564 static void do_gro(struct sge_eth_rxq *rxq, const struct pkt_gl *gl, in do_gro() argument
1567 struct adapter *adapter = rxq->rspq.adapter; in do_gro()
1573 skb = napi_get_frags(&rxq->rspq.napi); in do_gro()
1576 rxq->stats.rx_drops++; in do_gro()
1585 skb_record_rx_queue(skb, rxq->rspq.idx); in do_gro()
1591 rxq->stats.vlan_ex++; in do_gro()
1593 ret = napi_gro_frags(&rxq->rspq.napi); in do_gro()
1596 rxq->stats.lro_pkts++; in do_gro()
1598 rxq->stats.lro_merged++; in do_gro()
1599 rxq->stats.pkts++; in do_gro()
[all …]

1234567