| /linux/drivers/net/ethernet/intel/fm10k/ |
| H A D | fm10k_debugfs.c | 116 struct fm10k_q_vector *q_vector = ring->q_vector; in fm10k_dbg_desc_open() local 120 if (ring < q_vector->rx.ring) in fm10k_dbg_desc_open() 150 void fm10k_dbg_q_vector_init(struct fm10k_q_vector *q_vector) in fm10k_dbg_q_vector_init() argument 152 struct fm10k_intfc *interface = q_vector->interface; in fm10k_dbg_q_vector_init() 160 snprintf(name, sizeof(name), "q_vector.%03d", q_vector->v_idx); in fm10k_dbg_q_vector_init() 162 q_vector->dbg_q_vector = debugfs_create_dir(name, interface->dbg_intfc); in fm10k_dbg_q_vector_init() 165 for (i = 0; i < q_vector->tx.count; i++) { in fm10k_dbg_q_vector_init() 166 struct fm10k_ring *ring = &q_vector->tx.ring[i]; in fm10k_dbg_q_vector_init() 171 q_vector->dbg_q_vector, ring, in fm10k_dbg_q_vector_init() 176 for (i = 0; i < q_vector->rx.count; i++) { in fm10k_dbg_q_vector_init() [all …]
|
| H A D | fm10k.h | 104 struct fm10k_q_vector *q_vector;/* backpointer to host q_vector */ member 338 struct fm10k_q_vector *q_vector[MAX_Q_VECTORS]; member 545 void fm10k_dbg_q_vector_init(struct fm10k_q_vector *q_vector); 546 void fm10k_dbg_q_vector_exit(struct fm10k_q_vector *q_vector); 552 static inline void fm10k_dbg_q_vector_init(struct fm10k_q_vector *q_vector) {} in fm10k_dbg_q_vector_init() argument 553 static inline void fm10k_dbg_q_vector_exit(struct fm10k_q_vector *q_vector) {} in fm10k_dbg_q_vector_exit() argument
|
| /linux/drivers/net/ethernet/intel/idpf/ |
| H A D | idpf_txrx.c | 558 .xdp = idpf_xdp_enabled(bufq->q_vector->vport), in idpf_rx_hdr_buf_alloc_all() 559 .nid = idpf_q_vector_to_mem(bufq->q_vector), in idpf_rx_hdr_buf_alloc_all() 563 ret = libeth_rx_fq_create(&fq, &bufq->q_vector->napi); in idpf_rx_hdr_buf_alloc_all() 701 .nid = idpf_q_vector_to_mem(rxq->q_vector), in idpf_rx_bufs_init_singleq() 705 ret = libeth_rx_fq_create(&fq, &rxq->q_vector->napi); in idpf_rx_bufs_init_singleq() 760 .xdp = idpf_xdp_enabled(bufq->q_vector->vport), in idpf_rx_bufs_init() 761 .nid = idpf_q_vector_to_mem(bufq->q_vector), in idpf_rx_bufs_init() 768 ret = libeth_rx_fq_create(&fq, &bufq->q_vector->napi); in idpf_rx_bufs_init() 992 for (u32 j = 0; j < bufq->q_vector->num_bufq; j++) { in idpf_init_queue_set() 997 bufqs = bufq->q_vector in idpf_init_queue_set() 1185 struct idpf_q_vector *q_vector; idpf_qp_enable() local 1247 struct idpf_q_vector *q_vector; idpf_qp_disable() local 1281 struct idpf_q_vector *q_vector = idpf_find_rxq_vec(vport, qid); idpf_qp_switch() local 3760 struct idpf_q_vector *q_vector = (struct idpf_q_vector *)data; idpf_vport_intr_clean_queues() local 3802 struct idpf_q_vector *q_vector = &vport->q_vectors[v_idx]; idpf_vport_intr_rel() local 3820 idpf_q_vector_set_napi(struct idpf_q_vector * q_vector,bool link) idpf_q_vector_set_napi() argument 3844 struct idpf_q_vector *q_vector = &vport->q_vectors[vector]; idpf_vport_intr_rel_irq() local 3865 struct idpf_q_vector *q_vector = vport->q_vectors; idpf_vport_intr_dis_irq_all() local 3878 idpf_vport_intr_buildreg_itr(struct idpf_q_vector * q_vector) idpf_vport_intr_buildreg_itr() argument 3917 idpf_update_dim_sample(struct idpf_q_vector * q_vector,struct dim_sample * dim_sample,struct dim * dim,u64 packets,u64 bytes) idpf_update_dim_sample() argument 3941 idpf_net_dim(struct idpf_q_vector * q_vector) idpf_net_dim() argument 3992 idpf_vport_intr_update_itr_ena_irq(struct idpf_q_vector * q_vector) idpf_vport_intr_update_itr_ena_irq() argument 4019 struct idpf_q_vector *q_vector = &vport->q_vectors[vector]; idpf_vport_intr_req_irq() local 4066 idpf_vport_intr_write_itr(struct idpf_q_vector * q_vector,u16 itr,bool tx) idpf_vport_intr_write_itr() argument 4135 struct idpf_q_vector *q_vector; idpf_tx_dim_work() local 4161 struct idpf_q_vector *q_vector; idpf_rx_dim_work() local 4205 struct idpf_q_vector *q_vector = &vport->q_vectors[q_idx]; idpf_vport_intr_napi_ena_all() local 4291 struct idpf_q_vector *q_vector = idpf_vport_splitq_napi_poll() local 4488 struct idpf_q_vector *q_vector = &vport->q_vectors[v_idx]; idpf_vport_intr_napi_add_all() local 4509 struct idpf_q_vector *q_vector; idpf_vport_intr_alloc() local [all...] |
| H A D | idpf_txrx.h | 576 struct idpf_q_vector *q_vector; member 712 struct idpf_q_vector *q_vector; member 791 struct idpf_q_vector *q_vector; member 851 struct idpf_q_vector *q_vector; member 976 static inline int idpf_q_vector_to_mem(const struct idpf_q_vector *q_vector) in idpf_q_vector_to_mem() argument 980 if (!q_vector) in idpf_q_vector_to_mem() 983 cpu = cpumask_first(&q_vector->napi.config->affinity_mask); in idpf_q_vector_to_mem() 1047 static inline void idpf_vport_intr_set_wb_on_itr(struct idpf_q_vector *q_vector) in idpf_vport_intr_set_wb_on_itr() argument 1051 if (q_vector->wb_on_itr) in idpf_vport_intr_set_wb_on_itr() 1054 q_vector->wb_on_itr = true; in idpf_vport_intr_set_wb_on_itr() [all …]
|
| H A D | idpf_ethtool.c | 1304 return vport->rxq_grps->singleq.rxqs[q_num]->q_vector; in idpf_find_rxq_vec() 1309 return vport->rxq_grps[q_grp].splitq.rxq_sets[q_idx]->rxq.q_vector; in idpf_find_rxq_vec() 1325 return vport->txqs[q_num]->q_vector; in idpf_find_txq_vec() 1329 return vport->txq_grps[q_grp].complq->q_vector; in idpf_find_txq_vec() 1339 const struct idpf_q_vector *q_vector, in __idpf_get_q_coalesce() argument 1344 IDPF_ITR_IS_DYNAMIC(q_vector->rx_intr_mode); in __idpf_get_q_coalesce() 1345 ec->rx_coalesce_usecs = q_vector->rx_itr_value; in __idpf_get_q_coalesce() 1348 IDPF_ITR_IS_DYNAMIC(q_vector->tx_intr_mode); in __idpf_get_q_coalesce() 1349 ec->tx_coalesce_usecs = q_vector->tx_itr_value; in __idpf_get_q_coalesce()
|
| /linux/drivers/net/ethernet/intel/iavf/ |
| H A D | iavf_txrx.c | 157 static void iavf_force_wb(struct iavf_vsi *vsi, struct iavf_q_vector *q_vector) in iavf_force_wb() argument 166 IAVF_VFINT_DYN_CTLN1(q_vector->reg_idx), in iavf_force_wb() 209 iavf_force_wb(vsi, tx_ring->q_vector); in iavf_detect_recover_hung() 328 tx_ring->q_vector->tx.total_bytes += total_bytes; in iavf_clean_tx_irq() 329 tx_ring->q_vector->tx.total_packets += total_packets; in iavf_clean_tx_irq() 376 struct iavf_q_vector *q_vector) in iavf_enable_wb_on_itr() argument 378 u16 flags = q_vector->tx.ring[0].flags; in iavf_enable_wb_on_itr() 384 if (q_vector->arm_wb_state) in iavf_enable_wb_on_itr() 391 IAVF_VFINT_DYN_CTLN1(q_vector->reg_idx), val); in iavf_enable_wb_on_itr() 392 q_vector->arm_wb_state = true; in iavf_enable_wb_on_itr() [all …]
|
| H A D | iavf_main.c | 437 struct iavf_q_vector *q_vector = data; in iavf_msix_clean_rings() local 439 if (!q_vector->tx.ring && !q_vector->rx.ring) in iavf_msix_clean_rings() 442 napi_schedule_irqoff(&q_vector->napi); in iavf_msix_clean_rings() 456 struct iavf_q_vector *q_vector = &adapter->q_vectors[v_idx]; in iavf_map_vector_to_rxq() local 460 rx_ring->q_vector = q_vector; in iavf_map_vector_to_rxq() 461 rx_ring->next = q_vector->rx.ring; in iavf_map_vector_to_rxq() 463 q_vector->rx.ring = rx_ring; in iavf_map_vector_to_rxq() 464 q_vector->rx.count++; in iavf_map_vector_to_rxq() 465 q_vector->rx.next_update = jiffies + 1; in iavf_map_vector_to_rxq() 466 q_vector->rx.target_itr = ITR_TO_REG(rx_ring->itr_setting); in iavf_map_vector_to_rxq() [all …]
|
| /linux/drivers/net/ethernet/intel/ixgbevf/ |
| H A D | ixgbevf_main.c | 113 static void ixgbevf_set_itr(struct ixgbevf_q_vector *q_vector); 268 static bool ixgbevf_clean_tx_irq(struct ixgbevf_q_vector *q_vector, in ixgbevf_clean_tx_irq() argument 271 struct ixgbevf_adapter *adapter = q_vector->adapter; in ixgbevf_clean_tx_irq() 367 q_vector->tx.total_bytes += total_bytes; in ixgbevf_clean_tx_irq() 368 q_vector->tx.total_packets += total_packets; in ixgbevf_clean_tx_irq() 433 static void ixgbevf_rx_skb(struct ixgbevf_q_vector *q_vector, in ixgbevf_rx_skb() argument 436 napi_gro_receive(&q_vector->napi, skb); in ixgbevf_rx_skb() 883 skb = napi_alloc_skb(&rx_ring->q_vector->napi, IXGBEVF_RX_HDR_SIZE); in ixgbevf_construct_skb() 1116 static int ixgbevf_clean_rx_irq(struct ixgbevf_q_vector *q_vector, in ixgbevf_clean_rx_irq() argument 1121 struct ixgbevf_adapter *adapter = q_vector->adapter; in ixgbevf_clean_rx_irq() [all …]
|
| H A D | ixgbevf.h | 92 struct ixgbevf_q_vector *q_vector; /* backpointer to q_vector */ member 308 struct ixgbevf_q_vector *q_vector[MAX_MSIX_Q_VECTORS]; member 453 extern void ixgbevf_write_eitr(struct ixgbevf_q_vector *q_vector);
|
| /linux/drivers/net/ethernet/intel/ixgbe/ |
| H A D | ixgbe_txrx_common.h | 24 void ixgbe_rx_skb(struct ixgbe_q_vector *q_vector, 40 int ixgbe_clean_rx_irq_zc(struct ixgbe_q_vector *q_vector, 44 bool ixgbe_clean_xdp_tx_irq(struct ixgbe_q_vector *q_vector, 50 struct ixgbe_q_vector *q_vector, u64 pkts, 53 struct ixgbe_q_vector *q_vector, u64 pkts,
|
| H A D | ixgbe_xsk.c | 223 skb = napi_alloc_skb(&rx_ring->q_vector->napi, totalsize); in ixgbe_construct_skb_zc() 247 int ixgbe_clean_rx_irq_zc(struct ixgbe_q_vector *q_vector, in ixgbe_clean_rx_irq_zc() argument 252 struct ixgbe_adapter *adapter = q_vector->adapter; in ixgbe_clean_rx_irq_zc() 349 ixgbe_rx_skb(q_vector, skb); in ixgbe_clean_rx_irq_zc() 361 ixgbe_update_rx_ring_stats(rx_ring, q_vector, total_rx_packets, in ixgbe_clean_rx_irq_zc() 458 bool ixgbe_clean_xdp_tx_irq(struct ixgbe_q_vector *q_vector, in ixgbe_clean_xdp_tx_irq() argument 499 ixgbe_update_tx_ring_stats(tx_ring, q_vector, total_packets, in ixgbe_clean_xdp_tx_irq() 508 return ixgbe_xmit_zc(tx_ring, q_vector->tx.work_limit); in ixgbe_clean_xdp_tx_irq() 533 if (!napi_if_scheduled_mark_missed(&ring->q_vector->napi)) { in ixgbe_xsk_wakeup() 534 u64 eics = BIT_ULL(ring->q_vector->v_idx); in ixgbe_xsk_wakeup()
|
| H A D | ixgbe_main.c | 1171 struct ixgbe_q_vector *q_vector, u64 pkts, in ixgbe_update_tx_ring_stats() argument 1178 q_vector->tx.total_bytes += bytes; in ixgbe_update_tx_ring_stats() 1179 q_vector->tx.total_packets += pkts; in ixgbe_update_tx_ring_stats() 1190 struct ixgbe_q_vector *q_vector, u64 pkts, in ixgbe_update_rx_ring_stats() argument 1197 q_vector->rx.total_bytes += bytes; in ixgbe_update_rx_ring_stats() 1198 q_vector->rx.total_packets += pkts; in ixgbe_update_rx_ring_stats() 1349 static bool ixgbe_clean_tx_irq(struct ixgbe_q_vector *q_vector, in ixgbe_clean_tx_irq() argument 1352 struct ixgbe_adapter *adapter = q_vector->adapter; in ixgbe_clean_tx_irq() 1356 unsigned int budget = q_vector->tx.work_limit; in ixgbe_clean_tx_irq() 1445 ixgbe_update_tx_ring_stats(tx_ring, q_vector, total_packets, in ixgbe_clean_tx_irq() [all …]
|
| H A D | ixgbe_ptp.c | 880 void ixgbe_ptp_rx_pktstamp(struct ixgbe_q_vector *q_vector, in ixgbe_ptp_rx_pktstamp() argument 896 ixgbe_ptp_convert_to_hwtstamp(q_vector->adapter, skb_hwtstamps(skb), in ixgbe_ptp_rx_pktstamp() 909 void ixgbe_ptp_rx_rgtstamp(struct ixgbe_q_vector *q_vector, in ixgbe_ptp_rx_rgtstamp() argument 918 if (!q_vector || !q_vector->adapter) in ixgbe_ptp_rx_rgtstamp() 921 adapter = q_vector->adapter; in ixgbe_ptp_rx_rgtstamp()
|
| H A D | ixgbe.h | 358 struct ixgbe_q_vector *q_vector; /* backpointer to host q_vector */ member 725 struct ixgbe_q_vector *q_vector[MAX_Q_VECTORS]; member 1072 ixgbe_ptp_rx_pktstamp(rx_ring->q_vector, skb); in ixgbe_ptp_rx_hwtstamp() 1079 ixgbe_ptp_rx_rgtstamp(rx_ring->q_vector, skb); in ixgbe_ptp_rx_hwtstamp()
|
| /linux/drivers/net/ethernet/intel/i40e/ |
| H A D | i40e_txrx.c | 904 i40e_force_wb(vsi, tx_ring->q_vector); in i40e_detect_recover_hung() 1058 struct i40e_q_vector *q_vector) in i40e_enable_wb_on_itr() argument 1060 u16 flags = q_vector->tx.ring[0].flags; in i40e_enable_wb_on_itr() 1066 if (q_vector->arm_wb_state) in i40e_enable_wb_on_itr() 1074 I40E_PFINT_DYN_CTLN(q_vector->reg_idx), in i40e_enable_wb_on_itr() 1082 q_vector->arm_wb_state = true; in i40e_enable_wb_on_itr() 1091 void i40e_force_wb(struct i40e_vsi *vsi, struct i40e_q_vector *q_vector) in i40e_force_wb() argument 1101 I40E_PFINT_DYN_CTLN(q_vector->reg_idx), val); in i40e_force_wb() 1113 static inline bool i40e_container_is_rx(struct i40e_q_vector *q_vector, in i40e_container_is_rx() argument 1116 return &q_vector->rx == rc; in i40e_container_is_rx() [all …]
|
| H A D | i40e_txrx_common.h | 54 tx_ring->q_vector->tx.total_bytes += total_bytes; in i40e_update_tx_stats() 55 tx_ring->q_vector->tx.total_packets += total_packets; in i40e_update_tx_stats()
|
| H A D | i40e_main.c | 385 I40E_PFINT_DYN_CTLN(tx_ring->q_vector->v_idx + in i40e_tx_timeout() 3397 if (!ring->q_vector || !ring->netdev || ring->ch) in i40e_config_xps_tx_ring() 3404 cpu = cpumask_local_spread(ring->q_vector->v_idx, -1); in i40e_config_xps_tx_ring() 3583 ring->q_vector->napi.napi_id, in i40e_configure_rx_ring() 3595 ring->q_vector->napi.napi_id, in i40e_configure_rx_ring() 3879 struct i40e_q_vector *q_vector = vsi->q_vectors[i]; in i40e_vsi_configure_msix() local 3881 q_vector->rx.next_update = jiffies + 1; in i40e_vsi_configure_msix() 3882 q_vector->rx.target_itr = in i40e_vsi_configure_msix() 3885 q_vector->rx.target_itr >> 1); in i40e_vsi_configure_msix() 3886 q_vector->rx.current_itr = q_vector->rx.target_itr; in i40e_vsi_configure_msix() [all …]
|
| H A D | i40e_xsk.c | 305 skb = napi_alloc_skb(&rx_ring->q_vector->napi, totalsize); in i40e_construct_skb_zc() 390 napi_gro_receive(&rx_ring->q_vector->napi, skb); in i40e_handle_xdp_result_zc() 721 if (!napi_if_scheduled_mark_missed(&ring->q_vector->napi)) in i40e_xsk_wakeup() 722 i40e_force_wb(vsi, ring->q_vector); in i40e_xsk_wakeup()
|
| H A D | i40e_txrx.h | 394 struct i40e_q_vector *q_vector; /* Backreference to associated vector */ member 472 void i40e_force_wb(struct i40e_vsi *vsi, struct i40e_q_vector *q_vector);
|
| /linux/drivers/net/ethernet/wangxun/libwx/ |
| H A D | wx_vf_common.c | 82 struct wx_q_vector *q_vector = wx->q_vector[vector]; in wx_request_msix_irqs_vf() local 85 if (q_vector->tx.ring && q_vector->rx.ring) in wx_request_msix_irqs_vf() 86 snprintf(q_vector->name, sizeof(q_vector->name) - 1, in wx_request_msix_irqs_vf() 93 q_vector->name, q_vector); in wx_request_msix_irqs_vf() 96 q_vector->name, err); in wx_request_msix_irqs_vf() 114 wx->q_vector[vector]); in wx_request_msix_irqs_vf()
|
| /linux/drivers/net/ethernet/intel/igb/ |
| H A D | igb_xsk.c | 48 napi_disable(&rx_ring->q_vector->napi); in igb_txrx_ring_disable() 80 napi_enable(&rx_ring->q_vector->napi); in igb_txrx_ring_enable() 276 skb = napi_alloc_skb(&rx_ring->q_vector->napi, totalsize); in igb_construct_skb_zc() 341 int igb_clean_rx_irq_zc(struct igb_q_vector *q_vector, in igb_clean_rx_irq_zc() argument 344 struct igb_adapter *adapter = q_vector->adapter; in igb_clean_rx_irq_zc() 346 struct igb_ring *rx_ring = q_vector->rx.ring; in igb_clean_rx_irq_zc() 383 ts_hdr_len = igb_ptp_rx_pktstamp(rx_ring->q_vector, in igb_clean_rx_irq_zc() 436 napi_gro_receive(&q_vector->napi, skb); in igb_clean_rx_irq_zc() 447 igb_update_rx_stats(q_vector, total_packets, total_bytes); in igb_clean_rx_irq_zc() 551 if (!napi_if_scheduled_mark_missed(&ring->q_vector->napi)) { in igb_xsk_wakeup() [all …]
|
| H A D | igb_ethtool.c | 2230 struct igb_q_vector *q_vector = adapter->q_vector[i]; in igb_set_coalesce() local 2231 q_vector->tx.work_limit = adapter->tx_work_limit; in igb_set_coalesce() 2232 if (q_vector->rx.ring) in igb_set_coalesce() 2233 q_vector->itr_val = adapter->rx_itr_setting; in igb_set_coalesce() 2235 q_vector->itr_val = adapter->tx_itr_setting; in igb_set_coalesce() 2236 if (q_vector->itr_val && q_vector->itr_val <= 3) in igb_set_coalesce() 2237 q_vector->itr_val = IGB_START_ITR; in igb_set_coalesce() 2238 q_vector->set_itr = 1; in igb_set_coalesce()
|
| /linux/drivers/net/ethernet/intel/igc/ |
| H A D | igc_xdp.c | 32 napi_disable(&adapter->rx_ring[i]->q_vector->napi); in igc_xdp_set_prog() 47 napi_enable(&adapter->rx_ring[i]->q_vector->napi); in igc_xdp_set_prog() 92 napi = &rx_ring->q_vector->napi; in igc_xdp_enable_pool() 138 napi = &rx_ring->q_vector->napi; in igc_xdp_disable_pool()
|
| H A D | igc_ethtool.c | 946 struct igc_q_vector *q_vector = adapter->q_vector[i]; in igc_ethtool_set_coalesce() local 948 q_vector->tx.work_limit = adapter->tx_work_limit; in igc_ethtool_set_coalesce() 949 if (q_vector->rx.ring) in igc_ethtool_set_coalesce() 950 q_vector->itr_val = adapter->rx_itr_setting; in igc_ethtool_set_coalesce() 952 q_vector->itr_val = adapter->tx_itr_setting; in igc_ethtool_set_coalesce() 953 if (q_vector->itr_val && q_vector->itr_val <= 3) in igc_ethtool_set_coalesce() 954 q_vector->itr_val = IGC_START_ITR; in igc_ethtool_set_coalesce() 955 q_vector in igc_ethtool_set_coalesce() [all...] |
| /linux/drivers/net/ethernet/intel/ice/ |
| H A D | ice_lib.h | 95 void ice_write_intrl(struct ice_q_vector *q_vector, u8 intrl); 97 void ice_set_q_vector_intrl(struct ice_q_vector *q_vector);
|