Home
last modified time | relevance | path

Searched refs:vmdq (Results 1 – 10 of 10) sorted by relevance

/linux/drivers/net/ethernet/intel/ixgbe/
H A Dixgbe_lib.c22 struct ixgbe_ring_feature *vmdq = &adapter->ring_feature[RING_F_VMDQ]; in ixgbe_cache_ring_dcb_sriov() local
36 reg_idx = vmdq->offset * __ALIGN_MASK(1, ~vmdq->mask); in ixgbe_cache_ring_dcb_sriov()
39 if ((reg_idx & ~vmdq->mask) >= tcs) { in ixgbe_cache_ring_dcb_sriov()
41 reg_idx = __ALIGN_MASK(reg_idx, ~vmdq->mask); in ixgbe_cache_ring_dcb_sriov()
47 reg_idx = vmdq->offset * __ALIGN_MASK(1, ~vmdq->mask); in ixgbe_cache_ring_dcb_sriov()
50 if ((reg_idx & ~vmdq->mask) >= tcs) in ixgbe_cache_ring_dcb_sriov()
51 reg_idx = __ALIGN_MASK(reg_idx, ~vmdq->mask); in ixgbe_cache_ring_dcb_sriov()
66 u16 queues_per_pool = __ALIGN_MASK(1, ~vmdq->mask); in ixgbe_cache_ring_dcb_sriov()
69 reg_idx = (vmdq->offset + vmdq->indices) * queues_per_pool; in ixgbe_cache_ring_dcb_sriov()
71 reg_idx = __ALIGN_MASK(reg_idx, ~vmdq->mask) + fcoe_tc; in ixgbe_cache_ring_dcb_sriov()
[all …]
H A Dixgbe_common.h47 int ixgbe_set_rar_generic(struct ixgbe_hw *hw, u32 index, u8 *addr, u32 vmdq,
66 int ixgbe_set_vmdq_generic(struct ixgbe_hw *hw, u32 rar, u32 vmdq);
67 int ixgbe_set_vmdq_san_mac_generic(struct ixgbe_hw *hw, u32 vmdq);
68 int ixgbe_clear_vmdq_generic(struct ixgbe_hw *hw, u32 rar, u32 vmdq);
H A Dixgbe_sriov.c698 struct ixgbe_ring_feature *vmdq = &adapter->ring_feature[RING_F_VMDQ]; in ixgbe_vf_reset_event() local
700 u32 q_per_pool = __ALIGN_MASK(1, ~vmdq->mask); in ixgbe_vf_reset_event()
809 struct ixgbe_ring_feature *vmdq = &adapter->ring_feature[RING_F_VMDQ]; in ixgbe_write_qde() local
810 u32 q_per_pool = __ALIGN_MASK(1, ~vmdq->mask); in ixgbe_write_qde()
879 struct ixgbe_ring_feature *vmdq = &adapter->ring_feature[RING_F_VMDQ]; in ixgbe_vf_reset_msg() local
885 u32 q_per_pool = __ALIGN_MASK(1, ~vmdq->mask); in ixgbe_vf_reset_msg()
1060 struct ixgbe_ring_feature *vmdq = &adapter->ring_feature[RING_F_VMDQ]; in ixgbe_get_vf_queues() local
1077 msgbuf[IXGBE_VF_TX_QUEUES] = __ALIGN_MASK(1, ~vmdq->mask); in ixgbe_get_vf_queues()
1078 msgbuf[IXGBE_VF_RX_QUEUES] = __ALIGN_MASK(1, ~vmdq->mask); in ixgbe_get_vf_queues()
1572 struct ixgbe_ring_feature *vmdq = &adapter->ring_feature[RING_F_VMDQ]; in ixgbe_set_vf_rate_limit() local
[all …]
H A Dixgbe_common.c1826 int ixgbe_set_rar_generic(struct ixgbe_hw *hw, u32 index, u8 *addr, u32 vmdq, in ixgbe_set_rar_generic() argument
1839 hw->mac.ops.set_vmdq(hw, index, vmdq); in ixgbe_set_rar_generic()
2945 int ixgbe_clear_vmdq_generic(struct ixgbe_hw *hw, u32 rar, u32 vmdq) in ixgbe_clear_vmdq_generic() argument
2965 if (vmdq == IXGBE_CLEAR_VMDQ_ALL) { in ixgbe_clear_vmdq_generic()
2974 } else if (vmdq < 32) { in ixgbe_clear_vmdq_generic()
2975 mpsar_lo &= ~BIT(vmdq); in ixgbe_clear_vmdq_generic()
2978 mpsar_hi &= ~BIT(vmdq - 32); in ixgbe_clear_vmdq_generic()
2996 int ixgbe_set_vmdq_generic(struct ixgbe_hw *hw, u32 rar, u32 vmdq) in ixgbe_set_vmdq_generic() argument
3007 if (vmdq < 32) { in ixgbe_set_vmdq_generic()
3009 mpsar |= BIT(vmdq); in ixgbe_set_vmdq_generic()
[all …]
H A Dixgbe_main.c7668 struct ixgbe_ring_feature *vmdq = &adapter->ring_feature[RING_F_VMDQ]; in ixgbe_vf_tx_pending() local
7669 u32 q_per_pool = __ALIGN_MASK(1, ~vmdq->mask); in ixgbe_vf_tx_pending()
9441 struct ixgbe_ring_feature *vmdq = &adapter->ring_feature[RING_F_VMDQ]; in handle_redirect_action() local
9451 *queue = vf * __ALIGN_MASK(1, ~vmdq->mask); in handle_redirect_action()
/linux/drivers/net/ethernet/intel/ixgbevf/
H A Dvf.c433 u32 vmdq) in ixgbevf_set_rar_vf() argument
469 u32 vmdq) in ixgbevf_hv_set_rar_vf() argument
/linux/drivers/net/ethernet/intel/i40e/
H A Di40e_type.h221 bool vmdq; member
H A Di40e_common.c2900 p->vmdq = true; in i40e_parse_discover_capabilities()
H A Di40e_main.c12829 if (pf->hw.func_caps.vmdq && num_online_cpus() != 1) { in i40e_sw_init()
/linux/drivers/net/ethernet/wangxun/libwx/
H A Dwx_hw.c692 static int wx_clear_vmdq(struct wx *wx, u32 rar, u32 __maybe_unused vmdq) in wx_clear_vmdq() argument