Lines Matching refs:i40e
603 i40e_alloc_dma_buffer(i40e_t *i40e, i40e_dma_buffer_t *dmap, in i40e_alloc_dma_buffer() argument
621 ret = ddi_dma_alloc_handle(i40e->i40e_dip, attrsp, DDI_DMA_DONTWAIT, in i40e_alloc_dma_buffer()
624 i40e_error(i40e, "failed to allocate dma handle for I/O " in i40e_alloc_dma_buffer()
637 i40e_error(i40e, "failed to allocate %ld bytes of DMA for I/O " in i40e_alloc_dma_buffer()
658 i40e_error(i40e, "failed to allocate %ld bytes of DMA for I/O " in i40e_alloc_dma_buffer()
705 i40e_alloc_rx_data(i40e_t *i40e, i40e_trqpair_t *itrq) in i40e_alloc_rx_data() argument
713 rxd->rxd_i40e = i40e; in i40e_alloc_rx_data()
715 rxd->rxd_ring_size = i40e->i40e_rx_ring_size; in i40e_alloc_rx_data()
716 rxd->rxd_free_list_size = i40e->i40e_rx_ring_size; in i40e_alloc_rx_data()
723 i40e_error(i40e, "failed to allocate RX work list for a ring " in i40e_alloc_rx_data()
732 i40e_error(i40e, "failed to allocate a %d entry RX free list " in i40e_alloc_rx_data()
740 i40e_error(i40e, "failed to allocate a %d entry rcb area for " in i40e_alloc_rx_data()
769 i40e_t *i40e = rxd->rxd_i40e; in i40e_free_rx_dma() local
775 mutex_enter(&i40e->i40e_rx_pending_lock); in i40e_free_rx_dma()
801 atomic_inc_32(&i40e->i40e_rx_pending); in i40e_free_rx_dma()
804 mutex_exit(&i40e->i40e_rx_pending_lock); in i40e_free_rx_dma()
817 i40e_t *i40e = rxd->rxd_i40e; in i40e_alloc_rx_dma() local
824 if (i40e_alloc_dma_buffer(i40e, &rxd->rxd_desc_area, in i40e_alloc_rx_dma()
825 &i40e->i40e_static_dma_attr, &i40e->i40e_desc_acc_attr, B_FALSE, in i40e_alloc_rx_dma()
827 i40e_error(i40e, "failed to allocate DMA resources " in i40e_alloc_rx_dma()
838 dmasz = i40e->i40e_rx_buf_size; in i40e_alloc_rx_dma()
851 if (i40e_alloc_dma_buffer(i40e, dmap, in i40e_alloc_rx_dma()
852 &i40e->i40e_static_dma_attr, &i40e->i40e_buf_acc_attr, in i40e_alloc_rx_dma()
854 i40e_error(i40e, "failed to allocate RX dma buffer"); in i40e_alloc_rx_dma()
933 i40e_t *i40e = itrq->itrq_i40e; in i40e_alloc_tx_dma() local
935 itrq->itrq_tx_ring_size = i40e->i40e_tx_ring_size; in i40e_alloc_tx_dma()
936 itrq->itrq_tx_free_list_size = i40e->i40e_tx_ring_size + in i40e_alloc_tx_dma()
937 (i40e->i40e_tx_ring_size >> 1); in i40e_alloc_tx_dma()
946 if (i40e_alloc_dma_buffer(i40e, &itrq->itrq_desc_area, in i40e_alloc_tx_dma()
947 &i40e->i40e_static_dma_attr, &i40e->i40e_desc_acc_attr, in i40e_alloc_tx_dma()
949 i40e_error(i40e, "failed to allocate DMA resources for TX " in i40e_alloc_tx_dma()
964 i40e_error(i40e, "failed to allocate a %d entry TX work list " in i40e_alloc_tx_dma()
972 i40e_error(i40e, "failed to allocate a %d entry TX free list " in i40e_alloc_tx_dma()
984 i40e_error(i40e, "failed to allocate a %d entry tcb area for " in i40e_alloc_tx_dma()
992 dmasz = i40e->i40e_tx_buf_size; in i40e_alloc_tx_dma()
1003 ret = ddi_dma_alloc_handle(i40e->i40e_dip, in i40e_alloc_tx_dma()
1004 &i40e->i40e_txbind_dma_attr, DDI_DMA_DONTWAIT, NULL, in i40e_alloc_tx_dma()
1007 i40e_error(i40e, "failed to allocate DMA handle for TX " in i40e_alloc_tx_dma()
1014 ret = ddi_dma_alloc_handle(i40e->i40e_dip, in i40e_alloc_tx_dma()
1015 &i40e->i40e_txbind_lso_dma_attr, DDI_DMA_DONTWAIT, NULL, in i40e_alloc_tx_dma()
1018 i40e_error(i40e, "failed to allocate DMA handle for TX " in i40e_alloc_tx_dma()
1025 if (i40e_alloc_dma_buffer(i40e, &tcb->tcb_dma, in i40e_alloc_tx_dma()
1026 &i40e->i40e_static_dma_attr, &i40e->i40e_buf_acc_attr, in i40e_alloc_tx_dma()
1028 i40e_error(i40e, "failed to allocate %ld bytes of " in i40e_alloc_tx_dma()
1053 i40e_t *i40e = itrq->itrq_i40e; in i40e_free_ring_mem() local
1074 mutex_enter(&i40e->i40e_rx_pending_lock); in i40e_free_ring_mem()
1080 mutex_exit(&i40e->i40e_rx_pending_lock); in i40e_free_ring_mem()
1116 i40e_init_dma_attrs(i40e_t *i40e, boolean_t fma) in i40e_init_dma_attrs() argument
1118 bcopy(&i40e_g_static_dma_attr, &i40e->i40e_static_dma_attr, in i40e_init_dma_attrs()
1120 bcopy(&i40e_g_txbind_dma_attr, &i40e->i40e_txbind_dma_attr, in i40e_init_dma_attrs()
1122 bcopy(&i40e_g_txbind_lso_dma_attr, &i40e->i40e_txbind_lso_dma_attr, in i40e_init_dma_attrs()
1124 bcopy(&i40e_g_desc_acc_attr, &i40e->i40e_desc_acc_attr, in i40e_init_dma_attrs()
1126 bcopy(&i40e_g_buf_acc_attr, &i40e->i40e_buf_acc_attr, in i40e_init_dma_attrs()
1130 i40e->i40e_static_dma_attr.dma_attr_flags |= DDI_DMA_FLAGERR; in i40e_init_dma_attrs()
1131 i40e->i40e_txbind_dma_attr.dma_attr_flags |= DDI_DMA_FLAGERR; in i40e_init_dma_attrs()
1132 i40e->i40e_txbind_lso_dma_attr.dma_attr_flags |= in i40e_init_dma_attrs()
1135 i40e->i40e_static_dma_attr.dma_attr_flags &= ~DDI_DMA_FLAGERR; in i40e_init_dma_attrs()
1136 i40e->i40e_txbind_dma_attr.dma_attr_flags &= ~DDI_DMA_FLAGERR; in i40e_init_dma_attrs()
1137 i40e->i40e_txbind_lso_dma_attr.dma_attr_flags &= in i40e_init_dma_attrs()
1183 i40e_t *i40e; in i40e_rx_recycle() local
1188 i40e = rxd->rxd_i40e; in i40e_rx_recycle()
1218 mutex_enter(&i40e->i40e_rx_pending_lock); in i40e_rx_recycle()
1220 atomic_dec_32(&i40e->i40e_rx_pending); in i40e_rx_recycle()
1228 cv_broadcast(&i40e->i40e_rx_pending_cv); in i40e_rx_recycle()
1231 mutex_exit(&i40e->i40e_rx_pending_lock); in i40e_rx_recycle()
1240 i40e_t *i40e = rxd->rxd_i40e; in i40e_rx_bind() local
1270 ddi_fm_service_impact(i40e->i40e_dip, DDI_SERVICE_DEGRADED); in i40e_rx_bind()
1271 atomic_or_32(&i40e->i40e_state, I40E_ERROR); in i40e_rx_bind()
1297 i40e_t *i40e = rxd->rxd_i40e; in i40e_rx_copy() local
1307 ddi_fm_service_impact(i40e->i40e_dip, DDI_SERVICE_DEGRADED); in i40e_rx_copy()
1308 atomic_or_32(&i40e->i40e_state, I40E_ERROR); in i40e_rx_copy()
1438 i40e_t *i40e; in i40e_ring_rx() local
1450 i40e = itrq->itrq_i40e; in i40e_ring_rx()
1451 hw = &i40e->i40e_hw_space; in i40e_ring_rx()
1453 if (!(i40e->i40e_state & I40E_STARTED) || in i40e_ring_rx()
1454 (i40e->i40e_state & I40E_OVERTEMP) || in i40e_ring_rx()
1455 (i40e->i40e_state & I40E_SUSPENDED) || in i40e_ring_rx()
1456 (i40e->i40e_state & I40E_ERROR)) in i40e_ring_rx()
1467 ddi_fm_service_impact(i40e->i40e_dip, DDI_SERVICE_DEGRADED); in i40e_ring_rx()
1468 atomic_or_32(&i40e->i40e_state, I40E_ERROR); in i40e_ring_rx()
1548 if (plen >= i40e->i40e_rx_dma_min) in i40e_ring_rx()
1554 if (i40e->i40e_rx_hcksum_enable) in i40e_ring_rx()
1589 if (rx_frames > i40e->i40e_rx_limit_per_intr) { in i40e_ring_rx()
1604 ddi_fm_service_impact(i40e->i40e_dip, DDI_SERVICE_DEGRADED); in i40e_ring_rx()
1605 atomic_or_32(&i40e->i40e_state, I40E_ERROR); in i40e_ring_rx()
1610 ddi_acc_handle_t rh = i40e->i40e_osdep_space.ios_reg_handle; in i40e_ring_rx()
1616 ddi_fm_service_impact(i40e->i40e_dip, in i40e_ring_rx()
1618 atomic_or_32(&i40e->i40e_state, I40E_ERROR); in i40e_ring_rx()
1669 i40e_tx_context(i40e_t *i40e, i40e_trqpair_t *itrq, mblk_t *mp, in i40e_tx_context() argument
1677 if (i40e->i40e_tx_hcksum_enable != B_TRUE) in i40e_tx_context()
1925 i40e_t *i40e = itrq->itrq_i40e; in i40e_tx_recycle_ring() local
1934 mac_tx_ring_update(i40e->i40e_mac_hdl, in i40e_tx_recycle_ring()
1954 ddi_fm_service_impact(i40e->i40e_dip, DDI_SERVICE_DEGRADED); in i40e_tx_recycle_ring()
1955 atomic_or_32(&i40e->i40e_state, I40E_ERROR); in i40e_tx_recycle_ring()
2003 itrq->itrq_desc_free > i40e->i40e_tx_block_thresh) { in i40e_tx_recycle_ring()
2006 mac_tx_ring_update(i40e->i40e_mac_hdl, itrq->itrq_mactxring); in i40e_tx_recycle_ring()
2182 i40e_t *i40e = itrq->itrq_i40e; in i40e_non_lso_chain() local
2186 ASSERT3U(msgsize(mp), <, i40e->i40e_tx_buf_size); in i40e_non_lso_chain()
2197 if (nmp_len < i40e->i40e_tx_dma_min || force_copy) { in i40e_non_lso_chain()
2222 i40e_error(i40e, "dma bind failed!"); in i40e_non_lso_chain()
2389 i40e_t *i40e = itrq->itrq_i40e; in i40e_lso_chain() local
2435 if (mp_len < i40e->i40e_tx_dma_min) { in i40e_lso_chain()
2457 if (mp_len < i40e->i40e_tx_dma_min || force_copy) { in i40e_lso_chain()
2537 i40e_error(i40e, "dma bind failed!"); in i40e_lso_chain()
2725 i40e_t *i40e = itrq->itrq_i40e; in i40e_ring_tx() local
2726 i40e_hw_t *hw = &i40e->i40e_hw_space; in i40e_ring_tx()
2731 if (!(i40e->i40e_state & I40E_STARTED) || in i40e_ring_tx()
2732 (i40e->i40e_state & I40E_OVERTEMP) || in i40e_ring_tx()
2733 (i40e->i40e_state & I40E_SUSPENDED) || in i40e_ring_tx()
2734 (i40e->i40e_state & I40E_ERROR) || in i40e_ring_tx()
2735 (i40e->i40e_link_state != LINK_STATE_UP) || in i40e_ring_tx()
2752 if (i40e_tx_context(i40e, itrq, mp, &meo, &tctx) < 0) { in i40e_ring_tx()
2807 if (itrq->itrq_desc_free < i40e->i40e_tx_block_thresh || in i40e_ring_tx()
2882 if (i40e_check_acc_handle(i40e->i40e_osdep_space.ios_reg_handle) != in i40e_ring_tx()
2889 ddi_fm_service_impact(i40e->i40e_dip, DDI_SERVICE_DEGRADED); in i40e_ring_tx()
2890 atomic_or_32(&i40e->i40e_state, I40E_ERROR); in i40e_ring_tx()