Searched refs:RGE_SEND_SLOTS (Results 1 – 5 of 5) sorted by relevance
392 next = NEXT(slot, RGE_SEND_SLOTS); in rge_send_claim()404 ASSERT(slot < RGE_SEND_SLOTS); in rge_send_claim()405 ASSERT(next < RGE_SEND_SLOTS); in rge_send_claim()434 tc_tail = LAST(tc_tail, RGE_SEND_SLOTS); in rge_send_recycle()456 rgep->tc_next = NEXT(tc_tail, RGE_SEND_SLOTS); in rge_send_recycle()459 n += RGE_SEND_SLOTS; in rge_send_recycle()462 ASSERT(rgep->tx_free <= RGE_SEND_SLOTS); in rge_send_recycle()643 ASSERT(rgep->tx_free < RGE_SEND_SLOTS); in rge_send()670 if (rgep->chipid.is_pcie && rgep->tx_free != RGE_SEND_SLOTS) { in rge_reschedule()
258 txdescsize = RGE_SEND_SLOTS * sizeof (rge_bd_t); in rge_alloc_bufs()318 for (slot = 0; slot < RGE_SEND_SLOTS; slot++) { in rge_reinit_send_ring()324 if (slot == (RGE_SEND_SLOTS - 1)) in rge_reinit_send_ring()334 rgep->tx_free = RGE_SEND_SLOTS; in rge_reinit_send_ring()400 for (slot = 0; slot < RGE_SEND_SLOTS; ++slot) { in rge_fini_send_ring()405 kmem_free(rgep->sw_sbds, RGE_SEND_SLOTS * sizeof (sw_sbd_t)); in rge_fini_send_ring()480 ssbdp = kmem_zalloc(RGE_SEND_SLOTS*sizeof (*ssbdp), KM_SLEEP); in rge_init_send_ring()491 for (slot = 0; slot < RGE_SEND_SLOTS; slot++) { in rge_init_send_ring()
550 #define RGE_SEND_SLOTS 1024 macro
330 if ((slot) == (RGE_SEND_SLOTS -1)) { \
1521 ((int_status & TX_OK_INT) && rgep->tx_free < RGE_SEND_SLOTS/8)) { in rge_intr()1729 if (rgep->tx_free < RGE_SEND_SLOTS) in rge_chip_cyclic()