Lines Matching full:stmp

153 	KASSERT((*pstmp)->flags == 0, ("stmp flags are not 0"));  in sfxge_next_stmp()
155 &txq->stmp[txq->ptr_mask])) in sfxge_next_stmp()
156 *pstmp = &txq->stmp[0]; in sfxge_next_stmp()
226 struct sfxge_tx_mapping *stmp; in sfxge_tx_qcomplete() local
231 stmp = &txq->stmp[id]; in sfxge_tx_qcomplete()
232 if (stmp->flags & TX_BUF_UNMAP) { in sfxge_tx_qcomplete()
233 bus_dmamap_unload(txq->packet_dma_tag, stmp->map); in sfxge_tx_qcomplete()
234 if (stmp->flags & TX_BUF_MBUF) { in sfxge_tx_qcomplete()
235 struct mbuf *m = stmp->u.mbuf; in sfxge_tx_qcomplete()
240 free(stmp->u.heap_buf, M_SFXGE); in sfxge_tx_qcomplete()
242 stmp->flags = 0; in sfxge_tx_qcomplete()
401 struct sfxge_tx_mapping *stmp; in sfxge_tx_queue_mbuf() local
437 stmp = &txq->stmp[id]; in sfxge_tx_queue_mbuf()
438 rc = bus_dmamap_load_mbuf_sg(txq->packet_dma_tag, stmp->map, in sfxge_tx_queue_mbuf()
449 stmp->map, mbuf, in sfxge_tx_queue_mbuf()
456 bus_dmamap_sync(txq->packet_dma_tag, stmp->map, BUS_DMASYNC_PREWRITE); in sfxge_tx_queue_mbuf()
458 used_map = &stmp->map; in sfxge_tx_queue_mbuf()
468 n_extra_descs = sfxge_tx_maybe_toggle_cksum_offload(txq, mbuf, &stmp); in sfxge_tx_queue_mbuf()
469 n_extra_descs += sfxge_tx_maybe_insert_tag(txq, mbuf, &stmp); in sfxge_tx_queue_mbuf()
476 stmp = &txq->stmp[(rc - 1) & txq->ptr_mask]; in sfxge_tx_queue_mbuf()
494 sfxge_next_stmp(txq, &stmp); in sfxge_tx_queue_mbuf()
504 if (used_map != &stmp->map) { in sfxge_tx_queue_mbuf()
505 map = stmp->map; in sfxge_tx_queue_mbuf()
506 stmp->map = *used_map; in sfxge_tx_queue_mbuf()
510 stmp->u.mbuf = mbuf; in sfxge_tx_queue_mbuf()
511 stmp->flags = TX_BUF_UNMAP | TX_BUF_MBUF; in sfxge_tx_queue_mbuf()
1240 KASSERT(txq->stmp[id].flags == 0, ("stmp flags are not 0")); in tso_start_new_packet()
1258 KASSERT(txq->stmp[id].flags == 0, ("stmp flags are not 0")); in tso_start_new_packet()
1268 KASSERT(txq->stmp[id].flags == 0, ("stmp flags are not 0")); in tso_start_new_packet()
1282 KASSERT(txq->stmp[id].flags == 0, in tso_start_new_packet()
1283 ("stmp flags are not 0")); in tso_start_new_packet()
1285 struct sfxge_tx_mapping *stmp = &txq->stmp[id]; in tso_start_new_packet() local
1291 rc = bus_dmamap_load(txq->packet_dma_tag, stmp->map, in tso_start_new_packet()
1299 stmp->map); in tso_start_new_packet()
1305 map = stmp->map; in tso_start_new_packet()
1308 stmp->u.heap_buf = header; in tso_start_new_packet()
1309 stmp->flags = TX_BUF_UNMAP; in tso_start_new_packet()
1388 KASSERT(txq->stmp[id].flags == 0, ("stmp flags are not 0")); in sfxge_tx_queue_tso()
1790 bus_dmamap_destroy(txq->packet_dma_tag, txq->stmp[nmaps].map); in sfxge_tx_qfini()
1791 free(txq->stmp, M_SFXGE); in sfxge_tx_qfini()
1851 txq->stmp = malloc(sizeof(struct sfxge_tx_mapping) * sc->txq_entries, in sfxge_tx_qinit()
1855 &txq->stmp[nmaps].map); in sfxge_tx_qinit()
1919 bus_dmamap_destroy(txq->packet_dma_tag, txq->stmp[nmaps].map); in sfxge_tx_qinit()
1920 free(txq->stmp, M_SFXGE); in sfxge_tx_qinit()