Home
last modified time | relevance | path

Searched refs:ndesc (Results 1 – 25 of 28) sorted by relevance

12

/linux/drivers/net/ethernet/fungible/funeth/
H A Dfuneth_tx.c349 unsigned int ndesc; in fun_start_xmit() local
357 ndesc = write_pkt_desc(skb, q, tls_len); in fun_start_xmit()
358 if (unlikely(!ndesc)) { in fun_start_xmit()
363 q->prod_cnt += ndesc; in fun_start_xmit()
425 unsigned int npkts = 0, nbytes = 0, ndesc = 0; in fun_txq_reclaim() local
448 ndesc += pkt_desc; in fun_txq_reclaim()
454 q->cons_cnt += ndesc; in fun_txq_reclaim()
486 unsigned int npkts = 0, ndesc = 0, head, reclaim_idx; in fun_xdpq_clean() local
505 ndesc += pkt_desc; in fun_xdpq_clean()
510 q->cons_cnt += ndesc; in fun_xdpq_clean()
[all …]
H A Dfuneth_txrx.h255 unsigned int ndesc, struct fun_irq *irq, int state,
/linux/drivers/net/ethernet/altera/
H A Daltera_sgdma.c13 struct sgdma_descrip __iomem *ndesc,
173 struct sgdma_descrip __iomem *ndesc = &descbase[1]; in sgdma_tx_buffer() local
180 ndesc, /* next descriptor */ in sgdma_tx_buffer()
181 sgdma_txphysaddr(priv, ndesc), in sgdma_tx_buffer()
294 struct sgdma_descrip __iomem *ndesc, in sgdma_setup_descrip() argument
305 u32 ctrl = csrrd8(ndesc, sgdma_descroffs(control)); in sgdma_setup_descrip()
307 csrwr8(ctrl, ndesc, sgdma_descroffs(control)); in sgdma_setup_descrip()
342 struct sgdma_descrip __iomem *ndesc = &descbase[1]; in sgdma_async_read() local
353 ndesc, /* next descriptor */ in sgdma_async_read()
354 sgdma_rxphysaddr(priv, ndesc), in sgdma_async_read()
/linux/drivers/net/ethernet/socionext/
H A Dsni_ave.c233 u32 ndesc; /* number of descriptor */ member
681 u32 proc_idx, done_idx, ndesc, cmdsts; in ave_tx_complete() local
688 ndesc = priv->tx.ndesc; in ave_tx_complete()
721 done_idx = (done_idx + 1) % ndesc; in ave_tx_complete()
748 u32 ndesc, cmdsts; in ave_rx_receive() local
752 ndesc = priv->rx.ndesc; in ave_rx_receive()
753 restpkt = ((proc_idx + ndesc - 1) - done_idx) % ndesc; in ave_rx_receive()
768 proc_idx = (proc_idx + 1) % ndesc; in ave_rx_receive()
792 proc_idx = (proc_idx + 1) % ndesc; in ave_rx_receive()
807 done_idx = (done_idx + 1) % ndesc; in ave_rx_receive()
[all …]
/linux/drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/
H A Dchcr_ktls.c844 u32 len, cpl = 0, ndesc, wr_len, wr_mid = 0; in chcr_ktls_xmit_tcb_cpls() local
852 ndesc = DIV_ROUND_UP(len, 64); in chcr_ktls_xmit_tcb_cpls()
854 credits = chcr_txq_avail(&q->q) - ndesc; in chcr_ktls_xmit_tcb_cpls()
922 ndesc = DIV_ROUND_UP(len, 64); in chcr_ktls_xmit_tcb_cpls()
923 chcr_txq_advance(&q->q, ndesc); in chcr_ktls_xmit_tcb_cpls()
924 cxgb4_ring_tx_db(tx_info->adap, &q->q, ndesc); in chcr_ktls_xmit_tcb_cpls()
995 unsigned int ndesc; in chcr_ktls_write_tcp_options() local
1016 ndesc = DIV_ROUND_UP(len16, 4); in chcr_ktls_write_tcp_options()
1018 credits = chcr_txq_avail(&q->q) - ndesc; in chcr_ktls_write_tcp_options()
1079 chcr_txq_advance(&q->q, ndesc); in chcr_ktls_write_tcp_options()
[all …]
/linux/drivers/net/ethernet/chelsio/inline_crypto/ch_ipsec/
H A Dchcr_ipsec.c571 u16 ndesc; in ch_ipsec_crypto_wreq() local
580 ndesc = DIV_ROUND_UP(flits, 2); in ch_ipsec_crypto_wreq()
596 wr_mid = FW_CRYPTO_LOOKASIDE_WR_LEN16_V(ndesc); in ch_ipsec_crypto_wreq()
609 wr->req.ulptx.len = htonl(ndesc - 1); in ch_ipsec_crypto_wreq()
706 unsigned int last_desc, ndesc, flits = 0; in ch_ipsec_xmit() local
736 ndesc = flits_to_desc(flits); in ch_ipsec_xmit()
737 credits = txq_avail(&q->q) - ndesc; in ch_ipsec_xmit()
743 dev->name, qidx, credits, ndesc, txq_avail(&q->q), in ch_ipsec_xmit()
748 last_desc = q->q.pidx + ndesc - 1; in ch_ipsec_xmit()
786 txq_advance(&q->q, ndesc); in ch_ipsec_xmit()
[all …]
/linux/drivers/usb/mon/
H A Dmon_bin.c115 unsigned int ndesc; /* Actual number of ISO descriptors */ member
453 struct urb *urb, unsigned int ndesc) in mon_bin_collate_isodesc() argument
460 while (ndesc-- != 0) { in mon_bin_collate_isodesc()
471 unsigned int offset, struct urb *urb, char ev_type, unsigned int ndesc) in mon_bin_get_isodesc() argument
477 while (ndesc-- != 0) { in mon_bin_get_isodesc()
500 unsigned int ndesc, lendesc; in mon_bin_event() local
518 ndesc = 0; in mon_bin_event()
520 ndesc = ISODESC_MAX; in mon_bin_event()
522 ndesc = urb->number_of_packets; in mon_bin_event()
525 length = mon_bin_collate_isodesc(rp, urb, ndesc); in mon_bin_event()
[all …]
H A Dmon_text.c200 int i, ndesc; in mon_text_event() local
233 if ((ndesc = urb->number_of_packets) > ISODESC_MAX) in mon_text_event()
234 ndesc = ISODESC_MAX; in mon_text_event()
237 for (i = 0; i < ndesc; i++) { in mon_text_event()
591 int ndesc; /* Display this many */ in mon_text_read_isodesc() local
597 ndesc = ep->numdesc; in mon_text_read_isodesc()
598 if (ndesc > ISODESC_MAX) in mon_text_read_isodesc()
599 ndesc = ISODESC_MAX; in mon_text_read_isodesc()
600 if (ndesc < 0) in mon_text_read_isodesc()
601 ndesc = 0; in mon_text_read_isodesc()
[all …]
/linux/drivers/net/ethernet/chelsio/cxgb4/
H A Dsge.c1491 unsigned int last_desc, flits, ndesc; in cxgb4_eth_xmit() local
1548 ndesc = flits_to_desc(flits); in cxgb4_eth_xmit()
1549 credits = txq_avail(&q->q) - ndesc; in cxgb4_eth_xmit()
1565 last_desc = q->q.pidx + ndesc - 1; in cxgb4_eth_xmit()
1720 txq_advance(&q->q, ndesc); in cxgb4_eth_xmit()
1722 cxgb4_ring_tx_db(adap, &q->q, ndesc); in cxgb4_eth_xmit()
1813 unsigned int last_desc, flits, ndesc; in cxgb4_vf_eth_xmit() local
1857 ndesc = flits_to_desc(flits); in cxgb4_vf_eth_xmit()
1858 credits = txq_avail(&txq->q) - ndesc; in cxgb4_vf_eth_xmit()
1873 last_desc = txq->q.pidx + ndesc - 1; in cxgb4_vf_eth_xmit()
[all …]
H A Dcxgb4_tc_mqprio.c108 eosw_txq->ndesc = CXGB4_EOSW_TXQ_DEFAULT_DESC_NUM; in cxgb4_init_eosw_txq()
125 cxgb4_eosw_txq_free_desc(adap, eosw_txq, eosw_txq->ndesc); in cxgb4_clean_eosw_txq()
/linux/drivers/net/ethernet/chelsio/cxgb3/
H A Dsge.c1087 static void write_wr_hdr_sgl(unsigned int ndesc, struct sk_buff *skb, in write_wr_hdr_sgl() argument
1105 if (likely(ndesc == 1)) { in write_wr_hdr_sgl()
1128 ndesc--; in write_wr_hdr_sgl()
1158 WARN_ON(ndesc != 0); in write_wr_hdr_sgl()
1179 struct sge_txq *q, unsigned int ndesc, in write_tx_pkt_wr() argument
1237 sgp = ndesc == 1 ? (struct sg_ent *)&d->flit[flits] : sgl; in write_tx_pkt_wr()
1240 write_wr_hdr_sgl(ndesc, skb, d, pidx, q, sgl, flits, sgl_flits, gen, in write_tx_pkt_wr()
1263 unsigned int ndesc, pidx, credits, gen, compl; in t3_eth_xmit() local
1288 ndesc = calc_tx_descs(skb); in t3_eth_xmit()
1290 if (unlikely(credits < ndesc)) { in t3_eth_xmit()
[all …]
/linux/drivers/dma/
H A Didma64.c195 static struct idma64_desc *idma64_alloc_desc(unsigned int ndesc) in idma64_alloc_desc() argument
203 desc->hw = kzalloc_objs(*desc->hw, ndesc, GFP_NOWAIT); in idma64_alloc_desc()
217 if (desc->ndesc) { in idma64_desc_free()
218 unsigned int i = desc->ndesc; in idma64_desc_free()
280 unsigned int i = desc->ndesc; in idma64_desc_fill()
320 desc->ndesc = i; in idma64_prep_slave_sg()
329 desc->ndesc = sg_len; in idma64_prep_slave_sg()
362 } while (++i < desc->ndesc); in idma64_active_desc_size()
H A Didma64.h117 unsigned int ndesc; member
/linux/drivers/vhost/
H A Dnet.c594 unsigned int *ndesc) in vhost_net_tx_get_vq_desc() argument
601 out_num, in_num, NULL, NULL, ndesc); in vhost_net_tx_get_vq_desc()
613 out_num, in_num, NULL, NULL, ndesc); in vhost_net_tx_get_vq_desc()
645 unsigned int *ndesc) in get_tx_bufs() argument
651 busyloop_intr, ndesc); in get_tx_bufs()
770 unsigned int ndesc = 0; in handle_tx_copy() local
779 &busyloop_intr, &ndesc); in handle_tx_copy()
811 vhost_discard_vq_desc(vq, 1, ndesc); in handle_tx_copy()
834 vhost_discard_vq_desc(vq, 1, ndesc); in handle_tx_copy()
873 unsigned int ndesc = 0; in handle_tx_zerocopy() local
[all …]
H A Dvhost.h239 unsigned int *ndesc);
242 unsigned int ndesc);
H A Dvhost.c2819 unsigned int *ndesc) in vhost_get_vq_desc_n() argument
2937 if (ndesc) in vhost_get_vq_desc_n()
2938 *ndesc = c; in vhost_get_vq_desc_n()
2976 unsigned int ndesc) in vhost_discard_vq_desc() argument
2978 vq->next_avail_head -= ndesc; in vhost_discard_vq_desc()
/linux/drivers/net/ethernet/airoha/
H A Dairoha_eth.c546 while (q->queued < q->ndesc - 1) { in airoha_qdma_fill_rx_queue()
558 q->head = (q->head + 1) % q->ndesc; in airoha_qdma_fill_rx_queue()
624 q->tail = (q->tail + 1) % q->ndesc; in airoha_qdma_rx_process()
738 struct airoha_qdma *qdma, int ndesc) in airoha_qdma_init_rx_queue() argument
755 q->ndesc = ndesc; in airoha_qdma_init_rx_queue()
758 q->entry = devm_kzalloc(eth->dev, q->ndesc * sizeof(*q->entry), in airoha_qdma_init_rx_queue()
771 q->desc = dmam_alloc_coherent(eth->dev, q->ndesc * sizeof(*q->desc), in airoha_qdma_init_rx_queue()
781 FIELD_PREP(RX_RING_SIZE_MASK, ndesc)); in airoha_qdma_init_rx_queue()
783 thr = clamp(ndesc >> 3, 1, 32); in airoha_qdma_init_rx_queue()
806 q->tail = (q->tail + 1) % q->ndesc; in airoha_qdma_cleanup_rx_queue()
[all …]
/linux/drivers/net/wireless/ath/ath9k/
H A Dinit.c284 int nbuf, int ndesc, bool is_tx) in ath_descdma_setup() argument
291 name, nbuf, ndesc); in ath_descdma_setup()
307 dd->dd_desc_len = desc_len * nbuf * ndesc; in ath_descdma_setup()
347 for (i = 0; i < nbuf; i++, bf++, ds += (desc_len * ndesc)) { in ath_descdma_setup()
363 ds += (desc_len * ndesc); in ath_descdma_setup()
378 for (i = 0; i < nbuf; i++, bf++, ds += (desc_len * ndesc)) { in ath_descdma_setup()
394 ds += (desc_len * ndesc); in ath_descdma_setup()
H A Dath9k.h79 int nbuf, int ndesc, bool is_tx);
/linux/drivers/net/ethernet/chelsio/cxgb4vf/
H A Dsge.c1162 unsigned int flits, ndesc; in t4vf_eth_xmit() local
1213 ndesc = flits_to_desc(flits); in t4vf_eth_xmit()
1214 credits = txq_avail(&txq->q) - ndesc; in t4vf_eth_xmit()
1364 ndesc, credits, txq->q.pidx, skb->len, ssi->nr_frags); in t4vf_eth_xmit()
1435 last_desc = tq->pidx + ndesc - 1; in t4vf_eth_xmit()
1446 txq_advance(&txq->q, ndesc); in t4vf_eth_xmit()
1448 ring_tx_db(adapter, &txq->q, ndesc); in t4vf_eth_xmit()
/linux/drivers/infiniband/ulp/srp/
H A Dib_srp.h347 unsigned int ndesc; member
H A Dib_srp.c1407 state->ndesc++; in srp_map_desc()
1599 for (i = 0; i < state->ndesc; i++) in srp_check_mapping()
1608 state->ndesc, state->nmdesc); in srp_check_mapping()
1737 if (state.ndesc == 1) { in srp_map_data()
1749 if (unlikely(target->cmd_sg_cnt < state.ndesc && in srp_map_data()
1757 count = min(state.ndesc, target->cmd_sg_cnt); in srp_map_data()
1758 table_len = state.ndesc * sizeof (struct srp_direct_buf); in srp_map_data()
/linux/drivers/net/ethernet/atheros/
H A Dag71xx.c1452 int i, ring_mask, ndesc, split; in ag71xx_fill_dma_desc() local
1456 ndesc = 0; in ag71xx_fill_dma_desc()
1465 i = (ring->curr + ndesc) & ring_mask; in ag71xx_fill_dma_desc()
1489 if (!ndesc) in ag71xx_fill_dma_desc()
1493 ndesc++; in ag71xx_fill_dma_desc()
1496 return ndesc; in ag71xx_fill_dma_desc()
/linux/drivers/crypto/inside-secure/
H A Dsafexcel.c1016 int ret, i, nreq, ndesc, tot_descs, handled = 0; in safexcel_handle_result_descriptor() local
1032 ndesc = ctx->handle_result(priv, ring, req, in safexcel_handle_result_descriptor()
1034 if (ndesc < 0) { in safexcel_handle_result_descriptor()
1036 ndesc); in safexcel_handle_result_descriptor()
1046 tot_descs += ndesc; in safexcel_handle_result_descriptor()
/linux/drivers/net/ethernet/rdc/
H A Dr6040.c166 __le32 ndesc; /* 8-B */ member
302 desc->ndesc = cpu_to_le32(mapping); in r6040_init_ring_desc()
307 desc->ndesc = cpu_to_le32(desc_dma); in r6040_init_ring_desc()

12