Searched refs:nb_entries (Results 1 – 3 of 3) sorted by relevance
/linux/net/xdp/ |
H A D | xsk_queue.h | 240 u32 cached_cons = q->cached_cons, nb_entries = 0; in xskq_cons_read_desc_batch() local 247 while (cached_cons != q->cached_prod && nb_entries < max) { in xskq_cons_read_desc_batch() 252 descs[nb_entries] = ring->desc[idx]; in xskq_cons_read_desc_batch() 254 parse_desc(q, pool, &descs[nb_entries], &parsed); in xskq_cons_read_desc_batch() 268 nb_entries++; in xskq_cons_read_desc_batch() 394 u32 nb_entries) in xskq_prod_write_addr_batch() argument 401 for (i = 0; i < nb_entries; i++) in xskq_prod_write_addr_batch() 434 static inline void xskq_prod_submit_n(struct xsk_queue *q, u32 nb_entries) in xskq_prod_submit_n() argument 436 __xskq_prod_submit(q, q->ring->producer + nb_entries); in xskq_prod_submit_n()
|
H A D | xsk_buff_pool.c | 580 u32 i, cached_cons, nb_entries; in xp_alloc_new_from_fq() 587 nb_entries = max; in xp_alloc_new_from_fq() 600 nb_entries--; in xp_alloc_new_from_fq() 611 return nb_entries; in xp_alloc_new_from_fq() 614 static u32 xp_alloc_reused(struct xsk_buff_pool *pool, struct xdp_buff **xdp, u32 nb_entries) 619 nb_entries = min_t(u32, nb_entries, pool->free_list_cnt); in xp_alloc_reused() 621 i = nb_entries; in xp_alloc_reused() 629 pool->free_list_cnt -= nb_entries; in xp_alloc_reused() 631 return nb_entries; in xp_alloc_reused() 581 u32 i, cached_cons, nb_entries; xp_alloc_new_from_fq() local 615 xp_alloc_reused(struct xsk_buff_pool * pool,struct xdp_buff ** xdp,u32 nb_entries) xp_alloc_reused() argument [all...] |
H A D | xsk.c | 398 void xsk_tx_completed(struct xsk_buff_pool *pool, u32 nb_entries) in xsk_tx_completed() argument 400 xskq_prod_submit_n(pool->cq, nb_entries); in xsk_tx_completed()
|