Lines Matching refs:netmem_ref
381 static void page_pool_return_netmem(struct page_pool *pool, netmem_ref netmem);
383 static noinline netmem_ref page_pool_refill_alloc_cache(struct page_pool *pool) in page_pool_refill_alloc_cache()
386 netmem_ref netmem; in page_pool_refill_alloc_cache()
407 netmem = (__force netmem_ref)__ptr_ring_consume(r); in page_pool_refill_alloc_cache()
436 static netmem_ref __page_pool_get_cached(struct page_pool *pool) in __page_pool_get_cached()
438 netmem_ref netmem; in __page_pool_get_cached()
453 netmem_ref netmem, in __page_pool_dma_sync_for_device()
467 netmem_ref netmem, in page_pool_dma_sync_for_device()
481 netmem_ref netmem, gfp_t gfp) in page_pool_register_dma_index()
506 netmem_ref netmem) in page_pool_release_dma_index()
530 static bool page_pool_dma_map(struct page_pool *pool, netmem_ref netmem, gfp_t gfp) in page_pool_dma_map()
595 static noinline netmem_ref __page_pool_alloc_netmems_slow(struct page_pool *pool, in __page_pool_alloc_netmems_slow()
601 netmem_ref netmem; in __page_pool_alloc_netmems_slow()
659 netmem_ref page_pool_alloc_netmems(struct page_pool *pool, gfp_t gfp) in page_pool_alloc_netmems()
661 netmem_ref netmem; in page_pool_alloc_netmems()
708 void page_pool_set_pp_info(struct page_pool *pool, netmem_ref netmem) in page_pool_set_pp_info()
734 void page_pool_clear_pp_info(netmem_ref netmem) in page_pool_clear_pp_info()
751 netmem_ref netmem) in __page_pool_release_netmem_dma()
778 static void page_pool_return_netmem(struct page_pool *pool, netmem_ref netmem) in page_pool_return_netmem()
805 static bool page_pool_recycle_in_ring(struct page_pool *pool, netmem_ref netmem) in page_pool_recycle_in_ring()
824 static bool page_pool_recycle_in_cache(netmem_ref netmem, in page_pool_recycle_in_cache()
838 static bool __page_pool_page_can_be_recycled(netmem_ref netmem) in __page_pool_page_can_be_recycled()
851 static __always_inline netmem_ref
852 __page_pool_put_page(struct page_pool *pool, netmem_ref netmem, in __page_pool_put_page()
924 void page_pool_put_unrefed_netmem(struct page_pool *pool, netmem_ref netmem, in page_pool_put_unrefed_netmem()
949 netmem_ref *bulk, in page_pool_recycle_ring_bulk()
995 void page_pool_put_netmem_bulk(netmem_ref *data, u32 count) in page_pool_put_netmem_bulk()
1000 netmem_ref netmem = netmem_compound_head(data[i]); in page_pool_put_netmem_bulk()
1008 netmem_ref bulk[XDP_BULK_QUEUE_SIZE]; in page_pool_put_netmem_bulk()
1017 netmem_ref netmem = data[i]; in page_pool_put_netmem_bulk()
1047 static netmem_ref page_pool_drain_frag(struct page_pool *pool, in page_pool_drain_frag()
1048 netmem_ref netmem) in page_pool_drain_frag()
1068 netmem_ref netmem = pool->frag_page; in page_pool_free_frag()
1078 netmem_ref page_pool_alloc_frag_netmem(struct page_pool *pool, in page_pool_alloc_frag_netmem()
1083 netmem_ref netmem = pool->frag_page; in page_pool_alloc_frag_netmem()
1133 netmem_ref netmem; in page_pool_empty_ring()
1136 while ((netmem = (__force netmem_ref)ptr_ring_consume_bh(&pool->ring))) { in page_pool_empty_ring()
1159 netmem_ref netmem; in page_pool_empty_alloc_cache_once()
1335 netmem_ref netmem; in page_pool_update_nid()
1358 netmem_ref netmem = net_iov_to_netmem(niov); in net_mp_niov_set_page_pool()
1371 netmem_ref netmem = net_iov_to_netmem(niov); in net_mp_niov_clear_page_pool()