Home
last modified time | relevance | path

Searched refs:pprm (Results 1 – 3 of 3) sorted by relevance

/linux/drivers/infiniband/hw/irdma/
H A Dpble.h118 int irdma_prm_add_pble_mem(struct irdma_pble_prm *pprm,
120 int irdma_prm_get_pbles(struct irdma_pble_prm *pprm,
123 void irdma_prm_return_pbles(struct irdma_pble_prm *pprm,
H A Dutils.c2099 int irdma_prm_add_pble_mem(struct irdma_pble_prm *pprm, in irdma_prm_add_pble_mem() argument
2107 sizeofbitmap = (u64)pchunk->size >> pprm->pble_shift; in irdma_prm_add_pble_mem()
2115 pprm->total_pble_alloc += pchunk->size >> 3; in irdma_prm_add_pble_mem()
2116 pprm->free_pble_cnt += pchunk->size >> 3; in irdma_prm_add_pble_mem()
2129 int irdma_prm_get_pbles(struct irdma_pble_prm *pprm, in irdma_prm_get_pbles() argument
2136 struct list_head *chunk_entry = pprm->clist.next; in irdma_prm_get_pbles()
2142 bits_needed = DIV_ROUND_UP_ULL(mem_size, BIT_ULL(pprm->pble_shift)); in irdma_prm_get_pbles()
2144 spin_lock_irqsave(&pprm->prm_lock, flags); in irdma_prm_get_pbles()
2145 while (chunk_entry != &pprm->clist) { in irdma_prm_get_pbles()
2158 spin_unlock_irqrestore(&pprm->prm_lock, flags); in irdma_prm_get_pbles()
[all …]
/linux/drivers/net/ethernet/microsoft/mana/
H A Dmana_en.c2605 struct page_pool_params pprm = {}; in mana_create_page_pool() local
2608 pprm.pool_size = mpc->rx_queue_size / rxq->frag_count + 1; in mana_create_page_pool()
2609 pprm.nid = gc->numa_node; in mana_create_page_pool()
2610 pprm.napi = &rxq->rx_cq.napi; in mana_create_page_pool()
2611 pprm.netdev = rxq->ndev; in mana_create_page_pool()
2612 pprm.order = get_order(rxq->alloc_size); in mana_create_page_pool()
2613 pprm.queue_idx = rxq->rxq_idx; in mana_create_page_pool()
2614 pprm.dev = gc->dev; in mana_create_page_pool()
2620 pprm.flags = PP_FLAG_DMA_MAP | PP_FLAG_DMA_SYNC_DEV; in mana_create_page_pool()
2621 pprm.max_len = PAGE_SIZE; in mana_create_page_pool()
[all …]