/linux/include/rdma/ |
H A D | ib_umem.h | 62 static inline size_t ib_umem_num_dma_blocks(struct ib_umem *umem, in ib_umem_num_dma_blocks() function 72 return ib_umem_num_dma_blocks(umem, PAGE_SIZE); in ib_umem_num_pages() 82 biter->__sg_numblocks = ib_umem_num_dma_blocks(umem, pgsz); in __rdma_umem_block_iter_start()
|
/linux/drivers/infiniband/hw/vmw_pvrdma/ |
H A D | pvrdma_srq.c | 155 srq->npages = ib_umem_num_dma_blocks(srq->umem, PAGE_SIZE); in pvrdma_create_srq()
|
H A D | pvrdma_mr.c | 136 npages = ib_umem_num_dma_blocks(umem, PAGE_SIZE); in pvrdma_reg_user_mr()
|
H A D | pvrdma_cq.c | 149 npages = ib_umem_num_dma_blocks(cq->umem, PAGE_SIZE); in pvrdma_create_cq()
|
H A D | pvrdma_qp.c | 294 ib_umem_num_dma_blocks(qp->sumem, PAGE_SIZE); in pvrdma_create_qp() 296 qp->npages_recv = ib_umem_num_dma_blocks( in pvrdma_create_qp()
|
/linux/drivers/infiniband/hw/hns/ |
H A D | hns_roce_mr.c | 969 ba_cnt = ib_umem_num_dma_blocks(mtr->umem, buf_pg_sz); in get_best_hop_num() 1036 ib_umem_num_dma_blocks(mtr->umem, buf_pg_sz) : in mtr_init_buf_cfg() 1050 r->count = ib_umem_num_dma_blocks(mtr->umem, buf_pg_sz); in mtr_init_buf_cfg()
|
/linux/drivers/infiniband/hw/mlx4/ |
H A D | srq.c | 123 dev->dev, ib_umem_num_dma_blocks(srq->umem, PAGE_SIZE), in mlx4_ib_create_srq()
|
/linux/drivers/infiniband/hw/bnxt_re/ |
H A D | qplib_res.c | 117 pages = ib_umem_num_dma_blocks(sginfo->umem, sginfo->pgsize); in __alloc_pbl() 218 npages = ib_umem_num_dma_blocks(hwq_attr->sginfo->umem, in bnxt_qplib_alloc_init_hwq()
|
H A D | ib_verbs.c | 4221 umem_pgs = ib_umem_num_dma_blocks(umem, page_size); in __bnxt_re_user_reg_mr()
|
/linux/drivers/infiniband/hw/mlx5/ |
H A D | cq.c | 772 ncont = ib_umem_num_dma_blocks(cq->buf.umem, page_size); in create_cq_user() 1322 npas = ib_umem_num_dma_blocks(cq->resize_umem, page_size); in mlx5_ib_resize_cq()
|
H A D | umr.c | 686 ib_umem_num_dma_blocks(mr->umem, 1 << mr->page_shift), in _mlx5r_umr_update_mr_pas()
|
H A D | srq_cmd.c | 103 ib_umem_num_dma_blocks(in->umem, page_size) * sizeof(u64))) in __set_srq_page_size()
|
H A D | mr.c | 1128 rb_key.ndescs = ib_umem_num_dma_blocks(umem, page_size); in alloc_cacheable_mr() 1249 roundup(ib_umem_num_dma_blocks(umem, page_size), 2); in reg_create() 1751 ib_umem_num_dma_blocks(new_umem, *page_size); in can_use_umr_rereg_pas()
|
H A D | qp.c | 909 rwq->rq_num_pas = ib_umem_num_dma_blocks(rwq->umem, page_size); in create_user_rq() 1017 ncont = ib_umem_num_dma_blocks(ubuffer->umem, page_size); in _create_user_qp() 1370 ib_umem_num_dma_blocks(sq->ubuffer.umem, page_size); in create_raw_packet_qp_sq() 1459 sizeof(u64) * ib_umem_num_dma_blocks(umem, page_size); in create_raw_packet_qp_rq()
|
/linux/drivers/infiniband/hw/cxgb4/ |
H A D | mem.c | 535 n = ib_umem_num_dma_blocks(mhp->umem, 1 << shift); in c4iw_reg_user_mr()
|
/linux/drivers/infiniband/hw/mana/ |
H A D | main.c | 366 num_pages_total = ib_umem_num_dma_blocks(umem, page_sz); in mana_ib_gd_create_dma_region()
|
/linux/drivers/infiniband/hw/mthca/ |
H A D | mthca_provider.c | 862 n = ib_umem_num_dma_blocks(mr->umem, PAGE_SIZE); in mthca_reg_user_mr()
|
/linux/drivers/infiniband/hw/qedr/ |
H A D | verbs.c | 809 fw_pages = ib_umem_num_dma_blocks(q->umem, 1 << FW_PAGE_SHIFT); in qedr_init_user_queue() 2984 ib_umem_num_dma_blocks(mr->umem, PAGE_SIZE), 1); in qedr_reg_user_mr()
|
/linux/drivers/infiniband/hw/irdma/ |
H A D | verbs.c | 2928 iwmr->page_cnt = ib_umem_num_dma_blocks(region, iwmr->page_size); in irdma_alloc_iwmr() 3205 iwmr->page_cnt = ib_umem_num_dma_blocks(region, iwmr->page_size); in irdma_rereg_mr_trans()
|
/linux/drivers/infiniband/hw/ocrdma/ |
H A D | ocrdma_verbs.c | 871 dev, mr, ib_umem_num_dma_blocks(mr->umem, PAGE_SIZE)); in ocrdma_reg_user_mr()
|