Searched refs:umem_dmabuf (Results 1 – 7 of 7) sorted by relevance
| /linux/drivers/infiniband/core/ |
| H A D | umem_dmabuf.c | 15 int ib_umem_dmabuf_map_pages(struct ib_umem_dmabuf *umem_dmabuf) in ib_umem_dmabuf_map_pages() argument 24 dma_resv_assert_held(umem_dmabuf->attach->dmabuf->resv); in ib_umem_dmabuf_map_pages() 26 if (umem_dmabuf->revoked) in ib_umem_dmabuf_map_pages() 29 if (umem_dmabuf->sgt) in ib_umem_dmabuf_map_pages() 32 sgt = dma_buf_map_attachment(umem_dmabuf->attach, in ib_umem_dmabuf_map_pages() 39 start = ALIGN_DOWN(umem_dmabuf->umem.address, PAGE_SIZE); in ib_umem_dmabuf_map_pages() 40 end = ALIGN(umem_dmabuf->umem.address + umem_dmabuf->umem.length, in ib_umem_dmabuf_map_pages() 48 umem_dmabuf->first_sg = sg; in ib_umem_dmabuf_map_pages() 49 umem_dmabuf in ib_umem_dmabuf_map_pages() 86 ib_umem_dmabuf_unmap_pages(struct ib_umem_dmabuf * umem_dmabuf) ib_umem_dmabuf_unmap_pages() argument 124 struct ib_umem_dmabuf *umem_dmabuf; ib_umem_dmabuf_get_with_dma_device() local 190 struct ib_umem_dmabuf *umem_dmabuf = attach->importer_priv; ib_umem_dmabuf_unsupported_move_notify() local 207 struct ib_umem_dmabuf *umem_dmabuf; ib_umem_dmabuf_get_pinned_with_dma_device() local 248 ib_umem_dmabuf_revoke(struct ib_umem_dmabuf * umem_dmabuf) ib_umem_dmabuf_revoke() argument 266 ib_umem_dmabuf_release(struct ib_umem_dmabuf * umem_dmabuf) ib_umem_dmabuf_release() argument [all...] |
| H A D | uverbs_std_types_cq.c | 69 struct ib_umem_dmabuf *umem_dmabuf; in UVERBS_HANDLER() local 163 umem_dmabuf = ib_umem_dmabuf_get_pinned(ib_dev, buffer_offset, buffer_length, in UVERBS_HANDLER() 165 if (IS_ERR(umem_dmabuf)) { in UVERBS_HANDLER() 166 ret = PTR_ERR(umem_dmabuf); in UVERBS_HANDLER() 169 umem = &umem_dmabuf->umem; in UVERBS_HANDLER()
|
| /linux/drivers/infiniband/hw/mlx5/ |
| H A D | odp.c | 835 struct ib_umem_dmabuf *umem_dmabuf = to_ib_umem_dmabuf(mr->umem); in pagefault_dmabuf_mr() local 847 dma_resv_lock(umem_dmabuf->attach->dmabuf->resv, NULL); in pagefault_dmabuf_mr() 848 err = ib_umem_dmabuf_map_pages(umem_dmabuf); in pagefault_dmabuf_mr() 850 dma_resv_unlock(umem_dmabuf->attach->dmabuf->resv); in pagefault_dmabuf_mr() 854 page_size = mlx5_umem_dmabuf_find_best_pgsz(umem_dmabuf, access_mode); in pagefault_dmabuf_mr() 856 ib_umem_dmabuf_unmap_pages(umem_dmabuf); in pagefault_dmabuf_mr() 873 dma_resv_unlock(umem_dmabuf->attach->dmabuf->resv); in pagefault_dmabuf_mr()
|
| H A D | devx.c | 2260 struct ib_umem_dmabuf *umem_dmabuf; in devx_umem_get() local 2268 umem_dmabuf = ib_umem_dmabuf_get_pinned( in devx_umem_get() 2270 if (IS_ERR(umem_dmabuf)) in devx_umem_get() 2271 return PTR_ERR(umem_dmabuf); in devx_umem_get() 2272 obj->umem = &umem_dmabuf->umem; in devx_umem_get()
|
| /linux/drivers/infiniband/hw/efa/ |
| H A D | efa_verbs.c | 1766 struct ib_umem_dmabuf *umem_dmabuf; in efa_reg_user_mr_dmabuf() local 1781 umem_dmabuf = ib_umem_dmabuf_get_pinned(ibpd->device, start, length, fd, in efa_reg_user_mr_dmabuf() 1783 if (IS_ERR(umem_dmabuf)) { in efa_reg_user_mr_dmabuf() 1784 err = PTR_ERR(umem_dmabuf); in efa_reg_user_mr_dmabuf() 1786 umem_dmabuf); in efa_reg_user_mr_dmabuf() 1790 mr->umem = &umem_dmabuf->umem; in efa_reg_user_mr_dmabuf()
|
| /linux/drivers/infiniband/hw/irdma/ |
| H A D | verbs.c | 3600 struct ib_umem_dmabuf *umem_dmabuf; in irdma_reg_user_mr_dmabuf() 3610 umem_dmabuf = ib_umem_dmabuf_get_pinned(pd->device, start, len, fd, access); in irdma_reg_user_mr_dmabuf() 3611 if (IS_ERR(umem_dmabuf)) { in irdma_reg_user_mr_dmabuf() 3613 umem_dmabuf); in irdma_reg_user_mr_dmabuf() 3614 return ERR_CAST(umem_dmabuf); in irdma_reg_user_mr_dmabuf() 3617 iwmr = irdma_alloc_iwmr(&umem_dmabuf->umem, pd, virt, IRDMA_MEMREG_TYPE_MEM); in irdma_reg_user_mr_dmabuf() 3633 ib_umem_release(&umem_dmabuf->umem); in irdma_reg_user_mr_dmabuf() 3597 struct ib_umem_dmabuf *umem_dmabuf; irdma_reg_user_mr_dmabuf() local
|
| /linux/drivers/infiniband/hw/bnxt_re/ |
| H A D | ib_verbs.c | 4337 struct ib_umem_dmabuf *umem_dmabuf; in bnxt_re_alloc_ucontext() 4344 umem_dmabuf = ib_umem_dmabuf_get_pinned(&rdev->ibdev, start, length, in bnxt_re_alloc_ucontext() 4346 if (IS_ERR(umem_dmabuf)) in bnxt_re_alloc_ucontext() 4347 return ERR_CAST(umem_dmabuf); in bnxt_re_alloc_ucontext() 4349 umem = &umem_dmabuf->umem; in bnxt_re_alloc_ucontext() 4310 struct ib_umem_dmabuf *umem_dmabuf; bnxt_re_reg_user_mr_dmabuf() local
|