Home
last modified time | relevance | path

Searched refs:ucmd (Results 1 – 25 of 34) sorted by relevance

12

/linux/drivers/iommu/iommufd/
H A Dioas.c38 int iommufd_ioas_alloc_ioctl(struct iommufd_ucmd *ucmd) in iommufd_ioas_alloc_ioctl() argument
40 struct iommu_ioas_alloc *cmd = ucmd->cmd; in iommufd_ioas_alloc_ioctl()
47 ioas = iommufd_ioas_alloc(ucmd->ictx); in iommufd_ioas_alloc_ioctl()
52 rc = iommufd_ucmd_respond(ucmd, sizeof(*cmd)); in iommufd_ioas_alloc_ioctl()
56 down_read(&ucmd->ictx->ioas_creation_lock); in iommufd_ioas_alloc_ioctl()
57 iommufd_object_finalize(ucmd->ictx, &ioas->obj); in iommufd_ioas_alloc_ioctl()
58 up_read(&ucmd->ictx->ioas_creation_lock); in iommufd_ioas_alloc_ioctl()
62 iommufd_object_abort_and_destroy(ucmd->ictx, &ioas->obj); in iommufd_ioas_alloc_ioctl()
66 int iommufd_ioas_iova_ranges(struct iommufd_ucmd *ucmd) in iommufd_ioas_iova_ranges() argument
69 struct iommu_ioas_iova_ranges *cmd = ucmd->cmd; in iommufd_ioas_iova_ranges()
[all …]
H A Dselftest.c99 void iommufd_test_syz_conv_iova_id(struct iommufd_ucmd *ucmd, in iommufd_test_syz_conv_iova_id() argument
108 ioas = iommufd_get_ioas(ucmd->ictx, ioas_id); in iommufd_test_syz_conv_iova_id()
112 iommufd_put_object(ucmd->ictx, &ioas->obj); in iommufd_test_syz_conv_iova_id()
958 __get_md_pagetable(struct iommufd_ucmd *ucmd, u32 mockpt_id, u32 hwpt_type) in __get_md_pagetable() argument
962 obj = iommufd_get_object(ucmd->ictx, mockpt_id, hwpt_type); in __get_md_pagetable()
969 get_md_pagetable(struct iommufd_ucmd *ucmd, u32 mockpt_id, in get_md_pagetable() argument
974 hwpt = __get_md_pagetable(ucmd, mockpt_id, IOMMUFD_OBJ_HWPT_PAGING); in get_md_pagetable()
979 iommufd_put_object(ucmd->ictx, &hwpt->obj); in get_md_pagetable()
987 get_md_pagetable_nested(struct iommufd_ucmd *ucmd, u32 mockpt_id, in get_md_pagetable_nested() argument
992 hwpt = __get_md_pagetable(ucmd, mockpt_i in get_md_pagetable_nested()
1077 iommufd_test_mock_domain(struct iommufd_ucmd * ucmd,struct iommu_test_cmd * cmd) iommufd_test_mock_domain() argument
1159 iommufd_test_mock_domain_replace(struct iommufd_ucmd * ucmd,unsigned int device_id,u32 pt_id,struct iommu_test_cmd * cmd) iommufd_test_mock_domain_replace() argument
1183 iommufd_test_add_reserved(struct iommufd_ucmd * ucmd,unsigned int mockpt_id,unsigned long start,size_t length) iommufd_test_add_reserved() argument
1207 iommufd_test_md_check_pa(struct iommufd_ucmd * ucmd,unsigned int mockpt_id,unsigned long iova,size_t length,void __user * uptr) iommufd_test_md_check_pa() argument
1265 iommufd_test_md_check_refs(struct iommufd_ucmd * ucmd,void __user * uptr,size_t length,unsigned int refs) iommufd_test_md_check_refs() argument
1299 iommufd_test_md_check_iotlb(struct iommufd_ucmd * ucmd,u32 mockpt_id,unsigned int iotlb_id,u32 iotlb) iommufd_test_md_check_iotlb() argument
1319 iommufd_test_dev_check_cache(struct iommufd_ucmd * ucmd,u32 idev_id,unsigned int cache_id,u32 cache) iommufd_test_dev_check_cache() argument
1391 iommufd_test_access_item_destroy(struct iommufd_ucmd * ucmd,unsigned int access_id,unsigned int item_id) iommufd_test_access_item_destroy() argument
1468 iommufd_test_create_access(struct iommufd_ucmd * ucmd,unsigned int ioas_id,unsigned int flags) iommufd_test_create_access() argument
1522 iommufd_test_access_replace_ioas(struct iommufd_ucmd * ucmd,unsigned int access_id,unsigned int ioas_id) iommufd_test_access_replace_ioas() argument
1560 iommufd_test_access_pages(struct iommufd_ucmd * ucmd,unsigned int access_id,unsigned long iova,size_t length,void __user * uptr,u32 flags) iommufd_test_access_pages() argument
1655 iommufd_test_access_rw(struct iommufd_ucmd * ucmd,unsigned int access_id,unsigned long iova,size_t length,void __user * ubuf,unsigned int flags) iommufd_test_access_rw() argument
1714 iommufd_test_dirty(struct iommufd_ucmd * ucmd,unsigned int mockpt_id,unsigned long iova,size_t length,unsigned long page_size,void __user * uptr,u32 flags) iommufd_test_dirty() argument
1768 iommufd_test_trigger_iopf(struct iommufd_ucmd * ucmd,struct iommu_test_cmd * cmd) iommufd_test_trigger_iopf() argument
1793 iommufd_test_trigger_vevent(struct iommufd_ucmd * ucmd,struct iommu_test_cmd * cmd) iommufd_test_trigger_vevent() argument
1822 iommufd_get_hwpt(struct iommufd_ucmd * ucmd,u32 id) iommufd_get_hwpt() argument
1839 iommufd_test_pasid_check_hwpt(struct iommufd_ucmd * ucmd,struct iommu_test_cmd * cmd) iommufd_test_pasid_check_hwpt() argument
1885 iommufd_test_pasid_attach(struct iommufd_ucmd * ucmd,struct iommu_test_cmd * cmd) iommufd_test_pasid_attach() argument
1909 iommufd_test_pasid_replace(struct iommufd_ucmd * ucmd,struct iommu_test_cmd * cmd) iommufd_test_pasid_replace() argument
1931 iommufd_test_pasid_detach(struct iommufd_ucmd * ucmd,struct iommu_test_cmd * cmd) iommufd_test_pasid_detach() argument
2022 iommufd_test_dmabuf_get(struct iommufd_ucmd * ucmd,unsigned int open_flags,size_t len) iommufd_test_dmabuf_get() argument
2065 iommufd_test_dmabuf_revoke(struct iommufd_ucmd * ucmd,int fd,bool revoked) iommufd_test_dmabuf_revoke() argument
2092 iommufd_test(struct iommufd_ucmd * ucmd) iommufd_test() argument
[all...]
H A Dviommu.c17 int iommufd_viommu_alloc_ioctl(struct iommufd_ucmd *ucmd) in iommufd_viommu_alloc_ioctl() argument
19 struct iommu_viommu_alloc *cmd = ucmd->cmd; in iommufd_viommu_alloc_ioctl()
35 idev = iommufd_get_device(ucmd, cmd->dev_id); in iommufd_viommu_alloc_ioctl()
60 hwpt_paging = iommufd_get_hwpt_paging(ucmd, cmd->hwpt_id); in iommufd_viommu_alloc_ioctl()
72 ucmd, viommu_size, IOMMUFD_OBJ_VIOMMU); in iommufd_viommu_alloc_ioctl()
80 viommu->ictx = ucmd->ictx; in iommufd_viommu_alloc_ioctl()
104 rc = iommufd_ucmd_respond(ucmd, sizeof(*cmd)); in iommufd_viommu_alloc_ioctl()
107 iommufd_put_object(ucmd->ictx, &hwpt_paging->common.obj); in iommufd_viommu_alloc_ioctl()
109 iommufd_put_object(ucmd->ictx, &idev->obj); in iommufd_viommu_alloc_ioctl()
143 int iommufd_vdevice_alloc_ioctl(struct iommufd_ucmd *ucmd) in iommufd_vdevice_alloc_ioctl() argument
[all …]
H A Dmain.c66 struct iommufd_object *_iommufd_object_alloc_ucmd(struct iommufd_ucmd *ucmd, in _iommufd_object_alloc_ucmd() argument
73 if (WARN_ON(ucmd->new_obj)) in _iommufd_object_alloc_ucmd()
85 new_obj = _iommufd_object_alloc(ucmd->ictx, size, type); in _iommufd_object_alloc_ucmd()
89 ucmd->new_obj = new_obj; in _iommufd_object_alloc_ucmd()
288 static int iommufd_destroy(struct iommufd_ucmd *ucmd) in iommufd_destroy() argument
290 struct iommu_destroy *cmd = ucmd->cmd; in iommufd_destroy()
292 return iommufd_object_remove(ucmd->ictx, NULL, cmd->id, 0); in iommufd_destroy()
389 static int iommufd_option(struct iommufd_ucmd *ucmd) in iommufd_option() argument
391 struct iommu_option *cmd = ucmd->cmd; in iommufd_option()
399 rc = iommufd_option_rlimit_mode(cmd, ucmd->ictx); in iommufd_option()
[all …]
H A Dvfio_compat.c123 int iommufd_vfio_ioas(struct iommufd_ucmd *ucmd) in iommufd_vfio_ioas() argument
125 struct iommu_vfio_ioas *cmd = ucmd->cmd; in iommufd_vfio_ioas()
132 ioas = get_compat_ioas(ucmd->ictx); in iommufd_vfio_ioas()
136 iommufd_put_object(ucmd->ictx, &ioas->obj); in iommufd_vfio_ioas()
137 return iommufd_ucmd_respond(ucmd, sizeof(*cmd)); in iommufd_vfio_ioas()
140 ioas = iommufd_get_ioas(ucmd->ictx, cmd->ioas_id); in iommufd_vfio_ioas()
143 xa_lock(&ucmd->ictx->objects); in iommufd_vfio_ioas()
144 ucmd->ictx->vfio_ioas = ioas; in iommufd_vfio_ioas()
145 xa_unlock(&ucmd->ictx->objects); in iommufd_vfio_ioas()
146 iommufd_put_object(ucmd->ictx, &ioas->obj); in iommufd_vfio_ioas()
[all …]
H A Deventq.c417 int iommufd_fault_alloc(struct iommufd_ucmd *ucmd) in iommufd_fault_alloc() argument
419 struct iommu_fault_alloc *cmd = ucmd->cmd; in iommufd_fault_alloc()
427 fault = __iommufd_object_alloc_ucmd(ucmd, fault, IOMMUFD_OBJ_FAULT, in iommufd_fault_alloc()
436 ucmd->ictx, &iommufd_fault_fops); in iommufd_fault_alloc()
443 rc = iommufd_ucmd_respond(ucmd, sizeof(*cmd)); in iommufd_fault_alloc()
475 int iommufd_veventq_alloc(struct iommufd_ucmd *ucmd) in iommufd_veventq_alloc() argument
477 struct iommu_veventq_alloc *cmd = ucmd->cmd; in iommufd_veventq_alloc()
489 viommu = iommufd_get_viommu(ucmd, cmd->viommu_id); in iommufd_veventq_alloc()
500 veventq = __iommufd_object_alloc(ucmd->ictx, veventq, in iommufd_veventq_alloc()
516 ucmd->ictx, &iommufd_veventq_fops); in iommufd_veventq_alloc()
[all …]
/linux/drivers/fwctl/
H A Dmain.c33 static int ucmd_respond(struct fwctl_ucmd *ucmd, size_t cmd_len) in ucmd_respond() argument
35 if (copy_to_user(ucmd->ubuffer, ucmd->cmd, in ucmd_respond()
36 min_t(size_t, ucmd->user_size, cmd_len))) in ucmd_respond()
56 static int fwctl_cmd_info(struct fwctl_ucmd *ucmd) in fwctl_cmd_info() argument
58 struct fwctl_device *fwctl = ucmd->uctx->fwctl; in fwctl_cmd_info()
59 struct fwctl_info *cmd = ucmd->cmd; in fwctl_cmd_info()
71 fwctl->ops->info(ucmd->uctx, &driver_info_len); in fwctl_cmd_info()
83 return ucmd_respond(ucmd, sizeof(*cmd)); in fwctl_cmd_info()
86 static int fwctl_cmd_rpc(struct fwctl_ucmd *ucmd) in fwctl_cmd_rpc() argument
88 struct fwctl_device *fwctl = ucmd->uctx->fwctl; in fwctl_cmd_rpc()
[all …]
/linux/kernel/liveupdate/
H A Dluo_core.c280 static int luo_ioctl_create_session(struct luo_ucmd *ucmd) in luo_ioctl_create_session()
282 struct liveupdate_ioctl_create_session *argp = ucmd->cmd; in luo_ioctl_create_session()
294 err = luo_ucmd_respond(ucmd, sizeof(*argp)); in luo_ioctl_create_session()
310 static int luo_ioctl_retrieve_session(struct luo_ucmd *ucmd) in luo_ioctl_retrieve_session()
312 struct liveupdate_ioctl_retrieve_session *argp = ucmd->cmd; in luo_ioctl_retrieve_session()
324 err = luo_ucmd_respond(ucmd, sizeof(*argp)); in luo_ioctl_retrieve_session()
377 int (*execute)(struct luo_ucmd *ucmd);
400 struct luo_ucmd ucmd = {}; in luo_ioctl()
409 ucmd.ubuffer = (void __user *)arg; in luo_ioctl()
410 err = get_user(ucmd in luo_ioctl()
276 luo_ioctl_create_session(struct luo_ucmd * ucmd) luo_ioctl_create_session() argument
306 luo_ioctl_retrieve_session(struct luo_ucmd * ucmd) luo_ioctl_retrieve_session() argument
396 struct luo_ucmd ucmd = {}; luo_ioctl() local
[all...]
H A Dluo_session.c231 struct luo_ucmd *ucmd) in luo_session_preserve_fd() argument
233 struct liveupdate_session_preserve_fd *argp = ucmd->cmd; in luo_session_preserve_fd()
241 err = luo_ucmd_respond(ucmd, sizeof(*argp)); in luo_session_preserve_fd()
249 struct luo_ucmd *ucmd) in luo_session_retrieve_fd() argument
251 struct liveupdate_session_retrieve_fd *argp = ucmd->cmd; in luo_session_retrieve_fd()
264 err = luo_ucmd_respond(ucmd, sizeof(*argp)); in luo_session_retrieve_fd()
281 struct luo_ucmd *ucmd) in luo_session_finish() argument
283 struct liveupdate_session_finish *argp = ucmd->cmd; in luo_session_finish()
289 return luo_ucmd_respond(ucmd, sizeof(*argp)); in luo_session_finish()
302 int (*execute)(struct luo_session *session, struct luo_ucmd *ucmd);
329 struct luo_ucmd ucmd = {}; luo_session_ioctl() local
[all...]
H A Dluo_internal.h20 static inline int luo_ucmd_respond(struct luo_ucmd *ucmd, in luo_ucmd_respond() argument
27 if (copy_to_user(ucmd->ubuffer, ucmd->cmd, in luo_ucmd_respond()
28 min_t(size_t, ucmd->user_size, kernel_cmd_size))) { in luo_ucmd_respond()
/linux/drivers/infiniband/hw/mana/
H A Dwq.c14 struct mana_ib_create_wq ucmd = {}; in mana_ib_create_wq() local
18 if (udata->inlen < sizeof(ucmd)) in mana_ib_create_wq()
21 err = ib_copy_from_udata(&ucmd, udata, min(sizeof(ucmd), udata->inlen)); in mana_ib_create_wq()
32 ibdev_dbg(&mdev->ib_dev, "ucmd wq_buf_addr 0x%llx\n", ucmd.wq_buf_addr); in mana_ib_create_wq()
34 err = mana_ib_create_queue(mdev, ucmd.wq_buf_addr, ucmd.wq_buf_size, &wq->queue); in mana_ib_create_wq()
42 wq->wq_buf_size = ucmd.wq_buf_size; in mana_ib_create_wq()
H A Dqp.c100 struct mana_ib_create_qp_rss ucmd = {}; in mana_ib_create_qp_rss() local
114 if (!udata || udata->inlen < sizeof(ucmd)) in mana_ib_create_qp_rss()
117 ret = ib_copy_from_udata(&ucmd, udata, min(sizeof(ucmd), udata->inlen)); in mana_ib_create_qp_rss()
147 if (ucmd.rx_hash_function != MANA_IB_RX_HASH_FUNC_TOEPLITZ) { in mana_ib_create_qp_rss()
150 ucmd.rx_hash_function); in mana_ib_create_qp_rss()
155 port = ucmd.port; in mana_ib_create_qp_rss()
165 ucmd.rx_hash_function, port); in mana_ib_create_qp_rss()
228 ucmd.rx_hash_key_len, in mana_ib_create_qp_rss()
229 ucmd in mana_ib_create_qp_rss()
275 struct mana_ib_create_qp ucmd = {}; mana_ib_create_qp_raw() local
533 struct mana_ib_create_rc_qp ucmd = {}; mana_ib_create_rc_qp() local
[all...]
H A Dcq.c16 struct mana_ib_create_cq ucmd = {}; in mana_ib_create_cq() local
33 err = ib_copy_from_udata(&ucmd, udata, min(sizeof(ucmd), udata->inlen)); in mana_ib_create_cq()
46 err = mana_ib_create_queue(mdev, ucmd.buf_addr, cq->cqe * COMP_ENTRY_SIZE, in mana_ib_create_cq()
/linux/drivers/infiniband/hw/mthca/
H A Dmthca_provider.c395 struct mthca_create_srq ucmd; in mthca_create_srq() local
405 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) in mthca_create_srq()
409 context->db_tab, ucmd.db_index, in mthca_create_srq()
410 ucmd.db_page); in mthca_create_srq()
415 srq->mr.ibmr.lkey = ucmd.lkey; in mthca_create_srq()
416 srq->db_index = ucmd.db_index; in mthca_create_srq()
424 context->db_tab, ucmd.db_index); in mthca_create_srq()
460 struct mthca_create_qp ucmd; in mthca_create_qp() local
474 if (ib_copy_from_udata(&ucmd, udat in mthca_create_qp()
582 struct mthca_create_cq ucmd; mthca_create_cq() local
701 struct mthca_resize_cq ucmd; mthca_resize_cq() local
836 struct mthca_reg_mr ucmd; mthca_reg_user_mr() local
[all...]
/linux/drivers/infiniband/hw/mlx5/
H A Dqp.c434 int has_rq, struct mlx5_ib_qp *qp, struct mlx5_ib_create_qp *ucmd) in set_rq_size() argument
452 if (ucmd) { in set_rq_size()
453 qp->rq.wqe_cnt = ucmd->rq_wqe_count; in set_rq_size()
454 if (ucmd->rq_wqe_shift > BITS_PER_BYTE * sizeof(ucmd->rq_wqe_shift)) in set_rq_size()
456 qp->rq.wqe_shift = ucmd->rq_wqe_shift; in set_rq_size()
638 struct mlx5_ib_create_qp *ucmd, in set_user_buf_size() argument
650 if (ucmd->sq_wqe_count && !is_power_of_2(ucmd->sq_wqe_count)) { in set_user_buf_size()
652 ucmd in set_user_buf_size()
882 create_user_rq(struct mlx5_ib_dev * dev,struct ib_pd * pd,struct ib_udata * udata,struct mlx5_ib_rwq * rwq,struct mlx5_ib_create_wq * ucmd) create_user_rq() argument
946 _create_user_qp(struct mlx5_ib_dev * dev,struct ib_pd * pd,struct mlx5_ib_qp * qp,struct ib_udata * udata,struct ib_qp_init_attr * attr,u32 ** in,struct mlx5_ib_create_qp_resp * resp,int * inlen,struct mlx5_ib_qp_base * base,struct mlx5_ib_create_qp * ucmd) _create_user_qp() argument
1706 void *ucmd; global() member
1718 struct mlx5_ib_create_qp_rss *ucmd = params->ucmd; create_rss_raw_qp_tir() local
2075 struct mlx5_ib_create_qp *ucmd = params->ucmd; create_dci() local
2236 struct mlx5_ib_create_qp *ucmd = params->ucmd; create_user_qp() local
2736 struct mlx5_ib_create_qp *ucmd = params->ucmd; create_dct() local
2898 process_vendor_flags(struct mlx5_ib_dev * dev,struct mlx5_ib_qp * qp,void * ucmd,struct ib_qp_init_attr * attr) process_vendor_flags() argument
3061 size_t ucmd = sizeof(struct mlx5_ib_create_qp); process_udata_size() local
3191 struct mlx5_ib_create_qp *ucmd = params->ucmd; get_qp_uidx() local
4125 __mlx5_ib_modify_qp(struct ib_qp * ibqp,const struct ib_qp_attr * attr,int attr_mask,enum ib_qp_state cur_state,enum ib_qp_state new_state,const struct mlx5_ib_modify_qp * ucmd,struct mlx5_ib_modify_qp_resp * resp,struct ib_udata * udata) __mlx5_ib_modify_qp() argument
4526 mlx5_ib_modify_dct(struct ib_qp * ibqp,struct ib_qp_attr * attr,int attr_mask,struct mlx5_ib_modify_qp * ucmd,struct ib_udata * udata) mlx5_ib_modify_dct() argument
4696 struct mlx5_ib_modify_qp ucmd = {}; mlx5_ib_modify_qp() local
5352 set_user_rq_size(struct mlx5_ib_dev * dev,struct ib_wq_init_attr * wq_init_attr,struct mlx5_ib_create_wq * ucmd,struct mlx5_ib_rwq * rwq) set_user_rq_size() argument
5391 struct mlx5_ib_create_wq ucmd = {}; prepare_user_rq() local
5629 struct mlx5_ib_modify_wq ucmd = {}; mlx5_ib_modify_wq() local
[all...]
H A Dcq.c723 struct mlx5_ib_create_cq ucmd = {}; in create_cq_user() local
734 ucmdlen = min(udata->inlen, sizeof(ucmd)); in create_cq_user()
738 if (ib_copy_from_udata(&ucmd, udata, ucmdlen)) in create_cq_user()
741 if ((ucmd.flags & ~(MLX5_IB_CREATE_CQ_FLAGS_CQE_128B_PAD | in create_cq_user()
746 if ((ucmd.cqe_size != 64 && ucmd.cqe_size != 128) || in create_cq_user()
747 ucmd.reserved0 || ucmd.reserved1) in create_cq_user()
750 *cqe_size = ucmd.cqe_size; in create_cq_user()
753 ib_umem_get(&dev->ib_dev, ucmd in create_cq_user()
1179 struct mlx5_ib_resize_cq ucmd; resize_user() local
[all...]
H A Dsrq.c48 struct mlx5_ib_create_srq ucmd = {}; in create_srq_user() local
55 ucmdlen = min(udata->inlen, sizeof(ucmd)); in create_srq_user()
57 if (ib_copy_from_udata(&ucmd, udata, ucmdlen)) { in create_srq_user()
62 if (ucmd.reserved0 || ucmd.reserved1) in create_srq_user()
65 if (udata->inlen > sizeof(ucmd) && in create_srq_user()
66 !ib_is_udata_cleared(udata, sizeof(ucmd), in create_srq_user()
67 udata->inlen - sizeof(ucmd))) in create_srq_user()
71 err = get_srq_user_index(ucontext, &ucmd, udata->inlen, &uidx); in create_srq_user()
76 srq->wq_sig = !!(ucmd in create_srq_user()
[all...]
/linux/drivers/infiniband/hw/hns/
H A Dhns_roce_srq.c345 struct hns_roce_ib_create_srq ucmd = {}; in alloc_srq_buf() local
349 ret = ib_copy_from_udata(&ucmd, udata, in alloc_srq_buf()
350 min(udata->inlen, sizeof(ucmd))); in alloc_srq_buf()
359 ret = alloc_srq_idx(hr_dev, srq, udata, ucmd.que_addr); in alloc_srq_buf()
363 ret = alloc_srq_wqe_buf(hr_dev, srq, udata, ucmd.buf_addr); in alloc_srq_buf()
391 struct hns_roce_ib_create_srq *ucmd) in get_srq_ucmd() argument
396 ret = ib_copy_from_udata(ucmd, udata, min(udata->inlen, sizeof(*ucmd))); in get_srq_ucmd()
428 struct hns_roce_ib_create_srq ucmd = {}; in alloc_srq_db() local
433 ret = get_srq_ucmd(srq, udata, &ucmd); in alloc_srq_db()
438 (ucmd.req_cap_flags & HNS_ROCE_SRQ_CAP_RECORD_DB)) { in alloc_srq_db()
[all …]
/linux/drivers/infiniband/hw/mlx4/
H A Dqp.c448 struct mlx4_ib_create_qp *ucmd) in set_user_sq_size() argument
453 if (check_shl_overflow(1, ucmd->log_sq_bb_count, &cnt) || in set_user_sq_size()
456 if (ucmd->log_sq_stride > in set_user_sq_size()
458 ucmd->log_sq_stride < MLX4_IB_MIN_SQ_STRIDE) in set_user_sq_size()
461 qp->sq.wqe_cnt = 1 << ucmd->log_sq_bb_count; in set_user_sq_size()
462 qp->sq.wqe_shift = ucmd->log_sq_stride; in set_user_sq_size()
552 struct mlx4_ib_create_qp_rss *ucmd) in set_qp_rss() argument
557 if ((ucmd->rx_hash_function == MLX4_IB_RX_HASH_FUNC_TOEPLITZ) && in set_qp_rss()
559 memcpy(rss_ctx->rss_key, ucmd->rx_hash_key, in set_qp_rss()
566 if (ucmd in set_qp_rss()
654 create_qp_rss(struct mlx4_ib_dev * dev,struct ib_qp_init_attr * init_attr,struct mlx4_ib_create_qp_rss * ucmd,struct mlx4_ib_qp * qp) create_qp_rss() argument
712 struct mlx4_ib_create_qp_rss ucmd = {}; _mlx4_ib_create_qp_rss() local
1069 struct mlx4_ib_create_qp ucmd; create_qp_common() local
4133 struct mlx4_ib_create_wq ucmd; mlx4_ib_create_wq() local
4271 struct mlx4_ib_modify_wq ucmd = {}; mlx4_ib_modify_wq() local
[all...]
H A Dsrq.c112 struct mlx4_ib_create_srq ucmd; in mlx4_ib_create_srq() local
114 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) in mlx4_ib_create_srq()
118 ib_umem_get(ib_srq->device, ucmd.buf_addr, buf_size, 0); in mlx4_ib_create_srq()
132 err = mlx4_ib_db_map_user(udata, ucmd.db_addr, &srq->db); in mlx4_ib_create_srq()
H A Dcq.c210 struct mlx4_ib_create_cq ucmd; in mlx4_ib_create_cq() local
212 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) { in mlx4_ib_create_cq()
217 buf_addr = (void *)(unsigned long)ucmd.buf_addr; in mlx4_ib_create_cq()
219 ucmd.buf_addr, entries); in mlx4_ib_create_cq()
223 err = mlx4_ib_db_map_user(udata, ucmd.db_addr, &cq->db); in mlx4_ib_create_cq()
322 struct mlx4_ib_resize_cq ucmd; in mlx4_alloc_resize_umem() local
328 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) in mlx4_alloc_resize_umem()
336 ucmd in mlx4_alloc_resize_umem()
[all...]
/linux/drivers/dma-buf/
H A Ddma-heap.c131 static long dma_heap_ioctl(struct file *file, unsigned int ucmd, in dma_heap_ioctl() argument
138 int nr = _IOC_NR(ucmd); in dma_heap_ioctl()
150 out_size = _IOC_SIZE(ucmd); in dma_heap_ioctl()
152 if ((ucmd & kcmd & IOC_IN) == 0) in dma_heap_ioctl()
154 if ((ucmd & kcmd & IOC_OUT) == 0) in dma_heap_ioctl()
/linux/drivers/infiniband/hw/vmw_pvrdma/
H A Dpvrdma_srq.c109 struct pvrdma_create_srq ucmd; in pvrdma_create_srq() local
144 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) { in pvrdma_create_srq()
149 srq->umem = ib_umem_get(ibsrq->device, ucmd.buf_addr, ucmd.buf_size, 0); in pvrdma_create_srq()
H A Dpvrdma_cq.c117 struct pvrdma_create_cq ucmd; in pvrdma_create_cq() local
137 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) { in pvrdma_create_cq()
142 cq->umem = ib_umem_get(ibdev, ucmd.buf_addr, ucmd.buf_size, in pvrdma_create_cq()
/linux/arch/mips/cavium-octeon/
H A Docteon-platform.c37 u32 ucmd; in octeon2_usb_reset() local
44 ucmd = cvmx_read64_uint32(CVMX_UAHCX_EHCI_USBCMD); in octeon2_usb_reset()
45 ucmd &= ~CMD_RUN; in octeon2_usb_reset()
46 cvmx_write64_uint32(CVMX_UAHCX_EHCI_USBCMD, ucmd); in octeon2_usb_reset()
48 ucmd |= CMD_RESET; in octeon2_usb_reset()
49 cvmx_write64_uint32(CVMX_UAHCX_EHCI_USBCMD, ucmd); in octeon2_usb_reset()
50 ucmd = cvmx_read64_uint32(CVMX_UAHCX_OHCI_USBCMD); in octeon2_usb_reset()
51 ucmd |= CMD_RUN; in octeon2_usb_reset()
52 cvmx_write64_uint32(CVMX_UAHCX_OHCI_USBCMD, ucmd); in octeon2_usb_reset()

12