/* * Copyright (c) 2005 Topspin Communications. All rights reserved. * Copyright (c) 2005 PathScale, Inc. All rights reserved. * Copyright (c) 2006 Cisco Systems, Inc. All rights reserved. * * This software is available to you under a choice of one of two * licenses. You may choose to be licensed under the terms of the GNU * General Public License (GPL) Version 2, available from the file * COPYING in the main directory of this source tree, or the * OpenIB.org BSD license below: * * Redistribution and use in source and binary forms, with or * without modification, are permitted provided that the following * conditions are met: * * - Redistributions of source code must retain the above * copyright notice, this list of conditions and the following * disclaimer. * * - Redistributions in binary form must reproduce the above * copyright notice, this list of conditions and the following * disclaimer in the documentation and/or other materials * provided with the distribution. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE * SOFTWARE. */ #include #include #include #include #include #include #include #include "ibverbs.h" #include int ibv_cmd_get_context(struct ibv_context *context, struct ibv_get_context *cmd, size_t cmd_size, struct ibv_get_context_resp *resp, size_t resp_size) { if (abi_ver < IB_USER_VERBS_MIN_ABI_VERSION) return ENOSYS; IBV_INIT_CMD_RESP(cmd, cmd_size, GET_CONTEXT, resp, resp_size); if (write(context->cmd_fd, cmd, cmd_size) != cmd_size) return errno; (void) VALGRIND_MAKE_MEM_DEFINED(resp, resp_size); context->async_fd = resp->async_fd; context->num_comp_vectors = resp->num_comp_vectors; return 0; } static void copy_query_dev_fields(struct ibv_device_attr *device_attr, struct ibv_query_device_resp *resp, uint64_t *raw_fw_ver) { *raw_fw_ver = resp->fw_ver; device_attr->node_guid = resp->node_guid; device_attr->sys_image_guid = resp->sys_image_guid; device_attr->max_mr_size = resp->max_mr_size; device_attr->page_size_cap = resp->page_size_cap; device_attr->vendor_id = resp->vendor_id; device_attr->vendor_part_id = resp->vendor_part_id; device_attr->hw_ver = resp->hw_ver; device_attr->max_qp = resp->max_qp; device_attr->max_qp_wr = resp->max_qp_wr; device_attr->device_cap_flags = resp->device_cap_flags; device_attr->max_sge = resp->max_sge; device_attr->max_sge_rd = resp->max_sge_rd; device_attr->max_cq = resp->max_cq; device_attr->max_cqe = resp->max_cqe; device_attr->max_mr = resp->max_mr; device_attr->max_pd = resp->max_pd; device_attr->max_qp_rd_atom = resp->max_qp_rd_atom; device_attr->max_ee_rd_atom = resp->max_ee_rd_atom; device_attr->max_res_rd_atom = resp->max_res_rd_atom; device_attr->max_qp_init_rd_atom = resp->max_qp_init_rd_atom; device_attr->max_ee_init_rd_atom = resp->max_ee_init_rd_atom; device_attr->atomic_cap = resp->atomic_cap; device_attr->max_ee = resp->max_ee; device_attr->max_rdd = resp->max_rdd; device_attr->max_mw = resp->max_mw; device_attr->max_raw_ipv6_qp = resp->max_raw_ipv6_qp; device_attr->max_raw_ethy_qp = resp->max_raw_ethy_qp; device_attr->max_mcast_grp = resp->max_mcast_grp; device_attr->max_mcast_qp_attach = resp->max_mcast_qp_attach; device_attr->max_total_mcast_qp_attach = resp->max_total_mcast_qp_attach; device_attr->max_ah = resp->max_ah; device_attr->max_fmr = resp->max_fmr; device_attr->max_map_per_fmr = resp->max_map_per_fmr; device_attr->max_srq = resp->max_srq; device_attr->max_srq_wr = resp->max_srq_wr; device_attr->max_srq_sge = resp->max_srq_sge; device_attr->max_pkeys = resp->max_pkeys; device_attr->local_ca_ack_delay = resp->local_ca_ack_delay; device_attr->phys_port_cnt = resp->phys_port_cnt; } int ibv_cmd_query_device(struct ibv_context *context, struct ibv_device_attr *device_attr, uint64_t *raw_fw_ver, struct ibv_query_device *cmd, size_t cmd_size) { struct ibv_query_device_resp resp; IBV_INIT_CMD_RESP(cmd, cmd_size, QUERY_DEVICE, &resp, sizeof resp); if (write(context->cmd_fd, cmd, cmd_size) != cmd_size) return errno; (void) VALGRIND_MAKE_MEM_DEFINED(&resp, sizeof resp); memset(device_attr->fw_ver, 0, sizeof device_attr->fw_ver); copy_query_dev_fields(device_attr, &resp, raw_fw_ver); return 0; } int ibv_cmd_query_device_ex(struct ibv_context *context, const struct ibv_query_device_ex_input *input, struct ibv_device_attr_ex *attr, size_t attr_size, uint64_t *raw_fw_ver, struct ibv_query_device_ex *cmd, size_t cmd_core_size, size_t cmd_size, struct ibv_query_device_resp_ex *resp, size_t resp_core_size, size_t resp_size) { int err; if (input && input->comp_mask) return EINVAL; if (attr_size < offsetof(struct ibv_device_attr_ex, comp_mask) + sizeof(attr->comp_mask)) return EINVAL; if (resp_core_size < offsetof(struct ibv_query_device_resp_ex, response_length) + sizeof(resp->response_length)) return EINVAL; IBV_INIT_CMD_RESP_EX_V(cmd, cmd_core_size, cmd_size, QUERY_DEVICE_EX, resp, resp_core_size, resp_size); cmd->comp_mask = 0; cmd->reserved = 0; memset(attr->orig_attr.fw_ver, 0, sizeof(attr->orig_attr.fw_ver)); memset(&attr->comp_mask, 0, attr_size - sizeof(attr->orig_attr)); err = write(context->cmd_fd, cmd, cmd_size); if (err != cmd_size) return errno; (void)VALGRIND_MAKE_MEM_DEFINED(resp, resp_size); copy_query_dev_fields(&attr->orig_attr, &resp->base, raw_fw_ver); /* Report back supported comp_mask bits. For now no comp_mask bit is * defined */ attr->comp_mask = resp->comp_mask & 0; if (attr_size >= offsetof(struct ibv_device_attr_ex, odp_caps) + sizeof(attr->odp_caps)) { if (resp->response_length >= offsetof(struct ibv_query_device_resp_ex, odp_caps) + sizeof(resp->odp_caps)) { attr->odp_caps.general_caps = resp->odp_caps.general_caps; attr->odp_caps.per_transport_caps.rc_odp_caps = resp->odp_caps.per_transport_caps.rc_odp_caps; attr->odp_caps.per_transport_caps.uc_odp_caps = resp->odp_caps.per_transport_caps.uc_odp_caps; attr->odp_caps.per_transport_caps.ud_odp_caps = resp->odp_caps.per_transport_caps.ud_odp_caps; } } if (attr_size >= offsetof(struct ibv_device_attr_ex, completion_timestamp_mask) + sizeof(attr->completion_timestamp_mask)) { if (resp->response_length >= offsetof(struct ibv_query_device_resp_ex, timestamp_mask) + sizeof(resp->timestamp_mask)) attr->completion_timestamp_mask = resp->timestamp_mask; } if (attr_size >= offsetof(struct ibv_device_attr_ex, hca_core_clock) + sizeof(attr->hca_core_clock)) { if (resp->response_length >= offsetof(struct ibv_query_device_resp_ex, hca_core_clock) + sizeof(resp->hca_core_clock)) attr->hca_core_clock = resp->hca_core_clock; } if (attr_size >= offsetof(struct ibv_device_attr_ex, device_cap_flags_ex) + sizeof(attr->device_cap_flags_ex)) { if (resp->response_length >= offsetof(struct ibv_query_device_resp_ex, device_cap_flags_ex) + sizeof(resp->device_cap_flags_ex)) attr->device_cap_flags_ex = resp->device_cap_flags_ex; } if (attr_size >= offsetof(struct ibv_device_attr_ex, rss_caps) + sizeof(attr->rss_caps)) { if (resp->response_length >= offsetof(struct ibv_query_device_resp_ex, rss_caps) + sizeof(resp->rss_caps)) { attr->rss_caps.supported_qpts = resp->rss_caps.supported_qpts; attr->rss_caps.max_rwq_indirection_tables = resp->rss_caps.max_rwq_indirection_tables; attr->rss_caps.max_rwq_indirection_table_size = resp->rss_caps.max_rwq_indirection_table_size; } } if (attr_size >= offsetof(struct ibv_device_attr_ex, max_wq_type_rq) + sizeof(attr->max_wq_type_rq)) { if (resp->response_length >= offsetof(struct ibv_query_device_resp_ex, max_wq_type_rq) + sizeof(resp->max_wq_type_rq)) attr->max_wq_type_rq = resp->max_wq_type_rq; } if (attr_size >= offsetof(struct ibv_device_attr_ex, raw_packet_caps) + sizeof(attr->raw_packet_caps)) { if (resp->response_length >= offsetof(struct ibv_query_device_resp_ex, raw_packet_caps) + sizeof(resp->raw_packet_caps)) attr->raw_packet_caps = resp->raw_packet_caps; } return 0; } int ibv_cmd_query_port(struct ibv_context *context, uint8_t port_num, struct ibv_port_attr *port_attr, struct ibv_query_port *cmd, size_t cmd_size) { struct ibv_query_port_resp resp; IBV_INIT_CMD_RESP(cmd, cmd_size, QUERY_PORT, &resp, sizeof resp); cmd->port_num = port_num; memset(cmd->reserved, 0, sizeof cmd->reserved); if (write(context->cmd_fd, cmd, cmd_size) != cmd_size) return errno; (void) VALGRIND_MAKE_MEM_DEFINED(&resp, sizeof resp); port_attr->state = resp.state; port_attr->max_mtu = resp.max_mtu; port_attr->active_mtu = resp.active_mtu; port_attr->gid_tbl_len = resp.gid_tbl_len; port_attr->port_cap_flags = resp.port_cap_flags; port_attr->max_msg_sz = resp.max_msg_sz; port_attr->bad_pkey_cntr = resp.bad_pkey_cntr; port_attr->qkey_viol_cntr = resp.qkey_viol_cntr; port_attr->pkey_tbl_len = resp.pkey_tbl_len; port_attr->lid = resp.lid; port_attr->sm_lid = resp.sm_lid; port_attr->lmc = resp.lmc; port_attr->max_vl_num = resp.max_vl_num; port_attr->sm_sl = resp.sm_sl; port_attr->subnet_timeout = resp.subnet_timeout; port_attr->init_type_reply = resp.init_type_reply; port_attr->active_width = resp.active_width; port_attr->active_speed = resp.active_speed; port_attr->phys_state = resp.phys_state; port_attr->link_layer = resp.link_layer; return 0; } int ibv_cmd_alloc_pd(struct ibv_context *context, struct ibv_pd *pd, struct ibv_alloc_pd *cmd, size_t cmd_size, struct ibv_alloc_pd_resp *resp, size_t resp_size) { IBV_INIT_CMD_RESP(cmd, cmd_size, ALLOC_PD, resp, resp_size); if (write(context->cmd_fd, cmd, cmd_size) != cmd_size) return errno; (void) VALGRIND_MAKE_MEM_DEFINED(resp, resp_size); pd->handle = resp->pd_handle; pd->context = context; return 0; } int ibv_cmd_dealloc_pd(struct ibv_pd *pd) { struct ibv_dealloc_pd cmd; IBV_INIT_CMD(&cmd, sizeof cmd, DEALLOC_PD); cmd.pd_handle = pd->handle; if (write(pd->context->cmd_fd, &cmd, sizeof cmd) != sizeof cmd) return errno; return 0; } int ibv_cmd_open_xrcd(struct ibv_context *context, struct verbs_xrcd *xrcd, int vxrcd_size, struct ibv_xrcd_init_attr *attr, struct ibv_open_xrcd *cmd, size_t cmd_size, struct ibv_open_xrcd_resp *resp, size_t resp_size) { IBV_INIT_CMD_RESP(cmd, cmd_size, OPEN_XRCD, resp, resp_size); if (attr->comp_mask >= IBV_XRCD_INIT_ATTR_RESERVED) return ENOSYS; if (!(attr->comp_mask & IBV_XRCD_INIT_ATTR_FD) || !(attr->comp_mask & IBV_XRCD_INIT_ATTR_OFLAGS)) return EINVAL; cmd->fd = attr->fd; cmd->oflags = attr->oflags; if (write(context->cmd_fd, cmd, cmd_size) != cmd_size) return errno; (void) VALGRIND_MAKE_MEM_DEFINED(resp, resp_size); xrcd->xrcd.context = context; xrcd->comp_mask = 0; if (vext_field_avail(struct verbs_xrcd, handle, vxrcd_size)) { xrcd->comp_mask = VERBS_XRCD_HANDLE; xrcd->handle = resp->xrcd_handle; } return 0; } int ibv_cmd_close_xrcd(struct verbs_xrcd *xrcd) { struct ibv_close_xrcd cmd; IBV_INIT_CMD(&cmd, sizeof cmd, CLOSE_XRCD); cmd.xrcd_handle = xrcd->handle; if (write(xrcd->xrcd.context->cmd_fd, &cmd, sizeof cmd) != sizeof cmd) return errno; return 0; } int ibv_cmd_reg_mr(struct ibv_pd *pd, void *addr, size_t length, uint64_t hca_va, int access, struct ibv_mr *mr, struct ibv_reg_mr *cmd, size_t cmd_size, struct ibv_reg_mr_resp *resp, size_t resp_size) { IBV_INIT_CMD_RESP(cmd, cmd_size, REG_MR, resp, resp_size); cmd->start = (uintptr_t) addr; cmd->length = length; cmd->hca_va = hca_va; cmd->pd_handle = pd->handle; cmd->access_flags = access; if (write(pd->context->cmd_fd, cmd, cmd_size) != cmd_size) return errno; (void) VALGRIND_MAKE_MEM_DEFINED(resp, resp_size); mr->handle = resp->mr_handle; mr->lkey = resp->lkey; mr->rkey = resp->rkey; mr->context = pd->context; return 0; } int ibv_cmd_rereg_mr(struct ibv_mr *mr, uint32_t flags, void *addr, size_t length, uint64_t hca_va, int access, struct ibv_pd *pd, struct ibv_rereg_mr *cmd, size_t cmd_sz, struct ibv_rereg_mr_resp *resp, size_t resp_sz) { IBV_INIT_CMD_RESP(cmd, cmd_sz, REREG_MR, resp, resp_sz); cmd->mr_handle = mr->handle; cmd->flags = flags; cmd->start = (uintptr_t)addr; cmd->length = length; cmd->hca_va = hca_va; cmd->pd_handle = (flags & IBV_REREG_MR_CHANGE_PD) ? pd->handle : 0; cmd->access_flags = access; if (write(mr->context->cmd_fd, cmd, cmd_sz) != cmd_sz) return errno; (void)VALGRIND_MAKE_MEM_DEFINED(resp, resp_sz); mr->lkey = resp->lkey; mr->rkey = resp->rkey; if (flags & IBV_REREG_MR_CHANGE_PD) mr->context = pd->context; return 0; } int ibv_cmd_dereg_mr(struct ibv_mr *mr) { struct ibv_dereg_mr cmd; IBV_INIT_CMD(&cmd, sizeof cmd, DEREG_MR); cmd.mr_handle = mr->handle; if (write(mr->context->cmd_fd, &cmd, sizeof cmd) != sizeof cmd) return errno; return 0; } int ibv_cmd_alloc_mw(struct ibv_pd *pd, enum ibv_mw_type type, struct ibv_mw *mw, struct ibv_alloc_mw *cmd, size_t cmd_size, struct ibv_alloc_mw_resp *resp, size_t resp_size) { IBV_INIT_CMD_RESP(cmd, cmd_size, ALLOC_MW, resp, resp_size); cmd->pd_handle = pd->handle; cmd->mw_type = type; memset(cmd->reserved, 0, sizeof(cmd->reserved)); if (write(pd->context->cmd_fd, cmd, cmd_size) != cmd_size) return errno; (void) VALGRIND_MAKE_MEM_DEFINED(resp, resp_size); mw->context = pd->context; mw->pd = pd; mw->rkey = resp->rkey; mw->handle = resp->mw_handle; mw->type = type; return 0; } int ibv_cmd_dealloc_mw(struct ibv_mw *mw, struct ibv_dealloc_mw *cmd, size_t cmd_size) { IBV_INIT_CMD(cmd, cmd_size, DEALLOC_MW); cmd->mw_handle = mw->handle; cmd->reserved = 0; if (write(mw->context->cmd_fd, cmd, cmd_size) != cmd_size) return errno; return 0; } int ibv_cmd_create_cq(struct ibv_context *context, int cqe, struct ibv_comp_channel *channel, int comp_vector, struct ibv_cq *cq, struct ibv_create_cq *cmd, size_t cmd_size, struct ibv_create_cq_resp *resp, size_t resp_size) { IBV_INIT_CMD_RESP(cmd, cmd_size, CREATE_CQ, resp, resp_size); cmd->user_handle = (uintptr_t) cq; cmd->cqe = cqe; cmd->comp_vector = comp_vector; cmd->comp_channel = channel ? channel->fd : -1; cmd->reserved = 0; if (write(context->cmd_fd, cmd, cmd_size) != cmd_size) return errno; (void) VALGRIND_MAKE_MEM_DEFINED(resp, resp_size); cq->handle = resp->cq_handle; cq->cqe = resp->cqe; cq->context = context; return 0; } int ibv_cmd_create_cq_ex(struct ibv_context *context, struct ibv_cq_init_attr_ex *cq_attr, struct ibv_cq_ex *cq, struct ibv_create_cq_ex *cmd, size_t cmd_core_size, size_t cmd_size, struct ibv_create_cq_resp_ex *resp, size_t resp_core_size, size_t resp_size) { int err; memset(cmd, 0, cmd_core_size); IBV_INIT_CMD_RESP_EX_V(cmd, cmd_core_size, cmd_size, CREATE_CQ_EX, resp, resp_core_size, resp_size); if (cq_attr->comp_mask & ~(IBV_CQ_INIT_ATTR_MASK_RESERVED - 1)) return EINVAL; cmd->user_handle = (uintptr_t)cq; cmd->cqe = cq_attr->cqe; cmd->comp_vector = cq_attr->comp_vector; cmd->comp_channel = cq_attr->channel ? cq_attr->channel->fd : -1; cmd->comp_mask = 0; if (cmd_core_size >= offsetof(struct ibv_create_cq_ex, flags) + sizeof(cmd->flags)) { if ((cq_attr->comp_mask & IBV_CQ_INIT_ATTR_MASK_FLAGS) && (cq_attr->flags & ~(IBV_CREATE_CQ_ATTR_RESERVED - 1))) return EOPNOTSUPP; if (cq_attr->wc_flags & IBV_WC_EX_WITH_COMPLETION_TIMESTAMP) cmd->flags |= IBV_CREATE_CQ_EX_KERNEL_FLAG_COMPLETION_TIMESTAMP; } err = write(context->cmd_fd, cmd, cmd_size); if (err != cmd_size) return errno; (void)VALGRIND_MAKE_MEM_DEFINED(resp, resp_size); cq->handle = resp->base.cq_handle; cq->cqe = resp->base.cqe; cq->context = context; return 0; } int ibv_cmd_poll_cq(struct ibv_cq *ibcq, int ne, struct ibv_wc *wc) { struct ibv_poll_cq cmd; struct ibv_poll_cq_resp *resp; int i; int rsize; int ret; rsize = sizeof *resp + ne * sizeof(struct ibv_kern_wc); resp = malloc(rsize); if (!resp) return -1; IBV_INIT_CMD_RESP(&cmd, sizeof cmd, POLL_CQ, resp, rsize); cmd.cq_handle = ibcq->handle; cmd.ne = ne; if (write(ibcq->context->cmd_fd, &cmd, sizeof cmd) != sizeof cmd) { ret = -1; goto out; } (void) VALGRIND_MAKE_MEM_DEFINED(resp, rsize); for (i = 0; i < resp->count; i++) { wc[i].wr_id = resp->wc[i].wr_id; wc[i].status = resp->wc[i].status; wc[i].opcode = resp->wc[i].opcode; wc[i].vendor_err = resp->wc[i].vendor_err; wc[i].byte_len = resp->wc[i].byte_len; wc[i].imm_data = resp->wc[i].imm_data; wc[i].qp_num = resp->wc[i].qp_num; wc[i].src_qp = resp->wc[i].src_qp; wc[i].wc_flags = resp->wc[i].wc_flags; wc[i].pkey_index = resp->wc[i].pkey_index; wc[i].slid = resp->wc[i].slid; wc[i].sl = resp->wc[i].sl; wc[i].dlid_path_bits = resp->wc[i].dlid_path_bits; } ret = resp->count; out: free(resp); return ret; } int ibv_cmd_req_notify_cq(struct ibv_cq *ibcq, int solicited_only) { struct ibv_req_notify_cq cmd; IBV_INIT_CMD(&cmd, sizeof cmd, REQ_NOTIFY_CQ); cmd.cq_handle = ibcq->handle; cmd.solicited = !!solicited_only; if (write(ibcq->context->cmd_fd, &cmd, sizeof cmd) != sizeof cmd) return errno; return 0; } int ibv_cmd_resize_cq(struct ibv_cq *cq, int cqe, struct ibv_resize_cq *cmd, size_t cmd_size, struct ibv_resize_cq_resp *resp, size_t resp_size) { IBV_INIT_CMD_RESP(cmd, cmd_size, RESIZE_CQ, resp, resp_size); cmd->cq_handle = cq->handle; cmd->cqe = cqe; if (write(cq->context->cmd_fd, cmd, cmd_size) != cmd_size) return errno; (void) VALGRIND_MAKE_MEM_DEFINED(resp, resp_size); cq->cqe = resp->cqe; return 0; } int ibv_cmd_destroy_cq(struct ibv_cq *cq) { struct ibv_destroy_cq cmd; struct ibv_destroy_cq_resp resp; IBV_INIT_CMD_RESP(&cmd, sizeof cmd, DESTROY_CQ, &resp, sizeof resp); cmd.cq_handle = cq->handle; cmd.reserved = 0; if (write(cq->context->cmd_fd, &cmd, sizeof cmd) != sizeof cmd) return errno; (void) VALGRIND_MAKE_MEM_DEFINED(&resp, sizeof resp); pthread_mutex_lock(&cq->mutex); while (cq->comp_events_completed != resp.comp_events_reported || cq->async_events_completed != resp.async_events_reported) pthread_cond_wait(&cq->cond, &cq->mutex); pthread_mutex_unlock(&cq->mutex); return 0; } int ibv_cmd_create_srq(struct ibv_pd *pd, struct ibv_srq *srq, struct ibv_srq_init_attr *attr, struct ibv_create_srq *cmd, size_t cmd_size, struct ibv_create_srq_resp *resp, size_t resp_size) { IBV_INIT_CMD_RESP(cmd, cmd_size, CREATE_SRQ, resp, resp_size); cmd->user_handle = (uintptr_t) srq; cmd->pd_handle = pd->handle; cmd->max_wr = attr->attr.max_wr; cmd->max_sge = attr->attr.max_sge; cmd->srq_limit = attr->attr.srq_limit; if (write(pd->context->cmd_fd, cmd, cmd_size) != cmd_size) return errno; (void) VALGRIND_MAKE_MEM_DEFINED(resp, resp_size); srq->handle = resp->srq_handle; srq->context = pd->context; if (abi_ver > 5) { attr->attr.max_wr = resp->max_wr; attr->attr.max_sge = resp->max_sge; } else { struct ibv_create_srq_resp_v5 *resp_v5 = (struct ibv_create_srq_resp_v5 *) resp; memmove((void *) resp + sizeof *resp, (void *) resp_v5 + sizeof *resp_v5, resp_size - sizeof *resp); } return 0; } int ibv_cmd_create_srq_ex(struct ibv_context *context, struct verbs_srq *srq, int vsrq_sz, struct ibv_srq_init_attr_ex *attr_ex, struct ibv_create_xsrq *cmd, size_t cmd_size, struct ibv_create_srq_resp *resp, size_t resp_size) { struct verbs_xrcd *vxrcd = NULL; int ret = 0; IBV_INIT_CMD_RESP(cmd, cmd_size, CREATE_XSRQ, resp, resp_size); if (attr_ex->comp_mask >= IBV_SRQ_INIT_ATTR_RESERVED) return ENOSYS; if (!(attr_ex->comp_mask & IBV_SRQ_INIT_ATTR_PD)) return EINVAL; cmd->user_handle = (uintptr_t) srq; cmd->pd_handle = attr_ex->pd->handle; cmd->max_wr = attr_ex->attr.max_wr; cmd->max_sge = attr_ex->attr.max_sge; cmd->srq_limit = attr_ex->attr.srq_limit; cmd->srq_type = (attr_ex->comp_mask & IBV_SRQ_INIT_ATTR_TYPE) ? attr_ex->srq_type : IBV_SRQT_BASIC; if (attr_ex->comp_mask & IBV_SRQ_INIT_ATTR_XRCD) { if (!(attr_ex->comp_mask & IBV_SRQ_INIT_ATTR_CQ)) return EINVAL; vxrcd = container_of(attr_ex->xrcd, struct verbs_xrcd, xrcd); cmd->xrcd_handle = vxrcd->handle; cmd->cq_handle = attr_ex->cq->handle; } ret = pthread_mutex_init(&srq->srq.mutex, NULL); if (ret) goto err; ret = pthread_cond_init(&srq->srq.cond, NULL); if (ret) goto err_mutex; if (write(context->cmd_fd, cmd, cmd_size) != cmd_size) { ret = errno; goto err_cond; } (void) VALGRIND_MAKE_MEM_DEFINED(resp, resp_size); srq->srq.handle = resp->srq_handle; srq->srq.context = context; srq->srq.srq_context = attr_ex->srq_context; srq->srq.pd = attr_ex->pd; srq->srq.events_completed = 0; /* * check that the last field is available. * If it is than all the others exist as well */ if (vext_field_avail(struct verbs_srq, srq_num, vsrq_sz)) { srq->comp_mask = IBV_SRQ_INIT_ATTR_TYPE; srq->srq_type = (attr_ex->comp_mask & IBV_SRQ_INIT_ATTR_TYPE) ? attr_ex->srq_type : IBV_SRQT_BASIC; if (srq->srq_type == IBV_SRQT_XRC) { srq->comp_mask |= VERBS_SRQ_NUM; srq->srq_num = resp->srqn; } if (attr_ex->comp_mask & IBV_SRQ_INIT_ATTR_XRCD) { srq->comp_mask |= VERBS_SRQ_XRCD; srq->xrcd = vxrcd; } if (attr_ex->comp_mask & IBV_SRQ_INIT_ATTR_CQ) { srq->comp_mask |= VERBS_SRQ_CQ; srq->cq = attr_ex->cq; } } attr_ex->attr.max_wr = resp->max_wr; attr_ex->attr.max_sge = resp->max_sge; return 0; err_cond: pthread_cond_destroy(&srq->srq.cond); err_mutex: pthread_mutex_destroy(&srq->srq.mutex); err: return ret; } static int ibv_cmd_modify_srq_v3(struct ibv_srq *srq, struct ibv_srq_attr *srq_attr, int srq_attr_mask, struct ibv_modify_srq *new_cmd, size_t new_cmd_size) { struct ibv_modify_srq_v3 *cmd; size_t cmd_size; cmd_size = sizeof *cmd + new_cmd_size - sizeof *new_cmd; cmd = alloca(cmd_size); memcpy(cmd->driver_data, new_cmd->driver_data, new_cmd_size - sizeof *new_cmd); IBV_INIT_CMD(cmd, cmd_size, MODIFY_SRQ); cmd->srq_handle = srq->handle; cmd->attr_mask = srq_attr_mask; cmd->max_wr = srq_attr->max_wr; cmd->srq_limit = srq_attr->srq_limit; cmd->max_sge = 0; cmd->reserved = 0; if (write(srq->context->cmd_fd, cmd, cmd_size) != cmd_size) return errno; return 0; } int ibv_cmd_modify_srq(struct ibv_srq *srq, struct ibv_srq_attr *srq_attr, int srq_attr_mask, struct ibv_modify_srq *cmd, size_t cmd_size) { if (abi_ver == 3) return ibv_cmd_modify_srq_v3(srq, srq_attr, srq_attr_mask, cmd, cmd_size); IBV_INIT_CMD(cmd, cmd_size, MODIFY_SRQ); cmd->srq_handle = srq->handle; cmd->attr_mask = srq_attr_mask; cmd->max_wr = srq_attr->max_wr; cmd->srq_limit = srq_attr->srq_limit; if (write(srq->context->cmd_fd, cmd, cmd_size) != cmd_size) return errno; return 0; } int ibv_cmd_query_srq(struct ibv_srq *srq, struct ibv_srq_attr *srq_attr, struct ibv_query_srq *cmd, size_t cmd_size) { struct ibv_query_srq_resp resp; IBV_INIT_CMD_RESP(cmd, cmd_size, QUERY_SRQ, &resp, sizeof resp); cmd->srq_handle = srq->handle; cmd->reserved = 0; if (write(srq->context->cmd_fd, cmd, cmd_size) != cmd_size) return errno; (void) VALGRIND_MAKE_MEM_DEFINED(&resp, sizeof resp); srq_attr->max_wr = resp.max_wr; srq_attr->max_sge = resp.max_sge; srq_attr->srq_limit = resp.srq_limit; return 0; } int ibv_cmd_destroy_srq(struct ibv_srq *srq) { struct ibv_destroy_srq cmd; struct ibv_destroy_srq_resp resp; IBV_INIT_CMD_RESP(&cmd, sizeof cmd, DESTROY_SRQ, &resp, sizeof resp); cmd.srq_handle = srq->handle; cmd.reserved = 0; if (write(srq->context->cmd_fd, &cmd, sizeof cmd) != sizeof cmd) return errno; (void) VALGRIND_MAKE_MEM_DEFINED(&resp, sizeof resp); pthread_mutex_lock(&srq->mutex); while (srq->events_completed != resp.events_reported) pthread_cond_wait(&srq->cond, &srq->mutex); pthread_mutex_unlock(&srq->mutex); pthread_cond_destroy(&srq->cond); pthread_mutex_destroy(&srq->mutex); return 0; } static int create_qp_ex_common(struct verbs_qp *qp, struct ibv_qp_init_attr_ex *qp_attr, struct verbs_xrcd *vxrcd, struct ibv_create_qp_common *cmd) { cmd->user_handle = (uintptr_t)qp; if (qp_attr->comp_mask & IBV_QP_INIT_ATTR_XRCD) { vxrcd = container_of(qp_attr->xrcd, struct verbs_xrcd, xrcd); cmd->pd_handle = vxrcd->handle; } else { if (!(qp_attr->comp_mask & IBV_QP_INIT_ATTR_PD)) return EINVAL; cmd->pd_handle = qp_attr->pd->handle; if (qp_attr->comp_mask & IBV_QP_INIT_ATTR_IND_TABLE) { if (cmd->max_recv_wr || cmd->max_recv_sge || cmd->recv_cq_handle || qp_attr->srq) return EINVAL; /* send_cq is optinal */ if (qp_attr->cap.max_send_wr) cmd->send_cq_handle = qp_attr->send_cq->handle; } else { cmd->send_cq_handle = qp_attr->send_cq->handle; if (qp_attr->qp_type != IBV_QPT_XRC_SEND) { cmd->recv_cq_handle = qp_attr->recv_cq->handle; cmd->srq_handle = qp_attr->srq ? qp_attr->srq->handle : 0; } } } cmd->max_send_wr = qp_attr->cap.max_send_wr; cmd->max_recv_wr = qp_attr->cap.max_recv_wr; cmd->max_send_sge = qp_attr->cap.max_send_sge; cmd->max_recv_sge = qp_attr->cap.max_recv_sge; cmd->max_inline_data = qp_attr->cap.max_inline_data; cmd->sq_sig_all = qp_attr->sq_sig_all; cmd->qp_type = qp_attr->qp_type; cmd->is_srq = !!qp_attr->srq; cmd->reserved = 0; return 0; } static int create_qp_handle_resp_common_cleanup(struct verbs_qp *qp) { pthread_cond_destroy(&qp->qp.cond); pthread_mutex_destroy(&qp->qp.mutex); } static int create_qp_handle_resp_common_init(struct verbs_qp *qp) { int ret = 0; ret = pthread_mutex_init(&qp->qp.mutex, NULL); if (ret) return ret; ret = pthread_cond_init(&qp->qp.cond, NULL); if (ret) goto err; return ret; err: pthread_mutex_destroy(&qp->qp.mutex); return ret; } static void create_qp_handle_resp_common(struct ibv_context *context, struct verbs_qp *qp, struct ibv_qp_init_attr_ex *qp_attr, struct ibv_create_qp_resp *resp, struct verbs_xrcd *vxrcd, int vqp_sz) { if (abi_ver > 3) { qp_attr->cap.max_recv_sge = resp->max_recv_sge; qp_attr->cap.max_send_sge = resp->max_send_sge; qp_attr->cap.max_recv_wr = resp->max_recv_wr; qp_attr->cap.max_send_wr = resp->max_send_wr; qp_attr->cap.max_inline_data = resp->max_inline_data; } qp->qp.handle = resp->qp_handle; qp->qp.qp_num = resp->qpn; qp->qp.context = context; qp->qp.qp_context = qp_attr->qp_context; qp->qp.pd = qp_attr->pd; qp->qp.send_cq = qp_attr->send_cq; qp->qp.recv_cq = qp_attr->recv_cq; qp->qp.srq = qp_attr->srq; qp->qp.qp_type = qp_attr->qp_type; qp->qp.state = IBV_QPS_RESET; qp->qp.events_completed = 0; qp->comp_mask = 0; if (vext_field_avail(struct verbs_qp, xrcd, vqp_sz) && (qp_attr->comp_mask & IBV_QP_INIT_ATTR_XRCD)) { qp->comp_mask |= VERBS_QP_XRCD; qp->xrcd = vxrcd; } } enum { CREATE_QP_EX2_SUP_CREATE_FLAGS = IBV_QP_CREATE_BLOCK_SELF_MCAST_LB | IBV_QP_CREATE_SCATTER_FCS | IBV_QP_CREATE_CVLAN_STRIPPING, }; int ibv_cmd_create_qp_ex2(struct ibv_context *context, struct verbs_qp *qp, int vqp_sz, struct ibv_qp_init_attr_ex *qp_attr, struct ibv_create_qp_ex *cmd, size_t cmd_core_size, size_t cmd_size, struct ibv_create_qp_resp_ex *resp, size_t resp_core_size, size_t resp_size) { struct verbs_xrcd *vxrcd = NULL; int err; if (qp_attr->comp_mask >= IBV_QP_INIT_ATTR_RESERVED) return EINVAL; if (resp_core_size < offsetof(struct ibv_create_qp_resp_ex, response_length) + sizeof(resp->response_length)) return EINVAL; memset(cmd, 0, cmd_core_size); IBV_INIT_CMD_RESP_EX_V(cmd, cmd_core_size, cmd_size, CREATE_QP_EX, resp, resp_core_size, resp_size); err = create_qp_ex_common(qp, qp_attr, vxrcd, &cmd->base); if (err) return err; if (qp_attr->comp_mask & IBV_QP_INIT_ATTR_CREATE_FLAGS) { if (qp_attr->create_flags & ~CREATE_QP_EX2_SUP_CREATE_FLAGS) return EINVAL; if (cmd_core_size < offsetof(struct ibv_create_qp_ex, create_flags) + sizeof(qp_attr->create_flags)) return EINVAL; cmd->create_flags = qp_attr->create_flags; } if (qp_attr->comp_mask & IBV_QP_INIT_ATTR_IND_TABLE) { if (cmd_core_size < offsetof(struct ibv_create_qp_ex, ind_tbl_handle) + sizeof(cmd->ind_tbl_handle)) return EINVAL; cmd->ind_tbl_handle = qp_attr->rwq_ind_tbl->ind_tbl_handle; cmd->comp_mask = IBV_CREATE_QP_EX_KERNEL_MASK_IND_TABLE; } err = create_qp_handle_resp_common_init(qp); if (err) return err; err = write(context->cmd_fd, cmd, cmd_size); if (err != cmd_size) { err = errno; goto err; } (void)VALGRIND_MAKE_MEM_DEFINED(resp, resp_size); create_qp_handle_resp_common(context, qp, qp_attr, &resp->base, vxrcd, vqp_sz); return 0; err: create_qp_handle_resp_common_cleanup(qp); return err; } int ibv_cmd_create_qp_ex(struct ibv_context *context, struct verbs_qp *qp, int vqp_sz, struct ibv_qp_init_attr_ex *attr_ex, struct ibv_create_qp *cmd, size_t cmd_size, struct ibv_create_qp_resp *resp, size_t resp_size) { struct verbs_xrcd *vxrcd = NULL; int err; IBV_INIT_CMD_RESP(cmd, cmd_size, CREATE_QP, resp, resp_size); if (attr_ex->comp_mask > (IBV_QP_INIT_ATTR_XRCD | IBV_QP_INIT_ATTR_PD)) return ENOSYS; err = create_qp_ex_common(qp, attr_ex, vxrcd, (struct ibv_create_qp_common *)&cmd->user_handle); if (err) return err; err = create_qp_handle_resp_common_init(qp); if (err) return err; err = write(context->cmd_fd, cmd, cmd_size); if (err != cmd_size) { err = errno; goto err; } (void)VALGRIND_MAKE_MEM_DEFINED(resp, resp_size); if (abi_ver == 4) { struct ibv_create_qp_resp_v4 *resp_v4 = (struct ibv_create_qp_resp_v4 *)resp; memmove((void *)resp + sizeof *resp, (void *)resp_v4 + sizeof *resp_v4, resp_size - sizeof *resp); } else if (abi_ver <= 3) { struct ibv_create_qp_resp_v3 *resp_v3 = (struct ibv_create_qp_resp_v3 *)resp; memmove((void *)resp + sizeof *resp, (void *)resp_v3 + sizeof *resp_v3, resp_size - sizeof *resp); } create_qp_handle_resp_common(context, qp, attr_ex, resp, vxrcd, vqp_sz); return 0; err: create_qp_handle_resp_common_cleanup(qp); return err; } int ibv_cmd_create_qp(struct ibv_pd *pd, struct ibv_qp *qp, struct ibv_qp_init_attr *attr, struct ibv_create_qp *cmd, size_t cmd_size, struct ibv_create_qp_resp *resp, size_t resp_size) { IBV_INIT_CMD_RESP(cmd, cmd_size, CREATE_QP, resp, resp_size); cmd->user_handle = (uintptr_t) qp; cmd->pd_handle = pd->handle; cmd->send_cq_handle = attr->send_cq->handle; cmd->recv_cq_handle = attr->recv_cq->handle; cmd->srq_handle = attr->srq ? attr->srq->handle : 0; cmd->max_send_wr = attr->cap.max_send_wr; cmd->max_recv_wr = attr->cap.max_recv_wr; cmd->max_send_sge = attr->cap.max_send_sge; cmd->max_recv_sge = attr->cap.max_recv_sge; cmd->max_inline_data = attr->cap.max_inline_data; cmd->sq_sig_all = attr->sq_sig_all; cmd->qp_type = attr->qp_type; cmd->is_srq = !!attr->srq; cmd->reserved = 0; if (write(pd->context->cmd_fd, cmd, cmd_size) != cmd_size) return errno; (void) VALGRIND_MAKE_MEM_DEFINED(resp, resp_size); qp->handle = resp->qp_handle; qp->qp_num = resp->qpn; qp->context = pd->context; if (abi_ver > 3) { attr->cap.max_recv_sge = resp->max_recv_sge; attr->cap.max_send_sge = resp->max_send_sge; attr->cap.max_recv_wr = resp->max_recv_wr; attr->cap.max_send_wr = resp->max_send_wr; attr->cap.max_inline_data = resp->max_inline_data; } if (abi_ver == 4) { struct ibv_create_qp_resp_v4 *resp_v4 = (struct ibv_create_qp_resp_v4 *) resp; memmove((void *) resp + sizeof *resp, (void *) resp_v4 + sizeof *resp_v4, resp_size - sizeof *resp); } else if (abi_ver <= 3) { struct ibv_create_qp_resp_v3 *resp_v3 = (struct ibv_create_qp_resp_v3 *) resp; memmove((void *) resp + sizeof *resp, (void *) resp_v3 + sizeof *resp_v3, resp_size - sizeof *resp); } return 0; } int ibv_cmd_open_qp(struct ibv_context *context, struct verbs_qp *qp, int vqp_sz, struct ibv_qp_open_attr *attr, struct ibv_open_qp *cmd, size_t cmd_size, struct ibv_create_qp_resp *resp, size_t resp_size) { int err = 0; struct verbs_xrcd *xrcd; IBV_INIT_CMD_RESP(cmd, cmd_size, OPEN_QP, resp, resp_size); if (attr->comp_mask >= IBV_QP_OPEN_ATTR_RESERVED) return ENOSYS; if (!(attr->comp_mask & IBV_QP_OPEN_ATTR_XRCD) || !(attr->comp_mask & IBV_QP_OPEN_ATTR_NUM) || !(attr->comp_mask & IBV_QP_OPEN_ATTR_TYPE)) return EINVAL; xrcd = container_of(attr->xrcd, struct verbs_xrcd, xrcd); cmd->user_handle = (uintptr_t) qp; cmd->pd_handle = xrcd->handle; cmd->qpn = attr->qp_num; cmd->qp_type = attr->qp_type; err = pthread_mutex_init(&qp->qp.mutex, NULL); if (err) return err; err = pthread_cond_init(&qp->qp.cond, NULL); if (err) goto err_mutex; err = write(context->cmd_fd, cmd, cmd_size); if (err != cmd_size) { err = errno; goto err_cond; } (void) VALGRIND_MAKE_MEM_DEFINED(resp, resp_size); qp->qp.handle = resp->qp_handle; qp->qp.context = context; qp->qp.qp_context = attr->qp_context; qp->qp.pd = NULL; qp->qp.send_cq = NULL; qp->qp.recv_cq = NULL; qp->qp.srq = NULL; qp->qp.qp_num = attr->qp_num; qp->qp.qp_type = attr->qp_type; qp->qp.state = IBV_QPS_UNKNOWN; qp->qp.events_completed = 0; qp->comp_mask = 0; if (vext_field_avail(struct verbs_qp, xrcd, vqp_sz)) { qp->comp_mask = VERBS_QP_XRCD; qp->xrcd = xrcd; } return 0; err_cond: pthread_cond_destroy(&qp->qp.cond); err_mutex: pthread_mutex_destroy(&qp->qp.mutex); return err; } int ibv_cmd_query_qp(struct ibv_qp *qp, struct ibv_qp_attr *attr, int attr_mask, struct ibv_qp_init_attr *init_attr, struct ibv_query_qp *cmd, size_t cmd_size) { struct ibv_query_qp_resp resp; /* * Masks over IBV_QP_DEST_QPN are not supported by * that not extended command. */ if (attr_mask & ~((IBV_QP_DEST_QPN << 1) - 1)) return EOPNOTSUPP; IBV_INIT_CMD_RESP(cmd, cmd_size, QUERY_QP, &resp, sizeof resp); cmd->qp_handle = qp->handle; cmd->attr_mask = attr_mask; if (write(qp->context->cmd_fd, cmd, cmd_size) != cmd_size) return errno; (void) VALGRIND_MAKE_MEM_DEFINED(&resp, sizeof resp); attr->qkey = resp.qkey; attr->rq_psn = resp.rq_psn; attr->sq_psn = resp.sq_psn; attr->dest_qp_num = resp.dest_qp_num; attr->qp_access_flags = resp.qp_access_flags; attr->pkey_index = resp.pkey_index; attr->alt_pkey_index = resp.alt_pkey_index; attr->qp_state = resp.qp_state; attr->cur_qp_state = resp.cur_qp_state; attr->path_mtu = resp.path_mtu; attr->path_mig_state = resp.path_mig_state; attr->sq_draining = resp.sq_draining; attr->max_rd_atomic = resp.max_rd_atomic; attr->max_dest_rd_atomic = resp.max_dest_rd_atomic; attr->min_rnr_timer = resp.min_rnr_timer; attr->port_num = resp.port_num; attr->timeout = resp.timeout; attr->retry_cnt = resp.retry_cnt; attr->rnr_retry = resp.rnr_retry; attr->alt_port_num = resp.alt_port_num; attr->alt_timeout = resp.alt_timeout; attr->cap.max_send_wr = resp.max_send_wr; attr->cap.max_recv_wr = resp.max_recv_wr; attr->cap.max_send_sge = resp.max_send_sge; attr->cap.max_recv_sge = resp.max_recv_sge; attr->cap.max_inline_data = resp.max_inline_data; memcpy(attr->ah_attr.grh.dgid.raw, resp.dest.dgid, 16); attr->ah_attr.grh.flow_label = resp.dest.flow_label; attr->ah_attr.dlid = resp.dest.dlid; attr->ah_attr.grh.sgid_index = resp.dest.sgid_index; attr->ah_attr.grh.hop_limit = resp.dest.hop_limit; attr->ah_attr.grh.traffic_class = resp.dest.traffic_class; attr->ah_attr.sl = resp.dest.sl; attr->ah_attr.src_path_bits = resp.dest.src_path_bits; attr->ah_attr.static_rate = resp.dest.static_rate; attr->ah_attr.is_global = resp.dest.is_global; attr->ah_attr.port_num = resp.dest.port_num; memcpy(attr->alt_ah_attr.grh.dgid.raw, resp.alt_dest.dgid, 16); attr->alt_ah_attr.grh.flow_label = resp.alt_dest.flow_label; attr->alt_ah_attr.dlid = resp.alt_dest.dlid; attr->alt_ah_attr.grh.sgid_index = resp.alt_dest.sgid_index; attr->alt_ah_attr.grh.hop_limit = resp.alt_dest.hop_limit; attr->alt_ah_attr.grh.traffic_class = resp.alt_dest.traffic_class; attr->alt_ah_attr.sl = resp.alt_dest.sl; attr->alt_ah_attr.src_path_bits = resp.alt_dest.src_path_bits; attr->alt_ah_attr.static_rate = resp.alt_dest.static_rate; attr->alt_ah_attr.is_global = resp.alt_dest.is_global; attr->alt_ah_attr.port_num = resp.alt_dest.port_num; init_attr->qp_context = qp->qp_context; init_attr->send_cq = qp->send_cq; init_attr->recv_cq = qp->recv_cq; init_attr->srq = qp->srq; init_attr->qp_type = qp->qp_type; init_attr->cap.max_send_wr = resp.max_send_wr; init_attr->cap.max_recv_wr = resp.max_recv_wr; init_attr->cap.max_send_sge = resp.max_send_sge; init_attr->cap.max_recv_sge = resp.max_recv_sge; init_attr->cap.max_inline_data = resp.max_inline_data; init_attr->sq_sig_all = resp.sq_sig_all; return 0; } static void copy_modify_qp_fields(struct ibv_qp *qp, struct ibv_qp_attr *attr, int attr_mask, struct ibv_modify_qp_common *cmd) { cmd->qp_handle = qp->handle; cmd->attr_mask = attr_mask; if (attr_mask & IBV_QP_STATE) cmd->qp_state = attr->qp_state; if (attr_mask & IBV_QP_CUR_STATE) cmd->cur_qp_state = attr->cur_qp_state; if (attr_mask & IBV_QP_EN_SQD_ASYNC_NOTIFY) cmd->en_sqd_async_notify = attr->en_sqd_async_notify; if (attr_mask & IBV_QP_ACCESS_FLAGS) cmd->qp_access_flags = attr->qp_access_flags; if (attr_mask & IBV_QP_PKEY_INDEX) cmd->pkey_index = attr->pkey_index; if (attr_mask & IBV_QP_PORT) cmd->port_num = attr->port_num; if (attr_mask & IBV_QP_QKEY) cmd->qkey = attr->qkey; if (attr_mask & IBV_QP_AV) { memcpy(cmd->dest.dgid, attr->ah_attr.grh.dgid.raw, 16); cmd->dest.flow_label = attr->ah_attr.grh.flow_label; cmd->dest.dlid = attr->ah_attr.dlid; cmd->dest.reserved = 0; cmd->dest.sgid_index = attr->ah_attr.grh.sgid_index; cmd->dest.hop_limit = attr->ah_attr.grh.hop_limit; cmd->dest.traffic_class = attr->ah_attr.grh.traffic_class; cmd->dest.sl = attr->ah_attr.sl; cmd->dest.src_path_bits = attr->ah_attr.src_path_bits; cmd->dest.static_rate = attr->ah_attr.static_rate; cmd->dest.is_global = attr->ah_attr.is_global; cmd->dest.port_num = attr->ah_attr.port_num; } if (attr_mask & IBV_QP_PATH_MTU) cmd->path_mtu = attr->path_mtu; if (attr_mask & IBV_QP_TIMEOUT) cmd->timeout = attr->timeout; if (attr_mask & IBV_QP_RETRY_CNT) cmd->retry_cnt = attr->retry_cnt; if (attr_mask & IBV_QP_RNR_RETRY) cmd->rnr_retry = attr->rnr_retry; if (attr_mask & IBV_QP_RQ_PSN) cmd->rq_psn = attr->rq_psn; if (attr_mask & IBV_QP_MAX_QP_RD_ATOMIC) cmd->max_rd_atomic = attr->max_rd_atomic; if (attr_mask & IBV_QP_ALT_PATH) { cmd->alt_pkey_index = attr->alt_pkey_index; cmd->alt_port_num = attr->alt_port_num; cmd->alt_timeout = attr->alt_timeout; memcpy(cmd->alt_dest.dgid, attr->alt_ah_attr.grh.dgid.raw, 16); cmd->alt_dest.flow_label = attr->alt_ah_attr.grh.flow_label; cmd->alt_dest.dlid = attr->alt_ah_attr.dlid; cmd->alt_dest.reserved = 0; cmd->alt_dest.sgid_index = attr->alt_ah_attr.grh.sgid_index; cmd->alt_dest.hop_limit = attr->alt_ah_attr.grh.hop_limit; cmd->alt_dest.traffic_class = attr->alt_ah_attr.grh.traffic_class; cmd->alt_dest.sl = attr->alt_ah_attr.sl; cmd->alt_dest.src_path_bits = attr->alt_ah_attr.src_path_bits; cmd->alt_dest.static_rate = attr->alt_ah_attr.static_rate; cmd->alt_dest.is_global = attr->alt_ah_attr.is_global; cmd->alt_dest.port_num = attr->alt_ah_attr.port_num; } if (attr_mask & IBV_QP_MIN_RNR_TIMER) cmd->min_rnr_timer = attr->min_rnr_timer; if (attr_mask & IBV_QP_SQ_PSN) cmd->sq_psn = attr->sq_psn; if (attr_mask & IBV_QP_MAX_DEST_RD_ATOMIC) cmd->max_dest_rd_atomic = attr->max_dest_rd_atomic; if (attr_mask & IBV_QP_PATH_MIG_STATE) cmd->path_mig_state = attr->path_mig_state; if (attr_mask & IBV_QP_DEST_QPN) cmd->dest_qp_num = attr->dest_qp_num; cmd->reserved[0] = cmd->reserved[1] = 0; } int ibv_cmd_modify_qp(struct ibv_qp *qp, struct ibv_qp_attr *attr, int attr_mask, struct ibv_modify_qp *cmd, size_t cmd_size) { /* * Masks over IBV_QP_DEST_QPN are only supported by * ibv_cmd_modify_qp_ex. */ if (attr_mask & ~((IBV_QP_DEST_QPN << 1) - 1)) return EOPNOTSUPP; IBV_INIT_CMD(cmd, cmd_size, MODIFY_QP); copy_modify_qp_fields(qp, attr, attr_mask, &cmd->base); if (write(qp->context->cmd_fd, cmd, cmd_size) != cmd_size) return errno; return 0; } int ibv_cmd_modify_qp_ex(struct ibv_qp *qp, struct ibv_qp_attr *attr, int attr_mask, struct ibv_modify_qp_ex *cmd, size_t cmd_core_size, size_t cmd_size, struct ibv_modify_qp_resp_ex *resp, size_t resp_core_size, size_t resp_size) { if (resp_core_size < offsetof(struct ibv_modify_qp_resp_ex, response_length) + sizeof(resp->response_length)) return EINVAL; IBV_INIT_CMD_RESP_EX_V(cmd, cmd_core_size, cmd_size, MODIFY_QP_EX, resp, resp_core_size, resp_size); copy_modify_qp_fields(qp, attr, attr_mask, &cmd->base); if (attr_mask & IBV_QP_RATE_LIMIT) { if (cmd_size >= offsetof(struct ibv_modify_qp_ex, rate_limit) + sizeof(cmd->rate_limit)) cmd->rate_limit = attr->rate_limit; else return EINVAL; } if (write(qp->context->cmd_fd, cmd, cmd_size) != cmd_size) return errno; (void)VALGRIND_MAKE_MEM_DEFINED(resp, resp_size); return 0; } int ibv_cmd_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr, struct ibv_send_wr **bad_wr) { struct ibv_post_send *cmd; struct ibv_post_send_resp resp; struct ibv_send_wr *i; struct ibv_kern_send_wr *n, *tmp; struct ibv_sge *s; unsigned wr_count = 0; unsigned sge_count = 0; int cmd_size; int ret = 0; for (i = wr; i; i = i->next) { wr_count++; sge_count += i->num_sge; } cmd_size = sizeof *cmd + wr_count * sizeof *n + sge_count * sizeof *s; cmd = alloca(cmd_size); IBV_INIT_CMD_RESP(cmd, cmd_size, POST_SEND, &resp, sizeof resp); cmd->qp_handle = ibqp->handle; cmd->wr_count = wr_count; cmd->sge_count = sge_count; cmd->wqe_size = sizeof *n; n = (struct ibv_kern_send_wr *) ((void *) cmd + sizeof *cmd); s = (struct ibv_sge *) (n + wr_count); tmp = n; for (i = wr; i; i = i->next) { tmp->wr_id = i->wr_id; tmp->num_sge = i->num_sge; tmp->opcode = i->opcode; tmp->send_flags = i->send_flags; tmp->imm_data = i->imm_data; if (ibqp->qp_type == IBV_QPT_UD) { tmp->wr.ud.ah = i->wr.ud.ah->handle; tmp->wr.ud.remote_qpn = i->wr.ud.remote_qpn; tmp->wr.ud.remote_qkey = i->wr.ud.remote_qkey; } else { switch (i->opcode) { case IBV_WR_RDMA_WRITE: case IBV_WR_RDMA_WRITE_WITH_IMM: case IBV_WR_RDMA_READ: tmp->wr.rdma.remote_addr = i->wr.rdma.remote_addr; tmp->wr.rdma.rkey = i->wr.rdma.rkey; break; case IBV_WR_ATOMIC_CMP_AND_SWP: case IBV_WR_ATOMIC_FETCH_AND_ADD: tmp->wr.atomic.remote_addr = i->wr.atomic.remote_addr; tmp->wr.atomic.compare_add = i->wr.atomic.compare_add; tmp->wr.atomic.swap = i->wr.atomic.swap; tmp->wr.atomic.rkey = i->wr.atomic.rkey; break; default: break; } } if (tmp->num_sge) { memcpy(s, i->sg_list, tmp->num_sge * sizeof *s); s += tmp->num_sge; } tmp++; } resp.bad_wr = 0; if (write(ibqp->context->cmd_fd, cmd, cmd_size) != cmd_size) ret = errno; (void) VALGRIND_MAKE_MEM_DEFINED(&resp, sizeof resp); wr_count = resp.bad_wr; if (wr_count) { i = wr; while (--wr_count) i = i->next; *bad_wr = i; } else if (ret) *bad_wr = wr; return ret; } int ibv_cmd_post_recv(struct ibv_qp *ibqp, struct ibv_recv_wr *wr, struct ibv_recv_wr **bad_wr) { struct ibv_post_recv *cmd; struct ibv_post_recv_resp resp; struct ibv_recv_wr *i; struct ibv_kern_recv_wr *n, *tmp; struct ibv_sge *s; unsigned wr_count = 0; unsigned sge_count = 0; int cmd_size; int ret = 0; for (i = wr; i; i = i->next) { wr_count++; sge_count += i->num_sge; } cmd_size = sizeof *cmd + wr_count * sizeof *n + sge_count * sizeof *s; cmd = alloca(cmd_size); IBV_INIT_CMD_RESP(cmd, cmd_size, POST_RECV, &resp, sizeof resp); cmd->qp_handle = ibqp->handle; cmd->wr_count = wr_count; cmd->sge_count = sge_count; cmd->wqe_size = sizeof *n; n = (struct ibv_kern_recv_wr *) ((void *) cmd + sizeof *cmd); s = (struct ibv_sge *) (n + wr_count); tmp = n; for (i = wr; i; i = i->next) { tmp->wr_id = i->wr_id; tmp->num_sge = i->num_sge; if (tmp->num_sge) { memcpy(s, i->sg_list, tmp->num_sge * sizeof *s); s += tmp->num_sge; } tmp++; } resp.bad_wr = 0; if (write(ibqp->context->cmd_fd, cmd, cmd_size) != cmd_size) ret = errno; (void) VALGRIND_MAKE_MEM_DEFINED(&resp, sizeof resp); wr_count = resp.bad_wr; if (wr_count) { i = wr; while (--wr_count) i = i->next; *bad_wr = i; } else if (ret) *bad_wr = wr; return ret; } int ibv_cmd_post_srq_recv(struct ibv_srq *srq, struct ibv_recv_wr *wr, struct ibv_recv_wr **bad_wr) { struct ibv_post_srq_recv *cmd; struct ibv_post_srq_recv_resp resp; struct ibv_recv_wr *i; struct ibv_kern_recv_wr *n, *tmp; struct ibv_sge *s; unsigned wr_count = 0; unsigned sge_count = 0; int cmd_size; int ret = 0; for (i = wr; i; i = i->next) { wr_count++; sge_count += i->num_sge; } cmd_size = sizeof *cmd + wr_count * sizeof *n + sge_count * sizeof *s; cmd = alloca(cmd_size); IBV_INIT_CMD_RESP(cmd, cmd_size, POST_SRQ_RECV, &resp, sizeof resp); cmd->srq_handle = srq->handle; cmd->wr_count = wr_count; cmd->sge_count = sge_count; cmd->wqe_size = sizeof *n; n = (struct ibv_kern_recv_wr *) ((void *) cmd + sizeof *cmd); s = (struct ibv_sge *) (n + wr_count); tmp = n; for (i = wr; i; i = i->next) { tmp->wr_id = i->wr_id; tmp->num_sge = i->num_sge; if (tmp->num_sge) { memcpy(s, i->sg_list, tmp->num_sge * sizeof *s); s += tmp->num_sge; } tmp++; } resp.bad_wr = 0; if (write(srq->context->cmd_fd, cmd, cmd_size) != cmd_size) ret = errno; (void) VALGRIND_MAKE_MEM_DEFINED(&resp, sizeof resp); wr_count = resp.bad_wr; if (wr_count) { i = wr; while (--wr_count) i = i->next; *bad_wr = i; } else if (ret) *bad_wr = wr; return ret; } int ibv_cmd_create_ah(struct ibv_pd *pd, struct ibv_ah *ah, struct ibv_ah_attr *attr, struct ibv_create_ah_resp *resp, size_t resp_size) { struct ibv_create_ah cmd; IBV_INIT_CMD_RESP(&cmd, sizeof cmd, CREATE_AH, resp, resp_size); cmd.user_handle = (uintptr_t) ah; cmd.pd_handle = pd->handle; cmd.attr.dlid = attr->dlid; cmd.attr.sl = attr->sl; cmd.attr.src_path_bits = attr->src_path_bits; cmd.attr.static_rate = attr->static_rate; cmd.attr.is_global = attr->is_global; cmd.attr.port_num = attr->port_num; cmd.attr.grh.flow_label = attr->grh.flow_label; cmd.attr.grh.sgid_index = attr->grh.sgid_index; cmd.attr.grh.hop_limit = attr->grh.hop_limit; cmd.attr.grh.traffic_class = attr->grh.traffic_class; memcpy(cmd.attr.grh.dgid, attr->grh.dgid.raw, 16); if (write(pd->context->cmd_fd, &cmd, sizeof cmd) != sizeof cmd) return errno; (void) VALGRIND_MAKE_MEM_DEFINED(resp, resp_size); ah->handle = resp->handle; ah->context = pd->context; return 0; } int ibv_cmd_destroy_ah(struct ibv_ah *ah) { struct ibv_destroy_ah cmd; IBV_INIT_CMD(&cmd, sizeof cmd, DESTROY_AH); cmd.ah_handle = ah->handle; if (write(ah->context->cmd_fd, &cmd, sizeof cmd) != sizeof cmd) return errno; return 0; } int ibv_cmd_destroy_qp(struct ibv_qp *qp) { struct ibv_destroy_qp cmd; struct ibv_destroy_qp_resp resp; IBV_INIT_CMD_RESP(&cmd, sizeof cmd, DESTROY_QP, &resp, sizeof resp); cmd.qp_handle = qp->handle; cmd.reserved = 0; if (write(qp->context->cmd_fd, &cmd, sizeof cmd) != sizeof cmd) return errno; (void) VALGRIND_MAKE_MEM_DEFINED(&resp, sizeof resp); pthread_mutex_lock(&qp->mutex); while (qp->events_completed != resp.events_reported) pthread_cond_wait(&qp->cond, &qp->mutex); pthread_mutex_unlock(&qp->mutex); pthread_cond_destroy(&qp->cond); pthread_mutex_destroy(&qp->mutex); return 0; } int ibv_cmd_attach_mcast(struct ibv_qp *qp, const union ibv_gid *gid, uint16_t lid) { struct ibv_attach_mcast cmd; IBV_INIT_CMD(&cmd, sizeof cmd, ATTACH_MCAST); memcpy(cmd.gid, gid->raw, sizeof cmd.gid); cmd.qp_handle = qp->handle; cmd.mlid = lid; cmd.reserved = 0; if (write(qp->context->cmd_fd, &cmd, sizeof cmd) != sizeof cmd) return errno; return 0; } int ibv_cmd_detach_mcast(struct ibv_qp *qp, const union ibv_gid *gid, uint16_t lid) { struct ibv_detach_mcast cmd; IBV_INIT_CMD(&cmd, sizeof cmd, DETACH_MCAST); memcpy(cmd.gid, gid->raw, sizeof cmd.gid); cmd.qp_handle = qp->handle; cmd.mlid = lid; cmd.reserved = 0; if (write(qp->context->cmd_fd, &cmd, sizeof cmd) != sizeof cmd) return errno; return 0; } static int buffer_is_zero(char *addr, ssize_t size) { return addr[0] == 0 && !memcmp(addr, addr + 1, size - 1); } static int get_filters_size(struct ibv_flow_spec *ib_spec, struct ibv_kern_spec *kern_spec, int *ib_filter_size, int *kern_filter_size, enum ibv_flow_spec_type type) { void *ib_spec_filter_mask; int curr_kern_filter_size; int min_filter_size; *ib_filter_size = (ib_spec->hdr.size - sizeof(ib_spec->hdr)) / 2; switch (type) { case IBV_FLOW_SPEC_IPV4_EXT: min_filter_size = offsetof(struct ibv_kern_ipv4_ext_filter, flags) + sizeof(kern_spec->ipv4_ext.mask.flags); curr_kern_filter_size = min_filter_size; ib_spec_filter_mask = (void *)&ib_spec->ipv4_ext.val + *ib_filter_size; break; case IBV_FLOW_SPEC_IPV6: min_filter_size = offsetof(struct ibv_kern_ipv6_filter, hop_limit) + sizeof(kern_spec->ipv6.mask.hop_limit); curr_kern_filter_size = min_filter_size; ib_spec_filter_mask = (void *)&ib_spec->ipv6.val + *ib_filter_size; break; case IBV_FLOW_SPEC_VXLAN_TUNNEL: min_filter_size = offsetof(struct ibv_kern_tunnel_filter, tunnel_id) + sizeof(kern_spec->tunnel.mask.tunnel_id); curr_kern_filter_size = min_filter_size; ib_spec_filter_mask = (void *)&ib_spec->tunnel.val + *ib_filter_size; break; default: return EINVAL; } if (*ib_filter_size < min_filter_size) return EINVAL; if (*ib_filter_size > curr_kern_filter_size && !buffer_is_zero(ib_spec_filter_mask + curr_kern_filter_size, *ib_filter_size - curr_kern_filter_size)) return EOPNOTSUPP; *kern_filter_size = min_t(int, curr_kern_filter_size, *ib_filter_size); return 0; } static int ib_spec_to_kern_spec(struct ibv_flow_spec *ib_spec, struct ibv_kern_spec *kern_spec) { int kern_filter_size; int ib_filter_size; int ret; kern_spec->hdr.type = ib_spec->hdr.type; switch (kern_spec->hdr.type) { case IBV_FLOW_SPEC_ETH: case IBV_FLOW_SPEC_ETH | IBV_FLOW_SPEC_INNER: kern_spec->eth.size = sizeof(struct ibv_kern_spec_eth); memcpy(&kern_spec->eth.val, &ib_spec->eth.val, sizeof(struct ibv_flow_eth_filter)); memcpy(&kern_spec->eth.mask, &ib_spec->eth.mask, sizeof(struct ibv_flow_eth_filter)); break; case IBV_FLOW_SPEC_IPV4: case IBV_FLOW_SPEC_IPV4 | IBV_FLOW_SPEC_INNER: kern_spec->ipv4.size = sizeof(struct ibv_kern_spec_ipv4); memcpy(&kern_spec->ipv4.val, &ib_spec->ipv4.val, sizeof(struct ibv_flow_ipv4_filter)); memcpy(&kern_spec->ipv4.mask, &ib_spec->ipv4.mask, sizeof(struct ibv_flow_ipv4_filter)); break; case IBV_FLOW_SPEC_IPV4_EXT: case IBV_FLOW_SPEC_IPV4_EXT | IBV_FLOW_SPEC_INNER: ret = get_filters_size(ib_spec, kern_spec, &ib_filter_size, &kern_filter_size, IBV_FLOW_SPEC_IPV4_EXT); if (ret) return ret; kern_spec->hdr.type = IBV_FLOW_SPEC_IPV4 | (IBV_FLOW_SPEC_INNER & ib_spec->hdr.type); kern_spec->ipv4_ext.size = sizeof(struct ibv_kern_spec_ipv4_ext); memcpy(&kern_spec->ipv4_ext.val, &ib_spec->ipv4_ext.val, kern_filter_size); memcpy(&kern_spec->ipv4_ext.mask, (void *)&ib_spec->ipv4_ext.val + ib_filter_size, kern_filter_size); break; case IBV_FLOW_SPEC_IPV6: case IBV_FLOW_SPEC_IPV6 | IBV_FLOW_SPEC_INNER: ret = get_filters_size(ib_spec, kern_spec, &ib_filter_size, &kern_filter_size, IBV_FLOW_SPEC_IPV6); if (ret) return ret; kern_spec->ipv6.size = sizeof(struct ibv_kern_spec_ipv6); memcpy(&kern_spec->ipv6.val, &ib_spec->ipv6.val, kern_filter_size); memcpy(&kern_spec->ipv6.mask, (void *)&ib_spec->ipv6.val + ib_filter_size, kern_filter_size); break; case IBV_FLOW_SPEC_TCP: case IBV_FLOW_SPEC_UDP: case IBV_FLOW_SPEC_TCP | IBV_FLOW_SPEC_INNER: case IBV_FLOW_SPEC_UDP | IBV_FLOW_SPEC_INNER: kern_spec->tcp_udp.size = sizeof(struct ibv_kern_spec_tcp_udp); memcpy(&kern_spec->tcp_udp.val, &ib_spec->tcp_udp.val, sizeof(struct ibv_flow_ipv4_filter)); memcpy(&kern_spec->tcp_udp.mask, &ib_spec->tcp_udp.mask, sizeof(struct ibv_flow_tcp_udp_filter)); break; case IBV_FLOW_SPEC_VXLAN_TUNNEL: ret = get_filters_size(ib_spec, kern_spec, &ib_filter_size, &kern_filter_size, IBV_FLOW_SPEC_VXLAN_TUNNEL); if (ret) return ret; kern_spec->tunnel.size = sizeof(struct ibv_kern_spec_tunnel); memcpy(&kern_spec->tunnel.val, &ib_spec->tunnel.val, kern_filter_size); memcpy(&kern_spec->tunnel.mask, (void *)&ib_spec->tunnel.val + ib_filter_size, kern_filter_size); break; case IBV_FLOW_SPEC_ACTION_TAG: kern_spec->flow_tag.size = sizeof(struct ibv_kern_spec_action_tag); kern_spec->flow_tag.tag_id = ib_spec->flow_tag.tag_id; break; case IBV_FLOW_SPEC_ACTION_DROP: kern_spec->drop.size = sizeof(struct ibv_kern_spec_action_drop); break; default: return EINVAL; } return 0; } struct ibv_flow *ibv_cmd_create_flow(struct ibv_qp *qp, struct ibv_flow_attr *flow_attr) { struct ibv_create_flow *cmd; struct ibv_create_flow_resp resp; struct ibv_flow *flow_id; size_t cmd_size; size_t written_size; int i, err; void *kern_spec; void *ib_spec; cmd_size = sizeof(*cmd) + (flow_attr->num_of_specs * sizeof(struct ibv_kern_spec)); cmd = alloca(cmd_size); flow_id = malloc(sizeof(*flow_id)); if (!flow_id) return NULL; memset(cmd, 0, cmd_size); cmd->qp_handle = qp->handle; cmd->flow_attr.type = flow_attr->type; cmd->flow_attr.priority = flow_attr->priority; cmd->flow_attr.num_of_specs = flow_attr->num_of_specs; cmd->flow_attr.port = flow_attr->port; cmd->flow_attr.flags = flow_attr->flags; kern_spec = cmd + 1; ib_spec = flow_attr + 1; for (i = 0; i < flow_attr->num_of_specs; i++) { err = ib_spec_to_kern_spec(ib_spec, kern_spec); if (err) { errno = err; goto err; } cmd->flow_attr.size += ((struct ibv_kern_spec *)kern_spec)->hdr.size; kern_spec += ((struct ibv_kern_spec *)kern_spec)->hdr.size; ib_spec += ((struct ibv_flow_spec *)ib_spec)->hdr.size; } written_size = sizeof(*cmd) + cmd->flow_attr.size; IBV_INIT_CMD_RESP_EX_VCMD(cmd, written_size, written_size, CREATE_FLOW, &resp, sizeof(resp)); if (write(qp->context->cmd_fd, cmd, written_size) != written_size) goto err; (void) VALGRIND_MAKE_MEM_DEFINED(&resp, sizeof(resp)); flow_id->context = qp->context; flow_id->handle = resp.flow_handle; return flow_id; err: free(flow_id); return NULL; } int ibv_cmd_destroy_flow(struct ibv_flow *flow_id) { struct ibv_destroy_flow cmd; int ret = 0; memset(&cmd, 0, sizeof(cmd)); IBV_INIT_CMD_EX(&cmd, sizeof(cmd), DESTROY_FLOW); cmd.flow_handle = flow_id->handle; if (write(flow_id->context->cmd_fd, &cmd, sizeof(cmd)) != sizeof(cmd)) ret = errno; free(flow_id); return ret; } int ibv_cmd_create_wq(struct ibv_context *context, struct ibv_wq_init_attr *wq_init_attr, struct ibv_wq *wq, struct ibv_create_wq *cmd, size_t cmd_core_size, size_t cmd_size, struct ibv_create_wq_resp *resp, size_t resp_core_size, size_t resp_size) { int err; if (wq_init_attr->comp_mask >= IBV_WQ_INIT_ATTR_RESERVED) return EINVAL; IBV_INIT_CMD_RESP_EX_V(cmd, cmd_core_size, cmd_size, CREATE_WQ, resp, resp_core_size, resp_size); cmd->user_handle = (uintptr_t)wq; cmd->pd_handle = wq_init_attr->pd->handle; cmd->cq_handle = wq_init_attr->cq->handle; cmd->wq_type = wq_init_attr->wq_type; cmd->max_sge = wq_init_attr->max_sge; cmd->max_wr = wq_init_attr->max_wr; cmd->comp_mask = 0; if (cmd_core_size >= offsetof(struct ibv_create_wq, create_flags) + sizeof(cmd->create_flags)) { if (wq_init_attr->comp_mask & IBV_WQ_INIT_ATTR_FLAGS) { if (wq_init_attr->create_flags & ~(IBV_WQ_FLAGS_RESERVED - 1)) return EOPNOTSUPP; cmd->create_flags = wq_init_attr->create_flags; } } err = write(context->cmd_fd, cmd, cmd_size); if (err != cmd_size) return errno; (void) VALGRIND_MAKE_MEM_DEFINED(resp, resp_size); if (resp->response_length < resp_core_size) return EINVAL; wq->handle = resp->wq_handle; wq_init_attr->max_wr = resp->max_wr; wq_init_attr->max_sge = resp->max_sge; wq->wq_num = resp->wqn; wq->context = context; wq->cq = wq_init_attr->cq; wq->pd = wq_init_attr->pd; wq->wq_type = wq_init_attr->wq_type; return 0; } int ibv_cmd_modify_wq(struct ibv_wq *wq, struct ibv_wq_attr *attr, struct ibv_modify_wq *cmd, size_t cmd_core_size, size_t cmd_size) { if (attr->attr_mask >= IBV_WQ_ATTR_RESERVED) return EINVAL; memset(cmd, 0, cmd_core_size); IBV_INIT_CMD_EX(cmd, cmd_size, MODIFY_WQ); cmd->curr_wq_state = attr->curr_wq_state; cmd->wq_state = attr->wq_state; if (cmd_core_size >= offsetof(struct ibv_modify_wq, flags_mask) + sizeof(cmd->flags_mask)) { if (attr->attr_mask & IBV_WQ_ATTR_FLAGS) { if (attr->flags_mask & ~(IBV_WQ_FLAGS_RESERVED - 1)) return EOPNOTSUPP; cmd->flags = attr->flags; cmd->flags_mask = attr->flags_mask; } } cmd->wq_handle = wq->handle; cmd->attr_mask = attr->attr_mask; if (write(wq->context->cmd_fd, cmd, cmd_size) != cmd_size) return errno; if (attr->attr_mask & IBV_WQ_ATTR_STATE) wq->state = attr->wq_state; return 0; } int ibv_cmd_destroy_wq(struct ibv_wq *wq) { struct ibv_destroy_wq cmd; struct ibv_destroy_wq_resp resp; int ret = 0; memset(&cmd, 0, sizeof(cmd)); memset(&resp, 0, sizeof(resp)); IBV_INIT_CMD_RESP_EX(&cmd, sizeof(cmd), DESTROY_WQ, &resp, sizeof(resp)); cmd.wq_handle = wq->handle; if (write(wq->context->cmd_fd, &cmd, sizeof(cmd)) != sizeof(cmd)) return errno; if (resp.response_length < sizeof(resp)) return EINVAL; pthread_mutex_lock(&wq->mutex); while (wq->events_completed != resp.events_reported) pthread_cond_wait(&wq->cond, &wq->mutex); pthread_mutex_unlock(&wq->mutex); return ret; } int ibv_cmd_create_rwq_ind_table(struct ibv_context *context, struct ibv_rwq_ind_table_init_attr *init_attr, struct ibv_rwq_ind_table *rwq_ind_table, struct ibv_create_rwq_ind_table *cmd, size_t cmd_core_size, size_t cmd_size, struct ibv_create_rwq_ind_table_resp *resp, size_t resp_core_size, size_t resp_size) { int err, i; uint32_t required_tbl_size, alloc_tbl_size; uint32_t *tbl_start; int num_tbl_entries; if (init_attr->comp_mask >= IBV_CREATE_IND_TABLE_RESERVED) return EINVAL; alloc_tbl_size = cmd_core_size - sizeof(*cmd); num_tbl_entries = 1 << init_attr->log_ind_tbl_size; /* Data must be u64 aligned */ required_tbl_size = (num_tbl_entries * sizeof(uint32_t)) < sizeof(uint64_t) ? sizeof(uint64_t) : (num_tbl_entries * sizeof(uint32_t)); if (alloc_tbl_size < required_tbl_size) return EINVAL; tbl_start = (uint32_t *)((uint8_t *)cmd + sizeof(*cmd)); for (i = 0; i < num_tbl_entries; i++) tbl_start[i] = init_attr->ind_tbl[i]->handle; IBV_INIT_CMD_RESP_EX_V(cmd, cmd_core_size, cmd_size, CREATE_RWQ_IND_TBL, resp, resp_core_size, resp_size); cmd->log_ind_tbl_size = init_attr->log_ind_tbl_size; cmd->comp_mask = 0; err = write(context->cmd_fd, cmd, cmd_size); if (err != cmd_size) return errno; (void) VALGRIND_MAKE_MEM_DEFINED(resp, resp_size); if (resp->response_length < resp_core_size) return EINVAL; rwq_ind_table->ind_tbl_handle = resp->ind_tbl_handle; rwq_ind_table->ind_tbl_num = resp->ind_tbl_num; rwq_ind_table->context = context; return 0; } int ibv_cmd_destroy_rwq_ind_table(struct ibv_rwq_ind_table *rwq_ind_table) { struct ibv_destroy_rwq_ind_table cmd; int ret = 0; memset(&cmd, 0, sizeof(cmd)); IBV_INIT_CMD_EX(&cmd, sizeof(cmd), DESTROY_RWQ_IND_TBL); cmd.ind_tbl_handle = rwq_ind_table->ind_tbl_handle; if (write(rwq_ind_table->context->cmd_fd, &cmd, sizeof(cmd)) != sizeof(cmd)) ret = errno; return ret; }