| /linux/drivers/infiniband/ulp/ipoib/ |
| H A D | ipoib_verbs.c | 42 struct ib_qp_attr *qp_attr = NULL; in ipoib_mcast_attach() local 55 qp_attr = kmalloc_obj(*qp_attr); in ipoib_mcast_attach() 56 if (!qp_attr) in ipoib_mcast_attach() 60 qp_attr->qkey = qkey; in ipoib_mcast_attach() 61 ret = ib_modify_qp(priv->qp, qp_attr, IB_QP_QKEY); in ipoib_mcast_attach() 74 kfree(qp_attr); in ipoib_mcast_attach() 93 struct ib_qp_attr qp_attr; in ipoib_init_qp() local 99 qp_attr.qp_state = IB_QPS_INIT; in ipoib_init_qp() 100 qp_attr.qkey = 0; in ipoib_init_qp() 101 qp_attr.port_num = priv->port; in ipoib_init_qp() [all …]
|
| H A D | ipoib_cm.c | 278 struct ib_qp_attr qp_attr; in ipoib_cm_modify_rx_qp() local 281 qp_attr.qp_state = IB_QPS_INIT; in ipoib_cm_modify_rx_qp() 282 ret = ib_cm_init_qp_attr(cm_id, &qp_attr, &qp_attr_mask); in ipoib_cm_modify_rx_qp() 287 ret = ib_modify_qp(qp, &qp_attr, qp_attr_mask); in ipoib_cm_modify_rx_qp() 292 qp_attr.qp_state = IB_QPS_RTR; in ipoib_cm_modify_rx_qp() 293 ret = ib_cm_init_qp_attr(cm_id, &qp_attr, &qp_attr_mask); in ipoib_cm_modify_rx_qp() 298 qp_attr.rq_psn = psn; in ipoib_cm_modify_rx_qp() 299 ret = ib_modify_qp(qp, &qp_attr, qp_attr_mask); in ipoib_cm_modify_rx_qp() 313 qp_attr.qp_state = IB_QPS_RTS; in ipoib_cm_modify_rx_qp() 314 ret = ib_cm_init_qp_attr(cm_id, &qp_attr, &qp_attr_mask); in ipoib_cm_modify_rx_qp() [all …]
|
| H A D | ipoib_ib.c | 352 struct ib_qp_attr qp_attr; in ipoib_qp_state_validate_work() local 356 ret = ib_query_qp(priv->qp, &qp_attr, IB_QP_STATE, &query_init_attr); in ipoib_qp_state_validate_work() 363 __func__, priv->qp->qp_num, qp_attr.qp_state); in ipoib_qp_state_validate_work() 366 if (qp_attr.qp_state == IB_QPS_SQE) { in ipoib_qp_state_validate_work() 367 qp_attr.qp_state = IB_QPS_RTS; in ipoib_qp_state_validate_work() 369 ret = ib_modify_qp(priv->qp, &qp_attr, IB_QP_STATE); in ipoib_qp_state_validate_work() 379 priv->qp->qp_num, qp_attr.qp_state); in ipoib_qp_state_validate_work() 764 struct ib_qp_attr qp_attr; in check_qp_movement_and_print() local 768 ret = ib_query_qp(qp, &qp_attr, IB_QP_STATE, &query_init_attr); in check_qp_movement_and_print() 774 if (new_state == IB_QPS_ERR && qp_attr.qp_state == IB_QPS_RESET) in check_qp_movement_and_print() [all …]
|
| /linux/net/smc/ |
| H A D | smc_ib.c | 49 struct ib_qp_attr qp_attr; in smc_ib_modify_qp_init() local 51 memset(&qp_attr, 0, sizeof(qp_attr)); in smc_ib_modify_qp_init() 52 qp_attr.qp_state = IB_QPS_INIT; in smc_ib_modify_qp_init() 53 qp_attr.pkey_index = 0; in smc_ib_modify_qp_init() 54 qp_attr.port_num = lnk->ibport; in smc_ib_modify_qp_init() 55 qp_attr.qp_access_flags = IB_ACCESS_LOCAL_WRITE in smc_ib_modify_qp_init() 57 return ib_modify_qp(lnk->roce_qp, &qp_attr, in smc_ib_modify_qp_init() 67 struct ib_qp_attr qp_attr; in smc_ib_modify_qp_rtr() local 70 memset(&qp_attr, 0, sizeof(qp_attr)); in smc_ib_modify_qp_rtr() 71 qp_attr.qp_state = IB_QPS_RTR; in smc_ib_modify_qp_rtr() [all …]
|
| H A D | smc_wr.c | 525 struct ib_qp_attr *attr = &lnk->qp_attr; in smc_wr_remember_qp_attr() 551 lnk->qp_attr.cap.max_send_wr); in smc_wr_remember_qp_attr() 553 lnk->qp_attr.cap.max_recv_wr); in smc_wr_remember_qp_attr() 558 bool send_inline = (lnk->qp_attr.cap.max_inline_data > SMC_WR_TX_SIZE); in smc_wr_init_sge()
|
| /linux/net/sunrpc/xprtrdma/ |
| H A D | svc_rdma_transport.c | 415 struct ib_qp_init_attr qp_attr; in svc_rdma_accept() 498 memset(&qp_attr, 0, sizeof qp_attr); in svc_rdma_accept() 499 qp_attr.event_handler = qp_event_handler; in svc_rdma_accept() 500 qp_attr.qp_context = &newxprt->sc_xprt; in svc_rdma_accept() 501 qp_attr.port_num = newxprt->sc_port_num; in svc_rdma_accept() 502 qp_attr.cap.max_rdma_ctxs = ctxts; in svc_rdma_accept() 503 qp_attr.cap.max_send_wr = newxprt->sc_sq_depth - ctxts; in svc_rdma_accept() 504 qp_attr.cap.max_recv_wr = rq_depth; in svc_rdma_accept() 505 qp_attr in svc_rdma_accept() 414 struct ib_qp_init_attr qp_attr; svc_rdma_accept() local [all...] |
| /linux/net/9p/ |
| H A D | trans_rdma.c | 530 struct ib_qp_init_attr qp_attr; in rdma_create_trans() local 596 memset(&qp_attr, 0, sizeof qp_attr); in rdma_create_trans() 597 qp_attr.event_handler = qp_event_handler; in rdma_create_trans() 598 qp_attr.qp_context = client; in rdma_create_trans() 599 qp_attr.cap.max_send_wr = opts.sq_depth; in rdma_create_trans() 600 qp_attr.cap.max_recv_wr = opts.rq_depth; in rdma_create_trans() 601 qp_attr.cap.max_send_sge = P9_RDMA_SEND_SGE; in rdma_create_trans() 602 qp_attr.cap.max_recv_sge = P9_RDMA_RECV_SGE; in rdma_create_trans() 603 qp_attr.sq_sig_type = IB_SIGNAL_REQ_WR; in rdma_create_trans() 604 qp_attr.qp_type = IB_QPT_RC; in rdma_create_trans() [all …]
|
| /linux/drivers/infiniband/hw/mthca/ |
| H A D | mthca_qp.c | 430 int mthca_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, int qp_attr_mask, in mthca_query_qp() argument 444 qp_attr->qp_state = IB_QPS_RESET; in mthca_query_qp() 465 qp_attr->qp_state = qp->state; in mthca_query_qp() 466 qp_attr->path_mtu = context->mtu_msgmax >> 5; in mthca_query_qp() 467 qp_attr->path_mig_state = in mthca_query_qp() 469 qp_attr->qkey = be32_to_cpu(context->qkey); in mthca_query_qp() 470 qp_attr->rq_psn = be32_to_cpu(context->rnr_nextrecvpsn) & 0xffffff; in mthca_query_qp() 471 qp_attr->sq_psn = be32_to_cpu(context->next_send_psn) & 0xffffff; in mthca_query_qp() 472 qp_attr->dest_qp_num = be32_to_cpu(context->remote_qpn) & 0xffffff; in mthca_query_qp() 473 qp_attr->qp_access_flags = in mthca_query_qp() [all …]
|
| H A D | mthca_dev.h | 519 int mthca_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, int qp_attr_mask,
|
| /linux/drivers/infiniband/core/ |
| H A D | iwcm.c | 269 struct ib_qp_attr qp_attr; in iwcm_modify_qp_err() local 274 qp_attr.qp_state = IB_QPS_ERR; in iwcm_modify_qp_err() 275 return ib_modify_qp(qp, &qp_attr, IB_QP_STATE); in iwcm_modify_qp_err() 284 struct ib_qp_attr qp_attr; in iwcm_modify_qp_sqd() local 287 qp_attr.qp_state = IB_QPS_SQD; in iwcm_modify_qp_sqd() 288 return ib_modify_qp(qp, &qp_attr, IB_QP_STATE); in iwcm_modify_qp_sqd() 1091 struct ib_qp_attr *qp_attr, in iwcm_init_qp_init_attr() argument 1104 qp_attr->qp_access_flags = IB_ACCESS_REMOTE_WRITE| in iwcm_init_qp_init_attr() 1117 struct ib_qp_attr *qp_attr, in iwcm_init_qp_rts_attr() argument 1141 struct ib_qp_attr *qp_attr, in iw_cm_init_qp_attr() argument [all …]
|
| H A D | security.c | 332 const struct ib_qp_attr *qp_attr, in get_new_pps() argument 343 new_pps->main.port_num = qp_attr->port_num; in get_new_pps() 348 new_pps->main.pkey_index = qp_attr->pkey_index; in get_new_pps() 358 new_pps->alt.port_num = qp_attr->alt_port_num; in get_new_pps() 359 new_pps->alt.pkey_index = qp_attr->alt_pkey_index; in get_new_pps() 572 struct ib_qp_attr *qp_attr, in ib_security_modify_qp() argument 587 rdma_protocol_ib(real_qp->device, qp_attr->port_num) && in ib_security_modify_qp() 601 qp_attr, in ib_security_modify_qp() 626 qp_attr, in ib_security_modify_qp()
|
| H A D | cma.c | 1086 struct ib_qp_attr qp_attr; in cma_init_ud_qp() local 1089 qp_attr.qp_state = IB_QPS_INIT; in cma_init_ud_qp() 1090 ret = rdma_init_qp_attr(&id_priv->id, &qp_attr, &qp_attr_mask); in cma_init_ud_qp() 1094 ret = ib_modify_qp(qp, &qp_attr, qp_attr_mask); in cma_init_ud_qp() 1098 qp_attr.qp_state = IB_QPS_RTR; in cma_init_ud_qp() 1099 ret = ib_modify_qp(qp, &qp_attr, IB_QP_STATE); in cma_init_ud_qp() 1103 qp_attr.qp_state = IB_QPS_RTS; in cma_init_ud_qp() 1104 qp_attr.sq_psn = 0; in cma_init_ud_qp() 1105 ret = ib_modify_qp(qp, &qp_attr, IB_QP_STATE | IB_QP_SQ_PSN); in cma_init_ud_qp() 1112 struct ib_qp_attr qp_attr; in cma_init_conn_qp() local [all …]
|
| H A D | cm.c | 4079 struct ib_qp_attr *qp_attr, in cm_init_qp_init_attr() argument 4098 qp_attr->qp_access_flags = IB_ACCESS_REMOTE_WRITE; in cm_init_qp_init_attr() 4107 qp_attr->qp_access_flags |= IB_ACCESS_REMOTE_READ | in cm_init_qp_init_attr() 4111 qp_attr->pkey_index = cm_id_priv->av.pkey_index; in cm_init_qp_init_attr() 4113 qp_attr->port_num = cm_id_priv->av.port->port_num; in cm_init_qp_init_attr() 4126 struct ib_qp_attr *qp_attr, in cm_init_qp_rtr_attr() argument 4143 qp_attr->ah_attr = cm_id_priv->av.ah_attr; in cm_init_qp_rtr_attr() 4144 if ((qp_attr->ah_attr.type == RDMA_AH_ATTR_TYPE_IB) && in cm_init_qp_rtr_attr() 4147 qp_attr->ah_attr.ib.dlid = cm_id_priv->av.dlid_datapath; in cm_init_qp_rtr_attr() 4148 qp_attr->path_mtu = cm_id_priv->path_mtu; in cm_init_qp_rtr_attr() [all …]
|
| H A D | core_priv.h | 230 struct ib_qp_attr *qp_attr, 257 struct ib_qp_attr *qp_attr, in ib_security_modify_qp() argument 262 qp_attr, in ib_security_modify_qp()
|
| H A D | ucma.c | 1287 struct ib_qp_attr qp_attr; in ucma_init_qp_attr() local 1304 memset(&qp_attr, 0, sizeof qp_attr); in ucma_init_qp_attr() 1305 qp_attr.qp_state = cmd.qp_state; in ucma_init_qp_attr() 1307 ret = rdma_init_qp_attr(ctx->cm_id, &qp_attr, &resp.qp_attr_mask); in ucma_init_qp_attr() 1312 ib_copy_qp_attr_to_user(ctx->cm_id->device, &resp, &qp_attr); in ucma_init_qp_attr()
|
| /linux/drivers/infiniband/hw/efa/ |
| H A D | efa_verbs.c | 348 int efa_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, in efa_query_qp() argument 369 memset(qp_attr, 0, sizeof(*qp_attr)); in efa_query_qp() 377 qp_attr->qp_state = result.qp_state; in efa_query_qp() 378 qp_attr->qkey = result.qkey; in efa_query_qp() 379 qp_attr->sq_psn = result.sq_psn; in efa_query_qp() 380 qp_attr->sq_draining = result.sq_draining; in efa_query_qp() 381 qp_attr->port_num = 1; in efa_query_qp() 382 qp_attr->rnr_retry = result.rnr_retry; in efa_query_qp() 384 qp_attr->cap.max_send_wr = qp->max_send_wr; in efa_query_qp() 385 qp_attr->cap.max_recv_wr = qp->max_recv_wr; in efa_query_qp() [all …]
|
| /linux/drivers/nvme/target/ |
| H A D | rdma.c | 1264 struct ib_qp_init_attr qp_attr = { }; in nvmet_rdma_create_queue_ib() local 1282 qp_attr.qp_context = queue; in nvmet_rdma_create_queue_ib() 1283 qp_attr.event_handler = nvmet_rdma_qp_event; in nvmet_rdma_create_queue_ib() 1284 qp_attr.send_cq = queue->cq; in nvmet_rdma_create_queue_ib() 1285 qp_attr.recv_cq = queue->cq; in nvmet_rdma_create_queue_ib() 1286 qp_attr.sq_sig_type = IB_SIGNAL_REQ_WR; in nvmet_rdma_create_queue_ib() 1287 qp_attr.qp_type = IB_QPT_RC; in nvmet_rdma_create_queue_ib() 1289 qp_attr.cap.max_send_wr = queue->send_queue_size + 1; in nvmet_rdma_create_queue_ib() 1292 qp_attr.cap.max_rdma_ctxs = queue->send_queue_size * factor; in nvmet_rdma_create_queue_ib() 1293 qp_attr.cap.max_send_sge = max(ndev->device->attrs.max_sge_rd, in nvmet_rdma_create_queue_ib() [all …]
|
| /linux/fs/smb/smbdirect/ |
| H A D | connection.c | 245 struct ib_qp_init_attr qp_attr; in smbdirect_connection_create_qp() local 280 memset(&qp_attr, 0, sizeof(qp_attr)); in smbdirect_connection_create_qp() 281 qp_attr.cap = qp_cap; in smbdirect_connection_create_qp() 282 qp_attr.port_num = sc->rdma.cm_id->port_num; in smbdirect_connection_create_qp() 283 rdma_send_wr = smbdirect_rdma_rw_send_wrs(sc->ib.dev, &qp_attr); in smbdirect_connection_create_qp() 375 memset(&qp_attr, 0, sizeof(qp_attr)); in smbdirect_connection_create_qp() 376 qp_attr.event_handler = smbdirect_connection_qp_event_handler; in smbdirect_connection_create_qp() 377 qp_attr.qp_context = sc; in smbdirect_connection_create_qp() 378 qp_attr.cap = qp_cap; in smbdirect_connection_create_qp() 379 qp_attr.sq_sig_type = IB_SIGNAL_REQ_WR; in smbdirect_connection_create_qp() [all …]
|
| /linux/drivers/infiniband/hw/hns/ |
| H A D | hns_roce_hw_v2.c | 5787 static int hns_roce_v2_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, in hns_roce_v2_query_qp() argument 5799 memset(qp_attr, 0, sizeof(*qp_attr)); in hns_roce_v2_query_qp() 5805 qp_attr->qp_state = IB_QPS_RESET; in hns_roce_v2_query_qp() 5827 qp_attr->qp_state = (enum ib_qp_state)hr_qp->state; in hns_roce_v2_query_qp() 5828 qp_attr->path_mtu = (enum ib_mtu)hr_reg_read(&context, QPC_MTU); in hns_roce_v2_query_qp() 5829 qp_attr->path_mig_state = IB_MIG_ARMED; in hns_roce_v2_query_qp() 5830 qp_attr->ah_attr.type = RDMA_AH_ATTR_TYPE_ROCE; in hns_roce_v2_query_qp() 5832 qp_attr->qkey = le32_to_cpu(context.qkey_xrcd); in hns_roce_v2_query_qp() 5834 qp_attr->rq_psn = hr_reg_read(&context, QPC_RX_REQ_EPSN); in hns_roce_v2_query_qp() 5835 qp_attr->sq_psn = (u32)hr_reg_read(&context, QPC_SQ_CUR_PSN); in hns_roce_v2_query_qp() [all …]
|
| /linux/drivers/infiniband/ulp/srp/ |
| H A D | ib_srp.c | 2304 static uint32_t srp_compute_rq_tmo(struct ib_qp_attr *qp_attr, int attr_mask) in srp_compute_rq_tmo() argument 2323 T_tr_ns = 4096 * (1ULL << qp_attr->timeout); in srp_compute_rq_tmo() 2324 max_compl_time_ms = qp_attr->retry_cnt * 4 * T_tr_ns; in srp_compute_rq_tmo() 2336 struct ib_qp_attr *qp_attr = NULL; in srp_cm_rep_handler() local 2389 qp_attr = kmalloc_obj(*qp_attr); in srp_cm_rep_handler() 2390 if (!qp_attr) in srp_cm_rep_handler() 2393 qp_attr->qp_state = IB_QPS_RTR; in srp_cm_rep_handler() 2394 ret = ib_cm_init_qp_attr(cm_id, qp_attr, &attr_mask); in srp_cm_rep_handler() 2398 ret = ib_modify_qp(ch->qp, qp_attr, attr_mask); in srp_cm_rep_handler() 2402 qp_attr->qp_state = IB_QPS_RTS; in srp_cm_rep_handler() [all …]
|
| /linux/drivers/infiniband/hw/mlx5/ |
| H A D | gsi.c | 347 int mlx5_ib_gsi_query_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr, in mlx5_ib_gsi_query_qp() argument 355 ret = ib_query_qp(gsi->rx_qp, qp_attr, qp_attr_mask, qp_init_attr); in mlx5_ib_gsi_query_qp()
|
| /linux/include/rdma/ |
| H A D | rdma_cm.h | 274 int rdma_init_qp_attr(struct rdma_cm_id *id, struct ib_qp_attr *qp_attr,
|
| H A D | ib_cm.h | 507 struct ib_qp_attr *qp_attr,
|
| /linux/drivers/net/ethernet/ibm/ehea/ |
| H A D | ehea_qmr.c | 493 struct ehea_qp_init_attr *qp_attr = &qp->init_attr; in ehea_destroy_qp_res() local 504 if (qp_attr->rq_count > 1) in ehea_destroy_qp_res() 506 if (qp_attr->rq_count > 2) in ehea_destroy_qp_res()
|