| /linux/drivers/infiniband/hw/irdma/ |
| H A D | uda.h | 39 int irdma_sc_access_ah(struct irdma_sc_cqp *cqp, struct irdma_ah_info *info, 41 int irdma_access_mcast_grp(struct irdma_sc_cqp *cqp, 50 static inline int irdma_sc_create_ah(struct irdma_sc_cqp *cqp, in irdma_sc_create_ah() argument 53 return irdma_sc_access_ah(cqp, info, IRDMA_CQP_OP_CREATE_ADDR_HANDLE, in irdma_sc_create_ah() 57 static inline int irdma_sc_destroy_ah(struct irdma_sc_cqp *cqp, in irdma_sc_destroy_ah() argument 60 return irdma_sc_access_ah(cqp, info, IRDMA_CQP_OP_DESTROY_ADDR_HANDLE, in irdma_sc_destroy_ah() 64 static inline int irdma_sc_create_mcast_grp(struct irdma_sc_cqp *cqp, in irdma_sc_create_mcast_grp() argument 68 return irdma_access_mcast_grp(cqp, info, IRDMA_CQP_OP_CREATE_MCAST_GRP, in irdma_sc_create_mcast_grp() 72 static inline int irdma_sc_modify_mcast_grp(struct irdma_sc_cqp *cqp, in irdma_sc_modify_mcast_grp() argument 76 return irdma_access_mcast_grp(cqp, info, IRDMA_CQP_OP_MODIFY_MCAST_GRP, in irdma_sc_modify_mcast_grp() [all …]
|
| H A D | utils.c | 427 struct irdma_cqp_request *irdma_alloc_and_get_cqp_request(struct irdma_cqp *cqp, in irdma_alloc_and_get_cqp_request() argument 433 spin_lock_irqsave(&cqp->req_lock, flags); in irdma_alloc_and_get_cqp_request() 434 if (!list_empty(&cqp->cqp_avail_reqs)) { in irdma_alloc_and_get_cqp_request() 435 cqp_request = list_first_entry(&cqp->cqp_avail_reqs, in irdma_alloc_and_get_cqp_request() 439 spin_unlock_irqrestore(&cqp->req_lock, flags); in irdma_alloc_and_get_cqp_request() 449 ibdev_dbg(to_ibdev(cqp->sc_cqp.dev), "ERR: CQP Request Fail: No Memory"); in irdma_alloc_and_get_cqp_request() 475 void irdma_free_cqp_request(struct irdma_cqp *cqp, in irdma_free_cqp_request() argument 488 spin_lock_irqsave(&cqp->req_lock, flags); in irdma_free_cqp_request() 489 list_add_tail(&cqp_request->list, &cqp->cqp_avail_reqs); in irdma_free_cqp_request() 490 spin_unlock_irqrestore(&cqp->req_lock, flags); in irdma_free_cqp_request() [all …]
|
| H A D | uda.c | 20 int irdma_sc_access_ah(struct irdma_sc_cqp *cqp, struct irdma_ah_info *info, in irdma_sc_access_ah() argument 26 wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch); in irdma_sc_access_ah() 69 FIELD_PREP(IRDMA_UDA_CQPSQ_MAV_WQEVALID, cqp->polarity) | in irdma_sc_access_ah() 78 irdma_sc_cqp_post_sq(cqp); in irdma_sc_access_ah() 115 int irdma_access_mcast_grp(struct irdma_sc_cqp *cqp, in irdma_access_mcast_grp() argument 122 ibdev_dbg(to_ibdev(cqp->dev), "WQE: mg_id out of range\n"); in irdma_access_mcast_grp() 126 wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch); in irdma_access_mcast_grp() 128 ibdev_dbg(to_ibdev(cqp->dev), "WQE: ring full\n"); in irdma_access_mcast_grp() 157 FIELD_PREP(IRDMA_UDA_CQPSQ_MG_WQEVALID, cqp->polarity) | in irdma_access_mcast_grp() 168 irdma_sc_cqp_post_sq(cqp); in irdma_access_mcast_grp()
|
| H A D | verbs.c | 251 cqp_request = irdma_alloc_and_get_cqp_request(&iwdev->rf->cqp, true); in irdma_alloc_push_page() 263 cqp_info->in.u.manage_push_page.cqp = &iwdev->rf->cqp.sc_cqp; in irdma_alloc_push_page() 273 irdma_put_cqp_request(&iwdev->rf->cqp, cqp_request); in irdma_alloc_push_page() 772 cqp_request = irdma_alloc_and_get_cqp_request(&rf->cqp, true); in irdma_cqp_create_qp_cmd() 787 irdma_put_cqp_request(&rf->cqp, cqp_request); in irdma_cqp_create_qp_cmd() 2122 cqp_request = irdma_alloc_and_get_cqp_request(&rf->cqp, true); in irdma_resize_cq() 2140 irdma_put_cqp_request(&rf->cqp, cqp_request); in irdma_resize_cq() 2223 cqp_request = irdma_alloc_and_get_cqp_request(&rf->cqp, true); in irdma_modify_srq() 2239 irdma_put_cqp_request(&rf->cqp, cqp_request); in irdma_modify_srq() 2400 cqp_request = irdma_alloc_and_get_cqp_request(&rf->cqp, true); in irdma_create_srq() [all …]
|
| /linux/drivers/scsi/lpfc/ |
| H A D | lpfc_sli4.h | 1093 int lpfc_cq_create_set(struct lpfc_hba *phba, struct lpfc_queue **cqp, 1103 struct lpfc_queue **drqp, struct lpfc_queue **cqp,
|
| H A D | lpfc_sli.c | 16468 lpfc_cq_create_set(struct lpfc_hba *phba, struct lpfc_queue **cqp, in lpfc_cq_create_set() argument 16485 if (!cqp || !hdwq || !numcq) in lpfc_cq_create_set() 16493 length += ((numcq * cqp[0]->page_count) * in lpfc_cq_create_set() 16511 cq = cqp[idx]; in lpfc_cq_create_set() 16685 cq = cqp[idx]; in lpfc_cq_create_set() 17485 struct lpfc_queue **drqp, struct lpfc_queue **cqp, in lpfc_mrq_create() argument 17500 if (!hrqp || !drqp || !cqp || !numrq) in lpfc_mrq_create() 17536 cq = cqp[idx]; in lpfc_mrq_create()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| H A D | en_main.c | 2167 &cparam->xdp_sq.cqp, ccp, &xdpsq->cq); in mlx5e_open_xdpredirect_sq() 2385 err = mlx5e_open_cq(c->mdev, params->tx_cq_moderation, &cparam->txq_sq.cqp, in mlx5e_open_tx_cqs() 2598 err = mlx5e_open_cq(c->mdev, icocq_moder, &cparam->async_icosq.cqp, ccp, in mlx5e_open_async_icosq() 2636 err = mlx5e_open_cq(c->mdev, icocq_moder, &cparam->icosq.cqp, &ccp, in mlx5e_open_queues() 2653 err = mlx5e_open_cq(c->mdev, params->rx_cq_moderation, &cparam->rq.cqp, &ccp, in mlx5e_open_queues() 2658 err = c->xdp ? mlx5e_open_cq(c->mdev, params->tx_cq_moderation, &cparam->xdp_sq.cqp, in mlx5e_open_queues()
|