| /linux/drivers/infiniband/hw/mthca/ |
| H A D | mthca_cq.c | 181 return cqe_sw(get_cqe(cq, cq->cons_index & cq->ibcq.cqe)); in next_cqe_sw() 230 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in mthca_cq_completion() 254 event.element.cq = &cq->ibcq; in mthca_cq_event() 255 if (cq->ibcq.event_handler) in mthca_cq_event() 256 cq->ibcq.event_handler(&event, cq->ibcq.cq_context); in mthca_cq_event() 290 cqe_sw(get_cqe(cq, prod_index & cq->ibcq.cqe)); in mthca_cq_clean() 292 if (prod_index == cq->cons_index + cq->ibcq.cqe) in mthca_cq_clean() 304 cqe = get_cqe(cq, prod_index & cq->ibcq.cqe); in mthca_cq_clean() 310 memcpy(get_cqe(cq, (prod_index + nfreed) & cq->ibcq.cqe), in mthca_cq_clean() 316 set_cqe_hw(get_cqe(cq, (cq->cons_index + i) & cq->ibcq.cqe)); in mthca_cq_clean() [all …]
|
| H A D | mthca_provider.h | 184 struct ib_cq ibcq; member 301 static inline struct mthca_cq *to_mcq(struct ib_cq *ibcq) in to_mcq() argument 303 return container_of(ibcq, struct mthca_cq, ibcq); in to_mcq()
|
| H A D | mthca_provider.c | 578 static int mthca_create_cq(struct ib_cq *ibcq, in mthca_create_cq() argument 583 struct ib_device *ibdev = ibcq->device; in mthca_create_cq() 616 cq = to_mcq(ibcq); in mthca_create_cq() 701 static int mthca_resize_cq(struct ib_cq *ibcq, unsigned int entries, in mthca_resize_cq() argument 704 struct mthca_dev *dev = to_mdev(ibcq->device); in mthca_resize_cq() 705 struct mthca_cq *cq = to_mcq(ibcq); in mthca_resize_cq() 716 if (entries == ibcq->cqe + 1) { in mthca_resize_cq() 755 tcqe = cq->ibcq.cqe; in mthca_resize_cq() 757 cq->ibcq.cqe = cq->resize_buf->cqe; in mthca_resize_cq() 769 ibcq->cqe = entries - 1; in mthca_resize_cq() [all …]
|
| H A D | mthca_dev.h | 483 int mthca_poll_cq(struct ib_cq *ibcq, int num_entries,
|
| /linux/include/rdma/ |
| H A D | rdmavt_cq.h | 47 struct ib_cq ibcq; member 60 static inline struct rvt_cq *ibcq_to_rvtcq(struct ib_cq *ibcq) in ibcq_to_rvtcq() argument 62 return container_of(ibcq, struct rvt_cq, ibcq); in ibcq_to_rvtcq()
|
| /linux/drivers/infiniband/hw/cxgb4/ |
| H A D | ev.c | 105 event.device = chp->ibcq.device; in post_qp_event() 107 event.element.cq = &chp->ibcq; in post_qp_event() 115 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in post_qp_event() 234 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in c4iw_ev_handler()
|
| H A D | cq.c | 949 int c4iw_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc) in c4iw_poll_cq() argument 956 chp = to_c4iw_cq(ibcq); in c4iw_poll_cq() 997 int c4iw_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr, in c4iw_create_cq() argument 1001 struct ib_device *ibdev = ibcq->device; in c4iw_create_cq() 1004 struct c4iw_dev *rhp = to_c4iw_dev(ibcq->device); in c4iw_create_cq() 1005 struct c4iw_cq *chp = to_c4iw_cq(ibcq); in c4iw_create_cq() 1087 chp->ibcq.cqe = entries - 2; in c4iw_create_cq() 1167 int c4iw_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags flags) in c4iw_arm_cq() argument 1173 chp = to_c4iw_cq(ibcq); in c4iw_arm_cq()
|
| H A D | qp.c | 1008 (*schp->ibcq.comp_handler)(&schp->ibcq, in complete_sq_drain_wr() 1009 schp->ibcq.cq_context); in complete_sq_drain_wr() 1059 (*rchp->ibcq.comp_handler)(&rchp->ibcq, in complete_rq_drain_wr() 1060 rchp->ibcq.cq_context); in complete_rq_drain_wr() 1645 (*rchp->ibcq.comp_handler)(&rchp->ibcq, in __flush_qp() 1646 rchp->ibcq.cq_context); in __flush_qp() 1652 (*rchp->ibcq.comp_handler)(&rchp->ibcq, in __flush_qp() 1653 rchp->ibcq.cq_context); in __flush_qp() 1658 (*schp->ibcq.comp_handler)(&schp->ibcq, in __flush_qp() 1659 schp->ibcq.cq_context); in __flush_qp() [all …]
|
| H A D | restrack.c | 372 int c4iw_fill_res_cq_entry(struct sk_buff *msg, struct ib_cq *ibcq) in c4iw_fill_res_cq_entry() argument 374 struct c4iw_cq *chp = to_c4iw_cq(ibcq); in c4iw_fill_res_cq_entry() 382 if (ibcq->uobject) in c4iw_fill_res_cq_entry()
|
| H A D | provider.c | 504 INIT_RDMA_OBJ_SIZE(ib_cq, c4iw_cq, ibcq),
|
| /linux/drivers/infiniband/hw/qedr/ |
| H A D | qedr_roce_cm.c | 82 cq->ibcq.comp_handler ? "Yes" : "No"); in qedr_ll2_complete_tx_packet() 92 if (cq->ibcq.comp_handler) in qedr_ll2_complete_tx_packet() 93 (*cq->ibcq.comp_handler) (&cq->ibcq, cq->ibcq.cq_context); in qedr_ll2_complete_tx_packet() 121 if (cq->ibcq.comp_handler) in qedr_ll2_complete_rx_packet() 122 (*cq->ibcq.comp_handler) (&cq->ibcq, cq->ibcq.cq_context); in qedr_ll2_complete_rx_packet() 668 int qedr_gsi_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc) in qedr_gsi_poll_cq() argument 670 struct qedr_dev *dev = get_qedr_dev(ibcq->device); in qedr_gsi_poll_cq() 671 struct qedr_cq *cq = get_qedr_cq(ibcq); in qedr_gsi_poll_cq()
|
| H A D | main.c | 233 INIT_RDMA_OBJ_SIZE(ib_cq, qedr_cq, ibcq), 472 if (!cq->destroyed && cq->ibcq.comp_handler) in qedr_irq_handler() 473 (*cq->ibcq.comp_handler) in qedr_irq_handler() 474 (&cq->ibcq, cq->ibcq.cq_context); in qedr_irq_handler() 653 struct ib_cq *ibcq; in qedr_affiliated_event() local 724 ibcq = &cq->ibcq; in qedr_affiliated_event() 725 if (ibcq->event_handler) { in qedr_affiliated_event() 726 event.device = ibcq->device; in qedr_affiliated_event() 727 event.element.cq = ibcq; in qedr_affiliated_event() 728 ibcq->event_handler(&event, ibcq->cq_context); in qedr_affiliated_event()
|
| H A D | qedr_roce_cm.h | 48 int qedr_gsi_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
|
| /linux/drivers/infiniband/hw/ionic/ |
| H A D | ionic_admin.c | 492 static void ionic_rdma_admincq_comp(struct ib_cq *ibcq, void *cq_context) in ionic_rdma_admincq_comp() argument 527 vcq->ibcq.device = &dev->ibdev; in ionic_create_rdma_admincq() 528 vcq->ibcq.comp_handler = ionic_rdma_admincq_comp; in ionic_create_rdma_admincq() 529 vcq->ibcq.event_handler = ionic_rdma_admincq_event; in ionic_create_rdma_admincq() 530 atomic_set(&vcq->ibcq.usecnt, 0); in ionic_create_rdma_admincq() 809 if (cq->vcq->ibcq.comp_handler) in ionic_cq_event() 810 cq->vcq->ibcq.comp_handler(&cq->vcq->ibcq, in ionic_cq_event() 811 cq->vcq->ibcq.cq_context); in ionic_cq_event() 815 if (cq->vcq->ibcq.event_handler) { in ionic_cq_event() 818 ibev.element.cq = &cq->vcq->ibcq; in ionic_cq_event() [all …]
|
| H A D | ionic_controlpath.c | 85 struct ionic_ibdev *dev = to_ionic_ibdev(vcq->ibcq.device); in ionic_create_cq_common() 1214 int ionic_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr, in ionic_create_cq() argument 1217 struct ionic_ibdev *dev = to_ionic_ibdev(ibcq->device); in ionic_create_cq() 1221 struct ionic_vcq *vcq = to_ionic_vcq(ibcq); in ionic_create_cq() 1261 vcq->ibcq.cqe = attr->cqe; in ionic_create_cq() 1289 int ionic_destroy_cq(struct ib_cq *ibcq, struct ib_udata *udata) in ionic_destroy_cq() argument 1291 struct ionic_ibdev *dev = to_ionic_ibdev(ibcq->device); in ionic_destroy_cq() 1292 struct ionic_vcq *vcq = to_ionic_vcq(ibcq); in ionic_destroy_cq() 2395 if (cq->flush && cq->vcq->ibcq.comp_handler) in ionic_notify_flush_cq() 2396 cq->vcq->ibcq.comp_handler(&cq->vcq->ibcq, in ionic_notify_flush_cq() [all …]
|
| /linux/drivers/infiniband/hw/vmw_pvrdma/ |
| H A D | pvrdma_main.c | 227 INIT_RDMA_OBJ_SIZE(ib_cq, pvrdma_cq, ibcq), 359 if (cq && cq->ibcq.event_handler) { in pvrdma_cq_event() 360 struct ib_cq *ibcq = &cq->ibcq; in pvrdma_cq_event() local 363 e.device = ibcq->device; in pvrdma_cq_event() 364 e.element.cq = ibcq; in pvrdma_cq_event() 366 ibcq->event_handler(&e, ibcq->cq_context); in pvrdma_cq_event() 518 if (cq && cq->ibcq.comp_handler) in pvrdma_intrx_handler() 519 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in pvrdma_intrx_handler()
|
| /linux/drivers/infiniband/hw/efa/ |
| H A D | efa_verbs.c | 162 static inline struct efa_cq *to_ecq(struct ib_cq *ibcq) in to_ecq() argument 164 return container_of(ibcq, struct efa_cq, ibcq); in to_ecq() 1056 int efa_destroy_cq(struct ib_cq *ibcq, struct ib_udata *udata) in efa_destroy_cq() argument 1058 struct efa_dev *dev = to_edev(ibcq->device); in efa_destroy_cq() 1059 struct efa_cq *cq = to_ecq(ibcq); in efa_destroy_cq() 1113 int efa_create_user_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr, in efa_create_user_cq() argument 1122 struct ib_device *ibdev = ibcq->device; in efa_create_user_cq() 1125 struct efa_cq *cq = to_ecq(ibcq); in efa_create_user_cq() 1175 if (ibcq->umem) { in efa_create_user_cq() 1176 if (ibcq->umem->length < cq->size) { in efa_create_user_cq() [all …]
|
| /linux/drivers/infiniband/hw/irdma/ |
| H A D | utils.c | 813 void irdma_cq_add_ref(struct ib_cq *ibcq) in irdma_cq_add_ref() argument 815 struct irdma_cq *iwcq = to_iwcq(ibcq); in irdma_cq_add_ref() 820 void irdma_cq_rem_ref(struct ib_cq *ibcq) in irdma_cq_rem_ref() argument 822 struct ib_device *ibdev = ibcq->device; in irdma_cq_rem_ref() 824 struct irdma_cq *iwcq = to_iwcq(ibcq); in irdma_cq_rem_ref() 2379 ibdev_dbg(iwcq->ibcq.device, in irdma_generated_cmpls() 2406 if (!cq->ibcq.comp_handler) in irdma_comp_handler() 2409 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in irdma_comp_handler() 2454 ibdev_dbg(iwscq->ibcq.device, in irdma_generate_flush_completions() 2491 ibdev_dbg(iwrcq->ibcq.device, in irdma_generate_flush_completions()
|
| H A D | verbs.h | 142 struct ib_cq ibcq; member
|
| H A D | verbs.c | 2018 static int irdma_resize_cq(struct ib_cq *ibcq, unsigned int entries, in irdma_resize_cq() argument 2022 struct irdma_cq *iwcq = to_iwcq(ibcq); in irdma_resize_cq() 2038 iwdev = to_iwdev(ibcq->device); in irdma_resize_cq() 2155 ibcq->cqe = info.cq_size - 1; in irdma_resize_cq() 2473 static int irdma_create_cq(struct ib_cq *ibcq, in irdma_create_cq() argument 2480 struct ib_device *ibdev = ibcq->device; in irdma_create_cq() 2483 struct irdma_cq *iwcq = to_iwcq(ibcq); in irdma_create_cq() 2524 iwcq->ibcq.cqe = info.cq_uk_init_info.cq_size; in irdma_create_cq() 4481 iwdev = to_iwdev(iwcq->ibcq.device); in __irdma_poll_cq() 4558 static int irdma_poll_cq(struct ib_cq *ibcq, int num_entries, in irdma_poll_cq() argument [all …]
|
| /linux/drivers/infiniband/hw/usnic/ |
| H A D | usnic_ib.h | 65 struct ib_cq ibcq; member
|
| H A D | usnic_ib_main.c | 379 INIT_RDMA_OBJ_SIZE(ib_cq, usnic_ib_cq, ibcq),
|
| /linux/drivers/infiniband/hw/ocrdma/ |
| H A D | ocrdma_hw.c | 721 ib_evt.element.cq = &cq->ibcq; in ocrdma_dispatch_ibevent() 727 ib_evt.element.cq = &cq->ibcq; in ocrdma_dispatch_ibevent() 787 if (cq->ibcq.event_handler) in ocrdma_dispatch_ibevent() 788 cq->ibcq.event_handler(&ib_evt, cq->ibcq.cq_context); in ocrdma_dispatch_ibevent() 958 if (bcq && bcq->ibcq.comp_handler) { in ocrdma_qp_buddy_cq_handler() 960 (*bcq->ibcq.comp_handler) (&bcq->ibcq, bcq->ibcq.cq_context); in ocrdma_qp_buddy_cq_handler() 977 if (cq->ibcq.comp_handler) { in ocrdma_qp_cq_handler() 979 (*cq->ibcq.comp_handler) (&cq->ibcq, cq->ibcq.cq_context); in ocrdma_qp_cq_handler()
|
| /linux/drivers/infiniband/hw/hns/ |
| H A D | hns_roce_hw_v2.c | 912 static void handle_drain_completion(struct ib_cq *ibcq, in handle_drain_completion() argument 917 struct hns_roce_cq *hr_cq = to_hr_cq(ibcq); in handle_drain_completion() 921 if (ibcq->poll_ctx == IB_POLL_DIRECT) { in handle_drain_completion() 923 ib_process_cq_direct(ibcq, -1); in handle_drain_completion() 942 switch (ibcq->poll_ctx) { in handle_drain_completion() 944 irq_poll_disable(&ibcq->iop); in handle_drain_completion() 945 irq_poll_enable(&ibcq->iop); in handle_drain_completion() 949 cancel_work_sync(&ibcq->work); in handle_drain_completion() 956 if (ibcq->comp_handler) in handle_drain_completion() 957 ibcq->comp_handler(ibcq, ibcq->cq_context); in handle_drain_completion() [all …]
|