Home
last modified time | relevance | path

Searched refs:ibcq (Results 1 – 25 of 48) sorted by relevance

12

/linux/drivers/infiniband/sw/rdmavt/
H A Dcq.c55 if (head >= (unsigned)cq->ibcq.cqe) { in rvt_cq_enter()
56 head = cq->ibcq.cqe; in rvt_cq_enter()
69 if (cq->ibcq.event_handler) { in rvt_cq_enter()
72 ev.device = cq->ibcq.device; in rvt_cq_enter()
73 ev.element.cq = &cq->ibcq; in rvt_cq_enter()
75 cq->ibcq.event_handler(&ev, cq->ibcq.cq_context); in rvt_cq_enter()
141 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in send_complete()
159 int rvt_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr, in rvt_create_cq() argument
163 struct ib_device *ibdev = ibcq->device; in rvt_create_cq()
165 struct rvt_cq *cq = ibcq_to_rvtcq(ibcq); in rvt_create_cq()
[all …]
H A Dcq.h12 int rvt_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr,
14 int rvt_destroy_cq(struct ib_cq *ibcq, struct ib_udata *udata);
15 int rvt_req_notify_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags notify_flags);
16 int rvt_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata);
17 int rvt_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry);
/linux/drivers/infiniband/hw/mana/
H A Dcq.c8 int mana_ib_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr, in mana_ib_create_cq() argument
12 struct mana_ib_cq *cq = container_of(ibcq, struct mana_ib_cq, ibcq); in mana_ib_create_cq()
15 struct ib_device *ibdev = ibcq->device; in mana_ib_create_cq()
110 int mana_ib_destroy_cq(struct ib_cq *ibcq, struct ib_udata *udata) in mana_ib_destroy_cq() argument
112 struct mana_ib_cq *cq = container_of(ibcq, struct mana_ib_cq, ibcq); in mana_ib_destroy_cq()
113 struct ib_device *ibdev = ibcq->device; in mana_ib_destroy_cq()
134 if (cq->ibcq.comp_handler) in mana_ib_cq_handler()
135 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in mana_ib_cq_handler()
177 int mana_ib_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags flags) in mana_ib_arm_cq() argument
179 struct mana_ib_cq *cq = container_of(ibcq, struct mana_ib_cq, ibcq); in mana_ib_arm_cq()
[all …]
H A Dqp.c108 struct ib_cq *ibcq; in mana_ib_create_qp_rss() local
182 ibcq = ibwq->cq; in mana_ib_create_qp_rss()
183 cq = container_of(ibcq, struct mana_ib_cq, ibcq); in mana_ib_create_qp_rss()
248 ibcq = ibwq->cq; in mana_ib_create_qp_rss()
250 cq = container_of(ibcq, struct mana_ib_cq, ibcq); in mana_ib_create_qp_rss()
270 container_of(attr->send_cq, struct mana_ib_cq, ibcq); in mana_ib_create_qp_raw()
604 struct mana_ib_cq *send_cq = container_of(qp->ibqp.send_cq, struct mana_ib_cq, ibcq); in mana_add_qp_to_cqs()
605 struct mana_ib_cq *recv_cq = container_of(qp->ibqp.recv_cq, struct mana_ib_cq, ibcq); in mana_add_qp_to_cqs()
619 struct mana_ib_cq *send_cq = container_of(qp->ibqp.send_cq, struct mana_ib_cq, ibcq); in mana_remove_qp_from_cqs()
620 struct mana_ib_cq *recv_cq = container_of(qp->ibqp.recv_cq, struct mana_ib_cq, ibcq); in mana_remove_qp_from_cqs()
/linux/drivers/infiniband/hw/mlx4/
H A Dcq.c45 struct ib_cq *ibcq = &to_mibcq(cq)->ibcq; in mlx4_ib_cq_comp() local
46 ibcq->comp_handler(ibcq, ibcq->cq_context); in mlx4_ib_cq_comp()
52 struct ib_cq *ibcq; in mlx4_ib_cq_event() local
60 ibcq = &to_mibcq(cq)->ibcq; in mlx4_ib_cq_event()
61 if (ibcq->event_handler) { in mlx4_ib_cq_event()
62 event.device = ibcq->device; in mlx4_ib_cq_event()
64 event.element.cq = ibcq; in mlx4_ib_cq_event()
65 ibcq->event_handler(&event, ibcq->cq_context); in mlx4_ib_cq_event()
81 struct mlx4_cqe *cqe = get_cqe(cq, n & cq->ibcq.cqe); in get_sw_cqe()
85 !!(n & (cq->ibcq.cqe + 1))) ? NULL : cqe; in get_sw_cqe()
[all …]
/linux/drivers/infiniband/hw/vmw_pvrdma/
H A Dpvrdma_cq.c63 int pvrdma_req_notify_cq(struct ib_cq *ibcq, in pvrdma_req_notify_cq() argument
66 struct pvrdma_dev *dev = to_vdev(ibcq->device); in pvrdma_req_notify_cq()
67 struct pvrdma_cq *cq = to_vcq(ibcq); in pvrdma_req_notify_cq()
83 cq->ibcq.cqe, &head); in pvrdma_req_notify_cq()
101 int pvrdma_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr, in pvrdma_create_cq() argument
105 struct ib_device *ibdev = ibcq->device; in pvrdma_create_cq()
108 struct pvrdma_cq *cq = to_vcq(ibcq); in pvrdma_create_cq()
133 cq->ibcq.cqe = entries; in pvrdma_create_cq()
196 cq->ibcq.cqe = resp->cqe; in pvrdma_create_cq()
210 pvrdma_destroy_cq(&cq->ibcq, udata); in pvrdma_create_cq()
[all …]
H A Dpvrdma_main.c227 INIT_RDMA_OBJ_SIZE(ib_cq, pvrdma_cq, ibcq),
359 if (cq && cq->ibcq.event_handler) { in pvrdma_cq_event()
360 struct ib_cq *ibcq = &cq->ibcq; in pvrdma_cq_event() local
363 e.device = ibcq->device; in pvrdma_cq_event()
364 e.element.cq = ibcq; in pvrdma_cq_event()
366 ibcq->event_handler(&e, ibcq->cq_context); in pvrdma_cq_event()
518 if (cq && cq->ibcq.comp_handler) in pvrdma_intrx_handler()
519 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in pvrdma_intrx_handler()
H A Dpvrdma.h87 struct ib_cq ibcq; member
274 static inline struct pvrdma_cq *to_vcq(struct ib_cq *ibcq) in to_vcq() argument
276 return container_of(ibcq, struct pvrdma_cq, ibcq); in to_vcq()
/linux/drivers/infiniband/sw/rxe/
H A Drxe_cq.c65 cq->ibcq.cqe = cqe; in rxe_cq_from_init()
79 cq->ibcq.cqe = cqe; in rxe_cq_resize_queue()
98 if (cq->ibcq.event_handler) { in rxe_cq_post()
99 ev.device = cq->ibcq.device; in rxe_cq_post()
100 ev.element.cq = &cq->ibcq; in rxe_cq_post()
102 cq->ibcq.event_handler(&ev, cq->ibcq.cq_context); in rxe_cq_post()
116 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in rxe_cq_post()
H A Drxe_verbs.c1075 static int rxe_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr, in rxe_create_cq() argument
1079 struct ib_device *dev = ibcq->device; in rxe_create_cq()
1081 struct rxe_cq *cq = to_rcq(ibcq); in rxe_create_cq()
1130 static int rxe_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata) in rxe_resize_cq() argument
1132 struct rxe_cq *cq = to_rcq(ibcq); in rxe_resize_cq()
1133 struct rxe_dev *rxe = to_rdev(ibcq->device); in rxe_resize_cq()
1165 static int rxe_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc) in rxe_poll_cq() argument
1168 struct rxe_cq *cq = to_rcq(ibcq); in rxe_poll_cq()
1186 static int rxe_peek_cq(struct ib_cq *ibcq, int wc_cnt) in rxe_peek_cq() argument
1188 struct rxe_cq *cq = to_rcq(ibcq); in rxe_peek_cq()
[all …]
/linux/drivers/infiniband/hw/mthca/
H A Dmthca_cq.c181 return cqe_sw(get_cqe(cq, cq->cons_index & cq->ibcq.cqe)); in next_cqe_sw()
230 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in mthca_cq_completion()
254 event.element.cq = &cq->ibcq; in mthca_cq_event()
255 if (cq->ibcq.event_handler) in mthca_cq_event()
256 cq->ibcq.event_handler(&event, cq->ibcq.cq_context); in mthca_cq_event()
290 cqe_sw(get_cqe(cq, prod_index & cq->ibcq.cqe)); in mthca_cq_clean()
292 if (prod_index == cq->cons_index + cq->ibcq.cqe) in mthca_cq_clean()
304 cqe = get_cqe(cq, prod_index & cq->ibcq.cqe); in mthca_cq_clean()
310 memcpy(get_cqe(cq, (prod_index + nfreed) & cq->ibcq.cqe), in mthca_cq_clean()
316 set_cqe_hw(get_cqe(cq, (cq->cons_index + i) & cq->ibcq.cqe)); in mthca_cq_clean()
[all …]
H A Dmthca_provider.h184 struct ib_cq ibcq; member
301 static inline struct mthca_cq *to_mcq(struct ib_cq *ibcq) in to_mcq() argument
303 return container_of(ibcq, struct mthca_cq, ibcq); in to_mcq()
H A Dmthca_provider.c576 static int mthca_create_cq(struct ib_cq *ibcq, in mthca_create_cq() argument
581 struct ib_device *ibdev = ibcq->device; in mthca_create_cq()
613 cq = to_mcq(ibcq); in mthca_create_cq()
698 static int mthca_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata) in mthca_resize_cq() argument
700 struct mthca_dev *dev = to_mdev(ibcq->device); in mthca_resize_cq()
701 struct mthca_cq *cq = to_mcq(ibcq); in mthca_resize_cq()
712 if (entries == ibcq->cqe + 1) { in mthca_resize_cq()
752 tcqe = cq->ibcq.cqe; in mthca_resize_cq()
754 cq->ibcq.cqe = cq->resize_buf->cqe; in mthca_resize_cq()
766 ibcq->cqe = entries - 1; in mthca_resize_cq()
[all …]
/linux/drivers/infiniband/hw/mlx5/
H A Dcq.c46 struct ib_cq *ibcq = &to_mibcq(cq)->ibcq; in mlx5_ib_cq_comp() local
48 ibcq->comp_handler(ibcq, ibcq->cq_context); in mlx5_ib_cq_comp()
54 struct mlx5_ib_dev *dev = to_mdev(cq->ibcq.device); in mlx5_ib_cq_event()
55 struct ib_cq *ibcq = &cq->ibcq; in mlx5_ib_cq_event() local
64 if (ibcq->event_handler) { in mlx5_ib_cq_event()
67 event.element.cq = ibcq; in mlx5_ib_cq_event()
68 ibcq->event_handler(&event, ibcq->cq_context); in mlx5_ib_cq_event()
84 void *cqe = get_cqe(cq, n & cq->ibcq.cqe); in get_sw_cqe()
90 !((cqe64->op_own & MLX5_CQE_OWNER_MASK) ^ !!(n & (cq->ibcq.cqe + 1)))) { in get_sw_cqe()
454 struct mlx5_ib_dev *dev = to_mdev(cq->ibcq.device); in mlx5_poll_one()
[all …]
/linux/include/rdma/
H A Drdmavt_cq.h47 struct ib_cq ibcq; member
60 static inline struct rvt_cq *ibcq_to_rvtcq(struct ib_cq *ibcq) in ibcq_to_rvtcq() argument
62 return container_of(ibcq, struct rvt_cq, ibcq); in ibcq_to_rvtcq()
/linux/drivers/infiniband/hw/cxgb4/
H A Dev.c105 event.device = chp->ibcq.device; in post_qp_event()
107 event.element.cq = &chp->ibcq; in post_qp_event()
115 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in post_qp_event()
234 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in c4iw_ev_handler()
H A Dcq.c949 int c4iw_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc) in c4iw_poll_cq() argument
956 chp = to_c4iw_cq(ibcq); in c4iw_poll_cq()
997 int c4iw_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr, in c4iw_create_cq() argument
1001 struct ib_device *ibdev = ibcq->device; in c4iw_create_cq()
1004 struct c4iw_dev *rhp = to_c4iw_dev(ibcq->device); in c4iw_create_cq()
1005 struct c4iw_cq *chp = to_c4iw_cq(ibcq); in c4iw_create_cq()
1087 chp->ibcq.cqe = entries - 2; in c4iw_create_cq()
1167 int c4iw_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags flags) in c4iw_arm_cq() argument
1173 chp = to_c4iw_cq(ibcq); in c4iw_arm_cq()
H A Dqp.c1008 (*schp->ibcq.comp_handler)(&schp->ibcq, in complete_sq_drain_wr()
1009 schp->ibcq.cq_context); in complete_sq_drain_wr()
1059 (*rchp->ibcq.comp_handler)(&rchp->ibcq, in complete_rq_drain_wr()
1060 rchp->ibcq.cq_context); in complete_rq_drain_wr()
1645 (*rchp->ibcq.comp_handler)(&rchp->ibcq, in __flush_qp()
1646 rchp->ibcq.cq_context); in __flush_qp()
1652 (*rchp->ibcq.comp_handler)(&rchp->ibcq, in __flush_qp()
1653 rchp->ibcq.cq_context); in __flush_qp()
1658 (*schp->ibcq.comp_handler)(&schp->ibcq, in __flush_qp()
1659 schp->ibcq.cq_context); in __flush_qp()
[all …]
/linux/drivers/infiniband/hw/qedr/
H A Dqedr_roce_cm.c82 cq->ibcq.comp_handler ? "Yes" : "No"); in qedr_ll2_complete_tx_packet()
92 if (cq->ibcq.comp_handler) in qedr_ll2_complete_tx_packet()
93 (*cq->ibcq.comp_handler) (&cq->ibcq, cq->ibcq.cq_context); in qedr_ll2_complete_tx_packet()
121 if (cq->ibcq.comp_handler) in qedr_ll2_complete_rx_packet()
122 (*cq->ibcq.comp_handler) (&cq->ibcq, cq->ibcq.cq_context); in qedr_ll2_complete_rx_packet()
668 int qedr_gsi_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc) in qedr_gsi_poll_cq() argument
670 struct qedr_dev *dev = get_qedr_dev(ibcq->device); in qedr_gsi_poll_cq()
671 struct qedr_cq *cq = get_qedr_cq(ibcq); in qedr_gsi_poll_cq()
H A Dmain.c233 INIT_RDMA_OBJ_SIZE(ib_cq, qedr_cq, ibcq),
472 if (!cq->destroyed && cq->ibcq.comp_handler) in qedr_irq_handler()
473 (*cq->ibcq.comp_handler) in qedr_irq_handler()
474 (&cq->ibcq, cq->ibcq.cq_context); in qedr_irq_handler()
653 struct ib_cq *ibcq; in qedr_affiliated_event() local
724 ibcq = &cq->ibcq; in qedr_affiliated_event()
725 if (ibcq->event_handler) { in qedr_affiliated_event()
726 event.device = ibcq->device; in qedr_affiliated_event()
727 event.element.cq = ibcq; in qedr_affiliated_event()
728 ibcq->event_handler(&event, ibcq->cq_context); in qedr_affiliated_event()
H A Dqedr_roce_cm.h48 int qedr_gsi_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
/linux/drivers/infiniband/hw/ionic/
H A Dionic_admin.c492 static void ionic_rdma_admincq_comp(struct ib_cq *ibcq, void *cq_context) in ionic_rdma_admincq_comp() argument
527 vcq->ibcq.device = &dev->ibdev; in ionic_create_rdma_admincq()
528 vcq->ibcq.comp_handler = ionic_rdma_admincq_comp; in ionic_create_rdma_admincq()
529 vcq->ibcq.event_handler = ionic_rdma_admincq_event; in ionic_create_rdma_admincq()
530 atomic_set(&vcq->ibcq.usecnt, 0); in ionic_create_rdma_admincq()
809 if (cq->vcq->ibcq.comp_handler) in ionic_cq_event()
810 cq->vcq->ibcq.comp_handler(&cq->vcq->ibcq, in ionic_cq_event()
811 cq->vcq->ibcq.cq_context); in ionic_cq_event()
815 if (cq->vcq->ibcq.event_handler) { in ionic_cq_event()
818 ibev.element.cq = &cq->vcq->ibcq; in ionic_cq_event()
[all …]
/linux/drivers/infiniband/hw/ocrdma/
H A Docrdma_verbs.c968 int ocrdma_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr, in ocrdma_create_cq() argument
972 struct ib_device *ibdev = ibcq->device; in ocrdma_create_cq()
974 struct ocrdma_cq *cq = get_ocrdma_cq(ibcq); in ocrdma_create_cq()
1017 int ocrdma_resize_cq(struct ib_cq *ibcq, int new_cnt, in ocrdma_resize_cq() argument
1021 struct ocrdma_cq *cq = get_ocrdma_cq(ibcq); in ocrdma_resize_cq()
1027 ibcq->cqe = new_cnt; in ocrdma_resize_cq()
1037 struct ocrdma_dev *dev = get_ocrdma_dev(cq->ibcq.device); in ocrdma_flush_cq()
1057 int ocrdma_destroy_cq(struct ib_cq *ibcq, struct ib_udata *udata) in ocrdma_destroy_cq() argument
1059 struct ocrdma_cq *cq = get_ocrdma_cq(ibcq); in ocrdma_destroy_cq()
1061 struct ocrdma_dev *dev = get_ocrdma_dev(ibcq->device); in ocrdma_destroy_cq()
[all …]
/linux/drivers/infiniband/hw/erdma/
H A Derdma_verbs.c194 struct erdma_dev *dev = to_edev(cq->ibcq.device); in create_cq_cmd()
206 if (rdma_is_kernel_res(&cq->ibcq.res)) { in create_cq_cmd()
1316 int erdma_destroy_cq(struct ib_cq *ibcq, struct ib_udata *udata) in erdma_destroy_cq() argument
1318 struct erdma_cq *cq = to_ecq(ibcq); in erdma_destroy_cq()
1319 struct erdma_dev *dev = to_edev(ibcq->device); in erdma_destroy_cq()
1334 if (rdma_is_kernel_res(&cq->ibcq.res)) { in erdma_destroy_cq()
1909 struct erdma_dev *dev = to_edev(cq->ibcq.device); in erdma_init_user_cq()
1928 struct erdma_dev *dev = to_edev(cq->ibcq.device); in erdma_init_kernel_cq()
1954 int erdma_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr, in erdma_create_cq() argument
1958 struct erdma_cq *cq = to_ecq(ibcq); in erdma_create_cq()
[all …]
/linux/drivers/infiniband/hw/efa/
H A Defa_main.c91 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in efa_process_comp_eqe()
399 INIT_RDMA_OBJ_SIZE(ib_cq, efa_cq, ibcq),

12