Home
last modified time | relevance | path

Searched refs:cqe (Results 1 – 25 of 219) sorted by relevance

123456789

/linux/drivers/infiniband/hw/mthca/
H A Dmthca_cq.c174 static inline struct mthca_cqe *cqe_sw(struct mthca_cqe *cqe) in cqe_sw() argument
176 return MTHCA_CQ_ENTRY_OWNER_HW & cqe->owner ? NULL : cqe; in cqe_sw()
181 return cqe_sw(get_cqe(cq, cq->cons_index & cq->ibcq.cqe)); in next_cqe_sw()
184 static inline void set_cqe_hw(struct mthca_cqe *cqe) in set_cqe_hw() argument
186 cqe->owner = MTHCA_CQ_ENTRY_OWNER_HW; in set_cqe_hw()
191 __be32 *cqe = cqe_ptr; in dump_cqe() local
193 (void) cqe; /* avoid warning if mthca_dbg compiled away... */ in dump_cqe()
195 be32_to_cpu(cqe[0]), be32_to_cpu(cqe[1]), be32_to_cpu(cqe[2]), in dump_cqe()
196 be32_to_cpu(cqe[3]), be32_to_cpu(cqe[4]), be32_to_cpu(cqe[5]), in dump_cqe()
197 be32_to_cpu(cqe[6]), be32_to_cpu(cqe[7])); in dump_cqe()
[all …]
/linux/drivers/infiniband/hw/mlx4/
H A Dcq.c81 struct mlx4_cqe *cqe = get_cqe(cq, n & cq->ibcq.cqe); in get_sw_cqe() local
82 struct mlx4_cqe *tcqe = ((cq->buf.entry_size == 64) ? (cqe + 1) : cqe); in get_sw_cqe()
85 !!(n & (cq->ibcq.cqe + 1))) ? NULL : cqe; in get_sw_cqe()
133 static void mlx4_ib_free_cq_buf(struct mlx4_ib_dev *dev, struct mlx4_ib_cq_buf *buf, int cqe) in mlx4_ib_free_cq_buf() argument
135 mlx4_buf_free(dev->dev, (cqe + 1) * buf->entry_size, &buf->buf); in mlx4_ib_free_cq_buf()
145 int entries = attr->cqe; in mlx4_ib_get_cq_umem()
158 if (attr->cqe > de in mlx4_ib_get_cq_umem()
140 mlx4_ib_get_cq_umem(struct mlx4_ib_dev * dev,struct mlx4_ib_cq_buf * buf,struct ib_umem ** umem,u64 buf_addr,int cqe) mlx4_ib_get_cq_umem() argument
361 struct mlx4_cqe *cqe, *new_cqe; mlx4_ib_cq_resize_copy_cqes() local
505 dump_cqe(void * cqe) dump_cqe() argument
515 mlx4_ib_handle_error_cqe(struct mlx4_err_cqe * cqe,struct ib_wc * wc) mlx4_ib_handle_error_cqe() argument
586 use_tunnel_data(struct mlx4_ib_qp * qp,struct mlx4_ib_cq * cq,struct ib_wc * wc,unsigned tail,struct mlx4_cqe * cqe,int is_eth) use_tunnel_data() argument
665 struct mlx4_cqe *cqe; mlx4_ib_poll_one() local
924 struct mlx4_cqe *cqe, *dest; __mlx4_ib_cq_clean() local
[all...]
/linux/drivers/infiniband/hw/vmw_pvrdma/
H A Dpvrdma_cq.c83 cq->ibcq.cqe, &head); in pvrdma_req_notify_cq()
106 int entries = attr->cqe; in pvrdma_create_cq()
133 cq->ibcq.cqe = entries; in pvrdma_create_cq()
186 cmd->cqe = entries; in pvrdma_create_cq()
195 cq->ibcq.cqe = resp->cqe; in pvrdma_create_cq()
288 cq->ibcq.cqe, &head); in _pvrdma_flush_cqe()
293 cq->ibcq.cqe); in _pvrdma_flush_cqe()
294 struct pvrdma_cqe *cqe; in _pvrdma_flush_cqe()
298 (cq->ibcq.cqe in _pvrdma_flush_cqe()
295 struct pvrdma_cqe *cqe; _pvrdma_flush_cqe() local
330 struct pvrdma_cqe *cqe; pvrdma_poll_one() local
[all...]
/linux/drivers/infiniband/sw/rxe/
H A Drxe_cq.c11 int rxe_cq_from_init(struct rxe_dev *rxe, struct rxe_cq *cq, int cqe, in rxe_cq_chk_attr()
19 cq->queue = rxe_queue_init(rxe, &cqe, in rxe_cq_chk_attr()
34 cq->ibcq.cqe = cqe; in rxe_cq_chk_attr()
38 int rxe_cq_resize_queue(struct rxe_cq *cq, int cqe, in rxe_cq_chk_attr()
44 err = rxe_queue_resize(cq->queue, (unsigned int *)&cqe, in rxe_cq_from_init()
48 cq->ibcq.cqe = cqe; in rxe_cq_from_init()
54 int rxe_cq_post(struct rxe_cq *cq, struct rxe_cqe *cqe, int solicited) in rxe_cq_from_init()
78 memcpy(addr, cqe, sizeo in rxe_cq_resize_queue()
12 rxe_cq_chk_attr(struct rxe_dev * rxe,struct rxe_cq * cq,int cqe,int comp_vector) rxe_cq_chk_attr() argument
42 rxe_cq_from_init(struct rxe_dev * rxe,struct rxe_cq * cq,int cqe,int comp_vector,struct ib_udata * udata,struct rxe_create_cq_resp __user * uresp) rxe_cq_from_init() argument
69 rxe_cq_resize_queue(struct rxe_cq * cq,int cqe,struct rxe_resize_cq_resp __user * uresp,struct ib_udata * udata) rxe_cq_resize_queue() argument
85 rxe_cq_post(struct rxe_cq * cq,struct rxe_cqe * cqe,int solicited) rxe_cq_post() argument
[all...]
/linux/drivers/net/ethernet/mellanox/mlx5/core/
H A Den_rx.c66 struct mlx5_cqe64 *cqe, u16 cqe_bcnt, u32 head_offset,
70 struct mlx5_cqe64 *cqe, u16 cqe_bcnt, u32 head_offset,
72 static void mlx5e_handle_rx_cqe(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe);
73 static void mlx5e_handle_rx_cqe_mpwrq(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe);
74 static void mlx5e_handle_rx_cqe_mpwrq_shampo(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe);
91 struct mlx5_cqe64 *cqe) in mlx5e_read_enhanced_title_slot() argument
96 memcpy(title, cqe, sizeof(struct mlx5_cqe64)); in mlx5e_read_enhanced_title_slot()
139 struct mlx5_cqe64 *cqe = mlx5_cqwq_get_wqe(wq, ci); in mlx5e_cqes_update_owner() local
141 cqe->op_own = op_own; in mlx5e_cqes_update_owner()
147 struct mlx5_cqe64 *cqe in mlx5e_cqes_update_owner() local
197 mlx5e_decompress_enhanced_cqe(struct mlx5e_rq * rq,struct mlx5_cqwq * wq,struct mlx5_cqe64 * cqe,int budget_rem) mlx5e_decompress_enhanced_cqe() argument
790 struct mlx5_cqe64 *cqe; mlx5e_poll_ico_cq() local
962 mlx5e_lro_update_tcp_hdr(struct mlx5_cqe64 * cqe,struct tcphdr * tcp) mlx5e_lro_update_tcp_hdr() argument
979 mlx5e_lro_update_hdr(struct sk_buff * skb,struct mlx5_cqe64 * cqe,u32 cqe_bcnt) mlx5e_lro_update_hdr() argument
1073 mlx5e_shampo_get_hd_buf_info(struct mlx5e_rq * rq,struct mlx5_cqe64 * cqe,struct mlx5e_dma_info ** di,u32 * head_offset) mlx5e_shampo_get_hd_buf_info() argument
1087 mlx5e_shampo_get_hdr(struct mlx5e_rq * rq,struct mlx5_cqe64 * cqe,int len) mlx5e_shampo_get_hdr() argument
1101 mlx5e_shampo_update_fin_psh_flags(struct mlx5e_rq * rq,struct mlx5_cqe64 * cqe,struct tcphdr * skb_tcp_hd) mlx5e_shampo_update_fin_psh_flags() argument
1116 mlx5e_shampo_update_ipv4_tcp_hdr(struct mlx5e_rq * rq,struct iphdr * ipv4,struct mlx5_cqe64 * cqe,bool match) mlx5e_shampo_update_ipv4_tcp_hdr() argument
1144 mlx5e_shampo_update_ipv6_tcp_hdr(struct mlx5e_rq * rq,struct ipv6hdr * ipv6,struct mlx5_cqe64 * cqe,bool match) mlx5e_shampo_update_ipv6_tcp_hdr() argument
1164 mlx5e_shampo_update_hdr(struct mlx5e_rq * rq,struct mlx5_cqe64 * cqe,bool match) mlx5e_shampo_update_hdr() argument
1197 mlx5e_skb_set_hash(struct mlx5_cqe64 * cqe,struct sk_buff * skb) mlx5e_skb_set_hash() argument
1323 mlx5e_handle_csum(struct net_device * netdev,struct mlx5_cqe64 * cqe,struct mlx5e_rq * rq,struct sk_buff * skb,bool lro) mlx5e_handle_csum() argument
1398 mlx5e_build_rx_skb(struct mlx5_cqe64 * cqe,u32 cqe_bcnt,struct mlx5e_rq * rq,struct sk_buff * skb) mlx5e_build_rx_skb() argument
1467 mlx5e_shampo_complete_rx_cqe(struct mlx5e_rq * rq,struct mlx5_cqe64 * cqe,u32 cqe_bcnt,struct sk_buff * skb) mlx5e_shampo_complete_rx_cqe() argument
1490 mlx5e_complete_rx_cqe(struct mlx5e_rq * rq,struct mlx5_cqe64 * cqe,u32 cqe_bcnt,struct sk_buff * skb) mlx5e_complete_rx_cqe() argument
1522 mlx5e_fill_mxbuf(struct mlx5e_rq * rq,struct mlx5_cqe64 * cqe,void * va,u16 headroom,u32 frame_sz,u32 len,struct mlx5e_xdp_buff * mxbuf) mlx5e_fill_mxbuf() argument
1534 mlx5e_skb_from_cqe_linear(struct mlx5e_rq * rq,struct mlx5e_wqe_frag_info * wi,struct mlx5_cqe64 * cqe,u32 cqe_bcnt) mlx5e_skb_from_cqe_linear() argument
1582 mlx5e_skb_from_cqe_nonlinear(struct mlx5e_rq * rq,struct mlx5e_wqe_frag_info * wi,struct mlx5_cqe64 * cqe,u32 cqe_bcnt) mlx5e_skb_from_cqe_nonlinear() argument
1677 trigger_report(struct mlx5e_rq * rq,struct mlx5_cqe64 * cqe) trigger_report() argument
1689 mlx5e_handle_rx_err_cqe(struct mlx5e_rq * rq,struct mlx5_cqe64 * cqe) mlx5e_handle_rx_err_cqe() argument
1695 mlx5e_handle_rx_cqe(struct mlx5e_rq * rq,struct mlx5_cqe64 * cqe) mlx5e_handle_rx_cqe() argument
1740 mlx5e_handle_rx_cqe_rep(struct mlx5e_rq * rq,struct mlx5_cqe64 * cqe) mlx5e_handle_rx_cqe_rep() argument
1784 mlx5e_handle_rx_cqe_mpwrq_rep(struct mlx5e_rq * rq,struct mlx5_cqe64 * cqe) mlx5e_handle_rx_cqe_mpwrq_rep() argument
1865 mlx5e_skb_from_cqe_mpwrq_nonlinear(struct mlx5e_rq * rq,struct mlx5e_mpw_info * wi,struct mlx5_cqe64 * cqe,u16 cqe_bcnt,u32 head_offset,u32 page_idx) mlx5e_skb_from_cqe_mpwrq_nonlinear() argument
2045 mlx5e_skb_from_cqe_mpwrq_linear(struct mlx5e_rq * rq,struct mlx5e_mpw_info * wi,struct mlx5_cqe64 * cqe,u16 cqe_bcnt,u32 head_offset,u32 page_idx) mlx5e_skb_from_cqe_mpwrq_linear() argument
2103 mlx5e_skb_from_cqe_shampo(struct mlx5e_rq * rq,struct mlx5e_mpw_info * wi,struct mlx5_cqe64 * cqe,u16 header_index) mlx5e_skb_from_cqe_shampo() argument
2143 mlx5e_shampo_flush_skb(struct mlx5e_rq * rq,struct mlx5_cqe64 * cqe,bool match) mlx5e_shampo_flush_skb() argument
2175 mlx5e_handle_rx_cqe_mpwrq_shampo(struct mlx5e_rq * rq,struct mlx5_cqe64 * cqe) mlx5e_handle_rx_cqe_mpwrq_shampo() argument
2285 mlx5e_handle_rx_cqe_mpwrq(struct mlx5e_rq * rq,struct mlx5_cqe64 * cqe) mlx5e_handle_rx_cqe_mpwrq() argument
2349 struct mlx5_cqe64 *cqe, *title_cqe = NULL; mlx5e_rx_cq_process_enhanced_cqe_comp() local
2398 struct mlx5_cqe64 *cqe; mlx5e_rx_cq_process_basic_cqe_comp() local
2462 mlx5i_complete_rx_cqe(struct mlx5e_rq * rq,struct mlx5_cqe64 * cqe,u32 cqe_bcnt,struct sk_buff * skb) mlx5i_complete_rx_cqe() argument
2544 mlx5i_handle_rx_cqe(struct mlx5e_rq * rq,struct mlx5_cqe64 * cqe) mlx5i_handle_rx_cqe() argument
2634 mlx5e_trap_handle_rx_cqe(struct mlx5e_rq * rq,struct mlx5_cqe64 * cqe) mlx5e_trap_handle_rx_cqe() argument
[all...]
/linux/drivers/infiniband/sw/siw/
H A Dsiw_cq.c50 struct siw_cqe *cqe; in siw_reap_cqe() local
55 cqe = &cq->queue[cq->cq_get % cq->num_cqe]; in siw_reap_cqe()
56 if (READ_ONCE(cqe->flags) & SIW_WQE_VALID) { in siw_reap_cqe()
58 wc->wr_id = cqe->id; in siw_reap_cqe()
59 wc->byte_len = cqe->bytes; in siw_reap_cqe()
67 if (cqe->flags & SIW_WQE_REM_INVAL) { in siw_reap_cqe()
68 wc->ex.invalidate_rkey = cqe->inval_stag; in siw_reap_cqe()
71 wc->qp = cqe->base_qp; in siw_reap_cqe()
72 wc->opcode = map_wc_opcode[cqe->opcode]; in siw_reap_cqe()
73 wc->status = map_cqe_status[cqe->status].ib; in siw_reap_cqe()
[all …]
/linux/drivers/net/ethernet/mellanox/mlxsw/
H A Dpci_hw.h116 static inline u32 mlxsw_pci_cqe_##name##_get(enum mlxsw_pci_cqe_v v, char *cqe) \
121 return mlxsw_pci_cqe##v0##_##name##_get(cqe); \
123 return mlxsw_pci_cqe##v1##_##name##_get(cqe); \
125 return mlxsw_pci_cqe##v2##_##name##_get(cqe); \
129 char *cqe, u32 val) \
134 mlxsw_pci_cqe##v0##_##name##_set(cqe, val); \
137 mlxsw_pci_cqe##v1##_##name##_set(cqe, val); \
140 mlxsw_pci_cqe##v2##_##name##_set(cqe, val); \
158 MLXSW_ITEM32(pci, cqe, system_port, 0x00, 0, 16);
169 MLXSW_ITEM32(pci, cqe, wqe_counter, 0x04, 16, 16);
[all …]
/linux/drivers/infiniband/hw/cxgb4/
H A Dcq.c186 struct t4_cqe cqe; in insert_recv_cqe() local
190 memset(&cqe, 0, sizeof(cqe)); in insert_recv_cqe()
191 cqe.header = cpu_to_be32(CQE_STATUS_V(T4_ERR_SWFLUSH) | in insert_recv_cqe()
196 cqe.bits_type_ts = cpu_to_be64(CQE_GENBIT_V((u64)cq->gen)); in insert_recv_cqe()
198 cqe.u.srcqe.abs_rqe_idx = cpu_to_be32(srqidx); in insert_recv_cqe()
199 cq->sw_queue[cq->sw_pidx] = cqe; in insert_recv_cqe()
220 struct t4_cqe cqe; in insert_sq_cqe() local
224 memset(&cqe, 0, sizeof(cqe)); in insert_sq_cqe()
225 cqe.header = cpu_to_be32(CQE_STATUS_V(T4_ERR_SWFLUSH) | in insert_sq_cqe()
230 CQE_WRID_SQ_IDX(&cqe) = swcqe->idx; in insert_sq_cqe()
[all …]
/linux/drivers/infiniband/hw/mlx5/
H A Dcq.c84 void *cqe = get_cqe(cq, n & cq->ibcq.cqe); in get_sw_cqe() local
87 cqe64 = (cq->mcq.cqe_sz == 64) ? cqe : cqe + 64; in get_sw_cqe()
90 !((cqe64->op_own & MLX5_CQE_OWNER_MASK) ^ !!(n & (cq->ibcq.cqe + 1)))) { in get_sw_cqe()
91 return cqe; in get_sw_cqe()
120 static void handle_good_req(struct ib_wc *wc, struct mlx5_cqe64 *cqe, in handle_good_req() argument
124 switch (be32_to_cpu(cqe->sop_drop_qpn) >> 24) { in handle_good_req()
140 wc->byte_len = be32_to_cpu(cqe->byte_cnt); in handle_good_req()
169 static void handle_responder(struct ib_wc *wc, struct mlx5_cqe64 *cqe, in handle_responder() argument
273 dump_cqe(struct mlx5_ib_dev * dev,struct mlx5_err_cqe * cqe,struct ib_wc * wc,const char * level) dump_cqe() argument
283 mlx5_handle_error_cqe(struct mlx5_ib_dev * dev,struct mlx5_err_cqe * cqe,struct ib_wc * wc) mlx5_handle_error_cqe() argument
365 get_sig_err_item(struct mlx5_sig_err_cqe * cqe,struct ib_sig_err * item) get_sig_err_item() argument
462 void *cqe; mlx5_poll_one() local
877 void *cqe; init_cq_frag_buf() local
1097 void *cqe, *dest; __mlx5_ib_cq_clean() local
[all...]
/linux/tools/testing/selftests/net/
H A Dio_uring_zerocopy_tx.c97 struct io_uring_cqe *cqe; in do_tx() local
157 ret = io_uring_wait_cqe(&ring, &cqe); in do_tx()
159 error(1, -ret, "wait cqe"); in do_tx()
161 if (cqe->user_data != NONZC_TAG && in do_tx()
162 cqe->user_data != ZC_TAG) in do_tx()
163 error(1, EINVAL, "invalid cqe->user_data"); in do_tx()
165 if (cqe->flags & IORING_CQE_F_NOTIF) { in do_tx()
166 if (cqe->flags & IORING_CQE_F_MORE) in do_tx()
175 if (cqe->flags & IORING_CQE_F_MORE) { in do_tx()
176 if (cqe in do_tx()
[all...]
/linux/drivers/infiniband/hw/mana/
H A Dcq.c34 if ((!is_rnic_cq && attr->cqe > mdev->adapter_caps.max_qp_wr) || in mana_ib_create_cq()
35 attr->cqe > U32_MAX / COMP_ENTRY_SIZE) { in mana_ib_create_cq()
36 ibdev_dbg(ibdev, "CQE %d exceeding limit\n", attr->cqe); in mana_ib_create_cq()
40 cq->cqe = attr->cqe; in mana_ib_create_cq()
41 err = mana_ib_create_queue(mdev, ucmd.buf_addr, cq->cqe * COMP_ENTRY_SIZE, in mana_ib_create_cq()
52 if (attr->cqe > U32_MAX / COMP_ENTRY_SIZE / 2 + 1) { in mana_ib_create_cq()
53 ibdev_dbg(ibdev, "CQE %d exceeding limit\n", attr->cqe); in mana_ib_create_cq()
56 buf_size = MANA_PAGE_ALIGN(roundup_pow_of_two(attr->cqe * COMP_ENTRY_SIZE)); in mana_ib_create_cq()
57 cq->cqe in mana_ib_create_cq()
189 handle_ud_sq_cqe(struct mana_ib_qp * qp,struct gdma_comp * cqe) handle_ud_sq_cqe() argument
205 handle_ud_rq_cqe(struct mana_ib_qp * qp,struct gdma_comp * cqe) handle_ud_rq_cqe() argument
223 mana_handle_cqe(struct mana_ib_dev * mdev,struct gdma_comp * cqe) mana_handle_cqe() argument
[all...]
/linux/drivers/infiniband/sw/rdmavt/
H A Dcq.c55 if (head >= (unsigned)cq->ibcq.cqe) { in rvt_cq_enter()
56 head = cq->ibcq.cqe; in rvt_cq_enter()
169 unsigned int entries = attr->cqe; in rvt_create_cq()
239 * ib_create_cq() will initialize cq->ibcq except for cq->ibcq.cqe. in rvt_create_cq()
251 cq->ibcq.cqe = entries; in rvt_create_cq()
340 int rvt_resize_cq(struct ib_cq *ibcq, unsigned int cqe, struct ib_udata *udata) in rvt_resize_cq() argument
352 if (cqe > rdi->dparms.props.max_cqe) in rvt_resize_cq()
359 sz = sizeof(struct ib_uverbs_wc) * (cqe + 1); in rvt_resize_cq()
365 sz = sizeof(struct ib_wc) * (cqe + 1); in rvt_resize_cq()
395 if (head > (u32)cq->ibcq.cqe) in rvt_resize_cq()
[all...]
/linux/drivers/infiniband/hw/bnxt_re/
H A Dqplib_fp.c1444 struct cq_req *cqe = (struct cq_req *)hw_cqe;
1446 if (qp == le64_to_cpu(cqe->qp_handle)) in bnxt_qplib_query_qp()
1447 cqe->qp_handle = 0; in bnxt_qplib_query_qp()
1454 struct cq_res_rc *cqe = (struct cq_res_rc *)hw_cqe; in bnxt_qplib_query_qp()
1456 if (qp == le64_to_cpu(cqe->qp_handle)) in bnxt_qplib_query_qp()
1457 cqe->qp_handle = 0; in bnxt_qplib_query_qp()
2260 struct bnxt_qplib_cqe *cqe; in bnxt_qplib_create_cq()
2266 cqe = *pcqe; in bnxt_qplib_create_cq()
2276 memset(cqe, 0, sizeof(*cqe)); in bnxt_qplib_create_cq()
1557 struct cq_req *cqe = (struct cq_req *)hw_cqe; __clean_cq() local
1567 struct cq_res_rc *cqe = (struct cq_res_rc *)hw_cqe; __clean_cq() local
2373 struct bnxt_qplib_cqe *cqe; __flush_sq() local
2414 struct bnxt_qplib_cqe *cqe; __flush_rq() local
2603 struct bnxt_qplib_cqe *cqe; bnxt_qplib_cq_process_req() local
2729 struct bnxt_qplib_cqe *cqe; bnxt_qplib_cq_process_res_rc() local
2809 struct bnxt_qplib_cqe *cqe; bnxt_qplib_cq_process_res_ud() local
2909 struct bnxt_qplib_cqe *cqe; bnxt_qplib_cq_process_res_raweth_qp1() local
3001 struct bnxt_qplib_cqe *cqe; bnxt_qplib_cq_process_terminal() local
3113 bnxt_qplib_process_flush_list(struct bnxt_qplib_cq * cq,struct bnxt_qplib_cqe * cqe,int num_cqes) bnxt_qplib_process_flush_list() argument
3135 bnxt_qplib_poll_cq(struct bnxt_qplib_cq * cq,struct bnxt_qplib_cqe * cqe,int num_cqes,struct bnxt_qplib_qp ** lib_qp) bnxt_qplib_poll_cq() argument
[all...]
/linux/io_uring/
H A Dfdinfo.c156 struct io_uring_cqe *cqe; in __io_uring_show_fdinfo() local
159 cqe = &r->cqes[(cq_head & cq_mask)]; in __io_uring_show_fdinfo()
160 if (cqe->flags & IORING_CQE_F_32 || ctx->flags & IORING_SETUP_CQE32) in __io_uring_show_fdinfo()
163 cq_head & cq_mask, cqe->user_data, cqe->res, in __io_uring_show_fdinfo()
164 cqe->flags); in __io_uring_show_fdinfo()
167 cqe->big_cqe[0], cqe->big_cqe[1]); in __io_uring_show_fdinfo()
244 struct io_uring_cqe *cqe = &ocqe->cqe; in __io_uring_show_fdinfo() local
247 cqe->user_data, cqe->res, cqe->flags); in __io_uring_show_fdinfo()
H A Dcmd_net.c76 struct io_uring_cqe cqe[2]; in io_process_timestamp_skb()
91 cqe->user_data = 0; in io_uring_cmd_timestamp()
92 cqe->res = tskey; in io_uring_cmd_timestamp()
93 cqe->flags = IORING_CQE_F_MORE | ctx_cqe32_flags(cmd_to_io_kiocb(cmd)->ctx); in io_uring_cmd_timestamp()
94 cqe->flags |= tstype << IORING_TIMESTAMP_TYPE_SHIFT; in io_uring_cmd_timestamp()
96 cqe->flags |= IORING_CQE_F_TSTAMP_HW; in io_uring_cmd_timestamp()
98 iots = (struct io_timespec *)&cqe[1]; in io_uring_cmd_timestamp()
101 return io_uring_cmd_post_mshot_cqe32(cmd, issue_flags, cqe); in io_uring_cmd_timestamp()
61 struct io_uring_cqe cqe[2]; io_process_timestamp_skb() local
/linux/drivers/net/ethernet/qlogic/qede/
H A Dqede_fp.c651 struct eth_fast_path_rx_tpa_start_cqe *cqe) in qede_set_gro_params() argument
653 u16 parsing_flags = le16_to_cpu(cqe->pars_flags.flags); in qede_set_gro_params()
661 skb_shinfo(skb)->gso_size = __le16_to_cpu(cqe->len_on_first_bd) - in qede_set_gro_params()
662 cqe->header_len; in qede_set_gro_params()
834 struct eth_fast_path_rx_tpa_start_cqe *cqe) in qede_tpa_start() argument
836 struct qede_agg_info *tpa_info = &rxq->tpa_info[cqe->tpa_agg_index]; in qede_tpa_start()
841 pad = cqe->placement_offset + rxq->rx_headroom; in qede_tpa_start()
844 le16_to_cpu(cqe->len_on_first_bd), in qede_tpa_start()
865 if ((le16_to_cpu(cqe->pars_flags.flags) >> in qede_tpa_start()
868 tpa_info->vlan_tag = le16_to_cpu(cqe->vlan_tag); in qede_tpa_start()
[all …]
/linux/drivers/net/ethernet/marvell/octeontx2/nic/
H A Dotx2_txrx.c42 struct nix_cqe_rx_s *cqe,
107 struct nix_cqe_tx_s *cqe, in otx2_xdp_snd_pkt_handler() argument
110 struct nix_send_comp_s *snd_comp = &cqe->comp; in otx2_xdp_snd_pkt_handler()
128 struct nix_cqe_tx_s *cqe, in otx2_snd_pkt_handler() argument
131 struct nix_send_comp_s *snd_comp = &cqe->comp; in otx2_snd_pkt_handler()
225 struct nix_cqe_rx_s *cqe, struct sk_buff *skb) in otx2_set_rxhash() argument
241 hash = cqe->hdr.flow_tag; in otx2_set_rxhash()
246 static void otx2_free_rcv_seg(struct otx2_nic *pfvf, struct nix_cqe_rx_s *cqe, in otx2_free_rcv_seg() argument
249 struct nix_rx_sg_s *sg = &cqe->sg; in otx2_free_rcv_seg()
255 end = start + ((cqe->parse.desc_sizem1 + 1) * 16); in otx2_free_rcv_seg()
[all …]
/linux/drivers/infiniband/ulp/iser/
H A Discsi_iser.h244 struct ib_cqe cqe; member
271 struct ib_cqe cqe; member
291 struct ib_cqe cqe; member
552 iser_rx(struct ib_cqe *cqe) in iser_rx() argument
554 return container_of(cqe, struct iser_rx_desc, cqe); in iser_rx()
558 iser_tx(struct ib_cqe *cqe) in iser_tx() argument
560 return container_of(cqe, struct iser_tx_desc, cqe); in iser_tx()
564 iser_login(struct ib_cqe *cqe) in iser_login() argument
566 return container_of(cqe, struct iser_login_desc, cqe); in iser_login()
/linux/drivers/net/ethernet/huawei/hinic/
H A Dhinic_hw_qp.c324 cqe_size = wq->q_depth * sizeof(*rq->cqe); in alloc_rq_cqe()
325 rq->cqe = vzalloc(cqe_size); in alloc_rq_cqe()
326 if (!rq->cqe) in alloc_rq_cqe()
335 rq->cqe[i] = dma_alloc_coherent(&pdev->dev, in alloc_rq_cqe()
336 sizeof(*rq->cqe[i]), in alloc_rq_cqe()
338 if (!rq->cqe[i]) in alloc_rq_cqe()
346 dma_free_coherent(&pdev->dev, sizeof(*rq->cqe[j]), rq->cqe[j], in alloc_rq_cqe()
352 vfree(rq->cqe); in alloc_rq_cqe()
368 dma_free_coherent(&pdev->dev, sizeof(*rq->cqe[i]), rq->cqe[i], in free_rq_cqe()
372 vfree(rq->cqe); in free_rq_cqe()
[all …]
/linux/drivers/scsi/qedi/
H A Dqedi_fw.c31 union iscsi_cqe *cqe, in qedi_process_logout_resp() argument
42 cqe_logout_response = &cqe->cqe_common.iscsi_hdr.logout_response; in qedi_process_logout_resp()
50 resp_hdr->itt = build_itt(cqe->cqe_solicited.itid, conn->session->age); in qedi_process_logout_resp()
82 union iscsi_cqe *cqe, in qedi_process_text_resp() argument
97 cqe_text_response = &cqe->cqe_common.iscsi_hdr.text_response; in qedi_process_text_resp()
109 resp_hdr_ptr->itt = build_itt(cqe->cqe_solicited.itid, in qedi_process_text_resp()
178 union iscsi_cqe *cqe, in qedi_process_tmf_resp() argument
190 cqe_tmp_response = &cqe->cqe_common.iscsi_hdr.tmf_response; in qedi_process_tmf_resp()
214 resp_hdr_ptr->itt = build_itt(cqe->cqe_solicited.itid, in qedi_process_tmf_resp()
258 union iscsi_cqe *cqe, in qedi_process_login_resp() argument
[all …]
/linux/drivers/net/ethernet/mellanox/mlx5/core/en_accel/
H A Dmacsec.h32 static inline bool mlx5e_macsec_is_rx_flow(struct mlx5_cqe64 *cqe) in mlx5e_macsec_is_rx_flow() argument
34 return MLX5_MACSEC_METADATA_MARKER(be32_to_cpu(cqe->ft_metadata)); in mlx5e_macsec_is_rx_flow()
38 struct mlx5_cqe64 *cqe);
46 static inline bool mlx5e_macsec_is_rx_flow(struct mlx5_cqe64 *cqe) { return false; } in mlx5e_macsec_is_rx_flow() argument
49 struct mlx5_cqe64 *cqe) in mlx5e_macsec_offload_handle_rx_skb() argument
/linux/drivers/infiniband/hw/ocrdma/
H A Docrdma_verbs.c972 int entries = attr->cqe; in ocrdma_create_cq()
1025 ibcq->cqe = new_cnt; in ocrdma_resize_cq()
1036 struct ocrdma_cqe *cqe = NULL; in ocrdma_flush_cq()
1038 cqe = cq->va; in ocrdma_flush_cq() local
1046 if (is_cqe_valid(cq, cqe)) in ocrdma_flush_cq()
1048 cqe++; in ocrdma_flush_cq()
1404 /* syncronize with wqe, rqe posting and cqe processing contexts */ in ocrdma_modify_qp()
1590 /* discard the cqe for a given QP */
1596 struct ocrdma_cqe *cqe; in ocrdma_discard_cqes() local
1605 * we don't complete out of order cqe in ocrdma_discard_cqes()
2441 ocrdma_set_cqe_status_flushed(struct ocrdma_qp * qp,struct ocrdma_cqe * cqe) ocrdma_set_cqe_status_flushed() argument
2472 ocrdma_update_err_cqe(struct ib_wc * ibwc,struct ocrdma_cqe * cqe,struct ocrdma_qp * qp,int status) ocrdma_update_err_cqe() argument
2494 ocrdma_update_err_rcqe(struct ib_wc * ibwc,struct ocrdma_cqe * cqe,struct ocrdma_qp * qp,int status) ocrdma_update_err_rcqe() argument
2504 ocrdma_update_err_scqe(struct ib_wc * ibwc,struct ocrdma_cqe * cqe,struct ocrdma_qp * qp,int status) ocrdma_update_err_scqe() argument
2515 ocrdma_poll_err_scqe(struct ocrdma_qp * qp,struct ocrdma_cqe * cqe,struct ib_wc * ibwc,bool * polled,bool * stop) ocrdma_poll_err_scqe() argument
2559 ocrdma_poll_success_scqe(struct ocrdma_qp * qp,struct ocrdma_cqe * cqe,struct ib_wc * ibwc,bool * polled) ocrdma_poll_success_scqe() argument
2584 ocrdma_poll_scqe(struct ocrdma_qp * qp,struct ocrdma_cqe * cqe,struct ib_wc * ibwc,bool * polled,bool * stop) ocrdma_poll_scqe() argument
2601 ocrdma_update_ud_rcqe(struct ocrdma_dev * dev,struct ib_wc * ibwc,struct ocrdma_cqe * cqe) ocrdma_update_ud_rcqe() argument
2628 ocrdma_update_free_srq_cqe(struct ib_wc * ibwc,struct ocrdma_cqe * cqe,struct ocrdma_qp * qp) ocrdma_update_free_srq_cqe() argument
2647 ocrdma_poll_err_rcqe(struct ocrdma_qp * qp,struct ocrdma_cqe * cqe,struct ib_wc * ibwc,bool * polled,bool * stop,int status) ocrdma_poll_err_rcqe() argument
2683 ocrdma_poll_success_rcqe(struct ocrdma_qp * qp,struct ocrdma_cqe * cqe,struct ib_wc * ibwc) ocrdma_poll_success_rcqe() argument
2716 ocrdma_poll_rcqe(struct ocrdma_qp * qp,struct ocrdma_cqe * cqe,struct ib_wc * ibwc,bool * polled,bool * stop) ocrdma_poll_rcqe() argument
2742 ocrdma_change_cq_phase(struct ocrdma_cq * cq,struct ocrdma_cqe * cqe,u16 cur_getp) ocrdma_change_cq_phase() argument
2763 struct ocrdma_cqe *cqe; ocrdma_poll_hwcq() local
[all...]
/linux/drivers/net/ethernet/mellanox/mlx4/
H A Den_rx.c596 static int check_csum(struct mlx4_cqe *cqe, struct sk_buff *skb, void *va, in check_csum() argument
614 hw_checksum = csum_unfold((__force __sum16)cqe->checksum); in check_csum()
616 if (cqe->vlan_my_qpn & cpu_to_be32(MLX4_CQE_CVLAN_PRESENT_MASK) && in check_csum()
623 if (cqe->status & cpu_to_be16(MLX4_CQE_STATUS_IPV6)) in check_csum()
637 struct mlx4_cqe *cqe; member
651 mlx4_en_get_cqe_ts(_ctx->cqe)); in mlx4_en_xdp_rx_timestamp()
659 struct mlx4_cqe *cqe = _ctx->cqe; in mlx4_en_xdp_rx_hash() local
666 *hash = be32_to_cpu(cqe->immed_rss_invalid); in mlx4_en_xdp_rx_hash()
667 status = cqe->status; in mlx4_en_xdp_rx_hash()
676 if (cqe->ipv6_ext_mask) in mlx4_en_xdp_rx_hash()
[all …]
/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/
H A Dsend.c349 struct mlx5_cqe64 *cqe) in hws_send_engine_dump_error_cqe() argument
351 u8 wqe_opcode = cqe ? be32_to_cpu(cqe->sop_drop_qpn) >> 24 : 0; in hws_send_engine_dump_error_cqe()
353 u32 opcode = cqe ? get_cqe_opcode(cqe) : 0; in hws_send_engine_dump_error_cqe()
396 if (!cqe) { in hws_send_engine_dump_error_cqe()
408 struct mlx5_err_cqe *err_cqe = (struct mlx5_err_cqe *)cqe; in hws_send_engine_dump_error_cqe()
430 HWS_PTR_TO_ID(rule), be32_to_cpu(cqe->byte_cnt)); in hws_send_engine_dump_error_cqe()
434 (be32_to_cpu(cqe->byte_cnt) & 0x80000000) ? in hws_send_engine_dump_error_cqe()
439 ((be32_to_cpu(cqe->byte_cnt) & 0x00000003) == 1) ? in hws_send_engine_dump_error_cqe()
441 ((be32_to_cpu(cqe->byte_cnt) & 0x00000003) == 2) ? in hws_send_engine_dump_error_cqe()
445 HWS_PTR_TO_ID(rule), be32_to_cpu(cqe->sop_drop_qpn)); in hws_send_engine_dump_error_cqe()
[all …]
/linux/include/trace/events/
H A Dio_uring.h124 __entry->user_data = req->cqe.user_data;
161 __entry->user_data = req->cqe.user_data;
202 __entry->data = req->cqe.user_data;
302 __entry->user_data = req->cqe.user_data;
323 TP_PROTO(struct io_ring_ctx *ctx, void *req, struct io_uring_cqe *cqe),
325 TP_ARGS(ctx, req, cqe),
340 __entry->user_data = cqe->user_data;
341 __entry->res = cqe->res;
342 __entry->cflags = cqe->flags;
343 …__entry->extra1 = ctx->flags & IORING_SETUP_CQE32 || cqe->flags & IORING_CQE_F_32 ? cqe->big_cqe[…
[all …]

123456789