/linux/drivers/infiniband/sw/rxe/ |
H A D | rxe_req.c | 17 struct rxe_send_wqe *wqe, int npsn) in retry_first_write_send() 39 struct rxe_send_wqe *wqe; in req_retry() local 120 struct rxe_send_wqe *wqe; in req_check_sq_drain_done() local 176 struct rxe_send_wqe *wqe; in req_next_wqe() local 205 static int rxe_wqe_is_fenced(struct rxe_qp *qp, struct rxe_send_wqe *wqe) in rxe_wqe_is_fenced() 353 static int next_opcode(struct rxe_qp *qp, struct rxe_send_wqe *wqe, in next_opcode() 383 static inline int check_init_depth(struct rxe_qp *qp, struct rxe_send_wqe *wqe) in check_init_depth() 415 struct rxe_send_wqe *wqe, in init_req_packet() 499 struct rxe_send_wqe *wqe, struct rxe_pkt_info *pkt, in finish_packet() 542 struct rxe_send_wqe *wqe, in update_wqe_state() [all …]
|
H A D | rxe_comp.c | 141 struct rxe_send_wqe *wqe; in get_wqe() local 174 struct rxe_send_wqe *wqe) in check_psn() 217 struct rxe_send_wqe *wqe) in check_ack() 356 struct rxe_send_wqe *wqe) in do_read() 376 struct rxe_send_wqe *wqe) in do_atomic() 393 static void make_send_cqe(struct rxe_qp *qp, struct rxe_send_wqe *wqe, in make_send_cqe() 440 static void do_complete(struct rxe_qp *qp, struct rxe_send_wqe *wqe) in do_complete() 501 struct rxe_send_wqe *wqe) in complete_ack() 525 struct rxe_send_wqe *wqe) in complete_wqe() 557 static int flush_send_wqe(struct rxe_qp *qp, struct rxe_send_wqe *wqe) in flush_send_wqe() [all …]
|
H A D | rxe_mw.c | 50 static int rxe_check_bind_mw(struct rxe_qp *qp, struct rxe_send_wqe *wqe, in rxe_check_bind_mw() 135 static void rxe_do_bind_mw(struct rxe_qp *qp, struct rxe_send_wqe *wqe, in rxe_do_bind_mw() 164 int rxe_bind_mw(struct rxe_qp *qp, struct rxe_send_wqe *wqe) in rxe_bind_mw()
|
H A D | rxe_resp.c | 263 struct rxe_recv_wqe *wqe; in get_srq_wqe() local 1075 struct rxe_recv_wqe *wqe = qp->resp.wqe; in do_complete() local 1444 static int flush_recv_wqe(struct rxe_qp *qp, struct rxe_recv_wqe *wqe) in flush_recv_wqe() 1475 struct rxe_recv_wqe *wqe; in flush_recv_queue() local
|
/linux/drivers/infiniband/hw/irdma/ |
H A D | ctrl.c | 186 __le64 *wqe; in irdma_sc_add_arp_cache_entry() local 222 __le64 *wqe; in irdma_sc_del_arp_cache_entry() local 256 __le64 *wqe; in irdma_sc_manage_apbvt_entry() local 304 __le64 *wqe; in irdma_sc_manage_qhash_table_entry() local 452 __le64 *wqe; in irdma_sc_qp_create() local 502 __le64 *wqe; in irdma_sc_qp_modify() local 572 __le64 *wqe; in irdma_sc_qp_destroy() local 764 __le64 *wqe; in irdma_sc_alloc_local_mac_entry() local 799 __le64 *wqe; in irdma_sc_add_local_mac_entry() local 837 __le64 *wqe; in irdma_sc_del_local_mac_entry() local [all …]
|
H A D | uk.c | 15 static void irdma_set_fragment(__le64 *wqe, u32 offset, struct ib_sge *sge, in irdma_set_fragment() 39 static void irdma_set_fragment_gen_1(__le64 *wqe, u32 offset, in irdma_set_fragment_gen_1() 61 __le64 *wqe; in irdma_nop_1() local 158 __le64 *wqe; in irdma_qp_get_next_send_wqe() local 208 __le64 *wqe; in irdma_qp_get_next_recv_wqe() local 236 __le64 *wqe; in irdma_uk_rdma_write() local 335 __le64 *wqe; in irdma_uk_rdma_read() local 410 __le64 *wqe; in irdma_uk_send() local 501 static void irdma_set_mw_bind_wqe_gen_1(__le64 *wqe, in irdma_set_mw_bind_wqe_gen_1() 518 static void irdma_copy_inline_data_gen_1(u8 *wqe, struct ib_sge *sge_list, in irdma_copy_inline_data_gen_1() [all …]
|
H A D | uda.c | 23 __le64 *wqe; in irdma_sc_access_ah() local 119 __le64 *wqe; in irdma_access_mcast_grp() local
|
/linux/drivers/infiniband/sw/siw/ |
H A D | siw_qp_rx.c | 169 struct siw_wqe *wqe = &frx->wqe_active; in siw_rresp_check_ntoh() local 281 struct siw_wqe *wqe = &frx->wqe_active; in siw_send_check_ntoh() local 334 struct siw_wqe *wqe = NULL; in siw_rqe_get() local 439 struct siw_wqe *wqe; in siw_proc_send() local 565 struct siw_wqe *wqe = rx_wqe(frx); in siw_proc_write() local 742 struct siw_wqe *wqe = NULL; in siw_orqe_start_rx() local 786 struct siw_wqe *wqe = rx_wqe(frx); in siw_proc_rresp() local 1201 struct siw_wqe *wqe = rx_wqe(qp->rx_fpdu); in siw_rdmap_complete() local
|
H A D | siw_qp_tx.c | 51 struct siw_wqe *wqe = &c_tx->wqe_active; in siw_try_1seg() local 119 struct siw_wqe *wqe = &c_tx->wqe_active; in siw_qp_prepare_tx() local 440 struct siw_wqe *wqe = &c_tx->wqe_active; in siw_tx_hdt() local 706 static void siw_prepare_fpdu(struct siw_qp *qp, struct siw_wqe *wqe) in siw_prepare_fpdu() 763 static int siw_check_sgl_tx(struct ib_pd *pd, struct siw_wqe *wqe, in siw_check_sgl_tx() 793 static int siw_qp_sq_proc_tx(struct siw_qp *qp, struct siw_wqe *wqe) in siw_qp_sq_proc_tx() 978 static int siw_qp_sq_proc_local(struct siw_qp *qp, struct siw_wqe *wqe) in siw_qp_sq_proc_local() 1025 struct siw_wqe *wqe = tx_wqe(qp); in siw_qp_sq_process() local
|
/linux/drivers/infiniband/hw/hfi1/ |
H A D | rc.c | 394 struct rvt_swqe *wqe; in hfi1_make_rc_req() local 1418 struct rvt_swqe *wqe) in update_num_rd_atomic() 1461 struct rvt_swqe *wqe = rvt_get_swqe_ptr(qp, n); in reset_psn() local 1567 struct rvt_swqe *wqe = rvt_get_swqe_ptr(qp, qp->s_acked); in hfi1_restart_rc() local 1637 struct rvt_swqe *wqe; in reset_sending_psn() local 1697 struct rvt_swqe *wqe; in hfi1_rc_send_complete() local 1825 struct rvt_swqe *wqe, in do_rc_completion() 1970 struct rvt_swqe *wqe; in do_rc_ack() local 2266 struct rvt_swqe *wqe; in rdma_seq_err() local 2310 struct rvt_swqe *wqe; in rc_rcv_resp() local
|
H A D | tid_rdma.c | 378 struct rvt_swqe *wqe = rvt_get_swqe_ptr(qp, i); in hfi1_qp_priv_init() local 416 struct rvt_swqe *wqe; in hfi1_qp_priv_tid_free() local 1621 void __trdma_clean_swqe(struct rvt_qp *qp, struct rvt_swqe *wqe) in __trdma_clean_swqe() 1703 u32 hfi1_build_tid_rdma_read_packet(struct rvt_swqe *wqe, in hfi1_build_tid_rdma_read_packet() 1787 u32 hfi1_build_tid_rdma_read_req(struct rvt_qp *qp, struct rvt_swqe *wqe, in hfi1_build_tid_rdma_read_req() 2418 struct rvt_swqe *wqe; in find_tid_request() local 2582 struct rvt_swqe *wqe; in hfi1_kern_read_tid_flow_free() local 2630 struct rvt_qp *qp, struct rvt_swqe *wqe) in restart_tid_rdma_read_req() 2663 struct rvt_swqe *wqe; in handle_read_kdeth_eflags() local 3046 void hfi1_tid_rdma_restart_req(struct rvt_qp *qp, struct rvt_swqe *wqe, in hfi1_tid_rdma_restart_req() [all …]
|
H A D | ud.c | 224 static void hfi1_make_bth_deth(struct rvt_qp *qp, struct rvt_swqe *wqe, in hfi1_make_bth_deth() 262 struct rvt_swqe *wqe) in hfi1_make_ud_req_9B() 333 struct rvt_swqe *wqe) in hfi1_make_ud_req_16B() 441 struct rvt_swqe *wqe; in hfi1_make_ud_req() local
|
H A D | rc.h | 35 static inline u32 restart_sge(struct rvt_sge_state *ss, struct rvt_swqe *wqe, in restart_sge()
|
/linux/drivers/infiniband/sw/rdmavt/ |
H A D | qp.c | 591 struct rvt_swqe *wqe = rvt_get_swqe_ptr(qp, qp->s_last); in rvt_clear_mr_refs() local 621 static bool rvt_swqe_has_lkey(struct rvt_swqe *wqe, u32 lkey) in rvt_swqe_has_lkey() 644 struct rvt_swqe *wqe = rvt_get_swqe_ptr(qp, s_last); in rvt_qp_sends_has_lkey() local 979 struct rvt_swqe *wqe; in free_ud_wq_attr() local 999 struct rvt_swqe *wqe; in alloc_ud_wq_attr() local 1796 struct rvt_rwqe *wqe; in rvt_post_recv() local 1963 struct rvt_swqe *wqe; in rvt_post_one_wr() local 2211 struct rvt_rwqe *wqe; in rvt_post_srq_recv() local 2266 static int init_sge(struct rvt_qp *qp, struct rvt_rwqe *wqe) in init_sge() 2349 struct rvt_rwqe *wqe; in rvt_get_rwqe() local [all …]
|
H A D | rc.c | 162 u32 rvt_restart_sge(struct rvt_sge_state *ss, struct rvt_swqe *wqe, u32 len) in rvt_restart_sge()
|
/linux/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_hw_cmdq.c | 54 #define CMDQ_WQE_HEADER(wqe) ((struct hinic_cmdq_header *)(wqe)) argument 177 static void cmdq_prepare_wqe_ctrl(struct hinic_cmdq_wqe *wqe, int wrapped, in cmdq_prepare_wqe_ctrl() 234 static void cmdq_set_direct_wqe_data(struct hinic_cmdq_direct_wqe *wqe, in cmdq_set_direct_wqe_data() 243 static void cmdq_set_lcmd_wqe(struct hinic_cmdq_wqe *wqe, in cmdq_set_lcmd_wqe() 271 static void cmdq_set_direct_wqe(struct hinic_cmdq_wqe *wqe, in cmdq_set_direct_wqe() 529 struct hinic_cmdq_wqe *wqe) in clear_wqe_complete_bit() 563 struct hinic_cmdq_wqe *wqe) in cmdq_arm_ceq_handler()
|
H A D | hinic_hw_wq.c | 73 #define WQE_IN_RANGE(wqe, start, end) \ argument 77 #define WQE_SHADOW_PAGE(wq, wqe) \ argument 874 static inline bool wqe_shadow(struct hinic_wq *wq, struct hinic_hw_wqe *wqe) in wqe_shadow() 888 void hinic_write_wqe(struct hinic_wq *wq, struct hinic_hw_wqe *wqe, in hinic_write_wqe()
|
/linux/drivers/infiniband/hw/mthca/ |
H A D | mthca_srq.c | 92 static inline int *wqe_to_link(void *wqe) in wqe_to_link() 158 void *wqe; in mthca_alloc_srq_buf() local 495 void *wqe; in mthca_tavor_post_srq_recv() local 588 void *wqe; in mthca_arbel_post_srq_recv() local
|
H A D | mthca_cq.c | 126 __be32 wqe; member 140 __be32 wqe; member 541 u32 wqe = be32_to_cpu(cqe->wqe); in mthca_poll_one() local 547 s32 wqe; in mthca_poll_one() local
|
/linux/drivers/infiniband/hw/bnxt_re/ |
H A D | ib_verbs.c | 469 struct bnxt_qplib_swqe *wqe = &fence->bind_wqe; in bnxt_re_create_fence_wqe() local 501 struct bnxt_qplib_swqe wqe; in bnxt_re_bind_fence_mw() local 1975 struct bnxt_qplib_swqe wqe; in bnxt_re_post_srq_recv() local 2341 struct bnxt_qplib_swqe *wqe, in bnxt_re_build_qp1_send_v2() 2515 struct bnxt_qplib_swqe *wqe, in bnxt_re_build_qp1_shadow_qp_recv() 2564 struct bnxt_qplib_swqe *wqe) in bnxt_re_build_send_wqe() 2602 struct bnxt_qplib_swqe *wqe) in bnxt_re_build_rdma_wqe() 2634 struct bnxt_qplib_swqe *wqe) in bnxt_re_build_atomic_wqe() 2661 struct bnxt_qplib_swqe *wqe) in bnxt_re_build_inv_wqe() 2675 struct bnxt_qplib_swqe *wqe) in bnxt_re_build_reg_wqe() [all …]
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
H A D | ktls_txrx.c | 74 mlx5e_ktls_build_static_params(struct mlx5e_set_tls_static_params_wqe *wqe, in mlx5e_ktls_build_static_params() 117 mlx5e_ktls_build_progress_params(struct mlx5e_set_tls_progress_params_wqe *wqe, in mlx5e_ktls_build_progress_params()
|
/linux/drivers/scsi/lpfc/ |
H A D | lpfc_nvme.c | 399 union lpfc_wqe128 *wqe; in lpfc_nvme_gen_req() local 711 struct lpfc_iocbq *wqe, *next_wqe; in __lpfc_nvme_ls_abort() local 831 union lpfc_wqe128 *wqe; in lpfc_nvme_adj_fcp_sgls() local 1216 union lpfc_wqe128 *wqe = &pwqeq->wqe; in lpfc_nvme_prep_io_cmd() local 1342 union lpfc_wqe128 *wqe = &lpfc_ncmd->cur_iocbq.wqe; in lpfc_nvme_prep_io_dma() local 2043 union lpfc_wqe128 *wqe; in lpfc_get_nvme_buf() local
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/ |
H A D | rx.c | 183 struct mlx5e_rx_wqe_cyc *wqe; in mlx5e_xsk_alloc_rx_wqes_batched() local 206 struct mlx5e_rx_wqe_cyc *wqe; in mlx5e_xsk_alloc_rx_wqes() local
|
/linux/drivers/scsi/bnx2fc/ |
H A D | bnx2fc_hwi.c | 626 static void bnx2fc_process_unsol_compl(struct bnx2fc_rport *tgt, u16 wqe) in bnx2fc_process_unsol_compl() 859 void bnx2fc_process_cq_compl(struct bnx2fc_rport *tgt, u16 wqe, in bnx2fc_process_cq_compl() 970 static struct bnx2fc_work *bnx2fc_alloc_work(struct bnx2fc_rport *tgt, u16 wqe, in bnx2fc_alloc_work() 991 static bool bnx2fc_pending_work(struct bnx2fc_rport *tgt, unsigned int wqe) in bnx2fc_pending_work() 1071 u16 wqe; in bnx2fc_process_new_cqes() local
|
/linux/drivers/infiniband/hw/mlx5/ |
H A D | odp.c | 68 } wqe; member 1171 void *wqe, in pagefault_data_segments() 1243 struct mlx5_ib_qp *qp, void **wqe, void **wqe_end, int wqe_length) in mlx5_ib_mr_initiator_pfault_handler() 1302 void **wqe, void **wqe_end, in mlx5_ib_mr_responder_pfault_handler_srq() 1320 void *wqe, void **wqe_end, in mlx5_ib_mr_responder_pfault_handler_rq() 1385 void *wqe, *wqe_start = NULL, *wqe_end = NULL; in mlx5_ib_mr_wqe_pfault_handler() local
|