Home
last modified time | relevance | path

Searched refs:wqe_cnt (Results 1 – 23 of 23) sorted by relevance

/linux/drivers/infiniband/hw/mlx5/
H A Dqp.c196 wq->offset, wq->wqe_cnt, in mlx5_ib_read_user_wqe_sq()
222 wq->wqe_cnt, wq->wqe_shift, in mlx5_ib_read_user_wqe_sq()
258 wq->offset, wq->wqe_cnt, in mlx5_ib_read_user_wqe_rq()
445 qp->rq.wqe_cnt = 0; in set_rq_size()
453 qp->rq.wqe_cnt = ucmd->rq_wqe_count; in set_rq_size()
465 qp->rq.max_post = qp->rq.wqe_cnt; in set_rq_size()
474 qp->rq.wqe_cnt = wq_size / wqe_size; in set_rq_size()
487 qp->rq.max_post = qp->rq.wqe_cnt; in set_rq_size()
616 qp->sq.wqe_cnt = wq_size / MLX5_SEND_WQE_BB; in calc_sq_size()
617 if (qp->sq.wqe_cnt > (1 << MLX5_CAP_GEN(dev->mdev, log_max_qp_sz))) { in calc_sq_size()
[all …]
H A Dsrq.h21 u32 wqe_cnt; member
H A Dwr.h59 idx = (sq->cur_post + (wqe_sz >> 2)) & (sq->wqe_cnt - 1); in handle_post_send_edge()
H A Dcq.c200 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in handle_responder()
350 idx = tail & (qp->sq.wqe_cnt - 1); in handle_atomics()
416 idx &= (wq->wqe_cnt - 1); in sw_comp()
507 idx = wqe_ctr & (wq->wqe_cnt - 1); in mlx5_poll_one()
539 idx = wqe_ctr & (wq->wqe_cnt - 1); in mlx5_poll_one()
552 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in mlx5_poll_one()
H A Dwr.c716 idx = (idx + 1) & (qp->sq.wqe_cnt - 1); in dump_wqe()
732 *idx = qp->sq.cur_post & (qp->sq.wqe_cnt - 1); in mlx5r_begin_wqe()
783 (qp->sq.wqe_cnt - 1)) : in mlx5r_finish_wqe()
1231 ind = qp->rq.head & (qp->rq.wqe_cnt - 1); in mlx5_ib_post_recv()
1266 ind = (ind + 1) & (qp->rq.wqe_cnt - 1); in mlx5_ib_post_recv()
/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/
H A Dsend.c254 u16 wqe_cnt) in hws_send_engine_retry_post_send() argument
280 idx = wqe_cnt & send_sq->buf_mask; in hws_send_engine_retry_post_send()
287 idx = (wqe_cnt + 1) & send_sq->buf_mask; in hws_send_engine_retry_post_send()
459 u16 wqe_cnt, in hws_send_engine_update_rule() argument
470 hws_send_engine_retry_post_send(queue, priv, wqe_cnt); in hws_send_engine_update_rule()
525 u16 wqe_cnt) in hws_send_engine_update() argument
538 hws_send_engine_update_rule(queue, priv, wqe_cnt, in hws_send_engine_update()
608 u16 wqe_cnt; in hws_send_engine_poll_cq() local
621 wqe_cnt = be16_to_cpu(cqe->wqe_counter) & sq->buf_mask; in hws_send_engine_poll_cq()
623 while (cq->poll_wqe != wqe_cnt) { in hws_send_engine_poll_cq()
[all …]
/linux/drivers/infiniband/hw/hns/
H A Dhns_roce_srq.c176 buf_attr.region[0].size = to_hr_hem_entries_size(srq->wqe_cnt, in alloc_srq_idx()
191 idx_que->bitmap = bitmap_zalloc(srq->wqe_cnt, GFP_KERNEL); in alloc_srq_idx()
231 buf_attr.region[0].size = to_hr_hem_entries_size(srq->wqe_cnt, in alloc_srq_wqe_buf()
254 srq->wrid = kvmalloc_array(srq->wqe_cnt, sizeof(u64), GFP_KERNEL); in alloc_srq_wrid()
307 srq->wqe_cnt = roundup_pow_of_two(attr->max_wr); in set_srq_basic_param()
310 attr->max_wr = srq->wqe_cnt; in set_srq_basic_param()
H A Dhns_roce_hw_v2.c726 wqe_idx = (qp->sq.head + nreq) & (qp->sq.wqe_cnt - 1); in hns_roce_v2_post_send()
739 ~(((qp->sq.head + nreq) >> ilog2(qp->sq.wqe_cnt)) & 0x1); in hns_roce_v2_post_send()
863 wqe_idx = (hr_qp->rq.head + nreq) & (hr_qp->rq.wqe_cnt - 1); in hns_roce_v2_post_recv()
891 wq->wrid[wq->head & (wq->wqe_cnt - 1)] = wr_id; in hns_roce_push_drain_wr()
1069 return idx_que->head - idx_que->tail >= srq->wqe_cnt; in hns_roce_srqwq_overflow()
1098 pos = find_first_zero_bit(idx_que->bitmap, srq->wqe_cnt); in get_srq_wqe_idx()
1099 if (unlikely(pos == srq->wqe_cnt)) in get_srq_wqe_idx()
1113 head = idx_que->head & (srq->wqe_cnt - 1); in fill_wqe_idx()
3967 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in sw_comp()
4251 (wq->wqe_cnt - 1); in hns_roce_v2_poll_one()
[all …]
/linux/drivers/infiniband/hw/mlx4/
H A Dqp.c202 buf = get_send_wqe(qp, n & (qp->sq.wqe_cnt - 1)); in stamp_send_wqe()
349 qp->rq.wqe_cnt = qp->rq.max_gs = 0; in set_rq_size()
360 qp->rq.wqe_cnt = roundup_pow_of_two(max(1U, cap->max_recv_wr)); in set_rq_size()
368 cap->max_recv_wr = qp->rq.max_post = qp->rq.wqe_cnt; in set_rq_size()
372 min(dev->dev->caps.max_wqes - MLX4_IB_SQ_MAX_SPARE, qp->rq.wqe_cnt); in set_rq_size()
416 qp->sq.wqe_cnt = roundup_pow_of_two(cap->max_send_wr + in set_kernel_sq_size()
425 qp->buf_size = (qp->rq.wqe_cnt << qp->rq.wqe_shift) + in set_kernel_sq_size()
426 (qp->sq.wqe_cnt << qp->sq.wqe_shift); in set_kernel_sq_size()
429 qp->sq.offset = qp->rq.wqe_cnt << qp->rq.wqe_shift; in set_kernel_sq_size()
431 qp->rq.offset = qp->sq.wqe_cnt << qp->sq.wqe_shift; in set_kernel_sq_size()
[all …]
H A Dcq.c626 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in mlx4_ib_qp_sw_comp()
742 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in mlx4_ib_poll_one()
756 tail = wq->tail & (wq->wqe_cnt - 1); in mlx4_ib_poll_one()
/linux/drivers/infiniband/hw/vmw_pvrdma/
H A Dpvrdma_qp.c140 qp->rq.wqe_cnt = roundup_pow_of_two(max(1U, req_cap->max_recv_wr)); in pvrdma_set_rq_size()
144 req_cap->max_recv_wr = qp->rq.wqe_cnt; in pvrdma_set_rq_size()
150 qp->npages_recv = (qp->rq.wqe_cnt * qp->rq.wqe_size + PAGE_SIZE - 1) / in pvrdma_set_rq_size()
165 qp->sq.wqe_cnt = roundup_pow_of_two(max(1U, req_cap->max_send_wr)); in pvrdma_set_sq_size()
169 req_cap->max_send_wr = qp->sq.wqe_cnt; in pvrdma_set_sq_size()
177 (qp->sq.wqe_cnt * qp->sq.wqe_size + PAGE_SIZE - 1) / in pvrdma_set_sq_size()
696 qp->sq.ring, qp->sq.wqe_cnt, &tail))) { in pvrdma_post_send()
849 qp->sq.wqe_cnt); in pvrdma_post_send()
914 qp->rq.ring, qp->rq.wqe_cnt, &tail))) { in pvrdma_post_recv()
940 qp->rq.wqe_cnt); in pvrdma_post_recv()
H A Dpvrdma.h154 int wqe_cnt; member
169 int wqe_cnt; member
/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/sws/
H A Ddr_send.c197 (dr_cq->qp->sq.wqe_cnt - 1); in dr_parse_cqe()
203 (dr_cq->qp->sq.wqe_cnt - 1); in dr_parse_cqe()
270 dr_qp->rq.wqe_cnt = 256; in dr_create_rc_qp()
274 dr_qp->sq.wqe_cnt = roundup_pow_of_two(attr->max_send_wr); in dr_create_rc_qp()
277 MLX5_SET(qpc, temp_qpc, log_rq_size, ilog2(dr_qp->rq.wqe_cnt)); in dr_create_rc_qp()
278 MLX5_SET(qpc, temp_qpc, log_sq_size, ilog2(dr_qp->sq.wqe_cnt)); in dr_create_rc_qp()
286 dr_qp->sq.wqe_head = kcalloc(dr_qp->sq.wqe_cnt, in dr_create_rc_qp()
317 MLX5_SET(qpc, qpc, log_rq_size, ilog2(dr_qp->rq.wqe_cnt)); in dr_create_rc_qp()
319 MLX5_SET(qpc, qpc, log_sq_size, ilog2(dr_qp->sq.wqe_cnt)); in dr_create_rc_qp()
446 idx = dr_qp->sq.pc & (dr_qp->sq.wqe_cnt - 1); in dr_rdma_segments()
H A Ddr_types.h1426 unsigned int wqe_cnt; member
1432 unsigned int wqe_cnt; member
/linux/drivers/net/ethernet/huawei/hinic3/
H A Dhinic3_cmdq.c797 u16 wqe_cnt, wqe_idx, i; in hinic3_cmdq_flush_sync_cmd() local
803 wqe_cnt = hinic3_wq_get_used(wq); in hinic3_cmdq_flush_sync_cmd()
804 for (i = 0; i < wqe_cnt; i++) { in hinic3_cmdq_flush_sync_cmd()
/linux/drivers/vfio/pci/mlx5/
H A Dcmd.h149 unsigned int wqe_cnt; member
H A Dcmd.c1245 qp->rq.wqe_cnt = roundup_pow_of_two(max_recv_wr); in mlx5vf_create_rc_qp()
1247 log_rq_sz = ilog2(qp->rq.wqe_cnt); in mlx5vf_create_rc_qp()
1314 WARN_ON(qp->rq.pc - qp->rq.cc >= qp->rq.wqe_cnt); in mlx5vf_post_recv()
1315 ix = qp->rq.pc & (qp->rq.wqe_cnt - 1); in mlx5vf_post_recv()
1352 for (i = 0; i < qp->rq.wqe_cnt; i++) { in mlx5vf_activate_qp()
1665 ix = be16_to_cpu(cqe->wqe_counter) & (qp->rq.wqe_cnt - 1); in mlx5vf_rq_cqe()
/linux/drivers/infiniband/hw/mthca/
H A Dmthca_srq.c56 __be16 wqe_cnt; member
68 __be16 wqe_cnt; member
/linux/include/net/mana/
H A Dgdma.h113 u64 wqe_cnt : 8;
111 u64 wqe_cnt : 8; global() member
/linux/drivers/net/ethernet/mellanox/mlx4/
H A Dmlx4.h378 __be16 wqe_cnt; member
/linux/drivers/net/ethernet/microsoft/mana/
H A Dgdma_main.c356 e.rq.wqe_cnt = num_req; in mana_gd_ring_doorbell()
/linux/drivers/infiniband/hw/bnxt_re/
H A Dib_verbs.c2837 qp->qplib_qp.wqe_cnt == BNXT_RE_UD_QP_HW_STALL) { in bnxt_ud_qp_hw_stall_workaround()
2844 qp->qplib_qp.wqe_cnt = 0; in bnxt_ud_qp_hw_stall_workaround()
H A Dqplib_fp.c2095 qp->wqe_cnt++; in bnxt_qplib_post_send()