Searched refs:sq_size (Results 1 – 15 of 15) sorted by relevance
| /linux/drivers/scsi/bnx2i/ |
| H A D | bnx2i_init.c | 61 unsigned int sq_size; variable 62 module_param(sq_size, int, 0664); 63 MODULE_PARM_DESC(sq_size, "Configure SQ size"); 471 if (sq_size && !is_power_of_2(sq_size)) in bnx2i_mod_init() 472 sq_size = roundup_pow_of_two(sq_size); in bnx2i_mod_init()
|
| H A D | bnx2i_sysfs.c | 140 static DEVICE_ATTR(sq_size, S_IRUGO | S_IWUSR,
|
| H A D | bnx2i_iscsi.c | 844 if (sq_size && sq_size <= BNX2I_5770X_SQ_WQES_MAX) in bnx2i_alloc_hba() 845 hba->max_sqes = sq_size; in bnx2i_alloc_hba() 849 if (sq_size && sq_size <= BNX2I_570X_SQ_WQES_MAX) in bnx2i_alloc_hba() 850 hba->max_sqes = sq_size; in bnx2i_alloc_hba()
|
| H A D | bnx2i.h | 795 extern unsigned int sq_size;
|
| /linux/include/uapi/rdma/ |
| H A D | cxgb4-abi.h | 82 __u32 sq_size; member
|
| /linux/io_uring/ |
| H A D | register.c | 556 rd.size = PAGE_ALIGN(rl->sq_size); in io_register_resize_rings() 603 size_t sq_size; in io_register_resize_rings() local 606 sq_size = sizeof(struct io_uring_sqe); in io_register_resize_rings() 611 sq_size <<= 1; in io_register_resize_rings() 615 memcpy(&n.sq_sqes[index & dst_mask], &o.sq_sqes[index & src_mask], sq_size); in io_register_resize_rings()
|
| H A D | io_uring.h | 24 size_t sq_size; member
|
| H A D | io_uring.c | 2097 rl->sq_size = array_size(sqe_size, sq_entries); in rings_size() 2098 if (rl->sq_size == SIZE_MAX) in rings_size() 2743 rd.size = PAGE_ALIGN(rl->sq_size); in io_allocate_scq_urings()
|
| /linux/drivers/infiniband/sw/siw/ |
| H A D | siw.h | 257 u32 sq_size; member 626 struct siw_sqe *sqe = &qp->sendq[qp->sq_get % qp->attrs.sq_size]; in siw_sq_empty() 633 struct siw_sqe *sqe = &qp->sendq[qp->sq_get % qp->attrs.sq_size]; in sq_get_next()
|
| H A D | siw_qp.c | 1204 while (qp->attrs.sq_size) { in siw_sq_flush() 1205 sqe = &qp->sendq[qp->sq_get % qp->attrs.sq_size]; in siw_sq_flush()
|
| /linux/drivers/infiniband/hw/irdma/ |
| H A D | uk.c | 1491 int irdma_get_sqdepth(struct irdma_uk_attrs *uk_attrs, u32 sq_size, u8 shift, in irdma_get_sqdepth() argument 1496 irdma_round_up_wq(((u64)sq_size << shift) + IRDMA_SQ_RSVD); in irdma_get_sqdepth() 1664 status = irdma_get_sqdepth(ukinfo->uk_attrs, ukinfo->sq_size, in irdma_uk_calc_depth_shift_sq() 1725 qp->sq_size = info->sq_size; in irdma_uk_qp_init() 1727 sq_ring_size = qp->sq_size << info->sq_shift; in irdma_uk_qp_init()
|
| H A D | verbs.c | 659 ukinfo->sq_size = init_attr->cap.max_send_wr; in irdma_setup_umode_qp() 678 ukinfo->sq_size = ukinfo->sq_depth >> ukinfo->sq_shift; in irdma_setup_umode_qp() 752 ukinfo->sq_size = ukinfo->sq_depth >> ukinfo->sq_shift; in irdma_setup_kmode_qp() 990 init_info.qp_uk_init_info.sq_size = init_attr->cap.max_send_wr; in irdma_create_qp() 1122 uresp.actual_sq_size = init_info.qp_uk_init_info.sq_size; in irdma_create_qp()
|
| /linux/drivers/dma/ |
| H A D | hisi_dma.c | 585 size_t sq_size = sizeof(struct hisi_dma_sqe) * hdma_dev->chan_depth; in hisi_dma_alloc_qps_mem() local 593 chan->sq = dmam_alloc_coherent(dev, sq_size, &chan->sq_dma, in hisi_dma_alloc_qps_mem()
|
| /linux/drivers/net/ethernet/mellanox/mlx4/ |
| H A D | resource_tracker.c | 2721 int sq_size; in qp_get_mtt_size() local 2728 sq_size = 1 << (log_sq_size + log_sq_sride + 4); in qp_get_mtt_size() 2730 total_mem = sq_size + rq_size; in qp_get_mtt_size()
|
| /linux/drivers/infiniband/hw/cxgb4/ |
| H A D | qp.c | 2258 uresp.sq_size = qhp->wq.sq.size; in c4iw_create_qp()
|