| /linux/drivers/scsi/snic/ |
| H A D | snic_res.c | 110 snic->cq_count = svnic_dev_get_res_count(snic->vdev, RES_TYPE_CQ); in snic_get_res_counts() 111 SNIC_BUG_ON(snic->cq_count == 0); in snic_get_res_counts() 125 for (i = 0; i < snic->cq_count; i++) in snic_free_vnic_res() 157 snic->cq_count, in snic_alloc_vnic_res() 183 SNIC_BUG_ON(snic->cq_count != 2 * snic->wq_count); in snic_alloc_vnic_res() 185 for (i = snic->wq_count; i < snic->cq_count; i++) { in snic_alloc_vnic_res() 216 for (i = 0; i < snic->cq_count; i++) { in snic_alloc_vnic_res()
|
| H A D | snic_isr.c | 156 if (snic->wq_count < n || snic->cq_count < n + m) in snic_set_intr_mode() 163 snic->cq_count = n + m; in snic_set_intr_mode()
|
| H A D | snic.h | 279 unsigned int cq_count; member
|
| H A D | snic_main.c | 238 for (i = 0; i < snic->cq_count; i++) in snic_cleanup()
|
| /linux/drivers/scsi/fnic/ |
| H A D | fnic_isr.c | 259 fnic->wq_copy_count, fnic->cq_count); in fnic_set_intr_mode_msix() 289 if (fnic->cq_count != vec_count - 1) { in fnic_set_intr_mode_msix() 292 fnic->cq_count, vec_count); in fnic_set_intr_mode_msix() 293 fnic->cq_count = vec_count - 1; in fnic_set_intr_mode_msix() 306 fnic->wq_count, fnic->cq_count); in fnic_set_intr_mode_msix() 343 fnic->cq_count >= 3 && in fnic_set_intr_mode() 350 fnic->cq_count = 3; in fnic_set_intr_mode() 371 fnic->cq_count >= 3 && in fnic_set_intr_mode() 377 fnic->cq_count = 3; in fnic_set_intr_mode()
|
| H A D | fnic_res.c | 195 fnic->cq_count = vnic_dev_get_res_count(fnic->vdev, RES_TYPE_CQ); in fnic_get_res_counts() 203 dev_info(&fnic->pdev->dev, "vNIC fw resources cq_count: %d\n", fnic->cq_count); in fnic_get_res_counts() 220 for (i = 0; i < fnic->cq_count; i++) in fnic_free_vnic_resources() 251 fnic->cq_count, fnic->intr_count, in fnic_alloc_vnic_resources() 377 for (i = 0; i < fnic->cq_count; i++) { in fnic_alloc_vnic_resources()
|
| H A D | fnic.h | 381 unsigned int cq_count; member
|
| H A D | fnic_main.c | 549 for (i = 0; i < fnic->cq_count; i++) in fnic_cleanup()
|
| /linux/drivers/net/ethernet/cisco/enic/ |
| H A D | enic_res.c | 189 for (i = 0; i < enic->cq_count; i++) in enic_free_vnic_resources() 205 enic->cq_count = enic->cq_avail; in enic_get_res_counts() 271 for (i = 0; i < enic->cq_count; i++) { in enic_init_vnic_resources() 331 enic->cq_count, enic->intr_count, in enic_alloc_vnic_resources() 374 for (i = 0; i < enic->cq_count; i++) { in enic_alloc_vnic_resources()
|
| H A D | enic_main.c | 1829 for (i = 0; i < enic->cq_count; i++) in enic_stop() 2291 enic->cq_count = 2; in enic_adjust_resources() 2317 enic->cq_count = enic->rq_count + enic->wq_count; in enic_adjust_resources() 2318 enic->intr_count = enic->cq_count + ENIC_MSIX_RESERVED_INTR; in enic_adjust_resources()
|
| /linux/drivers/infiniband/hw/bnxt_re/ |
| H A D | hw_counters.h | 131 atomic_t cq_count; member
|
| H A D | main.c | 245 ctx->cq_count = min_t(u32, BNXT_RE_MAX_CQ_COUNT, attr->max_cq); in bnxt_re_limit_pf_res() 269 vf_res->max_cq_per_vf = (qplib_ctx->cq_count * vf_pct) / num_vf; in bnxt_re_limit_vf_res() 1448 atomic_set(&rdev->stats.res.cq_count, 0); in bnxt_re_dev_add()
|
| H A D | qplib_res.c | 515 hwq_attr.depth = ctx->cq_count; in bnxt_qplib_alloc_hwctx()
|
| /linux/drivers/net/ethernet/fungible/funcore/ |
| H A D | fun_dev.c | 563 unsigned int cq_count, sq_count, num_dbs; in fun_get_dev_limits() local 569 cq_count = rc; in fun_get_dev_limits() 579 if (cq_count < 2 || sq_count < 2 + !!fdev->admin_q->rq_depth) in fun_get_dev_limits() 587 fdev->max_qid = min3(cq_count, sq_count, num_dbs / 2) - 1; in fun_get_dev_limits()
|
| /linux/drivers/scsi/be2iscsi/ |
| H A D | be.h | 88 u32 cq_count; member
|
| H A D | be_main.c | 2076 pbe_eq->cq_count += ret; in be_iopoll() 5209 pbe_eq->cq_count < aic->eq_prev) { in beiscsi_eqd_update_work() 5211 aic->eq_prev = pbe_eq->cq_count; in beiscsi_eqd_update_work() 5215 pps = (((u32)(pbe_eq->cq_count - aic->eq_prev) * 1000) / delta); in beiscsi_eqd_update_work() 5224 aic->eq_prev = pbe_eq->cq_count; in beiscsi_eqd_update_work()
|
| /linux/drivers/scsi/elx/efct/ |
| H A D | efct_hw_queues.c | 25 hw->cq_count = 0; in efct_hw_init_queues() 169 cq->instance = eq->hw->cq_count++; in efct_hw_new_cq() 214 cq->instance = hw->cq_count++; in efct_hw_new_cq_set()
|
| H A D | efct_hw.h | 393 u32 cq_count; member
|
| H A D | efct_hw.c | 1092 for (i = 0; i < hw->cq_count; i++) { in efct_hw_init() 2228 for (i = 0; i < hw->cq_count; i++) in efct_hw_eq_process() 3470 for (i = 0; i < hw->cq_count; i++) in efct_hw_teardown()
|
| /linux/include/linux/qed/ |
| H A D | qed_rdma_if.h | 443 u64 cq_count; member
|
| /linux/drivers/infiniband/hw/mlx5/ |
| H A D | cq.c | 1155 int mlx5_ib_modify_cq(struct ib_cq *cq, u16 cq_count, u16 cq_period) in mlx5_ib_modify_cq() argument 1168 cq_period, cq_count); in mlx5_ib_modify_cq()
|
| /linux/drivers/net/ethernet/mellanox/mlxsw/ |
| H A D | pci.c | 1121 u8 cqn, cq_count; in mlxsw_pci_eq_tasklet() local 1141 cq_count = mlxsw_pci->num_cqs; in mlxsw_pci_eq_tasklet() 1142 for_each_set_bit(cqn, active_cqns, cq_count) { in mlxsw_pci_eq_tasklet()
|
| /linux/include/uapi/rdma/ |
| H A D | ib_user_verbs.h | 1320 __u16 cq_count; member
|
| /linux/drivers/infiniband/core/ |
| H A D | verbs.c | 2172 int rdma_set_cq_moderation(struct ib_cq *cq, u16 cq_count, u16 cq_period) in rdma_set_cq_moderation() argument 2178 cq->device->ops.modify_cq(cq, cq_count, in rdma_set_cq_moderation()
|
| /linux/include/rdma/ |
| H A D | ib_verbs.h | 2517 int (*modify_cq)(struct ib_cq *cq, u16 cq_count, u16 cq_period); 4050 int rdma_set_cq_moderation(struct ib_cq *cq, u16 cq_count, u16 cq_period);
|