| /linux/drivers/nvme/target/ |
| H A D | fabrics-cmd.c | 176 u16 sqsize = le16_to_cpu(c->sqsize); in nvmet_install_queue() local 181 if (!sqsize) { in nvmet_install_queue() 183 req->error_loc = offsetof(struct nvmf_connect_command, sqsize); in nvmet_install_queue() 184 req->cqe->result.u32 = IPO_IATTR_CONNECT_SQE(sqsize); in nvmet_install_queue() 196 if (qid && sqsize > mqes) { in nvmet_install_queue() 198 sqsize, mqes, ctrl->cntlid); in nvmet_install_queue() 199 req->error_loc = offsetof(struct nvmf_connect_command, sqsize); in nvmet_install_queue() 200 req->cqe->result.u32 = IPO_IATTR_CONNECT_SQE(sqsize); in nvmet_install_queue() 220 nvmet_cq_setup(ctrl, req->cq, qid, sqsize + 1); in nvmet_install_queue() 221 nvmet_sq_setup(ctrl, req->sq, qid, sqsize + 1); in nvmet_install_queue()
|
| H A D | fc.c | 134 u16 sqsize; member 624 for (i = 0; i < queue->sqsize; fod++, i++) { in nvmet_fc_prep_fcp_iodlist() 659 for (i = 0; i < queue->sqsize; fod++, i++) { in nvmet_fc_destroy_fcp_iodlist() 783 u16 qid, u16 sqsize) in nvmet_fc_alloc_target_queue() argument 791 queue = kzalloc_flex(*queue, fod, sqsize); in nvmet_fc_alloc_target_queue() 802 queue->sqsize = sqsize; in nvmet_fc_alloc_target_queue() 880 for (i = 0; i < queue->sqsize; fod++, i++) { in nvmet_fc_delete_target_queue() 1694 be16_to_cpu(rqst->assoc_cmd.sqsize))) in nvmet_fc_ls_create_association() 1705 be16_to_cpu(rqst->assoc_cmd.sqsize)); in nvmet_fc_ls_create_association() 1784 be16_to_cpu(rqst->connect_cmd.sqsize))) in nvmet_fc_ls_create_connection() [all …]
|
| H A D | loop.c | 610 ctrl->ctrl.sqsize = opts->queue_size - 1; in nvme_loop_create_ctrl()
|
| /linux/include/linux/ |
| H A D | nvme-fc.h | 260 __be16 sqsize; member 288 __be16 sqsize; member
|
| H A D | nvme.h | 1707 __le16 sqsize; member
|
| /linux/drivers/nvme/host/ |
| H A D | rdma.c | 740 ctrl->ctrl.sqsize + 1); in nvme_rdma_alloc_io_queues() 1029 if (ctrl->ctrl.opts->queue_size > ctrl->ctrl.sqsize + 1) { in nvme_rdma_setup_ctrl() 1032 ctrl->ctrl.opts->queue_size, ctrl->ctrl.sqsize + 1); in nvme_rdma_setup_ctrl() 1040 if (ctrl->ctrl.sqsize + 1 > max_queue_size) { in nvme_rdma_setup_ctrl() 1043 ctrl->ctrl.sqsize + 1, max_queue_size); in nvme_rdma_setup_ctrl() 1044 ctrl->ctrl.sqsize = max_queue_size - 1; in nvme_rdma_setup_ctrl() 1047 if (ctrl->ctrl.sqsize + 1 > ctrl->ctrl.maxcmd) { in nvme_rdma_setup_ctrl() 1050 ctrl->ctrl.sqsize + 1, ctrl->ctrl.maxcmd); in nvme_rdma_setup_ctrl() 1051 ctrl->ctrl.sqsize = ctrl->ctrl.maxcmd - 1; in nvme_rdma_setup_ctrl() 1871 priv.hsqsize = cpu_to_le16(queue->ctrl->ctrl.sqsize); in nvme_rdma_route_resolved() [all …]
|
| H A D | fc.c | 1170 assoc_rqst->assoc_cmd.sqsize = cpu_to_be16(qsize - 1); in nvme_fc_connect_admin_queue() 1292 conn_rqst->connect_cmd.sqsize = cpu_to_be16(qsize - 1); in nvme_fc_connect_queue() 2888 ret = nvme_fc_create_hw_io_queues(ctrl, ctrl->ctrl.sqsize + 1); in nvme_fc_create_io_queues() 2892 ret = nvme_fc_connect_io_queues(ctrl, ctrl->ctrl.sqsize + 1); in nvme_fc_create_io_queues() 2948 ret = nvme_fc_create_hw_io_queues(ctrl, ctrl->ctrl.sqsize + 1); in nvme_fc_recreate_io_queues() 2952 ret = nvme_fc_connect_io_queues(ctrl, ctrl->ctrl.sqsize + 1); in nvme_fc_recreate_io_queues() 3124 ctrl->ctrl.sqsize = opts->queue_size - 1; in nvme_fc_create_association() 3495 ctrl->ctrl.sqsize = opts->queue_size - 1; in nvme_fc_alloc_ctrl()
|
| H A D | fabrics.c | 419 cmd->connect.sqsize = cpu_to_le16(ctrl->sqsize); in nvmf_connect_cmd_prep() 421 cmd->connect.sqsize = cpu_to_le16(NVME_AQ_DEPTH - 1); in nvmf_connect_cmd_prep()
|
| H A D | tcp.c | 2396 if (opts->queue_size > ctrl->sqsize + 1) in nvme_tcp_setup_ctrl() 2399 opts->queue_size, ctrl->sqsize + 1); in nvme_tcp_setup_ctrl() 2401 if (ctrl->sqsize + 1 > ctrl->maxcmd) { in nvme_tcp_setup_ctrl() 2404 ctrl->sqsize + 1, ctrl->maxcmd); in nvme_tcp_setup_ctrl() 2405 ctrl->sqsize = ctrl->maxcmd - 1; in nvme_tcp_setup_ctrl() 2903 ctrl->ctrl.sqsize = opts->queue_size - 1; in nvme_tcp_alloc_ctrl()
|
| H A D | sysfs.c | 388 nvme_show_int_function(sqsize);
|
| H A D | nvme.h | 387 u16 sqsize; member
|
| H A D | core.c | 3511 ctrl->maxcmd = ctrl->sqsize + 1; in nvme_check_ctrl_fabric_info() 3674 ctrl->sqsize = min_t(u16, NVME_CAP_MQES(ctrl->cap), ctrl->sqsize); in nvme_init_ctrl_finish() 4915 set->queue_depth = min_t(unsigned, ctrl->sqsize, BLK_MQ_MAX_DEPTH - 1); in nvme_alloc_io_tag_set()
|
| H A D | apple.c | 1158 anv->ctrl.sqsize = in apple_nvme_reset_work()
|
| H A D | pci.c | 2968 dev->ctrl.sqsize = result - 1; in nvme_setup_io_queues() 3223 dev->ctrl.sqsize = dev->q_depth - 1; /* 0's based queue depth */ in nvme_pci_enable()
|
| /linux/drivers/infiniband/hw/irdma/ |
| H A D | hw.c | 1024 u32 sqsize = IRDMA_CQP_SW_SQSIZE_2048; in irdma_create_cqp() local 1032 cqp->cqp_requests = kzalloc_objs(*cqp->cqp_requests, sqsize); in irdma_create_cqp() 1036 cqp->scratch_array = kcalloc(sqsize, sizeof(*cqp->scratch_array), GFP_KERNEL); in irdma_create_cqp() 1042 cqp->oop_op_array = kzalloc_objs(*cqp->oop_op_array, sqsize); in irdma_create_cqp() 1050 cqp->sq.size = ALIGN(sizeof(struct irdma_cqp_sq_wqe) * sqsize, in irdma_create_cqp() 1068 cqp_init_info.sq_size = sqsize; in irdma_create_cqp() 1110 for (i = 0; i < sqsize; i++) { in irdma_create_cqp()
|
| /linux/drivers/infiniband/hw/cxgb4/ |
| H A D | qp.c | 2124 unsigned int sqsize, rqsize = 0; in c4iw_create_qp() local 2154 sqsize = attrs->cap.max_send_wr + 1; in c4iw_create_qp() 2155 if (sqsize < 8) in c4iw_create_qp() 2156 sqsize = 8; in c4iw_create_qp() 2162 qhp->wq.sq.size = sqsize; in c4iw_create_qp() 2164 (sqsize + rhp->rdev.hw_queue.t4_eq_status_entries) * in c4iw_create_qp() 2188 attrs->cap.max_send_wr = sqsize - 1; in c4iw_create_qp()
|
| /linux/drivers/infiniband/sw/rdmavt/ |
| H A D | qp.c | 1038 size_t sqsize; in rvt_create_qp() local 1064 sqsize = in rvt_create_qp() 1078 swq = vzalloc_node(array_size(sz, sqsize), rdi->dparms.node); in rvt_create_qp() 1151 qp->s_size = sqsize; in rvt_create_qp()
|