/linux/drivers/nvme/target/ |
H A D | fc.c | 134 u16 sqsize; member 641 for (i = 0; i < queue->sqsize; fod++, i++) { in nvmet_fc_prep_fcp_iodlist() 676 for (i = 0; i < queue->sqsize; fod++, i++) { in nvmet_fc_destroy_fcp_iodlist() 800 u16 qid, u16 sqsize) in nvmet_fc_alloc_target_queue() argument 808 queue = kzalloc(struct_size(queue, fod, sqsize), GFP_KERNEL); in nvmet_fc_alloc_target_queue() 819 queue->sqsize = sqsize; in nvmet_fc_alloc_target_queue() 895 for (i = 0; i < queue->sqsize; fod++, i++) { in nvmet_fc_delete_target_queue() 1694 be16_to_cpu(rqst->assoc_cmd.sqsize))) in nvmet_fc_ls_create_association() 1705 be16_to_cpu(rqst->assoc_cmd.sqsize)); in nvmet_fc_ls_create_association() 1785 be16_to_cpu(rqst->connect_cmd.sqsize))) in nvmet_fc_ls_create_connection() [all …]
|
H A D | loop.c | 599 ctrl->ctrl.sqsize = opts->queue_size - 1; in nvme_loop_create_ctrl()
|
/linux/include/linux/ |
H A D | nvme-fc.h | 260 __be16 sqsize; member 288 __be16 sqsize; member
|
H A D | nvme.h | 1603 __le16 sqsize; member
|
/linux/drivers/nvme/host/ |
H A D | rdma.c | 740 ctrl->ctrl.sqsize + 1); in nvme_rdma_alloc_io_queues() 1029 if (ctrl->ctrl.opts->queue_size > ctrl->ctrl.sqsize + 1) { in nvme_rdma_setup_ctrl() 1032 ctrl->ctrl.opts->queue_size, ctrl->ctrl.sqsize + 1); in nvme_rdma_setup_ctrl() 1040 if (ctrl->ctrl.sqsize + 1 > max_queue_size) { in nvme_rdma_setup_ctrl() 1043 ctrl->ctrl.sqsize + 1, max_queue_size); in nvme_rdma_setup_ctrl() 1044 ctrl->ctrl.sqsize = max_queue_size - 1; in nvme_rdma_setup_ctrl() 1047 if (ctrl->ctrl.sqsize + 1 > ctrl->ctrl.maxcmd) { in nvme_rdma_setup_ctrl() 1050 ctrl->ctrl.sqsize + 1, ctrl->ctrl.maxcmd); in nvme_rdma_setup_ctrl() 1051 ctrl->ctrl.sqsize = ctrl->ctrl.maxcmd - 1; in nvme_rdma_setup_ctrl() 1872 priv.hsqsize = cpu_to_le16(queue->ctrl->ctrl.sqsize); in nvme_rdma_route_resolved() [all …]
|
H A D | fc.c | 1210 assoc_rqst->assoc_cmd.sqsize = cpu_to_be16(qsize - 1); in nvme_fc_connect_admin_queue() 1332 conn_rqst->connect_cmd.sqsize = cpu_to_be16(qsize - 1); in nvme_fc_connect_queue() 2923 ret = nvme_fc_create_hw_io_queues(ctrl, ctrl->ctrl.sqsize + 1); in nvme_fc_create_io_queues() 2927 ret = nvme_fc_connect_io_queues(ctrl, ctrl->ctrl.sqsize + 1); in nvme_fc_create_io_queues() 2983 ret = nvme_fc_create_hw_io_queues(ctrl, ctrl->ctrl.sqsize + 1); in nvme_fc_recreate_io_queues() 2987 ret = nvme_fc_connect_io_queues(ctrl, ctrl->ctrl.sqsize + 1); in nvme_fc_recreate_io_queues() 3154 ctrl->ctrl.sqsize = opts->queue_size - 1; in nvme_fc_create_association() 3516 ctrl->ctrl.sqsize = opts->queue_size - 1; in nvme_fc_alloc_ctrl()
|
H A D | fabrics.c | 419 cmd->connect.sqsize = cpu_to_le16(ctrl->sqsize); in nvmf_connect_cmd_prep() 421 cmd->connect.sqsize = cpu_to_le16(NVME_AQ_DEPTH - 1); in nvmf_connect_cmd_prep()
|
H A D | tcp.c | 2222 if (opts->queue_size > ctrl->sqsize + 1) in nvme_tcp_setup_ctrl() 2225 opts->queue_size, ctrl->sqsize + 1); in nvme_tcp_setup_ctrl() 2227 if (ctrl->sqsize + 1 > ctrl->maxcmd) { in nvme_tcp_setup_ctrl() 2230 ctrl->sqsize + 1, ctrl->maxcmd); in nvme_tcp_setup_ctrl() 2231 ctrl->sqsize = ctrl->maxcmd - 1; in nvme_tcp_setup_ctrl() 2721 ctrl->ctrl.sqsize = opts->queue_size - 1; in nvme_tcp_alloc_ctrl()
|
H A D | sysfs.c | 333 nvme_show_int_function(sqsize);
|
H A D | nvme.h | 333 u16 sqsize; member
|
H A D | core.c | 3266 ctrl->maxcmd = ctrl->sqsize + 1; in nvme_check_ctrl_fabric_info() 3429 ctrl->sqsize = min_t(u16, NVME_CAP_MQES(ctrl->cap), ctrl->sqsize); in nvme_init_ctrl_finish() 4631 set->queue_depth = min_t(unsigned, ctrl->sqsize, BLK_MQ_MAX_DEPTH - 1); in nvme_alloc_io_tag_set()
|
H A D | apple.c | 1100 anv->ctrl.sqsize = in apple_nvme_reset_work()
|
H A D | pci.c | 2487 dev->ctrl.sqsize = result - 1; in nvme_setup_io_queues() 2739 dev->ctrl.sqsize = dev->q_depth - 1; /* 0's based queue depth */ in nvme_pci_enable()
|
/linux/drivers/infiniband/hw/irdma/ |
H A D | hw.c | 929 u32 sqsize = IRDMA_CQP_SW_SQSIZE_2048; in irdma_create_cqp() local 937 cqp->cqp_requests = kcalloc(sqsize, sizeof(*cqp->cqp_requests), GFP_KERNEL); in irdma_create_cqp() 941 cqp->scratch_array = kcalloc(sqsize, sizeof(*cqp->scratch_array), GFP_KERNEL); in irdma_create_cqp() 949 cqp->sq.size = ALIGN(sizeof(struct irdma_cqp_sq_wqe) * sqsize, in irdma_create_cqp() 967 cqp_init_info.sq_size = sqsize; in irdma_create_cqp() 1005 for (i = 0; i < sqsize; i++) { in irdma_create_cqp()
|
/linux/drivers/infiniband/hw/cxgb4/ |
H A D | qp.c | 2119 unsigned int sqsize, rqsize = 0; in c4iw_create_qp() local 2149 sqsize = attrs->cap.max_send_wr + 1; in c4iw_create_qp() 2150 if (sqsize < 8) in c4iw_create_qp() 2151 sqsize = 8; in c4iw_create_qp() 2157 qhp->wq.sq.size = sqsize; in c4iw_create_qp() 2159 (sqsize + rhp->rdev.hw_queue.t4_eq_status_entries) * in c4iw_create_qp() 2183 attrs->cap.max_send_wr = sqsize - 1; in c4iw_create_qp()
|
/linux/drivers/infiniband/sw/rdmavt/ |
H A D | qp.c | 1039 size_t sqsize; in rvt_create_qp() local 1065 sqsize = in rvt_create_qp() 1079 swq = vzalloc_node(array_size(sz, sqsize), rdi->dparms.node); in rvt_create_qp() 1153 qp->s_size = sqsize; in rvt_create_qp()
|