Home
last modified time | relevance | path

Searched refs:sgls (Results 1 – 11 of 11) sorted by relevance

/linux/drivers/nvme/target/
H A Dpci-epf.c997 struct nvme_sgl_desc *sgls; in nvmet_pci_epf_get_sgl_segment() local
1013 sgls = buf; in nvmet_pci_epf_get_sgl_segment()
1015 if (sgls[nr_descs - 1].type == (NVME_SGL_FMT_SEG_DESC << 4) || in nvmet_pci_epf_get_sgl_segment()
1016 sgls[nr_descs - 1].type == (NVME_SGL_FMT_LAST_SEG_DESC << 4)) { in nvmet_pci_epf_get_sgl_segment()
1022 *desc = sgls[nr_descs - 1]; in nvmet_pci_epf_get_sgl_segment()
1031 return sgls; in nvmet_pci_epf_get_sgl_segment()
1039 struct nvme_sgl_desc *sgls = NULL; in nvmet_pci_epf_iod_parse_sgl_segments() local
1054 sgls = nvmet_pci_epf_get_sgl_segment(ctrl, &seg, &nr_sgls); in nvmet_pci_epf_iod_parse_sgl_segments()
1055 if (!sgls) { in nvmet_pci_epf_iod_parse_sgl_segments()
1072 if (sgls[i].type != (NVME_SGL_FMT_DATA_DESC << 4)) { in nvmet_pci_epf_iod_parse_sgl_segments()
[all …]
H A Dpassthru.c140 id->sgls = cpu_to_le32(1 << 0); /* we always support SGLs */ in nvmet_passthru_override_id_ctrl()
142 id->sgls |= cpu_to_le32(1 << 2); in nvmet_passthru_override_id_ctrl()
144 id->sgls |= cpu_to_le32(1 << 20); in nvmet_passthru_override_id_ctrl()
H A Dadmin-cmd.c759 id->sgls = cpu_to_le32(NVME_CTRL_SGLS_BYTE_ALIGNED); in nvmet_execute_identify_ctrl()
761 id->sgls |= cpu_to_le32(NVME_CTRL_SGLS_KSDBDS); in nvmet_execute_identify_ctrl()
763 id->sgls |= cpu_to_le32(NVME_CTRL_SGLS_SAOS); in nvmet_execute_identify_ctrl()
/linux/drivers/misc/genwqe/
H A Dcard_dev.c846 if (req->sgls[i].sgl != NULL) in ddcb_cmd_cleanup()
847 genwqe_free_sync_sgl(cd, &req->sgls[i]); in ddcb_cmd_cleanup()
958 rc = genwqe_alloc_sync_sgl(cd, &req->sgls[i], in ddcb_cmd_fixups()
964 genwqe_setup_sgl(cd, &req->sgls[i], in ddcb_cmd_fixups()
968 cpu_to_be64(req->sgls[i].sgl_dma_addr); in ddcb_cmd_fixups()
H A Dcard_base.h387 struct genwqe_sgl sgls[DDCB_FIXUPS]; member
/linux/drivers/nvme/host/
H A Dnvme.h338 u32 sgls; member
1131 return ctrl->sgls & (NVME_CTRL_SGLS_BYTE_ALIGNED | in nvme_ctrl_sgl_supported()
1139 return ctrl->sgls & NVME_CTRL_SGLS_MSDS; in nvme_ctrl_meta_sgl_supported()
H A Drdma.c1022 if (!(ctrl->ctrl.sgls & NVME_CTRL_SGLS_KSDBDS)) { in nvme_rdma_setup_ctrl()
1054 if (ctrl->ctrl.sgls & NVME_CTRL_SGLS_SAOS) in nvme_rdma_setup_ctrl()
H A Dcore.c3383 ctrl->sgls = le32_to_cpu(id->sgls); in nvme_init_identify()
/linux/drivers/scsi/elx/efct/
H A Defct_hw.c610 struct efc_dma **sgls = NULL; in efct_hw_init_prereg_io() local
614 sgls = kmalloc_array(sgls_per_request, sizeof(*sgls), GFP_KERNEL); in efct_hw_init_prereg_io()
615 if (!sgls) in efct_hw_init_prereg_io()
623 kfree(sgls); in efct_hw_init_prereg_io()
641 sgls[n] = hw->io[idx + n]->sgl; in efct_hw_init_prereg_io()
645 hw->io[idx]->indicator, n, sgls, NULL, &req)) { in efct_hw_init_prereg_io()
667 kfree(sgls); in efct_hw_init_prereg_io()
/linux/drivers/infiniband/core/
H A Duverbs_cmd.c2004 const struct ib_sge __user *sgls; in ib_uverbs_post_send() local
2015 sgls = uverbs_request_next_ptr(&iter, in ib_uverbs_post_send()
2018 if (IS_ERR(sgls)) in ib_uverbs_post_send()
2019 return PTR_ERR(sgls); in ib_uverbs_post_send()
2146 if (copy_from_user(next->sg_list, sgls + sg_ind, in ib_uverbs_post_send()
2197 const struct ib_sge __user *sgls; in ib_uverbs_unmarshall_recv() local
2206 sgls = uverbs_request_next_ptr(iter, size_mul(sge_count, in ib_uverbs_unmarshall_recv()
2208 if (IS_ERR(sgls)) in ib_uverbs_unmarshall_recv()
2209 return ERR_CAST(sgls); in ib_uverbs_unmarshall_recv()
2260 if (copy_from_user(next->sg_list, sgls + sg_ind, in ib_uverbs_unmarshall_recv()
/linux/include/linux/
H A Dnvme.h379 __le32 sgls; member