Searched refs:sgls (Results 1 – 11 of 11) sorted by relevance
/linux/drivers/nvme/target/ |
H A D | pci-epf.c | 997 struct nvme_sgl_desc *sgls; in nvmet_pci_epf_get_sgl_segment() local 1013 sgls = buf; in nvmet_pci_epf_get_sgl_segment() 1015 if (sgls[nr_descs - 1].type == (NVME_SGL_FMT_SEG_DESC << 4) || in nvmet_pci_epf_get_sgl_segment() 1016 sgls[nr_descs - 1].type == (NVME_SGL_FMT_LAST_SEG_DESC << 4)) { in nvmet_pci_epf_get_sgl_segment() 1022 *desc = sgls[nr_descs - 1]; in nvmet_pci_epf_get_sgl_segment() 1031 return sgls; in nvmet_pci_epf_get_sgl_segment() 1039 struct nvme_sgl_desc *sgls = NULL; in nvmet_pci_epf_iod_parse_sgl_segments() local 1054 sgls = nvmet_pci_epf_get_sgl_segment(ctrl, &seg, &nr_sgls); in nvmet_pci_epf_iod_parse_sgl_segments() 1055 if (!sgls) { in nvmet_pci_epf_iod_parse_sgl_segments() 1072 if (sgls[i].type != (NVME_SGL_FMT_DATA_DESC << 4)) { in nvmet_pci_epf_iod_parse_sgl_segments() [all …]
|
H A D | passthru.c | 140 id->sgls = cpu_to_le32(1 << 0); /* we always support SGLs */ in nvmet_passthru_override_id_ctrl() 142 id->sgls |= cpu_to_le32(1 << 2); in nvmet_passthru_override_id_ctrl() 144 id->sgls |= cpu_to_le32(1 << 20); in nvmet_passthru_override_id_ctrl()
|
H A D | admin-cmd.c | 759 id->sgls = cpu_to_le32(NVME_CTRL_SGLS_BYTE_ALIGNED); in nvmet_execute_identify_ctrl() 761 id->sgls |= cpu_to_le32(NVME_CTRL_SGLS_KSDBDS); in nvmet_execute_identify_ctrl() 763 id->sgls |= cpu_to_le32(NVME_CTRL_SGLS_SAOS); in nvmet_execute_identify_ctrl()
|
/linux/drivers/misc/genwqe/ |
H A D | card_dev.c | 846 if (req->sgls[i].sgl != NULL) in ddcb_cmd_cleanup() 847 genwqe_free_sync_sgl(cd, &req->sgls[i]); in ddcb_cmd_cleanup() 958 rc = genwqe_alloc_sync_sgl(cd, &req->sgls[i], in ddcb_cmd_fixups() 964 genwqe_setup_sgl(cd, &req->sgls[i], in ddcb_cmd_fixups() 968 cpu_to_be64(req->sgls[i].sgl_dma_addr); in ddcb_cmd_fixups()
|
H A D | card_base.h | 387 struct genwqe_sgl sgls[DDCB_FIXUPS]; member
|
/linux/drivers/nvme/host/ |
H A D | nvme.h | 338 u32 sgls; member 1131 return ctrl->sgls & (NVME_CTRL_SGLS_BYTE_ALIGNED | in nvme_ctrl_sgl_supported() 1139 return ctrl->sgls & NVME_CTRL_SGLS_MSDS; in nvme_ctrl_meta_sgl_supported()
|
H A D | rdma.c | 1022 if (!(ctrl->ctrl.sgls & NVME_CTRL_SGLS_KSDBDS)) { in nvme_rdma_setup_ctrl() 1054 if (ctrl->ctrl.sgls & NVME_CTRL_SGLS_SAOS) in nvme_rdma_setup_ctrl()
|
H A D | core.c | 3383 ctrl->sgls = le32_to_cpu(id->sgls); in nvme_init_identify()
|
/linux/drivers/scsi/elx/efct/ |
H A D | efct_hw.c | 610 struct efc_dma **sgls = NULL; in efct_hw_init_prereg_io() local 614 sgls = kmalloc_array(sgls_per_request, sizeof(*sgls), GFP_KERNEL); in efct_hw_init_prereg_io() 615 if (!sgls) in efct_hw_init_prereg_io() 623 kfree(sgls); in efct_hw_init_prereg_io() 641 sgls[n] = hw->io[idx + n]->sgl; in efct_hw_init_prereg_io() 645 hw->io[idx]->indicator, n, sgls, NULL, &req)) { in efct_hw_init_prereg_io() 667 kfree(sgls); in efct_hw_init_prereg_io()
|
/linux/drivers/infiniband/core/ |
H A D | uverbs_cmd.c | 2004 const struct ib_sge __user *sgls; in ib_uverbs_post_send() local 2015 sgls = uverbs_request_next_ptr(&iter, in ib_uverbs_post_send() 2018 if (IS_ERR(sgls)) in ib_uverbs_post_send() 2019 return PTR_ERR(sgls); in ib_uverbs_post_send() 2146 if (copy_from_user(next->sg_list, sgls + sg_ind, in ib_uverbs_post_send() 2197 const struct ib_sge __user *sgls; in ib_uverbs_unmarshall_recv() local 2206 sgls = uverbs_request_next_ptr(iter, size_mul(sge_count, in ib_uverbs_unmarshall_recv() 2208 if (IS_ERR(sgls)) in ib_uverbs_unmarshall_recv() 2209 return ERR_CAST(sgls); in ib_uverbs_unmarshall_recv() 2260 if (copy_from_user(next->sg_list, sgls + sg_ind, in ib_uverbs_unmarshall_recv()
|
/linux/include/linux/ |
H A D | nvme.h | 379 __le32 sgls; member
|