Searched refs:data_sgl (Results 1 – 4 of 4) sorted by relevance
/linux/drivers/mtd/nand/ |
H A D | qpic_common.c | 50 (sizeof(*bam_txn->data_sgl) * QPIC_PER_CW_DATA_SGL)); in qcom_alloc_bam_transaction() 67 bam_txn->data_sgl = bam_txn_buf; in qcom_alloc_bam_transaction() 94 sg_init_table(bam_txn->data_sgl, nandc->max_cwperpage * in qcom_clear_bam_transaction() 172 sgl = &bam_txn->data_sgl[bam_txn->tx_sgl_start]; in qcom_prepare_bam_async_desc() 178 sgl = &bam_txn->data_sgl[bam_txn->rx_sgl_start]; in qcom_prepare_bam_async_desc() 317 sg_set_buf(&bam_txn->data_sgl[bam_txn->rx_sgl_pos], in qcom_prep_bam_dma_desc_data() 326 sg_set_buf(&bam_txn->data_sgl[bam_txn->tx_sgl_pos], in qcom_prep_bam_dma_desc_data()
|
/linux/drivers/nvme/host/ |
H A D | rdma.c | 74 struct nvme_rdma_sgl data_sgl; member 1226 ib_dma_unmap_sg(ibdev, req->data_sgl.sg_table.sgl, req->data_sgl.nents, in nvme_rdma_dma_unmap_req() 1228 sg_free_table_chained(&req->data_sgl.sg_table, NVME_INLINE_SG_CNT); in nvme_rdma_dma_unmap_req() 1274 for_each_sg(req->data_sgl.sg_table.sgl, sgl, count, i) { in nvme_rdma_map_sg_inline() 1295 sg->addr = cpu_to_le64(sg_dma_address(req->data_sgl.sg_table.sgl)); in nvme_rdma_map_sg_single() 1296 put_unaligned_le24(sg_dma_len(req->data_sgl.sg_table.sgl), sg->length); in nvme_rdma_map_sg_single() 1317 nr = ib_map_mr_sg(req->mr, req->data_sgl.sg_table.sgl, count, NULL, in nvme_rdma_map_sg_fr() 1410 struct nvme_rdma_sgl *sgl = &req->data_sgl; in nvme_rdma_map_sg_pi() 1472 req->data_sgl.sg_table.sgl = (struct scatterlist *)(req + 1); in nvme_rdma_dma_map_req() 1473 ret = sg_alloc_table_chained(&req->data_sgl.sg_table, in nvme_rdma_dma_map_req() [all …]
|
/linux/include/linux/mtd/ |
H A D | nand-qpic-common.h | 259 struct scatterlist *data_sgl; member
|
/linux/drivers/nvme/target/ |
H A D | pci-epf.c | 148 struct scatterlist data_sgl; member 1138 sg_init_table(&iod->data_sgl, 1); in nvmet_pci_epf_alloc_iod_data_buf() 1139 iod->data_sgt.sgl = &iod->data_sgl; in nvmet_pci_epf_alloc_iod_data_buf()
|