xref: /linux/drivers/misc/bcm-vk/bcm_vk_sg.c (revision 7ec462100ef9142344ddbf86f2c3008b97acddbe)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Copyright 2018-2020 Broadcom.
4  */
5 #include <linux/dma-mapping.h>
6 #include <linux/mm.h>
7 #include <linux/pagemap.h>
8 #include <linux/pgtable.h>
9 #include <linux/vmalloc.h>
10 
11 #include <asm/page.h>
12 #include <linux/unaligned.h>
13 
14 #include <uapi/linux/misc/bcm_vk.h>
15 
16 #include "bcm_vk.h"
17 #include "bcm_vk_msg.h"
18 #include "bcm_vk_sg.h"
19 
20 /*
21  * Valkyrie has a hardware limitation of 16M transfer size.
22  * So limit the SGL chunks to 16M.
23  */
24 #define BCM_VK_MAX_SGL_CHUNK SZ_16M
25 
26 static int bcm_vk_dma_alloc(struct device *dev,
27 			    struct bcm_vk_dma *dma,
28 			    int dir,
29 			    struct _vk_data *vkdata);
30 static int bcm_vk_dma_free(struct device *dev, struct bcm_vk_dma *dma);
31 
32 /* Uncomment to dump SGLIST */
33 /* #define BCM_VK_DUMP_SGLIST */
34 
bcm_vk_dma_alloc(struct device * dev,struct bcm_vk_dma * dma,int direction,struct _vk_data * vkdata)35 static int bcm_vk_dma_alloc(struct device *dev,
36 			    struct bcm_vk_dma *dma,
37 			    int direction,
38 			    struct _vk_data *vkdata)
39 {
40 	dma_addr_t addr, sg_addr;
41 	int err;
42 	int i;
43 	int offset;
44 	u32 size;
45 	u32 remaining_size;
46 	u32 transfer_size;
47 	u64 data;
48 	unsigned long first, last;
49 	struct _vk_data *sgdata;
50 
51 	/* Get 64-bit user address */
52 	data = get_unaligned(&vkdata->address);
53 
54 	/* offset into first page */
55 	offset = offset_in_page(data);
56 
57 	/* Calculate number of pages */
58 	first = (data & PAGE_MASK) >> PAGE_SHIFT;
59 	last  = ((data + vkdata->size - 1) & PAGE_MASK) >> PAGE_SHIFT;
60 	dma->nr_pages = last - first + 1;
61 
62 	/* Allocate DMA pages */
63 	dma->pages = kmalloc_array(dma->nr_pages,
64 				   sizeof(struct page *),
65 				   GFP_KERNEL);
66 	if (!dma->pages)
67 		return -ENOMEM;
68 
69 	dev_dbg(dev, "Alloc DMA Pages [0x%llx+0x%x => %d pages]\n",
70 		data, vkdata->size, dma->nr_pages);
71 
72 	dma->direction = direction;
73 
74 	/* Get user pages into memory */
75 	err = get_user_pages_fast(data & PAGE_MASK,
76 				  dma->nr_pages,
77 				  direction == DMA_FROM_DEVICE,
78 				  dma->pages);
79 	if (err != dma->nr_pages) {
80 		dma->nr_pages = (err >= 0) ? err : 0;
81 		dev_err(dev, "get_user_pages_fast, err=%d [%d]\n",
82 			err, dma->nr_pages);
83 		return err < 0 ? err : -EINVAL;
84 	}
85 
86 	/* Max size of sg list is 1 per mapped page + fields at start */
87 	dma->sglen = (dma->nr_pages * sizeof(*sgdata)) +
88 		     (sizeof(u32) * SGLIST_VKDATA_START);
89 
90 	/* Allocate sglist */
91 	dma->sglist = dma_alloc_coherent(dev,
92 					 dma->sglen,
93 					 &dma->handle,
94 					 GFP_KERNEL);
95 	if (!dma->sglist)
96 		return -ENOMEM;
97 
98 	dma->sglist[SGLIST_NUM_SG] = 0;
99 	dma->sglist[SGLIST_TOTALSIZE] = vkdata->size;
100 	remaining_size = vkdata->size;
101 	sgdata = (struct _vk_data *)&dma->sglist[SGLIST_VKDATA_START];
102 
103 	/* Map all pages into DMA */
104 	size = min_t(size_t, PAGE_SIZE - offset, remaining_size);
105 	remaining_size -= size;
106 	sg_addr = dma_map_page(dev,
107 			       dma->pages[0],
108 			       offset,
109 			       size,
110 			       dma->direction);
111 	transfer_size = size;
112 	if (unlikely(dma_mapping_error(dev, sg_addr))) {
113 		__free_page(dma->pages[0]);
114 		return -EIO;
115 	}
116 
117 	for (i = 1; i < dma->nr_pages; i++) {
118 		size = min_t(size_t, PAGE_SIZE, remaining_size);
119 		remaining_size -= size;
120 		addr = dma_map_page(dev,
121 				    dma->pages[i],
122 				    0,
123 				    size,
124 				    dma->direction);
125 		if (unlikely(dma_mapping_error(dev, addr))) {
126 			__free_page(dma->pages[i]);
127 			return -EIO;
128 		}
129 
130 		/*
131 		 * Compress SG list entry when pages are contiguous
132 		 * and transfer size less or equal to BCM_VK_MAX_SGL_CHUNK
133 		 */
134 		if ((addr == (sg_addr + transfer_size)) &&
135 		    ((transfer_size + size) <= BCM_VK_MAX_SGL_CHUNK)) {
136 			/* pages are contiguous, add to same sg entry */
137 			transfer_size += size;
138 		} else {
139 			/* pages are not contiguous, write sg entry */
140 			sgdata->size = transfer_size;
141 			put_unaligned(sg_addr, (u64 *)&sgdata->address);
142 			dma->sglist[SGLIST_NUM_SG]++;
143 
144 			/* start new sg entry */
145 			sgdata++;
146 			sg_addr = addr;
147 			transfer_size = size;
148 		}
149 	}
150 	/* Write last sg list entry */
151 	sgdata->size = transfer_size;
152 	put_unaligned(sg_addr, (u64 *)&sgdata->address);
153 	dma->sglist[SGLIST_NUM_SG]++;
154 
155 	/* Update pointers and size field to point to sglist */
156 	put_unaligned((u64)dma->handle, &vkdata->address);
157 	vkdata->size = (dma->sglist[SGLIST_NUM_SG] * sizeof(*sgdata)) +
158 		       (sizeof(u32) * SGLIST_VKDATA_START);
159 
160 #ifdef BCM_VK_DUMP_SGLIST
161 	dev_dbg(dev,
162 		"sgl 0x%llx handle 0x%llx, sglen: 0x%x sgsize: 0x%x\n",
163 		(u64)dma->sglist,
164 		dma->handle,
165 		dma->sglen,
166 		vkdata->size);
167 	for (i = 0; i < vkdata->size / sizeof(u32); i++)
168 		dev_dbg(dev, "i:0x%x 0x%x\n", i, dma->sglist[i]);
169 #endif
170 
171 	return 0;
172 }
173 
bcm_vk_sg_alloc(struct device * dev,struct bcm_vk_dma * dma,int dir,struct _vk_data * vkdata,int num)174 int bcm_vk_sg_alloc(struct device *dev,
175 		    struct bcm_vk_dma *dma,
176 		    int dir,
177 		    struct _vk_data *vkdata,
178 		    int num)
179 {
180 	int i;
181 	int rc = -EINVAL;
182 
183 	/* Convert user addresses to DMA SG List */
184 	for (i = 0; i < num; i++) {
185 		if (vkdata[i].size && vkdata[i].address) {
186 			/*
187 			 * If both size and address are non-zero
188 			 * then DMA alloc.
189 			 */
190 			rc = bcm_vk_dma_alloc(dev,
191 					      &dma[i],
192 					      dir,
193 					      &vkdata[i]);
194 		} else if (vkdata[i].size ||
195 			   vkdata[i].address) {
196 			/*
197 			 * If one of size and address are zero
198 			 * there is a problem.
199 			 */
200 			dev_err(dev,
201 				"Invalid vkdata %x 0x%x 0x%llx\n",
202 				i, vkdata[i].size, vkdata[i].address);
203 			rc = -EINVAL;
204 		} else {
205 			/*
206 			 * If size and address are both zero
207 			 * don't convert, but return success.
208 			 */
209 			rc = 0;
210 		}
211 
212 		if (rc)
213 			goto fail_alloc;
214 	}
215 	return rc;
216 
217 fail_alloc:
218 	while (i > 0) {
219 		i--;
220 		if (dma[i].sglist)
221 			bcm_vk_dma_free(dev, &dma[i]);
222 	}
223 	return rc;
224 }
225 
bcm_vk_dma_free(struct device * dev,struct bcm_vk_dma * dma)226 static int bcm_vk_dma_free(struct device *dev, struct bcm_vk_dma *dma)
227 {
228 	dma_addr_t addr;
229 	int i;
230 	int num_sg;
231 	u32 size;
232 	struct _vk_data *vkdata;
233 
234 	dev_dbg(dev, "free sglist=%p sglen=0x%x\n", dma->sglist, dma->sglen);
235 
236 	/* Unmap all pages in the sglist */
237 	num_sg = dma->sglist[SGLIST_NUM_SG];
238 	vkdata = (struct _vk_data *)&dma->sglist[SGLIST_VKDATA_START];
239 	for (i = 0; i < num_sg; i++) {
240 		size = vkdata[i].size;
241 		addr = get_unaligned(&vkdata[i].address);
242 
243 		dma_unmap_page(dev, addr, size, dma->direction);
244 	}
245 
246 	/* Free allocated sglist */
247 	dma_free_coherent(dev, dma->sglen, dma->sglist, dma->handle);
248 
249 	/* Release lock on all pages */
250 	for (i = 0; i < dma->nr_pages; i++)
251 		put_page(dma->pages[i]);
252 
253 	/* Free allocated dma pages */
254 	kfree(dma->pages);
255 	dma->sglist = NULL;
256 
257 	return 0;
258 }
259 
bcm_vk_sg_free(struct device * dev,struct bcm_vk_dma * dma,int num,int * proc_cnt)260 int bcm_vk_sg_free(struct device *dev, struct bcm_vk_dma *dma, int num,
261 		   int *proc_cnt)
262 {
263 	int i;
264 
265 	*proc_cnt = 0;
266 	/* Unmap and free all pages and sglists */
267 	for (i = 0; i < num; i++) {
268 		if (dma[i].sglist) {
269 			bcm_vk_dma_free(dev, &dma[i]);
270 			*proc_cnt += 1;
271 		}
272 	}
273 
274 	return 0;
275 }
276