xref: /linux/drivers/dma/ste_dma40_ll.c (revision 4b132aacb0768ac1e652cf517097ea6f237214b9)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * Copyright (C) ST-Ericsson SA 2007-2010
4  * Author: Per Forlin <per.forlin@stericsson.com> for ST-Ericsson
5  * Author: Jonas Aaberg <jonas.aberg@stericsson.com> for ST-Ericsson
6  */
7 
8 #include <linux/kernel.h>
9 #include <linux/dmaengine.h>
10 
11 #include "ste_dma40.h"
12 #include "ste_dma40_ll.h"
13 
14 static u8 d40_width_to_bits(enum dma_slave_buswidth width)
15 {
16 	if (width == DMA_SLAVE_BUSWIDTH_1_BYTE)
17 		return STEDMA40_ESIZE_8_BIT;
18 	else if (width == DMA_SLAVE_BUSWIDTH_2_BYTES)
19 		return STEDMA40_ESIZE_16_BIT;
20 	else if (width == DMA_SLAVE_BUSWIDTH_8_BYTES)
21 		return STEDMA40_ESIZE_64_BIT;
22 	else
23 		return STEDMA40_ESIZE_32_BIT;
24 }
25 
26 /* Sets up proper LCSP1 and LCSP3 register for a logical channel */
27 void d40_log_cfg(struct stedma40_chan_cfg *cfg,
28 		 u32 *lcsp1, u32 *lcsp3)
29 {
30 	u32 l3 = 0; /* dst */
31 	u32 l1 = 0; /* src */
32 
33 	/* src is mem? -> increase address pos */
34 	if (cfg->dir ==  DMA_MEM_TO_DEV ||
35 	    cfg->dir ==  DMA_MEM_TO_MEM)
36 		l1 |= BIT(D40_MEM_LCSP1_SCFG_INCR_POS);
37 
38 	/* dst is mem? -> increase address pos */
39 	if (cfg->dir ==  DMA_DEV_TO_MEM ||
40 	    cfg->dir ==  DMA_MEM_TO_MEM)
41 		l3 |= BIT(D40_MEM_LCSP3_DCFG_INCR_POS);
42 
43 	/* src is hw? -> master port 1 */
44 	if (cfg->dir ==  DMA_DEV_TO_MEM ||
45 	    cfg->dir ==  DMA_DEV_TO_DEV)
46 		l1 |= BIT(D40_MEM_LCSP1_SCFG_MST_POS);
47 
48 	/* dst is hw? -> master port 1 */
49 	if (cfg->dir ==  DMA_MEM_TO_DEV ||
50 	    cfg->dir ==  DMA_DEV_TO_DEV)
51 		l3 |= BIT(D40_MEM_LCSP3_DCFG_MST_POS);
52 
53 	l3 |= BIT(D40_MEM_LCSP3_DCFG_EIM_POS);
54 	l3 |= cfg->dst_info.psize << D40_MEM_LCSP3_DCFG_PSIZE_POS;
55 	l3 |= d40_width_to_bits(cfg->dst_info.data_width)
56 		<< D40_MEM_LCSP3_DCFG_ESIZE_POS;
57 
58 	l1 |= BIT(D40_MEM_LCSP1_SCFG_EIM_POS);
59 	l1 |= cfg->src_info.psize << D40_MEM_LCSP1_SCFG_PSIZE_POS;
60 	l1 |= d40_width_to_bits(cfg->src_info.data_width)
61 		<< D40_MEM_LCSP1_SCFG_ESIZE_POS;
62 
63 	*lcsp1 = l1;
64 	*lcsp3 = l3;
65 
66 }
67 
68 void d40_phy_cfg(struct stedma40_chan_cfg *cfg, u32 *src_cfg, u32 *dst_cfg)
69 {
70 	u32 src = 0;
71 	u32 dst = 0;
72 
73 	if ((cfg->dir == DMA_DEV_TO_MEM) ||
74 	    (cfg->dir == DMA_DEV_TO_DEV)) {
75 		/* Set master port to 1 */
76 		src |= BIT(D40_SREG_CFG_MST_POS);
77 		src |= D40_TYPE_TO_EVENT(cfg->dev_type);
78 
79 		if (cfg->src_info.flow_ctrl == STEDMA40_NO_FLOW_CTRL)
80 			src |= BIT(D40_SREG_CFG_PHY_TM_POS);
81 		else
82 			src |= 3 << D40_SREG_CFG_PHY_TM_POS;
83 	}
84 	if ((cfg->dir == DMA_MEM_TO_DEV) ||
85 	    (cfg->dir == DMA_DEV_TO_DEV)) {
86 		/* Set master port to 1 */
87 		dst |= BIT(D40_SREG_CFG_MST_POS);
88 		dst |= D40_TYPE_TO_EVENT(cfg->dev_type);
89 
90 		if (cfg->dst_info.flow_ctrl == STEDMA40_NO_FLOW_CTRL)
91 			dst |= BIT(D40_SREG_CFG_PHY_TM_POS);
92 		else
93 			dst |= 3 << D40_SREG_CFG_PHY_TM_POS;
94 	}
95 	/* Interrupt on end of transfer for destination */
96 	dst |= BIT(D40_SREG_CFG_TIM_POS);
97 
98 	/* Generate interrupt on error */
99 	src |= BIT(D40_SREG_CFG_EIM_POS);
100 	dst |= BIT(D40_SREG_CFG_EIM_POS);
101 
102 	/* PSIZE */
103 	if (cfg->src_info.psize != STEDMA40_PSIZE_PHY_1) {
104 		src |= BIT(D40_SREG_CFG_PHY_PEN_POS);
105 		src |= cfg->src_info.psize << D40_SREG_CFG_PSIZE_POS;
106 	}
107 	if (cfg->dst_info.psize != STEDMA40_PSIZE_PHY_1) {
108 		dst |= BIT(D40_SREG_CFG_PHY_PEN_POS);
109 		dst |= cfg->dst_info.psize << D40_SREG_CFG_PSIZE_POS;
110 	}
111 
112 	/* Element size */
113 	src |= d40_width_to_bits(cfg->src_info.data_width)
114 		<< D40_SREG_CFG_ESIZE_POS;
115 	dst |= d40_width_to_bits(cfg->dst_info.data_width)
116 		<< D40_SREG_CFG_ESIZE_POS;
117 
118 	/* Set the priority bit to high for the physical channel */
119 	if (cfg->high_priority) {
120 		src |= BIT(D40_SREG_CFG_PRI_POS);
121 		dst |= BIT(D40_SREG_CFG_PRI_POS);
122 	}
123 
124 	if (cfg->src_info.big_endian)
125 		src |= BIT(D40_SREG_CFG_LBE_POS);
126 	if (cfg->dst_info.big_endian)
127 		dst |= BIT(D40_SREG_CFG_LBE_POS);
128 
129 	*src_cfg = src;
130 	*dst_cfg = dst;
131 }
132 
133 static int d40_phy_fill_lli(struct d40_phy_lli *lli,
134 			    dma_addr_t data,
135 			    u32 data_size,
136 			    dma_addr_t next_lli,
137 			    u32 reg_cfg,
138 			    struct stedma40_half_channel_info *info,
139 			    unsigned int flags)
140 {
141 	bool addr_inc = flags & LLI_ADDR_INC;
142 	bool term_int = flags & LLI_TERM_INT;
143 	unsigned int data_width = info->data_width;
144 	int psize = info->psize;
145 	int num_elems;
146 
147 	if (psize == STEDMA40_PSIZE_PHY_1)
148 		num_elems = 1;
149 	else
150 		num_elems = 2 << psize;
151 
152 	/* Must be aligned */
153 	if (!IS_ALIGNED(data, data_width))
154 		return -EINVAL;
155 
156 	/* Transfer size can't be smaller than (num_elms * elem_size) */
157 	if (data_size < num_elems * data_width)
158 		return -EINVAL;
159 
160 	/* The number of elements. IE now many chunks */
161 	lli->reg_elt = (data_size / data_width) << D40_SREG_ELEM_PHY_ECNT_POS;
162 
163 	/*
164 	 * Distance to next element sized entry.
165 	 * Usually the size of the element unless you want gaps.
166 	 */
167 	if (addr_inc)
168 		lli->reg_elt |= data_width << D40_SREG_ELEM_PHY_EIDX_POS;
169 
170 	/* Where the data is */
171 	lli->reg_ptr = data;
172 	lli->reg_cfg = reg_cfg;
173 
174 	/* If this scatter list entry is the last one, no next link */
175 	if (next_lli == 0)
176 		lli->reg_lnk = BIT(D40_SREG_LNK_PHY_TCP_POS);
177 	else
178 		lli->reg_lnk = next_lli;
179 
180 	/* Set/clear interrupt generation on this link item.*/
181 	if (term_int)
182 		lli->reg_cfg |= BIT(D40_SREG_CFG_TIM_POS);
183 	else
184 		lli->reg_cfg &= ~BIT(D40_SREG_CFG_TIM_POS);
185 
186 	/*
187 	 * Post link - D40_SREG_LNK_PHY_PRE_POS = 0
188 	 * Relink happens after transfer completion.
189 	 */
190 
191 	return 0;
192 }
193 
194 static int d40_seg_size(int size, int data_width1, int data_width2)
195 {
196 	u32 max_w = max(data_width1, data_width2);
197 	u32 min_w = min(data_width1, data_width2);
198 	u32 seg_max = ALIGN(STEDMA40_MAX_SEG_SIZE * min_w, max_w);
199 
200 	if (seg_max > STEDMA40_MAX_SEG_SIZE)
201 		seg_max -= max_w;
202 
203 	if (size <= seg_max)
204 		return size;
205 
206 	if (size <= 2 * seg_max)
207 		return ALIGN(size / 2, max_w);
208 
209 	return seg_max;
210 }
211 
212 static struct d40_phy_lli *
213 d40_phy_buf_to_lli(struct d40_phy_lli *lli, dma_addr_t addr, u32 size,
214 		   dma_addr_t lli_phys, dma_addr_t first_phys, u32 reg_cfg,
215 		   struct stedma40_half_channel_info *info,
216 		   struct stedma40_half_channel_info *otherinfo,
217 		   unsigned long flags)
218 {
219 	bool lastlink = flags & LLI_LAST_LINK;
220 	bool addr_inc = flags & LLI_ADDR_INC;
221 	bool term_int = flags & LLI_TERM_INT;
222 	bool cyclic = flags & LLI_CYCLIC;
223 	int err;
224 	dma_addr_t next = lli_phys;
225 	int size_rest = size;
226 	int size_seg = 0;
227 
228 	/*
229 	 * This piece may be split up based on d40_seg_size(); we only want the
230 	 * term int on the last part.
231 	 */
232 	if (term_int)
233 		flags &= ~LLI_TERM_INT;
234 
235 	do {
236 		size_seg = d40_seg_size(size_rest, info->data_width,
237 					otherinfo->data_width);
238 		size_rest -= size_seg;
239 
240 		if (size_rest == 0 && term_int)
241 			flags |= LLI_TERM_INT;
242 
243 		if (size_rest == 0 && lastlink)
244 			next = cyclic ? first_phys : 0;
245 		else
246 			next = ALIGN(next + sizeof(struct d40_phy_lli),
247 				     D40_LLI_ALIGN);
248 
249 		err = d40_phy_fill_lli(lli, addr, size_seg, next,
250 				       reg_cfg, info, flags);
251 
252 		if (err)
253 			goto err;
254 
255 		lli++;
256 		if (addr_inc)
257 			addr += size_seg;
258 	} while (size_rest);
259 
260 	return lli;
261 
262 err:
263 	return NULL;
264 }
265 
266 int d40_phy_sg_to_lli(struct scatterlist *sg,
267 		      int sg_len,
268 		      dma_addr_t target,
269 		      struct d40_phy_lli *lli_sg,
270 		      dma_addr_t lli_phys,
271 		      u32 reg_cfg,
272 		      struct stedma40_half_channel_info *info,
273 		      struct stedma40_half_channel_info *otherinfo,
274 		      unsigned long flags)
275 {
276 	int total_size = 0;
277 	int i;
278 	struct scatterlist *current_sg = sg;
279 	struct d40_phy_lli *lli = lli_sg;
280 	dma_addr_t l_phys = lli_phys;
281 
282 	if (!target)
283 		flags |= LLI_ADDR_INC;
284 
285 	for_each_sg(sg, current_sg, sg_len, i) {
286 		dma_addr_t sg_addr = sg_dma_address(current_sg);
287 		unsigned int len = sg_dma_len(current_sg);
288 		dma_addr_t dst = target ?: sg_addr;
289 
290 		total_size += sg_dma_len(current_sg);
291 
292 		if (i == sg_len - 1)
293 			flags |= LLI_TERM_INT | LLI_LAST_LINK;
294 
295 		l_phys = ALIGN(lli_phys + (lli - lli_sg) *
296 			       sizeof(struct d40_phy_lli), D40_LLI_ALIGN);
297 
298 		lli = d40_phy_buf_to_lli(lli, dst, len, l_phys, lli_phys,
299 					 reg_cfg, info, otherinfo, flags);
300 
301 		if (lli == NULL)
302 			return -EINVAL;
303 	}
304 
305 	return total_size;
306 }
307 
308 
309 /* DMA logical lli operations */
310 
311 static void d40_log_lli_link(struct d40_log_lli *lli_dst,
312 			     struct d40_log_lli *lli_src,
313 			     int next, unsigned int flags)
314 {
315 	bool interrupt = flags & LLI_TERM_INT;
316 	u32 slos = 0;
317 	u32 dlos = 0;
318 
319 	if (next != -EINVAL) {
320 		slos = next * 2;
321 		dlos = next * 2 + 1;
322 	}
323 
324 	if (interrupt) {
325 		lli_dst->lcsp13 |= D40_MEM_LCSP1_SCFG_TIM_MASK;
326 		lli_dst->lcsp13 |= D40_MEM_LCSP3_DTCP_MASK;
327 	}
328 
329 	lli_src->lcsp13 = (lli_src->lcsp13 & ~D40_MEM_LCSP1_SLOS_MASK) |
330 		(slos << D40_MEM_LCSP1_SLOS_POS);
331 
332 	lli_dst->lcsp13 = (lli_dst->lcsp13 & ~D40_MEM_LCSP1_SLOS_MASK) |
333 		(dlos << D40_MEM_LCSP1_SLOS_POS);
334 }
335 
336 void d40_log_lli_lcpa_write(struct d40_log_lli_full *lcpa,
337 			   struct d40_log_lli *lli_dst,
338 			   struct d40_log_lli *lli_src,
339 			   int next, unsigned int flags)
340 {
341 	d40_log_lli_link(lli_dst, lli_src, next, flags);
342 
343 	writel_relaxed(lli_src->lcsp02, &lcpa[0].lcsp0);
344 	writel_relaxed(lli_src->lcsp13, &lcpa[0].lcsp1);
345 	writel_relaxed(lli_dst->lcsp02, &lcpa[0].lcsp2);
346 	writel_relaxed(lli_dst->lcsp13, &lcpa[0].lcsp3);
347 }
348 
349 void d40_log_lli_lcla_write(struct d40_log_lli *lcla,
350 			   struct d40_log_lli *lli_dst,
351 			   struct d40_log_lli *lli_src,
352 			   int next, unsigned int flags)
353 {
354 	d40_log_lli_link(lli_dst, lli_src, next, flags);
355 
356 	writel_relaxed(lli_src->lcsp02, &lcla[0].lcsp02);
357 	writel_relaxed(lli_src->lcsp13, &lcla[0].lcsp13);
358 	writel_relaxed(lli_dst->lcsp02, &lcla[1].lcsp02);
359 	writel_relaxed(lli_dst->lcsp13, &lcla[1].lcsp13);
360 }
361 
362 static void d40_log_fill_lli(struct d40_log_lli *lli,
363 			     dma_addr_t data, u32 data_size,
364 			     u32 reg_cfg,
365 			     u32 data_width,
366 			     unsigned int flags)
367 {
368 	bool addr_inc = flags & LLI_ADDR_INC;
369 
370 	lli->lcsp13 = reg_cfg;
371 
372 	/* The number of elements to transfer */
373 	lli->lcsp02 = ((data_size / data_width) <<
374 		       D40_MEM_LCSP0_ECNT_POS) & D40_MEM_LCSP0_ECNT_MASK;
375 
376 	BUG_ON((data_size / data_width) > STEDMA40_MAX_SEG_SIZE);
377 
378 	/* 16 LSBs address of the current element */
379 	lli->lcsp02 |= data & D40_MEM_LCSP0_SPTR_MASK;
380 	/* 16 MSBs address of the current element */
381 	lli->lcsp13 |= data & D40_MEM_LCSP1_SPTR_MASK;
382 
383 	if (addr_inc)
384 		lli->lcsp13 |= D40_MEM_LCSP1_SCFG_INCR_MASK;
385 
386 }
387 
388 static struct d40_log_lli *d40_log_buf_to_lli(struct d40_log_lli *lli_sg,
389 				       dma_addr_t addr,
390 				       int size,
391 				       u32 lcsp13, /* src or dst*/
392 				       u32 data_width1,
393 				       u32 data_width2,
394 				       unsigned int flags)
395 {
396 	bool addr_inc = flags & LLI_ADDR_INC;
397 	struct d40_log_lli *lli = lli_sg;
398 	int size_rest = size;
399 	int size_seg = 0;
400 
401 	do {
402 		size_seg = d40_seg_size(size_rest, data_width1, data_width2);
403 		size_rest -= size_seg;
404 
405 		d40_log_fill_lli(lli,
406 				 addr,
407 				 size_seg,
408 				 lcsp13, data_width1,
409 				 flags);
410 		if (addr_inc)
411 			addr += size_seg;
412 		lli++;
413 	} while (size_rest);
414 
415 	return lli;
416 }
417 
418 int d40_log_sg_to_lli(struct scatterlist *sg,
419 		      int sg_len,
420 		      dma_addr_t dev_addr,
421 		      struct d40_log_lli *lli_sg,
422 		      u32 lcsp13, /* src or dst*/
423 		      u32 data_width1, u32 data_width2)
424 {
425 	int total_size = 0;
426 	struct scatterlist *current_sg = sg;
427 	int i;
428 	struct d40_log_lli *lli = lli_sg;
429 	unsigned long flags = 0;
430 
431 	if (!dev_addr)
432 		flags |= LLI_ADDR_INC;
433 
434 	for_each_sg(sg, current_sg, sg_len, i) {
435 		dma_addr_t sg_addr = sg_dma_address(current_sg);
436 		unsigned int len = sg_dma_len(current_sg);
437 		dma_addr_t addr = dev_addr ?: sg_addr;
438 
439 		total_size += sg_dma_len(current_sg);
440 
441 		lli = d40_log_buf_to_lli(lli, addr, len,
442 					 lcsp13,
443 					 data_width1,
444 					 data_width2,
445 					 flags);
446 	}
447 
448 	return total_size;
449 }
450