xref: /linux/drivers/soc/fsl/qbman/qman_priv.h (revision e5c86679d5e864947a52fb31e45a425dea3e7fa9)
1 /* Copyright 2008 - 2016 Freescale Semiconductor, Inc.
2  *
3  * Redistribution and use in source and binary forms, with or without
4  * modification, are permitted provided that the following conditions are met:
5  *     * Redistributions of source code must retain the above copyright
6  *	 notice, this list of conditions and the following disclaimer.
7  *     * Redistributions in binary form must reproduce the above copyright
8  *	 notice, this list of conditions and the following disclaimer in the
9  *	 documentation and/or other materials provided with the distribution.
10  *     * Neither the name of Freescale Semiconductor nor the
11  *	 names of its contributors may be used to endorse or promote products
12  *	 derived from this software without specific prior written permission.
13  *
14  * ALTERNATIVELY, this software may be distributed under the terms of the
15  * GNU General Public License ("GPL") as published by the Free Software
16  * Foundation, either version 2 of that License or (at your option) any
17  * later version.
18  *
19  * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
20  * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
21  * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
22  * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
23  * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
24  * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
25  * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
26  * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
27  * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
28  * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
29  */
30 
31 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
32 
33 #include "dpaa_sys.h"
34 
35 #include <soc/fsl/qman.h>
36 #include <linux/iommu.h>
37 
38 #if defined(CONFIG_FSL_PAMU)
39 #include <asm/fsl_pamu_stash.h>
40 #endif
41 
42 struct qm_mcr_querywq {
43 	u8 verb;
44 	u8 result;
45 	u16 channel_wq; /* ignores wq (3 lsbits): _res[0-2] */
46 	u8 __reserved[28];
47 	u32 wq_len[8];
48 } __packed;
49 
50 static inline u16 qm_mcr_querywq_get_chan(const struct qm_mcr_querywq *wq)
51 {
52 	return wq->channel_wq >> 3;
53 }
54 
55 struct __qm_mcr_querycongestion {
56 	u32 state[8];
57 };
58 
59 /* "Query Congestion Group State" */
60 struct qm_mcr_querycongestion {
61 	u8 verb;
62 	u8 result;
63 	u8 __reserved[30];
64 	/* Access this struct using qman_cgrs_get() */
65 	struct __qm_mcr_querycongestion state;
66 } __packed;
67 
68 /* "Query CGR" */
69 struct qm_mcr_querycgr {
70 	u8 verb;
71 	u8 result;
72 	u16 __reserved1;
73 	struct __qm_mc_cgr cgr; /* CGR fields */
74 	u8 __reserved2[6];
75 	u8 i_bcnt_hi;	/* high 8-bits of 40-bit "Instant" */
76 	__be32 i_bcnt_lo;	/* low 32-bits of 40-bit */
77 	u8 __reserved3[3];
78 	u8 a_bcnt_hi;	/* high 8-bits of 40-bit "Average" */
79 	__be32 a_bcnt_lo;	/* low 32-bits of 40-bit */
80 	__be32 cscn_targ_swp[4];
81 } __packed;
82 
83 static inline u64 qm_mcr_querycgr_i_get64(const struct qm_mcr_querycgr *q)
84 {
85 	return ((u64)q->i_bcnt_hi << 32) | be32_to_cpu(q->i_bcnt_lo);
86 }
87 static inline u64 qm_mcr_querycgr_a_get64(const struct qm_mcr_querycgr *q)
88 {
89 	return ((u64)q->a_bcnt_hi << 32) | be32_to_cpu(q->a_bcnt_lo);
90 }
91 
92 /* "Query FQ Non-Programmable Fields" */
93 
94 struct qm_mcr_queryfq_np {
95 	u8 verb;
96 	u8 result;
97 	u8 __reserved1;
98 	u8 state;		/* QM_MCR_NP_STATE_*** */
99 	u32 fqd_link;		/* 24-bit, _res2[24-31] */
100 	u16 odp_seq;		/* 14-bit, _res3[14-15] */
101 	u16 orp_nesn;		/* 14-bit, _res4[14-15] */
102 	u16 orp_ea_hseq;	/* 15-bit, _res5[15] */
103 	u16 orp_ea_tseq;	/* 15-bit, _res6[15] */
104 	u32 orp_ea_hptr;	/* 24-bit, _res7[24-31] */
105 	u32 orp_ea_tptr;	/* 24-bit, _res8[24-31] */
106 	u32 pfdr_hptr;		/* 24-bit, _res9[24-31] */
107 	u32 pfdr_tptr;		/* 24-bit, _res10[24-31] */
108 	u8 __reserved2[5];
109 	u8 is;			/* 1-bit, _res12[1-7] */
110 	u16 ics_surp;
111 	u32 byte_cnt;
112 	u32 frm_cnt;		/* 24-bit, _res13[24-31] */
113 	u32 __reserved3;
114 	u16 ra1_sfdr;		/* QM_MCR_NP_RA1_*** */
115 	u16 ra2_sfdr;		/* QM_MCR_NP_RA2_*** */
116 	u16 __reserved4;
117 	u16 od1_sfdr;		/* QM_MCR_NP_OD1_*** */
118 	u16 od2_sfdr;		/* QM_MCR_NP_OD2_*** */
119 	u16 od3_sfdr;		/* QM_MCR_NP_OD3_*** */
120 } __packed;
121 
122 #define QM_MCR_NP_STATE_FE		0x10
123 #define QM_MCR_NP_STATE_R		0x08
124 #define QM_MCR_NP_STATE_MASK		0x07	/* Reads FQD::STATE; */
125 #define QM_MCR_NP_STATE_OOS		0x00
126 #define QM_MCR_NP_STATE_RETIRED		0x01
127 #define QM_MCR_NP_STATE_TEN_SCHED	0x02
128 #define QM_MCR_NP_STATE_TRU_SCHED	0x03
129 #define QM_MCR_NP_STATE_PARKED		0x04
130 #define QM_MCR_NP_STATE_ACTIVE		0x05
131 #define QM_MCR_NP_PTR_MASK		0x07ff	/* for RA[12] & OD[123] */
132 #define QM_MCR_NP_RA1_NRA(v)		(((v) >> 14) & 0x3)	/* FQD::NRA */
133 #define QM_MCR_NP_RA2_IT(v)		(((v) >> 14) & 0x1)	/* FQD::IT */
134 #define QM_MCR_NP_OD1_NOD(v)		(((v) >> 14) & 0x3)	/* FQD::NOD */
135 #define QM_MCR_NP_OD3_NPC(v)		(((v) >> 14) & 0x3)	/* FQD::NPC */
136 
137 enum qm_mcr_queryfq_np_masks {
138 	qm_mcr_fqd_link_mask = BIT(24)-1,
139 	qm_mcr_odp_seq_mask = BIT(14)-1,
140 	qm_mcr_orp_nesn_mask = BIT(14)-1,
141 	qm_mcr_orp_ea_hseq_mask = BIT(15)-1,
142 	qm_mcr_orp_ea_tseq_mask = BIT(15)-1,
143 	qm_mcr_orp_ea_hptr_mask = BIT(24)-1,
144 	qm_mcr_orp_ea_tptr_mask = BIT(24)-1,
145 	qm_mcr_pfdr_hptr_mask = BIT(24)-1,
146 	qm_mcr_pfdr_tptr_mask = BIT(24)-1,
147 	qm_mcr_is_mask = BIT(1)-1,
148 	qm_mcr_frm_cnt_mask = BIT(24)-1,
149 };
150 #define qm_mcr_np_get(np, field) \
151 	((np)->field & (qm_mcr_##field##_mask))
152 
153 /* Congestion Groups */
154 
155 /*
156  * This wrapper represents a bit-array for the state of the 256 QMan congestion
157  * groups. Is also used as a *mask* for congestion groups, eg. so we ignore
158  * those that don't concern us. We harness the structure and accessor details
159  * already used in the management command to query congestion groups.
160  */
161 #define CGR_BITS_PER_WORD 5
162 #define CGR_WORD(x)	((x) >> CGR_BITS_PER_WORD)
163 #define CGR_BIT(x)	(BIT(31) >> ((x) & 0x1f))
164 #define CGR_NUM	(sizeof(struct __qm_mcr_querycongestion) << 3)
165 
166 struct qman_cgrs {
167 	struct __qm_mcr_querycongestion q;
168 };
169 
170 static inline void qman_cgrs_init(struct qman_cgrs *c)
171 {
172 	memset(c, 0, sizeof(*c));
173 }
174 
175 static inline void qman_cgrs_fill(struct qman_cgrs *c)
176 {
177 	memset(c, 0xff, sizeof(*c));
178 }
179 
180 static inline int qman_cgrs_get(struct qman_cgrs *c, u8 cgr)
181 {
182 	return c->q.state[CGR_WORD(cgr)] & CGR_BIT(cgr);
183 }
184 
185 static inline void qman_cgrs_cp(struct qman_cgrs *dest,
186 				const struct qman_cgrs *src)
187 {
188 	*dest = *src;
189 }
190 
191 static inline void qman_cgrs_and(struct qman_cgrs *dest,
192 			const struct qman_cgrs *a, const struct qman_cgrs *b)
193 {
194 	int ret;
195 	u32 *_d = dest->q.state;
196 	const u32 *_a = a->q.state;
197 	const u32 *_b = b->q.state;
198 
199 	for (ret = 0; ret < 8; ret++)
200 		*_d++ = *_a++ & *_b++;
201 }
202 
203 static inline void qman_cgrs_xor(struct qman_cgrs *dest,
204 			const struct qman_cgrs *a, const struct qman_cgrs *b)
205 {
206 	int ret;
207 	u32 *_d = dest->q.state;
208 	const u32 *_a = a->q.state;
209 	const u32 *_b = b->q.state;
210 
211 	for (ret = 0; ret < 8; ret++)
212 		*_d++ = *_a++ ^ *_b++;
213 }
214 
215 void qman_init_cgr_all(void);
216 
217 struct qm_portal_config {
218 	/*
219 	 * Corenet portal addresses;
220 	 * [0]==cache-enabled, [1]==cache-inhibited.
221 	 */
222 	void __iomem *addr_virt[2];
223 	struct device *dev;
224 	struct iommu_domain *iommu_domain;
225 	/* Allow these to be joined in lists */
226 	struct list_head list;
227 	/* User-visible portal configuration settings */
228 	/* portal is affined to this cpu */
229 	int cpu;
230 	/* portal interrupt line */
231 	int irq;
232 	/*
233 	 * the portal's dedicated channel id, used initialising
234 	 * frame queues to target this portal when scheduled
235 	 */
236 	u16 channel;
237 	/*
238 	 * mask of pool channels this portal has dequeue access to
239 	 * (using QM_SDQCR_CHANNELS_POOL(n) for the bitmask)
240 	 */
241 	u32 pools;
242 };
243 
244 /* Revision info (for errata and feature handling) */
245 #define QMAN_REV11 0x0101
246 #define QMAN_REV12 0x0102
247 #define QMAN_REV20 0x0200
248 #define QMAN_REV30 0x0300
249 #define QMAN_REV31 0x0301
250 extern u16 qman_ip_rev; /* 0 if uninitialised, otherwise QMAN_REVx */
251 
252 #define QM_FQID_RANGE_START 1 /* FQID 0 reserved for internal use */
253 extern struct gen_pool *qm_fqalloc; /* FQID allocator */
254 extern struct gen_pool *qm_qpalloc; /* pool-channel allocator */
255 extern struct gen_pool *qm_cgralloc; /* CGR ID allocator */
256 u32 qm_get_pools_sdqcr(void);
257 
258 int qman_wq_alloc(void);
259 void qman_liodn_fixup(u16 channel);
260 void qman_set_sdest(u16 channel, unsigned int cpu_idx);
261 
262 struct qman_portal *qman_create_affine_portal(
263 			const struct qm_portal_config *config,
264 			const struct qman_cgrs *cgrs);
265 const struct qm_portal_config *qman_destroy_affine_portal(void);
266 
267 /*
268  * qman_query_fq - Queries FQD fields (via h/w query command)
269  * @fq: the frame queue object to be queried
270  * @fqd: storage for the queried FQD fields
271  */
272 int qman_query_fq(struct qman_fq *fq, struct qm_fqd *fqd);
273 
274 /*
275  * For qman_volatile_dequeue(); Choose one PRECEDENCE. EXACT is optional. Use
276  * NUMFRAMES(n) (6-bit) or NUMFRAMES_TILLEMPTY to fill in the frame-count. Use
277  * FQID(n) to fill in the frame queue ID.
278  */
279 #define QM_VDQCR_PRECEDENCE_VDQCR	0x0
280 #define QM_VDQCR_PRECEDENCE_SDQCR	0x80000000
281 #define QM_VDQCR_EXACT			0x40000000
282 #define QM_VDQCR_NUMFRAMES_MASK		0x3f000000
283 #define QM_VDQCR_NUMFRAMES_SET(n)	(((n) & 0x3f) << 24)
284 #define QM_VDQCR_NUMFRAMES_GET(n)	(((n) >> 24) & 0x3f)
285 #define QM_VDQCR_NUMFRAMES_TILLEMPTY	QM_VDQCR_NUMFRAMES_SET(0)
286 
287 #define QMAN_VOLATILE_FLAG_WAIT	     0x00000001 /* wait if VDQCR is in use */
288 #define QMAN_VOLATILE_FLAG_WAIT_INT  0x00000002 /* if wait, interruptible? */
289 #define QMAN_VOLATILE_FLAG_FINISH    0x00000004 /* wait till VDQCR completes */
290 
291 /*
292  * qman_volatile_dequeue - Issue a volatile dequeue command
293  * @fq: the frame queue object to dequeue from
294  * @flags: a bit-mask of QMAN_VOLATILE_FLAG_*** options
295  * @vdqcr: bit mask of QM_VDQCR_*** options, as per qm_dqrr_vdqcr_set()
296  *
297  * Attempts to lock access to the portal's VDQCR volatile dequeue functionality.
298  * The function will block and sleep if QMAN_VOLATILE_FLAG_WAIT is specified and
299  * the VDQCR is already in use, otherwise returns non-zero for failure. If
300  * QMAN_VOLATILE_FLAG_FINISH is specified, the function will only return once
301  * the VDQCR command has finished executing (ie. once the callback for the last
302  * DQRR entry resulting from the VDQCR command has been called). If not using
303  * the FINISH flag, completion can be determined either by detecting the
304  * presence of the QM_DQRR_STAT_UNSCHEDULED and QM_DQRR_STAT_DQCR_EXPIRED bits
305  * in the "stat" parameter passed to the FQ's dequeue callback, or by waiting
306  * for the QMAN_FQ_STATE_VDQCR bit to disappear.
307  */
308 int qman_volatile_dequeue(struct qman_fq *fq, u32 flags, u32 vdqcr);
309 
310 int qman_alloc_fq_table(u32 num_fqids);
311 
312 /*   QMan s/w corenet portal, low-level i/face	 */
313 
314 /*
315  * For qm_dqrr_sdqcr_set(); Choose one SOURCE. Choose one COUNT. Choose one
316  * dequeue TYPE. Choose TOKEN (8-bit).
317  * If SOURCE == CHANNELS,
318  *   Choose CHANNELS_DEDICATED and/or CHANNELS_POOL(n).
319  *   You can choose DEDICATED_PRECEDENCE if the portal channel should have
320  *   priority.
321  * If SOURCE == SPECIFICWQ,
322  *     Either select the work-queue ID with SPECIFICWQ_WQ(), or select the
323  *     channel (SPECIFICWQ_DEDICATED or SPECIFICWQ_POOL()) and specify the
324  *     work-queue priority (0-7) with SPECIFICWQ_WQ() - either way, you get the
325  *     same value.
326  */
327 #define QM_SDQCR_SOURCE_CHANNELS	0x0
328 #define QM_SDQCR_SOURCE_SPECIFICWQ	0x40000000
329 #define QM_SDQCR_COUNT_EXACT1		0x0
330 #define QM_SDQCR_COUNT_UPTO3		0x20000000
331 #define QM_SDQCR_DEDICATED_PRECEDENCE	0x10000000
332 #define QM_SDQCR_TYPE_MASK		0x03000000
333 #define QM_SDQCR_TYPE_NULL		0x0
334 #define QM_SDQCR_TYPE_PRIO_QOS		0x01000000
335 #define QM_SDQCR_TYPE_ACTIVE_QOS	0x02000000
336 #define QM_SDQCR_TYPE_ACTIVE		0x03000000
337 #define QM_SDQCR_TOKEN_MASK		0x00ff0000
338 #define QM_SDQCR_TOKEN_SET(v)		(((v) & 0xff) << 16)
339 #define QM_SDQCR_TOKEN_GET(v)		(((v) >> 16) & 0xff)
340 #define QM_SDQCR_CHANNELS_DEDICATED	0x00008000
341 #define QM_SDQCR_SPECIFICWQ_MASK	0x000000f7
342 #define QM_SDQCR_SPECIFICWQ_DEDICATED	0x00000000
343 #define QM_SDQCR_SPECIFICWQ_POOL(n)	((n) << 4)
344 #define QM_SDQCR_SPECIFICWQ_WQ(n)	(n)
345 
346 /* For qm_dqrr_vdqcr_set(): use FQID(n) to fill in the frame queue ID */
347 #define QM_VDQCR_FQID_MASK		0x00ffffff
348 #define QM_VDQCR_FQID(n)		((n) & QM_VDQCR_FQID_MASK)
349 
350 /*
351  * Used by all portal interrupt registers except 'inhibit'
352  * Channels with frame availability
353  */
354 #define QM_PIRQ_DQAVAIL	0x0000ffff
355 
356 /* The DQAVAIL interrupt fields break down into these bits; */
357 #define QM_DQAVAIL_PORTAL	0x8000		/* Portal channel */
358 #define QM_DQAVAIL_POOL(n)	(0x8000 >> (n))	/* Pool channel, n==[1..15] */
359 #define QM_DQAVAIL_MASK		0xffff
360 /* This mask contains all the "irqsource" bits visible to API users */
361 #define QM_PIRQ_VISIBLE	(QM_PIRQ_SLOW | QM_PIRQ_DQRI)
362 
363 extern struct qman_portal *affine_portals[NR_CPUS];
364 extern struct qman_portal *qman_dma_portal;
365 const struct qm_portal_config *qman_get_qm_portal_config(
366 						struct qman_portal *portal);
367