xref: /linux/drivers/net/ethernet/mscc/ocelot_devlink.c (revision 8a922b7728a93d837954315c98b84f6b78de0c4f)
1 // SPDX-License-Identifier: (GPL-2.0 OR MIT)
2 /* Copyright 2020-2021 NXP
3  */
4 #include <net/devlink.h>
5 #include "ocelot.h"
6 
7 /* The queue system tracks four resource consumptions:
8  * Resource 0: Memory tracked per source port
9  * Resource 1: Frame references tracked per source port
10  * Resource 2: Memory tracked per destination port
11  * Resource 3: Frame references tracked per destination port
12  */
13 #define OCELOT_RESOURCE_SZ		256
14 #define OCELOT_NUM_RESOURCES		4
15 
16 #define BUF_xxxx_I			(0 * OCELOT_RESOURCE_SZ)
17 #define REF_xxxx_I			(1 * OCELOT_RESOURCE_SZ)
18 #define BUF_xxxx_E			(2 * OCELOT_RESOURCE_SZ)
19 #define REF_xxxx_E			(3 * OCELOT_RESOURCE_SZ)
20 
21 /* For each resource type there are 4 types of watermarks:
22  * Q_RSRV: reservation per QoS class per port
23  * PRIO_SHR: sharing watermark per QoS class across all ports
24  * P_RSRV: reservation per port
25  * COL_SHR: sharing watermark per color (drop precedence) across all ports
26  */
27 #define xxx_Q_RSRV_x			0
28 #define xxx_PRIO_SHR_x			216
29 #define xxx_P_RSRV_x			224
30 #define xxx_COL_SHR_x			254
31 
32 /* Reservation Watermarks
33  * ----------------------
34  *
35  * For setting up the reserved areas, egress watermarks exist per port and per
36  * QoS class for both ingress and egress.
37  */
38 
39 /*  Amount of packet buffer
40  *  |  per QoS class
41  *  |  |  reserved
42  *  |  |  |   per egress port
43  *  |  |  |   |
44  *  V  V  v   v
45  * BUF_Q_RSRV_E
46  */
47 #define BUF_Q_RSRV_E(port, prio) \
48 	(BUF_xxxx_E + xxx_Q_RSRV_x + OCELOT_NUM_TC * (port) + (prio))
49 
50 /*  Amount of packet buffer
51  *  |  for all port's traffic classes
52  *  |  |  reserved
53  *  |  |  |   per egress port
54  *  |  |  |   |
55  *  V  V  v   v
56  * BUF_P_RSRV_E
57  */
58 #define BUF_P_RSRV_E(port) \
59 	(BUF_xxxx_E + xxx_P_RSRV_x + (port))
60 
61 /*  Amount of packet buffer
62  *  |  per QoS class
63  *  |  |  reserved
64  *  |  |  |   per ingress port
65  *  |  |  |   |
66  *  V  V  v   v
67  * BUF_Q_RSRV_I
68  */
69 #define BUF_Q_RSRV_I(port, prio) \
70 	(BUF_xxxx_I + xxx_Q_RSRV_x + OCELOT_NUM_TC * (port) + (prio))
71 
72 /*  Amount of packet buffer
73  *  |  for all port's traffic classes
74  *  |  |  reserved
75  *  |  |  |   per ingress port
76  *  |  |  |   |
77  *  V  V  v   v
78  * BUF_P_RSRV_I
79  */
80 #define BUF_P_RSRV_I(port) \
81 	(BUF_xxxx_I + xxx_P_RSRV_x + (port))
82 
83 /*  Amount of frame references
84  *  |  per QoS class
85  *  |  |  reserved
86  *  |  |  |   per egress port
87  *  |  |  |   |
88  *  V  V  v   v
89  * REF_Q_RSRV_E
90  */
91 #define REF_Q_RSRV_E(port, prio) \
92 	(REF_xxxx_E + xxx_Q_RSRV_x + OCELOT_NUM_TC * (port) + (prio))
93 
94 /*  Amount of frame references
95  *  |  for all port's traffic classes
96  *  |  |  reserved
97  *  |  |  |   per egress port
98  *  |  |  |   |
99  *  V  V  v   v
100  * REF_P_RSRV_E
101  */
102 #define REF_P_RSRV_E(port) \
103 	(REF_xxxx_E + xxx_P_RSRV_x + (port))
104 
105 /*  Amount of frame references
106  *  |  per QoS class
107  *  |  |  reserved
108  *  |  |  |   per ingress port
109  *  |  |  |   |
110  *  V  V  v   v
111  * REF_Q_RSRV_I
112  */
113 #define REF_Q_RSRV_I(port, prio) \
114 	(REF_xxxx_I + xxx_Q_RSRV_x + OCELOT_NUM_TC * (port) + (prio))
115 
116 /*  Amount of frame references
117  *  |  for all port's traffic classes
118  *  |  |  reserved
119  *  |  |  |   per ingress port
120  *  |  |  |   |
121  *  V  V  v   v
122  * REF_P_RSRV_I
123  */
124 #define REF_P_RSRV_I(port) \
125 	(REF_xxxx_I + xxx_P_RSRV_x + (port))
126 
127 /* Sharing Watermarks
128  * ------------------
129  *
130  * The shared memory area is shared between all ports.
131  */
132 
133 /* Amount of buffer
134  *  |   per QoS class
135  *  |   |    from the shared memory area
136  *  |   |    |  for egress traffic
137  *  |   |    |  |
138  *  V   V    v  v
139  * BUF_PRIO_SHR_E
140  */
141 #define BUF_PRIO_SHR_E(prio) \
142 	(BUF_xxxx_E + xxx_PRIO_SHR_x + (prio))
143 
144 /* Amount of buffer
145  *  |   per color (drop precedence level)
146  *  |   |   from the shared memory area
147  *  |   |   |  for egress traffic
148  *  |   |   |  |
149  *  V   V   v  v
150  * BUF_COL_SHR_E
151  */
152 #define BUF_COL_SHR_E(dp) \
153 	(BUF_xxxx_E + xxx_COL_SHR_x + (1 - (dp)))
154 
155 /* Amount of buffer
156  *  |   per QoS class
157  *  |   |    from the shared memory area
158  *  |   |    |  for ingress traffic
159  *  |   |    |  |
160  *  V   V    v  v
161  * BUF_PRIO_SHR_I
162  */
163 #define BUF_PRIO_SHR_I(prio) \
164 	(BUF_xxxx_I + xxx_PRIO_SHR_x + (prio))
165 
166 /* Amount of buffer
167  *  |   per color (drop precedence level)
168  *  |   |   from the shared memory area
169  *  |   |   |  for ingress traffic
170  *  |   |   |  |
171  *  V   V   v  v
172  * BUF_COL_SHR_I
173  */
174 #define BUF_COL_SHR_I(dp) \
175 	(BUF_xxxx_I + xxx_COL_SHR_x + (1 - (dp)))
176 
177 /* Amount of frame references
178  *  |   per QoS class
179  *  |   |    from the shared area
180  *  |   |    |  for egress traffic
181  *  |   |    |  |
182  *  V   V    v  v
183  * REF_PRIO_SHR_E
184  */
185 #define REF_PRIO_SHR_E(prio) \
186 	(REF_xxxx_E + xxx_PRIO_SHR_x + (prio))
187 
188 /* Amount of frame references
189  *  |   per color (drop precedence level)
190  *  |   |   from the shared area
191  *  |   |   |  for egress traffic
192  *  |   |   |  |
193  *  V   V   v  v
194  * REF_COL_SHR_E
195  */
196 #define REF_COL_SHR_E(dp) \
197 	(REF_xxxx_E + xxx_COL_SHR_x + (1 - (dp)))
198 
199 /* Amount of frame references
200  *  |   per QoS class
201  *  |   |    from the shared area
202  *  |   |    |  for ingress traffic
203  *  |   |    |  |
204  *  V   V    v  v
205  * REF_PRIO_SHR_I
206  */
207 #define REF_PRIO_SHR_I(prio) \
208 	(REF_xxxx_I + xxx_PRIO_SHR_x + (prio))
209 
210 /* Amount of frame references
211  *  |   per color (drop precedence level)
212  *  |   |   from the shared area
213  *  |   |   |  for ingress traffic
214  *  |   |   |  |
215  *  V   V   v  v
216  * REF_COL_SHR_I
217  */
218 #define REF_COL_SHR_I(dp) \
219 	(REF_xxxx_I + xxx_COL_SHR_x + (1 - (dp)))
220 
221 static u32 ocelot_wm_read(struct ocelot *ocelot, int index)
222 {
223 	int wm = ocelot_read_gix(ocelot, QSYS_RES_CFG, index);
224 
225 	return ocelot->ops->wm_dec(wm);
226 }
227 
228 static void ocelot_wm_write(struct ocelot *ocelot, int index, u32 val)
229 {
230 	u32 wm = ocelot->ops->wm_enc(val);
231 
232 	ocelot_write_gix(ocelot, wm, QSYS_RES_CFG, index);
233 }
234 
235 static void ocelot_wm_status(struct ocelot *ocelot, int index, u32 *inuse,
236 			     u32 *maxuse)
237 {
238 	int res_stat = ocelot_read_gix(ocelot, QSYS_RES_STAT, index);
239 
240 	return ocelot->ops->wm_stat(res_stat, inuse, maxuse);
241 }
242 
243 /* The hardware comes out of reset with strange defaults: the sum of all
244  * reservations for frame memory is larger than the total buffer size.
245  * One has to wonder how can the reservation watermarks still guarantee
246  * anything under congestion.
247  * Bring some sense into the hardware by changing the defaults to disable all
248  * reservations and rely only on the sharing watermark for frames with drop
249  * precedence 0. The user can still explicitly request reservations per port
250  * and per port-tc through devlink-sb.
251  */
252 static void ocelot_disable_reservation_watermarks(struct ocelot *ocelot,
253 						  int port)
254 {
255 	int prio;
256 
257 	for (prio = 0; prio < OCELOT_NUM_TC; prio++) {
258 		ocelot_wm_write(ocelot, BUF_Q_RSRV_I(port, prio), 0);
259 		ocelot_wm_write(ocelot, BUF_Q_RSRV_E(port, prio), 0);
260 		ocelot_wm_write(ocelot, REF_Q_RSRV_I(port, prio), 0);
261 		ocelot_wm_write(ocelot, REF_Q_RSRV_E(port, prio), 0);
262 	}
263 
264 	ocelot_wm_write(ocelot, BUF_P_RSRV_I(port), 0);
265 	ocelot_wm_write(ocelot, BUF_P_RSRV_E(port), 0);
266 	ocelot_wm_write(ocelot, REF_P_RSRV_I(port), 0);
267 	ocelot_wm_write(ocelot, REF_P_RSRV_E(port), 0);
268 }
269 
270 /* We want the sharing watermarks to consume all nonreserved resources, for
271  * efficient resource utilization (a single traffic flow should be able to use
272  * up the entire buffer space and frame resources as long as there's no
273  * interference).
274  * The switch has 10 sharing watermarks per lookup: 8 per traffic class and 2
275  * per color (drop precedence).
276  * The trouble with configuring these sharing watermarks is that:
277  * (1) There's a risk that we overcommit the resources if we configure
278  *     (a) all 8 per-TC sharing watermarks to the max
279  *     (b) all 2 per-color sharing watermarks to the max
280  * (2) There's a risk that we undercommit the resources if we configure
281  *     (a) all 8 per-TC sharing watermarks to "max / 8"
282  *     (b) all 2 per-color sharing watermarks to "max / 2"
283  * So for Linux, let's just disable the sharing watermarks per traffic class
284  * (setting them to 0 will make them always exceeded), and rely only on the
285  * sharing watermark for drop priority 0. So frames with drop priority set to 1
286  * by QoS classification or policing will still be allowed, but only as long as
287  * the port and port-TC reservations are not exceeded.
288  */
289 static void ocelot_disable_tc_sharing_watermarks(struct ocelot *ocelot)
290 {
291 	int prio;
292 
293 	for (prio = 0; prio < OCELOT_NUM_TC; prio++) {
294 		ocelot_wm_write(ocelot, BUF_PRIO_SHR_I(prio), 0);
295 		ocelot_wm_write(ocelot, BUF_PRIO_SHR_E(prio), 0);
296 		ocelot_wm_write(ocelot, REF_PRIO_SHR_I(prio), 0);
297 		ocelot_wm_write(ocelot, REF_PRIO_SHR_E(prio), 0);
298 	}
299 }
300 
301 static void ocelot_get_buf_rsrv(struct ocelot *ocelot, u32 *buf_rsrv_i,
302 				u32 *buf_rsrv_e)
303 {
304 	int port, prio;
305 
306 	*buf_rsrv_i = 0;
307 	*buf_rsrv_e = 0;
308 
309 	for (port = 0; port <= ocelot->num_phys_ports; port++) {
310 		for (prio = 0; prio < OCELOT_NUM_TC; prio++) {
311 			*buf_rsrv_i += ocelot_wm_read(ocelot,
312 						      BUF_Q_RSRV_I(port, prio));
313 			*buf_rsrv_e += ocelot_wm_read(ocelot,
314 						      BUF_Q_RSRV_E(port, prio));
315 		}
316 
317 		*buf_rsrv_i += ocelot_wm_read(ocelot, BUF_P_RSRV_I(port));
318 		*buf_rsrv_e += ocelot_wm_read(ocelot, BUF_P_RSRV_E(port));
319 	}
320 
321 	*buf_rsrv_i *= OCELOT_BUFFER_CELL_SZ;
322 	*buf_rsrv_e *= OCELOT_BUFFER_CELL_SZ;
323 }
324 
325 static void ocelot_get_ref_rsrv(struct ocelot *ocelot, u32 *ref_rsrv_i,
326 				u32 *ref_rsrv_e)
327 {
328 	int port, prio;
329 
330 	*ref_rsrv_i = 0;
331 	*ref_rsrv_e = 0;
332 
333 	for (port = 0; port <= ocelot->num_phys_ports; port++) {
334 		for (prio = 0; prio < OCELOT_NUM_TC; prio++) {
335 			*ref_rsrv_i += ocelot_wm_read(ocelot,
336 						      REF_Q_RSRV_I(port, prio));
337 			*ref_rsrv_e += ocelot_wm_read(ocelot,
338 						      REF_Q_RSRV_E(port, prio));
339 		}
340 
341 		*ref_rsrv_i += ocelot_wm_read(ocelot, REF_P_RSRV_I(port));
342 		*ref_rsrv_e += ocelot_wm_read(ocelot, REF_P_RSRV_E(port));
343 	}
344 }
345 
346 /* Calculate all reservations, then set up the sharing watermark for DP=0 to
347  * consume the remaining resources up to the pool's configured size.
348  */
349 static void ocelot_setup_sharing_watermarks(struct ocelot *ocelot)
350 {
351 	u32 buf_rsrv_i, buf_rsrv_e;
352 	u32 ref_rsrv_i, ref_rsrv_e;
353 	u32 buf_shr_i, buf_shr_e;
354 	u32 ref_shr_i, ref_shr_e;
355 
356 	ocelot_get_buf_rsrv(ocelot, &buf_rsrv_i, &buf_rsrv_e);
357 	ocelot_get_ref_rsrv(ocelot, &ref_rsrv_i, &ref_rsrv_e);
358 
359 	buf_shr_i = ocelot->pool_size[OCELOT_SB_BUF][OCELOT_SB_POOL_ING] -
360 		    buf_rsrv_i;
361 	buf_shr_e = ocelot->pool_size[OCELOT_SB_BUF][OCELOT_SB_POOL_EGR] -
362 		    buf_rsrv_e;
363 	ref_shr_i = ocelot->pool_size[OCELOT_SB_REF][OCELOT_SB_POOL_ING] -
364 		    ref_rsrv_i;
365 	ref_shr_e = ocelot->pool_size[OCELOT_SB_REF][OCELOT_SB_POOL_EGR] -
366 		    ref_rsrv_e;
367 
368 	buf_shr_i /= OCELOT_BUFFER_CELL_SZ;
369 	buf_shr_e /= OCELOT_BUFFER_CELL_SZ;
370 
371 	ocelot_wm_write(ocelot, BUF_COL_SHR_I(0), buf_shr_i);
372 	ocelot_wm_write(ocelot, BUF_COL_SHR_E(0), buf_shr_e);
373 	ocelot_wm_write(ocelot, REF_COL_SHR_E(0), ref_shr_e);
374 	ocelot_wm_write(ocelot, REF_COL_SHR_I(0), ref_shr_i);
375 	ocelot_wm_write(ocelot, BUF_COL_SHR_I(1), 0);
376 	ocelot_wm_write(ocelot, BUF_COL_SHR_E(1), 0);
377 	ocelot_wm_write(ocelot, REF_COL_SHR_E(1), 0);
378 	ocelot_wm_write(ocelot, REF_COL_SHR_I(1), 0);
379 }
380 
381 /* Ensure that all reservations can be enforced */
382 static int ocelot_watermark_validate(struct ocelot *ocelot,
383 				     struct netlink_ext_ack *extack)
384 {
385 	u32 buf_rsrv_i, buf_rsrv_e;
386 	u32 ref_rsrv_i, ref_rsrv_e;
387 
388 	ocelot_get_buf_rsrv(ocelot, &buf_rsrv_i, &buf_rsrv_e);
389 	ocelot_get_ref_rsrv(ocelot, &ref_rsrv_i, &ref_rsrv_e);
390 
391 	if (buf_rsrv_i > ocelot->pool_size[OCELOT_SB_BUF][OCELOT_SB_POOL_ING]) {
392 		NL_SET_ERR_MSG_MOD(extack,
393 				   "Ingress frame reservations exceed pool size");
394 		return -ERANGE;
395 	}
396 	if (buf_rsrv_e > ocelot->pool_size[OCELOT_SB_BUF][OCELOT_SB_POOL_EGR]) {
397 		NL_SET_ERR_MSG_MOD(extack,
398 				   "Egress frame reservations exceed pool size");
399 		return -ERANGE;
400 	}
401 	if (ref_rsrv_i > ocelot->pool_size[OCELOT_SB_REF][OCELOT_SB_POOL_ING]) {
402 		NL_SET_ERR_MSG_MOD(extack,
403 				   "Ingress reference reservations exceed pool size");
404 		return -ERANGE;
405 	}
406 	if (ref_rsrv_e > ocelot->pool_size[OCELOT_SB_REF][OCELOT_SB_POOL_EGR]) {
407 		NL_SET_ERR_MSG_MOD(extack,
408 				   "Egress reference reservations exceed pool size");
409 		return -ERANGE;
410 	}
411 
412 	return 0;
413 }
414 
415 /* The hardware works like this:
416  *
417  *                         Frame forwarding decision taken
418  *                                       |
419  *                                       v
420  *       +--------------------+--------------------+--------------------+
421  *       |                    |                    |                    |
422  *       v                    v                    v                    v
423  * Ingress memory       Egress memory        Ingress frame        Egress frame
424  *     check                check           reference check      reference check
425  *       |                    |                    |                    |
426  *       v                    v                    v                    v
427  *  BUF_Q_RSRV_I   ok    BUF_Q_RSRV_E   ok    REF_Q_RSRV_I   ok     REF_Q_RSRV_E   ok
428  *(src port, prio) -+  (dst port, prio) -+  (src port, prio) -+   (dst port, prio) -+
429  *       |          |         |          |         |          |         |           |
430  *       |exceeded  |         |exceeded  |         |exceeded  |         |exceeded   |
431  *       v          |         v          |         v          |         v           |
432  *  BUF_P_RSRV_I  ok|    BUF_P_RSRV_E  ok|    REF_P_RSRV_I  ok|    REF_P_RSRV_E   ok|
433  *   (src port) ----+     (dst port) ----+     (src port) ----+     (dst port) -----+
434  *       |          |         |          |         |          |         |           |
435  *       |exceeded  |         |exceeded  |         |exceeded  |         |exceeded   |
436  *       v          |         v          |         v          |         v           |
437  * BUF_PRIO_SHR_I ok|   BUF_PRIO_SHR_E ok|   REF_PRIO_SHR_I ok|   REF_PRIO_SHR_E  ok|
438  *     (prio) ------+       (prio) ------+       (prio) ------+       (prio) -------+
439  *       |          |         |          |         |          |         |           |
440  *       |exceeded  |         |exceeded  |         |exceeded  |         |exceeded   |
441  *       v          |         v          |         v          |         v           |
442  * BUF_COL_SHR_I  ok|   BUF_COL_SHR_E  ok|   REF_COL_SHR_I  ok|   REF_COL_SHR_E   ok|
443  *      (dp) -------+        (dp) -------+        (dp) -------+        (dp) --------+
444  *       |          |         |          |         |          |         |           |
445  *       |exceeded  |         |exceeded  |         |exceeded  |         |exceeded   |
446  *       v          v         v          v         v          v         v           v
447  *      fail     success     fail     success     fail     success     fail      success
448  *       |          |         |          |         |          |         |           |
449  *       v          v         v          v         v          v         v           v
450  *       +-----+----+         +-----+----+         +-----+----+         +-----+-----+
451  *             |                    |                    |                    |
452  *             +-------> OR <-------+                    +-------> OR <-------+
453  *                        |                                        |
454  *                        v                                        v
455  *                        +----------------> AND <-----------------+
456  *                                            |
457  *                                            v
458  *                                    FIFO drop / accept
459  *
460  * We are modeling each of the 4 parallel lookups as a devlink-sb pool.
461  * At least one (ingress or egress) memory pool and one (ingress or egress)
462  * frame reference pool need to have resources for frame acceptance to succeed.
463  *
464  * The following watermarks are controlled explicitly through devlink-sb:
465  * BUF_Q_RSRV_I, BUF_Q_RSRV_E, REF_Q_RSRV_I, REF_Q_RSRV_E
466  * BUF_P_RSRV_I, BUF_P_RSRV_E, REF_P_RSRV_I, REF_P_RSRV_E
467  * The following watermarks are controlled implicitly through devlink-sb:
468  * BUF_COL_SHR_I, BUF_COL_SHR_E, REF_COL_SHR_I, REF_COL_SHR_E
469  * The following watermarks are unused and disabled:
470  * BUF_PRIO_SHR_I, BUF_PRIO_SHR_E, REF_PRIO_SHR_I, REF_PRIO_SHR_E
471  *
472  * This function overrides the hardware defaults with more sane ones (no
473  * reservations by default, let sharing use all resources) and disables the
474  * unused watermarks.
475  */
476 static void ocelot_watermark_init(struct ocelot *ocelot)
477 {
478 	int all_tcs = GENMASK(OCELOT_NUM_TC - 1, 0);
479 	int port;
480 
481 	ocelot_write(ocelot, all_tcs, QSYS_RES_QOS_MODE);
482 
483 	for (port = 0; port <= ocelot->num_phys_ports; port++)
484 		ocelot_disable_reservation_watermarks(ocelot, port);
485 
486 	ocelot_disable_tc_sharing_watermarks(ocelot);
487 	ocelot_setup_sharing_watermarks(ocelot);
488 }
489 
490 /* Watermark encode
491  * Bit 8:   Unit; 0:1, 1:16
492  * Bit 7-0: Value to be multiplied with unit
493  */
494 u16 ocelot_wm_enc(u16 value)
495 {
496 	WARN_ON(value >= 16 * BIT(8));
497 
498 	if (value >= BIT(8))
499 		return BIT(8) | (value / 16);
500 
501 	return value;
502 }
503 EXPORT_SYMBOL(ocelot_wm_enc);
504 
505 u16 ocelot_wm_dec(u16 wm)
506 {
507 	if (wm & BIT(8))
508 		return (wm & GENMASK(7, 0)) * 16;
509 
510 	return wm;
511 }
512 EXPORT_SYMBOL(ocelot_wm_dec);
513 
514 void ocelot_wm_stat(u32 val, u32 *inuse, u32 *maxuse)
515 {
516 	*inuse = (val & GENMASK(23, 12)) >> 12;
517 	*maxuse = val & GENMASK(11, 0);
518 }
519 EXPORT_SYMBOL(ocelot_wm_stat);
520 
521 /* Pool size and type are fixed up at runtime. Keeping this structure to
522  * look up the cell size multipliers.
523  */
524 static const struct devlink_sb_pool_info ocelot_sb_pool[] = {
525 	[OCELOT_SB_BUF] = {
526 		.cell_size = OCELOT_BUFFER_CELL_SZ,
527 		.threshold_type = DEVLINK_SB_THRESHOLD_TYPE_STATIC,
528 	},
529 	[OCELOT_SB_REF] = {
530 		.cell_size = 1,
531 		.threshold_type = DEVLINK_SB_THRESHOLD_TYPE_STATIC,
532 	},
533 };
534 
535 /* Returns the pool size configured through ocelot_sb_pool_set */
536 int ocelot_sb_pool_get(struct ocelot *ocelot, unsigned int sb_index,
537 		       u16 pool_index,
538 		       struct devlink_sb_pool_info *pool_info)
539 {
540 	if (sb_index >= OCELOT_SB_NUM)
541 		return -ENODEV;
542 	if (pool_index >= OCELOT_SB_POOL_NUM)
543 		return -ENODEV;
544 
545 	*pool_info = ocelot_sb_pool[sb_index];
546 	pool_info->size = ocelot->pool_size[sb_index][pool_index];
547 	if (pool_index)
548 		pool_info->pool_type = DEVLINK_SB_POOL_TYPE_INGRESS;
549 	else
550 		pool_info->pool_type = DEVLINK_SB_POOL_TYPE_EGRESS;
551 
552 	return 0;
553 }
554 EXPORT_SYMBOL(ocelot_sb_pool_get);
555 
556 /* The pool size received here configures the total amount of resources used on
557  * ingress (or on egress, depending upon the pool index). The pool size, minus
558  * the values for the port and port-tc reservations, is written into the
559  * COL_SHR(dp=0) sharing watermark.
560  */
561 int ocelot_sb_pool_set(struct ocelot *ocelot, unsigned int sb_index,
562 		       u16 pool_index, u32 size,
563 		       enum devlink_sb_threshold_type threshold_type,
564 		       struct netlink_ext_ack *extack)
565 {
566 	u32 old_pool_size;
567 	int err;
568 
569 	if (sb_index >= OCELOT_SB_NUM) {
570 		NL_SET_ERR_MSG_MOD(extack,
571 				   "Invalid sb, use 0 for buffers and 1 for frame references");
572 		return -ENODEV;
573 	}
574 	if (pool_index >= OCELOT_SB_POOL_NUM) {
575 		NL_SET_ERR_MSG_MOD(extack,
576 				   "Invalid pool, use 0 for ingress and 1 for egress");
577 		return -ENODEV;
578 	}
579 	if (threshold_type != DEVLINK_SB_THRESHOLD_TYPE_STATIC) {
580 		NL_SET_ERR_MSG_MOD(extack,
581 				   "Only static threshold supported");
582 		return -EOPNOTSUPP;
583 	}
584 
585 	old_pool_size = ocelot->pool_size[sb_index][pool_index];
586 	ocelot->pool_size[sb_index][pool_index] = size;
587 
588 	err = ocelot_watermark_validate(ocelot, extack);
589 	if (err) {
590 		ocelot->pool_size[sb_index][pool_index] = old_pool_size;
591 		return err;
592 	}
593 
594 	ocelot_setup_sharing_watermarks(ocelot);
595 
596 	return 0;
597 }
598 EXPORT_SYMBOL(ocelot_sb_pool_set);
599 
600 /* This retrieves the configuration made with ocelot_sb_port_pool_set */
601 int ocelot_sb_port_pool_get(struct ocelot *ocelot, int port,
602 			    unsigned int sb_index, u16 pool_index,
603 			    u32 *p_threshold)
604 {
605 	int wm_index;
606 
607 	switch (sb_index) {
608 	case OCELOT_SB_BUF:
609 		if (pool_index == OCELOT_SB_POOL_ING)
610 			wm_index = BUF_P_RSRV_I(port);
611 		else
612 			wm_index = BUF_P_RSRV_E(port);
613 		break;
614 	case OCELOT_SB_REF:
615 		if (pool_index == OCELOT_SB_POOL_ING)
616 			wm_index = REF_P_RSRV_I(port);
617 		else
618 			wm_index = REF_P_RSRV_E(port);
619 		break;
620 	default:
621 		return -ENODEV;
622 	}
623 
624 	*p_threshold = ocelot_wm_read(ocelot, wm_index);
625 	*p_threshold *= ocelot_sb_pool[sb_index].cell_size;
626 
627 	return 0;
628 }
629 EXPORT_SYMBOL(ocelot_sb_port_pool_get);
630 
631 /* This configures the P_RSRV per-port reserved resource watermark */
632 int ocelot_sb_port_pool_set(struct ocelot *ocelot, int port,
633 			    unsigned int sb_index, u16 pool_index,
634 			    u32 threshold, struct netlink_ext_ack *extack)
635 {
636 	int wm_index, err;
637 	u32 old_thr;
638 
639 	switch (sb_index) {
640 	case OCELOT_SB_BUF:
641 		if (pool_index == OCELOT_SB_POOL_ING)
642 			wm_index = BUF_P_RSRV_I(port);
643 		else
644 			wm_index = BUF_P_RSRV_E(port);
645 		break;
646 	case OCELOT_SB_REF:
647 		if (pool_index == OCELOT_SB_POOL_ING)
648 			wm_index = REF_P_RSRV_I(port);
649 		else
650 			wm_index = REF_P_RSRV_E(port);
651 		break;
652 	default:
653 		NL_SET_ERR_MSG_MOD(extack, "Invalid shared buffer");
654 		return -ENODEV;
655 	}
656 
657 	threshold /= ocelot_sb_pool[sb_index].cell_size;
658 
659 	old_thr = ocelot_wm_read(ocelot, wm_index);
660 	ocelot_wm_write(ocelot, wm_index, threshold);
661 
662 	err = ocelot_watermark_validate(ocelot, extack);
663 	if (err) {
664 		ocelot_wm_write(ocelot, wm_index, old_thr);
665 		return err;
666 	}
667 
668 	ocelot_setup_sharing_watermarks(ocelot);
669 
670 	return 0;
671 }
672 EXPORT_SYMBOL(ocelot_sb_port_pool_set);
673 
674 /* This retrieves the configuration done by ocelot_sb_tc_pool_bind_set */
675 int ocelot_sb_tc_pool_bind_get(struct ocelot *ocelot, int port,
676 			       unsigned int sb_index, u16 tc_index,
677 			       enum devlink_sb_pool_type pool_type,
678 			       u16 *p_pool_index, u32 *p_threshold)
679 {
680 	int wm_index;
681 
682 	switch (sb_index) {
683 	case OCELOT_SB_BUF:
684 		if (pool_type == DEVLINK_SB_POOL_TYPE_INGRESS)
685 			wm_index = BUF_Q_RSRV_I(port, tc_index);
686 		else
687 			wm_index = BUF_Q_RSRV_E(port, tc_index);
688 		break;
689 	case OCELOT_SB_REF:
690 		if (pool_type == DEVLINK_SB_POOL_TYPE_INGRESS)
691 			wm_index = REF_Q_RSRV_I(port, tc_index);
692 		else
693 			wm_index = REF_Q_RSRV_E(port, tc_index);
694 		break;
695 	default:
696 		return -ENODEV;
697 	}
698 
699 	*p_threshold = ocelot_wm_read(ocelot, wm_index);
700 	*p_threshold *= ocelot_sb_pool[sb_index].cell_size;
701 
702 	if (pool_type == DEVLINK_SB_POOL_TYPE_INGRESS)
703 		*p_pool_index = 0;
704 	else
705 		*p_pool_index = 1;
706 
707 	return 0;
708 }
709 EXPORT_SYMBOL(ocelot_sb_tc_pool_bind_get);
710 
711 /* This configures the Q_RSRV per-port-tc reserved resource watermark */
712 int ocelot_sb_tc_pool_bind_set(struct ocelot *ocelot, int port,
713 			       unsigned int sb_index, u16 tc_index,
714 			       enum devlink_sb_pool_type pool_type,
715 			       u16 pool_index, u32 threshold,
716 			       struct netlink_ext_ack *extack)
717 {
718 	int wm_index, err;
719 	u32 old_thr;
720 
721 	/* Paranoid check? */
722 	if (pool_index == OCELOT_SB_POOL_ING &&
723 	    pool_type != DEVLINK_SB_POOL_TYPE_INGRESS)
724 		return -EINVAL;
725 	if (pool_index == OCELOT_SB_POOL_EGR &&
726 	    pool_type != DEVLINK_SB_POOL_TYPE_EGRESS)
727 		return -EINVAL;
728 
729 	switch (sb_index) {
730 	case OCELOT_SB_BUF:
731 		if (pool_type == DEVLINK_SB_POOL_TYPE_INGRESS)
732 			wm_index = BUF_Q_RSRV_I(port, tc_index);
733 		else
734 			wm_index = BUF_Q_RSRV_E(port, tc_index);
735 		break;
736 	case OCELOT_SB_REF:
737 		if (pool_type == DEVLINK_SB_POOL_TYPE_INGRESS)
738 			wm_index = REF_Q_RSRV_I(port, tc_index);
739 		else
740 			wm_index = REF_Q_RSRV_E(port, tc_index);
741 		break;
742 	default:
743 		NL_SET_ERR_MSG_MOD(extack, "Invalid shared buffer");
744 		return -ENODEV;
745 	}
746 
747 	threshold /= ocelot_sb_pool[sb_index].cell_size;
748 
749 	old_thr = ocelot_wm_read(ocelot, wm_index);
750 	ocelot_wm_write(ocelot, wm_index, threshold);
751 	err = ocelot_watermark_validate(ocelot, extack);
752 	if (err) {
753 		ocelot_wm_write(ocelot, wm_index, old_thr);
754 		return err;
755 	}
756 
757 	ocelot_setup_sharing_watermarks(ocelot);
758 
759 	return 0;
760 }
761 EXPORT_SYMBOL(ocelot_sb_tc_pool_bind_set);
762 
763 /* The hardware does not support atomic snapshots, we'll read out the
764  * occupancy registers individually and have this as just a stub.
765  */
766 int ocelot_sb_occ_snapshot(struct ocelot *ocelot, unsigned int sb_index)
767 {
768 	return 0;
769 }
770 EXPORT_SYMBOL(ocelot_sb_occ_snapshot);
771 
772 /* The watermark occupancy registers are cleared upon read,
773  * so let's read them.
774  */
775 int ocelot_sb_occ_max_clear(struct ocelot *ocelot, unsigned int sb_index)
776 {
777 	u32 inuse, maxuse;
778 	int port, prio;
779 
780 	switch (sb_index) {
781 	case OCELOT_SB_BUF:
782 		for (port = 0; port <= ocelot->num_phys_ports; port++) {
783 			for (prio = 0; prio < OCELOT_NUM_TC; prio++) {
784 				ocelot_wm_status(ocelot, BUF_Q_RSRV_I(port, prio),
785 						 &inuse, &maxuse);
786 				ocelot_wm_status(ocelot, BUF_Q_RSRV_E(port, prio),
787 						 &inuse, &maxuse);
788 			}
789 			ocelot_wm_status(ocelot, BUF_P_RSRV_I(port),
790 					 &inuse, &maxuse);
791 			ocelot_wm_status(ocelot, BUF_P_RSRV_E(port),
792 					 &inuse, &maxuse);
793 		}
794 		break;
795 	case OCELOT_SB_REF:
796 		for (port = 0; port <= ocelot->num_phys_ports; port++) {
797 			for (prio = 0; prio < OCELOT_NUM_TC; prio++) {
798 				ocelot_wm_status(ocelot, REF_Q_RSRV_I(port, prio),
799 						 &inuse, &maxuse);
800 				ocelot_wm_status(ocelot, REF_Q_RSRV_E(port, prio),
801 						 &inuse, &maxuse);
802 			}
803 			ocelot_wm_status(ocelot, REF_P_RSRV_I(port),
804 					 &inuse, &maxuse);
805 			ocelot_wm_status(ocelot, REF_P_RSRV_E(port),
806 					 &inuse, &maxuse);
807 		}
808 		break;
809 	default:
810 		return -ENODEV;
811 	}
812 
813 	return 0;
814 }
815 EXPORT_SYMBOL(ocelot_sb_occ_max_clear);
816 
817 /* This retrieves the watermark occupancy for per-port P_RSRV watermarks */
818 int ocelot_sb_occ_port_pool_get(struct ocelot *ocelot, int port,
819 				unsigned int sb_index, u16 pool_index,
820 				u32 *p_cur, u32 *p_max)
821 {
822 	int wm_index;
823 
824 	switch (sb_index) {
825 	case OCELOT_SB_BUF:
826 		if (pool_index == OCELOT_SB_POOL_ING)
827 			wm_index = BUF_P_RSRV_I(port);
828 		else
829 			wm_index = BUF_P_RSRV_E(port);
830 		break;
831 	case OCELOT_SB_REF:
832 		if (pool_index == OCELOT_SB_POOL_ING)
833 			wm_index = REF_P_RSRV_I(port);
834 		else
835 			wm_index = REF_P_RSRV_E(port);
836 		break;
837 	default:
838 		return -ENODEV;
839 	}
840 
841 	ocelot_wm_status(ocelot, wm_index, p_cur, p_max);
842 	*p_cur *= ocelot_sb_pool[sb_index].cell_size;
843 	*p_max *= ocelot_sb_pool[sb_index].cell_size;
844 
845 	return 0;
846 }
847 EXPORT_SYMBOL(ocelot_sb_occ_port_pool_get);
848 
849 /* This retrieves the watermark occupancy for per-port-tc Q_RSRV watermarks */
850 int ocelot_sb_occ_tc_port_bind_get(struct ocelot *ocelot, int port,
851 				   unsigned int sb_index, u16 tc_index,
852 				   enum devlink_sb_pool_type pool_type,
853 				   u32 *p_cur, u32 *p_max)
854 {
855 	int wm_index;
856 
857 	switch (sb_index) {
858 	case OCELOT_SB_BUF:
859 		if (pool_type == DEVLINK_SB_POOL_TYPE_INGRESS)
860 			wm_index = BUF_Q_RSRV_I(port, tc_index);
861 		else
862 			wm_index = BUF_Q_RSRV_E(port, tc_index);
863 		break;
864 	case OCELOT_SB_REF:
865 		if (pool_type == DEVLINK_SB_POOL_TYPE_INGRESS)
866 			wm_index = REF_Q_RSRV_I(port, tc_index);
867 		else
868 			wm_index = REF_Q_RSRV_E(port, tc_index);
869 		break;
870 	default:
871 		return -ENODEV;
872 	}
873 
874 	ocelot_wm_status(ocelot, wm_index, p_cur, p_max);
875 	*p_cur *= ocelot_sb_pool[sb_index].cell_size;
876 	*p_max *= ocelot_sb_pool[sb_index].cell_size;
877 
878 	return 0;
879 }
880 EXPORT_SYMBOL(ocelot_sb_occ_tc_port_bind_get);
881 
882 int ocelot_devlink_sb_register(struct ocelot *ocelot)
883 {
884 	int err;
885 
886 	err = devlink_sb_register(ocelot->devlink, OCELOT_SB_BUF,
887 				  ocelot->packet_buffer_size, 1, 1,
888 				  OCELOT_NUM_TC, OCELOT_NUM_TC);
889 	if (err)
890 		return err;
891 
892 	err = devlink_sb_register(ocelot->devlink, OCELOT_SB_REF,
893 				  ocelot->num_frame_refs, 1, 1,
894 				  OCELOT_NUM_TC, OCELOT_NUM_TC);
895 	if (err) {
896 		devlink_sb_unregister(ocelot->devlink, OCELOT_SB_BUF);
897 		return err;
898 	}
899 
900 	ocelot->pool_size[OCELOT_SB_BUF][OCELOT_SB_POOL_ING] = ocelot->packet_buffer_size;
901 	ocelot->pool_size[OCELOT_SB_BUF][OCELOT_SB_POOL_EGR] = ocelot->packet_buffer_size;
902 	ocelot->pool_size[OCELOT_SB_REF][OCELOT_SB_POOL_ING] = ocelot->num_frame_refs;
903 	ocelot->pool_size[OCELOT_SB_REF][OCELOT_SB_POOL_EGR] = ocelot->num_frame_refs;
904 
905 	ocelot_watermark_init(ocelot);
906 
907 	return 0;
908 }
909 EXPORT_SYMBOL(ocelot_devlink_sb_register);
910 
911 void ocelot_devlink_sb_unregister(struct ocelot *ocelot)
912 {
913 	devlink_sb_unregister(ocelot->devlink, OCELOT_SB_BUF);
914 	devlink_sb_unregister(ocelot->devlink, OCELOT_SB_REF);
915 }
916 EXPORT_SYMBOL(ocelot_devlink_sb_unregister);
917