xref: /freebsd/sys/dev/bxe/ecore_init.h (revision cfd6422a5217410fbd66f7a7a8a64d9d85e61229)
1 /*-
2  * SPDX-License-Identifier: BSD-2-Clause
3  *
4  * Copyright (c) 2007-2017 QLogic Corporation. All rights reserved.
5  *
6  * Redistribution and use in source and binary forms, with or without
7  * modification, are permitted provided that the following conditions
8  * are met:
9  *
10  * 1. Redistributions of source code must retain the above copyright
11  *    notice, this list of conditions and the following disclaimer.
12  * 2. Redistributions in binary form must reproduce the above copyright
13  *    notice, this list of conditions and the following disclaimer in the
14  *    documentation and/or other materials provided with the distribution.
15  *
16  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
17  * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
18  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
19  * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS
20  * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
21  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
22  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
23  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
24  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
25  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF
26  * THE POSSIBILITY OF SUCH DAMAGE.
27  */
28 
29 #include <sys/cdefs.h>
30 __FBSDID("$FreeBSD$");
31 
32 #ifndef ECORE_INIT_H
33 #define ECORE_INIT_H
34 
35 /* Init operation types and structures */
36 enum {
37 	OP_RD = 0x1,	/* read a single register */
38 	OP_WR,		/* write a single register */
39 	OP_SW,		/* copy a string to the device */
40 	OP_ZR,		/* clear memory */
41 	OP_ZP,		/* unzip then copy with DMAE */
42 	OP_WR_64,	/* write 64 bit pattern */
43 	OP_WB,		/* copy a string using DMAE */
44 #ifndef FW_ZIP_SUPPORT
45 	OP_FW,		/* copy an array from fw data (only used with unzipped FW) */
46 #endif
47 	OP_WB_ZR,	/* Clear a string using DMAE or indirect-wr */
48 	OP_IF_MODE_OR,  /* Skip the following ops if all init modes don't match */
49 	OP_IF_MODE_AND, /* Skip the following ops if any init modes don't match */
50 	OP_IF_PHASE,
51 	OP_RT,
52 	OP_DELAY,
53 	OP_VERIFY,
54 	OP_MAX
55 };
56 
57 enum {
58 	STAGE_START,
59 	STAGE_END,
60 };
61 
62 /* Returns the index of start or end of a specific block stage in ops array*/
63 #define BLOCK_OPS_IDX(block, stage, end) \
64 	(2*(((block)*NUM_OF_INIT_PHASES) + (stage)) + (end))
65 
66 
67 /* structs for the various opcodes */
68 struct raw_op {
69 	uint32_t op:8;
70 	uint32_t offset:24;
71 	uint32_t raw_data;
72 };
73 
74 struct op_read {
75 	uint32_t op:8;
76 	uint32_t offset:24;
77 	uint32_t val;
78 };
79 
80 struct op_write {
81 	uint32_t op:8;
82 	uint32_t offset:24;
83 	uint32_t val;
84 };
85 
86 struct op_arr_write {
87 	uint32_t op:8;
88 	uint32_t offset:24;
89 #ifdef __BIG_ENDIAN
90 	uint16_t data_len;
91 	uint16_t data_off;
92 #else /* __LITTLE_ENDIAN */
93 	uint16_t data_off;
94 	uint16_t data_len;
95 #endif
96 };
97 
98 struct op_zero {
99 	uint32_t op:8;
100 	uint32_t offset:24;
101 	uint32_t len;
102 };
103 
104 struct op_if_mode {
105 	uint32_t op:8;
106 	uint32_t cmd_offset:24;
107 	uint32_t mode_bit_map;
108 };
109 
110 struct op_if_phase {
111 	uint32_t op:8;
112 	uint32_t cmd_offset:24;
113 	uint32_t phase_bit_map;
114 };
115 
116 struct op_delay {
117 	uint32_t op:8;
118 	uint32_t reserved:24;
119 	uint32_t delay;
120 };
121 
122 union init_op {
123 	struct op_read		read;
124 	struct op_write		write;
125 	struct op_arr_write	arr_wr;
126 	struct op_zero		zero;
127 	struct raw_op		raw;
128 	struct op_if_mode	if_mode;
129 	struct op_if_phase	if_phase;
130 	struct op_delay		delay;
131 };
132 
133 
134 /* Init Phases */
135 enum {
136 	PHASE_COMMON,
137 	PHASE_PORT0,
138 	PHASE_PORT1,
139 	PHASE_PF0,
140 	PHASE_PF1,
141 	PHASE_PF2,
142 	PHASE_PF3,
143 	PHASE_PF4,
144 	PHASE_PF5,
145 	PHASE_PF6,
146 	PHASE_PF7,
147 	NUM_OF_INIT_PHASES
148 };
149 
150 /* Init Modes */
151 enum {
152 	MODE_ASIC                      = 0x00000001,
153 	MODE_FPGA                      = 0x00000002,
154 	MODE_EMUL                      = 0x00000004,
155 	MODE_E2                        = 0x00000008,
156 	MODE_E3                        = 0x00000010,
157 	MODE_PORT2                     = 0x00000020,
158 	MODE_PORT4                     = 0x00000040,
159 	MODE_SF                        = 0x00000080,
160 	MODE_MF                        = 0x00000100,
161 	MODE_MF_SD                     = 0x00000200,
162 	MODE_MF_SI                     = 0x00000400,
163 	MODE_MF_AFEX                   = 0x00000800,
164 	MODE_E3_A0                     = 0x00001000,
165 	MODE_E3_B0                     = 0x00002000,
166 	MODE_COS3                      = 0x00004000,
167 	MODE_COS6                      = 0x00008000,
168 	MODE_LITTLE_ENDIAN             = 0x00010000,
169 	MODE_BIG_ENDIAN                = 0x00020000,
170 };
171 
172 /* Init Blocks */
173 enum {
174 	BLOCK_ATC,
175 	BLOCK_BRB1,
176 	BLOCK_CCM,
177 	BLOCK_CDU,
178 	BLOCK_CFC,
179 	BLOCK_CSDM,
180 	BLOCK_CSEM,
181 	BLOCK_DBG,
182 	BLOCK_DMAE,
183 	BLOCK_DORQ,
184 	BLOCK_HC,
185 	BLOCK_IGU,
186 	BLOCK_MISC,
187 	BLOCK_NIG,
188 	BLOCK_PBF,
189 	BLOCK_PGLUE_B,
190 	BLOCK_PRS,
191 	BLOCK_PXP2,
192 	BLOCK_PXP,
193 	BLOCK_QM,
194 	BLOCK_SRC,
195 	BLOCK_TCM,
196 	BLOCK_TM,
197 	BLOCK_TSDM,
198 	BLOCK_TSEM,
199 	BLOCK_UCM,
200 	BLOCK_UPB,
201 	BLOCK_USDM,
202 	BLOCK_USEM,
203 	BLOCK_XCM,
204 	BLOCK_XPB,
205 	BLOCK_XSDM,
206 	BLOCK_XSEM,
207 	BLOCK_MISC_AEU,
208 	NUM_OF_INIT_BLOCKS
209 };
210 
211 
212 
213 
214 
215 
216 
217 
218 /* Vnics per mode */
219 #define ECORE_PORT2_MODE_NUM_VNICS 4
220 
221 
222 /* QM queue numbers */
223 #define ECORE_ETH_Q		0
224 #define ECORE_TOE_Q		3
225 #define ECORE_TOE_ACK_Q		6
226 #define ECORE_ISCSI_Q		9
227 #define ECORE_ISCSI_ACK_Q	11
228 #define ECORE_FCOE_Q		10
229 
230 /* Vnics per mode */
231 #define ECORE_PORT4_MODE_NUM_VNICS 2
232 
233 /* COS offset for port1 in E3 B0 4port mode */
234 #define ECORE_E3B0_PORT1_COS_OFFSET 3
235 
236 /* QM Register addresses */
237 #define ECORE_Q_VOQ_REG_ADDR(pf_q_num)\
238 	(QM_REG_QVOQIDX_0 + 4 * (pf_q_num))
239 #define ECORE_VOQ_Q_REG_ADDR(cos, pf_q_num)\
240 	(QM_REG_VOQQMASK_0_LSB + 4 * ((cos) * 2 + ((pf_q_num) >> 5)))
241 #define ECORE_Q_CMDQ_REG_ADDR(pf_q_num)\
242 	(QM_REG_BYTECRDCMDQ_0 + 4 * ((pf_q_num) >> 4))
243 
244 /* extracts the QM queue number for the specified port and vnic */
245 #define ECORE_PF_Q_NUM(q_num, port, vnic)\
246 	((((port) << 1) | (vnic)) * 16 + (q_num))
247 
248 
249 /* Maps the specified queue to the specified COS */
250 static inline void ecore_map_q_cos(struct bxe_softc *sc, uint32_t q_num, uint32_t new_cos)
251 {
252 	/* find current COS mapping */
253 	uint32_t curr_cos = REG_RD(sc, QM_REG_QVOQIDX_0 + q_num * 4);
254 
255 	/* check if queue->COS mapping has changed */
256 	if (curr_cos != new_cos) {
257 		uint32_t num_vnics = ECORE_PORT2_MODE_NUM_VNICS;
258 		uint32_t reg_addr, reg_bit_map, vnic;
259 
260 		/* update parameters for 4port mode */
261 		if (INIT_MODE_FLAGS(sc) & MODE_PORT4) {
262 			num_vnics = ECORE_PORT4_MODE_NUM_VNICS;
263 			if (PORT_ID(sc)) {
264 				curr_cos += ECORE_E3B0_PORT1_COS_OFFSET;
265 				new_cos += ECORE_E3B0_PORT1_COS_OFFSET;
266 			}
267 		}
268 
269 		/* change queue mapping for each VNIC */
270 		for (vnic = 0; vnic < num_vnics; vnic++) {
271 			uint32_t pf_q_num =
272 				ECORE_PF_Q_NUM(q_num, PORT_ID(sc), vnic);
273 			uint32_t q_bit_map = 1 << (pf_q_num & 0x1f);
274 
275 			/* overwrite queue->VOQ mapping */
276 			REG_WR(sc, ECORE_Q_VOQ_REG_ADDR(pf_q_num), new_cos);
277 
278 			/* clear queue bit from current COS bit map */
279 			reg_addr = ECORE_VOQ_Q_REG_ADDR(curr_cos, pf_q_num);
280 			reg_bit_map = REG_RD(sc, reg_addr);
281 			REG_WR(sc, reg_addr, reg_bit_map & (~q_bit_map));
282 
283 			/* set queue bit in new COS bit map */
284 			reg_addr = ECORE_VOQ_Q_REG_ADDR(new_cos, pf_q_num);
285 			reg_bit_map = REG_RD(sc, reg_addr);
286 			REG_WR(sc, reg_addr, reg_bit_map | q_bit_map);
287 
288 			/* set/clear queue bit in command-queue bit map
289 			(E2/E3A0 only, valid COS values are 0/1) */
290 			if (!(INIT_MODE_FLAGS(sc) & MODE_E3_B0)) {
291 				reg_addr = ECORE_Q_CMDQ_REG_ADDR(pf_q_num);
292 				reg_bit_map = REG_RD(sc, reg_addr);
293 				q_bit_map = 1 << (2 * (pf_q_num & 0xf));
294 				reg_bit_map = new_cos ?
295 					      (reg_bit_map | q_bit_map) :
296 					      (reg_bit_map & (~q_bit_map));
297 				REG_WR(sc, reg_addr, reg_bit_map);
298 			}
299 		}
300 	}
301 }
302 
303 /* Configures the QM according to the specified per-traffic-type COSes */
304 static inline void ecore_dcb_config_qm(struct bxe_softc *sc, enum cos_mode mode,
305 				       struct priority_cos *traffic_cos)
306 {
307 	ecore_map_q_cos(sc, ECORE_FCOE_Q,
308 			traffic_cos[LLFC_TRAFFIC_TYPE_FCOE].cos);
309 	ecore_map_q_cos(sc, ECORE_ISCSI_Q,
310 			traffic_cos[LLFC_TRAFFIC_TYPE_ISCSI].cos);
311 	ecore_map_q_cos(sc, ECORE_ISCSI_ACK_Q,
312 		traffic_cos[LLFC_TRAFFIC_TYPE_ISCSI].cos);
313 	if (mode != STATIC_COS) {
314 		/* required only in OVERRIDE_COS mode */
315 		ecore_map_q_cos(sc, ECORE_ETH_Q,
316 				traffic_cos[LLFC_TRAFFIC_TYPE_NW].cos);
317 		ecore_map_q_cos(sc, ECORE_TOE_Q,
318 				traffic_cos[LLFC_TRAFFIC_TYPE_NW].cos);
319 		ecore_map_q_cos(sc, ECORE_TOE_ACK_Q,
320 				traffic_cos[LLFC_TRAFFIC_TYPE_NW].cos);
321 	}
322 }
323 
324 
325 /*
326  * congestion management port init api description
327  * the api works as follows:
328  * the driver should pass the cmng_init_input struct, the port_init function
329  * will prepare the required internal ram structure which will be passed back
330  * to the driver (cmng_init) that will write it into the internal ram.
331  *
332  * IMPORTANT REMARKS:
333  * 1. the cmng_init struct does not represent the contiguous internal ram
334  *    structure. the driver should use the XSTORM_CMNG_PERPORT_VARS_OFFSET
335  *    offset in order to write the port sub struct and the
336  *    PFID_FROM_PORT_AND_VNIC offset for writing the vnic sub struct (in other
337  *    words - don't use memcpy!).
338  * 2. although the cmng_init struct is filled for the maximal vnic number
339  *    possible, the driver should only write the valid vnics into the internal
340  *    ram according to the appropriate port mode.
341  */
342 #define BITS_TO_BYTES(x) ((x)/8)
343 
344 /* CMNG constants, as derived from system spec calculations */
345 
346 /* default MIN rate in case VNIC min rate is configured to zero- 100Mbps */
347 #define DEF_MIN_RATE 100
348 
349 /* resolution of the rate shaping timer - 400 usec */
350 #define RS_PERIODIC_TIMEOUT_USEC 400
351 
352 /*
353  *  number of bytes in single QM arbitration cycle -
354  *  coefficient for calculating the fairness timer
355  */
356 #define QM_ARB_BYTES 160000
357 
358 /* resolution of Min algorithm 1:100 */
359 #define MIN_RES 100
360 
361 /*
362  *  how many bytes above threshold for
363  *  the minimal credit of Min algorithm
364  */
365 #define MIN_ABOVE_THRESH 32768
366 
367 /*
368  *  Fairness algorithm integration time coefficient -
369  *  for calculating the actual Tfair
370  */
371 #define T_FAIR_COEF ((MIN_ABOVE_THRESH + QM_ARB_BYTES) * 8 * MIN_RES)
372 
373 /* Memory of fairness algorithm - 2 cycles */
374 #define FAIR_MEM 2
375 #define SAFC_TIMEOUT_USEC 52
376 
377 #define SDM_TICKS 4
378 
379 
380 static inline void ecore_init_max(const struct cmng_init_input *input_data,
381 				  uint32_t r_param, struct cmng_init *ram_data)
382 {
383 	uint32_t vnic;
384 	struct cmng_vnic *vdata = &ram_data->vnic;
385 	struct cmng_struct_per_port *pdata = &ram_data->port;
386 	/*
387 	 * rate shaping per-port variables
388 	 *  100 micro seconds in SDM ticks = 25
389 	 *  since each tick is 4 microSeconds
390 	 */
391 
392 	pdata->rs_vars.rs_periodic_timeout =
393 	RS_PERIODIC_TIMEOUT_USEC / SDM_TICKS;
394 
395 	/* this is the threshold below which no timer arming will occur.
396 	 *  1.25 coefficient is for the threshold to be a little bigger
397 	 *  then the real time to compensate for timer in-accuracy
398 	 */
399 	pdata->rs_vars.rs_threshold =
400 	(5 * RS_PERIODIC_TIMEOUT_USEC * r_param)/4;
401 
402 	/* rate shaping per-vnic variables */
403 	for (vnic = 0; vnic < ECORE_PORT2_MODE_NUM_VNICS; vnic++) {
404 		/* global vnic counter */
405 		vdata->vnic_max_rate[vnic].vn_counter.rate =
406 		input_data->vnic_max_rate[vnic];
407 		/*
408 		 * maximal Mbps for this vnic
409 		 * the quota in each timer period - number of bytes
410 		 * transmitted in this period
411 		 */
412 		vdata->vnic_max_rate[vnic].vn_counter.quota =
413 			RS_PERIODIC_TIMEOUT_USEC *
414 			(uint32_t)vdata->vnic_max_rate[vnic].vn_counter.rate / 8;
415 	}
416 
417 }
418 
419 static inline void ecore_init_max_per_vn(uint16_t vnic_max_rate,
420 				  struct rate_shaping_vars_per_vn *ram_data)
421 {
422 	/* global vnic counter */
423 	ram_data->vn_counter.rate = vnic_max_rate;
424 
425 	/*
426 	* maximal Mbps for this vnic
427 	* the quota in each timer period - number of bytes
428 	* transmitted in this period
429 	*/
430 	ram_data->vn_counter.quota =
431 		RS_PERIODIC_TIMEOUT_USEC * (uint32_t)vnic_max_rate / 8;
432 }
433 
434 static inline void ecore_init_min(const struct cmng_init_input *input_data,
435 				  uint32_t r_param, struct cmng_init *ram_data)
436 {
437 	uint32_t vnic, fair_periodic_timeout_usec, vnicWeightSum, tFair;
438 	struct cmng_vnic *vdata = &ram_data->vnic;
439 	struct cmng_struct_per_port *pdata = &ram_data->port;
440 
441 	/* this is the resolution of the fairness timer */
442 	fair_periodic_timeout_usec = QM_ARB_BYTES / r_param;
443 
444 	/*
445 	 * fairness per-port variables
446 	 * for 10G it is 1000usec. for 1G it is 10000usec.
447 	 */
448 	tFair = T_FAIR_COEF / input_data->port_rate;
449 
450 	/* this is the threshold below which we won't arm the timer anymore */
451 	pdata->fair_vars.fair_threshold = QM_ARB_BYTES;
452 
453 	/*
454 	 *  we multiply by 1e3/8 to get bytes/msec. We don't want the credits
455 	 *  to pass a credit of the T_FAIR*FAIR_MEM (algorithm resolution)
456 	 */
457 	pdata->fair_vars.upper_bound = r_param * tFair * FAIR_MEM;
458 
459 	/* since each tick is 4 microSeconds */
460 	pdata->fair_vars.fairness_timeout =
461 				fair_periodic_timeout_usec / SDM_TICKS;
462 
463 	/* calculate sum of weights */
464 	vnicWeightSum = 0;
465 
466 	for (vnic = 0; vnic < ECORE_PORT2_MODE_NUM_VNICS; vnic++)
467 		vnicWeightSum += input_data->vnic_min_rate[vnic];
468 
469 	/* global vnic counter */
470 	if (vnicWeightSum > 0) {
471 		/* fairness per-vnic variables */
472 		for (vnic = 0; vnic < ECORE_PORT2_MODE_NUM_VNICS; vnic++) {
473 			/*
474 			 *  this is the credit for each period of the fairness
475 			 *  algorithm - number of bytes in T_FAIR (this vnic
476 			 *  share of the port rate)
477 			 */
478 			vdata->vnic_min_rate[vnic].vn_credit_delta =
479 				((uint32_t)(input_data->vnic_min_rate[vnic]) * 100 *
480 				(T_FAIR_COEF / (8 * 100 * vnicWeightSum)));
481 			if (vdata->vnic_min_rate[vnic].vn_credit_delta <
482 			    pdata->fair_vars.fair_threshold +
483 			    MIN_ABOVE_THRESH) {
484 				vdata->vnic_min_rate[vnic].vn_credit_delta =
485 					pdata->fair_vars.fair_threshold +
486 					MIN_ABOVE_THRESH;
487 			}
488 		}
489 	}
490 }
491 
492 static inline void ecore_init_fw_wrr(const struct cmng_init_input *input_data,
493 				     uint32_t r_param, struct cmng_init *ram_data)
494 {
495 	uint32_t vnic, cos;
496 	uint32_t cosWeightSum = 0;
497 	struct cmng_vnic *vdata = &ram_data->vnic;
498 	struct cmng_struct_per_port *pdata = &ram_data->port;
499 
500 	for (cos = 0; cos < MAX_COS_NUMBER; cos++)
501 		cosWeightSum += input_data->cos_min_rate[cos];
502 
503 	if (cosWeightSum > 0) {
504 
505 		for (vnic = 0; vnic < ECORE_PORT2_MODE_NUM_VNICS; vnic++) {
506 			/*
507 			 *  Since cos and vnic shouldn't work together the rate
508 			 *  to divide between the coses is the port rate.
509 			 */
510 			uint32_t *ccd = vdata->vnic_min_rate[vnic].cos_credit_delta;
511 			for (cos = 0; cos < MAX_COS_NUMBER; cos++) {
512 				/*
513 				 * this is the credit for each period of
514 				 * the fairness algorithm - number of bytes
515 				 * in T_FAIR (this cos share of the vnic rate)
516 				 */
517 				ccd[cos] =
518 				    ((uint32_t)input_data->cos_min_rate[cos] * 100 *
519 				    (T_FAIR_COEF / (8 * 100 * cosWeightSum)));
520 				 if (ccd[cos] < pdata->fair_vars.fair_threshold
521 						+ MIN_ABOVE_THRESH) {
522 					ccd[cos] =
523 					    pdata->fair_vars.fair_threshold +
524 					    MIN_ABOVE_THRESH;
525 				}
526 			}
527 		}
528 	}
529 }
530 
531 static inline void ecore_init_safc(const struct cmng_init_input *input_data,
532 				   struct cmng_init *ram_data)
533 {
534 	/* in microSeconds */
535 	ram_data->port.safc_vars.safc_timeout_usec = SAFC_TIMEOUT_USEC;
536 }
537 
538 /* Congestion management port init */
539 static inline void ecore_init_cmng(const struct cmng_init_input *input_data,
540 				   struct cmng_init *ram_data)
541 {
542 	uint32_t r_param;
543 	ECORE_MEMSET(ram_data, 0,sizeof(struct cmng_init));
544 
545 	ram_data->port.flags = input_data->flags;
546 
547 	/*
548 	 *  number of bytes transmitted in a rate of 10Gbps
549 	 *  in one usec = 1.25KB.
550 	 */
551 	r_param = BITS_TO_BYTES(input_data->port_rate);
552 	ecore_init_max(input_data, r_param, ram_data);
553 	ecore_init_min(input_data, r_param, ram_data);
554 	ecore_init_fw_wrr(input_data, r_param, ram_data);
555 	ecore_init_safc(input_data, ram_data);
556 }
557 
558 
559 
560 
561 /* Returns the index of start or end of a specific block stage in ops array*/
562 #define BLOCK_OPS_IDX(block, stage, end) \
563 			(2*(((block)*NUM_OF_INIT_PHASES) + (stage)) + (end))
564 
565 
566 #define INITOP_SET		0	/* set the HW directly */
567 #define INITOP_CLEAR		1	/* clear the HW directly */
568 #define INITOP_INIT		2	/* set the init-value array */
569 
570 /****************************************************************************
571 * ILT management
572 ****************************************************************************/
573 struct ilt_line {
574 	ecore_dma_addr_t page_mapping;
575 	void *page;
576 	uint32_t size;
577 };
578 
579 struct ilt_client_info {
580 	uint32_t page_size;
581 	uint16_t start;
582 	uint16_t end;
583 	uint16_t client_num;
584 	uint16_t flags;
585 #define ILT_CLIENT_SKIP_INIT	0x1
586 #define ILT_CLIENT_SKIP_MEM	0x2
587 };
588 
589 struct ecore_ilt {
590 	uint32_t start_line;
591 	struct ilt_line		*lines;
592 	struct ilt_client_info	clients[4];
593 #define ILT_CLIENT_CDU	0
594 #define ILT_CLIENT_QM	1
595 #define ILT_CLIENT_SRC	2
596 #define ILT_CLIENT_TM	3
597 };
598 
599 /****************************************************************************
600 * SRC configuration
601 ****************************************************************************/
602 struct src_ent {
603 	uint8_t opaque[56];
604 	uint64_t next;
605 };
606 
607 /****************************************************************************
608 * Parity configuration
609 ****************************************************************************/
610 #define BLOCK_PRTY_INFO(block, en_mask, m1, m1h, m2, m3) \
611 { \
612 	block##_REG_##block##_PRTY_MASK, \
613 	block##_REG_##block##_PRTY_STS_CLR, \
614 	en_mask, {m1, m1h, m2, m3}, #block \
615 }
616 
617 #define BLOCK_PRTY_INFO_0(block, en_mask, m1, m1h, m2, m3) \
618 { \
619 	block##_REG_##block##_PRTY_MASK_0, \
620 	block##_REG_##block##_PRTY_STS_CLR_0, \
621 	en_mask, {m1, m1h, m2, m3}, #block"_0" \
622 }
623 
624 #define BLOCK_PRTY_INFO_1(block, en_mask, m1, m1h, m2, m3) \
625 { \
626 	block##_REG_##block##_PRTY_MASK_1, \
627 	block##_REG_##block##_PRTY_STS_CLR_1, \
628 	en_mask, {m1, m1h, m2, m3}, #block"_1" \
629 }
630 
631 static const struct {
632 	uint32_t mask_addr;
633 	uint32_t sts_clr_addr;
634 	uint32_t en_mask;		/* Mask to enable parity attentions */
635 	struct {
636 		uint32_t e1;		/* 57710 */
637 		uint32_t e1h;	/* 57711 */
638 		uint32_t e2;		/* 57712 */
639 		uint32_t e3;		/* 578xx */
640 	} reg_mask;		/* Register mask (all valid bits) */
641 	char name[8];		/* Block's longest name is 7 characters long
642 				 * (name + suffix)
643 				 */
644 } ecore_blocks_parity_data[] = {
645 	/* bit 19 masked */
646 	/* REG_WR(bp, PXP_REG_PXP_PRTY_MASK, 0x80000); */
647 	/* bit 5,18,20-31 */
648 	/* REG_WR(bp, PXP2_REG_PXP2_PRTY_MASK_0, 0xfff40020); */
649 	/* bit 5 */
650 	/* REG_WR(bp, PXP2_REG_PXP2_PRTY_MASK_1, 0x20);	*/
651 	/* REG_WR(bp, HC_REG_HC_PRTY_MASK, 0x0); */
652 	/* REG_WR(bp, MISC_REG_MISC_PRTY_MASK, 0x0); */
653 
654 	/* Block IGU, MISC, PXP and PXP2 parity errors as long as we don't
655 	 * want to handle "system kill" flow at the moment.
656 	 */
657 	BLOCK_PRTY_INFO(PXP, 0x7ffffff, 0x3ffffff, 0x3ffffff, 0x7ffffff,
658 			0x7ffffff),
659 	BLOCK_PRTY_INFO_0(PXP2,	0xffffffff, 0xffffffff, 0xffffffff, 0xffffffff,
660 			  0xffffffff),
661 	BLOCK_PRTY_INFO_1(PXP2,	0x1ffffff, 0x7f, 0x7f, 0x7ff, 0x1ffffff),
662 	BLOCK_PRTY_INFO(HC, 0x7, 0x7, 0x7, 0, 0),
663 	BLOCK_PRTY_INFO(NIG, 0xffffffff, 0x3fffffff, 0xffffffff, 0, 0),
664 	BLOCK_PRTY_INFO_0(NIG,	0xffffffff, 0, 0, 0xffffffff, 0xffffffff),
665 	BLOCK_PRTY_INFO_1(NIG,	0xffff, 0, 0, 0xff, 0xffff),
666 	BLOCK_PRTY_INFO(IGU, 0x7ff, 0, 0, 0x7ff, 0x7ff),
667 	BLOCK_PRTY_INFO(MISC, 0x1, 0x1, 0x1, 0x1, 0x1),
668 	BLOCK_PRTY_INFO(QM, 0, 0x1ff, 0xfff, 0xfff, 0xfff),
669 	BLOCK_PRTY_INFO(ATC, 0x1f, 0, 0, 0x1f, 0x1f),
670 	BLOCK_PRTY_INFO(PGLUE_B, 0x3, 0, 0, 0x3, 0x3),
671 	BLOCK_PRTY_INFO(DORQ, 0, 0x3, 0x3, 0x3, 0x3),
672 	{GRCBASE_UPB + PB_REG_PB_PRTY_MASK,
673 		GRCBASE_UPB + PB_REG_PB_PRTY_STS_CLR, 0xf,
674 		{0xf, 0xf, 0xf, 0xf}, "UPB"},
675 	{GRCBASE_XPB + PB_REG_PB_PRTY_MASK,
676 		GRCBASE_XPB + PB_REG_PB_PRTY_STS_CLR, 0,
677 		{0xf, 0xf, 0xf, 0xf}, "XPB"},
678 	BLOCK_PRTY_INFO(SRC, 0x4, 0x7, 0x7, 0x7, 0x7),
679 	BLOCK_PRTY_INFO(CDU, 0, 0x1f, 0x1f, 0x1f, 0x1f),
680 	BLOCK_PRTY_INFO(CFC, 0, 0xf, 0xf, 0xf, 0x3f),
681 	BLOCK_PRTY_INFO(DBG, 0, 0x1, 0x1, 0x1, 0x1),
682 	BLOCK_PRTY_INFO(DMAE, 0, 0xf, 0xf, 0xf, 0xf),
683 	BLOCK_PRTY_INFO(BRB1, 0, 0xf, 0xf, 0xf, 0xf),
684 	BLOCK_PRTY_INFO(PRS, (1<<6), 0xff, 0xff, 0xff, 0xff),
685 	BLOCK_PRTY_INFO(PBF, 0, 0, 0x3ffff, 0xfffff, 0xfffffff),
686 	BLOCK_PRTY_INFO(TM, 0, 0, 0x7f, 0x7f, 0x7f),
687 	BLOCK_PRTY_INFO(TSDM, 0x18, 0x7ff, 0x7ff, 0x7ff, 0x7ff),
688 	BLOCK_PRTY_INFO(CSDM, 0x8, 0x7ff, 0x7ff, 0x7ff, 0x7ff),
689 	BLOCK_PRTY_INFO(USDM, 0x38, 0x7ff, 0x7ff, 0x7ff, 0x7ff),
690 	BLOCK_PRTY_INFO(XSDM, 0x8, 0x7ff, 0x7ff, 0x7ff, 0x7ff),
691 	BLOCK_PRTY_INFO(TCM, 0, 0, 0x7ffffff, 0x7ffffff, 0x7ffffff),
692 	BLOCK_PRTY_INFO(CCM, 0, 0, 0x7ffffff, 0x7ffffff, 0x7ffffff),
693 	BLOCK_PRTY_INFO(UCM, 0, 0, 0x7ffffff, 0x7ffffff, 0x7ffffff),
694 	BLOCK_PRTY_INFO(XCM, 0, 0, 0x3fffffff, 0x3fffffff, 0x3fffffff),
695 	BLOCK_PRTY_INFO_0(TSEM, 0, 0xffffffff, 0xffffffff, 0xffffffff,
696 			  0xffffffff),
697 	BLOCK_PRTY_INFO_1(TSEM, 0, 0x3, 0x1f, 0x3f, 0x3f),
698 	BLOCK_PRTY_INFO_0(USEM, 0, 0xffffffff, 0xffffffff, 0xffffffff,
699 			  0xffffffff),
700 	BLOCK_PRTY_INFO_1(USEM, 0, 0x3, 0x1f, 0x1f, 0x1f),
701 	BLOCK_PRTY_INFO_0(CSEM, 0, 0xffffffff, 0xffffffff, 0xffffffff,
702 			  0xffffffff),
703 	BLOCK_PRTY_INFO_1(CSEM, 0, 0x3, 0x1f, 0x1f, 0x1f),
704 	BLOCK_PRTY_INFO_0(XSEM, 0, 0xffffffff, 0xffffffff, 0xffffffff,
705 			  0xffffffff),
706 	BLOCK_PRTY_INFO_1(XSEM, 0, 0x3, 0x1f, 0x3f, 0x3f),
707 };
708 
709 
710 /* [28] MCP Latched rom_parity
711  * [29] MCP Latched ump_rx_parity
712  * [30] MCP Latched ump_tx_parity
713  * [31] MCP Latched scpad_parity
714  */
715 #define MISC_AEU_ENABLE_MCP_PRTY_SUB_BITS	\
716 	(AEU_INPUTS_ATTN_BITS_MCP_LATCHED_ROM_PARITY | \
717 	 AEU_INPUTS_ATTN_BITS_MCP_LATCHED_UMP_RX_PARITY | \
718 	 AEU_INPUTS_ATTN_BITS_MCP_LATCHED_UMP_TX_PARITY)
719 
720 #define MISC_AEU_ENABLE_MCP_PRTY_BITS	\
721 	(MISC_AEU_ENABLE_MCP_PRTY_SUB_BITS | \
722 	 AEU_INPUTS_ATTN_BITS_MCP_LATCHED_SCPAD_PARITY)
723 
724 /* Below registers control the MCP parity attention output. When
725  * MISC_AEU_ENABLE_MCP_PRTY_BITS are set - attentions are
726  * enabled, when cleared - disabled.
727  */
728 static const struct {
729 	uint32_t addr;
730 	uint32_t bits;
731 } mcp_attn_ctl_regs[] = {
732 	{ MISC_REG_AEU_ENABLE4_FUNC_0_OUT_0,
733 		MISC_AEU_ENABLE_MCP_PRTY_BITS },
734 	{ MISC_REG_AEU_ENABLE4_NIG_0,
735 		MISC_AEU_ENABLE_MCP_PRTY_SUB_BITS },
736 	{ MISC_REG_AEU_ENABLE4_PXP_0,
737 		MISC_AEU_ENABLE_MCP_PRTY_SUB_BITS },
738 	{ MISC_REG_AEU_ENABLE4_FUNC_1_OUT_0,
739 		MISC_AEU_ENABLE_MCP_PRTY_BITS },
740 	{ MISC_REG_AEU_ENABLE4_NIG_1,
741 		MISC_AEU_ENABLE_MCP_PRTY_SUB_BITS },
742 	{ MISC_REG_AEU_ENABLE4_PXP_1,
743 		MISC_AEU_ENABLE_MCP_PRTY_SUB_BITS }
744 };
745 
746 static inline void ecore_set_mcp_parity(struct bxe_softc *sc, uint8_t enable)
747 {
748 	int i;
749 	uint32_t reg_val;
750 
751 	for (i = 0; i < ARRSIZE(mcp_attn_ctl_regs); i++) {
752 		reg_val = REG_RD(sc, mcp_attn_ctl_regs[i].addr);
753 
754 		if (enable)
755 			reg_val |= MISC_AEU_ENABLE_MCP_PRTY_BITS; /* Linux is using mcp_attn_ctl_regs[i].bits */
756 		else
757 			reg_val &= ~MISC_AEU_ENABLE_MCP_PRTY_BITS; /* Linux is using mcp_attn_ctl_regs[i].bits */
758 
759 		REG_WR(sc, mcp_attn_ctl_regs[i].addr, reg_val);
760 	}
761 }
762 
763 static inline uint32_t ecore_parity_reg_mask(struct bxe_softc *sc, int idx)
764 {
765 	if (CHIP_IS_E1(sc))
766 		return ecore_blocks_parity_data[idx].reg_mask.e1;
767 	else if (CHIP_IS_E1H(sc))
768 		return ecore_blocks_parity_data[idx].reg_mask.e1h;
769 	else if (CHIP_IS_E2(sc))
770 		return ecore_blocks_parity_data[idx].reg_mask.e2;
771 	else /* CHIP_IS_E3 */
772 		return ecore_blocks_parity_data[idx].reg_mask.e3;
773 }
774 
775 static inline void ecore_disable_blocks_parity(struct bxe_softc *sc)
776 {
777 	int i;
778 
779 	for (i = 0; i < ARRSIZE(ecore_blocks_parity_data); i++) {
780 		uint32_t dis_mask = ecore_parity_reg_mask(sc, i);
781 
782 		if (dis_mask) {
783 			REG_WR(sc, ecore_blocks_parity_data[i].mask_addr,
784 			       dis_mask);
785 			ECORE_MSG(sc, "Setting parity mask "
786 						 "for %s to\t\t0x%x\n",
787 				    ecore_blocks_parity_data[i].name, dis_mask);
788 		}
789 	}
790 
791 	/* Disable MCP parity attentions */
792 	ecore_set_mcp_parity(sc, FALSE);
793 }
794 
795 /**
796  * Clear the parity error status registers.
797  */
798 static inline void ecore_clear_blocks_parity(struct bxe_softc *sc)
799 {
800 	int i;
801 	uint32_t reg_val, mcp_aeu_bits =
802 		AEU_INPUTS_ATTN_BITS_MCP_LATCHED_ROM_PARITY |
803 		AEU_INPUTS_ATTN_BITS_MCP_LATCHED_SCPAD_PARITY |
804 		AEU_INPUTS_ATTN_BITS_MCP_LATCHED_UMP_RX_PARITY |
805 		AEU_INPUTS_ATTN_BITS_MCP_LATCHED_UMP_TX_PARITY;
806 
807 	/* Clear SEM_FAST parities */
808 	REG_WR(sc, XSEM_REG_FAST_MEMORY + SEM_FAST_REG_PARITY_RST, 0x1);
809 	REG_WR(sc, TSEM_REG_FAST_MEMORY + SEM_FAST_REG_PARITY_RST, 0x1);
810 	REG_WR(sc, USEM_REG_FAST_MEMORY + SEM_FAST_REG_PARITY_RST, 0x1);
811 	REG_WR(sc, CSEM_REG_FAST_MEMORY + SEM_FAST_REG_PARITY_RST, 0x1);
812 
813 	for (i = 0; i < ARRSIZE(ecore_blocks_parity_data); i++) {
814 		uint32_t reg_mask = ecore_parity_reg_mask(sc, i);
815 
816 		if (reg_mask) {
817 			reg_val = REG_RD(sc, ecore_blocks_parity_data[i].
818 					 sts_clr_addr);
819 			if (reg_val & reg_mask)
820 				ECORE_MSG(sc,
821 					   "Parity errors in %s: 0x%x\n",
822 					   ecore_blocks_parity_data[i].name,
823 					   reg_val & reg_mask);
824 		}
825 	}
826 
827 	/* Check if there were parity attentions in MCP */
828 	reg_val = REG_RD(sc, MISC_REG_AEU_AFTER_INVERT_4_MCP);
829 	if (reg_val & mcp_aeu_bits)
830 		ECORE_MSG(sc, "Parity error in MCP: 0x%x\n",
831 			   reg_val & mcp_aeu_bits);
832 
833 	/* Clear parity attentions in MCP:
834 	 * [7]  clears Latched rom_parity
835 	 * [8]  clears Latched ump_rx_parity
836 	 * [9]  clears Latched ump_tx_parity
837 	 * [10] clears Latched scpad_parity (both ports)
838 	 */
839 	REG_WR(sc, MISC_REG_AEU_CLR_LATCH_SIGNAL, 0x780);
840 }
841 
842 static inline void ecore_enable_blocks_parity(struct bxe_softc *sc)
843 {
844 	int i;
845 
846 	for (i = 0; i < ARRSIZE(ecore_blocks_parity_data); i++) {
847 		uint32_t reg_mask = ecore_parity_reg_mask(sc, i);
848 
849 		if (reg_mask)
850 			REG_WR(sc, ecore_blocks_parity_data[i].mask_addr,
851 				ecore_blocks_parity_data[i].en_mask & reg_mask);
852 	}
853 
854 	/* Enable MCP parity attentions */
855 	ecore_set_mcp_parity(sc, TRUE);
856 }
857 
858 
859 #endif /* ECORE_INIT_H */
860 
861