xref: /linux/drivers/edac/i10nm_base.c (revision 03f76ddff5b04a808ae16c06418460151e2fdd4b)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Driver for Intel(R) 10nm server memory controller.
4  * Copyright (c) 2019, Intel Corporation.
5  *
6  */
7 
8 #include <linux/kernel.h>
9 #include <linux/io.h>
10 #include <asm/cpu_device_id.h>
11 #include <asm/intel-family.h>
12 #include <asm/mce.h>
13 #include "edac_module.h"
14 #include "skx_common.h"
15 
16 #define I10NM_REVISION	"v0.0.6"
17 #define EDAC_MOD_STR	"i10nm_edac"
18 
19 /* Debug macros */
20 #define i10nm_printk(level, fmt, arg...)	\
21 	edac_printk(level, "i10nm", fmt, ##arg)
22 
23 #define I10NM_GET_SCK_BAR(d, reg)	\
24 	pci_read_config_dword((d)->uracu, 0xd0, &(reg))
25 #define I10NM_GET_IMC_BAR(d, i, reg)		\
26 	pci_read_config_dword((d)->uracu,	\
27 	(res_cfg->type == GNR ? 0xd4 : 0xd8) + (i) * 4, &(reg))
28 #define I10NM_GET_SAD(d, offset, i, reg)\
29 	pci_read_config_dword((d)->sad_all, (offset) + (i) * \
30 	(res_cfg->type == GNR ? 12 : 8), &(reg))
31 #define I10NM_GET_HBM_IMC_BAR(d, reg)	\
32 	pci_read_config_dword((d)->uracu, 0xd4, &(reg))
33 #define I10NM_GET_CAPID3_CFG(d, reg)	\
34 	pci_read_config_dword((d)->pcu_cr3,	\
35 	res_cfg->type == GNR ? 0x290 : 0x90, &(reg))
36 #define I10NM_GET_CAPID5_CFG(d, reg)	\
37 	pci_read_config_dword((d)->pcu_cr3,	\
38 	res_cfg->type == GNR ? 0x298 : 0x98, &(reg))
39 #define I10NM_GET_DIMMMTR(m, i, j)	\
40 	readl((m)->mbase + ((m)->hbm_mc ? 0x80c :	\
41 	(res_cfg->type == GNR ? 0xc0c : 0x2080c)) +	\
42 	(i) * (m)->chan_mmio_sz + (j) * 4)
43 #define I10NM_GET_MCDDRTCFG(m, i)	\
44 	readl((m)->mbase + ((m)->hbm_mc ? 0x970 : 0x20970) + \
45 	(i) * (m)->chan_mmio_sz)
46 #define I10NM_GET_MCMTR(m, i)		\
47 	readl((m)->mbase + ((m)->hbm_mc ? 0xef8 :	\
48 	(res_cfg->type == GNR ? 0xaf8 : 0x20ef8)) +	\
49 	(i) * (m)->chan_mmio_sz)
50 #define I10NM_GET_REG32(m, i, offset)	\
51 	readl((m)->mbase + (i) * (m)->chan_mmio_sz + (offset))
52 #define I10NM_GET_REG64(m, i, offset)	\
53 	readq((m)->mbase + (i) * (m)->chan_mmio_sz + (offset))
54 #define I10NM_SET_REG32(m, i, offset, v)	\
55 	writel(v, (m)->mbase + (i) * (m)->chan_mmio_sz + (offset))
56 
57 #define I10NM_GET_SCK_MMIO_BASE(reg)	(GET_BITFIELD(reg, 0, 28) << 23)
58 #define I10NM_GET_IMC_MMIO_OFFSET(reg)	(GET_BITFIELD(reg, 0, 10) << 12)
59 #define I10NM_GET_IMC_MMIO_SIZE(reg)	((GET_BITFIELD(reg, 13, 23) - \
60 					 GET_BITFIELD(reg, 0, 10) + 1) << 12)
61 #define I10NM_GET_HBM_IMC_MMIO_OFFSET(reg)	\
62 	((GET_BITFIELD(reg, 0, 10) << 12) + 0x140000)
63 
64 #define I10NM_GNR_IMC_MMIO_OFFSET	0x24c000
65 #define I10NM_GNR_D_IMC_MMIO_OFFSET	0x206000
66 #define I10NM_GNR_IMC_MMIO_SIZE		0x4000
67 #define I10NM_HBM_IMC_MMIO_SIZE		0x9000
68 #define I10NM_DDR_IMC_CH_CNT(reg)	GET_BITFIELD(reg, 21, 24)
69 #define I10NM_IS_HBM_PRESENT(reg)	GET_BITFIELD(reg, 27, 30)
70 #define I10NM_IS_HBM_IMC(reg)		GET_BITFIELD(reg, 29, 29)
71 
72 #define I10NM_MAX_SAD			16
73 #define I10NM_SAD_ENABLE(reg)		GET_BITFIELD(reg, 0, 0)
74 #define I10NM_SAD_NM_CACHEABLE(reg)	GET_BITFIELD(reg, 5, 5)
75 
76 static struct list_head *i10nm_edac_list;
77 
78 static struct res_config *res_cfg;
79 static int retry_rd_err_log;
80 static int decoding_via_mca;
81 static bool mem_cfg_2lm;
82 
83 static struct reg_rrl icx_reg_rrl_ddr = {
84 	.set_num = 2,
85 	.reg_num = 6,
86 	.modes = {LRE_SCRUB, LRE_DEMAND},
87 	.offsets = {
88 		{0x22c60, 0x22c54, 0x22c5c, 0x22c58, 0x22c28, 0x20ed8},
89 		{0x22e54, 0x22e60, 0x22e64, 0x22e58, 0x22e5c, 0x20ee0},
90 	},
91 	.widths		= {4, 4, 4, 4, 4, 8},
92 	.v_mask		= BIT(0),
93 	.uc_mask	= BIT(1),
94 	.over_mask	= BIT(2),
95 	.en_patspr_mask	= BIT(13),
96 	.noover_mask	= BIT(14),
97 	.en_mask	= BIT(15),
98 
99 	.cecnt_num	= 4,
100 	.cecnt_offsets	= {0x22c18, 0x22c1c, 0x22c20, 0x22c24},
101 	.cecnt_widths	= {4, 4, 4, 4},
102 };
103 
104 static struct reg_rrl spr_reg_rrl_ddr = {
105 	.set_num = 3,
106 	.reg_num = 6,
107 	.modes = {LRE_SCRUB, LRE_DEMAND, FRE_DEMAND},
108 	.offsets = {
109 		{0x22c60, 0x22c54, 0x22f08, 0x22c58, 0x22c28, 0x20ed8},
110 		{0x22e54, 0x22e60, 0x22f10, 0x22e58, 0x22e5c, 0x20ee0},
111 		{0x22c70, 0x22d80, 0x22f18, 0x22d58, 0x22c64, 0x20f10},
112 	},
113 	.widths		= {4, 4, 8, 4, 4, 8},
114 	.v_mask		= BIT(0),
115 	.uc_mask	= BIT(1),
116 	.over_mask	= BIT(2),
117 	.en_patspr_mask	= BIT(13),
118 	.noover_mask	= BIT(14),
119 	.en_mask	= BIT(15),
120 
121 	.cecnt_num	= 4,
122 	.cecnt_offsets	= {0x22c18, 0x22c1c, 0x22c20, 0x22c24},
123 	.cecnt_widths	= {4, 4, 4, 4},
124 };
125 
126 static struct reg_rrl spr_reg_rrl_hbm_pch0 = {
127 	.set_num = 2,
128 	.reg_num = 6,
129 	.modes = {LRE_SCRUB, LRE_DEMAND},
130 	.offsets = {
131 		{0x2860, 0x2854, 0x2b08, 0x2858, 0x2828, 0x0ed8},
132 		{0x2a54, 0x2a60, 0x2b10, 0x2a58, 0x2a5c, 0x0ee0},
133 	},
134 	.widths		= {4, 4, 8, 4, 4, 8},
135 	.v_mask		= BIT(0),
136 	.uc_mask	= BIT(1),
137 	.over_mask	= BIT(2),
138 	.en_patspr_mask	= BIT(13),
139 	.noover_mask	= BIT(14),
140 	.en_mask	= BIT(15),
141 
142 	.cecnt_num	= 4,
143 	.cecnt_offsets	= {0x2818, 0x281c, 0x2820, 0x2824},
144 	.cecnt_widths	= {4, 4, 4, 4},
145 };
146 
147 static struct reg_rrl spr_reg_rrl_hbm_pch1 = {
148 	.set_num = 2,
149 	.reg_num = 6,
150 	.modes = {LRE_SCRUB, LRE_DEMAND},
151 	.offsets = {
152 		{0x2c60, 0x2c54, 0x2f08, 0x2c58, 0x2c28, 0x0fa8},
153 		{0x2e54, 0x2e60, 0x2f10, 0x2e58, 0x2e5c, 0x0fb0},
154 	},
155 	.widths		= {4, 4, 8, 4, 4, 8},
156 	.v_mask		= BIT(0),
157 	.uc_mask	= BIT(1),
158 	.over_mask	= BIT(2),
159 	.en_patspr_mask	= BIT(13),
160 	.noover_mask	= BIT(14),
161 	.en_mask	= BIT(15),
162 
163 	.cecnt_num	= 4,
164 	.cecnt_offsets	= {0x2c18, 0x2c1c, 0x2c20, 0x2c24},
165 	.cecnt_widths	= {4, 4, 4, 4},
166 };
167 
168 static struct reg_rrl gnr_reg_rrl_ddr = {
169 	.set_num = 4,
170 	.reg_num = 6,
171 	.modes = {FRE_SCRUB, FRE_DEMAND, LRE_SCRUB, LRE_DEMAND},
172 	.offsets = {
173 		{0x2f10, 0x2f20, 0x2f30, 0x2f50, 0x2f60, 0xba0},
174 		{0x2f14, 0x2f24, 0x2f38, 0x2f54, 0x2f64, 0xba8},
175 		{0x2f18, 0x2f28, 0x2f40, 0x2f58, 0x2f68, 0xbb0},
176 		{0x2f1c, 0x2f2c, 0x2f48, 0x2f5c, 0x2f6c, 0xbb8},
177 	},
178 	.widths		= {4, 4, 8, 4, 4, 8},
179 	.v_mask		= BIT(0),
180 	.uc_mask	= BIT(1),
181 	.over_mask	= BIT(2),
182 	.en_patspr_mask	= BIT(14),
183 	.noover_mask	= BIT(15),
184 	.en_mask	= BIT(12),
185 
186 	.cecnt_num	= 8,
187 	.cecnt_offsets	= {0x2c10, 0x2c14, 0x2c18, 0x2c1c, 0x2c20, 0x2c24, 0x2c28, 0x2c2c},
188 	.cecnt_widths	= {4, 4, 4, 4, 4, 4, 4, 4},
189 };
190 
read_imc_reg(struct skx_imc * imc,int chan,u32 offset,u8 width)191 static u64 read_imc_reg(struct skx_imc *imc, int chan, u32 offset, u8 width)
192 {
193 	switch (width) {
194 	case 4:
195 		return I10NM_GET_REG32(imc, chan, offset);
196 	case 8:
197 		return I10NM_GET_REG64(imc, chan, offset);
198 	default:
199 		i10nm_printk(KERN_ERR, "Invalid readd RRL 0x%x width %d\n", offset, width);
200 		return 0;
201 	}
202 }
203 
write_imc_reg(struct skx_imc * imc,int chan,u32 offset,u8 width,u64 val)204 static void write_imc_reg(struct skx_imc *imc, int chan, u32 offset, u8 width, u64 val)
205 {
206 	switch (width) {
207 	case 4:
208 		return I10NM_SET_REG32(imc, chan, offset, (u32)val);
209 	default:
210 		i10nm_printk(KERN_ERR, "Invalid write RRL 0x%x width %d\n", offset, width);
211 	}
212 }
213 
enable_rrl(struct skx_imc * imc,int chan,struct reg_rrl * rrl,int rrl_set,bool enable,u32 * rrl_ctl)214 static void enable_rrl(struct skx_imc *imc, int chan, struct reg_rrl *rrl,
215 		       int rrl_set, bool enable, u32 *rrl_ctl)
216 {
217 	enum rrl_mode mode = rrl->modes[rrl_set];
218 	u32 offset = rrl->offsets[rrl_set][0], v;
219 	u8 width = rrl->widths[0];
220 	bool first, scrub;
221 
222 	/* First or last read error. */
223 	first = (mode == FRE_SCRUB || mode == FRE_DEMAND);
224 	/* Patrol scrub or on-demand read error. */
225 	scrub = (mode == FRE_SCRUB || mode == LRE_SCRUB);
226 
227 	v = read_imc_reg(imc, chan, offset, width);
228 
229 	if (enable) {
230 		/* Save default configurations. */
231 		*rrl_ctl = v;
232 		v &= ~rrl->uc_mask;
233 
234 		if (first)
235 			v |= rrl->noover_mask;
236 		else
237 			v &= ~rrl->noover_mask;
238 
239 		if (scrub)
240 			v |= rrl->en_patspr_mask;
241 		else
242 			v &= ~rrl->en_patspr_mask;
243 
244 		v |= rrl->en_mask;
245 	} else {
246 		/* Restore default configurations. */
247 		if (*rrl_ctl & rrl->uc_mask)
248 			v |= rrl->uc_mask;
249 
250 		if (first) {
251 			if (!(*rrl_ctl & rrl->noover_mask))
252 				v &= ~rrl->noover_mask;
253 		} else {
254 			if (*rrl_ctl & rrl->noover_mask)
255 				v |= rrl->noover_mask;
256 		}
257 
258 		if (scrub) {
259 			if (!(*rrl_ctl & rrl->en_patspr_mask))
260 				v &= ~rrl->en_patspr_mask;
261 		} else {
262 			if (*rrl_ctl & rrl->en_patspr_mask)
263 				v |= rrl->en_patspr_mask;
264 		}
265 
266 		if (!(*rrl_ctl & rrl->en_mask))
267 			v &= ~rrl->en_mask;
268 	}
269 
270 	write_imc_reg(imc, chan, offset, width, v);
271 }
272 
enable_rrls(struct skx_imc * imc,int chan,struct reg_rrl * rrl,bool enable,u32 * rrl_ctl)273 static void enable_rrls(struct skx_imc *imc, int chan, struct reg_rrl *rrl,
274 			bool enable, u32 *rrl_ctl)
275 {
276 	for (int i = 0; i < rrl->set_num; i++)
277 		enable_rrl(imc, chan, rrl, i, enable, rrl_ctl + i);
278 }
279 
enable_rrls_ddr(struct skx_imc * imc,bool enable)280 static void enable_rrls_ddr(struct skx_imc *imc, bool enable)
281 {
282 	struct reg_rrl *rrl_ddr = res_cfg->reg_rrl_ddr;
283 	int i, chan_num = res_cfg->ddr_chan_num;
284 	struct skx_channel *chan = imc->chan;
285 
286 	if (!imc->mbase)
287 		return;
288 
289 	for (i = 0; i < chan_num; i++)
290 		enable_rrls(imc, i, rrl_ddr, enable, chan[i].rrl_ctl[0]);
291 }
292 
enable_rrls_hbm(struct skx_imc * imc,bool enable)293 static void enable_rrls_hbm(struct skx_imc *imc, bool enable)
294 {
295 	struct reg_rrl **rrl_hbm = res_cfg->reg_rrl_hbm;
296 	int i, chan_num = res_cfg->hbm_chan_num;
297 	struct skx_channel *chan = imc->chan;
298 
299 	if (!imc->mbase || !imc->hbm_mc || !rrl_hbm[0] || !rrl_hbm[1])
300 		return;
301 
302 	for (i = 0; i < chan_num; i++) {
303 		enable_rrls(imc, i, rrl_hbm[0], enable, chan[i].rrl_ctl[0]);
304 		enable_rrls(imc, i, rrl_hbm[1], enable, chan[i].rrl_ctl[1]);
305 	}
306 }
307 
enable_retry_rd_err_log(bool enable)308 static void enable_retry_rd_err_log(bool enable)
309 {
310 	struct skx_dev *d;
311 	int i, imc_num;
312 
313 	edac_dbg(2, "\n");
314 
315 	list_for_each_entry(d, i10nm_edac_list, list) {
316 		imc_num  = res_cfg->ddr_imc_num;
317 		for (i = 0; i < imc_num; i++)
318 			enable_rrls_ddr(&d->imc[i], enable);
319 
320 		imc_num += res_cfg->hbm_imc_num;
321 		for (; i < imc_num; i++)
322 			enable_rrls_hbm(&d->imc[i], enable);
323 	}
324 }
325 
show_retry_rd_err_log(struct decoded_addr * res,char * msg,int len,bool scrub_err)326 static void show_retry_rd_err_log(struct decoded_addr *res, char *msg,
327 				  int len, bool scrub_err)
328 {
329 	int i, j, n, ch = res->channel, pch = res->cs & 1;
330 	struct skx_imc *imc = &res->dev->imc[res->imc];
331 	u64 log, corr, status_mask;
332 	struct reg_rrl *rrl;
333 	bool scrub;
334 	u32 offset;
335 	u8 width;
336 
337 	if (!imc->mbase)
338 		return;
339 
340 	rrl = imc->hbm_mc ? res_cfg->reg_rrl_hbm[pch] : res_cfg->reg_rrl_ddr;
341 
342 	if (!rrl)
343 		return;
344 
345 	status_mask = rrl->over_mask | rrl->uc_mask | rrl->v_mask;
346 
347 	n = scnprintf(msg, len, " retry_rd_err_log[");
348 	for (i = 0; i < rrl->set_num; i++) {
349 		scrub = (rrl->modes[i] == FRE_SCRUB || rrl->modes[i] == LRE_SCRUB);
350 		if (scrub_err != scrub)
351 			continue;
352 
353 		for (j = 0; j < rrl->reg_num && len - n > 0; j++) {
354 			offset = rrl->offsets[i][j];
355 			width = rrl->widths[j];
356 			log = read_imc_reg(imc, ch, offset, width);
357 
358 			if (width == 4)
359 				n += scnprintf(msg + n, len - n, "%.8llx ", log);
360 			else
361 				n += scnprintf(msg + n, len - n, "%.16llx ", log);
362 
363 			/* Clear RRL status if RRL in Linux control mode. */
364 			if (retry_rd_err_log == 2 && !j && (log & status_mask))
365 				write_imc_reg(imc, ch, offset, width, log & ~status_mask);
366 		}
367 	}
368 
369 	/* Move back one space. */
370 	n--;
371 	n += scnprintf(msg + n, len - n, "]");
372 
373 	if (len - n > 0) {
374 		n += scnprintf(msg + n, len - n, " correrrcnt[");
375 		for (i = 0; i < rrl->cecnt_num && len - n > 0; i++) {
376 			offset = rrl->cecnt_offsets[i];
377 			width = rrl->cecnt_widths[i];
378 			corr = read_imc_reg(imc, ch, offset, width);
379 
380 			/* CPUs {ICX,SPR} encode two counters per 4-byte CORRERRCNT register. */
381 			if (res_cfg->type <= SPR) {
382 				n += scnprintf(msg + n, len - n, "%.4llx %.4llx ",
383 					      corr & 0xffff, corr >> 16);
384 			} else {
385 			/* CPUs {GNR} encode one counter per CORRERRCNT register. */
386 				if (width == 4)
387 					n += scnprintf(msg + n, len - n, "%.8llx ", corr);
388 				else
389 					n += scnprintf(msg + n, len - n, "%.16llx ", corr);
390 			}
391 		}
392 
393 		/* Move back one space. */
394 		n--;
395 		n += scnprintf(msg + n, len - n, "]");
396 	}
397 }
398 
pci_get_dev_wrapper(int dom,unsigned int bus,unsigned int dev,unsigned int fun)399 static struct pci_dev *pci_get_dev_wrapper(int dom, unsigned int bus,
400 					   unsigned int dev, unsigned int fun)
401 {
402 	struct pci_dev *pdev;
403 
404 	pdev = pci_get_domain_bus_and_slot(dom, bus, PCI_DEVFN(dev, fun));
405 	if (!pdev) {
406 		edac_dbg(2, "No device %02x:%02x.%x\n",
407 			 bus, dev, fun);
408 		return NULL;
409 	}
410 
411 	if (unlikely(pci_enable_device(pdev) < 0)) {
412 		edac_dbg(2, "Failed to enable device %02x:%02x.%x\n",
413 			 bus, dev, fun);
414 		pci_dev_put(pdev);
415 		return NULL;
416 	}
417 
418 	return pdev;
419 }
420 
421 /**
422  * i10nm_get_imc_num() - Get the number of present DDR memory controllers.
423  *
424  * @cfg : The pointer to the structure of EDAC resource configurations.
425  *
426  * For Granite Rapids CPUs, the number of present DDR memory controllers read
427  * at runtime overwrites the value statically configured in @cfg->ddr_imc_num.
428  * For other CPUs, the number of present DDR memory controllers is statically
429  * configured in @cfg->ddr_imc_num.
430  *
431  * RETURNS : 0 on success, < 0 on failure.
432  */
i10nm_get_imc_num(struct res_config * cfg)433 static int i10nm_get_imc_num(struct res_config *cfg)
434 {
435 	int n, imc_num, chan_num = 0;
436 	struct skx_dev *d;
437 	u32 reg;
438 
439 	list_for_each_entry(d, i10nm_edac_list, list) {
440 		d->pcu_cr3 = pci_get_dev_wrapper(d->seg, d->bus[res_cfg->pcu_cr3_bdf.bus],
441 						 res_cfg->pcu_cr3_bdf.dev,
442 						 res_cfg->pcu_cr3_bdf.fun);
443 		if (!d->pcu_cr3)
444 			continue;
445 
446 		if (I10NM_GET_CAPID5_CFG(d, reg))
447 			continue;
448 
449 		n = I10NM_DDR_IMC_CH_CNT(reg);
450 
451 		if (!chan_num) {
452 			chan_num = n;
453 			edac_dbg(2, "Get DDR CH number: %d\n", chan_num);
454 		} else if (chan_num != n) {
455 			i10nm_printk(KERN_NOTICE, "Get DDR CH numbers: %d, %d\n", chan_num, n);
456 		}
457 	}
458 
459 	switch (cfg->type) {
460 	case GNR:
461 		/*
462 		 * One channel per DDR memory controller for Granite Rapids CPUs.
463 		 */
464 		imc_num = chan_num;
465 
466 		if (!imc_num) {
467 			i10nm_printk(KERN_ERR, "Invalid DDR MC number\n");
468 			return -ENODEV;
469 		}
470 
471 		if (cfg->ddr_imc_num != imc_num) {
472 			/*
473 			 * Update the configuration data to reflect the number of
474 			 * present DDR memory controllers.
475 			 */
476 			cfg->ddr_imc_num = imc_num;
477 			edac_dbg(2, "Set DDR MC number: %d", imc_num);
478 
479 			/* Release and reallocate skx_dev list with the updated number. */
480 			skx_remove();
481 			if (skx_get_all_bus_mappings(cfg, &i10nm_edac_list) <= 0)
482 				return -ENODEV;
483 		}
484 
485 		return 0;
486 	default:
487 		/*
488 		 * For other CPUs, the number of present DDR memory controllers
489 		 * is statically pre-configured in cfg->ddr_imc_num.
490 		 */
491 		return 0;
492 	}
493 }
494 
i10nm_check_2lm(struct res_config * cfg)495 static bool i10nm_check_2lm(struct res_config *cfg)
496 {
497 	struct skx_dev *d;
498 	u32 reg;
499 	int i;
500 
501 	list_for_each_entry(d, i10nm_edac_list, list) {
502 		d->sad_all = pci_get_dev_wrapper(d->seg, d->bus[res_cfg->sad_all_bdf.bus],
503 						 res_cfg->sad_all_bdf.dev,
504 						 res_cfg->sad_all_bdf.fun);
505 		if (!d->sad_all)
506 			continue;
507 
508 		for (i = 0; i < I10NM_MAX_SAD; i++) {
509 			I10NM_GET_SAD(d, cfg->sad_all_offset, i, reg);
510 			if (I10NM_SAD_ENABLE(reg) && I10NM_SAD_NM_CACHEABLE(reg)) {
511 				edac_dbg(2, "2-level memory configuration.\n");
512 				return true;
513 			}
514 		}
515 	}
516 
517 	return false;
518 }
519 
520 /*
521  * Check whether the error comes from DDRT by ICX/Tremont/SPR model specific error code.
522  * Refer to SDM vol3B 17.11.3/17.13.2 Intel IMC MC error codes for IA32_MCi_STATUS.
523  */
i10nm_mscod_is_ddrt(u32 mscod)524 static bool i10nm_mscod_is_ddrt(u32 mscod)
525 {
526 	switch (res_cfg->type) {
527 	case I10NM:
528 		switch (mscod) {
529 		case 0x0106: case 0x0107:
530 		case 0x0800: case 0x0804:
531 		case 0x0806 ... 0x0808:
532 		case 0x080a ... 0x080e:
533 		case 0x0810: case 0x0811:
534 		case 0x0816: case 0x081e:
535 		case 0x081f:
536 			return true;
537 		}
538 
539 		break;
540 	case SPR:
541 		switch (mscod) {
542 		case 0x0800: case 0x0804:
543 		case 0x0806 ... 0x0808:
544 		case 0x080a ... 0x080e:
545 		case 0x0810: case 0x0811:
546 		case 0x0816: case 0x081e:
547 		case 0x081f:
548 			return true;
549 		}
550 
551 		break;
552 	default:
553 		return false;
554 	}
555 
556 	return false;
557 }
558 
i10nm_mc_decode_available(struct mce * mce)559 static bool i10nm_mc_decode_available(struct mce *mce)
560 {
561 #define ICX_IMCx_CHy		0x06666000
562 	u8 bank;
563 
564 	if (!decoding_via_mca || mem_cfg_2lm)
565 		return false;
566 
567 	if ((mce->status & (MCI_STATUS_MISCV | MCI_STATUS_ADDRV))
568 			!= (MCI_STATUS_MISCV | MCI_STATUS_ADDRV))
569 		return false;
570 
571 	bank = mce->bank;
572 
573 	switch (res_cfg->type) {
574 	case I10NM:
575 		/* Check whether the bank is one of {13,14,17,18,21,22,25,26} */
576 		if (!(ICX_IMCx_CHy & (1 << bank)))
577 			return false;
578 		break;
579 	case SPR:
580 		if (bank < 13 || bank > 20)
581 			return false;
582 		break;
583 	default:
584 		return false;
585 	}
586 
587 	/* DDRT errors can't be decoded from MCA bank registers */
588 	if (MCI_MISC_ECC_MODE(mce->misc) == MCI_MISC_ECC_DDRT)
589 		return false;
590 
591 	if (i10nm_mscod_is_ddrt(MCI_STATUS_MSCOD(mce->status)))
592 		return false;
593 
594 	return true;
595 }
596 
i10nm_mc_decode(struct decoded_addr * res)597 static bool i10nm_mc_decode(struct decoded_addr *res)
598 {
599 	struct mce *m = res->mce;
600 	struct skx_dev *d;
601 	u8 bank;
602 
603 	if (!i10nm_mc_decode_available(m))
604 		return false;
605 
606 	list_for_each_entry(d, i10nm_edac_list, list) {
607 		if (d->imc[0].src_id == m->socketid) {
608 			res->socket = m->socketid;
609 			res->dev = d;
610 			break;
611 		}
612 	}
613 
614 	switch (res_cfg->type) {
615 	case I10NM:
616 		bank              = m->bank - 13;
617 		res->imc          = bank / 4;
618 		res->channel      = bank % 2;
619 		res->column       = GET_BITFIELD(m->misc, 9, 18) << 2;
620 		res->row          = GET_BITFIELD(m->misc, 19, 39);
621 		res->bank_group   = GET_BITFIELD(m->misc, 40, 41);
622 		res->bank_address = GET_BITFIELD(m->misc, 42, 43);
623 		res->bank_group  |= GET_BITFIELD(m->misc, 44, 44) << 2;
624 		res->rank         = GET_BITFIELD(m->misc, 56, 58);
625 		res->dimm         = res->rank >> 2;
626 		res->rank         = res->rank % 4;
627 		break;
628 	case SPR:
629 		bank              = m->bank - 13;
630 		res->imc          = bank / 2;
631 		res->channel      = bank % 2;
632 		res->column       = GET_BITFIELD(m->misc, 9, 18) << 2;
633 		res->row          = GET_BITFIELD(m->misc, 19, 36);
634 		res->bank_group   = GET_BITFIELD(m->misc, 37, 38);
635 		res->bank_address = GET_BITFIELD(m->misc, 39, 40);
636 		res->bank_group  |= GET_BITFIELD(m->misc, 41, 41) << 2;
637 		res->rank         = GET_BITFIELD(m->misc, 57, 57);
638 		res->dimm         = GET_BITFIELD(m->misc, 58, 58);
639 		break;
640 	default:
641 		return false;
642 	}
643 
644 	if (!res->dev) {
645 		skx_printk(KERN_ERR, "No device for src_id %d imc %d\n",
646 			   m->socketid, res->imc);
647 		return false;
648 	}
649 
650 	return true;
651 }
652 
653 /**
654  * get_gnr_mdev() - Get the PCI device of the @logical_idx-th DDR memory controller.
655  *
656  * @d            : The pointer to the structure of CPU socket EDAC device.
657  * @logical_idx  : The logical index of the present memory controller (0 ~ max present MC# - 1).
658  * @physical_idx : To store the corresponding physical index of @logical_idx.
659  *
660  * RETURNS       : The PCI device of the @logical_idx-th DDR memory controller, NULL on failure.
661  */
get_gnr_mdev(struct skx_dev * d,int logical_idx,int * physical_idx)662 static struct pci_dev *get_gnr_mdev(struct skx_dev *d, int logical_idx, int *physical_idx)
663 {
664 #define GNR_MAX_IMC_PCI_CNT	28
665 
666 	struct pci_dev *mdev;
667 	int i, logical = 0;
668 
669 	/*
670 	 * Detect present memory controllers from { PCI device: 8-5, function 7-1 }
671 	 */
672 	for (i = 0; i < GNR_MAX_IMC_PCI_CNT; i++) {
673 		mdev = pci_get_dev_wrapper(d->seg,
674 					   d->bus[res_cfg->ddr_mdev_bdf.bus],
675 					   res_cfg->ddr_mdev_bdf.dev + i / 7,
676 					   res_cfg->ddr_mdev_bdf.fun + i % 7);
677 
678 		if (mdev) {
679 			if (logical == logical_idx) {
680 				*physical_idx = i;
681 				return mdev;
682 			}
683 
684 			pci_dev_put(mdev);
685 			logical++;
686 		}
687 	}
688 
689 	return NULL;
690 }
691 
get_gnr_imc_mmio_offset(void)692 static u32 get_gnr_imc_mmio_offset(void)
693 {
694 	if (boot_cpu_data.x86_vfm == INTEL_GRANITERAPIDS_D)
695 		return I10NM_GNR_D_IMC_MMIO_OFFSET;
696 
697 	return I10NM_GNR_IMC_MMIO_OFFSET;
698 }
699 
700 /**
701  * get_ddr_munit() - Get the resource of the i-th DDR memory controller.
702  *
703  * @d      : The pointer to the structure of CPU socket EDAC device.
704  * @i      : The index of the CPU socket relative DDR memory controller.
705  * @offset : To store the MMIO offset of the i-th DDR memory controller.
706  * @size   : To store the MMIO size of the i-th DDR memory controller.
707  *
708  * RETURNS : The PCI device of the i-th DDR memory controller, NULL on failure.
709  */
get_ddr_munit(struct skx_dev * d,int i,u32 * offset,unsigned long * size)710 static struct pci_dev *get_ddr_munit(struct skx_dev *d, int i, u32 *offset, unsigned long *size)
711 {
712 	struct pci_dev *mdev;
713 	int physical_idx;
714 	u32 reg;
715 
716 	switch (res_cfg->type) {
717 	case GNR:
718 		if (I10NM_GET_IMC_BAR(d, 0, reg)) {
719 			i10nm_printk(KERN_ERR, "Failed to get mc0 bar\n");
720 			return NULL;
721 		}
722 
723 		mdev = get_gnr_mdev(d, i, &physical_idx);
724 		if (!mdev)
725 			return NULL;
726 
727 		*offset = I10NM_GET_IMC_MMIO_OFFSET(reg) +
728 			  get_gnr_imc_mmio_offset() +
729 			  physical_idx * I10NM_GNR_IMC_MMIO_SIZE;
730 		*size   = I10NM_GNR_IMC_MMIO_SIZE;
731 
732 		break;
733 	default:
734 		if (I10NM_GET_IMC_BAR(d, i, reg)) {
735 			i10nm_printk(KERN_ERR, "Failed to get mc%d bar\n", i);
736 			return NULL;
737 		}
738 
739 		mdev = pci_get_dev_wrapper(d->seg,
740 					   d->bus[res_cfg->ddr_mdev_bdf.bus],
741 					   res_cfg->ddr_mdev_bdf.dev + i,
742 					   res_cfg->ddr_mdev_bdf.fun);
743 		if (!mdev)
744 			return NULL;
745 
746 		*offset  = I10NM_GET_IMC_MMIO_OFFSET(reg);
747 		*size    = I10NM_GET_IMC_MMIO_SIZE(reg);
748 	}
749 
750 	return mdev;
751 }
752 
753 /**
754  * i10nm_imc_absent() - Check whether the memory controller @imc is absent
755  *
756  * @imc    : The pointer to the structure of memory controller EDAC device.
757  *
758  * RETURNS : true if the memory controller EDAC device is absent, false otherwise.
759  */
i10nm_imc_absent(struct skx_imc * imc)760 static bool i10nm_imc_absent(struct skx_imc *imc)
761 {
762 	u32 mcmtr;
763 	int i;
764 
765 	switch (res_cfg->type) {
766 	case SPR:
767 		for (i = 0; i < res_cfg->ddr_chan_num; i++) {
768 			mcmtr = I10NM_GET_MCMTR(imc, i);
769 			edac_dbg(1, "ch%d mcmtr reg %x\n", i, mcmtr);
770 			if (mcmtr != ~0)
771 				return false;
772 		}
773 
774 		/*
775 		 * Some workstations' absent memory controllers still
776 		 * appear as PCIe devices, misleading the EDAC driver.
777 		 * By observing that the MMIO registers of these absent
778 		 * memory controllers consistently hold the value of ~0.
779 		 *
780 		 * We identify a memory controller as absent by checking
781 		 * if its MMIO register "mcmtr" == ~0 in all its channels.
782 		 */
783 		return true;
784 	default:
785 		return false;
786 	}
787 }
788 
i10nm_get_ddr_munits(void)789 static int i10nm_get_ddr_munits(void)
790 {
791 	struct pci_dev *mdev;
792 	void __iomem *mbase;
793 	unsigned long size;
794 	struct skx_dev *d;
795 	int i, lmc, j = 0;
796 	u32 reg, off;
797 	u64 base;
798 
799 	list_for_each_entry(d, i10nm_edac_list, list) {
800 		d->util_all = pci_get_dev_wrapper(d->seg, d->bus[res_cfg->util_all_bdf.bus],
801 						  res_cfg->util_all_bdf.dev,
802 						  res_cfg->util_all_bdf.fun);
803 		if (!d->util_all)
804 			return -ENODEV;
805 
806 		d->uracu = pci_get_dev_wrapper(d->seg, d->bus[res_cfg->uracu_bdf.bus],
807 					       res_cfg->uracu_bdf.dev,
808 					       res_cfg->uracu_bdf.fun);
809 		if (!d->uracu)
810 			return -ENODEV;
811 
812 		if (I10NM_GET_SCK_BAR(d, reg)) {
813 			i10nm_printk(KERN_ERR, "Failed to socket bar\n");
814 			return -ENODEV;
815 		}
816 
817 		base = I10NM_GET_SCK_MMIO_BASE(reg);
818 		edac_dbg(2, "socket%d mmio base 0x%llx (reg 0x%x)\n",
819 			 j++, base, reg);
820 
821 		for (lmc = 0, i = 0; i < res_cfg->ddr_imc_num; i++) {
822 			mdev = get_ddr_munit(d, i, &off, &size);
823 
824 			if (i == 0 && !mdev) {
825 				i10nm_printk(KERN_ERR, "No IMC found\n");
826 				return -ENODEV;
827 			}
828 			if (!mdev)
829 				continue;
830 
831 			edac_dbg(2, "mc%d mmio base 0x%llx size 0x%lx (reg 0x%x)\n",
832 				 i, base + off, size, reg);
833 
834 			mbase = ioremap(base + off, size);
835 			if (!mbase) {
836 				i10nm_printk(KERN_ERR, "Failed to ioremap 0x%llx\n",
837 					     base + off);
838 				return -ENODEV;
839 			}
840 
841 			d->imc[lmc].mbase = mbase;
842 			if (i10nm_imc_absent(&d->imc[lmc])) {
843 				pci_dev_put(mdev);
844 				iounmap(mbase);
845 				d->imc[lmc].mbase = NULL;
846 				edac_dbg(2, "Skip absent mc%d\n", i);
847 				continue;
848 			} else {
849 				d->imc[lmc].mdev = mdev;
850 				if (res_cfg->type == SPR)
851 					skx_set_mc_mapping(d, i, lmc);
852 				lmc++;
853 			}
854 		}
855 	}
856 
857 	return 0;
858 }
859 
i10nm_check_hbm_imc(struct skx_dev * d)860 static bool i10nm_check_hbm_imc(struct skx_dev *d)
861 {
862 	u32 reg;
863 
864 	if (I10NM_GET_CAPID3_CFG(d, reg)) {
865 		i10nm_printk(KERN_ERR, "Failed to get capid3_cfg\n");
866 		return false;
867 	}
868 
869 	return I10NM_IS_HBM_PRESENT(reg) != 0;
870 }
871 
i10nm_get_hbm_munits(void)872 static int i10nm_get_hbm_munits(void)
873 {
874 	struct pci_dev *mdev;
875 	void __iomem *mbase;
876 	u32 reg, off, mcmtr;
877 	struct skx_dev *d;
878 	int i, lmc;
879 	u64 base;
880 
881 	list_for_each_entry(d, i10nm_edac_list, list) {
882 		if (!d->pcu_cr3)
883 			return -ENODEV;
884 
885 		if (!i10nm_check_hbm_imc(d)) {
886 			i10nm_printk(KERN_DEBUG, "No hbm memory\n");
887 			return -ENODEV;
888 		}
889 
890 		if (I10NM_GET_SCK_BAR(d, reg)) {
891 			i10nm_printk(KERN_ERR, "Failed to get socket bar\n");
892 			return -ENODEV;
893 		}
894 		base = I10NM_GET_SCK_MMIO_BASE(reg);
895 
896 		if (I10NM_GET_HBM_IMC_BAR(d, reg)) {
897 			i10nm_printk(KERN_ERR, "Failed to get hbm mc bar\n");
898 			return -ENODEV;
899 		}
900 		base += I10NM_GET_HBM_IMC_MMIO_OFFSET(reg);
901 
902 		lmc = res_cfg->ddr_imc_num;
903 
904 		for (i = 0; i < res_cfg->hbm_imc_num; i++) {
905 			mdev = pci_get_dev_wrapper(d->seg, d->bus[res_cfg->hbm_mdev_bdf.bus],
906 						   res_cfg->hbm_mdev_bdf.dev + i / 4,
907 						   res_cfg->hbm_mdev_bdf.fun + i % 4);
908 
909 			if (i == 0 && !mdev) {
910 				i10nm_printk(KERN_ERR, "No hbm mc found\n");
911 				return -ENODEV;
912 			}
913 			if (!mdev)
914 				continue;
915 
916 			d->imc[lmc].mdev = mdev;
917 			off = i * I10NM_HBM_IMC_MMIO_SIZE;
918 
919 			edac_dbg(2, "hbm mc%d mmio base 0x%llx size 0x%x\n",
920 				 lmc, base + off, I10NM_HBM_IMC_MMIO_SIZE);
921 
922 			mbase = ioremap(base + off, I10NM_HBM_IMC_MMIO_SIZE);
923 			if (!mbase) {
924 				pci_dev_put(d->imc[lmc].mdev);
925 				d->imc[lmc].mdev = NULL;
926 
927 				i10nm_printk(KERN_ERR, "Failed to ioremap for hbm mc 0x%llx\n",
928 					     base + off);
929 				return -ENOMEM;
930 			}
931 
932 			d->imc[lmc].mbase = mbase;
933 			d->imc[lmc].hbm_mc = true;
934 
935 			mcmtr = I10NM_GET_MCMTR(&d->imc[lmc], 0);
936 			if (!I10NM_IS_HBM_IMC(mcmtr)) {
937 				iounmap(d->imc[lmc].mbase);
938 				d->imc[lmc].mbase = NULL;
939 				d->imc[lmc].hbm_mc = false;
940 				pci_dev_put(d->imc[lmc].mdev);
941 				d->imc[lmc].mdev = NULL;
942 
943 				i10nm_printk(KERN_ERR, "This isn't an hbm mc!\n");
944 				return -ENODEV;
945 			}
946 
947 			lmc++;
948 		}
949 	}
950 
951 	return 0;
952 }
953 
954 static struct res_config i10nm_cfg0 = {
955 	.type			= I10NM,
956 	.decs_did		= 0x3452,
957 	.busno_cfg_offset	= 0xcc,
958 	.ddr_imc_num		= 4,
959 	.ddr_chan_num		= 2,
960 	.ddr_dimm_num		= 2,
961 	.ddr_chan_mmio_sz	= 0x4000,
962 	.sad_all_bdf		= {1, 29, 0},
963 	.pcu_cr3_bdf		= {1, 30, 3},
964 	.util_all_bdf		= {1, 29, 1},
965 	.uracu_bdf		= {0, 0, 1},
966 	.ddr_mdev_bdf		= {0, 12, 0},
967 	.hbm_mdev_bdf		= {0, 12, 1},
968 	.sad_all_offset		= 0x108,
969 	.reg_rrl_ddr		= &icx_reg_rrl_ddr,
970 };
971 
972 static struct res_config i10nm_cfg1 = {
973 	.type			= I10NM,
974 	.decs_did		= 0x3452,
975 	.busno_cfg_offset	= 0xd0,
976 	.ddr_imc_num		= 4,
977 	.ddr_chan_num		= 2,
978 	.ddr_dimm_num		= 2,
979 	.ddr_chan_mmio_sz	= 0x4000,
980 	.sad_all_bdf		= {1, 29, 0},
981 	.pcu_cr3_bdf		= {1, 30, 3},
982 	.util_all_bdf		= {1, 29, 1},
983 	.uracu_bdf		= {0, 0, 1},
984 	.ddr_mdev_bdf		= {0, 12, 0},
985 	.hbm_mdev_bdf		= {0, 12, 1},
986 	.sad_all_offset		= 0x108,
987 	.reg_rrl_ddr		= &icx_reg_rrl_ddr,
988 };
989 
990 static struct res_config spr_cfg = {
991 	.type			= SPR,
992 	.decs_did		= 0x3252,
993 	.busno_cfg_offset	= 0xd0,
994 	.ddr_imc_num		= 4,
995 	.ddr_chan_num		= 2,
996 	.ddr_dimm_num		= 2,
997 	.hbm_imc_num		= 16,
998 	.hbm_chan_num		= 2,
999 	.hbm_dimm_num		= 1,
1000 	.ddr_chan_mmio_sz	= 0x8000,
1001 	.hbm_chan_mmio_sz	= 0x4000,
1002 	.support_ddr5		= true,
1003 	.sad_all_bdf		= {1, 10, 0},
1004 	.pcu_cr3_bdf		= {1, 30, 3},
1005 	.util_all_bdf		= {1, 29, 1},
1006 	.uracu_bdf		= {0, 0, 1},
1007 	.ddr_mdev_bdf		= {0, 12, 0},
1008 	.hbm_mdev_bdf		= {0, 12, 1},
1009 	.sad_all_offset		= 0x300,
1010 	.reg_rrl_ddr		= &spr_reg_rrl_ddr,
1011 	.reg_rrl_hbm[0]		= &spr_reg_rrl_hbm_pch0,
1012 	.reg_rrl_hbm[1]		= &spr_reg_rrl_hbm_pch1,
1013 };
1014 
1015 static struct res_config gnr_cfg = {
1016 	.type			= GNR,
1017 	.decs_did		= 0x3252,
1018 	.busno_cfg_offset	= 0xd0,
1019 	.ddr_imc_num		= 12,
1020 	.ddr_chan_num		= 1,
1021 	.ddr_dimm_num		= 2,
1022 	.ddr_chan_mmio_sz	= 0x4000,
1023 	.support_ddr5		= true,
1024 	.sad_all_bdf		= {0, 13, 0},
1025 	.pcu_cr3_bdf		= {0, 5, 0},
1026 	.util_all_bdf		= {0, 13, 1},
1027 	.uracu_bdf		= {0, 0, 1},
1028 	.ddr_mdev_bdf		= {0, 5, 1},
1029 	.sad_all_offset		= 0x300,
1030 	.reg_rrl_ddr		= &gnr_reg_rrl_ddr,
1031 };
1032 
1033 static const struct x86_cpu_id i10nm_cpuids[] = {
1034 	X86_MATCH_VFM_STEPS(INTEL_ATOM_TREMONT_D, X86_STEP_MIN,		 0x3, &i10nm_cfg0),
1035 	X86_MATCH_VFM_STEPS(INTEL_ATOM_TREMONT_D,	   0x4,	X86_STEP_MAX, &i10nm_cfg1),
1036 	X86_MATCH_VFM_STEPS(INTEL_ICELAKE_X,	  X86_STEP_MIN,		 0x3, &i10nm_cfg0),
1037 	X86_MATCH_VFM_STEPS(INTEL_ICELAKE_X,		   0x4, X86_STEP_MAX, &i10nm_cfg1),
1038 	X86_MATCH_VFM(	    INTEL_ICELAKE_D,				      &i10nm_cfg1),
1039 
1040 	X86_MATCH_VFM(INTEL_SAPPHIRERAPIDS_X, &spr_cfg),
1041 	X86_MATCH_VFM(INTEL_EMERALDRAPIDS_X,  &spr_cfg),
1042 	X86_MATCH_VFM(INTEL_GRANITERAPIDS_X,  &gnr_cfg),
1043 	X86_MATCH_VFM(INTEL_GRANITERAPIDS_D,  &gnr_cfg),
1044 	X86_MATCH_VFM(INTEL_ATOM_CRESTMONT_X, &gnr_cfg),
1045 	X86_MATCH_VFM(INTEL_ATOM_CRESTMONT,   &gnr_cfg),
1046 	X86_MATCH_VFM(INTEL_ATOM_DARKMONT_X,  &gnr_cfg),
1047 	{}
1048 };
1049 MODULE_DEVICE_TABLE(x86cpu, i10nm_cpuids);
1050 
i10nm_check_ecc(struct skx_imc * imc,int chan)1051 static bool i10nm_check_ecc(struct skx_imc *imc, int chan)
1052 {
1053 	u32 mcmtr;
1054 
1055 	mcmtr = I10NM_GET_MCMTR(imc, chan);
1056 	edac_dbg(1, "ch%d mcmtr reg %x\n", chan, mcmtr);
1057 
1058 	return !!GET_BITFIELD(mcmtr, 2, 2);
1059 }
1060 
i10nm_channel_disabled(struct skx_imc * imc,int chan)1061 static bool i10nm_channel_disabled(struct skx_imc *imc, int chan)
1062 {
1063 	u32 mcmtr = I10NM_GET_MCMTR(imc, chan);
1064 
1065 	edac_dbg(1, "mc%d ch%d mcmtr reg %x\n", imc->mc, chan, mcmtr);
1066 
1067 	return (mcmtr == ~0 || GET_BITFIELD(mcmtr, 18, 18));
1068 }
1069 
i10nm_get_dimm_config(struct mem_ctl_info * mci,struct res_config * cfg)1070 static int i10nm_get_dimm_config(struct mem_ctl_info *mci,
1071 				 struct res_config *cfg)
1072 {
1073 	struct skx_pvt *pvt = mci->pvt_info;
1074 	struct skx_imc *imc = pvt->imc;
1075 	u32 mtr, mcddrtcfg = 0;
1076 	struct dimm_info *dimm;
1077 	int i, j, ndimms;
1078 
1079 	for (i = 0; i < imc->num_channels; i++) {
1080 		if (!imc->mbase)
1081 			continue;
1082 
1083 		if (i10nm_channel_disabled(imc, i)) {
1084 			edac_dbg(1, "mc%d ch%d is disabled.\n", imc->mc, i);
1085 			continue;
1086 		}
1087 
1088 		ndimms = 0;
1089 
1090 		if (res_cfg->type != GNR)
1091 			mcddrtcfg = I10NM_GET_MCDDRTCFG(imc, i);
1092 
1093 		for (j = 0; j < imc->num_dimms; j++) {
1094 			dimm = edac_get_dimm(mci, i, j, 0);
1095 			mtr = I10NM_GET_DIMMMTR(imc, i, j);
1096 			edac_dbg(1, "dimmmtr 0x%x mcddrtcfg 0x%x (mc%d ch%d dimm%d)\n",
1097 				 mtr, mcddrtcfg, imc->mc, i, j);
1098 
1099 			if (IS_DIMM_PRESENT(mtr))
1100 				ndimms += skx_get_dimm_info(mtr, 0, 0, dimm,
1101 							    imc, i, j, cfg);
1102 			else if (IS_NVDIMM_PRESENT(mcddrtcfg, j))
1103 				ndimms += skx_get_nvdimm_info(dimm, imc, i, j,
1104 							      EDAC_MOD_STR);
1105 		}
1106 		if (ndimms && !i10nm_check_ecc(imc, i)) {
1107 			i10nm_printk(KERN_ERR, "ECC is disabled on imc %d channel %d\n",
1108 				     imc->mc, i);
1109 			return -ENODEV;
1110 		}
1111 	}
1112 
1113 	return 0;
1114 }
1115 
1116 static struct notifier_block i10nm_mce_dec = {
1117 	.notifier_call	= skx_mce_check_error,
1118 	.priority	= MCE_PRIO_EDAC,
1119 };
1120 
i10nm_init(void)1121 static int __init i10nm_init(void)
1122 {
1123 	u8 mc = 0, src_id = 0;
1124 	const struct x86_cpu_id *id;
1125 	struct res_config *cfg;
1126 	const char *owner;
1127 	struct skx_dev *d;
1128 	int rc, i, off[3] = {0xd0, 0xc8, 0xcc};
1129 	u64 tolm, tohm;
1130 	int imc_num;
1131 
1132 	edac_dbg(2, "\n");
1133 
1134 	if (ghes_get_devices())
1135 		return -EBUSY;
1136 
1137 	owner = edac_get_owner();
1138 	if (owner && strncmp(owner, EDAC_MOD_STR, sizeof(EDAC_MOD_STR)))
1139 		return -EBUSY;
1140 
1141 	if (cpu_feature_enabled(X86_FEATURE_HYPERVISOR))
1142 		return -ENODEV;
1143 
1144 	id = x86_match_cpu(i10nm_cpuids);
1145 	if (!id)
1146 		return -ENODEV;
1147 
1148 	cfg = (struct res_config *)id->driver_data;
1149 	skx_set_res_cfg(cfg);
1150 	res_cfg = cfg;
1151 
1152 	rc = skx_get_hi_lo(0x09a2, off, &tolm, &tohm);
1153 	if (rc)
1154 		return rc;
1155 
1156 	rc = skx_get_all_bus_mappings(cfg, &i10nm_edac_list);
1157 	if (rc < 0)
1158 		goto fail;
1159 	if (rc == 0) {
1160 		i10nm_printk(KERN_ERR, "No memory controllers found\n");
1161 		return -ENODEV;
1162 	}
1163 
1164 	rc = i10nm_get_imc_num(cfg);
1165 	if (rc < 0)
1166 		goto fail;
1167 
1168 	mem_cfg_2lm = i10nm_check_2lm(cfg);
1169 	skx_set_mem_cfg(mem_cfg_2lm);
1170 
1171 	rc = i10nm_get_ddr_munits();
1172 
1173 	if (i10nm_get_hbm_munits() && rc)
1174 		goto fail;
1175 
1176 	imc_num = res_cfg->ddr_imc_num + res_cfg->hbm_imc_num;
1177 
1178 	list_for_each_entry(d, i10nm_edac_list, list) {
1179 		rc = skx_get_src_id(d, 0xf8, &src_id);
1180 		if (rc < 0)
1181 			goto fail;
1182 
1183 		edac_dbg(2, "src_id = %d\n", src_id);
1184 		for (i = 0; i < imc_num; i++) {
1185 			if (!d->imc[i].mdev)
1186 				continue;
1187 
1188 			d->imc[i].mc  = mc++;
1189 			d->imc[i].lmc = i;
1190 			d->imc[i].src_id = src_id;
1191 			if (d->imc[i].hbm_mc) {
1192 				d->imc[i].chan_mmio_sz = cfg->hbm_chan_mmio_sz;
1193 				d->imc[i].num_channels = cfg->hbm_chan_num;
1194 				d->imc[i].num_dimms    = cfg->hbm_dimm_num;
1195 			} else {
1196 				d->imc[i].chan_mmio_sz = cfg->ddr_chan_mmio_sz;
1197 				d->imc[i].num_channels = cfg->ddr_chan_num;
1198 				d->imc[i].num_dimms    = cfg->ddr_dimm_num;
1199 			}
1200 
1201 			rc = skx_register_mci(&d->imc[i], d->imc[i].mdev,
1202 					      "Intel_10nm Socket", EDAC_MOD_STR,
1203 					      i10nm_get_dimm_config, cfg);
1204 			if (rc < 0)
1205 				goto fail;
1206 		}
1207 	}
1208 
1209 	rc = skx_adxl_get();
1210 	if (rc)
1211 		goto fail;
1212 
1213 	opstate_init();
1214 	mce_register_decode_chain(&i10nm_mce_dec);
1215 	skx_setup_debug("i10nm_test");
1216 
1217 	if (retry_rd_err_log && res_cfg->reg_rrl_ddr) {
1218 		skx_set_decode(i10nm_mc_decode, show_retry_rd_err_log);
1219 		if (retry_rd_err_log == 2)
1220 			enable_retry_rd_err_log(true);
1221 	} else {
1222 		skx_set_decode(i10nm_mc_decode, NULL);
1223 	}
1224 
1225 	i10nm_printk(KERN_INFO, "%s\n", I10NM_REVISION);
1226 
1227 	return 0;
1228 fail:
1229 	skx_remove();
1230 	return rc;
1231 }
1232 
i10nm_exit(void)1233 static void __exit i10nm_exit(void)
1234 {
1235 	edac_dbg(2, "\n");
1236 
1237 	if (retry_rd_err_log && res_cfg->reg_rrl_ddr) {
1238 		skx_set_decode(NULL, NULL);
1239 		if (retry_rd_err_log == 2)
1240 			enable_retry_rd_err_log(false);
1241 	}
1242 
1243 	skx_teardown_debug();
1244 	mce_unregister_decode_chain(&i10nm_mce_dec);
1245 	skx_adxl_put();
1246 	skx_remove();
1247 }
1248 
1249 module_init(i10nm_init);
1250 module_exit(i10nm_exit);
1251 
set_decoding_via_mca(const char * buf,const struct kernel_param * kp)1252 static int set_decoding_via_mca(const char *buf, const struct kernel_param *kp)
1253 {
1254 	unsigned long val;
1255 	int ret;
1256 
1257 	ret = kstrtoul(buf, 0, &val);
1258 
1259 	if (ret || val > 1)
1260 		return -EINVAL;
1261 
1262 	if (val && mem_cfg_2lm) {
1263 		i10nm_printk(KERN_NOTICE, "Decoding errors via MCA banks for 2LM isn't supported yet\n");
1264 		return -EIO;
1265 	}
1266 
1267 	ret = param_set_int(buf, kp);
1268 
1269 	return ret;
1270 }
1271 
1272 static const struct kernel_param_ops decoding_via_mca_param_ops = {
1273 	.set = set_decoding_via_mca,
1274 	.get = param_get_int,
1275 };
1276 
1277 module_param_cb(decoding_via_mca, &decoding_via_mca_param_ops, &decoding_via_mca, 0644);
1278 MODULE_PARM_DESC(decoding_via_mca, "decoding_via_mca: 0=off(default), 1=enable");
1279 
1280 module_param(retry_rd_err_log, int, 0444);
1281 MODULE_PARM_DESC(retry_rd_err_log, "retry_rd_err_log: 0=off(default), 1=bios(Linux doesn't reset any control bits, but just reports values.), 2=linux(Linux tries to take control and resets mode bits, clear valid/UC bits after reading.)");
1282 
1283 MODULE_LICENSE("GPL v2");
1284 MODULE_DESCRIPTION("MC Driver for Intel 10nm server processors");
1285