xref: /linux/drivers/mtd/nand/raw/pl35x-nand-controller.c (revision 55d0969c451159cff86949b38c39171cab962069)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * ARM PL35X NAND flash controller driver
4  *
5  * Copyright (C) 2017 Xilinx, Inc
6  * Author:
7  *   Miquel Raynal <miquel.raynal@bootlin.com>
8  * Original work (rewritten):
9  *   Punnaiah Choudary Kalluri <punnaia@xilinx.com>
10  *   Naga Sureshkumar Relli <nagasure@xilinx.com>
11  */
12 
13 #include <linux/amba/bus.h>
14 #include <linux/err.h>
15 #include <linux/delay.h>
16 #include <linux/interrupt.h>
17 #include <linux/io.h>
18 #include <linux/ioport.h>
19 #include <linux/iopoll.h>
20 #include <linux/irq.h>
21 #include <linux/module.h>
22 #include <linux/moduleparam.h>
23 #include <linux/mtd/mtd.h>
24 #include <linux/mtd/rawnand.h>
25 #include <linux/mtd/partitions.h>
26 #include <linux/of.h>
27 #include <linux/platform_device.h>
28 #include <linux/slab.h>
29 #include <linux/clk.h>
30 
31 #define PL35X_NANDC_DRIVER_NAME "pl35x-nand-controller"
32 
33 /* SMC controller status register (RO) */
34 #define PL35X_SMC_MEMC_STATUS 0x0
35 #define   PL35X_SMC_MEMC_STATUS_RAW_INT_STATUS1	BIT(6)
36 /* SMC clear config register (WO) */
37 #define PL35X_SMC_MEMC_CFG_CLR 0xC
38 #define   PL35X_SMC_MEMC_CFG_CLR_INT_DIS_1	BIT(1)
39 #define   PL35X_SMC_MEMC_CFG_CLR_INT_CLR_1	BIT(4)
40 #define   PL35X_SMC_MEMC_CFG_CLR_ECC_INT_DIS_1	BIT(6)
41 /* SMC direct command register (WO) */
42 #define PL35X_SMC_DIRECT_CMD 0x10
43 #define   PL35X_SMC_DIRECT_CMD_NAND_CS (0x4 << 23)
44 #define   PL35X_SMC_DIRECT_CMD_UPD_REGS (0x2 << 21)
45 /* SMC set cycles register (WO) */
46 #define PL35X_SMC_CYCLES 0x14
47 #define   PL35X_SMC_NAND_TRC_CYCLES(x) ((x) << 0)
48 #define   PL35X_SMC_NAND_TWC_CYCLES(x) ((x) << 4)
49 #define   PL35X_SMC_NAND_TREA_CYCLES(x) ((x) << 8)
50 #define   PL35X_SMC_NAND_TWP_CYCLES(x) ((x) << 11)
51 #define   PL35X_SMC_NAND_TCLR_CYCLES(x) ((x) << 14)
52 #define   PL35X_SMC_NAND_TAR_CYCLES(x) ((x) << 17)
53 #define   PL35X_SMC_NAND_TRR_CYCLES(x) ((x) << 20)
54 /* SMC set opmode register (WO) */
55 #define PL35X_SMC_OPMODE 0x18
56 #define   PL35X_SMC_OPMODE_BW_8 0
57 #define   PL35X_SMC_OPMODE_BW_16 1
58 /* SMC ECC status register (RO) */
59 #define PL35X_SMC_ECC_STATUS 0x400
60 #define   PL35X_SMC_ECC_STATUS_ECC_BUSY BIT(6)
61 /* SMC ECC configuration register */
62 #define PL35X_SMC_ECC_CFG 0x404
63 #define   PL35X_SMC_ECC_CFG_MODE_MASK 0xC
64 #define   PL35X_SMC_ECC_CFG_MODE_BYPASS 0
65 #define   PL35X_SMC_ECC_CFG_MODE_APB BIT(2)
66 #define   PL35X_SMC_ECC_CFG_MODE_MEM BIT(3)
67 #define   PL35X_SMC_ECC_CFG_PGSIZE_MASK	0x3
68 /* SMC ECC command 1 register */
69 #define PL35X_SMC_ECC_CMD1 0x408
70 #define   PL35X_SMC_ECC_CMD1_WRITE(x) ((x) << 0)
71 #define   PL35X_SMC_ECC_CMD1_READ(x) ((x) << 8)
72 #define   PL35X_SMC_ECC_CMD1_READ_END(x) ((x) << 16)
73 #define   PL35X_SMC_ECC_CMD1_READ_END_VALID(x) ((x) << 24)
74 /* SMC ECC command 2 register */
75 #define PL35X_SMC_ECC_CMD2 0x40C
76 #define   PL35X_SMC_ECC_CMD2_WRITE_COL_CHG(x) ((x) << 0)
77 #define   PL35X_SMC_ECC_CMD2_READ_COL_CHG(x) ((x) << 8)
78 #define   PL35X_SMC_ECC_CMD2_READ_COL_CHG_END(x) ((x) << 16)
79 #define   PL35X_SMC_ECC_CMD2_READ_COL_CHG_END_VALID(x) ((x) << 24)
80 /* SMC ECC value registers (RO) */
81 #define PL35X_SMC_ECC_VALUE(x) (0x418 + (4 * (x)))
82 #define   PL35X_SMC_ECC_VALUE_IS_CORRECTABLE(x) ((x) & BIT(27))
83 #define   PL35X_SMC_ECC_VALUE_HAS_FAILED(x) ((x) & BIT(28))
84 #define   PL35X_SMC_ECC_VALUE_IS_VALID(x) ((x) & BIT(30))
85 
86 /* NAND AXI interface */
87 #define PL35X_SMC_CMD_PHASE 0
88 #define PL35X_SMC_CMD_PHASE_CMD0(x) ((x) << 3)
89 #define PL35X_SMC_CMD_PHASE_CMD1(x) ((x) << 11)
90 #define PL35X_SMC_CMD_PHASE_CMD1_VALID BIT(20)
91 #define PL35X_SMC_CMD_PHASE_ADDR(pos, x) ((x) << (8 * (pos)))
92 #define PL35X_SMC_CMD_PHASE_NADDRS(x) ((x) << 21)
93 #define PL35X_SMC_DATA_PHASE BIT(19)
94 #define PL35X_SMC_DATA_PHASE_ECC_LAST BIT(10)
95 #define PL35X_SMC_DATA_PHASE_CLEAR_CS BIT(21)
96 
97 #define PL35X_NAND_MAX_CS 1
98 #define PL35X_NAND_LAST_XFER_SZ 4
99 #define TO_CYCLES(ps, period_ns) (DIV_ROUND_UP((ps) / 1000, period_ns))
100 
101 #define PL35X_NAND_ECC_BITS_MASK 0xFFF
102 #define PL35X_NAND_ECC_BYTE_OFF_MASK 0x1FF
103 #define PL35X_NAND_ECC_BIT_OFF_MASK 0x7
104 
105 struct pl35x_nand_timings {
106 	unsigned int t_rc:4;
107 	unsigned int t_wc:4;
108 	unsigned int t_rea:3;
109 	unsigned int t_wp:3;
110 	unsigned int t_clr:3;
111 	unsigned int t_ar:3;
112 	unsigned int t_rr:4;
113 	unsigned int rsvd:8;
114 };
115 
116 struct pl35x_nand {
117 	struct list_head node;
118 	struct nand_chip chip;
119 	unsigned int cs;
120 	unsigned int addr_cycles;
121 	u32 ecc_cfg;
122 	u32 timings;
123 };
124 
125 /**
126  * struct pl35x_nandc - NAND flash controller driver structure
127  * @dev: Kernel device
128  * @conf_regs: SMC configuration registers for command phase
129  * @io_regs: NAND data registers for data phase
130  * @controller: Core NAND controller structure
131  * @chips: List of connected NAND chips
132  * @selected_chip: NAND chip currently selected by the controller
133  * @assigned_cs: List of assigned CS
134  * @ecc_buf: Temporary buffer to extract ECC bytes
135  */
136 struct pl35x_nandc {
137 	struct device *dev;
138 	void __iomem *conf_regs;
139 	void __iomem *io_regs;
140 	struct nand_controller controller;
141 	struct list_head chips;
142 	struct nand_chip *selected_chip;
143 	unsigned long assigned_cs;
144 	u8 *ecc_buf;
145 };
146 
147 static inline struct pl35x_nandc *to_pl35x_nandc(struct nand_controller *ctrl)
148 {
149 	return container_of(ctrl, struct pl35x_nandc, controller);
150 }
151 
152 static inline struct pl35x_nand *to_pl35x_nand(struct nand_chip *chip)
153 {
154 	return container_of(chip, struct pl35x_nand, chip);
155 }
156 
157 static int pl35x_ecc_ooblayout16_ecc(struct mtd_info *mtd, int section,
158 				     struct mtd_oob_region *oobregion)
159 {
160 	struct nand_chip *chip = mtd_to_nand(mtd);
161 
162 	if (section >= chip->ecc.steps)
163 		return -ERANGE;
164 
165 	oobregion->offset = (section * chip->ecc.bytes);
166 	oobregion->length = chip->ecc.bytes;
167 
168 	return 0;
169 }
170 
171 static int pl35x_ecc_ooblayout16_free(struct mtd_info *mtd, int section,
172 				      struct mtd_oob_region *oobregion)
173 {
174 	struct nand_chip *chip = mtd_to_nand(mtd);
175 
176 	if (section >= chip->ecc.steps)
177 		return -ERANGE;
178 
179 	oobregion->offset = (section * chip->ecc.bytes) + 8;
180 	oobregion->length = 8;
181 
182 	return 0;
183 }
184 
185 static const struct mtd_ooblayout_ops pl35x_ecc_ooblayout16_ops = {
186 	.ecc = pl35x_ecc_ooblayout16_ecc,
187 	.free = pl35x_ecc_ooblayout16_free,
188 };
189 
190 /* Generic flash bbt decriptors */
191 static u8 bbt_pattern[] = { 'B', 'b', 't', '0' };
192 static u8 mirror_pattern[] = { '1', 't', 'b', 'B' };
193 
194 static struct nand_bbt_descr bbt_main_descr = {
195 	.options = NAND_BBT_LASTBLOCK | NAND_BBT_CREATE | NAND_BBT_WRITE
196 		| NAND_BBT_2BIT | NAND_BBT_VERSION | NAND_BBT_PERCHIP,
197 	.offs = 4,
198 	.len = 4,
199 	.veroffs = 20,
200 	.maxblocks = 4,
201 	.pattern = bbt_pattern
202 };
203 
204 static struct nand_bbt_descr bbt_mirror_descr = {
205 	.options = NAND_BBT_LASTBLOCK | NAND_BBT_CREATE | NAND_BBT_WRITE
206 		| NAND_BBT_2BIT | NAND_BBT_VERSION | NAND_BBT_PERCHIP,
207 	.offs = 4,
208 	.len = 4,
209 	.veroffs = 20,
210 	.maxblocks = 4,
211 	.pattern = mirror_pattern
212 };
213 
214 static void pl35x_smc_update_regs(struct pl35x_nandc *nfc)
215 {
216 	writel(PL35X_SMC_DIRECT_CMD_NAND_CS |
217 	       PL35X_SMC_DIRECT_CMD_UPD_REGS,
218 	       nfc->conf_regs + PL35X_SMC_DIRECT_CMD);
219 }
220 
221 static int pl35x_smc_set_buswidth(struct pl35x_nandc *nfc, unsigned int bw)
222 {
223 	if (bw != PL35X_SMC_OPMODE_BW_8 && bw != PL35X_SMC_OPMODE_BW_16)
224 		return -EINVAL;
225 
226 	writel(bw, nfc->conf_regs + PL35X_SMC_OPMODE);
227 	pl35x_smc_update_regs(nfc);
228 
229 	return 0;
230 }
231 
232 static void pl35x_smc_clear_irq(struct pl35x_nandc *nfc)
233 {
234 	writel(PL35X_SMC_MEMC_CFG_CLR_INT_CLR_1,
235 	       nfc->conf_regs + PL35X_SMC_MEMC_CFG_CLR);
236 }
237 
238 static int pl35x_smc_wait_for_irq(struct pl35x_nandc *nfc)
239 {
240 	u32 reg;
241 	int ret;
242 
243 	ret = readl_poll_timeout(nfc->conf_regs + PL35X_SMC_MEMC_STATUS, reg,
244 				 reg & PL35X_SMC_MEMC_STATUS_RAW_INT_STATUS1,
245 				 10, 1000000);
246 	if (ret)
247 		dev_err(nfc->dev,
248 			"Timeout polling on NAND controller interrupt (0x%x)\n",
249 			reg);
250 
251 	pl35x_smc_clear_irq(nfc);
252 
253 	return ret;
254 }
255 
256 static int pl35x_smc_wait_for_ecc_done(struct pl35x_nandc *nfc)
257 {
258 	u32 reg;
259 	int ret;
260 
261 	ret = readl_poll_timeout(nfc->conf_regs + PL35X_SMC_ECC_STATUS, reg,
262 				 !(reg & PL35X_SMC_ECC_STATUS_ECC_BUSY),
263 				 10, 1000000);
264 	if (ret)
265 		dev_err(nfc->dev,
266 			"Timeout polling on ECC controller interrupt\n");
267 
268 	return ret;
269 }
270 
271 static int pl35x_smc_set_ecc_mode(struct pl35x_nandc *nfc,
272 				  struct nand_chip *chip,
273 				  unsigned int mode)
274 {
275 	struct pl35x_nand *plnand;
276 	u32 ecc_cfg;
277 
278 	ecc_cfg = readl(nfc->conf_regs + PL35X_SMC_ECC_CFG);
279 	ecc_cfg &= ~PL35X_SMC_ECC_CFG_MODE_MASK;
280 	ecc_cfg |= mode;
281 	writel(ecc_cfg, nfc->conf_regs + PL35X_SMC_ECC_CFG);
282 
283 	if (chip) {
284 		plnand = to_pl35x_nand(chip);
285 		plnand->ecc_cfg = ecc_cfg;
286 	}
287 
288 	if (mode != PL35X_SMC_ECC_CFG_MODE_BYPASS)
289 		return pl35x_smc_wait_for_ecc_done(nfc);
290 
291 	return 0;
292 }
293 
294 static void pl35x_smc_force_byte_access(struct nand_chip *chip,
295 					bool force_8bit)
296 {
297 	struct pl35x_nandc *nfc = to_pl35x_nandc(chip->controller);
298 	int ret;
299 
300 	if (!(chip->options & NAND_BUSWIDTH_16))
301 		return;
302 
303 	if (force_8bit)
304 		ret = pl35x_smc_set_buswidth(nfc, PL35X_SMC_OPMODE_BW_8);
305 	else
306 		ret = pl35x_smc_set_buswidth(nfc, PL35X_SMC_OPMODE_BW_16);
307 
308 	if (ret)
309 		dev_err(nfc->dev, "Error in Buswidth\n");
310 }
311 
312 static void pl35x_nand_select_target(struct nand_chip *chip,
313 				     unsigned int die_nr)
314 {
315 	struct pl35x_nandc *nfc = to_pl35x_nandc(chip->controller);
316 	struct pl35x_nand *plnand = to_pl35x_nand(chip);
317 
318 	if (chip == nfc->selected_chip)
319 		return;
320 
321 	/* Setup the timings */
322 	writel(plnand->timings, nfc->conf_regs + PL35X_SMC_CYCLES);
323 	pl35x_smc_update_regs(nfc);
324 
325 	/* Configure the ECC engine */
326 	writel(plnand->ecc_cfg, nfc->conf_regs + PL35X_SMC_ECC_CFG);
327 
328 	nfc->selected_chip = chip;
329 }
330 
331 static void pl35x_nand_read_data_op(struct nand_chip *chip, u8 *in,
332 				    unsigned int len, bool force_8bit,
333 				    unsigned int flags, unsigned int last_flags)
334 {
335 	struct pl35x_nandc *nfc = to_pl35x_nandc(chip->controller);
336 	unsigned int buf_end = len / 4;
337 	unsigned int in_start = round_down(len, 4);
338 	unsigned int data_phase_addr;
339 	u32 *buf32 = (u32 *)in;
340 	u8 *buf8 = (u8 *)in;
341 	int i;
342 
343 	if (force_8bit)
344 		pl35x_smc_force_byte_access(chip, true);
345 
346 	for (i = 0; i < buf_end; i++) {
347 		data_phase_addr = PL35X_SMC_DATA_PHASE + flags;
348 		if (i + 1 == buf_end)
349 			data_phase_addr = PL35X_SMC_DATA_PHASE + last_flags;
350 
351 		buf32[i] = readl(nfc->io_regs + data_phase_addr);
352 	}
353 
354 	/* No working extra flags on unaligned data accesses */
355 	for (i = in_start; i < len; i++)
356 		buf8[i] = readb(nfc->io_regs + PL35X_SMC_DATA_PHASE);
357 
358 	if (force_8bit)
359 		pl35x_smc_force_byte_access(chip, false);
360 }
361 
362 static void pl35x_nand_write_data_op(struct nand_chip *chip, const u8 *out,
363 				     int len, bool force_8bit,
364 				     unsigned int flags,
365 				     unsigned int last_flags)
366 {
367 	struct pl35x_nandc *nfc = to_pl35x_nandc(chip->controller);
368 	unsigned int buf_end = len / 4;
369 	unsigned int in_start = round_down(len, 4);
370 	const u32 *buf32 = (const u32 *)out;
371 	const u8 *buf8 = (const u8 *)out;
372 	unsigned int data_phase_addr;
373 	int i;
374 
375 	if (force_8bit)
376 		pl35x_smc_force_byte_access(chip, true);
377 
378 	for (i = 0; i < buf_end; i++) {
379 		data_phase_addr = PL35X_SMC_DATA_PHASE + flags;
380 		if (i + 1 == buf_end)
381 			data_phase_addr = PL35X_SMC_DATA_PHASE + last_flags;
382 
383 		writel(buf32[i], nfc->io_regs + data_phase_addr);
384 	}
385 
386 	/* No working extra flags on unaligned data accesses */
387 	for (i = in_start; i < len; i++)
388 		writeb(buf8[i], nfc->io_regs + PL35X_SMC_DATA_PHASE);
389 
390 	if (force_8bit)
391 		pl35x_smc_force_byte_access(chip, false);
392 }
393 
394 static int pl35x_nand_correct_data(struct pl35x_nandc *nfc, unsigned char *buf,
395 				   unsigned char *read_ecc,
396 				   unsigned char *calc_ecc)
397 {
398 	unsigned short ecc_odd, ecc_even, read_ecc_lower, read_ecc_upper;
399 	unsigned short calc_ecc_lower, calc_ecc_upper;
400 	unsigned short byte_addr, bit_addr;
401 
402 	read_ecc_lower = (read_ecc[0] | (read_ecc[1] << 8)) &
403 			 PL35X_NAND_ECC_BITS_MASK;
404 	read_ecc_upper = ((read_ecc[1] >> 4) | (read_ecc[2] << 4)) &
405 			 PL35X_NAND_ECC_BITS_MASK;
406 
407 	calc_ecc_lower = (calc_ecc[0] | (calc_ecc[1] << 8)) &
408 			 PL35X_NAND_ECC_BITS_MASK;
409 	calc_ecc_upper = ((calc_ecc[1] >> 4) | (calc_ecc[2] << 4)) &
410 			 PL35X_NAND_ECC_BITS_MASK;
411 
412 	ecc_odd = read_ecc_lower ^ calc_ecc_lower;
413 	ecc_even = read_ecc_upper ^ calc_ecc_upper;
414 
415 	/* No error */
416 	if (likely(!ecc_odd && !ecc_even))
417 		return 0;
418 
419 	/* One error in the main data; to be corrected */
420 	if (ecc_odd == (~ecc_even & PL35X_NAND_ECC_BITS_MASK)) {
421 		/* Bits [11:3] of error code give the byte offset */
422 		byte_addr = (ecc_odd >> 3) & PL35X_NAND_ECC_BYTE_OFF_MASK;
423 		/* Bits [2:0] of error code give the bit offset */
424 		bit_addr = ecc_odd & PL35X_NAND_ECC_BIT_OFF_MASK;
425 		/* Toggle the faulty bit */
426 		buf[byte_addr] ^= (BIT(bit_addr));
427 
428 		return 1;
429 	}
430 
431 	/* One error in the ECC data; no action needed */
432 	if (hweight32(ecc_odd | ecc_even) == 1)
433 		return 1;
434 
435 	return -EBADMSG;
436 }
437 
438 static void pl35x_nand_ecc_reg_to_array(struct nand_chip *chip, u32 ecc_reg,
439 					u8 *ecc_array)
440 {
441 	u32 ecc_value = ~ecc_reg;
442 	unsigned int ecc_byte;
443 
444 	for (ecc_byte = 0; ecc_byte < chip->ecc.bytes; ecc_byte++)
445 		ecc_array[ecc_byte] = ecc_value >> (8 * ecc_byte);
446 }
447 
448 static int pl35x_nand_read_eccbytes(struct pl35x_nandc *nfc,
449 				    struct nand_chip *chip, u8 *read_ecc)
450 {
451 	u32 ecc_value;
452 	int chunk;
453 
454 	for (chunk = 0; chunk < chip->ecc.steps;
455 	     chunk++, read_ecc += chip->ecc.bytes) {
456 		ecc_value = readl(nfc->conf_regs + PL35X_SMC_ECC_VALUE(chunk));
457 		if (!PL35X_SMC_ECC_VALUE_IS_VALID(ecc_value))
458 			return -EINVAL;
459 
460 		pl35x_nand_ecc_reg_to_array(chip, ecc_value, read_ecc);
461 	}
462 
463 	return 0;
464 }
465 
466 static int pl35x_nand_recover_data_hwecc(struct pl35x_nandc *nfc,
467 					 struct nand_chip *chip, u8 *data,
468 					 u8 *read_ecc)
469 {
470 	struct mtd_info *mtd = nand_to_mtd(chip);
471 	unsigned int max_bitflips = 0, chunk;
472 	u8 calc_ecc[3];
473 	u32 ecc_value;
474 	int stats;
475 
476 	for (chunk = 0; chunk < chip->ecc.steps;
477 	     chunk++, data += chip->ecc.size, read_ecc += chip->ecc.bytes) {
478 		/* Read ECC value for each chunk */
479 		ecc_value = readl(nfc->conf_regs + PL35X_SMC_ECC_VALUE(chunk));
480 
481 		if (!PL35X_SMC_ECC_VALUE_IS_VALID(ecc_value))
482 			return -EINVAL;
483 
484 		if (PL35X_SMC_ECC_VALUE_HAS_FAILED(ecc_value)) {
485 			mtd->ecc_stats.failed++;
486 			continue;
487 		}
488 
489 		pl35x_nand_ecc_reg_to_array(chip, ecc_value, calc_ecc);
490 		stats = pl35x_nand_correct_data(nfc, data, read_ecc, calc_ecc);
491 		if (stats < 0) {
492 			mtd->ecc_stats.failed++;
493 		} else {
494 			mtd->ecc_stats.corrected += stats;
495 			max_bitflips = max_t(unsigned int, max_bitflips, stats);
496 		}
497 	}
498 
499 	return max_bitflips;
500 }
501 
502 static int pl35x_nand_write_page_hwecc(struct nand_chip *chip,
503 				       const u8 *buf, int oob_required,
504 				       int page)
505 {
506 	struct pl35x_nandc *nfc = to_pl35x_nandc(chip->controller);
507 	struct pl35x_nand *plnand = to_pl35x_nand(chip);
508 	struct mtd_info *mtd = nand_to_mtd(chip);
509 	unsigned int first_row = (mtd->writesize <= 512) ? 1 : 2;
510 	unsigned int nrows = plnand->addr_cycles;
511 	u32 addr1 = 0, addr2 = 0, row;
512 	u32 cmd_addr;
513 	int i, ret;
514 	u8 status;
515 
516 	ret = pl35x_smc_set_ecc_mode(nfc, chip, PL35X_SMC_ECC_CFG_MODE_APB);
517 	if (ret)
518 		return ret;
519 
520 	cmd_addr = PL35X_SMC_CMD_PHASE |
521 		   PL35X_SMC_CMD_PHASE_NADDRS(plnand->addr_cycles) |
522 		   PL35X_SMC_CMD_PHASE_CMD0(NAND_CMD_SEQIN);
523 
524 	for (i = 0, row = first_row; row < nrows; i++, row++) {
525 		u8 addr = page >> ((i * 8) & 0xFF);
526 
527 		if (row < 4)
528 			addr1 |= PL35X_SMC_CMD_PHASE_ADDR(row, addr);
529 		else
530 			addr2 |= PL35X_SMC_CMD_PHASE_ADDR(row - 4, addr);
531 	}
532 
533 	/* Send the command and address cycles */
534 	writel(addr1, nfc->io_regs + cmd_addr);
535 	if (plnand->addr_cycles > 4)
536 		writel(addr2, nfc->io_regs + cmd_addr);
537 
538 	/* Write the data with the engine enabled */
539 	pl35x_nand_write_data_op(chip, buf, mtd->writesize, false,
540 				 0, PL35X_SMC_DATA_PHASE_ECC_LAST);
541 	ret = pl35x_smc_wait_for_ecc_done(nfc);
542 	if (ret)
543 		goto disable_ecc_engine;
544 
545 	/* Copy the HW calculated ECC bytes in the OOB buffer */
546 	ret = pl35x_nand_read_eccbytes(nfc, chip, nfc->ecc_buf);
547 	if (ret)
548 		goto disable_ecc_engine;
549 
550 	if (!oob_required)
551 		memset(chip->oob_poi, 0xFF, mtd->oobsize);
552 
553 	ret = mtd_ooblayout_set_eccbytes(mtd, nfc->ecc_buf, chip->oob_poi,
554 					 0, chip->ecc.total);
555 	if (ret)
556 		goto disable_ecc_engine;
557 
558 	/* Write the spare area with ECC bytes */
559 	pl35x_nand_write_data_op(chip, chip->oob_poi, mtd->oobsize, false, 0,
560 				 PL35X_SMC_CMD_PHASE_CMD1(NAND_CMD_PAGEPROG) |
561 				 PL35X_SMC_CMD_PHASE_CMD1_VALID |
562 				 PL35X_SMC_DATA_PHASE_CLEAR_CS);
563 	ret = pl35x_smc_wait_for_irq(nfc);
564 	if (ret)
565 		goto disable_ecc_engine;
566 
567 	/* Check write status on the chip side */
568 	ret = nand_status_op(chip, &status);
569 	if (ret)
570 		goto disable_ecc_engine;
571 
572 	if (status & NAND_STATUS_FAIL)
573 		ret = -EIO;
574 
575 disable_ecc_engine:
576 	pl35x_smc_set_ecc_mode(nfc, chip, PL35X_SMC_ECC_CFG_MODE_BYPASS);
577 
578 	return ret;
579 }
580 
581 /*
582  * This functions reads data and checks the data integrity by comparing hardware
583  * generated ECC values and read ECC values from spare area.
584  *
585  * There is a limitation with SMC controller: ECC_LAST must be set on the
586  * last data access to tell the ECC engine not to expect any further data.
587  * In practice, this implies to shrink the last data transfert by eg. 4 bytes,
588  * and doing a last 4-byte transfer with the additional bit set. The last block
589  * should be aligned with the end of an ECC block. Because of this limitation,
590  * it is not possible to use the core routines.
591  */
592 static int pl35x_nand_read_page_hwecc(struct nand_chip *chip,
593 				      u8 *buf, int oob_required, int page)
594 {
595 	const struct nand_sdr_timings *sdr =
596 		nand_get_sdr_timings(nand_get_interface_config(chip));
597 	struct pl35x_nandc *nfc = to_pl35x_nandc(chip->controller);
598 	struct pl35x_nand *plnand = to_pl35x_nand(chip);
599 	struct mtd_info *mtd = nand_to_mtd(chip);
600 	unsigned int first_row = (mtd->writesize <= 512) ? 1 : 2;
601 	unsigned int nrows = plnand->addr_cycles;
602 	unsigned int addr1 = 0, addr2 = 0, row;
603 	u32 cmd_addr;
604 	int i, ret;
605 
606 	ret = pl35x_smc_set_ecc_mode(nfc, chip, PL35X_SMC_ECC_CFG_MODE_APB);
607 	if (ret)
608 		return ret;
609 
610 	cmd_addr = PL35X_SMC_CMD_PHASE |
611 		   PL35X_SMC_CMD_PHASE_NADDRS(plnand->addr_cycles) |
612 		   PL35X_SMC_CMD_PHASE_CMD0(NAND_CMD_READ0) |
613 		   PL35X_SMC_CMD_PHASE_CMD1(NAND_CMD_READSTART) |
614 		   PL35X_SMC_CMD_PHASE_CMD1_VALID;
615 
616 	for (i = 0, row = first_row; row < nrows; i++, row++) {
617 		u8 addr = page >> ((i * 8) & 0xFF);
618 
619 		if (row < 4)
620 			addr1 |= PL35X_SMC_CMD_PHASE_ADDR(row, addr);
621 		else
622 			addr2 |= PL35X_SMC_CMD_PHASE_ADDR(row - 4, addr);
623 	}
624 
625 	/* Send the command and address cycles */
626 	writel(addr1, nfc->io_regs + cmd_addr);
627 	if (plnand->addr_cycles > 4)
628 		writel(addr2, nfc->io_regs + cmd_addr);
629 
630 	/* Wait the data to be available in the NAND cache */
631 	ndelay(PSEC_TO_NSEC(sdr->tRR_min));
632 	ret = pl35x_smc_wait_for_irq(nfc);
633 	if (ret)
634 		goto disable_ecc_engine;
635 
636 	/* Retrieve the raw data with the engine enabled */
637 	pl35x_nand_read_data_op(chip, buf, mtd->writesize, false,
638 				0, PL35X_SMC_DATA_PHASE_ECC_LAST);
639 	ret = pl35x_smc_wait_for_ecc_done(nfc);
640 	if (ret)
641 		goto disable_ecc_engine;
642 
643 	/* Retrieve the stored ECC bytes */
644 	pl35x_nand_read_data_op(chip, chip->oob_poi, mtd->oobsize, false,
645 				0, PL35X_SMC_DATA_PHASE_CLEAR_CS);
646 	ret = mtd_ooblayout_get_eccbytes(mtd, nfc->ecc_buf, chip->oob_poi, 0,
647 					 chip->ecc.total);
648 	if (ret)
649 		goto disable_ecc_engine;
650 
651 	pl35x_smc_set_ecc_mode(nfc, chip, PL35X_SMC_ECC_CFG_MODE_BYPASS);
652 
653 	/* Correct the data and report failures */
654 	return pl35x_nand_recover_data_hwecc(nfc, chip, buf, nfc->ecc_buf);
655 
656 disable_ecc_engine:
657 	pl35x_smc_set_ecc_mode(nfc, chip, PL35X_SMC_ECC_CFG_MODE_BYPASS);
658 
659 	return ret;
660 }
661 
662 static int pl35x_nand_exec_op(struct nand_chip *chip,
663 			      const struct nand_subop *subop)
664 {
665 	struct pl35x_nandc *nfc = to_pl35x_nandc(chip->controller);
666 	const struct nand_op_instr *instr, *data_instr = NULL;
667 	unsigned int rdy_tim_ms = 0, naddrs = 0, cmds = 0, last_flags = 0;
668 	u32 addr1 = 0, addr2 = 0, cmd0 = 0, cmd1 = 0, cmd_addr = 0;
669 	unsigned int op_id, len, offset, rdy_del_ns;
670 	int last_instr_type = -1;
671 	bool cmd1_valid = false;
672 	const u8 *addrs;
673 	int i, ret;
674 
675 	for (op_id = 0; op_id < subop->ninstrs; op_id++) {
676 		instr = &subop->instrs[op_id];
677 
678 		switch (instr->type) {
679 		case NAND_OP_CMD_INSTR:
680 			if (!cmds) {
681 				cmd0 = PL35X_SMC_CMD_PHASE_CMD0(instr->ctx.cmd.opcode);
682 			} else {
683 				cmd1 = PL35X_SMC_CMD_PHASE_CMD1(instr->ctx.cmd.opcode);
684 				if (last_instr_type != NAND_OP_DATA_OUT_INSTR)
685 					cmd1_valid = true;
686 			}
687 			cmds++;
688 			break;
689 
690 		case NAND_OP_ADDR_INSTR:
691 			offset = nand_subop_get_addr_start_off(subop, op_id);
692 			naddrs = nand_subop_get_num_addr_cyc(subop, op_id);
693 			addrs = &instr->ctx.addr.addrs[offset];
694 			cmd_addr |= PL35X_SMC_CMD_PHASE_NADDRS(naddrs);
695 
696 			for (i = offset; i < naddrs; i++) {
697 				if (i < 4)
698 					addr1 |= PL35X_SMC_CMD_PHASE_ADDR(i, addrs[i]);
699 				else
700 					addr2 |= PL35X_SMC_CMD_PHASE_ADDR(i - 4, addrs[i]);
701 			}
702 			break;
703 
704 		case NAND_OP_DATA_IN_INSTR:
705 		case NAND_OP_DATA_OUT_INSTR:
706 			data_instr = instr;
707 			len = nand_subop_get_data_len(subop, op_id);
708 			break;
709 
710 		case NAND_OP_WAITRDY_INSTR:
711 			rdy_tim_ms = instr->ctx.waitrdy.timeout_ms;
712 			rdy_del_ns = instr->delay_ns;
713 			break;
714 		}
715 
716 		last_instr_type = instr->type;
717 	}
718 
719 	/* Command phase */
720 	cmd_addr |= PL35X_SMC_CMD_PHASE | cmd0 | cmd1 |
721 		    (cmd1_valid ? PL35X_SMC_CMD_PHASE_CMD1_VALID : 0);
722 	writel(addr1, nfc->io_regs + cmd_addr);
723 	if (naddrs > 4)
724 		writel(addr2, nfc->io_regs + cmd_addr);
725 
726 	/* Data phase */
727 	if (data_instr && data_instr->type == NAND_OP_DATA_OUT_INSTR) {
728 		last_flags = PL35X_SMC_DATA_PHASE_CLEAR_CS;
729 		if (cmds == 2)
730 			last_flags |= cmd1 | PL35X_SMC_CMD_PHASE_CMD1_VALID;
731 
732 		pl35x_nand_write_data_op(chip, data_instr->ctx.data.buf.out,
733 					 len, data_instr->ctx.data.force_8bit,
734 					 0, last_flags);
735 	}
736 
737 	if (rdy_tim_ms) {
738 		ndelay(rdy_del_ns);
739 		ret = pl35x_smc_wait_for_irq(nfc);
740 		if (ret)
741 			return ret;
742 	}
743 
744 	if (data_instr && data_instr->type == NAND_OP_DATA_IN_INSTR)
745 		pl35x_nand_read_data_op(chip, data_instr->ctx.data.buf.in,
746 					len, data_instr->ctx.data.force_8bit,
747 					0, PL35X_SMC_DATA_PHASE_CLEAR_CS);
748 
749 	return 0;
750 }
751 
752 static const struct nand_op_parser pl35x_nandc_op_parser = NAND_OP_PARSER(
753 	NAND_OP_PARSER_PATTERN(pl35x_nand_exec_op,
754 			       NAND_OP_PARSER_PAT_CMD_ELEM(true),
755 			       NAND_OP_PARSER_PAT_ADDR_ELEM(true, 7),
756 			       NAND_OP_PARSER_PAT_CMD_ELEM(true),
757 			       NAND_OP_PARSER_PAT_WAITRDY_ELEM(true),
758 			       NAND_OP_PARSER_PAT_DATA_IN_ELEM(true, 2112)),
759 	NAND_OP_PARSER_PATTERN(pl35x_nand_exec_op,
760 			       NAND_OP_PARSER_PAT_CMD_ELEM(false),
761 			       NAND_OP_PARSER_PAT_ADDR_ELEM(false, 7),
762 			       NAND_OP_PARSER_PAT_DATA_OUT_ELEM(false, 2112),
763 			       NAND_OP_PARSER_PAT_CMD_ELEM(false),
764 			       NAND_OP_PARSER_PAT_WAITRDY_ELEM(true)),
765 	NAND_OP_PARSER_PATTERN(pl35x_nand_exec_op,
766 			       NAND_OP_PARSER_PAT_CMD_ELEM(false),
767 			       NAND_OP_PARSER_PAT_ADDR_ELEM(false, 7),
768 			       NAND_OP_PARSER_PAT_DATA_OUT_ELEM(false, 2112),
769 			       NAND_OP_PARSER_PAT_CMD_ELEM(true),
770 			       NAND_OP_PARSER_PAT_WAITRDY_ELEM(true)),
771 	);
772 
773 static int pl35x_nfc_exec_op(struct nand_chip *chip,
774 			     const struct nand_operation *op,
775 			     bool check_only)
776 {
777 	if (!check_only)
778 		pl35x_nand_select_target(chip, op->cs);
779 
780 	return nand_op_parser_exec_op(chip, &pl35x_nandc_op_parser,
781 				      op, check_only);
782 }
783 
784 static int pl35x_nfc_setup_interface(struct nand_chip *chip, int cs,
785 				     const struct nand_interface_config *conf)
786 {
787 	struct pl35x_nandc *nfc = to_pl35x_nandc(chip->controller);
788 	struct pl35x_nand *plnand = to_pl35x_nand(chip);
789 	struct pl35x_nand_timings tmgs = {};
790 	const struct nand_sdr_timings *sdr;
791 	unsigned int period_ns, val;
792 	struct clk *mclk;
793 
794 	sdr = nand_get_sdr_timings(conf);
795 	if (IS_ERR(sdr))
796 		return PTR_ERR(sdr);
797 
798 	mclk = of_clk_get_by_name(nfc->dev->parent->of_node, "memclk");
799 	if (IS_ERR(mclk)) {
800 		dev_err(nfc->dev, "Failed to retrieve SMC memclk\n");
801 		return PTR_ERR(mclk);
802 	}
803 
804 	/*
805 	 * SDR timings are given in pico-seconds while NFC timings must be
806 	 * expressed in NAND controller clock cycles. We use the TO_CYCLE()
807 	 * macro to convert from one to the other.
808 	 */
809 	period_ns = NSEC_PER_SEC / clk_get_rate(mclk);
810 
811 	/*
812 	 * PL35X SMC needs one extra read cycle in SDR Mode 5. This is not
813 	 * written anywhere in the datasheet but is an empirical observation.
814 	 */
815 	val = TO_CYCLES(sdr->tRC_min, period_ns);
816 	if (sdr->tRC_min <= 20000)
817 		val++;
818 
819 	tmgs.t_rc = val;
820 	if (tmgs.t_rc != val || tmgs.t_rc < 2)
821 		return -EINVAL;
822 
823 	val = TO_CYCLES(sdr->tWC_min, period_ns);
824 	tmgs.t_wc = val;
825 	if (tmgs.t_wc != val || tmgs.t_wc < 2)
826 		return -EINVAL;
827 
828 	/*
829 	 * For all SDR modes, PL35X SMC needs tREA_max being 1,
830 	 * this is also an empirical result.
831 	 */
832 	tmgs.t_rea = 1;
833 
834 	val = TO_CYCLES(sdr->tWP_min, period_ns);
835 	tmgs.t_wp = val;
836 	if (tmgs.t_wp != val || tmgs.t_wp < 1)
837 		return -EINVAL;
838 
839 	val = TO_CYCLES(sdr->tCLR_min, period_ns);
840 	tmgs.t_clr = val;
841 	if (tmgs.t_clr != val)
842 		return -EINVAL;
843 
844 	val = TO_CYCLES(sdr->tAR_min, period_ns);
845 	tmgs.t_ar = val;
846 	if (tmgs.t_ar != val)
847 		return -EINVAL;
848 
849 	val = TO_CYCLES(sdr->tRR_min, period_ns);
850 	tmgs.t_rr = val;
851 	if (tmgs.t_rr != val)
852 		return -EINVAL;
853 
854 	if (cs == NAND_DATA_IFACE_CHECK_ONLY)
855 		return 0;
856 
857 	plnand->timings = PL35X_SMC_NAND_TRC_CYCLES(tmgs.t_rc) |
858 			  PL35X_SMC_NAND_TWC_CYCLES(tmgs.t_wc) |
859 			  PL35X_SMC_NAND_TREA_CYCLES(tmgs.t_rea) |
860 			  PL35X_SMC_NAND_TWP_CYCLES(tmgs.t_wp) |
861 			  PL35X_SMC_NAND_TCLR_CYCLES(tmgs.t_clr) |
862 			  PL35X_SMC_NAND_TAR_CYCLES(tmgs.t_ar) |
863 			  PL35X_SMC_NAND_TRR_CYCLES(tmgs.t_rr);
864 
865 	return 0;
866 }
867 
868 static void pl35x_smc_set_ecc_pg_size(struct pl35x_nandc *nfc,
869 				      struct nand_chip *chip,
870 				      unsigned int pg_sz)
871 {
872 	struct pl35x_nand *plnand = to_pl35x_nand(chip);
873 	u32 sz;
874 
875 	switch (pg_sz) {
876 	case SZ_512:
877 		sz = 1;
878 		break;
879 	case SZ_1K:
880 		sz = 2;
881 		break;
882 	case SZ_2K:
883 		sz = 3;
884 		break;
885 	default:
886 		sz = 0;
887 		break;
888 	}
889 
890 	plnand->ecc_cfg = readl(nfc->conf_regs + PL35X_SMC_ECC_CFG);
891 	plnand->ecc_cfg &= ~PL35X_SMC_ECC_CFG_PGSIZE_MASK;
892 	plnand->ecc_cfg |= sz;
893 	writel(plnand->ecc_cfg, nfc->conf_regs + PL35X_SMC_ECC_CFG);
894 }
895 
896 static int pl35x_nand_init_hw_ecc_controller(struct pl35x_nandc *nfc,
897 					     struct nand_chip *chip)
898 {
899 	struct mtd_info *mtd = nand_to_mtd(chip);
900 	int ret = 0;
901 
902 	if (mtd->writesize < SZ_512 || mtd->writesize > SZ_2K) {
903 		dev_err(nfc->dev,
904 			"The hardware ECC engine is limited to pages up to 2kiB\n");
905 		return -EOPNOTSUPP;
906 	}
907 
908 	chip->ecc.strength = 1;
909 	chip->ecc.bytes = 3;
910 	chip->ecc.size = SZ_512;
911 	chip->ecc.steps = mtd->writesize / chip->ecc.size;
912 	chip->ecc.read_page = pl35x_nand_read_page_hwecc;
913 	chip->ecc.write_page = pl35x_nand_write_page_hwecc;
914 	chip->ecc.write_page_raw = nand_monolithic_write_page_raw;
915 	pl35x_smc_set_ecc_pg_size(nfc, chip, mtd->writesize);
916 
917 	nfc->ecc_buf = devm_kmalloc(nfc->dev, chip->ecc.bytes * chip->ecc.steps,
918 				    GFP_KERNEL);
919 	if (!nfc->ecc_buf)
920 		return -ENOMEM;
921 
922 	switch (mtd->oobsize) {
923 	case 16:
924 		/* Legacy Xilinx layout */
925 		mtd_set_ooblayout(mtd, &pl35x_ecc_ooblayout16_ops);
926 		chip->bbt_options |= NAND_BBT_NO_OOB_BBM;
927 		break;
928 	case 64:
929 		mtd_set_ooblayout(mtd, nand_get_large_page_ooblayout());
930 		break;
931 	default:
932 		dev_err(nfc->dev, "Unsupported OOB size\n");
933 		return -EOPNOTSUPP;
934 	}
935 
936 	return ret;
937 }
938 
939 static int pl35x_nand_attach_chip(struct nand_chip *chip)
940 {
941 	const struct nand_ecc_props *requirements =
942 		nanddev_get_ecc_requirements(&chip->base);
943 	struct pl35x_nandc *nfc = to_pl35x_nandc(chip->controller);
944 	struct pl35x_nand *plnand = to_pl35x_nand(chip);
945 	struct mtd_info *mtd = nand_to_mtd(chip);
946 	int ret;
947 
948 	if (chip->ecc.engine_type != NAND_ECC_ENGINE_TYPE_NONE &&
949 	    (!chip->ecc.size || !chip->ecc.strength)) {
950 		if (requirements->step_size && requirements->strength) {
951 			chip->ecc.size = requirements->step_size;
952 			chip->ecc.strength = requirements->strength;
953 		} else {
954 			dev_info(nfc->dev,
955 				 "No minimum ECC strength, using 1b/512B\n");
956 			chip->ecc.size = 512;
957 			chip->ecc.strength = 1;
958 		}
959 	}
960 
961 	if (mtd->writesize <= SZ_512)
962 		plnand->addr_cycles = 1;
963 	else
964 		plnand->addr_cycles = 2;
965 
966 	if (chip->options & NAND_ROW_ADDR_3)
967 		plnand->addr_cycles += 3;
968 	else
969 		plnand->addr_cycles += 2;
970 
971 	switch (chip->ecc.engine_type) {
972 	case NAND_ECC_ENGINE_TYPE_ON_DIE:
973 		/* Keep these legacy BBT descriptors for ON_DIE situations */
974 		chip->bbt_td = &bbt_main_descr;
975 		chip->bbt_md = &bbt_mirror_descr;
976 		fallthrough;
977 	case NAND_ECC_ENGINE_TYPE_NONE:
978 	case NAND_ECC_ENGINE_TYPE_SOFT:
979 		break;
980 	case NAND_ECC_ENGINE_TYPE_ON_HOST:
981 		ret = pl35x_nand_init_hw_ecc_controller(nfc, chip);
982 		if (ret)
983 			return ret;
984 		break;
985 	default:
986 		dev_err(nfc->dev, "Unsupported ECC mode: %d\n",
987 			chip->ecc.engine_type);
988 		return -EINVAL;
989 	}
990 
991 	return 0;
992 }
993 
994 static const struct nand_controller_ops pl35x_nandc_ops = {
995 	.attach_chip = pl35x_nand_attach_chip,
996 	.exec_op = pl35x_nfc_exec_op,
997 	.setup_interface = pl35x_nfc_setup_interface,
998 };
999 
1000 static int pl35x_nand_reset_state(struct pl35x_nandc *nfc)
1001 {
1002 	int ret;
1003 
1004 	/* Disable interrupts and clear their status */
1005 	writel(PL35X_SMC_MEMC_CFG_CLR_INT_CLR_1 |
1006 	       PL35X_SMC_MEMC_CFG_CLR_ECC_INT_DIS_1 |
1007 	       PL35X_SMC_MEMC_CFG_CLR_INT_DIS_1,
1008 	       nfc->conf_regs + PL35X_SMC_MEMC_CFG_CLR);
1009 
1010 	/* Set default bus width to 8-bit */
1011 	ret = pl35x_smc_set_buswidth(nfc, PL35X_SMC_OPMODE_BW_8);
1012 	if (ret)
1013 		return ret;
1014 
1015 	/* Ensure the ECC controller is bypassed by default */
1016 	ret = pl35x_smc_set_ecc_mode(nfc, NULL, PL35X_SMC_ECC_CFG_MODE_BYPASS);
1017 	if (ret)
1018 		return ret;
1019 
1020 	/*
1021 	 * Configure the commands that the ECC block uses to detect the
1022 	 * operations it should start/end.
1023 	 */
1024 	writel(PL35X_SMC_ECC_CMD1_WRITE(NAND_CMD_SEQIN) |
1025 	       PL35X_SMC_ECC_CMD1_READ(NAND_CMD_READ0) |
1026 	       PL35X_SMC_ECC_CMD1_READ_END(NAND_CMD_READSTART) |
1027 	       PL35X_SMC_ECC_CMD1_READ_END_VALID(NAND_CMD_READ1),
1028 	       nfc->conf_regs + PL35X_SMC_ECC_CMD1);
1029 	writel(PL35X_SMC_ECC_CMD2_WRITE_COL_CHG(NAND_CMD_RNDIN) |
1030 	       PL35X_SMC_ECC_CMD2_READ_COL_CHG(NAND_CMD_RNDOUT) |
1031 	       PL35X_SMC_ECC_CMD2_READ_COL_CHG_END(NAND_CMD_RNDOUTSTART) |
1032 	       PL35X_SMC_ECC_CMD2_READ_COL_CHG_END_VALID(NAND_CMD_READ1),
1033 	       nfc->conf_regs + PL35X_SMC_ECC_CMD2);
1034 
1035 	return 0;
1036 }
1037 
1038 static int pl35x_nand_chip_init(struct pl35x_nandc *nfc,
1039 				struct device_node *np)
1040 {
1041 	struct pl35x_nand *plnand;
1042 	struct nand_chip *chip;
1043 	struct mtd_info *mtd;
1044 	int cs, ret;
1045 
1046 	plnand = devm_kzalloc(nfc->dev, sizeof(*plnand), GFP_KERNEL);
1047 	if (!plnand)
1048 		return -ENOMEM;
1049 
1050 	ret = of_property_read_u32(np, "reg", &cs);
1051 	if (ret)
1052 		return ret;
1053 
1054 	if (cs >= PL35X_NAND_MAX_CS) {
1055 		dev_err(nfc->dev, "Wrong CS %d\n", cs);
1056 		return -EINVAL;
1057 	}
1058 
1059 	if (test_and_set_bit(cs, &nfc->assigned_cs)) {
1060 		dev_err(nfc->dev, "Already assigned CS %d\n", cs);
1061 		return -EINVAL;
1062 	}
1063 
1064 	plnand->cs = cs;
1065 
1066 	chip = &plnand->chip;
1067 	chip->options = NAND_BUSWIDTH_AUTO | NAND_USES_DMA | NAND_NO_SUBPAGE_WRITE;
1068 	chip->bbt_options = NAND_BBT_USE_FLASH;
1069 	chip->controller = &nfc->controller;
1070 	mtd = nand_to_mtd(chip);
1071 	mtd->dev.parent = nfc->dev;
1072 	nand_set_flash_node(chip, np);
1073 	if (!mtd->name) {
1074 		mtd->name = devm_kasprintf(nfc->dev, GFP_KERNEL,
1075 					   "%s", PL35X_NANDC_DRIVER_NAME);
1076 		if (!mtd->name) {
1077 			dev_err(nfc->dev, "Failed to allocate mtd->name\n");
1078 			return -ENOMEM;
1079 		}
1080 	}
1081 
1082 	ret = nand_scan(chip, 1);
1083 	if (ret)
1084 		return ret;
1085 
1086 	ret = mtd_device_register(mtd, NULL, 0);
1087 	if (ret) {
1088 		nand_cleanup(chip);
1089 		return ret;
1090 	}
1091 
1092 	list_add_tail(&plnand->node, &nfc->chips);
1093 
1094 	return ret;
1095 }
1096 
1097 static void pl35x_nand_chips_cleanup(struct pl35x_nandc *nfc)
1098 {
1099 	struct pl35x_nand *plnand, *tmp;
1100 	struct nand_chip *chip;
1101 	int ret;
1102 
1103 	list_for_each_entry_safe(plnand, tmp, &nfc->chips, node) {
1104 		chip = &plnand->chip;
1105 		ret = mtd_device_unregister(nand_to_mtd(chip));
1106 		WARN_ON(ret);
1107 		nand_cleanup(chip);
1108 		list_del(&plnand->node);
1109 	}
1110 }
1111 
1112 static int pl35x_nand_chips_init(struct pl35x_nandc *nfc)
1113 {
1114 	struct device_node *np = nfc->dev->of_node;
1115 	int nchips = of_get_child_count(np);
1116 	int ret;
1117 
1118 	if (!nchips || nchips > PL35X_NAND_MAX_CS) {
1119 		dev_err(nfc->dev, "Incorrect number of NAND chips (%d)\n",
1120 			nchips);
1121 		return -EINVAL;
1122 	}
1123 
1124 	for_each_child_of_node_scoped(np, nand_np) {
1125 		ret = pl35x_nand_chip_init(nfc, nand_np);
1126 		if (ret) {
1127 			pl35x_nand_chips_cleanup(nfc);
1128 			break;
1129 		}
1130 	}
1131 
1132 	return ret;
1133 }
1134 
1135 static int pl35x_nand_probe(struct platform_device *pdev)
1136 {
1137 	struct device *smc_dev = pdev->dev.parent;
1138 	struct amba_device *smc_amba = to_amba_device(smc_dev);
1139 	struct pl35x_nandc *nfc;
1140 	u32 ret;
1141 
1142 	nfc = devm_kzalloc(&pdev->dev, sizeof(*nfc), GFP_KERNEL);
1143 	if (!nfc)
1144 		return -ENOMEM;
1145 
1146 	nfc->dev = &pdev->dev;
1147 	nand_controller_init(&nfc->controller);
1148 	nfc->controller.ops = &pl35x_nandc_ops;
1149 	INIT_LIST_HEAD(&nfc->chips);
1150 
1151 	nfc->conf_regs = devm_ioremap_resource(&smc_amba->dev, &smc_amba->res);
1152 	if (IS_ERR(nfc->conf_regs))
1153 		return PTR_ERR(nfc->conf_regs);
1154 
1155 	nfc->io_regs = devm_platform_ioremap_resource(pdev, 0);
1156 	if (IS_ERR(nfc->io_regs))
1157 		return PTR_ERR(nfc->io_regs);
1158 
1159 	ret = pl35x_nand_reset_state(nfc);
1160 	if (ret)
1161 		return ret;
1162 
1163 	ret = pl35x_nand_chips_init(nfc);
1164 	if (ret)
1165 		return ret;
1166 
1167 	platform_set_drvdata(pdev, nfc);
1168 
1169 	return 0;
1170 }
1171 
1172 static void pl35x_nand_remove(struct platform_device *pdev)
1173 {
1174 	struct pl35x_nandc *nfc = platform_get_drvdata(pdev);
1175 
1176 	pl35x_nand_chips_cleanup(nfc);
1177 }
1178 
1179 static const struct of_device_id pl35x_nand_of_match[] = {
1180 	{ .compatible = "arm,pl353-nand-r2p1" },
1181 	{},
1182 };
1183 MODULE_DEVICE_TABLE(of, pl35x_nand_of_match);
1184 
1185 static struct platform_driver pl35x_nandc_driver = {
1186 	.probe = pl35x_nand_probe,
1187 	.remove_new = pl35x_nand_remove,
1188 	.driver = {
1189 		.name = PL35X_NANDC_DRIVER_NAME,
1190 		.of_match_table = pl35x_nand_of_match,
1191 	},
1192 };
1193 module_platform_driver(pl35x_nandc_driver);
1194 
1195 MODULE_AUTHOR("Xilinx, Inc.");
1196 MODULE_ALIAS("platform:" PL35X_NANDC_DRIVER_NAME);
1197 MODULE_DESCRIPTION("ARM PL35X NAND controller driver");
1198 MODULE_LICENSE("GPL");
1199