xref: /linux/drivers/ata/ahci_xgene.c (revision 63ba5b0fb4f54db256ec43b3062b2606b383055d)
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3  * AppliedMicro X-Gene SoC SATA Host Controller Driver
4  *
5  * Copyright (c) 2014, Applied Micro Circuits Corporation
6  * Author: Loc Ho <lho@apm.com>
7  *         Tuan Phan <tphan@apm.com>
8  *         Suman Tripathi <stripathi@apm.com>
9  *
10  * NOTE: PM support is not currently available.
11  */
12 #include <linux/acpi.h>
13 #include <linux/module.h>
14 #include <linux/platform_device.h>
15 #include <linux/ahci_platform.h>
16 #include <linux/of.h>
17 #include <linux/phy/phy.h>
18 #include "ahci.h"
19 
20 #define DRV_NAME "xgene-ahci"
21 
22 /* Max # of disk per a controller */
23 #define MAX_AHCI_CHN_PERCTR		2
24 
25 /* MUX CSR */
26 #define SATA_ENET_CONFIG_REG		0x00000000
27 #define  CFG_SATA_ENET_SELECT_MASK	0x00000001
28 
29 /* SATA core host controller CSR */
30 #define SLVRDERRATTRIBUTES		0x00000000
31 #define SLVWRERRATTRIBUTES		0x00000004
32 #define MSTRDERRATTRIBUTES		0x00000008
33 #define MSTWRERRATTRIBUTES		0x0000000c
34 #define BUSCTLREG			0x00000014
35 #define IOFMSTRWAUX			0x00000018
36 #define INTSTATUSMASK			0x0000002c
37 #define ERRINTSTATUS			0x00000030
38 #define ERRINTSTATUSMASK		0x00000034
39 
40 /* SATA host AHCI CSR */
41 #define PORTCFG				0x000000a4
42 #define  PORTADDR_SET(dst, src) \
43 		(((dst) & ~0x0000003f) | (((u32)(src)) & 0x0000003f))
44 #define PORTPHY1CFG		0x000000a8
45 #define PORTPHY1CFG_FRCPHYRDY_SET(dst, src) \
46 		(((dst) & ~0x00100000) | (((u32)(src) << 0x14) & 0x00100000))
47 #define PORTPHY2CFG			0x000000ac
48 #define PORTPHY3CFG			0x000000b0
49 #define PORTPHY4CFG			0x000000b4
50 #define PORTPHY5CFG			0x000000b8
51 #define SCTL0				0x0000012C
52 #define PORTPHY5CFG_RTCHG_SET(dst, src) \
53 		(((dst) & ~0xfff00000) | (((u32)(src) << 0x14) & 0xfff00000))
54 #define PORTAXICFG_EN_CONTEXT_SET(dst, src) \
55 		(((dst) & ~0x01000000) | (((u32)(src) << 0x18) & 0x01000000))
56 #define PORTAXICFG			0x000000bc
57 #define PORTAXICFG_OUTTRANS_SET(dst, src) \
58 		(((dst) & ~0x00f00000) | (((u32)(src) << 0x14) & 0x00f00000))
59 #define PORTRANSCFG			0x000000c8
60 #define PORTRANSCFG_RXWM_SET(dst, src)		\
61 		(((dst) & ~0x0000007f) | (((u32)(src)) & 0x0000007f))
62 
63 /* SATA host controller AXI CSR */
64 #define INT_SLV_TMOMASK			0x00000010
65 
66 /* SATA diagnostic CSR */
67 #define CFG_MEM_RAM_SHUTDOWN		0x00000070
68 #define BLOCK_MEM_RDY			0x00000074
69 
70 /* Max retry for link down */
71 #define MAX_LINK_DOWN_RETRY 3
72 
73 enum xgene_ahci_version {
74 	XGENE_AHCI_V1 = 1,
75 	XGENE_AHCI_V2,
76 };
77 
78 struct xgene_ahci_context {
79 	struct ahci_host_priv *hpriv;
80 	struct device *dev;
81 	u8 last_cmd[MAX_AHCI_CHN_PERCTR]; /* tracking the last command issued*/
82 	u32 class[MAX_AHCI_CHN_PERCTR]; /* tracking the class of device */
83 	void __iomem *csr_core;		/* Core CSR address of IP */
84 	void __iomem *csr_diag;		/* Diag CSR address of IP */
85 	void __iomem *csr_axi;		/* AXI CSR address of IP */
86 	void __iomem *csr_mux;		/* MUX CSR address of IP */
87 };
88 
89 static int xgene_ahci_init_memram(struct xgene_ahci_context *ctx)
90 {
91 	dev_dbg(ctx->dev, "Release memory from shutdown\n");
92 	writel(0x0, ctx->csr_diag + CFG_MEM_RAM_SHUTDOWN);
93 	readl(ctx->csr_diag + CFG_MEM_RAM_SHUTDOWN); /* Force a barrier */
94 	msleep(1);	/* reset may take up to 1ms */
95 	if (readl(ctx->csr_diag + BLOCK_MEM_RDY) != 0xFFFFFFFF) {
96 		dev_err(ctx->dev, "failed to release memory from shutdown\n");
97 		return -ENODEV;
98 	}
99 	return 0;
100 }
101 
102 /**
103  * xgene_ahci_poll_reg_val- Poll a register on a specific value.
104  * @ap : ATA port of interest.
105  * @reg : Register of interest.
106  * @val : Value to be attained.
107  * @interval : waiting interval for polling.
108  * @timeout : timeout for achieving the value.
109  */
110 static int xgene_ahci_poll_reg_val(struct ata_port *ap,
111 				   void __iomem *reg, unsigned int val,
112 				   unsigned int interval, unsigned int timeout)
113 {
114 	unsigned long deadline;
115 	unsigned int tmp;
116 
117 	tmp = ioread32(reg);
118 	deadline = ata_deadline(jiffies, timeout);
119 
120 	while (tmp != val && time_before(jiffies, deadline)) {
121 		ata_msleep(ap, interval);
122 		tmp = ioread32(reg);
123 	}
124 
125 	return tmp;
126 }
127 
128 /**
129  * xgene_ahci_restart_engine - Restart the dma engine.
130  * @ap : ATA port of interest
131  *
132  * Waits for completion of multiple commands and restarts
133  * the DMA engine inside the controller.
134  */
135 static int xgene_ahci_restart_engine(struct ata_port *ap)
136 {
137 	struct ahci_host_priv *hpriv = ap->host->private_data;
138 	struct ahci_port_priv *pp = ap->private_data;
139 	void __iomem *port_mmio = ahci_port_base(ap);
140 	u32 fbs;
141 
142 	/*
143 	 * In case of PMP multiple IDENTIFY DEVICE commands can be
144 	 * issued inside PxCI. So need to poll PxCI for the
145 	 * completion of outstanding IDENTIFY DEVICE commands before
146 	 * we restart the DMA engine.
147 	 */
148 	if (xgene_ahci_poll_reg_val(ap, port_mmio +
149 				    PORT_CMD_ISSUE, 0x0, 1, 100))
150 		  return -EBUSY;
151 
152 	hpriv->stop_engine(ap);
153 	ahci_start_fis_rx(ap);
154 
155 	/*
156 	 * Enable the PxFBS.FBS_EN bit as it
157 	 * gets cleared due to stopping the engine.
158 	 */
159 	if (pp->fbs_supported) {
160 		fbs = readl(port_mmio + PORT_FBS);
161 		writel(fbs | PORT_FBS_EN, port_mmio + PORT_FBS);
162 		fbs = readl(port_mmio + PORT_FBS);
163 	}
164 
165 	hpriv->start_engine(ap);
166 
167 	return 0;
168 }
169 
170 /**
171  * xgene_ahci_qc_issue - Issue commands to the device
172  * @qc: Command to issue
173  *
174  * Due to Hardware errata for IDENTIFY DEVICE command, the controller cannot
175  * clear the BSY bit after receiving the PIO setup FIS. This results in the dma
176  * state machine goes into the CMFatalErrorUpdate state and locks up. By
177  * restarting the dma engine, it removes the controller out of lock up state.
178  *
179  * Due to H/W errata, the controller is unable to save the PMP
180  * field fetched from command header before sending the H2D FIS.
181  * When the device returns the PMP port field in the D2H FIS, there is
182  * a mismatch and results in command completion failure. The
183  * workaround is to write the pmp value to PxFBS.DEV field before issuing
184  * any command to PMP.
185  */
186 static unsigned int xgene_ahci_qc_issue(struct ata_queued_cmd *qc)
187 {
188 	struct ata_port *ap = qc->ap;
189 	struct ahci_host_priv *hpriv = ap->host->private_data;
190 	struct xgene_ahci_context *ctx = hpriv->plat_data;
191 	int rc = 0;
192 	u32 port_fbs;
193 	void __iomem *port_mmio = ahci_port_base(ap);
194 
195 	/*
196 	 * Write the pmp value to PxFBS.DEV
197 	 * for case of Port Mulitplier.
198 	 */
199 	if (ctx->class[ap->port_no] == ATA_DEV_PMP) {
200 		port_fbs = readl(port_mmio + PORT_FBS);
201 		port_fbs &= ~PORT_FBS_DEV_MASK;
202 		port_fbs |= qc->dev->link->pmp << PORT_FBS_DEV_OFFSET;
203 		writel(port_fbs, port_mmio + PORT_FBS);
204 	}
205 
206 	if (unlikely((ctx->last_cmd[ap->port_no] == ATA_CMD_ID_ATA) ||
207 	    (ctx->last_cmd[ap->port_no] == ATA_CMD_PACKET) ||
208 	    (ctx->last_cmd[ap->port_no] == ATA_CMD_SMART)))
209 		xgene_ahci_restart_engine(ap);
210 
211 	rc = ahci_qc_issue(qc);
212 
213 	/* Save the last command issued */
214 	ctx->last_cmd[ap->port_no] = qc->tf.command;
215 
216 	return rc;
217 }
218 
219 static bool xgene_ahci_is_memram_inited(struct xgene_ahci_context *ctx)
220 {
221 	void __iomem *diagcsr = ctx->csr_diag;
222 
223 	return (readl(diagcsr + CFG_MEM_RAM_SHUTDOWN) == 0 &&
224 	        readl(diagcsr + BLOCK_MEM_RDY) == 0xFFFFFFFF);
225 }
226 
227 /**
228  * xgene_ahci_read_id - Read ID data from the specified device
229  * @dev: device
230  * @tf: proposed taskfile
231  * @id: data buffer
232  *
233  * This custom read ID function is required due to the fact that the HW
234  * does not support DEVSLP.
235  */
236 static unsigned int xgene_ahci_read_id(struct ata_device *dev,
237 				       struct ata_taskfile *tf, __le16 *id)
238 {
239 	u32 err_mask;
240 
241 	err_mask = ata_do_dev_read_id(dev, tf, id);
242 	if (err_mask)
243 		return err_mask;
244 
245 	/*
246 	 * Mask reserved area. Word78 spec of Link Power Management
247 	 * bit15-8: reserved
248 	 * bit7: NCQ autosence
249 	 * bit6: Software settings preservation supported
250 	 * bit5: reserved
251 	 * bit4: In-order sata delivery supported
252 	 * bit3: DIPM requests supported
253 	 * bit2: DMA Setup FIS Auto-Activate optimization supported
254 	 * bit1: DMA Setup FIX non-Zero buffer offsets supported
255 	 * bit0: Reserved
256 	 *
257 	 * Clear reserved bit 8 (DEVSLP bit) as we don't support DEVSLP
258 	 */
259 	id[ATA_ID_FEATURE_SUPP] &= cpu_to_le16(~(1 << 8));
260 
261 	return 0;
262 }
263 
264 static void xgene_ahci_set_phy_cfg(struct xgene_ahci_context *ctx, int channel)
265 {
266 	void __iomem *mmio = ctx->hpriv->mmio;
267 	u32 val;
268 
269 	dev_dbg(ctx->dev, "port configure mmio 0x%p channel %d\n",
270 		mmio, channel);
271 	val = readl(mmio + PORTCFG);
272 	val = PORTADDR_SET(val, channel == 0 ? 2 : 3);
273 	writel(val, mmio + PORTCFG);
274 	readl(mmio + PORTCFG);  /* Force a barrier */
275 	/* Disable fix rate */
276 	writel(0x0001fffe, mmio + PORTPHY1CFG);
277 	readl(mmio + PORTPHY1CFG); /* Force a barrier */
278 	writel(0x28183219, mmio + PORTPHY2CFG);
279 	readl(mmio + PORTPHY2CFG); /* Force a barrier */
280 	writel(0x13081008, mmio + PORTPHY3CFG);
281 	readl(mmio + PORTPHY3CFG); /* Force a barrier */
282 	writel(0x00480815, mmio + PORTPHY4CFG);
283 	readl(mmio + PORTPHY4CFG); /* Force a barrier */
284 	/* Set window negotiation */
285 	val = readl(mmio + PORTPHY5CFG);
286 	val = PORTPHY5CFG_RTCHG_SET(val, 0x300);
287 	writel(val, mmio + PORTPHY5CFG);
288 	readl(mmio + PORTPHY5CFG); /* Force a barrier */
289 	val = readl(mmio + PORTAXICFG);
290 	val = PORTAXICFG_EN_CONTEXT_SET(val, 0x1); /* Enable context mgmt */
291 	val = PORTAXICFG_OUTTRANS_SET(val, 0xe); /* Set outstanding */
292 	writel(val, mmio + PORTAXICFG);
293 	readl(mmio + PORTAXICFG); /* Force a barrier */
294 	/* Set the watermark threshold of the receive FIFO */
295 	val = readl(mmio + PORTRANSCFG);
296 	val = PORTRANSCFG_RXWM_SET(val, 0x30);
297 	writel(val, mmio + PORTRANSCFG);
298 }
299 
300 /**
301  * xgene_ahci_do_hardreset - Issue the actual COMRESET
302  * @link: link to reset
303  * @deadline: deadline jiffies for the operation
304  * @online: Return value to indicate if device online
305  *
306  * Due to the limitation of the hardware PHY, a difference set of setting is
307  * required for each supported disk speed - Gen3 (6.0Gbps), Gen2 (3.0Gbps),
308  * and Gen1 (1.5Gbps). Otherwise during long IO stress test, the PHY will
309  * report disparity error and etc. In addition, during COMRESET, there can
310  * be error reported in the register PORT_SCR_ERR. For SERR_DISPARITY and
311  * SERR_10B_8B_ERR, the PHY receiver line must be reseted. Also during long
312  * reboot cycle regression, sometimes the PHY reports link down even if the
313  * device is present because of speed negotiation failure. so need to retry
314  * the COMRESET to get the link up. The following algorithm is followed to
315  * proper configure the hardware PHY during COMRESET:
316  *
317  * Alg Part 1:
318  * 1. Start the PHY at Gen3 speed (default setting)
319  * 2. Issue the COMRESET
320  * 3. If no link, go to Alg Part 3
321  * 4. If link up, determine if the negotiated speed matches the PHY
322  *    configured speed
323  * 5. If they matched, go to Alg Part 2
324  * 6. If they do not matched and first time, configure the PHY for the linked
325  *    up disk speed and repeat step 2
326  * 7. Go to Alg Part 2
327  *
328  * Alg Part 2:
329  * 1. On link up, if there are any SERR_DISPARITY and SERR_10B_8B_ERR error
330  *    reported in the register PORT_SCR_ERR, then reset the PHY receiver line
331  * 2. Go to Alg Part 4
332  *
333  * Alg Part 3:
334  * 1. Check the PORT_SCR_STAT to see whether device presence detected but PHY
335  *    communication establishment failed and maximum link down attempts are
336  *    less than Max attempts 3 then goto Alg Part 1.
337  * 2. Go to Alg Part 4.
338  *
339  * Alg Part 4:
340  * 1. Clear any pending from register PORT_SCR_ERR.
341  *
342  * NOTE: For the initial version, we will NOT support Gen1/Gen2. In addition
343  *       and until the underlying PHY supports an method to reset the receiver
344  *       line, on detection of SERR_DISPARITY or SERR_10B_8B_ERR errors,
345  *       an warning message will be printed.
346  */
347 static int xgene_ahci_do_hardreset(struct ata_link *link,
348 				   unsigned long deadline, bool *online)
349 {
350 	const unsigned int *timing = sata_ehc_deb_timing(&link->eh_context);
351 	struct ata_port *ap = link->ap;
352 	struct ahci_host_priv *hpriv = ap->host->private_data;
353 	struct xgene_ahci_context *ctx = hpriv->plat_data;
354 	struct ahci_port_priv *pp = ap->private_data;
355 	u8 *d2h_fis = pp->rx_fis + RX_FIS_D2H_REG;
356 	void __iomem *port_mmio = ahci_port_base(ap);
357 	struct ata_taskfile tf;
358 	int link_down_retry = 0;
359 	int rc;
360 	u32 val, sstatus;
361 
362 	do {
363 		/* clear D2H reception area to properly wait for D2H FIS */
364 		ata_tf_init(link->device, &tf);
365 		tf.status = ATA_BUSY;
366 		ata_tf_to_fis(&tf, 0, 0, d2h_fis);
367 		rc = sata_link_hardreset(link, timing, deadline, online,
368 				 ahci_check_ready);
369 		if (*online) {
370 			val = readl(port_mmio + PORT_SCR_ERR);
371 			if (val & (SERR_DISPARITY | SERR_10B_8B_ERR))
372 				dev_warn(ctx->dev, "link has error\n");
373 			break;
374 		}
375 
376 		sata_scr_read(link, SCR_STATUS, &sstatus);
377 	} while (link_down_retry++ < MAX_LINK_DOWN_RETRY &&
378 		 (sstatus & 0xff) == 0x1);
379 
380 	/* clear all errors if any pending */
381 	val = readl(port_mmio + PORT_SCR_ERR);
382 	writel(val, port_mmio + PORT_SCR_ERR);
383 
384 	return rc;
385 }
386 
387 static int xgene_ahci_hardreset(struct ata_link *link, unsigned int *class,
388 				unsigned long deadline)
389 {
390 	struct ata_port *ap = link->ap;
391         struct ahci_host_priv *hpriv = ap->host->private_data;
392 	void __iomem *port_mmio = ahci_port_base(ap);
393 	bool online;
394 	int rc;
395 	u32 portcmd_saved;
396 	u32 portclb_saved;
397 	u32 portclbhi_saved;
398 	u32 portrxfis_saved;
399 	u32 portrxfishi_saved;
400 
401 	/* As hardreset resets these CSR, save it to restore later */
402 	portcmd_saved = readl(port_mmio + PORT_CMD);
403 	portclb_saved = readl(port_mmio + PORT_LST_ADDR);
404 	portclbhi_saved = readl(port_mmio + PORT_LST_ADDR_HI);
405 	portrxfis_saved = readl(port_mmio + PORT_FIS_ADDR);
406 	portrxfishi_saved = readl(port_mmio + PORT_FIS_ADDR_HI);
407 
408 	hpriv->stop_engine(ap);
409 
410 	rc = xgene_ahci_do_hardreset(link, deadline, &online);
411 
412 	/* As controller hardreset clears them, restore them */
413 	writel(portcmd_saved, port_mmio + PORT_CMD);
414 	writel(portclb_saved, port_mmio + PORT_LST_ADDR);
415 	writel(portclbhi_saved, port_mmio + PORT_LST_ADDR_HI);
416 	writel(portrxfis_saved, port_mmio + PORT_FIS_ADDR);
417 	writel(portrxfishi_saved, port_mmio + PORT_FIS_ADDR_HI);
418 
419 	hpriv->start_engine(ap);
420 
421 	if (online)
422 		*class = ahci_dev_classify(ap);
423 
424 	return rc;
425 }
426 
427 static void xgene_ahci_host_stop(struct ata_host *host)
428 {
429 	struct ahci_host_priv *hpriv = host->private_data;
430 
431 	ahci_platform_disable_resources(hpriv);
432 }
433 
434 /**
435  * xgene_ahci_pmp_softreset - Issue the softreset to the drives connected
436  *                            to Port Multiplier.
437  * @link: link to reset
438  * @class: Return value to indicate class of device
439  * @deadline: deadline jiffies for the operation
440  *
441  * Due to H/W errata, the controller is unable to save the PMP
442  * field fetched from command header before sending the H2D FIS.
443  * When the device returns the PMP port field in the D2H FIS, there is
444  * a mismatch and results in command completion failure. The workaround
445  * is to write the pmp value to PxFBS.DEV field before issuing any command
446  * to PMP.
447  */
448 static int xgene_ahci_pmp_softreset(struct ata_link *link, unsigned int *class,
449 			  unsigned long deadline)
450 {
451 	int pmp = sata_srst_pmp(link);
452 	struct ata_port *ap = link->ap;
453 	u32 rc;
454 	void __iomem *port_mmio = ahci_port_base(ap);
455 	u32 port_fbs;
456 
457 	/*
458 	 * Set PxFBS.DEV field with pmp
459 	 * value.
460 	 */
461 	port_fbs = readl(port_mmio + PORT_FBS);
462 	port_fbs &= ~PORT_FBS_DEV_MASK;
463 	port_fbs |= pmp << PORT_FBS_DEV_OFFSET;
464 	writel(port_fbs, port_mmio + PORT_FBS);
465 
466 	rc = ahci_do_softreset(link, class, pmp, deadline, ahci_check_ready);
467 
468 	return rc;
469 }
470 
471 /**
472  * xgene_ahci_softreset - Issue the softreset to the drive.
473  * @link: link to reset
474  * @class: Return value to indicate class of device
475  * @deadline: deadline jiffies for the operation
476  *
477  * Due to H/W errata, the controller is unable to save the PMP
478  * field fetched from command header before sending the H2D FIS.
479  * When the device returns the PMP port field in the D2H FIS, there is
480  * a mismatch and results in command completion failure. The workaround
481  * is to write the pmp value to PxFBS.DEV field before issuing any command
482  * to PMP. Here is the algorithm to detect PMP :
483  *
484  * 1. Save the PxFBS value
485  * 2. Program PxFBS.DEV with pmp value send by framework. Framework sends
486  *    0xF for both PMP/NON-PMP initially
487  * 3. Issue softreset
488  * 4. If signature class is PMP goto 6
489  * 5. restore the original PxFBS and goto 3
490  * 6. return
491  */
492 static int xgene_ahci_softreset(struct ata_link *link, unsigned int *class,
493 			  unsigned long deadline)
494 {
495 	int pmp = sata_srst_pmp(link);
496 	struct ata_port *ap = link->ap;
497 	struct ahci_host_priv *hpriv = ap->host->private_data;
498 	struct xgene_ahci_context *ctx = hpriv->plat_data;
499 	void __iomem *port_mmio = ahci_port_base(ap);
500 	u32 port_fbs;
501 	u32 port_fbs_save;
502 	u32 retry = 1;
503 	u32 rc;
504 
505 	port_fbs_save = readl(port_mmio + PORT_FBS);
506 
507 	/*
508 	 * Set PxFBS.DEV field with pmp
509 	 * value.
510 	 */
511 	port_fbs = readl(port_mmio + PORT_FBS);
512 	port_fbs &= ~PORT_FBS_DEV_MASK;
513 	port_fbs |= pmp << PORT_FBS_DEV_OFFSET;
514 	writel(port_fbs, port_mmio + PORT_FBS);
515 
516 softreset_retry:
517 	rc = ahci_do_softreset(link, class, pmp,
518 			       deadline, ahci_check_ready);
519 
520 	ctx->class[ap->port_no] = *class;
521 	if (*class != ATA_DEV_PMP) {
522 		/*
523 		 * Retry for normal drives without
524 		 * setting PxFBS.DEV field with pmp value.
525 		 */
526 		if (retry--) {
527 			writel(port_fbs_save, port_mmio + PORT_FBS);
528 			goto softreset_retry;
529 		}
530 	}
531 
532 	return rc;
533 }
534 
535 /**
536  * xgene_ahci_handle_broken_edge_irq - Handle the broken irq.
537  * @host: Host that recieved the irq
538  * @irq_masked: HOST_IRQ_STAT value
539  *
540  * For hardware with broken edge trigger latch
541  * the HOST_IRQ_STAT register misses the edge interrupt
542  * when clearing of HOST_IRQ_STAT register and hardware
543  * reporting the PORT_IRQ_STAT register at the
544  * same clock cycle.
545  * As such, the algorithm below outlines the workaround.
546  *
547  * 1. Read HOST_IRQ_STAT register and save the state.
548  * 2. Clear the HOST_IRQ_STAT register.
549  * 3. Read back the HOST_IRQ_STAT register.
550  * 4. If HOST_IRQ_STAT register equals to zero, then
551  *    traverse the rest of port's PORT_IRQ_STAT register
552  *    to check if an interrupt is triggered at that point else
553  *    go to step 6.
554  * 5. If PORT_IRQ_STAT register of rest ports is not equal to zero
555  *    then update the state of HOST_IRQ_STAT saved in step 1.
556  * 6. Handle port interrupts.
557  * 7. Exit
558  */
559 static int xgene_ahci_handle_broken_edge_irq(struct ata_host *host,
560 					     u32 irq_masked)
561 {
562 	struct ahci_host_priv *hpriv = host->private_data;
563 	void __iomem *port_mmio;
564 	int i;
565 
566 	if (!readl(hpriv->mmio + HOST_IRQ_STAT)) {
567 		for (i = 0; i < host->n_ports; i++) {
568 			if (irq_masked & (1 << i))
569 				continue;
570 
571 			port_mmio = ahci_port_base(host->ports[i]);
572 			if (readl(port_mmio + PORT_IRQ_STAT))
573 				irq_masked |= (1 << i);
574 		}
575 	}
576 
577 	return ahci_handle_port_intr(host, irq_masked);
578 }
579 
580 static irqreturn_t xgene_ahci_irq_intr(int irq, void *dev_instance)
581 {
582 	struct ata_host *host = dev_instance;
583 	struct ahci_host_priv *hpriv;
584 	unsigned int rc = 0;
585 	void __iomem *mmio;
586 	u32 irq_stat, irq_masked;
587 
588 	hpriv = host->private_data;
589 	mmio = hpriv->mmio;
590 
591 	/* sigh.  0xffffffff is a valid return from h/w */
592 	irq_stat = readl(mmio + HOST_IRQ_STAT);
593 	if (!irq_stat)
594 		return IRQ_NONE;
595 
596 	irq_masked = irq_stat & hpriv->port_map;
597 
598 	spin_lock(&host->lock);
599 
600 	/*
601 	 * HOST_IRQ_STAT behaves as edge triggered latch meaning that
602 	 * it should be cleared before all the port events are cleared.
603 	 */
604 	writel(irq_stat, mmio + HOST_IRQ_STAT);
605 
606 	rc = xgene_ahci_handle_broken_edge_irq(host, irq_masked);
607 
608 	spin_unlock(&host->lock);
609 
610 	return IRQ_RETVAL(rc);
611 }
612 
613 static struct ata_port_operations xgene_ahci_v1_ops = {
614 	.inherits = &ahci_ops,
615 	.host_stop = xgene_ahci_host_stop,
616 	.hardreset = xgene_ahci_hardreset,
617 	.read_id = xgene_ahci_read_id,
618 	.qc_issue = xgene_ahci_qc_issue,
619 	.softreset = xgene_ahci_softreset,
620 	.pmp_softreset = xgene_ahci_pmp_softreset
621 };
622 
623 static const struct ata_port_info xgene_ahci_v1_port_info = {
624 	.flags = AHCI_FLAG_COMMON | ATA_FLAG_PMP,
625 	.pio_mask = ATA_PIO4,
626 	.udma_mask = ATA_UDMA6,
627 	.port_ops = &xgene_ahci_v1_ops,
628 };
629 
630 static struct ata_port_operations xgene_ahci_v2_ops = {
631 	.inherits = &ahci_ops,
632 	.host_stop = xgene_ahci_host_stop,
633 	.hardreset = xgene_ahci_hardreset,
634 	.read_id = xgene_ahci_read_id,
635 };
636 
637 static const struct ata_port_info xgene_ahci_v2_port_info = {
638 	.flags = AHCI_FLAG_COMMON | ATA_FLAG_PMP,
639 	.pio_mask = ATA_PIO4,
640 	.udma_mask = ATA_UDMA6,
641 	.port_ops = &xgene_ahci_v2_ops,
642 };
643 
644 static int xgene_ahci_hw_init(struct ahci_host_priv *hpriv)
645 {
646 	struct xgene_ahci_context *ctx = hpriv->plat_data;
647 	int i;
648 	int rc;
649 	u32 val;
650 
651 	/* Remove IP RAM out of shutdown */
652 	rc = xgene_ahci_init_memram(ctx);
653 	if (rc)
654 		return rc;
655 
656 	for (i = 0; i < MAX_AHCI_CHN_PERCTR; i++)
657 		xgene_ahci_set_phy_cfg(ctx, i);
658 
659 	/* AXI disable Mask */
660 	writel(0xffffffff, hpriv->mmio + HOST_IRQ_STAT);
661 	readl(hpriv->mmio + HOST_IRQ_STAT); /* Force a barrier */
662 	writel(0, ctx->csr_core + INTSTATUSMASK);
663 	val = readl(ctx->csr_core + INTSTATUSMASK); /* Force a barrier */
664 	dev_dbg(ctx->dev, "top level interrupt mask 0x%X value 0x%08X\n",
665 		INTSTATUSMASK, val);
666 
667 	writel(0x0, ctx->csr_core + ERRINTSTATUSMASK);
668 	readl(ctx->csr_core + ERRINTSTATUSMASK); /* Force a barrier */
669 	writel(0x0, ctx->csr_axi + INT_SLV_TMOMASK);
670 	readl(ctx->csr_axi + INT_SLV_TMOMASK);
671 
672 	/* Enable AXI Interrupt */
673 	writel(0xffffffff, ctx->csr_core + SLVRDERRATTRIBUTES);
674 	writel(0xffffffff, ctx->csr_core + SLVWRERRATTRIBUTES);
675 	writel(0xffffffff, ctx->csr_core + MSTRDERRATTRIBUTES);
676 	writel(0xffffffff, ctx->csr_core + MSTWRERRATTRIBUTES);
677 
678 	/* Enable coherency */
679 	val = readl(ctx->csr_core + BUSCTLREG);
680 	val &= ~0x00000002;     /* Enable write coherency */
681 	val &= ~0x00000001;     /* Enable read coherency */
682 	writel(val, ctx->csr_core + BUSCTLREG);
683 
684 	val = readl(ctx->csr_core + IOFMSTRWAUX);
685 	val |= (1 << 3);        /* Enable read coherency */
686 	val |= (1 << 9);        /* Enable write coherency */
687 	writel(val, ctx->csr_core + IOFMSTRWAUX);
688 	val = readl(ctx->csr_core + IOFMSTRWAUX);
689 	dev_dbg(ctx->dev, "coherency 0x%X value 0x%08X\n",
690 		IOFMSTRWAUX, val);
691 
692 	return rc;
693 }
694 
695 static int xgene_ahci_mux_select(struct xgene_ahci_context *ctx)
696 {
697 	u32 val;
698 
699 	/* Check for optional MUX resource */
700 	if (!ctx->csr_mux)
701 		return 0;
702 
703 	val = readl(ctx->csr_mux + SATA_ENET_CONFIG_REG);
704 	val &= ~CFG_SATA_ENET_SELECT_MASK;
705 	writel(val, ctx->csr_mux + SATA_ENET_CONFIG_REG);
706 	val = readl(ctx->csr_mux + SATA_ENET_CONFIG_REG);
707 	return val & CFG_SATA_ENET_SELECT_MASK ? -1 : 0;
708 }
709 
710 static const struct scsi_host_template ahci_platform_sht = {
711 	AHCI_SHT(DRV_NAME),
712 };
713 
714 #ifdef CONFIG_ACPI
715 static const struct acpi_device_id xgene_ahci_acpi_match[] = {
716 	{ "APMC0D0D", XGENE_AHCI_V1},
717 	{ "APMC0D32", XGENE_AHCI_V2},
718 	{},
719 };
720 MODULE_DEVICE_TABLE(acpi, xgene_ahci_acpi_match);
721 #endif
722 
723 static const struct of_device_id xgene_ahci_of_match[] = {
724 	{.compatible = "apm,xgene-ahci", .data = (void *) XGENE_AHCI_V1},
725 	{.compatible = "apm,xgene-ahci-v2", .data = (void *) XGENE_AHCI_V2},
726 	{ /* sentinel */ }
727 };
728 MODULE_DEVICE_TABLE(of, xgene_ahci_of_match);
729 
730 static int xgene_ahci_probe(struct platform_device *pdev)
731 {
732 	struct device *dev = &pdev->dev;
733 	struct ahci_host_priv *hpriv;
734 	struct xgene_ahci_context *ctx;
735 	struct resource *res;
736 	enum xgene_ahci_version version = XGENE_AHCI_V1;
737 	const struct ata_port_info *ppi[] = { &xgene_ahci_v1_port_info,
738 					      &xgene_ahci_v2_port_info };
739 	int rc;
740 
741 	hpriv = ahci_platform_get_resources(pdev, 0);
742 	if (IS_ERR(hpriv))
743 		return PTR_ERR(hpriv);
744 
745 	ctx = devm_kzalloc(dev, sizeof(*ctx), GFP_KERNEL);
746 	if (!ctx)
747 		return -ENOMEM;
748 
749 	hpriv->plat_data = ctx;
750 	ctx->hpriv = hpriv;
751 	ctx->dev = dev;
752 
753 	/* Retrieve the IP core resource */
754 	ctx->csr_core = devm_platform_ioremap_resource(pdev, 1);
755 	if (IS_ERR(ctx->csr_core))
756 		return PTR_ERR(ctx->csr_core);
757 
758 	/* Retrieve the IP diagnostic resource */
759 	ctx->csr_diag = devm_platform_ioremap_resource(pdev, 2);
760 	if (IS_ERR(ctx->csr_diag))
761 		return PTR_ERR(ctx->csr_diag);
762 
763 	/* Retrieve the IP AXI resource */
764 	ctx->csr_axi = devm_platform_ioremap_resource(pdev, 3);
765 	if (IS_ERR(ctx->csr_axi))
766 		return PTR_ERR(ctx->csr_axi);
767 
768 	/* Retrieve the optional IP mux resource */
769 	res = platform_get_resource(pdev, IORESOURCE_MEM, 4);
770 	if (res) {
771 		void __iomem *csr = devm_ioremap_resource(dev, res);
772 		if (IS_ERR(csr))
773 			return PTR_ERR(csr);
774 
775 		ctx->csr_mux = csr;
776 	}
777 
778 	if (dev->of_node) {
779 		version = (enum xgene_ahci_version)of_device_get_match_data(dev);
780 	}
781 #ifdef CONFIG_ACPI
782 	else {
783 		const struct acpi_device_id *acpi_id;
784 		struct acpi_device_info *info;
785 		acpi_status status;
786 
787 		acpi_id = acpi_match_device(xgene_ahci_acpi_match, &pdev->dev);
788 		if (!acpi_id) {
789 			dev_warn(&pdev->dev, "No node entry in ACPI table. Assume version1\n");
790 			version = XGENE_AHCI_V1;
791 		} else if (acpi_id->driver_data) {
792 			version = (enum xgene_ahci_version) acpi_id->driver_data;
793 			status = acpi_get_object_info(ACPI_HANDLE(&pdev->dev), &info);
794 			if (ACPI_FAILURE(status)) {
795 				dev_warn(&pdev->dev, "%s: Error reading device info. Assume version1\n",
796 					__func__);
797 				version = XGENE_AHCI_V1;
798 			} else {
799 				if (info->valid & ACPI_VALID_CID)
800 					version = XGENE_AHCI_V2;
801 				kfree(info);
802 			}
803 		}
804 	}
805 #endif
806 
807 	dev_dbg(dev, "VAddr 0x%p Mmio VAddr 0x%p\n", ctx->csr_core,
808 		hpriv->mmio);
809 
810 	/* Select ATA */
811 	if ((rc = xgene_ahci_mux_select(ctx))) {
812 		dev_err(dev, "SATA mux selection failed error %d\n", rc);
813 		return -ENODEV;
814 	}
815 
816 	if (xgene_ahci_is_memram_inited(ctx)) {
817 		dev_info(dev, "skip clock and PHY initialization\n");
818 		goto skip_clk_phy;
819 	}
820 
821 	/* Due to errata, HW requires full toggle transition */
822 	rc = ahci_platform_enable_clks(hpriv);
823 	if (rc)
824 		goto disable_resources;
825 	ahci_platform_disable_clks(hpriv);
826 
827 	rc = ahci_platform_enable_resources(hpriv);
828 	if (rc)
829 		goto disable_resources;
830 
831 	/* Configure the host controller */
832 	xgene_ahci_hw_init(hpriv);
833 skip_clk_phy:
834 
835 	switch (version) {
836 	case XGENE_AHCI_V1:
837 		hpriv->flags = AHCI_HFLAG_NO_NCQ;
838 		break;
839 	case XGENE_AHCI_V2:
840 		hpriv->flags |= AHCI_HFLAG_YES_FBS;
841 		hpriv->irq_handler = xgene_ahci_irq_intr;
842 		break;
843 	default:
844 		break;
845 	}
846 
847 	rc = ahci_platform_init_host(pdev, hpriv, ppi[version - 1],
848 				     &ahci_platform_sht);
849 	if (rc)
850 		goto disable_resources;
851 
852 	dev_dbg(dev, "X-Gene SATA host controller initialized\n");
853 	return 0;
854 
855 disable_resources:
856 	ahci_platform_disable_resources(hpriv);
857 	return rc;
858 }
859 
860 static struct platform_driver xgene_ahci_driver = {
861 	.probe = xgene_ahci_probe,
862 	.remove_new = ata_platform_remove_one,
863 	.driver = {
864 		.name = DRV_NAME,
865 		.of_match_table = xgene_ahci_of_match,
866 		.acpi_match_table = ACPI_PTR(xgene_ahci_acpi_match),
867 	},
868 };
869 
870 module_platform_driver(xgene_ahci_driver);
871 
872 MODULE_DESCRIPTION("APM X-Gene AHCI SATA driver");
873 MODULE_AUTHOR("Loc Ho <lho@apm.com>");
874 MODULE_LICENSE("GPL");
875 MODULE_VERSION("0.4");
876