xref: /linux/drivers/spi/spi-pci1xxxx.c (revision 5ac79730324c6f37106ce397586020ffe6e8e234)
1 // SPDX-License-Identifier: GPL-2.0
2 // PCI1xxxx SPI driver
3 // Copyright (C) 2022 Microchip Technology Inc.
4 // Authors: Tharun Kumar P <tharunkumar.pasumarthi@microchip.com>
5 //          Kumaravel Thiagarajan <Kumaravel.Thiagarajan@microchip.com>
6 
7 
8 #include <linux/bitfield.h>
9 #include <linux/delay.h>
10 #include <linux/dma-mapping.h>
11 #include <linux/iopoll.h>
12 #include <linux/irq.h>
13 #include <linux/module.h>
14 #include <linux/msi.h>
15 #include <linux/pci_regs.h>
16 #include <linux/pci.h>
17 #include <linux/spinlock.h>
18 #include <linux/spi/spi.h>
19 #include "internals.h"
20 
21 #define DRV_NAME "spi-pci1xxxx"
22 
23 #define	SYS_FREQ_DEFAULT		(62500000)
24 
25 #define	PCI1XXXX_SPI_MAX_CLOCK_HZ	(30000000)
26 #define	PCI1XXXX_SPI_CLK_20MHZ		(20000000)
27 #define	PCI1XXXX_SPI_CLK_15MHZ		(15000000)
28 #define	PCI1XXXX_SPI_CLK_12MHZ		(12000000)
29 #define	PCI1XXXX_SPI_CLK_10MHZ		(10000000)
30 #define	PCI1XXXX_SPI_MIN_CLOCK_HZ	(2000000)
31 
32 #define	PCI1XXXX_SPI_BUFFER_SIZE	(320)
33 
34 #define	SPI_MST_CTL_DEVSEL_MASK		(GENMASK(27, 25))
35 #define	SPI_MST_CTL_CMD_LEN_MASK	(GENMASK(16, 8))
36 #define	SPI_MST_CTL_SPEED_MASK		(GENMASK(7, 5))
37 #define	SPI_MSI_VECTOR_SEL_MASK		(GENMASK(4, 4))
38 
39 #define	SPI_MST_CTL_FORCE_CE		(BIT(4))
40 #define	SPI_MST_CTL_MODE_SEL		(BIT(2))
41 #define	SPI_MST_CTL_GO			(BIT(0))
42 
43 #define SPI_PERI_ADDR_BASE		(0x160000)
44 #define SPI_SYSTEM_ADDR_BASE		(0x2000)
45 #define	SPI_MST1_ADDR_BASE		(0x800)
46 
47 #define DEV_REV_REG			(SPI_SYSTEM_ADDR_BASE + 0x00)
48 #define SPI_SYSLOCK_REG			(SPI_SYSTEM_ADDR_BASE + 0xA0)
49 #define SPI_CONFIG_PERI_ENABLE_REG	(SPI_SYSTEM_ADDR_BASE + 0x108)
50 
51 #define SPI_PERI_ENBLE_PF_MASK		(GENMASK(17, 16))
52 #define DEV_REV_MASK			(GENMASK(7, 0))
53 
54 #define SPI_SYSLOCK			BIT(4)
55 #define SPI0				(0)
56 #define SPI1				(1)
57 
58 /* DMA Related Registers */
59 #define SPI_DMA_ADDR_BASE		(0x1000)
60 #define SPI_DMA_GLOBAL_WR_ENGINE_EN	(SPI_DMA_ADDR_BASE + 0x0C)
61 #define SPI_DMA_WR_DOORBELL_REG		(SPI_DMA_ADDR_BASE + 0x10)
62 #define SPI_DMA_GLOBAL_RD_ENGINE_EN	(SPI_DMA_ADDR_BASE + 0x2C)
63 #define SPI_DMA_RD_DOORBELL_REG		(SPI_DMA_ADDR_BASE + 0x30)
64 #define SPI_DMA_INTR_WR_STS		(SPI_DMA_ADDR_BASE + 0x4C)
65 #define SPI_DMA_WR_INT_MASK		(SPI_DMA_ADDR_BASE + 0x54)
66 #define SPI_DMA_INTR_WR_CLR		(SPI_DMA_ADDR_BASE + 0x58)
67 #define SPI_DMA_ERR_WR_STS		(SPI_DMA_ADDR_BASE + 0x5C)
68 #define SPI_DMA_INTR_IMWR_WDONE_LOW	(SPI_DMA_ADDR_BASE + 0x60)
69 #define SPI_DMA_INTR_IMWR_WDONE_HIGH	(SPI_DMA_ADDR_BASE + 0x64)
70 #define SPI_DMA_INTR_IMWR_WABORT_LOW	(SPI_DMA_ADDR_BASE + 0x68)
71 #define SPI_DMA_INTR_IMWR_WABORT_HIGH	(SPI_DMA_ADDR_BASE + 0x6C)
72 #define SPI_DMA_INTR_WR_IMWR_DATA	(SPI_DMA_ADDR_BASE + 0x70)
73 #define SPI_DMA_INTR_RD_STS		(SPI_DMA_ADDR_BASE + 0xA0)
74 #define SPI_DMA_RD_INT_MASK		(SPI_DMA_ADDR_BASE + 0xA8)
75 #define SPI_DMA_INTR_RD_CLR		(SPI_DMA_ADDR_BASE + 0xAC)
76 #define SPI_DMA_ERR_RD_STS		(SPI_DMA_ADDR_BASE + 0xB8)
77 #define SPI_DMA_INTR_IMWR_RDONE_LOW	(SPI_DMA_ADDR_BASE + 0xCC)
78 #define SPI_DMA_INTR_IMWR_RDONE_HIGH	(SPI_DMA_ADDR_BASE + 0xD0)
79 #define SPI_DMA_INTR_IMWR_RABORT_LOW	(SPI_DMA_ADDR_BASE + 0xD4)
80 #define SPI_DMA_INTR_IMWR_RABORT_HIGH	(SPI_DMA_ADDR_BASE + 0xD8)
81 #define SPI_DMA_INTR_RD_IMWR_DATA	(SPI_DMA_ADDR_BASE + 0xDC)
82 
83 #define SPI_DMA_CH0_WR_BASE		(SPI_DMA_ADDR_BASE + 0x200)
84 #define SPI_DMA_CH0_RD_BASE		(SPI_DMA_ADDR_BASE + 0x300)
85 #define SPI_DMA_CH1_WR_BASE		(SPI_DMA_ADDR_BASE + 0x400)
86 #define SPI_DMA_CH1_RD_BASE		(SPI_DMA_ADDR_BASE + 0x500)
87 
88 #define SPI_DMA_CH_CTL1_OFFSET		(0x00)
89 #define SPI_DMA_CH_XFER_LEN_OFFSET	(0x08)
90 #define SPI_DMA_CH_SAR_LO_OFFSET	(0x0C)
91 #define SPI_DMA_CH_SAR_HI_OFFSET	(0x10)
92 #define SPI_DMA_CH_DAR_LO_OFFSET	(0x14)
93 #define SPI_DMA_CH_DAR_HI_OFFSET	(0x18)
94 
95 #define SPI_DMA_CH0_DONE_INT		BIT(0)
96 #define SPI_DMA_CH1_DONE_INT		BIT(1)
97 #define SPI_DMA_CH0_ABORT_INT		BIT(16)
98 #define SPI_DMA_CH1_ABORT_INT		BIT(17)
99 #define SPI_DMA_DONE_INT_MASK		(SPI_DMA_CH0_DONE_INT | SPI_DMA_CH1_DONE_INT)
100 #define SPI_DMA_ABORT_INT_MASK		(SPI_DMA_CH0_ABORT_INT | SPI_DMA_CH1_ABORT_INT)
101 #define DMA_CH_CONTROL_LIE		BIT(3)
102 #define DMA_CH_CONTROL_RIE		BIT(4)
103 #define DMA_INTR_EN			(DMA_CH_CONTROL_RIE | DMA_CH_CONTROL_LIE)
104 
105 /* x refers to SPI Host Controller HW instance id in the below macros - 0 or 1 */
106 
107 #define	SPI_MST_CMD_BUF_OFFSET(x)		(((x) * SPI_MST1_ADDR_BASE) + 0x00)
108 #define	SPI_MST_RSP_BUF_OFFSET(x)		(((x) * SPI_MST1_ADDR_BASE) + 0x200)
109 #define	SPI_MST_CTL_REG_OFFSET(x)		(((x) * SPI_MST1_ADDR_BASE) + 0x400)
110 #define	SPI_MST_EVENT_REG_OFFSET(x)		(((x) * SPI_MST1_ADDR_BASE) + 0x420)
111 #define	SPI_MST_EVENT_MASK_REG_OFFSET(x)	(((x) * SPI_MST1_ADDR_BASE) + 0x424)
112 #define	SPI_MST_PAD_CTL_REG_OFFSET(x)		(((x) * SPI_MST1_ADDR_BASE) + 0x460)
113 #define	SPIALERT_MST_DB_REG_OFFSET(x)		(((x) * SPI_MST1_ADDR_BASE) + 0x464)
114 #define	SPIALERT_MST_VAL_REG_OFFSET(x)		(((x) * SPI_MST1_ADDR_BASE) + 0x468)
115 #define	SPI_PCI_CTRL_REG_OFFSET(x)		(((x) * SPI_MST1_ADDR_BASE) + 0x480)
116 
117 #define PCI1XXXX_IRQ_FLAGS			(IRQF_NO_SUSPEND | IRQF_TRIGGER_NONE)
118 #define SPI_MAX_DATA_LEN			320
119 
120 #define PCI1XXXX_SPI_TIMEOUT			(msecs_to_jiffies(100))
121 #define SYSLOCK_RETRY_CNT			(1000)
122 #define SPI_DMA_ENGINE_EN			(0x1)
123 #define SPI_DMA_ENGINE_DIS			(0x0)
124 
125 #define SPI_INTR		BIT(8)
126 #define SPI_FORCE_CE		BIT(4)
127 
128 #define SPI_CHIP_SEL_COUNT 7
129 #define VENDOR_ID_MCHP 0x1055
130 
131 #define SPI_SUSPEND_CONFIG 0x101
132 #define SPI_RESUME_CONFIG 0x203
133 
134 struct pci1xxxx_spi_internal {
135 	u8 hw_inst;
136 	u8 clkdiv;
137 	int irq;
138 	int mode;
139 	bool spi_xfer_in_progress;
140 	void *rx_buf;
141 	bool dma_aborted_rd;
142 	u32 bytes_recvd;
143 	u32 tx_sgl_len;
144 	u32 rx_sgl_len;
145 	struct scatterlist *tx_sgl, *rx_sgl;
146 	bool dma_aborted_wr;
147 	struct completion spi_xfer_done;
148 	struct spi_controller *spi_host;
149 	struct pci1xxxx_spi *parent;
150 	struct spi_transfer *xfer;
151 	struct {
152 		unsigned int dev_sel : 3;
153 		unsigned int msi_vector_sel : 1;
154 	} prev_val;
155 };
156 
157 struct pci1xxxx_spi {
158 	struct pci_dev *dev;
159 	u8 total_hw_instances;
160 	u8 dev_rev;
161 	void __iomem *reg_base;
162 	void __iomem *dma_offset_bar;
163 	/* lock to safely access the DMA registers in isr */
164 	spinlock_t dma_reg_lock;
165 	bool can_dma;
166 	struct pci1xxxx_spi_internal *spi_int[] __counted_by(total_hw_instances);
167 };
168 
169 static const struct pci_device_id pci1xxxx_spi_pci_id_table[] = {
170 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa004, PCI_ANY_ID, 0x0001), 0, 0, 0x02},
171 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa004, PCI_ANY_ID, 0x0002), 0, 0, 0x01},
172 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa004, PCI_ANY_ID, 0x0003), 0, 0, 0x11},
173 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa004, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01},
174 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa014, PCI_ANY_ID, 0x0001), 0, 0, 0x02},
175 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa014, PCI_ANY_ID, 0x0002), 0, 0, 0x01},
176 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa014, PCI_ANY_ID, 0x0003), 0, 0, 0x11},
177 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa014, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01},
178 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa024, PCI_ANY_ID, 0x0001), 0, 0, 0x02},
179 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa024, PCI_ANY_ID, 0x0002), 0, 0, 0x01},
180 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa024, PCI_ANY_ID, 0x0003), 0, 0, 0x11},
181 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa024, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01},
182 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa034, PCI_ANY_ID, 0x0001), 0, 0, 0x02},
183 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa034, PCI_ANY_ID, 0x0002), 0, 0, 0x01},
184 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa034, PCI_ANY_ID, 0x0003), 0, 0, 0x11},
185 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa034, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01},
186 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa044, PCI_ANY_ID, 0x0001), 0, 0, 0x02},
187 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa044, PCI_ANY_ID, 0x0002), 0, 0, 0x01},
188 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa044, PCI_ANY_ID, 0x0003), 0, 0, 0x11},
189 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa044, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01},
190 	{ 0, }
191 };
192 
193 MODULE_DEVICE_TABLE(pci, pci1xxxx_spi_pci_id_table);
194 
195 static int pci1xxxx_set_sys_lock(struct pci1xxxx_spi *par)
196 {
197 	writel(SPI_SYSLOCK, par->reg_base + SPI_SYSLOCK_REG);
198 	return readl(par->reg_base + SPI_SYSLOCK_REG);
199 }
200 
201 static int pci1xxxx_acquire_sys_lock(struct pci1xxxx_spi *par)
202 {
203 	u32 regval;
204 
205 	return readx_poll_timeout(pci1xxxx_set_sys_lock, par, regval,
206 			   (regval & SPI_SYSLOCK), 100,
207 			   SYSLOCK_RETRY_CNT * 100);
208 }
209 
210 static void pci1xxxx_release_sys_lock(struct pci1xxxx_spi *par)
211 {
212 	writel(0x0, par->reg_base + SPI_SYSLOCK_REG);
213 }
214 
215 static int pci1xxxx_check_spi_can_dma(struct pci1xxxx_spi *spi_bus, int irq)
216 {
217 	struct pci_dev *pdev = spi_bus->dev;
218 	u32 pf_num;
219 	u32 regval;
220 	int ret;
221 
222 	/*
223 	 * DEV REV Registers is a system register, HW Syslock bit
224 	 * should be acquired before accessing the register
225 	 */
226 	ret = pci1xxxx_acquire_sys_lock(spi_bus);
227 	if (ret) {
228 		dev_err(&pdev->dev, "Error failed to acquire syslock\n");
229 		return ret;
230 	}
231 
232 	regval = readl(spi_bus->reg_base + DEV_REV_REG);
233 	spi_bus->dev_rev = regval & DEV_REV_MASK;
234 	if (spi_bus->dev_rev >= 0xC0) {
235 		regval = readl(spi_bus->reg_base +
236 			       SPI_CONFIG_PERI_ENABLE_REG);
237 		pf_num = regval & SPI_PERI_ENBLE_PF_MASK;
238 	}
239 
240 	pci1xxxx_release_sys_lock(spi_bus);
241 
242 	/*
243 	 * DMA is supported only from C0 and SPI can use DMA only if
244 	 * it is mapped to PF0
245 	 */
246 	if (spi_bus->dev_rev < 0xC0 || pf_num)
247 		return -EOPNOTSUPP;
248 
249 	/*
250 	 * DMA Supported only with MSI Interrupts
251 	 * One of the SPI instance's MSI vector address and data
252 	 * is used for DMA Interrupt
253 	 */
254 	if (!irq_get_msi_desc(irq)) {
255 		dev_warn(&pdev->dev, "Error MSI Interrupt not supported, will operate in PIO mode\n");
256 		return -EOPNOTSUPP;
257 	}
258 
259 	spi_bus->dma_offset_bar = pcim_iomap(pdev, 2, pci_resource_len(pdev, 2));
260 	if (!spi_bus->dma_offset_bar) {
261 		dev_warn(&pdev->dev, "Error failed to map dma bar, will operate in PIO mode\n");
262 		return -EOPNOTSUPP;
263 	}
264 
265 	if (dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(64))) {
266 		dev_warn(&pdev->dev, "Error failed to set DMA mask, will operate in PIO mode\n");
267 		pcim_iounmap(pdev, spi_bus->dma_offset_bar);
268 		spi_bus->dma_offset_bar = NULL;
269 		return -EOPNOTSUPP;
270 	}
271 
272 	return 0;
273 }
274 
275 static int pci1xxxx_spi_dma_init(struct pci1xxxx_spi *spi_bus, int irq)
276 {
277 	struct msi_msg msi;
278 	int ret;
279 
280 	ret = pci1xxxx_check_spi_can_dma(spi_bus, irq);
281 	if (ret)
282 		return ret;
283 
284 	spin_lock_init(&spi_bus->dma_reg_lock);
285 	get_cached_msi_msg(irq, &msi);
286 	writel(SPI_DMA_ENGINE_EN, spi_bus->dma_offset_bar + SPI_DMA_GLOBAL_WR_ENGINE_EN);
287 	writel(SPI_DMA_ENGINE_EN, spi_bus->dma_offset_bar + SPI_DMA_GLOBAL_RD_ENGINE_EN);
288 	writel(msi.address_hi, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_WDONE_HIGH);
289 	writel(msi.address_hi, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_WABORT_HIGH);
290 	writel(msi.address_hi, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_RDONE_HIGH);
291 	writel(msi.address_hi, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_RABORT_HIGH);
292 	writel(msi.address_lo, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_WDONE_LOW);
293 	writel(msi.address_lo, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_WABORT_LOW);
294 	writel(msi.address_lo, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_RDONE_LOW);
295 	writel(msi.address_lo, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_RABORT_LOW);
296 	writel(msi.data, spi_bus->dma_offset_bar + SPI_DMA_INTR_WR_IMWR_DATA);
297 	writel(msi.data, spi_bus->dma_offset_bar + SPI_DMA_INTR_RD_IMWR_DATA);
298 	dma_set_max_seg_size(&spi_bus->dev->dev, PCI1XXXX_SPI_BUFFER_SIZE);
299 	spi_bus->can_dma = true;
300 	return 0;
301 }
302 
303 static void pci1xxxx_spi_set_cs(struct spi_device *spi, bool enable)
304 {
305 	struct pci1xxxx_spi_internal *p = spi_controller_get_devdata(spi->controller);
306 	struct pci1xxxx_spi *par = p->parent;
307 	u32 regval;
308 
309 	/* Set the DEV_SEL bits of the SPI_MST_CTL_REG */
310 	regval = readl(par->reg_base + SPI_MST_CTL_REG_OFFSET(p->hw_inst));
311 	if (!enable) {
312 		regval |= SPI_FORCE_CE;
313 		regval &= ~SPI_MST_CTL_DEVSEL_MASK;
314 		regval |= (spi_get_chipselect(spi, 0) << 25);
315 	} else {
316 		regval &= ~SPI_FORCE_CE;
317 	}
318 	writel(regval, par->reg_base + SPI_MST_CTL_REG_OFFSET(p->hw_inst));
319 }
320 
321 static u8 pci1xxxx_get_clock_div(u32 hz)
322 {
323 	u8 val = 0;
324 
325 	if (hz >= PCI1XXXX_SPI_MAX_CLOCK_HZ)
326 		val = 2;
327 	else if ((hz < PCI1XXXX_SPI_MAX_CLOCK_HZ) && (hz >= PCI1XXXX_SPI_CLK_20MHZ))
328 		val = 3;
329 	else if ((hz < PCI1XXXX_SPI_CLK_20MHZ) && (hz >= PCI1XXXX_SPI_CLK_15MHZ))
330 		val = 4;
331 	else if ((hz < PCI1XXXX_SPI_CLK_15MHZ) && (hz >= PCI1XXXX_SPI_CLK_12MHZ))
332 		val = 5;
333 	else if ((hz < PCI1XXXX_SPI_CLK_12MHZ) && (hz >= PCI1XXXX_SPI_CLK_10MHZ))
334 		val = 6;
335 	else if ((hz < PCI1XXXX_SPI_CLK_10MHZ) && (hz >= PCI1XXXX_SPI_MIN_CLOCK_HZ))
336 		val = 7;
337 	else
338 		val = 2;
339 
340 	return val;
341 }
342 
343 static void pci1xxxx_spi_setup_dma_to_io(struct pci1xxxx_spi_internal *p,
344 					 dma_addr_t dma_addr, u32 len)
345 {
346 	void __iomem *base;
347 
348 	if (!p->hw_inst)
349 		base = p->parent->dma_offset_bar + SPI_DMA_CH0_RD_BASE;
350 	else
351 		base = p->parent->dma_offset_bar + SPI_DMA_CH1_RD_BASE;
352 
353 	writel(DMA_INTR_EN, base + SPI_DMA_CH_CTL1_OFFSET);
354 	writel(len, base + SPI_DMA_CH_XFER_LEN_OFFSET);
355 	writel(lower_32_bits(dma_addr), base + SPI_DMA_CH_SAR_LO_OFFSET);
356 	writel(upper_32_bits(dma_addr), base + SPI_DMA_CH_SAR_HI_OFFSET);
357 	/* Updated SPI Command Registers */
358 	writel(lower_32_bits(SPI_PERI_ADDR_BASE + SPI_MST_CMD_BUF_OFFSET(p->hw_inst)),
359 	       base + SPI_DMA_CH_DAR_LO_OFFSET);
360 	writel(upper_32_bits(SPI_PERI_ADDR_BASE + SPI_MST_CMD_BUF_OFFSET(p->hw_inst)),
361 	       base + SPI_DMA_CH_DAR_HI_OFFSET);
362 }
363 
364 static void pci1xxxx_spi_setup_dma_from_io(struct pci1xxxx_spi_internal *p,
365 					   dma_addr_t dma_addr, u32 len)
366 {
367 	void *base;
368 
369 	if (!p->hw_inst)
370 		base = p->parent->dma_offset_bar + SPI_DMA_CH0_WR_BASE;
371 	else
372 		base = p->parent->dma_offset_bar + SPI_DMA_CH1_WR_BASE;
373 
374 	writel(DMA_INTR_EN, base + SPI_DMA_CH_CTL1_OFFSET);
375 	writel(len, base + SPI_DMA_CH_XFER_LEN_OFFSET);
376 	writel(lower_32_bits(dma_addr), base + SPI_DMA_CH_DAR_LO_OFFSET);
377 	writel(upper_32_bits(dma_addr), base + SPI_DMA_CH_DAR_HI_OFFSET);
378 	writel(lower_32_bits(SPI_PERI_ADDR_BASE + SPI_MST_RSP_BUF_OFFSET(p->hw_inst)),
379 	       base + SPI_DMA_CH_SAR_LO_OFFSET);
380 	writel(upper_32_bits(SPI_PERI_ADDR_BASE + SPI_MST_RSP_BUF_OFFSET(p->hw_inst)),
381 	       base + SPI_DMA_CH_SAR_HI_OFFSET);
382 }
383 
384 static void pci1xxxx_spi_setup(struct pci1xxxx_spi *par, u8 hw_inst, u32 mode,
385 			       u8 clkdiv, u32 len)
386 {
387 	u32 regval;
388 
389 	regval = readl(par->reg_base + SPI_MST_CTL_REG_OFFSET(hw_inst));
390 	regval &= ~(SPI_MST_CTL_MODE_SEL | SPI_MST_CTL_CMD_LEN_MASK |
391 		    SPI_MST_CTL_SPEED_MASK);
392 
393 	if (mode == SPI_MODE_3)
394 		regval |= SPI_MST_CTL_MODE_SEL;
395 
396 	regval |= FIELD_PREP(SPI_MST_CTL_CMD_LEN_MASK, len);
397 	regval |= FIELD_PREP(SPI_MST_CTL_SPEED_MASK, clkdiv);
398 	writel(regval, par->reg_base + SPI_MST_CTL_REG_OFFSET(hw_inst));
399 }
400 
401 static void pci1xxxx_start_spi_xfer(struct pci1xxxx_spi_internal *p, u8 hw_inst)
402 {
403 	u32 regval;
404 
405 	regval = readl(p->parent->reg_base + SPI_MST_CTL_REG_OFFSET(hw_inst));
406 	regval |= SPI_MST_CTL_GO;
407 	writel(regval, p->parent->reg_base + SPI_MST_CTL_REG_OFFSET(hw_inst));
408 }
409 
410 static int pci1xxxx_spi_transfer_with_io(struct spi_controller *spi_ctlr,
411 					 struct spi_device *spi, struct spi_transfer *xfer)
412 {
413 	struct pci1xxxx_spi_internal *p = spi_controller_get_devdata(spi_ctlr);
414 	struct pci1xxxx_spi *par = p->parent;
415 	int len, loop_iter, transfer_len;
416 	unsigned long bytes_transfered;
417 	unsigned long bytes_recvd;
418 	unsigned long loop_count;
419 	u8 *rx_buf, result;
420 	const u8 *tx_buf;
421 	u32 regval;
422 	u8 clkdiv;
423 
424 	p->spi_xfer_in_progress = true;
425 	p->bytes_recvd = 0;
426 	clkdiv = pci1xxxx_get_clock_div(xfer->speed_hz);
427 	tx_buf = xfer->tx_buf;
428 	rx_buf = xfer->rx_buf;
429 	transfer_len = xfer->len;
430 	regval = readl(par->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
431 	writel(regval, par->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
432 
433 	if (tx_buf) {
434 		bytes_transfered = 0;
435 		bytes_recvd = 0;
436 		loop_count = transfer_len / SPI_MAX_DATA_LEN;
437 		if (transfer_len % SPI_MAX_DATA_LEN != 0)
438 			loop_count += 1;
439 
440 		for (loop_iter = 0; loop_iter < loop_count; loop_iter++) {
441 			len = SPI_MAX_DATA_LEN;
442 			if ((transfer_len % SPI_MAX_DATA_LEN != 0) &&
443 			    (loop_iter == loop_count - 1))
444 				len = transfer_len % SPI_MAX_DATA_LEN;
445 
446 			reinit_completion(&p->spi_xfer_done);
447 			memcpy_toio(par->reg_base + SPI_MST_CMD_BUF_OFFSET(p->hw_inst),
448 				    &tx_buf[bytes_transfered], len);
449 			bytes_transfered += len;
450 			pci1xxxx_spi_setup(par, p->hw_inst, spi->mode, clkdiv, len);
451 			pci1xxxx_start_spi_xfer(p, p->hw_inst);
452 
453 			/* Wait for DMA_TERM interrupt */
454 			result = wait_for_completion_timeout(&p->spi_xfer_done,
455 							     PCI1XXXX_SPI_TIMEOUT);
456 			if (!result)
457 				return -ETIMEDOUT;
458 
459 			if (rx_buf) {
460 				memcpy_fromio(&rx_buf[bytes_recvd], par->reg_base +
461 					      SPI_MST_RSP_BUF_OFFSET(p->hw_inst), len);
462 				bytes_recvd += len;
463 			}
464 		}
465 	}
466 	p->spi_xfer_in_progress = false;
467 
468 	return 0;
469 }
470 
471 static int pci1xxxx_spi_transfer_with_dma(struct spi_controller *spi_ctlr,
472 					  struct spi_device *spi,
473 					  struct spi_transfer *xfer)
474 {
475 	struct pci1xxxx_spi_internal *p = spi_controller_get_devdata(spi_ctlr);
476 	struct pci1xxxx_spi *par = p->parent;
477 	dma_addr_t rx_dma_addr = 0;
478 	dma_addr_t tx_dma_addr = 0;
479 	int ret = 0;
480 	u32 regval;
481 
482 	p->spi_xfer_in_progress = true;
483 	p->tx_sgl = xfer->tx_sg.sgl;
484 	p->rx_sgl = xfer->rx_sg.sgl;
485 	p->rx_buf = xfer->rx_buf;
486 	regval = readl(par->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
487 	writel(regval, par->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
488 
489 	if (!xfer->tx_buf || !p->tx_sgl) {
490 		ret = -EINVAL;
491 		goto error;
492 	}
493 	p->xfer = xfer;
494 	p->mode = spi->mode;
495 	p->clkdiv = pci1xxxx_get_clock_div(xfer->speed_hz);
496 	p->bytes_recvd = 0;
497 	p->rx_buf = xfer->rx_buf;
498 	regval = readl(par->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
499 	writel(regval, par->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
500 
501 	tx_dma_addr = sg_dma_address(p->tx_sgl);
502 	rx_dma_addr = sg_dma_address(p->rx_sgl);
503 	p->tx_sgl_len = sg_dma_len(p->tx_sgl);
504 	p->rx_sgl_len = sg_dma_len(p->rx_sgl);
505 	pci1xxxx_spi_setup(par, p->hw_inst, p->mode, p->clkdiv, p->tx_sgl_len);
506 	pci1xxxx_spi_setup_dma_to_io(p, (tx_dma_addr), p->tx_sgl_len);
507 	if (rx_dma_addr)
508 		pci1xxxx_spi_setup_dma_from_io(p, rx_dma_addr, p->rx_sgl_len);
509 	writel(p->hw_inst, par->dma_offset_bar + SPI_DMA_RD_DOORBELL_REG);
510 
511 	reinit_completion(&p->spi_xfer_done);
512 	/* Wait for DMA_TERM interrupt */
513 	ret = wait_for_completion_timeout(&p->spi_xfer_done, PCI1XXXX_SPI_TIMEOUT);
514 	if (!ret) {
515 		ret = -ETIMEDOUT;
516 		if (p->dma_aborted_rd) {
517 			writel(SPI_DMA_ENGINE_DIS,
518 			       par->dma_offset_bar + SPI_DMA_GLOBAL_RD_ENGINE_EN);
519 			/*
520 			 * DMA ENGINE reset takes time if any TLP
521 			 * completeion in progress, should wait
522 			 * till DMA Engine reset is completed.
523 			 */
524 			ret = readl_poll_timeout(par->dma_offset_bar +
525 						 SPI_DMA_GLOBAL_RD_ENGINE_EN, regval,
526 						 (regval == 0x0), 0, USEC_PER_MSEC);
527 			if (ret) {
528 				ret = -ECANCELED;
529 				goto error;
530 			}
531 			writel(SPI_DMA_ENGINE_EN,
532 			       par->dma_offset_bar + SPI_DMA_GLOBAL_RD_ENGINE_EN);
533 			p->dma_aborted_rd = false;
534 			ret = -ECANCELED;
535 		}
536 		if (p->dma_aborted_wr) {
537 			writel(SPI_DMA_ENGINE_DIS,
538 			       par->dma_offset_bar + SPI_DMA_GLOBAL_WR_ENGINE_EN);
539 
540 			/*
541 			 * DMA ENGINE reset takes time if any TLP
542 			 * completeion in progress, should wait
543 			 * till DMA Engine reset is completed.
544 			 */
545 			ret = readl_poll_timeout(par->dma_offset_bar +
546 						 SPI_DMA_GLOBAL_WR_ENGINE_EN, regval,
547 						 (regval == 0x0), 0, USEC_PER_MSEC);
548 			if (ret) {
549 				ret = -ECANCELED;
550 				goto error;
551 			}
552 
553 			writel(SPI_DMA_ENGINE_EN,
554 			       par->dma_offset_bar + SPI_DMA_GLOBAL_WR_ENGINE_EN);
555 			p->dma_aborted_wr = false;
556 			ret = -ECANCELED;
557 		}
558 		goto error;
559 	}
560 	ret = 0;
561 
562 error:
563 	p->spi_xfer_in_progress = false;
564 
565 	return ret;
566 }
567 
568 static int pci1xxxx_spi_transfer_one(struct spi_controller *spi_ctlr,
569 				     struct spi_device *spi, struct spi_transfer *xfer)
570 {
571 	if (spi_xfer_is_dma_mapped(spi_ctlr, spi, xfer))
572 		return pci1xxxx_spi_transfer_with_dma(spi_ctlr, spi, xfer);
573 	else
574 		return pci1xxxx_spi_transfer_with_io(spi_ctlr, spi, xfer);
575 }
576 
577 static irqreturn_t pci1xxxx_spi_isr_io(int irq, void *dev)
578 {
579 	struct pci1xxxx_spi_internal *p = dev;
580 	irqreturn_t spi_int_fired = IRQ_NONE;
581 	u32 regval;
582 
583 	/* Clear the SPI GO_BIT Interrupt */
584 	regval = readl(p->parent->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
585 	if (regval & SPI_INTR) {
586 		/* Clear xfer_done */
587 		if (p->parent->can_dma && p->rx_buf)
588 			writel(p->hw_inst, p->parent->dma_offset_bar +
589 			       SPI_DMA_WR_DOORBELL_REG);
590 		else
591 			complete(&p->parent->spi_int[p->hw_inst]->spi_xfer_done);
592 		spi_int_fired = IRQ_HANDLED;
593 	}
594 	writel(regval, p->parent->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
595 	return spi_int_fired;
596 }
597 
598 static void pci1xxxx_spi_setup_next_dma_transfer(struct pci1xxxx_spi_internal *p)
599 {
600 	dma_addr_t tx_dma_addr = 0;
601 	dma_addr_t rx_dma_addr = 0;
602 	u32 prev_len;
603 
604 	p->tx_sgl = sg_next(p->tx_sgl);
605 	if (p->rx_sgl)
606 		p->rx_sgl = sg_next(p->rx_sgl);
607 	if (!p->tx_sgl) {
608 		/* Clear xfer_done */
609 		complete(&p->spi_xfer_done);
610 	} else {
611 		tx_dma_addr = sg_dma_address(p->tx_sgl);
612 		prev_len = p->tx_sgl_len;
613 		p->tx_sgl_len = sg_dma_len(p->tx_sgl);
614 		if (prev_len != p->tx_sgl_len)
615 			pci1xxxx_spi_setup(p->parent,
616 					   p->hw_inst, p->mode, p->clkdiv, p->tx_sgl_len);
617 		pci1xxxx_spi_setup_dma_to_io(p, tx_dma_addr, p->tx_sgl_len);
618 		if (p->rx_sgl) {
619 			rx_dma_addr = sg_dma_address(p->rx_sgl);
620 			p->rx_sgl_len = sg_dma_len(p->rx_sgl);
621 			pci1xxxx_spi_setup_dma_from_io(p, rx_dma_addr, p->rx_sgl_len);
622 		}
623 		writel(p->hw_inst, p->parent->dma_offset_bar + SPI_DMA_RD_DOORBELL_REG);
624 	}
625 }
626 
627 static irqreturn_t pci1xxxx_spi_isr_dma(int irq, void *dev)
628 {
629 	struct pci1xxxx_spi_internal *p = dev;
630 	irqreturn_t spi_int_fired = IRQ_NONE;
631 	unsigned long flags;
632 	u32 regval;
633 
634 	spin_lock_irqsave(&p->parent->dma_reg_lock, flags);
635 	/* Clear the DMA RD INT and start spi xfer*/
636 	regval = readl(p->parent->dma_offset_bar + SPI_DMA_INTR_RD_STS);
637 	if (regval & SPI_DMA_DONE_INT_MASK) {
638 		if (regval & SPI_DMA_CH0_DONE_INT)
639 			pci1xxxx_start_spi_xfer(p, SPI0);
640 		if (regval & SPI_DMA_CH1_DONE_INT)
641 			pci1xxxx_start_spi_xfer(p, SPI1);
642 		spi_int_fired = IRQ_HANDLED;
643 	}
644 	if (regval & SPI_DMA_ABORT_INT_MASK) {
645 		p->dma_aborted_rd = true;
646 		spi_int_fired = IRQ_HANDLED;
647 	}
648 	writel(regval, p->parent->dma_offset_bar + SPI_DMA_INTR_RD_CLR);
649 
650 	/* Clear the DMA WR INT */
651 	regval = readl(p->parent->dma_offset_bar + SPI_DMA_INTR_WR_STS);
652 	if (regval & SPI_DMA_DONE_INT_MASK) {
653 		if (regval & SPI_DMA_CH0_DONE_INT)
654 			pci1xxxx_spi_setup_next_dma_transfer(p->parent->spi_int[SPI0]);
655 
656 		if (regval & SPI_DMA_CH1_DONE_INT)
657 			pci1xxxx_spi_setup_next_dma_transfer(p->parent->spi_int[SPI1]);
658 
659 		spi_int_fired = IRQ_HANDLED;
660 	}
661 	if (regval & SPI_DMA_ABORT_INT_MASK) {
662 		p->dma_aborted_wr = true;
663 		spi_int_fired = IRQ_HANDLED;
664 	}
665 	writel(regval, p->parent->dma_offset_bar + SPI_DMA_INTR_WR_CLR);
666 	spin_unlock_irqrestore(&p->parent->dma_reg_lock, flags);
667 
668 	/* Clear the SPI GO_BIT Interrupt */
669 	regval = readl(p->parent->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
670 	if (regval & SPI_INTR) {
671 		writel(p->hw_inst, p->parent->dma_offset_bar + SPI_DMA_WR_DOORBELL_REG);
672 		spi_int_fired = IRQ_HANDLED;
673 	}
674 	writel(regval, p->parent->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
675 	return spi_int_fired;
676 }
677 
678 static irqreturn_t pci1xxxx_spi_isr(int irq, void *dev)
679 {
680 	struct pci1xxxx_spi_internal *p = dev;
681 
682 	if (p->spi_host->can_dma(p->spi_host, NULL, p->xfer))
683 		return pci1xxxx_spi_isr_dma(irq, dev);
684 	else
685 		return pci1xxxx_spi_isr_io(irq, dev);
686 }
687 
688 static bool pci1xxxx_spi_can_dma(struct spi_controller *host,
689 				 struct spi_device *spi,
690 				 struct spi_transfer *xfer)
691 {
692 	struct pci1xxxx_spi_internal *p = spi_controller_get_devdata(host);
693 	struct pci1xxxx_spi *par = p->parent;
694 
695 	return par->can_dma;
696 }
697 
698 static int pci1xxxx_spi_probe(struct pci_dev *pdev, const struct pci_device_id *ent)
699 {
700 	u8 hw_inst_cnt, iter, start, only_sec_inst;
701 	struct pci1xxxx_spi_internal *spi_sub_ptr;
702 	struct device *dev = &pdev->dev;
703 	struct pci1xxxx_spi *spi_bus;
704 	struct spi_controller *spi_host;
705 	u32 regval;
706 	int ret;
707 
708 	hw_inst_cnt = ent->driver_data & 0x0f;
709 	start = (ent->driver_data & 0xf0) >> 4;
710 	if (start == 1)
711 		only_sec_inst = 1;
712 	else
713 		only_sec_inst = 0;
714 
715 	spi_bus = devm_kzalloc(&pdev->dev,
716 			       struct_size(spi_bus, spi_int, hw_inst_cnt),
717 			       GFP_KERNEL);
718 	if (!spi_bus)
719 		return -ENOMEM;
720 
721 	spi_bus->dev = pdev;
722 	spi_bus->total_hw_instances = hw_inst_cnt;
723 	pci_set_master(pdev);
724 
725 	for (iter = 0; iter < hw_inst_cnt; iter++) {
726 		spi_bus->spi_int[iter] = devm_kzalloc(&pdev->dev,
727 						      sizeof(struct pci1xxxx_spi_internal),
728 						      GFP_KERNEL);
729 		if (!spi_bus->spi_int[iter])
730 			return -ENOMEM;
731 		spi_sub_ptr = spi_bus->spi_int[iter];
732 		spi_sub_ptr->spi_host = devm_spi_alloc_host(dev, sizeof(struct spi_controller));
733 		if (!spi_sub_ptr->spi_host)
734 			return -ENOMEM;
735 
736 		spi_sub_ptr->parent = spi_bus;
737 		spi_sub_ptr->spi_xfer_in_progress = false;
738 
739 		if (!iter) {
740 			ret = pcim_enable_device(pdev);
741 			if (ret)
742 				return -ENOMEM;
743 
744 			ret = pci_request_regions(pdev, DRV_NAME);
745 			if (ret)
746 				return -ENOMEM;
747 
748 			spi_bus->reg_base = pcim_iomap(pdev, 0, pci_resource_len(pdev, 0));
749 			if (!spi_bus->reg_base) {
750 				ret = -EINVAL;
751 				goto error;
752 			}
753 
754 			ret = pci_alloc_irq_vectors(pdev, hw_inst_cnt, hw_inst_cnt,
755 						    PCI_IRQ_ALL_TYPES);
756 			if (ret < 0) {
757 				dev_err(&pdev->dev, "Error allocating MSI vectors\n");
758 				goto error;
759 			}
760 
761 			init_completion(&spi_sub_ptr->spi_xfer_done);
762 			/* Initialize Interrupts - SPI_INT */
763 			regval = readl(spi_bus->reg_base +
764 				       SPI_MST_EVENT_MASK_REG_OFFSET(spi_sub_ptr->hw_inst));
765 			regval &= ~SPI_INTR;
766 			writel(regval, spi_bus->reg_base +
767 			       SPI_MST_EVENT_MASK_REG_OFFSET(spi_sub_ptr->hw_inst));
768 			spi_sub_ptr->irq = pci_irq_vector(pdev, 0);
769 
770 			ret = devm_request_irq(&pdev->dev, spi_sub_ptr->irq,
771 					       pci1xxxx_spi_isr, PCI1XXXX_IRQ_FLAGS,
772 					       pci_name(pdev), spi_sub_ptr);
773 			if (ret < 0) {
774 				dev_err(&pdev->dev, "Unable to request irq : %d",
775 					spi_sub_ptr->irq);
776 				ret = -ENODEV;
777 				goto error;
778 			}
779 
780 			ret = pci1xxxx_spi_dma_init(spi_bus, spi_sub_ptr->irq);
781 			if (ret && ret != -EOPNOTSUPP)
782 				goto error;
783 
784 			/* This register is only applicable for 1st instance */
785 			regval = readl(spi_bus->reg_base + SPI_PCI_CTRL_REG_OFFSET(0));
786 			if (!only_sec_inst)
787 				regval |= (BIT(4));
788 			else
789 				regval &= ~(BIT(4));
790 
791 			writel(regval, spi_bus->reg_base + SPI_PCI_CTRL_REG_OFFSET(0));
792 		}
793 
794 		spi_sub_ptr->hw_inst = start++;
795 
796 		if (iter == 1) {
797 			init_completion(&spi_sub_ptr->spi_xfer_done);
798 			/* Initialize Interrupts - SPI_INT */
799 			regval = readl(spi_bus->reg_base +
800 			       SPI_MST_EVENT_MASK_REG_OFFSET(spi_sub_ptr->hw_inst));
801 			regval &= ~SPI_INTR;
802 			writel(regval, spi_bus->reg_base +
803 			       SPI_MST_EVENT_MASK_REG_OFFSET(spi_sub_ptr->hw_inst));
804 			spi_sub_ptr->irq = pci_irq_vector(pdev, iter);
805 			ret = devm_request_irq(&pdev->dev, spi_sub_ptr->irq,
806 					       pci1xxxx_spi_isr, PCI1XXXX_IRQ_FLAGS,
807 					       pci_name(pdev), spi_sub_ptr);
808 			if (ret < 0) {
809 				dev_err(&pdev->dev, "Unable to request irq : %d",
810 					spi_sub_ptr->irq);
811 				ret = -ENODEV;
812 				goto error;
813 			}
814 		}
815 
816 		spi_host = spi_sub_ptr->spi_host;
817 		spi_host->num_chipselect = SPI_CHIP_SEL_COUNT;
818 		spi_host->mode_bits = SPI_MODE_0 | SPI_MODE_3 | SPI_RX_DUAL |
819 				      SPI_TX_DUAL | SPI_LOOP;
820 		spi_host->can_dma = pci1xxxx_spi_can_dma;
821 		spi_host->transfer_one = pci1xxxx_spi_transfer_one;
822 
823 		spi_host->set_cs = pci1xxxx_spi_set_cs;
824 		spi_host->bits_per_word_mask = SPI_BPW_MASK(8);
825 		spi_host->max_speed_hz = PCI1XXXX_SPI_MAX_CLOCK_HZ;
826 		spi_host->min_speed_hz = PCI1XXXX_SPI_MIN_CLOCK_HZ;
827 		spi_host->flags = SPI_CONTROLLER_MUST_TX;
828 		spi_controller_set_devdata(spi_host, spi_sub_ptr);
829 		ret = devm_spi_register_controller(dev, spi_host);
830 		if (ret)
831 			goto error;
832 	}
833 	pci_set_drvdata(pdev, spi_bus);
834 
835 	return 0;
836 
837 error:
838 	pci_release_regions(pdev);
839 	return ret;
840 }
841 
842 static void store_restore_config(struct pci1xxxx_spi *spi_ptr,
843 				 struct pci1xxxx_spi_internal *spi_sub_ptr,
844 				 u8 inst, bool store)
845 {
846 	u32 regval;
847 
848 	if (store) {
849 		regval = readl(spi_ptr->reg_base +
850 			       SPI_MST_CTL_REG_OFFSET(spi_sub_ptr->hw_inst));
851 		regval &= SPI_MST_CTL_DEVSEL_MASK;
852 		spi_sub_ptr->prev_val.dev_sel = (regval >> 25) & 7;
853 		regval = readl(spi_ptr->reg_base +
854 			       SPI_PCI_CTRL_REG_OFFSET(spi_sub_ptr->hw_inst));
855 		regval &= SPI_MSI_VECTOR_SEL_MASK;
856 		spi_sub_ptr->prev_val.msi_vector_sel = (regval >> 4) & 1;
857 	} else {
858 		regval = readl(spi_ptr->reg_base + SPI_MST_CTL_REG_OFFSET(inst));
859 		regval &= ~SPI_MST_CTL_DEVSEL_MASK;
860 		regval |= (spi_sub_ptr->prev_val.dev_sel << 25);
861 		writel(regval,
862 		       spi_ptr->reg_base + SPI_MST_CTL_REG_OFFSET(inst));
863 		writel((spi_sub_ptr->prev_val.msi_vector_sel << 4),
864 			spi_ptr->reg_base + SPI_PCI_CTRL_REG_OFFSET(inst));
865 	}
866 }
867 
868 static int pci1xxxx_spi_resume(struct device *dev)
869 {
870 	struct pci1xxxx_spi *spi_ptr = dev_get_drvdata(dev);
871 	struct pci1xxxx_spi_internal *spi_sub_ptr;
872 	u32 regval = SPI_RESUME_CONFIG;
873 	u8 iter;
874 
875 	for (iter = 0; iter < spi_ptr->total_hw_instances; iter++) {
876 		spi_sub_ptr = spi_ptr->spi_int[iter];
877 		spi_controller_resume(spi_sub_ptr->spi_host);
878 		writel(regval, spi_ptr->reg_base +
879 		       SPI_MST_EVENT_MASK_REG_OFFSET(iter));
880 
881 		/* Restore config at resume */
882 		store_restore_config(spi_ptr, spi_sub_ptr, iter, 0);
883 	}
884 
885 	return 0;
886 }
887 
888 static int pci1xxxx_spi_suspend(struct device *dev)
889 {
890 	struct pci1xxxx_spi *spi_ptr = dev_get_drvdata(dev);
891 	struct pci1xxxx_spi_internal *spi_sub_ptr;
892 	u32 reg1 = SPI_SUSPEND_CONFIG;
893 	u8 iter;
894 
895 	for (iter = 0; iter < spi_ptr->total_hw_instances; iter++) {
896 		spi_sub_ptr = spi_ptr->spi_int[iter];
897 
898 		while (spi_sub_ptr->spi_xfer_in_progress)
899 			msleep(20);
900 
901 		/* Store existing config before suspend */
902 		store_restore_config(spi_ptr, spi_sub_ptr, iter, 1);
903 		spi_controller_suspend(spi_sub_ptr->spi_host);
904 		writel(reg1, spi_ptr->reg_base +
905 		       SPI_MST_EVENT_MASK_REG_OFFSET(iter));
906 	}
907 
908 	return 0;
909 }
910 
911 static DEFINE_SIMPLE_DEV_PM_OPS(spi_pm_ops, pci1xxxx_spi_suspend,
912 				pci1xxxx_spi_resume);
913 
914 static struct pci_driver pci1xxxx_spi_driver = {
915 	.name		= DRV_NAME,
916 	.id_table	= pci1xxxx_spi_pci_id_table,
917 	.probe		= pci1xxxx_spi_probe,
918 	.driver		=	{
919 		.pm = pm_sleep_ptr(&spi_pm_ops),
920 	},
921 };
922 
923 module_pci_driver(pci1xxxx_spi_driver);
924 
925 MODULE_DESCRIPTION("Microchip Technology Inc. pci1xxxx SPI bus driver");
926 MODULE_AUTHOR("Tharun Kumar P<tharunkumar.pasumarthi@microchip.com>");
927 MODULE_AUTHOR("Kumaravel Thiagarajan<kumaravel.thiagarajan@microchip.com>");
928 MODULE_LICENSE("GPL v2");
929