xref: /linux/drivers/spi/spi-pci1xxxx.c (revision 0262163136de813894cb172aa8ccf762b92e5fd7)
1 // SPDX-License-Identifier: GPL-2.0
2 // PCI1xxxx SPI driver
3 // Copyright (C) 2022 Microchip Technology Inc.
4 // Authors: Tharun Kumar P <tharunkumar.pasumarthi@microchip.com>
5 //          Kumaravel Thiagarajan <Kumaravel.Thiagarajan@microchip.com>
6 
7 
8 #include <linux/bitfield.h>
9 #include <linux/delay.h>
10 #include <linux/dma-mapping.h>
11 #include <linux/iopoll.h>
12 #include <linux/irq.h>
13 #include <linux/module.h>
14 #include <linux/msi.h>
15 #include <linux/pci_regs.h>
16 #include <linux/pci.h>
17 #include <linux/spinlock.h>
18 #include <linux/spi/spi.h>
19 #include "internals.h"
20 
21 #define DRV_NAME "spi-pci1xxxx"
22 
23 #define	SYS_FREQ_DEFAULT		(62500000)
24 
25 #define	PCI1XXXX_SPI_MAX_CLOCK_HZ	(30000000)
26 #define	PCI1XXXX_SPI_CLK_25MHZ		(25000000)
27 #define	PCI1XXXX_SPI_CLK_20MHZ		(20000000)
28 #define	PCI1XXXX_SPI_CLK_15MHZ		(15000000)
29 #define	PCI1XXXX_SPI_CLK_12MHZ		(12000000)
30 #define	PCI1XXXX_SPI_CLK_10MHZ		(10000000)
31 #define	PCI1XXXX_SPI_MIN_CLOCK_HZ	(2000000)
32 
33 #define	PCI1XXXX_SPI_BUFFER_SIZE	(320)
34 
35 #define	SPI_MST_CTL_DEVSEL_MASK		(GENMASK(27, 25))
36 #define	SPI_MST_CTL_CMD_LEN_MASK	(GENMASK(16, 8))
37 #define	SPI_MST_CTL_SPEED_MASK		(GENMASK(7, 5))
38 #define	SPI_MSI_VECTOR_SEL_MASK		(GENMASK(4, 4))
39 
40 #define	SPI_MST_CTL_FORCE_CE		(BIT(4))
41 #define	SPI_MST_CTL_MODE_SEL		(BIT(2))
42 #define	SPI_MST_CTL_GO			(BIT(0))
43 
44 #define SPI_PERI_ADDR_BASE		(0x160000)
45 #define SPI_SYSTEM_ADDR_BASE		(0x2000)
46 #define	SPI_MST1_ADDR_BASE		(0x800)
47 
48 #define DEV_REV_REG			(SPI_SYSTEM_ADDR_BASE + 0x00)
49 #define SPI_SYSLOCK_REG			(SPI_SYSTEM_ADDR_BASE + 0xA0)
50 #define SPI_CONFIG_PERI_ENABLE_REG	(SPI_SYSTEM_ADDR_BASE + 0x108)
51 
52 #define SPI_PERI_ENBLE_PF_MASK		(GENMASK(17, 16))
53 #define DEV_REV_MASK			(GENMASK(7, 0))
54 
55 #define SPI_SYSLOCK			BIT(4)
56 #define SPI0				(0)
57 #define SPI1				(1)
58 
59 /* DMA Related Registers */
60 #define SPI_DMA_ADDR_BASE		(0x1000)
61 #define SPI_DMA_GLOBAL_WR_ENGINE_EN	(SPI_DMA_ADDR_BASE + 0x0C)
62 #define SPI_DMA_WR_DOORBELL_REG		(SPI_DMA_ADDR_BASE + 0x10)
63 #define SPI_DMA_GLOBAL_RD_ENGINE_EN	(SPI_DMA_ADDR_BASE + 0x2C)
64 #define SPI_DMA_RD_DOORBELL_REG		(SPI_DMA_ADDR_BASE + 0x30)
65 #define SPI_DMA_INTR_WR_STS		(SPI_DMA_ADDR_BASE + 0x4C)
66 #define SPI_DMA_WR_INT_MASK		(SPI_DMA_ADDR_BASE + 0x54)
67 #define SPI_DMA_INTR_WR_CLR		(SPI_DMA_ADDR_BASE + 0x58)
68 #define SPI_DMA_ERR_WR_STS		(SPI_DMA_ADDR_BASE + 0x5C)
69 #define SPI_DMA_INTR_IMWR_WDONE_LOW	(SPI_DMA_ADDR_BASE + 0x60)
70 #define SPI_DMA_INTR_IMWR_WDONE_HIGH	(SPI_DMA_ADDR_BASE + 0x64)
71 #define SPI_DMA_INTR_IMWR_WABORT_LOW	(SPI_DMA_ADDR_BASE + 0x68)
72 #define SPI_DMA_INTR_IMWR_WABORT_HIGH	(SPI_DMA_ADDR_BASE + 0x6C)
73 #define SPI_DMA_INTR_WR_IMWR_DATA	(SPI_DMA_ADDR_BASE + 0x70)
74 #define SPI_DMA_INTR_RD_STS		(SPI_DMA_ADDR_BASE + 0xA0)
75 #define SPI_DMA_RD_INT_MASK		(SPI_DMA_ADDR_BASE + 0xA8)
76 #define SPI_DMA_INTR_RD_CLR		(SPI_DMA_ADDR_BASE + 0xAC)
77 #define SPI_DMA_ERR_RD_STS		(SPI_DMA_ADDR_BASE + 0xB8)
78 #define SPI_DMA_INTR_IMWR_RDONE_LOW	(SPI_DMA_ADDR_BASE + 0xCC)
79 #define SPI_DMA_INTR_IMWR_RDONE_HIGH	(SPI_DMA_ADDR_BASE + 0xD0)
80 #define SPI_DMA_INTR_IMWR_RABORT_LOW	(SPI_DMA_ADDR_BASE + 0xD4)
81 #define SPI_DMA_INTR_IMWR_RABORT_HIGH	(SPI_DMA_ADDR_BASE + 0xD8)
82 #define SPI_DMA_INTR_RD_IMWR_DATA	(SPI_DMA_ADDR_BASE + 0xDC)
83 
84 #define SPI_DMA_CH0_WR_BASE		(SPI_DMA_ADDR_BASE + 0x200)
85 #define SPI_DMA_CH0_RD_BASE		(SPI_DMA_ADDR_BASE + 0x300)
86 #define SPI_DMA_CH1_WR_BASE		(SPI_DMA_ADDR_BASE + 0x400)
87 #define SPI_DMA_CH1_RD_BASE		(SPI_DMA_ADDR_BASE + 0x500)
88 
89 #define SPI_DMA_CH_CTL1_OFFSET		(0x00)
90 #define SPI_DMA_CH_XFER_LEN_OFFSET	(0x08)
91 #define SPI_DMA_CH_SAR_LO_OFFSET	(0x0C)
92 #define SPI_DMA_CH_SAR_HI_OFFSET	(0x10)
93 #define SPI_DMA_CH_DAR_LO_OFFSET	(0x14)
94 #define SPI_DMA_CH_DAR_HI_OFFSET	(0x18)
95 
96 #define SPI_DMA_CH0_DONE_INT		BIT(0)
97 #define SPI_DMA_CH1_DONE_INT		BIT(1)
98 #define SPI_DMA_CH0_ABORT_INT		BIT(16)
99 #define SPI_DMA_CH1_ABORT_INT		BIT(17)
100 #define SPI_DMA_DONE_INT_MASK(x)	(1 << (x))
101 #define SPI_DMA_ABORT_INT_MASK(x)	(1 << (16 + (x)))
102 #define DMA_CH_CONTROL_LIE		BIT(3)
103 #define DMA_CH_CONTROL_RIE		BIT(4)
104 #define DMA_INTR_EN			(DMA_CH_CONTROL_RIE | DMA_CH_CONTROL_LIE)
105 
106 /* x refers to SPI Host Controller HW instance id in the below macros - 0 or 1 */
107 
108 #define	SPI_MST_CMD_BUF_OFFSET(x)		(((x) * SPI_MST1_ADDR_BASE) + 0x00)
109 #define	SPI_MST_RSP_BUF_OFFSET(x)		(((x) * SPI_MST1_ADDR_BASE) + 0x200)
110 #define	SPI_MST_CTL_REG_OFFSET(x)		(((x) * SPI_MST1_ADDR_BASE) + 0x400)
111 #define	SPI_MST_EVENT_REG_OFFSET(x)		(((x) * SPI_MST1_ADDR_BASE) + 0x420)
112 #define	SPI_MST_EVENT_MASK_REG_OFFSET(x)	(((x) * SPI_MST1_ADDR_BASE) + 0x424)
113 #define	SPI_MST_PAD_CTL_REG_OFFSET(x)		(((x) * SPI_MST1_ADDR_BASE) + 0x460)
114 #define	SPIALERT_MST_DB_REG_OFFSET(x)		(((x) * SPI_MST1_ADDR_BASE) + 0x464)
115 #define	SPIALERT_MST_VAL_REG_OFFSET(x)		(((x) * SPI_MST1_ADDR_BASE) + 0x468)
116 #define	SPI_PCI_CTRL_REG_OFFSET(x)		(((x) * SPI_MST1_ADDR_BASE) + 0x480)
117 
118 #define PCI1XXXX_IRQ_FLAGS			(IRQF_NO_SUSPEND | IRQF_TRIGGER_NONE)
119 #define SPI_MAX_DATA_LEN			320
120 
121 #define PCI1XXXX_SPI_TIMEOUT			(msecs_to_jiffies(100))
122 #define SYSLOCK_RETRY_CNT			(1000)
123 #define SPI_DMA_ENGINE_EN			(0x1)
124 #define SPI_DMA_ENGINE_DIS			(0x0)
125 
126 #define SPI_INTR		BIT(8)
127 #define SPI_FORCE_CE		BIT(4)
128 
129 #define SPI_CHIP_SEL_COUNT 7
130 #define VENDOR_ID_MCHP 0x1055
131 
132 #define SPI_SUSPEND_CONFIG 0x101
133 #define SPI_RESUME_CONFIG 0x203
134 
135 #define NUM_VEC_PER_INST 3
136 
137 struct pci1xxxx_spi_internal {
138 	u8 hw_inst;
139 	u8 clkdiv;
140 	int irq[NUM_VEC_PER_INST];
141 	int mode;
142 	bool spi_xfer_in_progress;
143 	atomic_t dma_completion_count;
144 	void *rx_buf;
145 	bool dma_aborted_rd;
146 	u32 bytes_recvd;
147 	u32 tx_sgl_len;
148 	u32 rx_sgl_len;
149 	struct scatterlist *tx_sgl, *rx_sgl;
150 	bool dma_aborted_wr;
151 	struct completion spi_xfer_done;
152 	struct spi_controller *spi_host;
153 	struct pci1xxxx_spi *parent;
154 	struct spi_transfer *xfer;
155 	struct {
156 		unsigned int dev_sel : 3;
157 		unsigned int msi_vector_sel : 1;
158 	} prev_val;
159 };
160 
161 struct pci1xxxx_spi {
162 	struct pci_dev *dev;
163 	u8 total_hw_instances;
164 	u8 dev_rev;
165 	void __iomem *reg_base;
166 	void __iomem *dma_offset_bar;
167 	/* lock to safely access the DMA RD registers in isr */
168 	spinlock_t dma_rd_reg_lock;
169 	/* lock to safely access the DMA RD registers in isr */
170 	spinlock_t dma_wr_reg_lock;
171 	bool can_dma;
172 	struct pci1xxxx_spi_internal *spi_int[] __counted_by(total_hw_instances);
173 };
174 
175 static const struct pci_device_id pci1xxxx_spi_pci_id_table[] = {
176 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa004, PCI_ANY_ID, 0x0001), 0, 0, 0x02},
177 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa004, PCI_ANY_ID, 0x0002), 0, 0, 0x01},
178 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa004, PCI_ANY_ID, 0x0003), 0, 0, 0x11},
179 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa004, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01},
180 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa014, PCI_ANY_ID, 0x0001), 0, 0, 0x02},
181 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa014, PCI_ANY_ID, 0x0002), 0, 0, 0x01},
182 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa014, PCI_ANY_ID, 0x0003), 0, 0, 0x11},
183 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa014, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01},
184 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa024, PCI_ANY_ID, 0x0001), 0, 0, 0x02},
185 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa024, PCI_ANY_ID, 0x0002), 0, 0, 0x01},
186 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa024, PCI_ANY_ID, 0x0003), 0, 0, 0x11},
187 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa024, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01},
188 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa034, PCI_ANY_ID, 0x0001), 0, 0, 0x02},
189 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa034, PCI_ANY_ID, 0x0002), 0, 0, 0x01},
190 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa034, PCI_ANY_ID, 0x0003), 0, 0, 0x11},
191 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa034, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01},
192 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa044, PCI_ANY_ID, 0x0001), 0, 0, 0x02},
193 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa044, PCI_ANY_ID, 0x0002), 0, 0, 0x01},
194 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa044, PCI_ANY_ID, 0x0003), 0, 0, 0x11},
195 	{ PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa044, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01},
196 	{ 0, }
197 };
198 
199 MODULE_DEVICE_TABLE(pci, pci1xxxx_spi_pci_id_table);
200 
201 static irqreturn_t pci1xxxx_spi_isr_dma_rd(int irq, void *dev);
202 static irqreturn_t pci1xxxx_spi_isr_dma_wr(int irq, void *dev);
203 
pci1xxxx_set_sys_lock(struct pci1xxxx_spi * par)204 static int pci1xxxx_set_sys_lock(struct pci1xxxx_spi *par)
205 {
206 	writel(SPI_SYSLOCK, par->reg_base + SPI_SYSLOCK_REG);
207 	return readl(par->reg_base + SPI_SYSLOCK_REG);
208 }
209 
pci1xxxx_acquire_sys_lock(struct pci1xxxx_spi * par)210 static int pci1xxxx_acquire_sys_lock(struct pci1xxxx_spi *par)
211 {
212 	u32 regval;
213 
214 	return readx_poll_timeout(pci1xxxx_set_sys_lock, par, regval,
215 			   (regval & SPI_SYSLOCK), 100,
216 			   SYSLOCK_RETRY_CNT * 100);
217 }
218 
pci1xxxx_release_sys_lock(struct pci1xxxx_spi * par)219 static void pci1xxxx_release_sys_lock(struct pci1xxxx_spi *par)
220 {
221 	writel(0x0, par->reg_base + SPI_SYSLOCK_REG);
222 }
223 
pci1xxxx_check_spi_can_dma(struct pci1xxxx_spi * spi_bus,int hw_inst,int num_vector)224 static int pci1xxxx_check_spi_can_dma(struct pci1xxxx_spi *spi_bus, int hw_inst, int num_vector)
225 {
226 	struct pci_dev *pdev = spi_bus->dev;
227 	u32 pf_num;
228 	u32 regval;
229 	int ret;
230 
231 	if (num_vector != hw_inst * NUM_VEC_PER_INST)
232 		return -EOPNOTSUPP;
233 
234 	/*
235 	 * DEV REV Registers is a system register, HW Syslock bit
236 	 * should be acquired before accessing the register
237 	 */
238 	ret = pci1xxxx_acquire_sys_lock(spi_bus);
239 	if (ret) {
240 		dev_err(&pdev->dev, "Error failed to acquire syslock\n");
241 		return ret;
242 	}
243 
244 	regval = readl(spi_bus->reg_base + DEV_REV_REG);
245 	spi_bus->dev_rev = regval & DEV_REV_MASK;
246 	if (spi_bus->dev_rev >= 0xC0) {
247 		regval = readl(spi_bus->reg_base +
248 			       SPI_CONFIG_PERI_ENABLE_REG);
249 		pf_num = regval & SPI_PERI_ENBLE_PF_MASK;
250 	}
251 
252 	pci1xxxx_release_sys_lock(spi_bus);
253 
254 	/*
255 	 * DMA is supported only from C0 and SPI can use DMA only if
256 	 * it is mapped to PF0
257 	 */
258 	if (spi_bus->dev_rev < 0xC0 || pf_num)
259 		return -EOPNOTSUPP;
260 
261 	spi_bus->dma_offset_bar = pcim_iomap(pdev, 2, pci_resource_len(pdev, 2));
262 	if (!spi_bus->dma_offset_bar) {
263 		dev_warn(&pdev->dev, "Error failed to map dma bar, will operate in PIO mode\n");
264 		return -EOPNOTSUPP;
265 	}
266 
267 	if (dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(64))) {
268 		dev_warn(&pdev->dev, "Error failed to set DMA mask, will operate in PIO mode\n");
269 		pcim_iounmap(pdev, spi_bus->dma_offset_bar);
270 		spi_bus->dma_offset_bar = NULL;
271 		return -EOPNOTSUPP;
272 	}
273 
274 	return 0;
275 }
276 
pci1xxxx_spi_dma_config(struct pci1xxxx_spi * spi_bus)277 static void pci1xxxx_spi_dma_config(struct pci1xxxx_spi *spi_bus)
278 {
279 	struct pci1xxxx_spi_internal *spi_sub_ptr;
280 	u8 iter, irq_index;
281 	struct msi_msg msi;
282 	u32 regval;
283 	u16 data;
284 
285 	irq_index = spi_bus->total_hw_instances;
286 	for (iter = 0; iter < spi_bus->total_hw_instances; iter++) {
287 		spi_sub_ptr = spi_bus->spi_int[iter];
288 		get_cached_msi_msg(spi_sub_ptr->irq[1], &msi);
289 		if (iter == 0) {
290 			writel(msi.address_hi, spi_bus->dma_offset_bar +
291 			       SPI_DMA_INTR_IMWR_WDONE_HIGH);
292 			writel(msi.address_hi, spi_bus->dma_offset_bar +
293 			       SPI_DMA_INTR_IMWR_WABORT_HIGH);
294 			writel(msi.address_hi, spi_bus->dma_offset_bar +
295 			       SPI_DMA_INTR_IMWR_RDONE_HIGH);
296 			writel(msi.address_hi, spi_bus->dma_offset_bar +
297 			       SPI_DMA_INTR_IMWR_RABORT_HIGH);
298 			writel(msi.address_lo, spi_bus->dma_offset_bar +
299 			       SPI_DMA_INTR_IMWR_WDONE_LOW);
300 			writel(msi.address_lo, spi_bus->dma_offset_bar +
301 			       SPI_DMA_INTR_IMWR_WABORT_LOW);
302 			writel(msi.address_lo, spi_bus->dma_offset_bar +
303 			       SPI_DMA_INTR_IMWR_RDONE_LOW);
304 			writel(msi.address_lo, spi_bus->dma_offset_bar +
305 			       SPI_DMA_INTR_IMWR_RABORT_LOW);
306 			writel(0, spi_bus->dma_offset_bar + SPI_DMA_INTR_WR_IMWR_DATA);
307 			writel(0, spi_bus->dma_offset_bar + SPI_DMA_INTR_RD_IMWR_DATA);
308 		}
309 		regval = readl(spi_bus->dma_offset_bar + SPI_DMA_INTR_WR_IMWR_DATA);
310 		data = msi.data + irq_index;
311 		writel((regval | (data << (iter * 16))), spi_bus->dma_offset_bar +
312 		       SPI_DMA_INTR_WR_IMWR_DATA);
313 		regval = readl(spi_bus->dma_offset_bar + SPI_DMA_INTR_WR_IMWR_DATA);
314 		irq_index++;
315 
316 		data = msi.data + irq_index;
317 		regval = readl(spi_bus->dma_offset_bar + SPI_DMA_INTR_RD_IMWR_DATA);
318 		writel(regval | (data << (iter * 16)), spi_bus->dma_offset_bar +
319 		       SPI_DMA_INTR_RD_IMWR_DATA);
320 		regval = readl(spi_bus->dma_offset_bar + SPI_DMA_INTR_RD_IMWR_DATA);
321 		irq_index++;
322 	}
323 }
324 
pci1xxxx_spi_dma_init(struct pci1xxxx_spi * spi_bus,int hw_inst,int num_vector)325 static int pci1xxxx_spi_dma_init(struct pci1xxxx_spi *spi_bus, int hw_inst, int num_vector)
326 {
327 	struct pci1xxxx_spi_internal *spi_sub_ptr;
328 	u8 iter, irq_index;
329 	int ret;
330 
331 	irq_index = hw_inst;
332 	ret = pci1xxxx_check_spi_can_dma(spi_bus, hw_inst, num_vector);
333 	if (ret)
334 		return ret;
335 
336 	spin_lock_init(&spi_bus->dma_rd_reg_lock);
337 	spin_lock_init(&spi_bus->dma_wr_reg_lock);
338 	writel(SPI_DMA_ENGINE_EN, spi_bus->dma_offset_bar + SPI_DMA_GLOBAL_WR_ENGINE_EN);
339 	writel(SPI_DMA_ENGINE_EN, spi_bus->dma_offset_bar + SPI_DMA_GLOBAL_RD_ENGINE_EN);
340 
341 	for (iter = 0; iter < hw_inst; iter++) {
342 		spi_sub_ptr = spi_bus->spi_int[iter];
343 		spi_sub_ptr->irq[1] = pci_irq_vector(spi_bus->dev, irq_index);
344 		ret = devm_request_irq(&spi_bus->dev->dev, spi_sub_ptr->irq[1],
345 				       pci1xxxx_spi_isr_dma_wr, PCI1XXXX_IRQ_FLAGS,
346 				       pci_name(spi_bus->dev), spi_sub_ptr);
347 		if (ret < 0)
348 			return ret;
349 
350 		irq_index++;
351 
352 		spi_sub_ptr->irq[2] = pci_irq_vector(spi_bus->dev, irq_index);
353 		ret = devm_request_irq(&spi_bus->dev->dev, spi_sub_ptr->irq[2],
354 				       pci1xxxx_spi_isr_dma_rd, PCI1XXXX_IRQ_FLAGS,
355 				       pci_name(spi_bus->dev), spi_sub_ptr);
356 		if (ret < 0)
357 			return ret;
358 
359 		irq_index++;
360 	}
361 	pci1xxxx_spi_dma_config(spi_bus);
362 	dma_set_max_seg_size(&spi_bus->dev->dev, PCI1XXXX_SPI_BUFFER_SIZE);
363 	spi_bus->can_dma = true;
364 	return 0;
365 }
366 
pci1xxxx_spi_set_cs(struct spi_device * spi,bool enable)367 static void pci1xxxx_spi_set_cs(struct spi_device *spi, bool enable)
368 {
369 	struct pci1xxxx_spi_internal *p = spi_controller_get_devdata(spi->controller);
370 	struct pci1xxxx_spi *par = p->parent;
371 	u32 regval;
372 
373 	/* Set the DEV_SEL bits of the SPI_MST_CTL_REG */
374 	regval = readl(par->reg_base + SPI_MST_CTL_REG_OFFSET(p->hw_inst));
375 	if (!enable) {
376 		regval |= SPI_FORCE_CE;
377 		regval &= ~SPI_MST_CTL_DEVSEL_MASK;
378 		regval |= (spi_get_chipselect(spi, 0) << 25);
379 	} else {
380 		regval &= ~SPI_FORCE_CE;
381 	}
382 	writel(regval, par->reg_base + SPI_MST_CTL_REG_OFFSET(p->hw_inst));
383 }
384 
pci1xxxx_get_clock_div(struct pci1xxxx_spi * par,u32 hz)385 static u8 pci1xxxx_get_clock_div(struct pci1xxxx_spi *par, u32 hz)
386 {
387 	u8 val = 0;
388 
389 	if (hz >= PCI1XXXX_SPI_MAX_CLOCK_HZ)
390 		val = 2;
391 	else if (par->dev_rev >= 0xC0 && hz >= PCI1XXXX_SPI_CLK_25MHZ)
392 		val = 1;
393 	else if ((hz < PCI1XXXX_SPI_MAX_CLOCK_HZ) && (hz >= PCI1XXXX_SPI_CLK_20MHZ))
394 		val = 3;
395 	else if ((hz < PCI1XXXX_SPI_CLK_20MHZ) && (hz >= PCI1XXXX_SPI_CLK_15MHZ))
396 		val = 4;
397 	else if ((hz < PCI1XXXX_SPI_CLK_15MHZ) && (hz >= PCI1XXXX_SPI_CLK_12MHZ))
398 		val = 5;
399 	else if ((hz < PCI1XXXX_SPI_CLK_12MHZ) && (hz >= PCI1XXXX_SPI_CLK_10MHZ))
400 		val = 6;
401 	else if ((hz < PCI1XXXX_SPI_CLK_10MHZ) && (hz >= PCI1XXXX_SPI_MIN_CLOCK_HZ))
402 		val = 7;
403 	else
404 		val = 2;
405 
406 	return val;
407 }
408 
pci1xxxx_spi_setup_dma_to_io(struct pci1xxxx_spi_internal * p,dma_addr_t dma_addr,u32 len)409 static void pci1xxxx_spi_setup_dma_to_io(struct pci1xxxx_spi_internal *p,
410 					 dma_addr_t dma_addr, u32 len)
411 {
412 	void __iomem *base;
413 
414 	if (!p->hw_inst)
415 		base = p->parent->dma_offset_bar + SPI_DMA_CH0_RD_BASE;
416 	else
417 		base = p->parent->dma_offset_bar + SPI_DMA_CH1_RD_BASE;
418 
419 	writel(DMA_INTR_EN, base + SPI_DMA_CH_CTL1_OFFSET);
420 	writel(len, base + SPI_DMA_CH_XFER_LEN_OFFSET);
421 	writel(lower_32_bits(dma_addr), base + SPI_DMA_CH_SAR_LO_OFFSET);
422 	writel(upper_32_bits(dma_addr), base + SPI_DMA_CH_SAR_HI_OFFSET);
423 	/* Updated SPI Command Registers */
424 	writel(lower_32_bits(SPI_PERI_ADDR_BASE + SPI_MST_CMD_BUF_OFFSET(p->hw_inst)),
425 	       base + SPI_DMA_CH_DAR_LO_OFFSET);
426 	writel(upper_32_bits(SPI_PERI_ADDR_BASE + SPI_MST_CMD_BUF_OFFSET(p->hw_inst)),
427 	       base + SPI_DMA_CH_DAR_HI_OFFSET);
428 }
429 
pci1xxxx_spi_setup_dma_from_io(struct pci1xxxx_spi_internal * p,dma_addr_t dma_addr,u32 len)430 static void pci1xxxx_spi_setup_dma_from_io(struct pci1xxxx_spi_internal *p,
431 					   dma_addr_t dma_addr, u32 len)
432 {
433 	void *base;
434 
435 	if (!p->hw_inst)
436 		base = p->parent->dma_offset_bar + SPI_DMA_CH0_WR_BASE;
437 	else
438 		base = p->parent->dma_offset_bar + SPI_DMA_CH1_WR_BASE;
439 
440 	writel(DMA_INTR_EN, base + SPI_DMA_CH_CTL1_OFFSET);
441 	writel(len, base + SPI_DMA_CH_XFER_LEN_OFFSET);
442 	writel(lower_32_bits(dma_addr), base + SPI_DMA_CH_DAR_LO_OFFSET);
443 	writel(upper_32_bits(dma_addr), base + SPI_DMA_CH_DAR_HI_OFFSET);
444 	writel(lower_32_bits(SPI_PERI_ADDR_BASE + SPI_MST_RSP_BUF_OFFSET(p->hw_inst)),
445 	       base + SPI_DMA_CH_SAR_LO_OFFSET);
446 	writel(upper_32_bits(SPI_PERI_ADDR_BASE + SPI_MST_RSP_BUF_OFFSET(p->hw_inst)),
447 	       base + SPI_DMA_CH_SAR_HI_OFFSET);
448 }
449 
pci1xxxx_spi_setup(struct pci1xxxx_spi * par,u8 hw_inst,u32 mode,u8 clkdiv,u32 len)450 static void pci1xxxx_spi_setup(struct pci1xxxx_spi *par, u8 hw_inst, u32 mode,
451 			       u8 clkdiv, u32 len)
452 {
453 	u32 regval;
454 
455 	regval = readl(par->reg_base + SPI_MST_CTL_REG_OFFSET(hw_inst));
456 	regval &= ~(SPI_MST_CTL_MODE_SEL | SPI_MST_CTL_CMD_LEN_MASK |
457 		    SPI_MST_CTL_SPEED_MASK);
458 
459 	if (mode == SPI_MODE_3)
460 		regval |= SPI_MST_CTL_MODE_SEL;
461 
462 	regval |= FIELD_PREP(SPI_MST_CTL_CMD_LEN_MASK, len);
463 	regval |= FIELD_PREP(SPI_MST_CTL_SPEED_MASK, clkdiv);
464 	writel(regval, par->reg_base + SPI_MST_CTL_REG_OFFSET(hw_inst));
465 }
466 
pci1xxxx_start_spi_xfer(struct pci1xxxx_spi_internal * p)467 static void pci1xxxx_start_spi_xfer(struct pci1xxxx_spi_internal *p)
468 {
469 	u32 regval;
470 
471 	atomic_set(&p->dma_completion_count, 0);
472 	regval = readl(p->parent->reg_base + SPI_MST_CTL_REG_OFFSET(p->hw_inst));
473 	regval |= SPI_MST_CTL_GO;
474 	writel(regval, p->parent->reg_base + SPI_MST_CTL_REG_OFFSET(p->hw_inst));
475 }
476 
pci1xxxx_spi_transfer_with_io(struct spi_controller * spi_ctlr,struct spi_device * spi,struct spi_transfer * xfer)477 static int pci1xxxx_spi_transfer_with_io(struct spi_controller *spi_ctlr,
478 					 struct spi_device *spi, struct spi_transfer *xfer)
479 {
480 	struct pci1xxxx_spi_internal *p = spi_controller_get_devdata(spi_ctlr);
481 	struct pci1xxxx_spi *par = p->parent;
482 	int len, loop_iter, transfer_len;
483 	unsigned long bytes_transfered;
484 	unsigned long bytes_recvd;
485 	unsigned long loop_count;
486 	u8 *rx_buf, result;
487 	const u8 *tx_buf;
488 	u32 regval;
489 	u8 clkdiv;
490 
491 	p->spi_xfer_in_progress = true;
492 	p->bytes_recvd = 0;
493 	clkdiv = pci1xxxx_get_clock_div(par, xfer->speed_hz);
494 	tx_buf = xfer->tx_buf;
495 	rx_buf = xfer->rx_buf;
496 	transfer_len = xfer->len;
497 	regval = readl(par->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
498 	writel(regval, par->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
499 
500 	if (tx_buf) {
501 		bytes_transfered = 0;
502 		bytes_recvd = 0;
503 		loop_count = transfer_len / SPI_MAX_DATA_LEN;
504 		if (transfer_len % SPI_MAX_DATA_LEN != 0)
505 			loop_count += 1;
506 
507 		for (loop_iter = 0; loop_iter < loop_count; loop_iter++) {
508 			len = SPI_MAX_DATA_LEN;
509 			if ((transfer_len % SPI_MAX_DATA_LEN != 0) &&
510 			    (loop_iter == loop_count - 1))
511 				len = transfer_len % SPI_MAX_DATA_LEN;
512 
513 			reinit_completion(&p->spi_xfer_done);
514 			memcpy_toio(par->reg_base + SPI_MST_CMD_BUF_OFFSET(p->hw_inst),
515 				    &tx_buf[bytes_transfered], len);
516 			bytes_transfered += len;
517 			pci1xxxx_spi_setup(par, p->hw_inst, spi->mode, clkdiv, len);
518 			pci1xxxx_start_spi_xfer(p);
519 
520 			/* Wait for DMA_TERM interrupt */
521 			result = wait_for_completion_timeout(&p->spi_xfer_done,
522 							     PCI1XXXX_SPI_TIMEOUT);
523 			if (!result)
524 				return -ETIMEDOUT;
525 
526 			if (rx_buf) {
527 				memcpy_fromio(&rx_buf[bytes_recvd], par->reg_base +
528 					      SPI_MST_RSP_BUF_OFFSET(p->hw_inst), len);
529 				bytes_recvd += len;
530 			}
531 		}
532 	}
533 	p->spi_xfer_in_progress = false;
534 
535 	return 0;
536 }
537 
pci1xxxx_spi_transfer_with_dma(struct spi_controller * spi_ctlr,struct spi_device * spi,struct spi_transfer * xfer)538 static int pci1xxxx_spi_transfer_with_dma(struct spi_controller *spi_ctlr,
539 					  struct spi_device *spi,
540 					  struct spi_transfer *xfer)
541 {
542 	struct pci1xxxx_spi_internal *p = spi_controller_get_devdata(spi_ctlr);
543 	struct pci1xxxx_spi *par = p->parent;
544 	dma_addr_t tx_dma_addr = 0;
545 	int ret = 0;
546 	u32 regval;
547 
548 	p->spi_xfer_in_progress = true;
549 	p->tx_sgl = xfer->tx_sg.sgl;
550 	p->rx_sgl = xfer->rx_sg.sgl;
551 	p->rx_buf = xfer->rx_buf;
552 	atomic_set(&p->dma_completion_count, 1);
553 	regval = readl(par->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
554 	writel(regval, par->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
555 
556 	if (!xfer->tx_buf || !p->tx_sgl) {
557 		ret = -EINVAL;
558 		goto error;
559 	}
560 	p->xfer = xfer;
561 	p->mode = spi->mode;
562 	p->clkdiv = pci1xxxx_get_clock_div(par, xfer->speed_hz);
563 	p->bytes_recvd = 0;
564 	p->rx_buf = xfer->rx_buf;
565 	regval = readl(par->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
566 	writel(regval, par->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
567 
568 	tx_dma_addr = sg_dma_address(p->tx_sgl);
569 	p->tx_sgl_len = sg_dma_len(p->tx_sgl);
570 	pci1xxxx_spi_setup(par, p->hw_inst, p->mode, p->clkdiv, p->tx_sgl_len);
571 	pci1xxxx_spi_setup_dma_to_io(p, (tx_dma_addr), p->tx_sgl_len);
572 	writel(p->hw_inst, par->dma_offset_bar + SPI_DMA_RD_DOORBELL_REG);
573 
574 	reinit_completion(&p->spi_xfer_done);
575 	/* Wait for DMA_TERM interrupt */
576 	ret = wait_for_completion_timeout(&p->spi_xfer_done, PCI1XXXX_SPI_TIMEOUT);
577 	if (!ret) {
578 		ret = -ETIMEDOUT;
579 		if (p->dma_aborted_rd) {
580 			writel(SPI_DMA_ENGINE_DIS,
581 			       par->dma_offset_bar + SPI_DMA_GLOBAL_RD_ENGINE_EN);
582 			/*
583 			 * DMA ENGINE reset takes time if any TLP
584 			 * completeion in progress, should wait
585 			 * till DMA Engine reset is completed.
586 			 */
587 			ret = readl_poll_timeout(par->dma_offset_bar +
588 						 SPI_DMA_GLOBAL_RD_ENGINE_EN, regval,
589 						 (regval == 0x0), 0, USEC_PER_MSEC);
590 			if (ret) {
591 				ret = -ECANCELED;
592 				goto error;
593 			}
594 			writel(SPI_DMA_ENGINE_EN,
595 			       par->dma_offset_bar + SPI_DMA_GLOBAL_RD_ENGINE_EN);
596 			p->dma_aborted_rd = false;
597 			ret = -ECANCELED;
598 		}
599 		if (p->dma_aborted_wr) {
600 			writel(SPI_DMA_ENGINE_DIS,
601 			       par->dma_offset_bar + SPI_DMA_GLOBAL_WR_ENGINE_EN);
602 
603 			/*
604 			 * DMA ENGINE reset takes time if any TLP
605 			 * completeion in progress, should wait
606 			 * till DMA Engine reset is completed.
607 			 */
608 			ret = readl_poll_timeout(par->dma_offset_bar +
609 						 SPI_DMA_GLOBAL_WR_ENGINE_EN, regval,
610 						 (regval == 0x0), 0, USEC_PER_MSEC);
611 			if (ret) {
612 				ret = -ECANCELED;
613 				goto error;
614 			}
615 
616 			writel(SPI_DMA_ENGINE_EN,
617 			       par->dma_offset_bar + SPI_DMA_GLOBAL_WR_ENGINE_EN);
618 			p->dma_aborted_wr = false;
619 			ret = -ECANCELED;
620 		}
621 		goto error;
622 	}
623 	ret = 0;
624 
625 error:
626 	p->spi_xfer_in_progress = false;
627 
628 	return ret;
629 }
630 
pci1xxxx_spi_transfer_one(struct spi_controller * spi_ctlr,struct spi_device * spi,struct spi_transfer * xfer)631 static int pci1xxxx_spi_transfer_one(struct spi_controller *spi_ctlr,
632 				     struct spi_device *spi, struct spi_transfer *xfer)
633 {
634 	if (spi_xfer_is_dma_mapped(spi_ctlr, spi, xfer))
635 		return pci1xxxx_spi_transfer_with_dma(spi_ctlr, spi, xfer);
636 	else
637 		return pci1xxxx_spi_transfer_with_io(spi_ctlr, spi, xfer);
638 }
639 
pci1xxxx_spi_isr_io(int irq,void * dev)640 static irqreturn_t pci1xxxx_spi_isr_io(int irq, void *dev)
641 {
642 	struct pci1xxxx_spi_internal *p = dev;
643 	irqreturn_t spi_int_fired = IRQ_NONE;
644 	u32 regval;
645 
646 	/* Clear the SPI GO_BIT Interrupt */
647 	regval = readl(p->parent->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
648 	if (regval & SPI_INTR) {
649 		/* Clear xfer_done */
650 		if (p->parent->can_dma && p->rx_buf)
651 			writel(p->hw_inst, p->parent->dma_offset_bar +
652 			       SPI_DMA_WR_DOORBELL_REG);
653 		else
654 			complete(&p->parent->spi_int[p->hw_inst]->spi_xfer_done);
655 		spi_int_fired = IRQ_HANDLED;
656 	}
657 	writel(regval, p->parent->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
658 	return spi_int_fired;
659 }
660 
pci1xxxx_spi_setup_next_dma_to_io_transfer(struct pci1xxxx_spi_internal * p)661 static void pci1xxxx_spi_setup_next_dma_to_io_transfer(struct pci1xxxx_spi_internal *p)
662 {
663 	dma_addr_t tx_dma_addr = 0;
664 	u32 prev_len;
665 
666 	p->tx_sgl = sg_next(p->tx_sgl);
667 	if (p->tx_sgl) {
668 		tx_dma_addr = sg_dma_address(p->tx_sgl);
669 		prev_len = p->tx_sgl_len;
670 		p->tx_sgl_len = sg_dma_len(p->tx_sgl);
671 		pci1xxxx_spi_setup_dma_to_io(p, tx_dma_addr, p->tx_sgl_len);
672 		writel(p->hw_inst, p->parent->dma_offset_bar + SPI_DMA_RD_DOORBELL_REG);
673 		if (prev_len != p->tx_sgl_len)
674 			pci1xxxx_spi_setup(p->parent,
675 					   p->hw_inst, p->mode, p->clkdiv, p->tx_sgl_len);
676 	}
677 }
678 
pci1xxxx_spi_setup_next_dma_from_io_transfer(struct pci1xxxx_spi_internal * p)679 static void pci1xxxx_spi_setup_next_dma_from_io_transfer(struct pci1xxxx_spi_internal *p)
680 {
681 	dma_addr_t rx_dma_addr = 0;
682 
683 	if (p->rx_sgl) {
684 		rx_dma_addr = sg_dma_address(p->rx_sgl);
685 		p->rx_sgl_len = sg_dma_len(p->rx_sgl);
686 		pci1xxxx_spi_setup_dma_from_io(p, rx_dma_addr, p->rx_sgl_len);
687 		writel(p->hw_inst, p->parent->dma_offset_bar + SPI_DMA_WR_DOORBELL_REG);
688 	}
689 }
690 
pci1xxxx_spi_isr_dma_rd(int irq,void * dev)691 static irqreturn_t pci1xxxx_spi_isr_dma_rd(int irq, void *dev)
692 {
693 	struct pci1xxxx_spi_internal *p = dev;
694 	irqreturn_t spi_int_fired = IRQ_NONE;
695 	unsigned long flags;
696 	u32 regval;
697 
698 	/* Clear the DMA RD INT and start spi xfer*/
699 	regval = readl(p->parent->dma_offset_bar + SPI_DMA_INTR_RD_STS);
700 	if (regval) {
701 		if (regval & SPI_DMA_DONE_INT_MASK(p->hw_inst)) {
702 			/* Start the SPI transfer only if both DMA read and write are completed */
703 			if (atomic_inc_return(&p->dma_completion_count) == 2)
704 				pci1xxxx_start_spi_xfer(p);
705 			spi_int_fired = IRQ_HANDLED;
706 		}
707 		if (regval & SPI_DMA_ABORT_INT_MASK(p->hw_inst)) {
708 			p->dma_aborted_rd = true;
709 			spi_int_fired = IRQ_HANDLED;
710 		}
711 		spin_lock_irqsave(&p->parent->dma_rd_reg_lock, flags);
712 		writel((SPI_DMA_DONE_INT_MASK(p->hw_inst) | SPI_DMA_ABORT_INT_MASK(p->hw_inst)),
713 		       p->parent->dma_offset_bar + SPI_DMA_INTR_RD_CLR);
714 		spin_unlock_irqrestore(&p->parent->dma_rd_reg_lock, flags);
715 	}
716 	return spi_int_fired;
717 }
718 
pci1xxxx_spi_isr_dma_wr(int irq,void * dev)719 static irqreturn_t pci1xxxx_spi_isr_dma_wr(int irq, void *dev)
720 {
721 	struct pci1xxxx_spi_internal *p = dev;
722 	irqreturn_t spi_int_fired = IRQ_NONE;
723 	unsigned long flags;
724 	u32 regval;
725 
726 	/* Clear the DMA WR INT */
727 	regval = readl(p->parent->dma_offset_bar + SPI_DMA_INTR_WR_STS);
728 	if (regval) {
729 		if (regval & SPI_DMA_DONE_INT_MASK(p->hw_inst)) {
730 			spi_int_fired = IRQ_HANDLED;
731 			if (sg_is_last(p->rx_sgl)) {
732 				complete(&p->spi_xfer_done);
733 			} else {
734 				p->rx_sgl =  sg_next(p->rx_sgl);
735 				if (atomic_inc_return(&p->dma_completion_count) == 2)
736 					pci1xxxx_start_spi_xfer(p);
737 			}
738 
739 		}
740 		if (regval & SPI_DMA_ABORT_INT_MASK(p->hw_inst)) {
741 			p->dma_aborted_wr = true;
742 			spi_int_fired = IRQ_HANDLED;
743 		}
744 		spin_lock_irqsave(&p->parent->dma_wr_reg_lock, flags);
745 		writel((SPI_DMA_DONE_INT_MASK(p->hw_inst) | SPI_DMA_ABORT_INT_MASK(p->hw_inst)),
746 		       p->parent->dma_offset_bar + SPI_DMA_INTR_WR_CLR);
747 		spin_unlock_irqrestore(&p->parent->dma_wr_reg_lock, flags);
748 	}
749 	return spi_int_fired;
750 }
751 
pci1xxxx_spi_isr_dma(int irq,void * dev)752 static irqreturn_t pci1xxxx_spi_isr_dma(int irq, void *dev)
753 {
754 	struct pci1xxxx_spi_internal *p = dev;
755 	irqreturn_t spi_int_fired = IRQ_NONE;
756 	u32 regval;
757 
758 	/* Clear the SPI GO_BIT Interrupt */
759 	regval = readl(p->parent->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
760 	if (regval & SPI_INTR) {
761 		pci1xxxx_spi_setup_next_dma_from_io_transfer(p);
762 		pci1xxxx_spi_setup_next_dma_to_io_transfer(p);
763 		spi_int_fired = IRQ_HANDLED;
764 		writel(regval, p->parent->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
765 	}
766 	return spi_int_fired;
767 }
768 
pci1xxxx_spi_isr(int irq,void * dev)769 static irqreturn_t pci1xxxx_spi_isr(int irq, void *dev)
770 {
771 	struct pci1xxxx_spi_internal *p = dev;
772 
773 	if (p->spi_host->can_dma(p->spi_host, NULL, p->xfer))
774 		return pci1xxxx_spi_isr_dma(irq, dev);
775 	else
776 		return pci1xxxx_spi_isr_io(irq, dev);
777 }
778 
pci1xxxx_spi_shared_isr(int irq,void * dev)779 static irqreturn_t pci1xxxx_spi_shared_isr(int irq, void *dev)
780 {
781 	struct pci1xxxx_spi *par = dev;
782 	u8 i = 0;
783 
784 	for (i = 0; i < par->total_hw_instances; i++)
785 		pci1xxxx_spi_isr(irq, par->spi_int[i]);
786 
787 	return IRQ_HANDLED;
788 }
789 
pci1xxxx_spi_can_dma(struct spi_controller * host,struct spi_device * spi,struct spi_transfer * xfer)790 static bool pci1xxxx_spi_can_dma(struct spi_controller *host,
791 				 struct spi_device *spi,
792 				 struct spi_transfer *xfer)
793 {
794 	struct pci1xxxx_spi_internal *p = spi_controller_get_devdata(host);
795 	struct pci1xxxx_spi *par = p->parent;
796 
797 	return par->can_dma;
798 }
799 
pci1xxxx_spi_probe(struct pci_dev * pdev,const struct pci_device_id * ent)800 static int pci1xxxx_spi_probe(struct pci_dev *pdev, const struct pci_device_id *ent)
801 {
802 	u8 hw_inst_cnt, iter, start, only_sec_inst;
803 	struct pci1xxxx_spi_internal *spi_sub_ptr;
804 	struct device *dev = &pdev->dev;
805 	struct pci1xxxx_spi *spi_bus;
806 	struct spi_controller *spi_host;
807 	int num_vector = 0;
808 	u32 regval;
809 	int ret;
810 
811 	hw_inst_cnt = ent->driver_data & 0x0f;
812 	start = (ent->driver_data & 0xf0) >> 4;
813 	if (start == 1)
814 		only_sec_inst = 1;
815 	else
816 		only_sec_inst = 0;
817 
818 	spi_bus = devm_kzalloc(&pdev->dev,
819 			       struct_size(spi_bus, spi_int, hw_inst_cnt),
820 			       GFP_KERNEL);
821 	if (!spi_bus)
822 		return -ENOMEM;
823 
824 	spi_bus->dev = pdev;
825 	spi_bus->total_hw_instances = hw_inst_cnt;
826 	pci_set_master(pdev);
827 
828 	for (iter = 0; iter < hw_inst_cnt; iter++) {
829 		spi_bus->spi_int[iter] = devm_kzalloc(&pdev->dev,
830 						      sizeof(struct pci1xxxx_spi_internal),
831 						      GFP_KERNEL);
832 		if (!spi_bus->spi_int[iter])
833 			return -ENOMEM;
834 		spi_sub_ptr = spi_bus->spi_int[iter];
835 		spi_sub_ptr->spi_host = devm_spi_alloc_host(dev, sizeof(struct spi_controller));
836 		if (!spi_sub_ptr->spi_host)
837 			return -ENOMEM;
838 
839 		spi_sub_ptr->parent = spi_bus;
840 		spi_sub_ptr->spi_xfer_in_progress = false;
841 
842 		if (!iter) {
843 			ret = pcim_enable_device(pdev);
844 			if (ret)
845 				return -ENOMEM;
846 
847 			ret = pcim_request_all_regions(pdev, DRV_NAME);
848 			if (ret)
849 				return -ENOMEM;
850 
851 			spi_bus->reg_base = pcim_iomap(pdev, 0, pci_resource_len(pdev, 0));
852 			if (!spi_bus->reg_base)
853 				return -EINVAL;
854 
855 			num_vector = pci_alloc_irq_vectors(pdev, 1, hw_inst_cnt * NUM_VEC_PER_INST,
856 							   PCI_IRQ_INTX | PCI_IRQ_MSI);
857 			if (num_vector < 0) {
858 				dev_err(&pdev->dev, "Error allocating MSI vectors\n");
859 				return num_vector;
860 			}
861 
862 			init_completion(&spi_sub_ptr->spi_xfer_done);
863 			/* Initialize Interrupts - SPI_INT */
864 			regval = readl(spi_bus->reg_base +
865 				       SPI_MST_EVENT_MASK_REG_OFFSET(spi_sub_ptr->hw_inst));
866 			regval &= ~SPI_INTR;
867 			writel(regval, spi_bus->reg_base +
868 			       SPI_MST_EVENT_MASK_REG_OFFSET(spi_sub_ptr->hw_inst));
869 			spi_sub_ptr->irq[0] = pci_irq_vector(pdev, 0);
870 
871 			if (num_vector >= hw_inst_cnt)
872 				ret = devm_request_irq(&pdev->dev, spi_sub_ptr->irq[0],
873 						       pci1xxxx_spi_isr, PCI1XXXX_IRQ_FLAGS,
874 						       pci_name(pdev), spi_sub_ptr);
875 			else
876 				ret = devm_request_irq(&pdev->dev, spi_sub_ptr->irq[0],
877 						       pci1xxxx_spi_shared_isr,
878 						       PCI1XXXX_IRQ_FLAGS | IRQF_SHARED,
879 						       pci_name(pdev), spi_bus);
880 			if (ret < 0) {
881 				dev_err(&pdev->dev, "Unable to request irq : %d",
882 					spi_sub_ptr->irq[0]);
883 				return -ENODEV;
884 			}
885 
886 			/* This register is only applicable for 1st instance */
887 			regval = readl(spi_bus->reg_base + SPI_PCI_CTRL_REG_OFFSET(0));
888 			if (!only_sec_inst)
889 				regval |= (BIT(4));
890 			else
891 				regval &= ~(BIT(4));
892 
893 			writel(regval, spi_bus->reg_base + SPI_PCI_CTRL_REG_OFFSET(0));
894 		}
895 
896 		spi_sub_ptr->hw_inst = start++;
897 
898 		if (iter == 1) {
899 			init_completion(&spi_sub_ptr->spi_xfer_done);
900 			/* Initialize Interrupts - SPI_INT */
901 			regval = readl(spi_bus->reg_base +
902 			       SPI_MST_EVENT_MASK_REG_OFFSET(spi_sub_ptr->hw_inst));
903 			regval &= ~SPI_INTR;
904 			writel(regval, spi_bus->reg_base +
905 			       SPI_MST_EVENT_MASK_REG_OFFSET(spi_sub_ptr->hw_inst));
906 			if (num_vector >= hw_inst_cnt) {
907 				spi_sub_ptr->irq[0] = pci_irq_vector(pdev, iter);
908 				ret = devm_request_irq(&pdev->dev, spi_sub_ptr->irq[0],
909 						       pci1xxxx_spi_isr, PCI1XXXX_IRQ_FLAGS,
910 						       pci_name(pdev), spi_sub_ptr);
911 				if (ret < 0) {
912 					dev_err(&pdev->dev, "Unable to request irq : %d",
913 						spi_sub_ptr->irq[0]);
914 					return -ENODEV;
915 				}
916 			}
917 		}
918 
919 		spi_host = spi_sub_ptr->spi_host;
920 		spi_host->num_chipselect = SPI_CHIP_SEL_COUNT;
921 		spi_host->mode_bits = SPI_MODE_0 | SPI_MODE_3 | SPI_RX_DUAL |
922 				      SPI_TX_DUAL | SPI_LOOP;
923 		spi_host->can_dma = pci1xxxx_spi_can_dma;
924 		spi_host->transfer_one = pci1xxxx_spi_transfer_one;
925 
926 		spi_host->set_cs = pci1xxxx_spi_set_cs;
927 		spi_host->bits_per_word_mask = SPI_BPW_MASK(8);
928 		spi_host->max_speed_hz = PCI1XXXX_SPI_MAX_CLOCK_HZ;
929 		spi_host->min_speed_hz = PCI1XXXX_SPI_MIN_CLOCK_HZ;
930 		spi_host->flags = SPI_CONTROLLER_MUST_TX;
931 		spi_controller_set_devdata(spi_host, spi_sub_ptr);
932 		ret = devm_spi_register_controller(dev, spi_host);
933 		if (ret)
934 			return ret;
935 	}
936 	ret = pci1xxxx_spi_dma_init(spi_bus, hw_inst_cnt, num_vector);
937 	if (ret && ret != -EOPNOTSUPP)
938 		return ret;
939 
940 	pci_set_drvdata(pdev, spi_bus);
941 
942 	return 0;
943 }
944 
store_restore_config(struct pci1xxxx_spi * spi_ptr,struct pci1xxxx_spi_internal * spi_sub_ptr,u8 inst,bool store)945 static void store_restore_config(struct pci1xxxx_spi *spi_ptr,
946 				 struct pci1xxxx_spi_internal *spi_sub_ptr,
947 				 u8 inst, bool store)
948 {
949 	u32 regval;
950 
951 	if (store) {
952 		regval = readl(spi_ptr->reg_base +
953 			       SPI_MST_CTL_REG_OFFSET(spi_sub_ptr->hw_inst));
954 		regval &= SPI_MST_CTL_DEVSEL_MASK;
955 		spi_sub_ptr->prev_val.dev_sel = (regval >> 25) & 7;
956 		regval = readl(spi_ptr->reg_base +
957 			       SPI_PCI_CTRL_REG_OFFSET(spi_sub_ptr->hw_inst));
958 		regval &= SPI_MSI_VECTOR_SEL_MASK;
959 		spi_sub_ptr->prev_val.msi_vector_sel = (regval >> 4) & 1;
960 	} else {
961 		regval = readl(spi_ptr->reg_base + SPI_MST_CTL_REG_OFFSET(inst));
962 		regval &= ~SPI_MST_CTL_DEVSEL_MASK;
963 		regval |= (spi_sub_ptr->prev_val.dev_sel << 25);
964 		writel(regval,
965 		       spi_ptr->reg_base + SPI_MST_CTL_REG_OFFSET(inst));
966 		writel((spi_sub_ptr->prev_val.msi_vector_sel << 4),
967 			spi_ptr->reg_base + SPI_PCI_CTRL_REG_OFFSET(inst));
968 	}
969 }
970 
pci1xxxx_spi_resume(struct device * dev)971 static int pci1xxxx_spi_resume(struct device *dev)
972 {
973 	struct pci1xxxx_spi *spi_ptr = dev_get_drvdata(dev);
974 	struct pci1xxxx_spi_internal *spi_sub_ptr;
975 	u32 regval = SPI_RESUME_CONFIG;
976 	u8 iter;
977 
978 	for (iter = 0; iter < spi_ptr->total_hw_instances; iter++) {
979 		spi_sub_ptr = spi_ptr->spi_int[iter];
980 		spi_controller_resume(spi_sub_ptr->spi_host);
981 		writel(regval, spi_ptr->reg_base +
982 		       SPI_MST_EVENT_MASK_REG_OFFSET(iter));
983 
984 		/* Restore config at resume */
985 		store_restore_config(spi_ptr, spi_sub_ptr, iter, 0);
986 	}
987 
988 	return 0;
989 }
990 
pci1xxxx_spi_suspend(struct device * dev)991 static int pci1xxxx_spi_suspend(struct device *dev)
992 {
993 	struct pci1xxxx_spi *spi_ptr = dev_get_drvdata(dev);
994 	struct pci1xxxx_spi_internal *spi_sub_ptr;
995 	u32 reg1 = SPI_SUSPEND_CONFIG;
996 	u8 iter;
997 
998 	for (iter = 0; iter < spi_ptr->total_hw_instances; iter++) {
999 		spi_sub_ptr = spi_ptr->spi_int[iter];
1000 
1001 		while (spi_sub_ptr->spi_xfer_in_progress)
1002 			msleep(20);
1003 
1004 		/* Store existing config before suspend */
1005 		store_restore_config(spi_ptr, spi_sub_ptr, iter, 1);
1006 		spi_controller_suspend(spi_sub_ptr->spi_host);
1007 		writel(reg1, spi_ptr->reg_base +
1008 		       SPI_MST_EVENT_MASK_REG_OFFSET(iter));
1009 	}
1010 
1011 	return 0;
1012 }
1013 
1014 static DEFINE_SIMPLE_DEV_PM_OPS(spi_pm_ops, pci1xxxx_spi_suspend,
1015 				pci1xxxx_spi_resume);
1016 
1017 static struct pci_driver pci1xxxx_spi_driver = {
1018 	.name		= DRV_NAME,
1019 	.id_table	= pci1xxxx_spi_pci_id_table,
1020 	.probe		= pci1xxxx_spi_probe,
1021 	.driver		=	{
1022 		.pm = pm_sleep_ptr(&spi_pm_ops),
1023 	},
1024 };
1025 
1026 module_pci_driver(pci1xxxx_spi_driver);
1027 
1028 MODULE_DESCRIPTION("Microchip Technology Inc. pci1xxxx SPI bus driver");
1029 MODULE_AUTHOR("Tharun Kumar P<tharunkumar.pasumarthi@microchip.com>");
1030 MODULE_AUTHOR("Kumaravel Thiagarajan<kumaravel.thiagarajan@microchip.com>");
1031 MODULE_LICENSE("GPL v2");
1032