xref: /linux/drivers/ata/pata_atiixp.c (revision 39f75da7bcc829ddc4d40bb60d0e95520de7898b)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * pata_atiixp.c 	- ATI PATA for new ATA layer
4  *			  (C) 2005 Red Hat Inc
5  *			  (C) 2009-2010 Bartlomiej Zolnierkiewicz
6  *
7  * Based on
8  *
9  *  linux/drivers/ide/pci/atiixp.c	Version 0.01-bart2	Feb. 26, 2004
10  *
11  *  Copyright (C) 2003 ATI Inc. <hyu@ati.com>
12  *  Copyright (C) 2004 Bartlomiej Zolnierkiewicz
13  *
14  */
15 
16 #include <linux/kernel.h>
17 #include <linux/module.h>
18 #include <linux/pci.h>
19 #include <linux/blkdev.h>
20 #include <linux/delay.h>
21 #include <scsi/scsi_host.h>
22 #include <linux/libata.h>
23 #include <linux/dmi.h>
24 
25 #define DRV_NAME "pata_atiixp"
26 #define DRV_VERSION "0.4.6"
27 
28 enum {
29 	ATIIXP_IDE_PIO_TIMING	= 0x40,
30 	ATIIXP_IDE_MWDMA_TIMING	= 0x44,
31 	ATIIXP_IDE_PIO_CONTROL	= 0x48,
32 	ATIIXP_IDE_PIO_MODE	= 0x4a,
33 	ATIIXP_IDE_UDMA_CONTROL	= 0x54,
34 	ATIIXP_IDE_UDMA_MODE 	= 0x56
35 };
36 
37 static const struct dmi_system_id attixp_cable_override_dmi_table[] = {
38 	{
39 		/* Board has onboard PATA<->SATA converters */
40 		.ident = "MSI E350DM-E33",
41 		.matches = {
42 			DMI_MATCH(DMI_BOARD_VENDOR, "MSI"),
43 			DMI_MATCH(DMI_BOARD_NAME, "E350DM-E33(MS-7720)"),
44 		},
45 	},
46 	{ }
47 };
48 
49 static int atiixp_cable_detect(struct ata_port *ap)
50 {
51 	struct pci_dev *pdev = to_pci_dev(ap->host->dev);
52 	u8 udma;
53 
54 	if (dmi_check_system(attixp_cable_override_dmi_table))
55 		return ATA_CBL_PATA40_SHORT;
56 
57 	/* Hack from drivers/ide/pci. Really we want to know how to do the
58 	   raw detection not play follow the bios mode guess */
59 	pci_read_config_byte(pdev, ATIIXP_IDE_UDMA_MODE + ap->port_no, &udma);
60 	if ((udma & 0x07) >= 0x04 || (udma & 0x70) >= 0x40)
61 		return  ATA_CBL_PATA80;
62 	return ATA_CBL_PATA40;
63 }
64 
65 static DEFINE_SPINLOCK(atiixp_lock);
66 
67 /**
68  *	atiixp_prereset	-	perform reset handling
69  *	@link: ATA link
70  *	@deadline: deadline jiffies for the operation
71  *
72  *	Reset sequence checking enable bits to see which ports are
73  *	active.
74  */
75 
76 static int atiixp_prereset(struct ata_link *link, unsigned long deadline)
77 {
78 	static const struct pci_bits atiixp_enable_bits[] = {
79 		{ 0x48, 1, 0x01, 0x00 },
80 		{ 0x48, 1, 0x08, 0x00 }
81 	};
82 
83 	struct ata_port *ap = link->ap;
84 	struct pci_dev *pdev = to_pci_dev(ap->host->dev);
85 
86 	if (!pci_test_config_bits(pdev, &atiixp_enable_bits[ap->port_no]))
87 		return -ENOENT;
88 
89 	return ata_sff_prereset(link, deadline);
90 }
91 
92 /**
93  *	atiixp_set_pio_timing	-	set initial PIO mode data
94  *	@ap: ATA interface
95  *	@adev: ATA device
96  *	@pio: Requested PIO
97  *
98  *	Called by both the pio and dma setup functions to set the controller
99  *	timings for PIO transfers. We must load both the mode number and
100  *	timing values into the controller.
101  */
102 
103 static void atiixp_set_pio_timing(struct ata_port *ap, struct ata_device *adev, int pio)
104 {
105 	static u8 pio_timings[5] = { 0x5D, 0x47, 0x34, 0x22, 0x20 };
106 
107 	struct pci_dev *pdev = to_pci_dev(ap->host->dev);
108 	int dn = 2 * ap->port_no + adev->devno;
109 	int timing_shift = (16 * ap->port_no) + 8 * (adev->devno ^ 1);
110 	u32 pio_timing_data;
111 	u16 pio_mode_data;
112 
113 	pci_read_config_word(pdev, ATIIXP_IDE_PIO_MODE, &pio_mode_data);
114 	pio_mode_data &= ~(0x7 << (4 * dn));
115 	pio_mode_data |= pio << (4 * dn);
116 	pci_write_config_word(pdev, ATIIXP_IDE_PIO_MODE, pio_mode_data);
117 
118 	pci_read_config_dword(pdev, ATIIXP_IDE_PIO_TIMING, &pio_timing_data);
119 	pio_timing_data &= ~(0xFF << timing_shift);
120 	pio_timing_data |= (pio_timings[pio] << timing_shift);
121 	pci_write_config_dword(pdev, ATIIXP_IDE_PIO_TIMING, pio_timing_data);
122 }
123 
124 /**
125  *	atiixp_set_piomode	-	set initial PIO mode data
126  *	@ap: ATA interface
127  *	@adev: ATA device
128  *
129  *	Called to do the PIO mode setup. We use a shared helper for this
130  *	as the DMA setup must also adjust the PIO timing information.
131  */
132 
133 static void atiixp_set_piomode(struct ata_port *ap, struct ata_device *adev)
134 {
135 	unsigned long flags;
136 	spin_lock_irqsave(&atiixp_lock, flags);
137 	atiixp_set_pio_timing(ap, adev, adev->pio_mode - XFER_PIO_0);
138 	spin_unlock_irqrestore(&atiixp_lock, flags);
139 }
140 
141 /**
142  *	atiixp_set_dmamode	-	set initial DMA mode data
143  *	@ap: ATA interface
144  *	@adev: ATA device
145  *
146  *	Called to do the DMA mode setup. We use timing tables for most
147  *	modes but must tune an appropriate PIO mode to match.
148  */
149 
150 static void atiixp_set_dmamode(struct ata_port *ap, struct ata_device *adev)
151 {
152 	static u8 mwdma_timings[5] = { 0x77, 0x21, 0x20 };
153 
154 	struct pci_dev *pdev = to_pci_dev(ap->host->dev);
155 	int dma = adev->dma_mode;
156 	int dn = 2 * ap->port_no + adev->devno;
157 	int wanted_pio;
158 	unsigned long flags;
159 
160 	spin_lock_irqsave(&atiixp_lock, flags);
161 
162 	if (adev->dma_mode >= XFER_UDMA_0) {
163 		u16 udma_mode_data;
164 
165 		dma -= XFER_UDMA_0;
166 
167 		pci_read_config_word(pdev, ATIIXP_IDE_UDMA_MODE, &udma_mode_data);
168 		udma_mode_data &= ~(0x7 << (4 * dn));
169 		udma_mode_data |= dma << (4 * dn);
170 		pci_write_config_word(pdev, ATIIXP_IDE_UDMA_MODE, udma_mode_data);
171 	} else {
172 		int timing_shift = (16 * ap->port_no) + 8 * (adev->devno ^ 1);
173 		u32 mwdma_timing_data;
174 
175 		dma -= XFER_MW_DMA_0;
176 
177 		pci_read_config_dword(pdev, ATIIXP_IDE_MWDMA_TIMING,
178 				      &mwdma_timing_data);
179 		mwdma_timing_data &= ~(0xFF << timing_shift);
180 		mwdma_timing_data |= (mwdma_timings[dma] << timing_shift);
181 		pci_write_config_dword(pdev, ATIIXP_IDE_MWDMA_TIMING,
182 				       mwdma_timing_data);
183 	}
184 	/*
185 	 *	We must now look at the PIO mode situation. We may need to
186 	 *	adjust the PIO mode to keep the timings acceptable
187 	 */
188 	if (adev->dma_mode >= XFER_MW_DMA_2)
189 		wanted_pio = 4;
190 	else if (adev->dma_mode == XFER_MW_DMA_1)
191 		wanted_pio = 3;
192 	else if (adev->dma_mode == XFER_MW_DMA_0)
193 		wanted_pio = 0;
194 	else BUG();
195 
196 	if (adev->pio_mode != wanted_pio)
197 		atiixp_set_pio_timing(ap, adev, wanted_pio);
198 	spin_unlock_irqrestore(&atiixp_lock, flags);
199 }
200 
201 /**
202  *	atiixp_bmdma_start	-	DMA start callback
203  *	@qc: Command in progress
204  *
205  *	When DMA begins we need to ensure that the UDMA control
206  *	register for the channel is correctly set.
207  *
208  *	Note: The host lock held by the libata layer protects
209  *	us from two channels both trying to set DMA bits at once
210  */
211 
212 static void atiixp_bmdma_start(struct ata_queued_cmd *qc)
213 {
214 	struct ata_port *ap = qc->ap;
215 	struct ata_device *adev = qc->dev;
216 
217 	struct pci_dev *pdev = to_pci_dev(ap->host->dev);
218 	int dn = (2 * ap->port_no) + adev->devno;
219 	u16 tmp16;
220 
221 	pci_read_config_word(pdev, ATIIXP_IDE_UDMA_CONTROL, &tmp16);
222 	if (ata_using_udma(adev))
223 		tmp16 |= (1 << dn);
224 	else
225 		tmp16 &= ~(1 << dn);
226 	pci_write_config_word(pdev, ATIIXP_IDE_UDMA_CONTROL, tmp16);
227 	ata_bmdma_start(qc);
228 }
229 
230 /**
231  *	atiixp_bmdma_stop	-	DMA stop callback
232  *	@qc: Command in progress
233  *
234  *	DMA has completed. Clear the UDMA flag as the next operations will
235  *	be PIO ones not UDMA data transfer.
236  *
237  *	Note: The host lock held by the libata layer protects
238  *	us from two channels both trying to set DMA bits at once
239  */
240 
241 static void atiixp_bmdma_stop(struct ata_queued_cmd *qc)
242 {
243 	struct ata_port *ap = qc->ap;
244 	struct pci_dev *pdev = to_pci_dev(ap->host->dev);
245 	int dn = (2 * ap->port_no) + qc->dev->devno;
246 	u16 tmp16;
247 
248 	pci_read_config_word(pdev, ATIIXP_IDE_UDMA_CONTROL, &tmp16);
249 	tmp16 &= ~(1 << dn);
250 	pci_write_config_word(pdev, ATIIXP_IDE_UDMA_CONTROL, tmp16);
251 	ata_bmdma_stop(qc);
252 }
253 
254 static struct scsi_host_template atiixp_sht = {
255 	ATA_BASE_SHT(DRV_NAME),
256 	.sg_tablesize		= LIBATA_DUMB_MAX_PRD,
257 	.dma_boundary		= ATA_DMA_BOUNDARY,
258 };
259 
260 static struct ata_port_operations atiixp_port_ops = {
261 	.inherits	= &ata_bmdma_port_ops,
262 
263 	.qc_prep 	= ata_bmdma_dumb_qc_prep,
264 	.bmdma_start 	= atiixp_bmdma_start,
265 	.bmdma_stop	= atiixp_bmdma_stop,
266 
267 	.prereset	= atiixp_prereset,
268 	.cable_detect	= atiixp_cable_detect,
269 	.set_piomode	= atiixp_set_piomode,
270 	.set_dmamode	= atiixp_set_dmamode,
271 };
272 
273 static int atiixp_init_one(struct pci_dev *pdev, const struct pci_device_id *id)
274 {
275 	static const struct ata_port_info info = {
276 		.flags = ATA_FLAG_SLAVE_POSS,
277 		.pio_mask = ATA_PIO4,
278 		.mwdma_mask = ATA_MWDMA12_ONLY,
279 		.udma_mask = ATA_UDMA5,
280 		.port_ops = &atiixp_port_ops
281 	};
282 	const struct ata_port_info *ppi[] = { &info, &info };
283 
284 	/* SB600 doesn't have secondary port wired */
285 	if (pdev->device == PCI_DEVICE_ID_ATI_IXP600_IDE)
286 		ppi[1] = &ata_dummy_port_info;
287 
288 	return ata_pci_bmdma_init_one(pdev, ppi, &atiixp_sht, NULL,
289 				      ATA_HOST_PARALLEL_SCAN);
290 }
291 
292 static const struct pci_device_id atiixp[] = {
293 	{ PCI_VDEVICE(ATI, PCI_DEVICE_ID_ATI_IXP200_IDE), },
294 	{ PCI_VDEVICE(ATI, PCI_DEVICE_ID_ATI_IXP300_IDE), },
295 	{ PCI_VDEVICE(ATI, PCI_DEVICE_ID_ATI_IXP400_IDE), },
296 	{ PCI_VDEVICE(ATI, PCI_DEVICE_ID_ATI_IXP600_IDE), },
297 	{ PCI_VDEVICE(ATI, PCI_DEVICE_ID_ATI_IXP700_IDE), },
298 	{ PCI_VDEVICE(AMD, PCI_DEVICE_ID_AMD_HUDSON2_IDE), },
299 
300 	{ },
301 };
302 
303 static struct pci_driver atiixp_pci_driver = {
304 	.name 		= DRV_NAME,
305 	.id_table	= atiixp,
306 	.probe 		= atiixp_init_one,
307 	.remove		= ata_pci_remove_one,
308 #ifdef CONFIG_PM_SLEEP
309 	.resume		= ata_pci_device_resume,
310 	.suspend	= ata_pci_device_suspend,
311 #endif
312 };
313 
314 module_pci_driver(atiixp_pci_driver);
315 
316 MODULE_AUTHOR("Alan Cox");
317 MODULE_DESCRIPTION("low-level driver for ATI IXP200/300/400");
318 MODULE_LICENSE("GPL");
319 MODULE_DEVICE_TABLE(pci, atiixp);
320 MODULE_VERSION(DRV_VERSION);
321