xref: /linux/drivers/mmc/host/alcor.c (revision 2f7932b011e7fb9f98732f95a68f6017d4d8c542)
1 // SPDX-License-Identifier: GPL-2.0+
2 /*
3  * Copyright (C) 2018 Oleksij Rempel <linux@rempel-privat.de>
4  *
5  * Driver for Alcor Micro AU6601 and AU6621 controllers
6  */
7 
8 /* Note: this driver was created without any documentation. Based
9  * on sniffing, testing and in some cases mimic of original driver.
10  * As soon as some one with documentation or more experience in SD/MMC, or
11  * reverse engineering then me, please review this driver and question every
12  * thing what I did. 2018 Oleksij Rempel <linux@rempel-privat.de>
13  */
14 
15 #include <linux/delay.h>
16 #include <linux/pci.h>
17 #include <linux/module.h>
18 #include <linux/io.h>
19 #include <linux/pm.h>
20 #include <linux/irq.h>
21 #include <linux/interrupt.h>
22 #include <linux/platform_device.h>
23 
24 #include <linux/mmc/host.h>
25 #include <linux/mmc/mmc.h>
26 
27 #include <linux/alcor_pci.h>
28 
29 enum alcor_cookie {
30 	COOKIE_UNMAPPED,
31 	COOKIE_PRE_MAPPED,
32 	COOKIE_MAPPED,
33 };
34 
35 struct alcor_pll_conf {
36 	unsigned int clk_src_freq;
37 	unsigned int clk_src_reg;
38 	unsigned int min_div;
39 	unsigned int max_div;
40 };
41 
42 struct alcor_sdmmc_host {
43 	struct  device *dev;
44 	struct alcor_pci_priv *alcor_pci;
45 
46 	struct mmc_host *mmc;
47 	struct mmc_request *mrq;
48 	struct mmc_command *cmd;
49 	struct mmc_data *data;
50 	unsigned int dma_on:1;
51 
52 	struct mutex cmd_mutex;
53 
54 	struct delayed_work timeout_work;
55 
56 	struct sg_mapping_iter sg_miter;	/* SG state for PIO */
57 	struct scatterlist *sg;
58 	unsigned int blocks;		/* remaining PIO blocks */
59 	int sg_count;
60 
61 	u32			irq_status_sd;
62 	unsigned char		cur_power_mode;
63 };
64 
65 static const struct alcor_pll_conf alcor_pll_cfg[] = {
66 	/* MHZ,		CLK src,		max div, min div */
67 	{ 31250000,	AU6601_CLK_31_25_MHZ,	1,	511},
68 	{ 48000000,	AU6601_CLK_48_MHZ,	1,	511},
69 	{125000000,	AU6601_CLK_125_MHZ,	1,	511},
70 	{384000000,	AU6601_CLK_384_MHZ,	1,	511},
71 };
72 
73 static inline void alcor_rmw8(struct alcor_sdmmc_host *host, unsigned int addr,
74 			       u8 clear, u8 set)
75 {
76 	struct alcor_pci_priv *priv = host->alcor_pci;
77 	u32 var;
78 
79 	var = alcor_read8(priv, addr);
80 	var &= ~clear;
81 	var |= set;
82 	alcor_write8(priv, var, addr);
83 }
84 
85 /* As soon as irqs are masked, some status updates may be missed.
86  * Use this with care.
87  */
88 static inline void alcor_mask_sd_irqs(struct alcor_sdmmc_host *host)
89 {
90 	struct alcor_pci_priv *priv = host->alcor_pci;
91 
92 	alcor_write32(priv, 0, AU6601_REG_INT_ENABLE);
93 }
94 
95 static inline void alcor_unmask_sd_irqs(struct alcor_sdmmc_host *host)
96 {
97 	struct alcor_pci_priv *priv = host->alcor_pci;
98 
99 	alcor_write32(priv, AU6601_INT_CMD_MASK | AU6601_INT_DATA_MASK |
100 		  AU6601_INT_CARD_INSERT | AU6601_INT_CARD_REMOVE |
101 		  AU6601_INT_OVER_CURRENT_ERR,
102 		  AU6601_REG_INT_ENABLE);
103 }
104 
105 static void alcor_reset(struct alcor_sdmmc_host *host, u8 val)
106 {
107 	struct alcor_pci_priv *priv = host->alcor_pci;
108 	int i;
109 
110 	alcor_write8(priv, val | AU6601_BUF_CTRL_RESET,
111 		      AU6601_REG_SW_RESET);
112 	for (i = 0; i < 100; i++) {
113 		if (!(alcor_read8(priv, AU6601_REG_SW_RESET) & val))
114 			return;
115 		udelay(50);
116 	}
117 	dev_err(host->dev, "%s: timeout\n", __func__);
118 }
119 
120 static void alcor_data_set_dma(struct alcor_sdmmc_host *host)
121 {
122 	struct alcor_pci_priv *priv = host->alcor_pci;
123 	u32 addr;
124 
125 	if (!host->sg_count)
126 		return;
127 
128 	if (!host->sg) {
129 		dev_err(host->dev, "have blocks, but no SG\n");
130 		return;
131 	}
132 
133 	if (!sg_dma_len(host->sg)) {
134 		dev_err(host->dev, "DMA SG len == 0\n");
135 		return;
136 	}
137 
138 
139 	addr = (u32)sg_dma_address(host->sg);
140 
141 	alcor_write32(priv, addr, AU6601_REG_SDMA_ADDR);
142 	host->sg = sg_next(host->sg);
143 	host->sg_count--;
144 }
145 
146 static void alcor_trigger_data_transfer(struct alcor_sdmmc_host *host)
147 {
148 	struct alcor_pci_priv *priv = host->alcor_pci;
149 	struct mmc_data *data = host->data;
150 	u8 ctrl = 0;
151 
152 	if (data->flags & MMC_DATA_WRITE)
153 		ctrl |= AU6601_DATA_WRITE;
154 
155 	if (data->host_cookie == COOKIE_MAPPED) {
156 		alcor_data_set_dma(host);
157 		ctrl |= AU6601_DATA_DMA_MODE;
158 		host->dma_on = 1;
159 		alcor_write32(priv, data->sg_count * 0x1000,
160 			       AU6601_REG_BLOCK_SIZE);
161 	} else {
162 		alcor_write32(priv, data->blksz, AU6601_REG_BLOCK_SIZE);
163 	}
164 
165 	alcor_write8(priv, ctrl | AU6601_DATA_START_XFER,
166 		      AU6601_DATA_XFER_CTRL);
167 }
168 
169 static void alcor_trf_block_pio(struct alcor_sdmmc_host *host, bool read)
170 {
171 	struct alcor_pci_priv *priv = host->alcor_pci;
172 	size_t blksize, len;
173 	u8 *buf;
174 
175 	if (!host->blocks)
176 		return;
177 
178 	if (host->dma_on) {
179 		dev_err(host->dev, "configured DMA but got PIO request.\n");
180 		return;
181 	}
182 
183 	if (!!(host->data->flags & MMC_DATA_READ) != read) {
184 		dev_err(host->dev, "got unexpected direction %i != %i\n",
185 			!!(host->data->flags & MMC_DATA_READ), read);
186 	}
187 
188 	if (!sg_miter_next(&host->sg_miter))
189 		return;
190 
191 	blksize = host->data->blksz;
192 	len = min(host->sg_miter.length, blksize);
193 
194 	dev_dbg(host->dev, "PIO, %s block size: 0x%zx\n",
195 		read ? "read" : "write", blksize);
196 
197 	host->sg_miter.consumed = len;
198 	host->blocks--;
199 
200 	buf = host->sg_miter.addr;
201 
202 	if (read)
203 		ioread32_rep(priv->iobase + AU6601_REG_BUFFER, buf, len >> 2);
204 	else
205 		iowrite32_rep(priv->iobase + AU6601_REG_BUFFER, buf, len >> 2);
206 
207 	sg_miter_stop(&host->sg_miter);
208 }
209 
210 static void alcor_prepare_sg_miter(struct alcor_sdmmc_host *host)
211 {
212 	unsigned int flags = SG_MITER_ATOMIC;
213 	struct mmc_data *data = host->data;
214 
215 	if (data->flags & MMC_DATA_READ)
216 		flags |= SG_MITER_TO_SG;
217 	else
218 		flags |= SG_MITER_FROM_SG;
219 	sg_miter_start(&host->sg_miter, data->sg, data->sg_len, flags);
220 }
221 
222 static void alcor_prepare_data(struct alcor_sdmmc_host *host,
223 			       struct mmc_command *cmd)
224 {
225 	struct alcor_pci_priv *priv = host->alcor_pci;
226 	struct mmc_data *data = cmd->data;
227 
228 	if (!data)
229 		return;
230 
231 
232 	host->data = data;
233 	host->data->bytes_xfered = 0;
234 	host->blocks = data->blocks;
235 	host->sg = data->sg;
236 	host->sg_count = data->sg_count;
237 	dev_dbg(host->dev, "prepare DATA: sg %i, blocks: %i\n",
238 			host->sg_count, host->blocks);
239 
240 	if (data->host_cookie != COOKIE_MAPPED)
241 		alcor_prepare_sg_miter(host);
242 
243 	alcor_write8(priv, 0, AU6601_DATA_XFER_CTRL);
244 }
245 
246 static void alcor_send_cmd(struct alcor_sdmmc_host *host,
247 			   struct mmc_command *cmd, bool set_timeout)
248 {
249 	struct alcor_pci_priv *priv = host->alcor_pci;
250 	unsigned long timeout = 0;
251 	u8 ctrl = 0;
252 
253 	host->cmd = cmd;
254 	alcor_prepare_data(host, cmd);
255 
256 	dev_dbg(host->dev, "send CMD. opcode: 0x%02x, arg; 0x%08x\n",
257 		cmd->opcode, cmd->arg);
258 	alcor_write8(priv, cmd->opcode | 0x40, AU6601_REG_CMD_OPCODE);
259 	alcor_write32be(priv, cmd->arg, AU6601_REG_CMD_ARG);
260 
261 	switch (mmc_resp_type(cmd)) {
262 	case MMC_RSP_NONE:
263 		ctrl = AU6601_CMD_NO_RESP;
264 		break;
265 	case MMC_RSP_R1:
266 		ctrl = AU6601_CMD_6_BYTE_CRC;
267 		break;
268 	case MMC_RSP_R1B:
269 		ctrl = AU6601_CMD_6_BYTE_CRC | AU6601_CMD_STOP_WAIT_RDY;
270 		break;
271 	case MMC_RSP_R2:
272 		ctrl = AU6601_CMD_17_BYTE_CRC;
273 		break;
274 	case MMC_RSP_R3:
275 		ctrl = AU6601_CMD_6_BYTE_WO_CRC;
276 		break;
277 	default:
278 		dev_err(host->dev, "%s: cmd->flag (0x%02x) is not valid\n",
279 			mmc_hostname(host->mmc), mmc_resp_type(cmd));
280 		break;
281 	}
282 
283 	if (set_timeout) {
284 		if (!cmd->data && cmd->busy_timeout)
285 			timeout = cmd->busy_timeout;
286 		else
287 			timeout = 10000;
288 
289 		schedule_delayed_work(&host->timeout_work,
290 				      msecs_to_jiffies(timeout));
291 	}
292 
293 	dev_dbg(host->dev, "xfer ctrl: 0x%02x; timeout: %lu\n", ctrl, timeout);
294 	alcor_write8(priv, ctrl | AU6601_CMD_START_XFER,
295 				 AU6601_CMD_XFER_CTRL);
296 }
297 
298 static void alcor_request_complete(struct alcor_sdmmc_host *host,
299 				   bool cancel_timeout)
300 {
301 	struct mmc_request *mrq;
302 
303 	/*
304 	 * If this work gets rescheduled while running, it will
305 	 * be run again afterwards but without any active request.
306 	 */
307 	if (!host->mrq)
308 		return;
309 
310 	if (cancel_timeout)
311 		cancel_delayed_work(&host->timeout_work);
312 
313 	mrq = host->mrq;
314 
315 	host->mrq = NULL;
316 	host->cmd = NULL;
317 	host->data = NULL;
318 	host->dma_on = 0;
319 
320 	mmc_request_done(host->mmc, mrq);
321 }
322 
323 static void alcor_finish_data(struct alcor_sdmmc_host *host)
324 {
325 	struct mmc_data *data;
326 
327 	data = host->data;
328 	host->data = NULL;
329 	host->dma_on = 0;
330 
331 	/*
332 	 * The specification states that the block count register must
333 	 * be updated, but it does not specify at what point in the
334 	 * data flow. That makes the register entirely useless to read
335 	 * back so we have to assume that nothing made it to the card
336 	 * in the event of an error.
337 	 */
338 	if (data->error)
339 		data->bytes_xfered = 0;
340 	else
341 		data->bytes_xfered = data->blksz * data->blocks;
342 
343 	/*
344 	 * Need to send CMD12 if -
345 	 * a) open-ended multiblock transfer (no CMD23)
346 	 * b) error in multiblock transfer
347 	 */
348 	if (data->stop &&
349 	    (data->error ||
350 	     !host->mrq->sbc)) {
351 
352 		/*
353 		 * The controller needs a reset of internal state machines
354 		 * upon error conditions.
355 		 */
356 		if (data->error)
357 			alcor_reset(host, AU6601_RESET_CMD | AU6601_RESET_DATA);
358 
359 		alcor_unmask_sd_irqs(host);
360 		alcor_send_cmd(host, data->stop, false);
361 		return;
362 	}
363 
364 	alcor_request_complete(host, 1);
365 }
366 
367 static void alcor_err_irq(struct alcor_sdmmc_host *host, u32 intmask)
368 {
369 	dev_dbg(host->dev, "ERR IRQ %x\n", intmask);
370 
371 	if (host->cmd) {
372 		if (intmask & AU6601_INT_CMD_TIMEOUT_ERR)
373 			host->cmd->error = -ETIMEDOUT;
374 		else
375 			host->cmd->error = -EILSEQ;
376 	}
377 
378 	if (host->data) {
379 		if (intmask & AU6601_INT_DATA_TIMEOUT_ERR)
380 			host->data->error = -ETIMEDOUT;
381 		else
382 			host->data->error = -EILSEQ;
383 
384 		host->data->bytes_xfered = 0;
385 	}
386 
387 	alcor_reset(host, AU6601_RESET_CMD | AU6601_RESET_DATA);
388 	alcor_request_complete(host, 1);
389 }
390 
391 static int alcor_cmd_irq_done(struct alcor_sdmmc_host *host, u32 intmask)
392 {
393 	struct alcor_pci_priv *priv = host->alcor_pci;
394 
395 	intmask &= AU6601_INT_CMD_END;
396 
397 	if (!intmask)
398 		return true;
399 
400 	/* got CMD_END but no CMD is in progress, wake thread an process the
401 	 * error
402 	 */
403 	if (!host->cmd)
404 		return false;
405 
406 	if (host->cmd->flags & MMC_RSP_PRESENT) {
407 		struct mmc_command *cmd = host->cmd;
408 
409 		cmd->resp[0] = alcor_read32be(priv, AU6601_REG_CMD_RSP0);
410 		dev_dbg(host->dev, "RSP0: 0x%04x\n", cmd->resp[0]);
411 		if (host->cmd->flags & MMC_RSP_136) {
412 			cmd->resp[1] =
413 				alcor_read32be(priv, AU6601_REG_CMD_RSP1);
414 			cmd->resp[2] =
415 				alcor_read32be(priv, AU6601_REG_CMD_RSP2);
416 			cmd->resp[3] =
417 				alcor_read32be(priv, AU6601_REG_CMD_RSP3);
418 			dev_dbg(host->dev, "RSP1,2,3: 0x%04x 0x%04x 0x%04x\n",
419 				cmd->resp[1], cmd->resp[2], cmd->resp[3]);
420 		}
421 
422 	}
423 
424 	host->cmd->error = 0;
425 
426 	/* Processed actual command. */
427 	if (!host->data)
428 		return false;
429 
430 	alcor_trigger_data_transfer(host);
431 	host->cmd = NULL;
432 	return true;
433 }
434 
435 static void alcor_cmd_irq_thread(struct alcor_sdmmc_host *host, u32 intmask)
436 {
437 	intmask &= AU6601_INT_CMD_END;
438 
439 	if (!intmask)
440 		return;
441 
442 	if (!host->cmd && intmask & AU6601_INT_CMD_END) {
443 		dev_dbg(host->dev, "Got command interrupt 0x%08x even though no command operation was in progress.\n",
444 			intmask);
445 	}
446 
447 	/* Processed actual command. */
448 	if (!host->data)
449 		alcor_request_complete(host, 1);
450 	else
451 		alcor_trigger_data_transfer(host);
452 	host->cmd = NULL;
453 }
454 
455 static int alcor_data_irq_done(struct alcor_sdmmc_host *host, u32 intmask)
456 {
457 	u32 tmp;
458 
459 	intmask &= AU6601_INT_DATA_MASK;
460 
461 	/* nothing here to do */
462 	if (!intmask)
463 		return 1;
464 
465 	/* we was too fast and got DATA_END after it was processed?
466 	 * lets ignore it for now.
467 	 */
468 	if (!host->data && intmask == AU6601_INT_DATA_END)
469 		return 1;
470 
471 	/* looks like an error, so lets handle it. */
472 	if (!host->data)
473 		return 0;
474 
475 	tmp = intmask & (AU6601_INT_READ_BUF_RDY | AU6601_INT_WRITE_BUF_RDY
476 			 | AU6601_INT_DMA_END);
477 	switch (tmp) {
478 	case 0:
479 		break;
480 	case AU6601_INT_READ_BUF_RDY:
481 		alcor_trf_block_pio(host, true);
482 		return 1;
483 	case AU6601_INT_WRITE_BUF_RDY:
484 		alcor_trf_block_pio(host, false);
485 		return 1;
486 	case AU6601_INT_DMA_END:
487 		if (!host->sg_count)
488 			break;
489 
490 		alcor_data_set_dma(host);
491 		break;
492 	default:
493 		dev_err(host->dev, "Got READ_BUF_RDY and WRITE_BUF_RDY at same time\n");
494 		break;
495 	}
496 
497 	if (intmask & AU6601_INT_DATA_END) {
498 		if (!host->dma_on && host->blocks) {
499 			alcor_trigger_data_transfer(host);
500 			return 1;
501 		} else {
502 			return 0;
503 		}
504 	}
505 
506 	return 1;
507 }
508 
509 static void alcor_data_irq_thread(struct alcor_sdmmc_host *host, u32 intmask)
510 {
511 	intmask &= AU6601_INT_DATA_MASK;
512 
513 	if (!intmask)
514 		return;
515 
516 	if (!host->data) {
517 		dev_dbg(host->dev, "Got data interrupt 0x%08x even though no data operation was in progress.\n",
518 			intmask);
519 		alcor_reset(host, AU6601_RESET_DATA);
520 		return;
521 	}
522 
523 	if (alcor_data_irq_done(host, intmask))
524 		return;
525 
526 	if ((intmask & AU6601_INT_DATA_END) || !host->blocks ||
527 	    (host->dma_on && !host->sg_count))
528 		alcor_finish_data(host);
529 }
530 
531 static void alcor_cd_irq(struct alcor_sdmmc_host *host, u32 intmask)
532 {
533 	dev_dbg(host->dev, "card %s\n",
534 		intmask & AU6601_INT_CARD_REMOVE ? "removed" : "inserted");
535 
536 	if (host->mrq) {
537 		dev_dbg(host->dev, "cancel all pending tasks.\n");
538 
539 		if (host->data)
540 			host->data->error = -ENOMEDIUM;
541 
542 		if (host->cmd)
543 			host->cmd->error = -ENOMEDIUM;
544 		else
545 			host->mrq->cmd->error = -ENOMEDIUM;
546 
547 		alcor_request_complete(host, 1);
548 	}
549 
550 	mmc_detect_change(host->mmc, msecs_to_jiffies(1));
551 }
552 
553 static irqreturn_t alcor_irq_thread(int irq, void *d)
554 {
555 	struct alcor_sdmmc_host *host = d;
556 	irqreturn_t ret = IRQ_HANDLED;
557 	u32 intmask, tmp;
558 
559 	mutex_lock(&host->cmd_mutex);
560 
561 	intmask = host->irq_status_sd;
562 
563 	/* some thing bad */
564 	if (unlikely(!intmask || AU6601_INT_ALL_MASK == intmask)) {
565 		dev_dbg(host->dev, "unexpected IRQ: 0x%04x\n", intmask);
566 		ret = IRQ_NONE;
567 		goto exit;
568 	}
569 
570 	tmp = intmask & (AU6601_INT_CMD_MASK | AU6601_INT_DATA_MASK);
571 	if (tmp) {
572 		if (tmp & AU6601_INT_ERROR_MASK)
573 			alcor_err_irq(host, tmp);
574 		else {
575 			alcor_cmd_irq_thread(host, tmp);
576 			alcor_data_irq_thread(host, tmp);
577 		}
578 		intmask &= ~(AU6601_INT_CMD_MASK | AU6601_INT_DATA_MASK);
579 	}
580 
581 	if (intmask & (AU6601_INT_CARD_INSERT | AU6601_INT_CARD_REMOVE)) {
582 		alcor_cd_irq(host, intmask);
583 		intmask &= ~(AU6601_INT_CARD_INSERT | AU6601_INT_CARD_REMOVE);
584 	}
585 
586 	if (intmask & AU6601_INT_OVER_CURRENT_ERR) {
587 		dev_warn(host->dev,
588 			 "warning: over current detected!\n");
589 		intmask &= ~AU6601_INT_OVER_CURRENT_ERR;
590 	}
591 
592 	if (intmask)
593 		dev_dbg(host->dev, "got not handled IRQ: 0x%04x\n", intmask);
594 
595 exit:
596 	mutex_unlock(&host->cmd_mutex);
597 	alcor_unmask_sd_irqs(host);
598 	return ret;
599 }
600 
601 
602 static irqreturn_t alcor_irq(int irq, void *d)
603 {
604 	struct alcor_sdmmc_host *host = d;
605 	struct alcor_pci_priv *priv = host->alcor_pci;
606 	u32 status, tmp;
607 	irqreturn_t ret;
608 	int cmd_done, data_done;
609 
610 	status = alcor_read32(priv, AU6601_REG_INT_STATUS);
611 	if (!status)
612 		return IRQ_NONE;
613 
614 	alcor_write32(priv, status, AU6601_REG_INT_STATUS);
615 
616 	tmp = status & (AU6601_INT_READ_BUF_RDY | AU6601_INT_WRITE_BUF_RDY
617 			| AU6601_INT_DATA_END | AU6601_INT_DMA_END
618 			| AU6601_INT_CMD_END);
619 	if (tmp == status) {
620 		cmd_done = alcor_cmd_irq_done(host, tmp);
621 		data_done = alcor_data_irq_done(host, tmp);
622 		/* use fast path for simple tasks */
623 		if (cmd_done && data_done) {
624 			ret = IRQ_HANDLED;
625 			goto alcor_irq_done;
626 		}
627 	}
628 
629 	host->irq_status_sd = status;
630 	ret = IRQ_WAKE_THREAD;
631 	alcor_mask_sd_irqs(host);
632 alcor_irq_done:
633 	return ret;
634 }
635 
636 static void alcor_set_clock(struct alcor_sdmmc_host *host, unsigned int clock)
637 {
638 	struct alcor_pci_priv *priv = host->alcor_pci;
639 	int i, diff = 0x7fffffff, tmp_clock = 0;
640 	u16 clk_src = 0;
641 	u8 clk_div = 0;
642 
643 	if (clock == 0) {
644 		alcor_write16(priv, 0, AU6601_CLK_SELECT);
645 		return;
646 	}
647 
648 	for (i = 0; i < ARRAY_SIZE(alcor_pll_cfg); i++) {
649 		unsigned int tmp_div, tmp_diff;
650 		const struct alcor_pll_conf *cfg = &alcor_pll_cfg[i];
651 
652 		tmp_div = DIV_ROUND_UP(cfg->clk_src_freq, clock);
653 		if (cfg->min_div > tmp_div || tmp_div > cfg->max_div)
654 			continue;
655 
656 		tmp_clock = DIV_ROUND_UP(cfg->clk_src_freq, tmp_div);
657 		tmp_diff = abs(clock - tmp_clock);
658 
659 		if (tmp_diff >= 0 && tmp_diff < diff) {
660 			diff = tmp_diff;
661 			clk_src = cfg->clk_src_reg;
662 			clk_div = tmp_div;
663 		}
664 	}
665 
666 	clk_src |= ((clk_div - 1) << 8);
667 	clk_src |= AU6601_CLK_ENABLE;
668 
669 	dev_dbg(host->dev, "set freq %d cal freq %d, use div %d, mod %x\n",
670 			clock, tmp_clock, clk_div, clk_src);
671 
672 	alcor_write16(priv, clk_src, AU6601_CLK_SELECT);
673 
674 }
675 
676 static void alcor_set_timing(struct mmc_host *mmc, struct mmc_ios *ios)
677 {
678 	struct alcor_sdmmc_host *host = mmc_priv(mmc);
679 
680 	if (ios->timing == MMC_TIMING_LEGACY) {
681 		alcor_rmw8(host, AU6601_CLK_DELAY,
682 			    AU6601_CLK_POSITIVE_EDGE_ALL, 0);
683 	} else {
684 		alcor_rmw8(host, AU6601_CLK_DELAY,
685 			    0, AU6601_CLK_POSITIVE_EDGE_ALL);
686 	}
687 }
688 
689 static void alcor_set_bus_width(struct mmc_host *mmc, struct mmc_ios *ios)
690 {
691 	struct alcor_sdmmc_host *host = mmc_priv(mmc);
692 	struct alcor_pci_priv *priv = host->alcor_pci;
693 
694 	if (ios->bus_width == MMC_BUS_WIDTH_1) {
695 		alcor_write8(priv, 0, AU6601_REG_BUS_CTRL);
696 	} else if (ios->bus_width == MMC_BUS_WIDTH_4) {
697 		alcor_write8(priv, AU6601_BUS_WIDTH_4BIT,
698 			      AU6601_REG_BUS_CTRL);
699 	} else
700 		dev_err(host->dev, "Unknown BUS mode\n");
701 
702 }
703 
704 static int alcor_card_busy(struct mmc_host *mmc)
705 {
706 	struct alcor_sdmmc_host *host = mmc_priv(mmc);
707 	struct alcor_pci_priv *priv = host->alcor_pci;
708 	u8 status;
709 
710 	/* Check whether dat[0:3] low */
711 	status = alcor_read8(priv, AU6601_DATA_PIN_STATE);
712 
713 	return !(status & AU6601_BUS_STAT_DAT_MASK);
714 }
715 
716 static int alcor_get_cd(struct mmc_host *mmc)
717 {
718 	struct alcor_sdmmc_host *host = mmc_priv(mmc);
719 	struct alcor_pci_priv *priv = host->alcor_pci;
720 	u8 detect;
721 
722 	detect = alcor_read8(priv, AU6601_DETECT_STATUS)
723 		& AU6601_DETECT_STATUS_M;
724 	/* check if card is present then send command and data */
725 	return (detect == AU6601_SD_DETECTED);
726 }
727 
728 static int alcor_get_ro(struct mmc_host *mmc)
729 {
730 	struct alcor_sdmmc_host *host = mmc_priv(mmc);
731 	struct alcor_pci_priv *priv = host->alcor_pci;
732 	u8 status;
733 
734 	/* get write protect pin status */
735 	status = alcor_read8(priv, AU6601_INTERFACE_MODE_CTRL);
736 
737 	return !!(status & AU6601_SD_CARD_WP);
738 }
739 
740 static void alcor_request(struct mmc_host *mmc, struct mmc_request *mrq)
741 {
742 	struct alcor_sdmmc_host *host = mmc_priv(mmc);
743 
744 	mutex_lock(&host->cmd_mutex);
745 
746 	host->mrq = mrq;
747 
748 	/* check if card is present then send command and data */
749 	if (alcor_get_cd(mmc))
750 		alcor_send_cmd(host, mrq->cmd, true);
751 	else {
752 		mrq->cmd->error = -ENOMEDIUM;
753 		alcor_request_complete(host, 1);
754 	}
755 
756 	mutex_unlock(&host->cmd_mutex);
757 }
758 
759 static void alcor_pre_req(struct mmc_host *mmc,
760 			   struct mmc_request *mrq)
761 {
762 	struct alcor_sdmmc_host *host = mmc_priv(mmc);
763 	struct mmc_data *data = mrq->data;
764 	struct mmc_command *cmd = mrq->cmd;
765 	struct scatterlist *sg;
766 	unsigned int i, sg_len;
767 
768 	if (!data || !cmd)
769 		return;
770 
771 	data->host_cookie = COOKIE_UNMAPPED;
772 
773 	/* FIXME: looks like the DMA engine works only with CMD18 */
774 	if (cmd->opcode != 18)
775 		return;
776 	/*
777 	 * We don't do DMA on "complex" transfers, i.e. with
778 	 * non-word-aligned buffers or lengths. Also, we don't bother
779 	 * with all the DMA setup overhead for short transfers.
780 	 */
781 	if (data->blocks * data->blksz < AU6601_MAX_DMA_BLOCK_SIZE)
782 		return;
783 
784 	if (data->blksz & 3)
785 		return;
786 
787 	for_each_sg(data->sg, sg, data->sg_len, i) {
788 		if (sg->length != AU6601_MAX_DMA_BLOCK_SIZE)
789 			return;
790 	}
791 
792 	/* This data might be unmapped at this time */
793 
794 	sg_len = dma_map_sg(host->dev, data->sg, data->sg_len,
795 			    mmc_get_dma_dir(data));
796 	if (sg_len)
797 		data->host_cookie = COOKIE_MAPPED;
798 
799 	data->sg_count = sg_len;
800 }
801 
802 static void alcor_post_req(struct mmc_host *mmc,
803 			    struct mmc_request *mrq,
804 			    int err)
805 {
806 	struct alcor_sdmmc_host *host = mmc_priv(mmc);
807 	struct mmc_data *data = mrq->data;
808 
809 	if (!data)
810 		return;
811 
812 	if (data->host_cookie == COOKIE_MAPPED) {
813 		dma_unmap_sg(host->dev,
814 			     data->sg,
815 			     data->sg_len,
816 			     mmc_get_dma_dir(data));
817 	}
818 
819 	data->host_cookie = COOKIE_UNMAPPED;
820 }
821 
822 static void alcor_set_power_mode(struct mmc_host *mmc, struct mmc_ios *ios)
823 {
824 	struct alcor_sdmmc_host *host = mmc_priv(mmc);
825 	struct alcor_pci_priv *priv = host->alcor_pci;
826 
827 	switch (ios->power_mode) {
828 	case MMC_POWER_OFF:
829 		alcor_set_clock(host, ios->clock);
830 		/* set all pins to input */
831 		alcor_write8(priv, 0, AU6601_OUTPUT_ENABLE);
832 		/* turn of VDD */
833 		alcor_write8(priv, 0, AU6601_POWER_CONTROL);
834 		break;
835 	case MMC_POWER_UP:
836 		break;
837 	case MMC_POWER_ON:
838 		/* This is most trickiest part. The order and timings of
839 		 * instructions seems to play important role. Any changes may
840 		 * confuse internal state engine if this HW.
841 		 * FIXME: If we will ever get access to documentation, then this
842 		 * part should be reviewed again.
843 		 */
844 
845 		/* enable SD card mode */
846 		alcor_write8(priv, AU6601_SD_CARD,
847 			      AU6601_ACTIVE_CTRL);
848 		/* set signal voltage to 3.3V */
849 		alcor_write8(priv, 0, AU6601_OPT);
850 		/* no documentation about clk delay, for now just try to mimic
851 		 * original driver.
852 		 */
853 		alcor_write8(priv, 0x20, AU6601_CLK_DELAY);
854 		/* set BUS width to 1 bit */
855 		alcor_write8(priv, 0, AU6601_REG_BUS_CTRL);
856 		/* set CLK first time */
857 		alcor_set_clock(host, ios->clock);
858 		/* power on VDD */
859 		alcor_write8(priv, AU6601_SD_CARD,
860 			      AU6601_POWER_CONTROL);
861 		/* wait until the CLK will get stable */
862 		mdelay(20);
863 		/* set CLK again, mimic original driver. */
864 		alcor_set_clock(host, ios->clock);
865 
866 		/* enable output */
867 		alcor_write8(priv, AU6601_SD_CARD,
868 			      AU6601_OUTPUT_ENABLE);
869 		/* The clk will not work on au6621. We need to trigger data
870 		 * transfer.
871 		 */
872 		alcor_write8(priv, AU6601_DATA_WRITE,
873 			      AU6601_DATA_XFER_CTRL);
874 		/* configure timeout. Not clear what exactly it means. */
875 		alcor_write8(priv, 0x7d, AU6601_TIME_OUT_CTRL);
876 		mdelay(100);
877 		break;
878 	default:
879 		dev_err(host->dev, "Unknown power parameter\n");
880 	}
881 }
882 
883 static void alcor_set_ios(struct mmc_host *mmc, struct mmc_ios *ios)
884 {
885 	struct alcor_sdmmc_host *host = mmc_priv(mmc);
886 
887 	mutex_lock(&host->cmd_mutex);
888 
889 	dev_dbg(host->dev, "set ios. bus width: %x, power mode: %x\n",
890 		ios->bus_width, ios->power_mode);
891 
892 	if (ios->power_mode != host->cur_power_mode) {
893 		alcor_set_power_mode(mmc, ios);
894 		host->cur_power_mode = ios->power_mode;
895 	} else {
896 		alcor_set_timing(mmc, ios);
897 		alcor_set_bus_width(mmc, ios);
898 		alcor_set_clock(host, ios->clock);
899 	}
900 
901 	mutex_unlock(&host->cmd_mutex);
902 }
903 
904 static int alcor_signal_voltage_switch(struct mmc_host *mmc,
905 				       struct mmc_ios *ios)
906 {
907 	struct alcor_sdmmc_host *host = mmc_priv(mmc);
908 
909 	mutex_lock(&host->cmd_mutex);
910 
911 	switch (ios->signal_voltage) {
912 	case MMC_SIGNAL_VOLTAGE_330:
913 		alcor_rmw8(host, AU6601_OPT, AU6601_OPT_SD_18V, 0);
914 		break;
915 	case MMC_SIGNAL_VOLTAGE_180:
916 		alcor_rmw8(host, AU6601_OPT, 0, AU6601_OPT_SD_18V);
917 		break;
918 	default:
919 		/* No signal voltage switch required */
920 		break;
921 	}
922 
923 	mutex_unlock(&host->cmd_mutex);
924 	return 0;
925 }
926 
927 static const struct mmc_host_ops alcor_sdc_ops = {
928 	.card_busy	= alcor_card_busy,
929 	.get_cd		= alcor_get_cd,
930 	.get_ro		= alcor_get_ro,
931 	.post_req	= alcor_post_req,
932 	.pre_req	= alcor_pre_req,
933 	.request	= alcor_request,
934 	.set_ios	= alcor_set_ios,
935 	.start_signal_voltage_switch = alcor_signal_voltage_switch,
936 };
937 
938 static void alcor_timeout_timer(struct work_struct *work)
939 {
940 	struct delayed_work *d = to_delayed_work(work);
941 	struct alcor_sdmmc_host *host = container_of(d, struct alcor_sdmmc_host,
942 						timeout_work);
943 	mutex_lock(&host->cmd_mutex);
944 
945 	dev_dbg(host->dev, "triggered timeout\n");
946 	if (host->mrq) {
947 		dev_err(host->dev, "Timeout waiting for hardware interrupt.\n");
948 
949 		if (host->data) {
950 			host->data->error = -ETIMEDOUT;
951 		} else {
952 			if (host->cmd)
953 				host->cmd->error = -ETIMEDOUT;
954 			else
955 				host->mrq->cmd->error = -ETIMEDOUT;
956 		}
957 
958 		alcor_reset(host, AU6601_RESET_CMD | AU6601_RESET_DATA);
959 		alcor_request_complete(host, 0);
960 	}
961 
962 	mmiowb();
963 	mutex_unlock(&host->cmd_mutex);
964 }
965 
966 static void alcor_hw_init(struct alcor_sdmmc_host *host)
967 {
968 	struct alcor_pci_priv *priv = host->alcor_pci;
969 	struct alcor_dev_cfg *cfg = priv->cfg;
970 
971 	/* FIXME: This part is a mimics HW init of original driver.
972 	 * If we will ever get access to documentation, then this part
973 	 * should be reviewed again.
974 	 */
975 
976 	/* reset command state engine */
977 	alcor_reset(host, AU6601_RESET_CMD);
978 
979 	alcor_write8(priv, 0, AU6601_DMA_BOUNDARY);
980 	/* enable sd card mode */
981 	alcor_write8(priv, AU6601_SD_CARD, AU6601_ACTIVE_CTRL);
982 
983 	/* set BUS width to 1 bit */
984 	alcor_write8(priv, 0, AU6601_REG_BUS_CTRL);
985 
986 	/* reset data state engine */
987 	alcor_reset(host, AU6601_RESET_DATA);
988 	/* Not sure if a voodoo with AU6601_DMA_BOUNDARY is really needed */
989 	alcor_write8(priv, 0, AU6601_DMA_BOUNDARY);
990 
991 	alcor_write8(priv, 0, AU6601_INTERFACE_MODE_CTRL);
992 	/* not clear what we are doing here. */
993 	alcor_write8(priv, 0x44, AU6601_PAD_DRIVE0);
994 	alcor_write8(priv, 0x44, AU6601_PAD_DRIVE1);
995 	alcor_write8(priv, 0x00, AU6601_PAD_DRIVE2);
996 
997 	/* for 6601 - dma_boundary; for 6621 - dma_page_cnt
998 	 * exact meaning of this register is not clear.
999 	 */
1000 	alcor_write8(priv, cfg->dma, AU6601_DMA_BOUNDARY);
1001 
1002 	/* make sure all pins are set to input and VDD is off */
1003 	alcor_write8(priv, 0, AU6601_OUTPUT_ENABLE);
1004 	alcor_write8(priv, 0, AU6601_POWER_CONTROL);
1005 
1006 	alcor_write8(priv, AU6601_DETECT_EN, AU6601_DETECT_STATUS);
1007 	/* now we should be safe to enable IRQs */
1008 	alcor_unmask_sd_irqs(host);
1009 }
1010 
1011 static void alcor_hw_uninit(struct alcor_sdmmc_host *host)
1012 {
1013 	struct alcor_pci_priv *priv = host->alcor_pci;
1014 
1015 	alcor_mask_sd_irqs(host);
1016 	alcor_reset(host, AU6601_RESET_CMD | AU6601_RESET_DATA);
1017 
1018 	alcor_write8(priv, 0, AU6601_DETECT_STATUS);
1019 
1020 	alcor_write8(priv, 0, AU6601_OUTPUT_ENABLE);
1021 	alcor_write8(priv, 0, AU6601_POWER_CONTROL);
1022 
1023 	alcor_write8(priv, 0, AU6601_OPT);
1024 }
1025 
1026 static void alcor_init_mmc(struct alcor_sdmmc_host *host)
1027 {
1028 	struct mmc_host *mmc = host->mmc;
1029 
1030 	mmc->f_min = AU6601_MIN_CLOCK;
1031 	mmc->f_max = AU6601_MAX_CLOCK;
1032 	mmc->ocr_avail = MMC_VDD_33_34;
1033 	mmc->caps = MMC_CAP_4_BIT_DATA | MMC_CAP_SD_HIGHSPEED
1034 		| MMC_CAP_UHS_SDR12 | MMC_CAP_UHS_SDR25 | MMC_CAP_UHS_SDR50
1035 		| MMC_CAP_UHS_SDR104 | MMC_CAP_UHS_DDR50;
1036 	mmc->caps2 = MMC_CAP2_NO_SDIO;
1037 	mmc->ops = &alcor_sdc_ops;
1038 
1039 	/* The hardware does DMA data transfer of 4096 bytes to/from a single
1040 	 * buffer address. Scatterlists are not supported, but upon DMA
1041 	 * completion (signalled via IRQ), the original vendor driver does
1042 	 * then immediately set up another DMA transfer of the next 4096
1043 	 * bytes.
1044 	 *
1045 	 * This means that we need to handle the I/O in 4096 byte chunks.
1046 	 * Lacking a way to limit the sglist entries to 4096 bytes, we instead
1047 	 * impose that only one segment is provided, with maximum size 4096,
1048 	 * which also happens to be the minimum size. This means that the
1049 	 * single-entry sglist handled by this driver can be handed directly
1050 	 * to the hardware, nice and simple.
1051 	 *
1052 	 * Unfortunately though, that means we only do 4096 bytes I/O per
1053 	 * MMC command. A future improvement would be to make the driver
1054 	 * accept sg lists and entries of any size, and simply iterate
1055 	 * through them 4096 bytes at a time.
1056 	 */
1057 	mmc->max_segs = AU6601_MAX_DMA_SEGMENTS;
1058 	mmc->max_seg_size = AU6601_MAX_DMA_BLOCK_SIZE;
1059 	mmc->max_req_size = mmc->max_seg_size;
1060 }
1061 
1062 static int alcor_pci_sdmmc_drv_probe(struct platform_device *pdev)
1063 {
1064 	struct alcor_pci_priv *priv = pdev->dev.platform_data;
1065 	struct mmc_host *mmc;
1066 	struct alcor_sdmmc_host *host;
1067 	int ret;
1068 
1069 	mmc = mmc_alloc_host(sizeof(*host), &pdev->dev);
1070 	if (!mmc) {
1071 		dev_err(&pdev->dev, "Can't allocate MMC\n");
1072 		return -ENOMEM;
1073 	}
1074 
1075 	host = mmc_priv(mmc);
1076 	host->mmc = mmc;
1077 	host->dev = &pdev->dev;
1078 	host->cur_power_mode = MMC_POWER_UNDEFINED;
1079 	host->alcor_pci = priv;
1080 
1081 	/* make sure irqs are disabled */
1082 	alcor_write32(priv, 0, AU6601_REG_INT_ENABLE);
1083 	alcor_write32(priv, 0, AU6601_MS_INT_ENABLE);
1084 
1085 	ret = devm_request_threaded_irq(&pdev->dev, priv->irq,
1086 			alcor_irq, alcor_irq_thread, IRQF_SHARED,
1087 			DRV_NAME_ALCOR_PCI_SDMMC, host);
1088 
1089 	if (ret) {
1090 		dev_err(&pdev->dev, "Failed to get irq for data line\n");
1091 		return ret;
1092 	}
1093 
1094 	mutex_init(&host->cmd_mutex);
1095 	INIT_DELAYED_WORK(&host->timeout_work, alcor_timeout_timer);
1096 
1097 	alcor_init_mmc(host);
1098 	alcor_hw_init(host);
1099 
1100 	dev_set_drvdata(&pdev->dev, host);
1101 	mmc_add_host(mmc);
1102 	return 0;
1103 }
1104 
1105 static int alcor_pci_sdmmc_drv_remove(struct platform_device *pdev)
1106 {
1107 	struct alcor_sdmmc_host *host = dev_get_drvdata(&pdev->dev);
1108 
1109 	if (cancel_delayed_work_sync(&host->timeout_work))
1110 		alcor_request_complete(host, 0);
1111 
1112 	alcor_hw_uninit(host);
1113 	mmc_remove_host(host->mmc);
1114 	mmc_free_host(host->mmc);
1115 
1116 	return 0;
1117 }
1118 
1119 #ifdef CONFIG_PM_SLEEP
1120 static int alcor_pci_sdmmc_suspend(struct device *dev)
1121 {
1122 	struct alcor_sdmmc_host *host = dev_get_drvdata(dev);
1123 
1124 	if (cancel_delayed_work_sync(&host->timeout_work))
1125 		alcor_request_complete(host, 0);
1126 
1127 	alcor_hw_uninit(host);
1128 
1129 	return 0;
1130 }
1131 
1132 static int alcor_pci_sdmmc_resume(struct device *dev)
1133 {
1134 	struct alcor_sdmmc_host *host = dev_get_drvdata(dev);
1135 
1136 	alcor_hw_init(host);
1137 
1138 	return 0;
1139 }
1140 #endif /* CONFIG_PM_SLEEP */
1141 
1142 static SIMPLE_DEV_PM_OPS(alcor_mmc_pm_ops, alcor_pci_sdmmc_suspend,
1143 			 alcor_pci_sdmmc_resume);
1144 
1145 static const struct platform_device_id alcor_pci_sdmmc_ids[] = {
1146 	{
1147 		.name = DRV_NAME_ALCOR_PCI_SDMMC,
1148 	}, {
1149 		/* sentinel */
1150 	}
1151 };
1152 MODULE_DEVICE_TABLE(platform, alcor_pci_sdmmc_ids);
1153 
1154 static struct platform_driver alcor_pci_sdmmc_driver = {
1155 	.probe		= alcor_pci_sdmmc_drv_probe,
1156 	.remove		= alcor_pci_sdmmc_drv_remove,
1157 	.id_table	= alcor_pci_sdmmc_ids,
1158 	.driver		= {
1159 		.name	= DRV_NAME_ALCOR_PCI_SDMMC,
1160 		.pm	= &alcor_mmc_pm_ops
1161 	},
1162 };
1163 module_platform_driver(alcor_pci_sdmmc_driver);
1164 
1165 MODULE_AUTHOR("Oleksij Rempel <linux@rempel-privat.de>");
1166 MODULE_DESCRIPTION("PCI driver for Alcor Micro AU6601 Secure Digital Host Controller Interface");
1167 MODULE_LICENSE("GPL");
1168