xref: /freebsd/sys/dev/ida/ida.c (revision 732a02b4e77866604a120a275c082bb6221bd2ff)
1 /*-
2  * SPDX-License-Identifier: BSD-2-Clause-FreeBSD
3  *
4  * Copyright (c) 1999,2000 Jonathan Lemon
5  * All rights reserved.
6  *
7  # Derived from the original IDA Compaq RAID driver, which is
8  * Copyright (c) 1996, 1997, 1998, 1999
9  *    Mark Dawson and David James. All rights reserved.
10  *
11  * Redistribution and use in source and binary forms, with or without
12  * modification, are permitted provided that the following conditions
13  * are met:
14  * 1. Redistributions of source code must retain the above copyright
15  *    notice, this list of conditions and the following disclaimer.
16  * 2. Redistributions in binary form must reproduce the above copyright
17  *    notice, this list of conditions and the following disclaimer in the
18  *    documentation and/or other materials provided with the distribution.
19  *
20  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
21  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
22  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
23  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
24  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
25  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
26  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
27  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
28  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
29  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
30  * SUCH DAMAGE.
31  */
32 
33 #include <sys/cdefs.h>
34 __FBSDID("$FreeBSD$");
35 
36 /*
37  * Generic driver for Compaq SMART RAID adapters.
38  */
39 
40 #include <sys/param.h>
41 #include <sys/kernel.h>
42 #include <sys/systm.h>
43 #include <sys/lock.h>
44 #include <sys/malloc.h>
45 #include <sys/mutex.h>
46 #include <sys/stat.h>
47 
48 #include <sys/bio.h>
49 #include <sys/bus.h>
50 #include <sys/conf.h>
51 #include <sys/endian.h>
52 
53 #include <machine/bus.h>
54 #include <sys/rman.h>
55 
56 #include <geom/geom_disk.h>
57 
58 #include <dev/ida/idareg.h>
59 #include <dev/ida/idavar.h>
60 #include <dev/ida/idaio.h>
61 
62 /* prototypes */
63 static int ida_alloc_qcbs(struct ida_softc *ida);
64 static void ida_done(struct ida_softc *ida, struct ida_qcb *qcb);
65 static void ida_start(struct ida_softc *ida);
66 static void ida_startio(struct ida_softc *ida);
67 static void ida_startup(void *arg);
68 static void ida_timeout(void *arg);
69 static int ida_wait(struct ida_softc *ida, struct ida_qcb *qcb);
70 
71 static d_ioctl_t ida_ioctl;
72 static struct cdevsw ida_cdevsw = {
73 	.d_version =	D_VERSION,
74 	.d_ioctl =	ida_ioctl,
75 	.d_name =	"ida",
76 };
77 
78 void
79 ida_free(struct ida_softc *ida)
80 {
81 	int i;
82 
83 	if (ida->ih != NULL)
84 		bus_teardown_intr(ida->dev, ida->irq, ida->ih);
85 
86 	mtx_lock(&ida->lock);
87 	callout_stop(&ida->ch);
88 	mtx_unlock(&ida->lock);
89 	callout_drain(&ida->ch);
90 
91 	if (ida->buffer_dmat) {
92 		for (i = 0; i < IDA_QCB_MAX; i++)
93 			bus_dmamap_destroy(ida->buffer_dmat, ida->qcbs[i].dmamap);
94 		bus_dma_tag_destroy(ida->buffer_dmat);
95 	}
96 
97 	if (ida->hwqcb_dmat) {
98 		if (ida->hwqcb_busaddr)
99 			bus_dmamap_unload(ida->hwqcb_dmat, ida->hwqcb_dmamap);
100 		if (ida->hwqcbs)
101 			bus_dmamem_free(ida->hwqcb_dmat, ida->hwqcbs,
102 			    ida->hwqcb_dmamap);
103 		bus_dma_tag_destroy(ida->hwqcb_dmat);
104 	}
105 
106 	if (ida->qcbs != NULL)
107 		free(ida->qcbs, M_DEVBUF);
108 
109 	if (ida->irq != NULL)
110 		bus_release_resource(ida->dev, ida->irq_res_type,
111 		    0, ida->irq);
112 
113 	if (ida->parent_dmat != NULL)
114 		bus_dma_tag_destroy(ida->parent_dmat);
115 
116 	if (ida->regs != NULL)
117 		bus_release_resource(ida->dev, ida->regs_res_type,
118 		    ida->regs_res_id, ida->regs);
119 
120 	mtx_destroy(&ida->lock);
121 }
122 
123 /*
124  * record bus address from bus_dmamap_load
125  */
126 static void
127 ida_dma_map_cb(void *arg, bus_dma_segment_t *segs, int nseg, int error)
128 {
129 	bus_addr_t *baddr;
130 
131 	baddr = (bus_addr_t *)arg;
132 	*baddr = segs->ds_addr;
133 }
134 
135 static __inline struct ida_qcb *
136 ida_get_qcb(struct ida_softc *ida)
137 {
138 	struct ida_qcb *qcb;
139 
140 	if ((qcb = SLIST_FIRST(&ida->free_qcbs)) != NULL) {
141 		SLIST_REMOVE_HEAD(&ida->free_qcbs, link.sle);
142 		bzero(qcb->hwqcb, sizeof(struct ida_hdr) + sizeof(struct ida_req));
143 	}
144 	return (qcb);
145 }
146 
147 static __inline void
148 ida_free_qcb(struct ida_softc *ida, struct ida_qcb *qcb)
149 {
150 
151 	qcb->state = QCB_FREE;
152 	qcb->buf = NULL;
153 	qcb->error = 0;
154 	SLIST_INSERT_HEAD(&ida->free_qcbs, qcb, link.sle);
155 }
156 
157 static __inline bus_addr_t
158 idahwqcbvtop(struct ida_softc *ida, struct ida_hardware_qcb *hwqcb)
159 {
160 	return (ida->hwqcb_busaddr +
161 	    ((bus_addr_t)hwqcb - (bus_addr_t)ida->hwqcbs));
162 }
163 
164 static __inline struct ida_qcb *
165 idahwqcbptov(struct ida_softc *ida, bus_addr_t hwqcb_addr)
166 {
167 	struct ida_hardware_qcb *hwqcb;
168 
169 	hwqcb = (struct ida_hardware_qcb *)
170 	    ((bus_addr_t)ida->hwqcbs + (hwqcb_addr - ida->hwqcb_busaddr));
171 	return (hwqcb->qcb);
172 }
173 
174 static int
175 ida_alloc_qcbs(struct ida_softc *ida)
176 {
177 	struct ida_qcb *qcb;
178 	int error, i;
179 
180 	for (i = 0; i < IDA_QCB_MAX; i++) {
181 		qcb = &ida->qcbs[i];
182 
183 		error = bus_dmamap_create(ida->buffer_dmat, /*flags*/0, &qcb->dmamap);
184 		if (error != 0)
185 			return (error);
186 
187 		qcb->ida = ida;
188 		qcb->flags = QCB_FREE;
189 		qcb->hwqcb = &ida->hwqcbs[i];
190 		qcb->hwqcb->qcb = qcb;
191 		qcb->hwqcb_busaddr = idahwqcbvtop(ida, qcb->hwqcb);
192 		SLIST_INSERT_HEAD(&ida->free_qcbs, qcb, link.sle);
193 	}
194 	return (0);
195 }
196 
197 int
198 ida_setup(struct ida_softc *ida)
199 {
200 	struct ida_controller_info cinfo;
201 	device_t child;
202 	int error, i, unit;
203 
204 	SLIST_INIT(&ida->free_qcbs);
205 	STAILQ_INIT(&ida->qcb_queue);
206 	bioq_init(&ida->bio_queue);
207 
208 	ida->qcbs = (struct ida_qcb *)
209 	    malloc(IDA_QCB_MAX * sizeof(struct ida_qcb), M_DEVBUF,
210 		M_NOWAIT | M_ZERO);
211 	if (ida->qcbs == NULL)
212 		return (ENOMEM);
213 
214 	/*
215 	 * Create our DMA tags
216 	 */
217 
218 	/* DMA tag for our hardware QCB structures */
219 	error = bus_dma_tag_create(
220 		/* parent	*/ ida->parent_dmat,
221 		/* alignment	*/ 1,
222 		/* boundary	*/ 0,
223 		/* lowaddr	*/ BUS_SPACE_MAXADDR,
224 		/* highaddr	*/ BUS_SPACE_MAXADDR,
225 		/* filter	*/ NULL,
226 		/* filterarg	*/ NULL,
227 		/* maxsize	*/ IDA_QCB_MAX * sizeof(struct ida_hardware_qcb),
228 		/* nsegments	*/ 1,
229 		/* maxsegsz	*/ BUS_SPACE_MAXSIZE_32BIT,
230 		/* flags	*/ 0,
231 		/* lockfunc	*/ NULL,
232 		/* lockarg	*/ NULL,
233 		&ida->hwqcb_dmat);
234 	if (error)
235 		return (ENOMEM);
236 
237 	/* DMA tag for mapping buffers into device space */
238 	error = bus_dma_tag_create(
239 		/* parent 	*/ ida->parent_dmat,
240 		/* alignment	*/ 1,
241 		/* boundary	*/ 0,
242 		/* lowaddr	*/ BUS_SPACE_MAXADDR,
243 		/* highaddr	*/ BUS_SPACE_MAXADDR,
244 		/* filter	*/ NULL,
245 		/* filterarg	*/ NULL,
246 		/* maxsize	*/ DFLTPHYS,
247 		/* nsegments	*/ IDA_NSEG,
248 		/* maxsegsz	*/ BUS_SPACE_MAXSIZE_32BIT,
249 		/* flags	*/ 0,
250 		/* lockfunc	*/ busdma_lock_mutex,
251 		/* lockarg	*/ &Giant,
252 		&ida->buffer_dmat);
253 	if (error)
254 		return (ENOMEM);
255 
256 	/* Allocation of hardware QCBs */
257 	/* XXX allocation is rounded to hardware page size */
258 	error = bus_dmamem_alloc(ida->hwqcb_dmat,
259 	    (void **)&ida->hwqcbs, BUS_DMA_NOWAIT, &ida->hwqcb_dmamap);
260 	if (error)
261 		return (ENOMEM);
262 
263 	/* And permanently map them in */
264 	bus_dmamap_load(ida->hwqcb_dmat, ida->hwqcb_dmamap,
265 	    ida->hwqcbs, IDA_QCB_MAX * sizeof(struct ida_hardware_qcb),
266 	    ida_dma_map_cb, &ida->hwqcb_busaddr, /*flags*/0);
267 
268 	bzero(ida->hwqcbs, IDA_QCB_MAX * sizeof(struct ida_hardware_qcb));
269 
270 	error = ida_alloc_qcbs(ida);
271 	if (error)
272 		return (error);
273 
274 	mtx_lock(&ida->lock);
275 	ida->cmd.int_enable(ida, 0);
276 
277 	error = ida_command(ida, CMD_GET_CTRL_INFO, &cinfo, sizeof(cinfo),
278 	    IDA_CONTROLLER, 0, DMA_DATA_IN);
279 	if (error) {
280 		mtx_unlock(&ida->lock);
281 		device_printf(ida->dev, "CMD_GET_CTRL_INFO failed.\n");
282 		return (error);
283 	}
284 
285 	device_printf(ida->dev, "drives=%d firm_rev=%c%c%c%c\n",
286 	    cinfo.num_drvs, cinfo.firm_rev[0], cinfo.firm_rev[1],
287 	    cinfo.firm_rev[2], cinfo.firm_rev[3]);
288 
289 	if (ida->flags & IDA_FIRMWARE) {
290 		int data;
291 
292 		error = ida_command(ida, CMD_START_FIRMWARE,
293 		    &data, sizeof(data), IDA_CONTROLLER, 0, DMA_DATA_IN);
294 		if (error) {
295 			mtx_unlock(&ida->lock);
296 			device_printf(ida->dev, "CMD_START_FIRMWARE failed.\n");
297 			return (error);
298 		}
299 	}
300 
301 	ida->cmd.int_enable(ida, 1);
302 	ida->flags |= IDA_ATTACHED;
303 	mtx_unlock(&ida->lock);
304 
305 	for (i = 0; i < cinfo.num_drvs; i++) {
306 		child = device_add_child(ida->dev, /*"idad"*/NULL, -1);
307 		if (child != NULL)
308 			device_set_ivars(child, (void *)(intptr_t)i);
309 	}
310 
311 	ida->ich.ich_func = ida_startup;
312 	ida->ich.ich_arg = ida;
313 	if (config_intrhook_establish(&ida->ich) != 0) {
314 		device_delete_children(ida->dev);
315 		device_printf(ida->dev, "Cannot establish configuration hook\n");
316 		return (error);
317 	}
318 
319 	unit = device_get_unit(ida->dev);
320 	ida->ida_dev_t = make_dev(&ida_cdevsw, unit,
321 				 UID_ROOT, GID_OPERATOR, S_IRUSR | S_IWUSR,
322 				 "ida%d", unit);
323 	ida->ida_dev_t->si_drv1 = ida;
324 
325 	return (0);
326 }
327 
328 static void
329 ida_startup(void *arg)
330 {
331 	struct ida_softc *ida;
332 
333 	ida = arg;
334 
335 	config_intrhook_disestablish(&ida->ich);
336 
337 	mtx_lock(&Giant);
338 	bus_generic_attach(ida->dev);
339 	mtx_unlock(&Giant);
340 }
341 
342 int
343 ida_detach(device_t dev)
344 {
345 	struct ida_softc *ida;
346 	int error;
347 
348 	ida = (struct ida_softc *)device_get_softc(dev);
349 
350 	error = bus_generic_detach(dev);
351 	if (error)
352 		return (error);
353 	error = device_delete_children(dev);
354 	if (error)
355 		return (error);
356 
357 	/*
358 	 * XXX
359 	 * before detaching, we must make sure that the system is
360 	 * quiescent; nothing mounted, no pending activity.
361 	 */
362 
363 	/*
364 	 * XXX
365 	 * now, how are we supposed to maintain a list of our drives?
366 	 * iterate over our "child devices"?
367 	 */
368 
369 	destroy_dev(ida->ida_dev_t);
370 	ida_free(ida);
371 	return (error);
372 }
373 
374 static void
375 ida_data_cb(void *arg, bus_dma_segment_t *segs, int nsegments, int error)
376 {
377 	struct ida_hardware_qcb *hwqcb;
378 	struct ida_softc *ida;
379 	struct ida_qcb *qcb;
380 	bus_dmasync_op_t op;
381 	int i;
382 
383 	qcb = arg;
384 	ida = qcb->ida;
385 	if (!dumping)
386 		mtx_assert(&ida->lock, MA_OWNED);
387 	if (error) {
388 		qcb->error = error;
389 		ida_done(ida, qcb);
390 		return;
391 	}
392 
393 	hwqcb = qcb->hwqcb;
394 	hwqcb->hdr.size = htole16((sizeof(struct ida_req) +
395 	    sizeof(struct ida_sgb) * IDA_NSEG) >> 2);
396 
397 	for (i = 0; i < nsegments; i++) {
398 		hwqcb->seg[i].addr = htole32(segs[i].ds_addr);
399 		hwqcb->seg[i].length = htole32(segs[i].ds_len);
400 	}
401 	hwqcb->req.sgcount = nsegments;
402 	if (qcb->flags & DMA_DATA_TRANSFER) {
403 		switch (qcb->flags & DMA_DATA_TRANSFER) {
404 		case DMA_DATA_TRANSFER:
405 			op = BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE;
406 			break;
407 		case DMA_DATA_IN:
408 			op = BUS_DMASYNC_PREREAD;
409 			break;
410 		default:
411 			KASSERT((qcb->flags & DMA_DATA_TRANSFER) ==
412 			    DMA_DATA_OUT, ("bad DMA data flags"));
413 			op = BUS_DMASYNC_PREWRITE;
414 			break;
415 		}
416 		bus_dmamap_sync(ida->buffer_dmat, qcb->dmamap, op);
417 	}
418 	bus_dmamap_sync(ida->hwqcb_dmat, ida->hwqcb_dmamap,
419 	    BUS_DMASYNC_PREWRITE | BUS_DMASYNC_PREREAD);
420 
421 	STAILQ_INSERT_TAIL(&ida->qcb_queue, qcb, link.stqe);
422 	ida_start(ida);
423 	ida->flags &= ~IDA_QFROZEN;
424 }
425 
426 static int
427 ida_map_qcb(struct ida_softc *ida, struct ida_qcb *qcb, void *data,
428     bus_size_t datasize)
429 {
430 	int error, flags;
431 
432 	if (ida->flags & IDA_INTERRUPTS)
433 		flags = BUS_DMA_WAITOK;
434 	else
435 		flags = BUS_DMA_NOWAIT;
436 	error = bus_dmamap_load(ida->buffer_dmat, qcb->dmamap, data, datasize,
437 	    ida_data_cb, qcb, flags);
438 	if (error == EINPROGRESS) {
439 		ida->flags |= IDA_QFROZEN;
440 		error = 0;
441 	}
442 	return (error);
443 }
444 
445 int
446 ida_command(struct ida_softc *ida, int command, void *data, int datasize,
447 	int drive, u_int32_t pblkno, int flags)
448 {
449 	struct ida_hardware_qcb *hwqcb;
450 	struct ida_qcb *qcb;
451 	int error;
452 
453 	if (!dumping)
454 		mtx_assert(&ida->lock, MA_OWNED);
455 	qcb = ida_get_qcb(ida);
456 
457 	if (qcb == NULL) {
458 		device_printf(ida->dev, "out of QCBs\n");
459 		return (EAGAIN);
460 	}
461 
462 	qcb->flags = flags | IDA_COMMAND;
463 	hwqcb = qcb->hwqcb;
464 	hwqcb->hdr.drive = drive;
465 	hwqcb->req.blkno = htole32(pblkno);
466 	hwqcb->req.bcount = htole16(howmany(datasize, DEV_BSIZE));
467 	hwqcb->req.command = command;
468 
469 	error = ida_map_qcb(ida, qcb, data, datasize);
470 	if (error == 0) {
471 		error = ida_wait(ida, qcb);
472 		/* Don't free QCB on a timeout in case it later completes. */
473 		if (error)
474 			return (error);
475 		error = qcb->error;
476 	}
477 
478 	/* XXX should have status returned here? */
479 	/* XXX have "status pointer" area in QCB? */
480 
481 	ida_free_qcb(ida, qcb);
482 	return (error);
483 }
484 
485 void
486 ida_submit_buf(struct ida_softc *ida, struct bio *bp)
487 {
488 	mtx_lock(&ida->lock);
489 	bioq_insert_tail(&ida->bio_queue, bp);
490 	ida_startio(ida);
491 	mtx_unlock(&ida->lock);
492 }
493 
494 static void
495 ida_startio(struct ida_softc *ida)
496 {
497 	struct ida_hardware_qcb *hwqcb;
498 	struct ida_qcb *qcb;
499 	struct idad_softc *drv;
500 	struct bio *bp;
501 	int error;
502 
503 	mtx_assert(&ida->lock, MA_OWNED);
504 	for (;;) {
505 		if (ida->flags & IDA_QFROZEN)
506 			return;
507 		bp = bioq_first(&ida->bio_queue);
508 		if (bp == NULL)
509 			return;				/* no more buffers */
510 
511 		qcb = ida_get_qcb(ida);
512 		if (qcb == NULL)
513 			return;				/* out of resources */
514 
515 		bioq_remove(&ida->bio_queue, bp);
516 		qcb->buf = bp;
517 		qcb->flags = bp->bio_cmd == BIO_READ ? DMA_DATA_IN : DMA_DATA_OUT;
518 
519 		hwqcb = qcb->hwqcb;
520 		drv = bp->bio_driver1;
521 		hwqcb->hdr.drive = drv->drive;
522 		hwqcb->req.blkno = bp->bio_pblkno;
523 		hwqcb->req.bcount = howmany(bp->bio_bcount, DEV_BSIZE);
524 		hwqcb->req.command = bp->bio_cmd == BIO_READ ? CMD_READ : CMD_WRITE;
525 
526 		error = ida_map_qcb(ida, qcb, bp->bio_data, bp->bio_bcount);
527 		if (error) {
528 			qcb->error = error;
529 			ida_done(ida, qcb);
530 		}
531 	}
532 }
533 
534 static void
535 ida_start(struct ida_softc *ida)
536 {
537 	struct ida_qcb *qcb;
538 
539 	if (!dumping)
540 		mtx_assert(&ida->lock, MA_OWNED);
541 	while ((qcb = STAILQ_FIRST(&ida->qcb_queue)) != NULL) {
542 		if (ida->cmd.fifo_full(ida))
543 			break;
544 		STAILQ_REMOVE_HEAD(&ida->qcb_queue, link.stqe);
545 		/*
546 		 * XXX
547 		 * place the qcb on an active list?
548 		 */
549 
550 		/* Set a timeout. */
551 		if (!ida->qactive && !dumping)
552 			callout_reset(&ida->ch, hz * 5, ida_timeout, ida);
553 		ida->qactive++;
554 
555 		qcb->state = QCB_ACTIVE;
556 		ida->cmd.submit(ida, qcb);
557 	}
558 }
559 
560 static int
561 ida_wait(struct ida_softc *ida, struct ida_qcb *qcb)
562 {
563 	struct ida_qcb *qcb_done = NULL;
564 	bus_addr_t completed;
565 	int delay;
566 
567 	if (!dumping)
568 		mtx_assert(&ida->lock, MA_OWNED);
569 	if (ida->flags & IDA_INTERRUPTS) {
570 		if (mtx_sleep(qcb, &ida->lock, PRIBIO, "idacmd", 5 * hz)) {
571 			qcb->state = QCB_TIMEDOUT;
572 			return (ETIMEDOUT);
573 		}
574 		return (0);
575 	}
576 
577 again:
578 	delay = 5 * 1000 * 100;			/* 5 sec delay */
579 	while ((completed = ida->cmd.done(ida)) == 0) {
580 		if (delay-- == 0) {
581 			qcb->state = QCB_TIMEDOUT;
582 			return (ETIMEDOUT);
583 		}
584 		DELAY(10);
585 	}
586 
587 	qcb_done = idahwqcbptov(ida, completed & ~3);
588 	if (qcb_done != qcb)
589 		goto again;
590 	ida_done(ida, qcb);
591 	return (0);
592 }
593 
594 void
595 ida_intr(void *data)
596 {
597 	struct ida_softc *ida;
598 	struct ida_qcb *qcb;
599 	bus_addr_t completed;
600 
601 	ida = (struct ida_softc *)data;
602 
603 	mtx_lock(&ida->lock);
604 	if (ida->cmd.int_pending(ida) == 0) {
605 		mtx_unlock(&ida->lock);
606 		return;				/* not our interrupt */
607 	}
608 
609 	while ((completed = ida->cmd.done(ida)) != 0) {
610 		qcb = idahwqcbptov(ida, completed & ~3);
611 
612 		if (qcb == NULL || qcb->state != QCB_ACTIVE) {
613 			device_printf(ida->dev,
614 			    "ignoring completion %jx\n", (intmax_t)completed);
615 			continue;
616 		}
617 		/* Handle "Bad Command List" errors. */
618 		if ((completed & 3) && (qcb->hwqcb->req.error == 0))
619 			qcb->hwqcb->req.error = CMD_REJECTED;
620 		ida_done(ida, qcb);
621 	}
622 	ida_startio(ida);
623 	mtx_unlock(&ida->lock);
624 }
625 
626 /*
627  * should switch out command type; may be status, not just I/O.
628  */
629 static void
630 ida_done(struct ida_softc *ida, struct ida_qcb *qcb)
631 {
632 	bus_dmasync_op_t op;
633 	int active, error = 0;
634 
635 	/*
636 	 * finish up command
637 	 */
638 	if (!dumping)
639 		mtx_assert(&ida->lock, MA_OWNED);
640 	active = (qcb->state != QCB_FREE);
641 	if (qcb->flags & DMA_DATA_TRANSFER && active) {
642 		switch (qcb->flags & DMA_DATA_TRANSFER) {
643 		case DMA_DATA_TRANSFER:
644 			op = BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE;
645 			break;
646 		case DMA_DATA_IN:
647 			op = BUS_DMASYNC_POSTREAD;
648 			break;
649 		default:
650 			KASSERT((qcb->flags & DMA_DATA_TRANSFER) ==
651 			    DMA_DATA_OUT, ("bad DMA data flags"));
652 			op = BUS_DMASYNC_POSTWRITE;
653 			break;
654 		}
655 		bus_dmamap_sync(ida->buffer_dmat, qcb->dmamap, op);
656 		bus_dmamap_unload(ida->buffer_dmat, qcb->dmamap);
657 	}
658 	if (active)
659 		bus_dmamap_sync(ida->hwqcb_dmat, ida->hwqcb_dmamap,
660 		    BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
661 
662 	if (qcb->hwqcb->req.error & SOFT_ERROR) {
663 		if (qcb->buf)
664 			device_printf(ida->dev, "soft %s error\n",
665 				qcb->buf->bio_cmd == BIO_READ ?
666 					"read" : "write");
667 		else
668 			device_printf(ida->dev, "soft error\n");
669 	}
670 	if (qcb->hwqcb->req.error & HARD_ERROR) {
671 		error = 1;
672 		if (qcb->buf)
673 			device_printf(ida->dev, "hard %s error\n",
674 				qcb->buf->bio_cmd == BIO_READ ?
675 					"read" : "write");
676 		else
677 			device_printf(ida->dev, "hard error\n");
678 	}
679 	if (qcb->hwqcb->req.error & CMD_REJECTED) {
680 		error = 1;
681 		device_printf(ida->dev, "invalid request\n");
682 	}
683 	if (qcb->error) {
684 		error = 1;
685 		device_printf(ida->dev, "request failed to map: %d\n", qcb->error);
686 	}
687 
688 	if (qcb->flags & IDA_COMMAND) {
689 		if (ida->flags & IDA_INTERRUPTS)
690 			wakeup(qcb);
691 		if (qcb->state == QCB_TIMEDOUT)
692 			ida_free_qcb(ida, qcb);
693 	} else {
694 		KASSERT(qcb->buf != NULL, ("ida_done(): qcb->buf is NULL!"));
695 		if (error)
696 			qcb->buf->bio_flags |= BIO_ERROR;
697 		idad_intr(qcb->buf);
698 		ida_free_qcb(ida, qcb);
699 	}
700 
701 	if (!active)
702 		return;
703 
704 	ida->qactive--;
705 	/* Reschedule or cancel timeout */
706 	if (ida->qactive)
707 		callout_reset(&ida->ch, hz * 5, ida_timeout, ida);
708 	else
709 		callout_stop(&ida->ch);
710 }
711 
712 static void
713 ida_timeout(void *arg)
714 {
715 	struct ida_softc *ida;
716 
717 	ida = (struct ida_softc *)arg;
718 	device_printf(ida->dev, "%s() qactive %d\n", __func__, ida->qactive);
719 
720 	if (ida->flags & IDA_INTERRUPTS)
721 		device_printf(ida->dev, "IDA_INTERRUPTS\n");
722 
723 	device_printf(ida->dev,	"\t   R_CMD_FIFO: %08x\n"
724 				"\t  R_DONE_FIFO: %08x\n"
725 				"\t   R_INT_MASK: %08x\n"
726 				"\t     R_STATUS: %08x\n"
727 				"\tR_INT_PENDING: %08x\n",
728 					ida_inl(ida, R_CMD_FIFO),
729 					ida_inl(ida, R_DONE_FIFO),
730 					ida_inl(ida, R_INT_MASK),
731 					ida_inl(ida, R_STATUS),
732 					ida_inl(ida, R_INT_PENDING));
733 
734 	return;
735 }
736 
737 /*
738  * IOCTL stuff follows.
739  */
740 struct cmd_info {
741 	int	cmd;
742 	int	len;
743 	int	flags;
744 };
745 static struct cmd_info *ida_cmd_lookup(int);
746 
747 static int
748 ida_ioctl (struct cdev *dev, u_long cmd, caddr_t addr, int32_t flag, struct thread *td)
749 {
750 	struct ida_softc *sc;
751 	struct ida_user_command *uc;
752 	struct cmd_info *ci;
753 	int len;
754 	int flags;
755 	int error;
756 	int data;
757 	void *daddr;
758 
759 	sc = (struct ida_softc *)dev->si_drv1;
760 	uc = (struct ida_user_command *)addr;
761 	error = 0;
762 
763 	switch (cmd) {
764 	case IDAIO_COMMAND:
765 		ci = ida_cmd_lookup(uc->command);
766 		if (ci == NULL) {
767 			error = EINVAL;
768 			break;
769 		}
770 		len = ci->len;
771 		flags = ci->flags;
772 		if (len)
773 			daddr = &uc->d.buf;
774 		else {
775 			daddr = &data;
776 			len = sizeof(data);
777 		}
778 		mtx_lock(&sc->lock);
779 		error = ida_command(sc, uc->command, daddr, len,
780 				    uc->drive, uc->blkno, flags);
781 		mtx_unlock(&sc->lock);
782 		break;
783 	default:
784 		error = ENOIOCTL;
785 		break;
786 	}
787 	return (error);
788 }
789 
790 static struct cmd_info ci_list[] = {
791 	{ CMD_GET_LOG_DRV_INFO,
792 			sizeof(struct ida_drive_info), DMA_DATA_IN },
793 	{ CMD_GET_CTRL_INFO,
794 			sizeof(struct ida_controller_info), DMA_DATA_IN },
795 	{ CMD_SENSE_DRV_STATUS,
796 			sizeof(struct ida_drive_status), DMA_DATA_IN },
797 	{ CMD_START_RECOVERY,		0, 0 },
798 	{ CMD_GET_PHYS_DRV_INFO,
799 			sizeof(struct ida_phys_drv_info), DMA_DATA_TRANSFER },
800 	{ CMD_BLINK_DRV_LEDS,
801 			sizeof(struct ida_blink_drv_leds), DMA_DATA_OUT },
802 	{ CMD_SENSE_DRV_LEDS,
803 			sizeof(struct ida_blink_drv_leds), DMA_DATA_IN },
804 	{ CMD_GET_LOG_DRV_EXT,
805 			sizeof(struct ida_drive_info_ext), DMA_DATA_IN },
806 	{ CMD_RESET_CTRL,		0, 0 },
807 	{ CMD_GET_CONFIG,		0, 0 },
808 	{ CMD_SET_CONFIG,		0, 0 },
809 	{ CMD_LABEL_LOG_DRV,
810 			sizeof(struct ida_label_logical), DMA_DATA_OUT },
811 	{ CMD_SET_SURFACE_DELAY,	0, 0 },
812 	{ CMD_SENSE_BUS_PARAMS,		0, 0 },
813 	{ CMD_SENSE_SUBSYS_INFO,	0, 0 },
814 	{ CMD_SENSE_SURFACE_ATS,	0, 0 },
815 	{ CMD_PASSTHROUGH,		0, 0 },
816 	{ CMD_RESET_SCSI_DEV,		0, 0 },
817 	{ CMD_PAUSE_BG_ACT,		0, 0 },
818 	{ CMD_RESUME_BG_ACT,		0, 0 },
819 	{ CMD_START_FIRMWARE,		0, 0 },
820 	{ CMD_SENSE_DRV_ERR_LOG,	0, 0 },
821 	{ CMD_START_CPM,		0, 0 },
822 	{ CMD_SENSE_CP,			0, 0 },
823 	{ CMD_STOP_CPM,			0, 0 },
824 	{ CMD_FLUSH_CACHE,		0, 0 },
825 	{ CMD_ACCEPT_MEDIA_EXCH,	0, 0 },
826 	{ 0, 0, 0 }
827 };
828 
829 static struct cmd_info *
830 ida_cmd_lookup (int command)
831 {
832 	struct cmd_info *ci;
833 
834 	ci = ci_list;
835 	while (ci->cmd) {
836 		if (ci->cmd == command)
837 			return (ci);
838 		ci++;
839 	}
840 	return (NULL);
841 }
842