xref: /freebsd/sys/dev/ata/ata-all.c (revision a3cf0ef5a295c885c895fabfd56470c0d1db322d)
1 /*-
2  * Copyright (c) 1998 - 2008 S�ren Schmidt <sos@FreeBSD.org>
3  * All rights reserved.
4  *
5  * Redistribution and use in source and binary forms, with or without
6  * modification, are permitted provided that the following conditions
7  * are met:
8  * 1. Redistributions of source code must retain the above copyright
9  *    notice, this list of conditions and the following disclaimer,
10  *    without modification, immediately at the beginning of the file.
11  * 2. Redistributions in binary form must reproduce the above copyright
12  *    notice, this list of conditions and the following disclaimer in the
13  *    documentation and/or other materials provided with the distribution.
14  *
15  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
16  * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
17  * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
18  * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
19  * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
20  * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
21  * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
22  * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
23  * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
24  * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
25  */
26 
27 #include <sys/cdefs.h>
28 __FBSDID("$FreeBSD$");
29 
30 #include "opt_ata.h"
31 #include <sys/param.h>
32 #include <sys/systm.h>
33 #include <sys/ata.h>
34 #include <sys/kernel.h>
35 #include <sys/module.h>
36 #include <sys/endian.h>
37 #include <sys/ctype.h>
38 #include <sys/conf.h>
39 #include <sys/bus.h>
40 #include <sys/bio.h>
41 #include <sys/malloc.h>
42 #include <sys/sysctl.h>
43 #include <sys/sema.h>
44 #include <sys/taskqueue.h>
45 #include <vm/uma.h>
46 #include <machine/stdarg.h>
47 #include <machine/resource.h>
48 #include <machine/bus.h>
49 #include <sys/rman.h>
50 #include <dev/ata/ata-all.h>
51 #include <dev/pci/pcivar.h>
52 #include <ata_if.h>
53 
54 #ifdef ATA_CAM
55 #include <cam/cam.h>
56 #include <cam/cam_ccb.h>
57 #include <cam/cam_sim.h>
58 #include <cam/cam_xpt_sim.h>
59 #include <cam/cam_debug.h>
60 #endif
61 
62 #ifndef ATA_CAM
63 /* device structure */
64 static  d_ioctl_t       ata_ioctl;
65 static struct cdevsw ata_cdevsw = {
66 	.d_version =    D_VERSION,
67 	.d_flags =      D_NEEDGIANT, /* we need this as newbus isn't mpsafe */
68 	.d_ioctl =      ata_ioctl,
69 	.d_name =       "ata",
70 };
71 #endif
72 
73 /* prototypes */
74 #ifndef ATA_CAM
75 static void ata_boot_attach(void);
76 static device_t ata_add_child(device_t, struct ata_device *, int);
77 #else
78 static void ataaction(struct cam_sim *sim, union ccb *ccb);
79 static void atapoll(struct cam_sim *sim);
80 #endif
81 static void ata_conn_event(void *, int);
82 static void bswap(int8_t *, int);
83 static void btrim(int8_t *, int);
84 static void bpack(int8_t *, int8_t *, int);
85 static void ata_interrupt_locked(void *data);
86 static void ata_periodic_poll(void *data);
87 
88 /* global vars */
89 MALLOC_DEFINE(M_ATA, "ata_generic", "ATA driver generic layer");
90 int (*ata_raid_ioctl_func)(u_long cmd, caddr_t data) = NULL;
91 struct intr_config_hook *ata_delayed_attach = NULL;
92 devclass_t ata_devclass;
93 uma_zone_t ata_request_zone;
94 uma_zone_t ata_composite_zone;
95 int ata_wc = 1;
96 int ata_setmax = 0;
97 int ata_dma_check_80pin = 1;
98 
99 /* local vars */
100 static int ata_dma = 1;
101 static int atapi_dma = 1;
102 
103 /* sysctl vars */
104 SYSCTL_NODE(_hw, OID_AUTO, ata, CTLFLAG_RD, 0, "ATA driver parameters");
105 TUNABLE_INT("hw.ata.ata_dma", &ata_dma);
106 SYSCTL_INT(_hw_ata, OID_AUTO, ata_dma, CTLFLAG_RDTUN, &ata_dma, 0,
107 	   "ATA disk DMA mode control");
108 TUNABLE_INT("hw.ata.ata_dma_check_80pin", &ata_dma_check_80pin);
109 SYSCTL_INT(_hw_ata, OID_AUTO, ata_dma_check_80pin,
110 	   CTLFLAG_RW, &ata_dma_check_80pin, 1,
111 	   "Check for 80pin cable before setting ATA DMA mode");
112 TUNABLE_INT("hw.ata.atapi_dma", &atapi_dma);
113 SYSCTL_INT(_hw_ata, OID_AUTO, atapi_dma, CTLFLAG_RDTUN, &atapi_dma, 0,
114 	   "ATAPI device DMA mode control");
115 TUNABLE_INT("hw.ata.wc", &ata_wc);
116 SYSCTL_INT(_hw_ata, OID_AUTO, wc, CTLFLAG_RDTUN, &ata_wc, 0,
117 	   "ATA disk write caching");
118 TUNABLE_INT("hw.ata.setmax", &ata_setmax);
119 SYSCTL_INT(_hw_ata, OID_AUTO, setmax, CTLFLAG_RDTUN, &ata_setmax, 0,
120 	   "ATA disk set max native address");
121 
122 /*
123  * newbus device interface related functions
124  */
125 int
126 ata_probe(device_t dev)
127 {
128     return 0;
129 }
130 
131 int
132 ata_attach(device_t dev)
133 {
134     struct ata_channel *ch = device_get_softc(dev);
135     int error, rid;
136 #ifdef ATA_CAM
137     struct cam_devq *devq;
138     const char *res;
139     char buf[64];
140     int i, mode;
141 #endif
142 
143     /* check that we have a virgin channel to attach */
144     if (ch->r_irq)
145 	return EEXIST;
146 
147     /* initialize the softc basics */
148     ch->dev = dev;
149     ch->state = ATA_IDLE;
150     bzero(&ch->state_mtx, sizeof(struct mtx));
151     mtx_init(&ch->state_mtx, "ATA state lock", NULL, MTX_DEF);
152     bzero(&ch->queue_mtx, sizeof(struct mtx));
153     mtx_init(&ch->queue_mtx, "ATA queue lock", NULL, MTX_DEF);
154     TAILQ_INIT(&ch->ata_queue);
155     TASK_INIT(&ch->conntask, 0, ata_conn_event, dev);
156 #ifdef ATA_CAM
157 	for (i = 0; i < 16; i++) {
158 		ch->user[i].mode = 0;
159 		snprintf(buf, sizeof(buf), "dev%d.mode", i);
160 		if (resource_string_value(device_get_name(dev),
161 		    device_get_unit(dev), buf, &res) == 0)
162 			mode = ata_str2mode(res);
163 		else if (resource_string_value(device_get_name(dev),
164 		    device_get_unit(dev), "mode", &res) == 0)
165 			mode = ata_str2mode(res);
166 		else
167 			mode = -1;
168 		if (mode >= 0)
169 			ch->user[i].mode = mode;
170 		if (ch->flags & ATA_SATA)
171 			ch->user[i].bytecount = 8192;
172 		else
173 			ch->user[i].bytecount = MAXPHYS;
174 		ch->curr[i] = ch->user[i];
175 	}
176 #endif
177 	callout_init(&ch->poll_callout, 1);
178 
179     /* reset the controller HW, the channel and device(s) */
180     while (ATA_LOCKING(dev, ATA_LF_LOCK) != ch->unit)
181 	pause("ataatch", 1);
182 #ifndef ATA_CAM
183     ATA_RESET(dev);
184 #endif
185     ATA_LOCKING(dev, ATA_LF_UNLOCK);
186 
187     /* allocate DMA resources if DMA HW present*/
188     if (ch->dma.alloc)
189 	ch->dma.alloc(dev);
190 
191     /* setup interrupt delivery */
192     rid = ATA_IRQ_RID;
193     ch->r_irq = bus_alloc_resource_any(dev, SYS_RES_IRQ, &rid,
194 				       RF_SHAREABLE | RF_ACTIVE);
195     if (!ch->r_irq) {
196 	device_printf(dev, "unable to allocate interrupt\n");
197 	return ENXIO;
198     }
199     if ((error = bus_setup_intr(dev, ch->r_irq, ATA_INTR_FLAGS, NULL,
200 				ata_interrupt, ch, &ch->ih))) {
201 	bus_release_resource(dev, SYS_RES_IRQ, rid, ch->r_irq);
202 	device_printf(dev, "unable to setup interrupt\n");
203 	return error;
204     }
205     if (ch->flags & ATA_PERIODIC_POLL)
206 	callout_reset(&ch->poll_callout, hz, ata_periodic_poll, ch);
207 
208 #ifndef ATA_CAM
209     /* probe and attach devices on this channel unless we are in early boot */
210     if (!ata_delayed_attach)
211 	ata_identify(dev);
212     return (0);
213 #else
214 	mtx_lock(&ch->state_mtx);
215 	/* Create the device queue for our SIM. */
216 	devq = cam_simq_alloc(1);
217 	if (devq == NULL) {
218 		device_printf(dev, "Unable to allocate simq\n");
219 		error = ENOMEM;
220 		goto err1;
221 	}
222 	/* Construct SIM entry */
223 	ch->sim = cam_sim_alloc(ataaction, atapoll, "ata", ch,
224 	    device_get_unit(dev), &ch->state_mtx, 1, 0, devq);
225 	if (ch->sim == NULL) {
226 		device_printf(dev, "unable to allocate sim\n");
227 		cam_simq_free(devq);
228 		error = ENOMEM;
229 		goto err1;
230 	}
231 	if (xpt_bus_register(ch->sim, dev, 0) != CAM_SUCCESS) {
232 		device_printf(dev, "unable to register xpt bus\n");
233 		error = ENXIO;
234 		goto err2;
235 	}
236 	if (xpt_create_path(&ch->path, /*periph*/NULL, cam_sim_path(ch->sim),
237 	    CAM_TARGET_WILDCARD, CAM_LUN_WILDCARD) != CAM_REQ_CMP) {
238 		device_printf(dev, "unable to create path\n");
239 		error = ENXIO;
240 		goto err3;
241 	}
242 	mtx_unlock(&ch->state_mtx);
243 	return (0);
244 
245 err3:
246 	xpt_bus_deregister(cam_sim_path(ch->sim));
247 err2:
248 	cam_sim_free(ch->sim, /*free_devq*/TRUE);
249 	ch->sim = NULL;
250 err1:
251 	bus_release_resource(dev, SYS_RES_IRQ, rid, ch->r_irq);
252 	mtx_unlock(&ch->state_mtx);
253 	if (ch->flags & ATA_PERIODIC_POLL)
254 		callout_drain(&ch->poll_callout);
255 	return (error);
256 #endif
257 }
258 
259 int
260 ata_detach(device_t dev)
261 {
262     struct ata_channel *ch = device_get_softc(dev);
263 #ifndef ATA_CAM
264     device_t *children;
265     int nchildren, i;
266 #endif
267 
268     /* check that we have a valid channel to detach */
269     if (!ch->r_irq)
270 	return ENXIO;
271 
272     /* grap the channel lock so no new requests gets launched */
273     mtx_lock(&ch->state_mtx);
274     ch->state |= ATA_STALL_QUEUE;
275     mtx_unlock(&ch->state_mtx);
276     if (ch->flags & ATA_PERIODIC_POLL)
277 	callout_drain(&ch->poll_callout);
278 
279 #ifndef ATA_CAM
280     /* detach & delete all children */
281     if (!device_get_children(dev, &children, &nchildren)) {
282 	for (i = 0; i < nchildren; i++)
283 	    if (children[i])
284 		device_delete_child(dev, children[i]);
285 	free(children, M_TEMP);
286     }
287 #endif
288     taskqueue_drain(taskqueue_thread, &ch->conntask);
289 
290 #ifdef ATA_CAM
291 	mtx_lock(&ch->state_mtx);
292 	xpt_async(AC_LOST_DEVICE, ch->path, NULL);
293 	xpt_free_path(ch->path);
294 	xpt_bus_deregister(cam_sim_path(ch->sim));
295 	cam_sim_free(ch->sim, /*free_devq*/TRUE);
296 	ch->sim = NULL;
297 	mtx_unlock(&ch->state_mtx);
298 #endif
299 
300     /* release resources */
301     bus_teardown_intr(dev, ch->r_irq, ch->ih);
302     bus_release_resource(dev, SYS_RES_IRQ, ATA_IRQ_RID, ch->r_irq);
303     ch->r_irq = NULL;
304 
305     /* free DMA resources if DMA HW present*/
306     if (ch->dma.free)
307 	ch->dma.free(dev);
308 
309     mtx_destroy(&ch->state_mtx);
310     mtx_destroy(&ch->queue_mtx);
311     return 0;
312 }
313 
314 static void
315 ata_conn_event(void *context, int dummy)
316 {
317 	device_t dev = (device_t)context;
318 #ifdef ATA_CAM
319 	struct ata_channel *ch = device_get_softc(dev);
320 	union ccb *ccb;
321 
322 	mtx_lock(&ch->state_mtx);
323 	if (ch->sim == NULL) {
324 		mtx_unlock(&ch->state_mtx);
325 		return;
326 	}
327 	ata_reinit(dev);
328 	if ((ccb = xpt_alloc_ccb_nowait()) == NULL)
329 		return;
330 	if (xpt_create_path(&ccb->ccb_h.path, NULL,
331 	    cam_sim_path(ch->sim),
332 	    CAM_TARGET_WILDCARD, CAM_LUN_WILDCARD) != CAM_REQ_CMP) {
333 		xpt_free_ccb(ccb);
334 		return;
335 	}
336 	xpt_rescan(ccb);
337 	mtx_unlock(&ch->state_mtx);
338 #else
339 	ata_reinit(dev);
340 #endif
341 }
342 
343 int
344 ata_reinit(device_t dev)
345 {
346     struct ata_channel *ch = device_get_softc(dev);
347     struct ata_request *request;
348 #ifndef ATA_CAM
349     device_t *children;
350     int nchildren, i;
351 
352     /* check that we have a valid channel to reinit */
353     if (!ch || !ch->r_irq)
354 	return ENXIO;
355 
356     if (bootverbose)
357 	device_printf(dev, "reiniting channel ..\n");
358 
359     /* poll for locking the channel */
360     while (ATA_LOCKING(dev, ATA_LF_LOCK) != ch->unit)
361 	pause("atarini", 1);
362 
363     /* catch eventual request in ch->running */
364     mtx_lock(&ch->state_mtx);
365     if (ch->state & ATA_STALL_QUEUE) {
366 	/* Recursive reinits and reinits during detach prohobited. */
367 	mtx_unlock(&ch->state_mtx);
368 	return (ENXIO);
369     }
370     if ((request = ch->running))
371 	callout_stop(&request->callout);
372     ch->running = NULL;
373 
374     /* unconditionally grap the channel lock */
375     ch->state |= ATA_STALL_QUEUE;
376     mtx_unlock(&ch->state_mtx);
377 
378     /* reset the controller HW, the channel and device(s) */
379     ATA_RESET(dev);
380 
381     /* reinit the children and delete any that fails */
382     if (!device_get_children(dev, &children, &nchildren)) {
383 	mtx_lock(&Giant);       /* newbus suckage it needs Giant */
384 	for (i = 0; i < nchildren; i++) {
385 	    /* did any children go missing ? */
386 	    if (children[i] && device_is_attached(children[i]) &&
387 		ATA_REINIT(children[i])) {
388 		/*
389 		 * if we had a running request and its device matches
390 		 * this child we need to inform the request that the
391 		 * device is gone.
392 		 */
393 		if (request && request->dev == children[i]) {
394 		    request->result = ENXIO;
395 		    device_printf(request->dev, "FAILURE - device detached\n");
396 
397 		    /* if not timeout finish request here */
398 		    if (!(request->flags & ATA_R_TIMEOUT))
399 			    ata_finish(request);
400 		    request = NULL;
401 		}
402 		device_delete_child(dev, children[i]);
403 	    }
404 	}
405 	free(children, M_TEMP);
406 	mtx_unlock(&Giant);     /* newbus suckage dealt with, release Giant */
407     }
408 
409     /* if we still have a good request put it on the queue again */
410     if (request && !(request->flags & ATA_R_TIMEOUT)) {
411 	device_printf(request->dev,
412 		      "WARNING - %s requeued due to channel reset",
413 		      ata_cmd2str(request));
414 	if (!(request->flags & (ATA_R_ATAPI | ATA_R_CONTROL)))
415 	    printf(" LBA=%ju", request->u.ata.lba);
416 	printf("\n");
417 	request->flags |= ATA_R_REQUEUE;
418 	ata_queue_request(request);
419     }
420 
421     /* we're done release the channel for new work */
422     mtx_lock(&ch->state_mtx);
423     ch->state = ATA_IDLE;
424     mtx_unlock(&ch->state_mtx);
425     ATA_LOCKING(dev, ATA_LF_UNLOCK);
426 
427     /* Add new children. */
428 /*    ata_identify(dev); */
429 
430     if (bootverbose)
431 	device_printf(dev, "reinit done ..\n");
432 
433     /* kick off requests on the queue */
434     ata_start(dev);
435 #else
436 	xpt_freeze_simq(ch->sim, 1);
437 	if ((request = ch->running)) {
438 		ch->running = NULL;
439 		if (ch->state == ATA_ACTIVE)
440 		    ch->state = ATA_IDLE;
441 		callout_stop(&request->callout);
442 		if (ch->dma.unload)
443 		    ch->dma.unload(request);
444 		request->result = ERESTART;
445 		ata_cam_end_transaction(dev, request);
446 	}
447 	/* reset the controller HW, the channel and device(s) */
448 	ATA_RESET(dev);
449 	/* Tell the XPT about the event */
450 	xpt_async(AC_BUS_RESET, ch->path, NULL);
451 	xpt_release_simq(ch->sim, TRUE);
452 #endif
453 	return(0);
454 }
455 
456 int
457 ata_suspend(device_t dev)
458 {
459     struct ata_channel *ch;
460 
461     /* check for valid device */
462     if (!dev || !(ch = device_get_softc(dev)))
463 	return ENXIO;
464 
465     if (ch->flags & ATA_PERIODIC_POLL)
466 	callout_drain(&ch->poll_callout);
467 #ifdef ATA_CAM
468 	mtx_lock(&ch->state_mtx);
469 	xpt_freeze_simq(ch->sim, 1);
470 	while (ch->state != ATA_IDLE)
471 		msleep(ch, &ch->state_mtx, PRIBIO, "atasusp", hz/100);
472 	mtx_unlock(&ch->state_mtx);
473 #else
474     /* wait for the channel to be IDLE or detached before suspending */
475     while (ch->r_irq) {
476 	mtx_lock(&ch->state_mtx);
477 	if (ch->state == ATA_IDLE) {
478 	    ch->state = ATA_ACTIVE;
479 	    mtx_unlock(&ch->state_mtx);
480 	    break;
481 	}
482 	mtx_unlock(&ch->state_mtx);
483 	tsleep(ch, PRIBIO, "atasusp", hz/10);
484     }
485     ATA_LOCKING(dev, ATA_LF_UNLOCK);
486 #endif
487     return(0);
488 }
489 
490 int
491 ata_resume(device_t dev)
492 {
493     struct ata_channel *ch;
494     int error;
495 
496     /* check for valid device */
497     if (!dev || !(ch = device_get_softc(dev)))
498 	return ENXIO;
499 
500 #ifdef ATA_CAM
501 	mtx_lock(&ch->state_mtx);
502 	error = ata_reinit(dev);
503 	xpt_release_simq(ch->sim, TRUE);
504 	mtx_unlock(&ch->state_mtx);
505 #else
506     /* reinit the devices, we dont know what mode/state they are in */
507     error = ata_reinit(dev);
508     /* kick off requests on the queue */
509     ata_start(dev);
510 #endif
511     if (ch->flags & ATA_PERIODIC_POLL)
512 	callout_reset(&ch->poll_callout, hz, ata_periodic_poll, ch);
513     return error;
514 }
515 
516 void
517 ata_interrupt(void *data)
518 {
519 #ifdef ATA_CAM
520     struct ata_channel *ch = (struct ata_channel *)data;
521 
522     mtx_lock(&ch->state_mtx);
523 #endif
524     ata_interrupt_locked(data);
525 #ifdef ATA_CAM
526     mtx_unlock(&ch->state_mtx);
527 #endif
528 }
529 
530 static void
531 ata_interrupt_locked(void *data)
532 {
533     struct ata_channel *ch = (struct ata_channel *)data;
534     struct ata_request *request;
535 
536 #ifndef ATA_CAM
537     mtx_lock(&ch->state_mtx);
538 #endif
539     do {
540 	/* ignore interrupt if its not for us */
541 	if (ch->hw.status && !ch->hw.status(ch->dev))
542 	    break;
543 
544 	/* do we have a running request */
545 	if (!(request = ch->running))
546 	    break;
547 
548 	ATA_DEBUG_RQ(request, "interrupt");
549 
550 	/* safetycheck for the right state */
551 	if (ch->state == ATA_IDLE) {
552 	    device_printf(request->dev, "interrupt on idle channel ignored\n");
553 	    break;
554 	}
555 
556 	/*
557 	 * we have the HW locks, so end the transaction for this request
558 	 * if it finishes immediately otherwise wait for next interrupt
559 	 */
560 	if (ch->hw.end_transaction(request) == ATA_OP_FINISHED) {
561 	    ch->running = NULL;
562 	    if (ch->state == ATA_ACTIVE)
563 		ch->state = ATA_IDLE;
564 #ifdef ATA_CAM
565 	    ata_cam_end_transaction(ch->dev, request);
566 #else
567 	    mtx_unlock(&ch->state_mtx);
568 	    ATA_LOCKING(ch->dev, ATA_LF_UNLOCK);
569 	    ata_finish(request);
570 #endif
571 	    return;
572 	}
573     } while (0);
574 #ifndef ATA_CAM
575     mtx_unlock(&ch->state_mtx);
576 #endif
577 }
578 
579 static void
580 ata_periodic_poll(void *data)
581 {
582     struct ata_channel *ch = (struct ata_channel *)data;
583 
584     callout_reset(&ch->poll_callout, hz, ata_periodic_poll, ch);
585     ata_interrupt(ch);
586 }
587 
588 void
589 ata_print_cable(device_t dev, u_int8_t *who)
590 {
591     device_printf(dev,
592                   "DMA limited to UDMA33, %s found non-ATA66 cable\n", who);
593 }
594 
595 int
596 ata_check_80pin(device_t dev, int mode)
597 {
598     struct ata_device *atadev = device_get_softc(dev);
599 
600     if (!ata_dma_check_80pin) {
601         if (bootverbose)
602             device_printf(dev, "Skipping 80pin cable check\n");
603         return mode;
604     }
605 
606     if (mode > ATA_UDMA2 && !(atadev->param.hwres & ATA_CABLE_ID)) {
607         ata_print_cable(dev, "device");
608         mode = ATA_UDMA2;
609     }
610     return mode;
611 }
612 
613 void
614 ata_setmode(device_t dev)
615 {
616 	struct ata_channel *ch = device_get_softc(device_get_parent(dev));
617 	struct ata_device *atadev = device_get_softc(dev);
618 	int error, mode, pmode;
619 
620 	mode = atadev->mode;
621 	do {
622 		pmode = mode = ata_limit_mode(dev, mode, ATA_DMA_MAX);
623 		mode = ATA_SETMODE(device_get_parent(dev), atadev->unit, mode);
624 		if ((ch->flags & (ATA_CHECKS_CABLE | ATA_SATA)) == 0)
625 			mode = ata_check_80pin(dev, mode);
626 	} while (pmode != mode); /* Interate till successfull negotiation. */
627 	error = ata_controlcmd(dev, ATA_SETFEATURES, ATA_SF_SETXFER, 0, mode);
628 	if (bootverbose)
629 	        device_printf(dev, "%ssetting %s\n",
630 		    (error) ? "FAILURE " : "", ata_mode2str(mode));
631 	atadev->mode = mode;
632 }
633 
634 /*
635  * device related interfaces
636  */
637 #ifndef ATA_CAM
638 static int
639 ata_ioctl(struct cdev *dev, u_long cmd, caddr_t data,
640 	  int32_t flag, struct thread *td)
641 {
642     device_t device, *children;
643     struct ata_ioc_devices *devices = (struct ata_ioc_devices *)data;
644     int *value = (int *)data;
645     int i, nchildren, error = ENOTTY;
646 
647     switch (cmd) {
648     case IOCATAGMAXCHANNEL:
649 	/* In case we have channel 0..n this will return n+1. */
650 	*value = devclass_get_maxunit(ata_devclass);
651 	error = 0;
652 	break;
653 
654     case IOCATAREINIT:
655 	if (*value >= devclass_get_maxunit(ata_devclass) ||
656 	    !(device = devclass_get_device(ata_devclass, *value)) ||
657 	    !device_is_attached(device))
658 	    return ENXIO;
659 	error = ata_reinit(device);
660 	break;
661 
662     case IOCATAATTACH:
663 	if (*value >= devclass_get_maxunit(ata_devclass) ||
664 	    !(device = devclass_get_device(ata_devclass, *value)) ||
665 	    !device_is_attached(device))
666 	    return ENXIO;
667 	error = DEVICE_ATTACH(device);
668 	break;
669 
670     case IOCATADETACH:
671 	if (*value >= devclass_get_maxunit(ata_devclass) ||
672 	    !(device = devclass_get_device(ata_devclass, *value)) ||
673 	    !device_is_attached(device))
674 	    return ENXIO;
675 	error = DEVICE_DETACH(device);
676 	break;
677 
678     case IOCATADEVICES:
679 	if (devices->channel >= devclass_get_maxunit(ata_devclass) ||
680 	    !(device = devclass_get_device(ata_devclass, devices->channel)) ||
681 	    !device_is_attached(device))
682 	    return ENXIO;
683 	bzero(devices->name[0], 32);
684 	bzero(&devices->params[0], sizeof(struct ata_params));
685 	bzero(devices->name[1], 32);
686 	bzero(&devices->params[1], sizeof(struct ata_params));
687 	if (!device_get_children(device, &children, &nchildren)) {
688 	    for (i = 0; i < nchildren; i++) {
689 		if (children[i] && device_is_attached(children[i])) {
690 		    struct ata_device *atadev = device_get_softc(children[i]);
691 
692 		    if (atadev->unit == ATA_MASTER) { /* XXX SOS PM */
693 			strncpy(devices->name[0],
694 				device_get_nameunit(children[i]), 32);
695 			bcopy(&atadev->param, &devices->params[0],
696 			      sizeof(struct ata_params));
697 		    }
698 		    if (atadev->unit == ATA_SLAVE) { /* XXX SOS PM */
699 			strncpy(devices->name[1],
700 				device_get_nameunit(children[i]), 32);
701 			bcopy(&atadev->param, &devices->params[1],
702 			      sizeof(struct ata_params));
703 		    }
704 		}
705 	    }
706 	    free(children, M_TEMP);
707 	    error = 0;
708 	}
709 	else
710 	    error = ENODEV;
711 	break;
712 
713     default:
714 	if (ata_raid_ioctl_func)
715 	    error = ata_raid_ioctl_func(cmd, data);
716     }
717     return error;
718 }
719 #endif
720 
721 int
722 ata_device_ioctl(device_t dev, u_long cmd, caddr_t data)
723 {
724     struct ata_device *atadev = device_get_softc(dev);
725     struct ata_channel *ch = device_get_softc(device_get_parent(dev));
726     struct ata_ioc_request *ioc_request = (struct ata_ioc_request *)data;
727     struct ata_params *params = (struct ata_params *)data;
728     int *mode = (int *)data;
729     struct ata_request *request;
730     caddr_t buf;
731     int error;
732 
733     switch (cmd) {
734     case IOCATAREQUEST:
735 	if (ioc_request->count >
736 	    (ch->dma.max_iosize ? ch->dma.max_iosize : DFLTPHYS)) {
737 		return (EFBIG);
738 	}
739 	if (!(buf = malloc(ioc_request->count, M_ATA, M_NOWAIT))) {
740 	    return ENOMEM;
741 	}
742 	if (!(request = ata_alloc_request())) {
743 	    free(buf, M_ATA);
744 	    return  ENOMEM;
745 	}
746 	request->dev = atadev->dev;
747 	if (ioc_request->flags & ATA_CMD_WRITE) {
748 	    error = copyin(ioc_request->data, buf, ioc_request->count);
749 	    if (error) {
750 		free(buf, M_ATA);
751 		ata_free_request(request);
752 		return error;
753 	    }
754 	}
755 	if (ioc_request->flags & ATA_CMD_ATAPI) {
756 	    request->flags = ATA_R_ATAPI;
757 	    bcopy(ioc_request->u.atapi.ccb, request->u.atapi.ccb, 16);
758 	}
759 	else {
760 	    request->u.ata.command = ioc_request->u.ata.command;
761 	    request->u.ata.feature = ioc_request->u.ata.feature;
762 	    request->u.ata.lba = ioc_request->u.ata.lba;
763 	    request->u.ata.count = ioc_request->u.ata.count;
764 	}
765 	request->timeout = ioc_request->timeout;
766 	request->data = buf;
767 	request->bytecount = ioc_request->count;
768 	request->transfersize = request->bytecount;
769 	if (ioc_request->flags & ATA_CMD_CONTROL)
770 	    request->flags |= ATA_R_CONTROL;
771 	if (ioc_request->flags & ATA_CMD_READ)
772 	    request->flags |= ATA_R_READ;
773 	if (ioc_request->flags & ATA_CMD_WRITE)
774 	    request->flags |= ATA_R_WRITE;
775 	ata_queue_request(request);
776 	if (request->flags & ATA_R_ATAPI) {
777 	    bcopy(&request->u.atapi.sense, &ioc_request->u.atapi.sense,
778 		  sizeof(struct atapi_sense));
779 	}
780 	else {
781 	    ioc_request->u.ata.command = request->u.ata.command;
782 	    ioc_request->u.ata.feature = request->u.ata.feature;
783 	    ioc_request->u.ata.lba = request->u.ata.lba;
784 	    ioc_request->u.ata.count = request->u.ata.count;
785 	}
786 	ioc_request->error = request->result;
787 	if (ioc_request->flags & ATA_CMD_READ)
788 	    error = copyout(buf, ioc_request->data, ioc_request->count);
789 	else
790 	    error = 0;
791 	free(buf, M_ATA);
792 	ata_free_request(request);
793 	return error;
794 
795     case IOCATAGPARM:
796 	ata_getparam(atadev, 0);
797 	bcopy(&atadev->param, params, sizeof(struct ata_params));
798 	return 0;
799 
800     case IOCATASMODE:
801 	atadev->mode = *mode;
802 	ata_setmode(dev);
803 	return 0;
804 
805     case IOCATAGMODE:
806 	*mode = atadev->mode |
807 	    (ATA_GETREV(device_get_parent(dev), atadev->unit) << 8);
808 	return 0;
809     case IOCATASSPINDOWN:
810 	atadev->spindown = *mode;
811 	return 0;
812     case IOCATAGSPINDOWN:
813 	*mode = atadev->spindown;
814 	return 0;
815     default:
816 	return ENOTTY;
817     }
818 }
819 
820 #ifndef ATA_CAM
821 static void
822 ata_boot_attach(void)
823 {
824     struct ata_channel *ch;
825     int ctlr;
826 
827     mtx_lock(&Giant);       /* newbus suckage it needs Giant */
828 
829     /* kick of probe and attach on all channels */
830     for (ctlr = 0; ctlr < devclass_get_maxunit(ata_devclass); ctlr++) {
831 	if ((ch = devclass_get_softc(ata_devclass, ctlr))) {
832 	    ata_identify(ch->dev);
833 	}
834     }
835 
836     /* release the hook that got us here, we are only needed once during boot */
837     if (ata_delayed_attach) {
838 	config_intrhook_disestablish(ata_delayed_attach);
839 	free(ata_delayed_attach, M_TEMP);
840 	ata_delayed_attach = NULL;
841     }
842 
843     mtx_unlock(&Giant);     /* newbus suckage dealt with, release Giant */
844 }
845 #endif
846 
847 /*
848  * misc support functions
849  */
850 #ifndef ATA_CAM
851 static device_t
852 ata_add_child(device_t parent, struct ata_device *atadev, int unit)
853 {
854     device_t child;
855 
856     if ((child = device_add_child(parent, NULL, unit))) {
857 	device_set_softc(child, atadev);
858 	device_quiet(child);
859 	atadev->dev = child;
860 	atadev->max_iosize = DEV_BSIZE;
861 	atadev->mode = ATA_PIO_MAX;
862     }
863     return child;
864 }
865 #endif
866 
867 int
868 ata_getparam(struct ata_device *atadev, int init)
869 {
870     struct ata_channel *ch = device_get_softc(device_get_parent(atadev->dev));
871     struct ata_request *request;
872     const char *res;
873     char buf[64];
874     u_int8_t command = 0;
875     int error = ENOMEM, retries = 2, mode = -1;
876 
877     if (ch->devices & (ATA_ATA_MASTER << atadev->unit))
878 	command = ATA_ATA_IDENTIFY;
879     if (ch->devices & (ATA_ATAPI_MASTER << atadev->unit))
880 	command = ATA_ATAPI_IDENTIFY;
881     if (!command)
882 	return ENXIO;
883 
884     while (retries-- > 0 && error) {
885 	if (!(request = ata_alloc_request()))
886 	    break;
887 	request->dev = atadev->dev;
888 	request->timeout = 1;
889 	request->retries = 0;
890 	request->u.ata.command = command;
891 	request->flags = (ATA_R_READ|ATA_R_AT_HEAD|ATA_R_DIRECT);
892 	if (!bootverbose)
893 	    request->flags |= ATA_R_QUIET;
894 	request->data = (void *)&atadev->param;
895 	request->bytecount = sizeof(struct ata_params);
896 	request->donecount = 0;
897 	request->transfersize = DEV_BSIZE;
898 	ata_queue_request(request);
899 	error = request->result;
900 	ata_free_request(request);
901     }
902 
903     if (!error && (isprint(atadev->param.model[0]) ||
904 		   isprint(atadev->param.model[1]))) {
905 	struct ata_params *atacap = &atadev->param;
906 	int16_t *ptr;
907 
908 	for (ptr = (int16_t *)atacap;
909 	     ptr < (int16_t *)atacap + sizeof(struct ata_params)/2; ptr++) {
910 	    *ptr = le16toh(*ptr);
911 	}
912 	if (!(!strncmp(atacap->model, "FX", 2) ||
913 	      !strncmp(atacap->model, "NEC", 3) ||
914 	      !strncmp(atacap->model, "Pioneer", 7) ||
915 	      !strncmp(atacap->model, "SHARP", 5))) {
916 	    bswap(atacap->model, sizeof(atacap->model));
917 	    bswap(atacap->revision, sizeof(atacap->revision));
918 	    bswap(atacap->serial, sizeof(atacap->serial));
919 	}
920 	btrim(atacap->model, sizeof(atacap->model));
921 	bpack(atacap->model, atacap->model, sizeof(atacap->model));
922 	btrim(atacap->revision, sizeof(atacap->revision));
923 	bpack(atacap->revision, atacap->revision, sizeof(atacap->revision));
924 	btrim(atacap->serial, sizeof(atacap->serial));
925 	bpack(atacap->serial, atacap->serial, sizeof(atacap->serial));
926 
927 	if (bootverbose)
928 	    printf("ata%d-%s: pio=%s wdma=%s udma=%s cable=%s wire\n",
929 		   device_get_unit(ch->dev),
930 		   ata_unit2str(atadev),
931 		   ata_mode2str(ata_pmode(atacap)),
932 		   ata_mode2str(ata_wmode(atacap)),
933 		   ata_mode2str(ata_umode(atacap)),
934 		   (atacap->hwres & ATA_CABLE_ID) ? "80":"40");
935 
936 	if (init) {
937 	    char buffer[64];
938 
939 	    sprintf(buffer, "%.40s/%.8s", atacap->model, atacap->revision);
940 	    device_set_desc_copy(atadev->dev, buffer);
941 	    if ((atadev->param.config & ATA_PROTO_ATAPI) &&
942 		(atadev->param.config != ATA_CFA_MAGIC1) &&
943 		(atadev->param.config != ATA_CFA_MAGIC2)) {
944 		if (atapi_dma &&
945 		    (atadev->param.config & ATA_DRQ_MASK) != ATA_DRQ_INTR &&
946 		    ata_umode(&atadev->param) >= ATA_UDMA2)
947 		    atadev->mode = ATA_DMA_MAX;
948 	    }
949 	    else {
950 		if (ata_dma &&
951 		    (ata_umode(&atadev->param) > 0 ||
952 		     ata_wmode(&atadev->param) > 0))
953 		    atadev->mode = ATA_DMA_MAX;
954 	    }
955 	    snprintf(buf, sizeof(buf), "dev%d.mode", atadev->unit);
956 	    if (resource_string_value(device_get_name(ch->dev),
957 	        device_get_unit(ch->dev), buf, &res) == 0)
958 		    mode = ata_str2mode(res);
959 	    else if (resource_string_value(device_get_name(ch->dev),
960 		device_get_unit(ch->dev), "mode", &res) == 0)
961 		    mode = ata_str2mode(res);
962 	    if (mode >= 0)
963 		    atadev->mode = mode;
964 	}
965     }
966     else {
967 	if (!error)
968 	    error = ENXIO;
969     }
970     return error;
971 }
972 
973 #ifndef ATA_CAM
974 int
975 ata_identify(device_t dev)
976 {
977     struct ata_channel *ch = device_get_softc(dev);
978     struct ata_device *atadev;
979     device_t *children;
980     device_t child, master = NULL;
981     int nchildren, i, n = ch->devices;
982 
983     if (bootverbose)
984 	device_printf(dev, "Identifying devices: %08x\n", ch->devices);
985 
986     mtx_lock(&Giant);
987     /* Skip existing devices. */
988     if (!device_get_children(dev, &children, &nchildren)) {
989 	for (i = 0; i < nchildren; i++) {
990 	    if (children[i] && (atadev = device_get_softc(children[i])))
991 		n &= ~((ATA_ATA_MASTER | ATA_ATAPI_MASTER) << atadev->unit);
992 	}
993 	free(children, M_TEMP);
994     }
995     /* Create new devices. */
996     if (bootverbose)
997 	device_printf(dev, "New devices: %08x\n", n);
998     if (n == 0) {
999 	mtx_unlock(&Giant);
1000 	return (0);
1001     }
1002     for (i = 0; i < ATA_PM; ++i) {
1003 	if (n & (((ATA_ATA_MASTER | ATA_ATAPI_MASTER) << i))) {
1004 	    int unit = -1;
1005 
1006 	    if (!(atadev = malloc(sizeof(struct ata_device),
1007 				  M_ATA, M_NOWAIT | M_ZERO))) {
1008 		device_printf(dev, "out of memory\n");
1009 		return ENOMEM;
1010 	    }
1011 	    atadev->unit = i;
1012 #ifdef ATA_STATIC_ID
1013 	    if (n & (ATA_ATA_MASTER << i))
1014 		unit = (device_get_unit(dev) << 1) + i;
1015 #endif
1016 	    if ((child = ata_add_child(dev, atadev, unit))) {
1017 		/*
1018 		 * PATA slave should be identified first, to allow
1019 		 * device cable detection on master to work properly.
1020 		 */
1021 		if (i == 0 && (n & ATA_PORTMULTIPLIER) == 0 &&
1022 			(n & ((ATA_ATA_MASTER | ATA_ATAPI_MASTER) << 1)) != 0) {
1023 		    master = child;
1024 		    continue;
1025 		}
1026 		if (ata_getparam(atadev, 1)) {
1027 		    device_delete_child(dev, child);
1028 		    free(atadev, M_ATA);
1029 		}
1030 	    }
1031 	    else
1032 		free(atadev, M_ATA);
1033 	}
1034     }
1035     if (master) {
1036 	atadev = device_get_softc(master);
1037 	if (ata_getparam(atadev, 1)) {
1038 	    device_delete_child(dev, master);
1039 	    free(atadev, M_ATA);
1040 	}
1041     }
1042     bus_generic_probe(dev);
1043     bus_generic_attach(dev);
1044     mtx_unlock(&Giant);
1045     return 0;
1046 }
1047 #endif
1048 
1049 void
1050 ata_default_registers(device_t dev)
1051 {
1052     struct ata_channel *ch = device_get_softc(dev);
1053 
1054     /* fill in the defaults from whats setup already */
1055     ch->r_io[ATA_ERROR].res = ch->r_io[ATA_FEATURE].res;
1056     ch->r_io[ATA_ERROR].offset = ch->r_io[ATA_FEATURE].offset;
1057     ch->r_io[ATA_IREASON].res = ch->r_io[ATA_COUNT].res;
1058     ch->r_io[ATA_IREASON].offset = ch->r_io[ATA_COUNT].offset;
1059     ch->r_io[ATA_STATUS].res = ch->r_io[ATA_COMMAND].res;
1060     ch->r_io[ATA_STATUS].offset = ch->r_io[ATA_COMMAND].offset;
1061     ch->r_io[ATA_ALTSTAT].res = ch->r_io[ATA_CONTROL].res;
1062     ch->r_io[ATA_ALTSTAT].offset = ch->r_io[ATA_CONTROL].offset;
1063 }
1064 
1065 void
1066 ata_modify_if_48bit(struct ata_request *request)
1067 {
1068     struct ata_channel *ch = device_get_softc(request->parent);
1069     struct ata_device *atadev = device_get_softc(request->dev);
1070 
1071     request->flags &= ~ATA_R_48BIT;
1072 
1073     if (((request->u.ata.lba + request->u.ata.count) >= ATA_MAX_28BIT_LBA ||
1074 	 request->u.ata.count > 256) &&
1075 	atadev->param.support.command2 & ATA_SUPPORT_ADDRESS48) {
1076 
1077 	/* translate command into 48bit version */
1078 	switch (request->u.ata.command) {
1079 	case ATA_READ:
1080 	    request->u.ata.command = ATA_READ48;
1081 	    break;
1082 	case ATA_READ_MUL:
1083 	    request->u.ata.command = ATA_READ_MUL48;
1084 	    break;
1085 	case ATA_READ_DMA:
1086 	    if (ch->flags & ATA_NO_48BIT_DMA) {
1087 		if (request->transfersize > DEV_BSIZE)
1088 		    request->u.ata.command = ATA_READ_MUL48;
1089 		else
1090 		    request->u.ata.command = ATA_READ48;
1091 		request->flags &= ~ATA_R_DMA;
1092 	    }
1093 	    else
1094 		request->u.ata.command = ATA_READ_DMA48;
1095 	    break;
1096 	case ATA_READ_DMA_QUEUED:
1097 	    if (ch->flags & ATA_NO_48BIT_DMA) {
1098 		if (request->transfersize > DEV_BSIZE)
1099 		    request->u.ata.command = ATA_READ_MUL48;
1100 		else
1101 		    request->u.ata.command = ATA_READ48;
1102 		request->flags &= ~ATA_R_DMA;
1103 	    }
1104 	    else
1105 		request->u.ata.command = ATA_READ_DMA_QUEUED48;
1106 	    break;
1107 	case ATA_WRITE:
1108 	    request->u.ata.command = ATA_WRITE48;
1109 	    break;
1110 	case ATA_WRITE_MUL:
1111 	    request->u.ata.command = ATA_WRITE_MUL48;
1112 	    break;
1113 	case ATA_WRITE_DMA:
1114 	    if (ch->flags & ATA_NO_48BIT_DMA) {
1115 		if (request->transfersize > DEV_BSIZE)
1116 		    request->u.ata.command = ATA_WRITE_MUL48;
1117 		else
1118 		    request->u.ata.command = ATA_WRITE48;
1119 		request->flags &= ~ATA_R_DMA;
1120 	    }
1121 	    else
1122 		request->u.ata.command = ATA_WRITE_DMA48;
1123 	    break;
1124 	case ATA_WRITE_DMA_QUEUED:
1125 	    if (ch->flags & ATA_NO_48BIT_DMA) {
1126 		if (request->transfersize > DEV_BSIZE)
1127 		    request->u.ata.command = ATA_WRITE_MUL48;
1128 		else
1129 		    request->u.ata.command = ATA_WRITE48;
1130 		request->u.ata.command = ATA_WRITE48;
1131 		request->flags &= ~ATA_R_DMA;
1132 	    }
1133 	    else
1134 		request->u.ata.command = ATA_WRITE_DMA_QUEUED48;
1135 	    break;
1136 	case ATA_FLUSHCACHE:
1137 	    request->u.ata.command = ATA_FLUSHCACHE48;
1138 	    break;
1139 	case ATA_SET_MAX_ADDRESS:
1140 	    request->u.ata.command = ATA_SET_MAX_ADDRESS48;
1141 	    break;
1142 	default:
1143 	    return;
1144 	}
1145 	request->flags |= ATA_R_48BIT;
1146     }
1147     else if (atadev->param.support.command2 & ATA_SUPPORT_ADDRESS48) {
1148 
1149 	/* translate command into 48bit version */
1150 	switch (request->u.ata.command) {
1151 	case ATA_FLUSHCACHE:
1152 	    request->u.ata.command = ATA_FLUSHCACHE48;
1153 	    break;
1154 	case ATA_READ_NATIVE_MAX_ADDRESS:
1155 	    request->u.ata.command = ATA_READ_NATIVE_MAX_ADDRESS48;
1156 	    break;
1157 	case ATA_SET_MAX_ADDRESS:
1158 	    request->u.ata.command = ATA_SET_MAX_ADDRESS48;
1159 	    break;
1160 	default:
1161 	    return;
1162 	}
1163 	request->flags |= ATA_R_48BIT;
1164     }
1165 }
1166 
1167 void
1168 ata_udelay(int interval)
1169 {
1170     /* for now just use DELAY, the timer/sleep subsytems are not there yet */
1171     if (1 || interval < (1000000/hz) || ata_delayed_attach)
1172 	DELAY(interval);
1173     else
1174 	pause("ataslp", interval/(1000000/hz));
1175 }
1176 
1177 char *
1178 ata_unit2str(struct ata_device *atadev)
1179 {
1180     struct ata_channel *ch = device_get_softc(device_get_parent(atadev->dev));
1181     static char str[8];
1182 
1183     if (ch->devices & ATA_PORTMULTIPLIER)
1184 	sprintf(str, "port%d", atadev->unit);
1185     else
1186 	sprintf(str, "%s", atadev->unit == ATA_MASTER ? "master" : "slave");
1187     return str;
1188 }
1189 
1190 const char *
1191 ata_mode2str(int mode)
1192 {
1193     switch (mode) {
1194     case -1: return "UNSUPPORTED";
1195     case ATA_PIO0: return "PIO0";
1196     case ATA_PIO1: return "PIO1";
1197     case ATA_PIO2: return "PIO2";
1198     case ATA_PIO3: return "PIO3";
1199     case ATA_PIO4: return "PIO4";
1200     case ATA_WDMA0: return "WDMA0";
1201     case ATA_WDMA1: return "WDMA1";
1202     case ATA_WDMA2: return "WDMA2";
1203     case ATA_UDMA0: return "UDMA16";
1204     case ATA_UDMA1: return "UDMA25";
1205     case ATA_UDMA2: return "UDMA33";
1206     case ATA_UDMA3: return "UDMA40";
1207     case ATA_UDMA4: return "UDMA66";
1208     case ATA_UDMA5: return "UDMA100";
1209     case ATA_UDMA6: return "UDMA133";
1210     case ATA_SA150: return "SATA150";
1211     case ATA_SA300: return "SATA300";
1212     default:
1213 	if (mode & ATA_DMA_MASK)
1214 	    return "BIOSDMA";
1215 	else
1216 	    return "BIOSPIO";
1217     }
1218 }
1219 
1220 int
1221 ata_str2mode(const char *str)
1222 {
1223 
1224 	if (!strcasecmp(str, "PIO0")) return (ATA_PIO0);
1225 	if (!strcasecmp(str, "PIO1")) return (ATA_PIO1);
1226 	if (!strcasecmp(str, "PIO2")) return (ATA_PIO2);
1227 	if (!strcasecmp(str, "PIO3")) return (ATA_PIO3);
1228 	if (!strcasecmp(str, "PIO4")) return (ATA_PIO4);
1229 	if (!strcasecmp(str, "WDMA0")) return (ATA_WDMA0);
1230 	if (!strcasecmp(str, "WDMA1")) return (ATA_WDMA1);
1231 	if (!strcasecmp(str, "WDMA2")) return (ATA_WDMA2);
1232 	if (!strcasecmp(str, "UDMA0")) return (ATA_UDMA0);
1233 	if (!strcasecmp(str, "UDMA16")) return (ATA_UDMA0);
1234 	if (!strcasecmp(str, "UDMA1")) return (ATA_UDMA1);
1235 	if (!strcasecmp(str, "UDMA25")) return (ATA_UDMA1);
1236 	if (!strcasecmp(str, "UDMA2")) return (ATA_UDMA2);
1237 	if (!strcasecmp(str, "UDMA33")) return (ATA_UDMA2);
1238 	if (!strcasecmp(str, "UDMA3")) return (ATA_UDMA3);
1239 	if (!strcasecmp(str, "UDMA44")) return (ATA_UDMA3);
1240 	if (!strcasecmp(str, "UDMA4")) return (ATA_UDMA4);
1241 	if (!strcasecmp(str, "UDMA66")) return (ATA_UDMA4);
1242 	if (!strcasecmp(str, "UDMA5")) return (ATA_UDMA5);
1243 	if (!strcasecmp(str, "UDMA100")) return (ATA_UDMA5);
1244 	if (!strcasecmp(str, "UDMA6")) return (ATA_UDMA6);
1245 	if (!strcasecmp(str, "UDMA133")) return (ATA_UDMA6);
1246 	return (-1);
1247 }
1248 
1249 const char *
1250 ata_satarev2str(int rev)
1251 {
1252 	switch (rev) {
1253 	case 0: return "";
1254 	case 1: return "SATA 1.5Gb/s";
1255 	case 2: return "SATA 3Gb/s";
1256 	case 3: return "SATA 6Gb/s";
1257 	case 0xff: return "SATA";
1258 	default: return "???";
1259 	}
1260 }
1261 
1262 int
1263 ata_atapi(device_t dev, int target)
1264 {
1265     struct ata_channel *ch = device_get_softc(dev);
1266 
1267     return (ch->devices & (ATA_ATAPI_MASTER << target));
1268 }
1269 
1270 int
1271 ata_pmode(struct ata_params *ap)
1272 {
1273     if (ap->atavalid & ATA_FLAG_64_70) {
1274 	if (ap->apiomodes & 0x02)
1275 	    return ATA_PIO4;
1276 	if (ap->apiomodes & 0x01)
1277 	    return ATA_PIO3;
1278     }
1279     if (ap->mwdmamodes & 0x04)
1280 	return ATA_PIO4;
1281     if (ap->mwdmamodes & 0x02)
1282 	return ATA_PIO3;
1283     if (ap->mwdmamodes & 0x01)
1284 	return ATA_PIO2;
1285     if ((ap->retired_piomode & ATA_RETIRED_PIO_MASK) == 0x200)
1286 	return ATA_PIO2;
1287     if ((ap->retired_piomode & ATA_RETIRED_PIO_MASK) == 0x100)
1288 	return ATA_PIO1;
1289     if ((ap->retired_piomode & ATA_RETIRED_PIO_MASK) == 0x000)
1290 	return ATA_PIO0;
1291     return ATA_PIO0;
1292 }
1293 
1294 int
1295 ata_wmode(struct ata_params *ap)
1296 {
1297     if (ap->mwdmamodes & 0x04)
1298 	return ATA_WDMA2;
1299     if (ap->mwdmamodes & 0x02)
1300 	return ATA_WDMA1;
1301     if (ap->mwdmamodes & 0x01)
1302 	return ATA_WDMA0;
1303     return -1;
1304 }
1305 
1306 int
1307 ata_umode(struct ata_params *ap)
1308 {
1309     if (ap->atavalid & ATA_FLAG_88) {
1310 	if (ap->udmamodes & 0x40)
1311 	    return ATA_UDMA6;
1312 	if (ap->udmamodes & 0x20)
1313 	    return ATA_UDMA5;
1314 	if (ap->udmamodes & 0x10)
1315 	    return ATA_UDMA4;
1316 	if (ap->udmamodes & 0x08)
1317 	    return ATA_UDMA3;
1318 	if (ap->udmamodes & 0x04)
1319 	    return ATA_UDMA2;
1320 	if (ap->udmamodes & 0x02)
1321 	    return ATA_UDMA1;
1322 	if (ap->udmamodes & 0x01)
1323 	    return ATA_UDMA0;
1324     }
1325     return -1;
1326 }
1327 
1328 int
1329 ata_limit_mode(device_t dev, int mode, int maxmode)
1330 {
1331     struct ata_device *atadev = device_get_softc(dev);
1332 
1333     if (maxmode && mode > maxmode)
1334 	mode = maxmode;
1335 
1336     if (mode >= ATA_UDMA0 && ata_umode(&atadev->param) > 0)
1337 	return min(mode, ata_umode(&atadev->param));
1338 
1339     if (mode >= ATA_WDMA0 && ata_wmode(&atadev->param) > 0)
1340 	return min(mode, ata_wmode(&atadev->param));
1341 
1342     if (mode > ata_pmode(&atadev->param))
1343 	return min(mode, ata_pmode(&atadev->param));
1344 
1345     return mode;
1346 }
1347 
1348 static void
1349 bswap(int8_t *buf, int len)
1350 {
1351     u_int16_t *ptr = (u_int16_t*)(buf + len);
1352 
1353     while (--ptr >= (u_int16_t*)buf)
1354 	*ptr = ntohs(*ptr);
1355 }
1356 
1357 static void
1358 btrim(int8_t *buf, int len)
1359 {
1360     int8_t *ptr;
1361 
1362     for (ptr = buf; ptr < buf+len; ++ptr)
1363 	if (!*ptr || *ptr == '_')
1364 	    *ptr = ' ';
1365     for (ptr = buf + len - 1; ptr >= buf && *ptr == ' '; --ptr)
1366 	*ptr = 0;
1367 }
1368 
1369 static void
1370 bpack(int8_t *src, int8_t *dst, int len)
1371 {
1372     int i, j, blank;
1373 
1374     for (i = j = blank = 0 ; i < len; i++) {
1375 	if (blank && src[i] == ' ') continue;
1376 	if (blank && src[i] != ' ') {
1377 	    dst[j++] = src[i];
1378 	    blank = 0;
1379 	    continue;
1380 	}
1381 	if (src[i] == ' ') {
1382 	    blank = 1;
1383 	    if (i == 0)
1384 		continue;
1385 	}
1386 	dst[j++] = src[i];
1387     }
1388     if (j < len)
1389 	dst[j] = 0x00;
1390 }
1391 
1392 #ifdef ATA_CAM
1393 void
1394 ata_cam_begin_transaction(device_t dev, union ccb *ccb)
1395 {
1396 	struct ata_channel *ch = device_get_softc(dev);
1397 	struct ata_request *request;
1398 
1399 	if (!(request = ata_alloc_request())) {
1400 		device_printf(dev, "FAILURE - out of memory in start\n");
1401 		ccb->ccb_h.status = CAM_REQ_INVALID;
1402 		xpt_done(ccb);
1403 		return;
1404 	}
1405 	bzero(request, sizeof(*request));
1406 
1407 	/* setup request */
1408 	request->dev = NULL;
1409 	request->parent = dev;
1410 	request->unit = ccb->ccb_h.target_id;
1411 	if (ccb->ccb_h.func_code == XPT_ATA_IO) {
1412 		request->data = ccb->ataio.data_ptr;
1413 		request->bytecount = ccb->ataio.dxfer_len;
1414 		request->u.ata.command = ccb->ataio.cmd.command;
1415 		request->u.ata.feature = ((uint16_t)ccb->ataio.cmd.features_exp << 8) |
1416 					  (uint16_t)ccb->ataio.cmd.features;
1417 		request->u.ata.count = ((uint16_t)ccb->ataio.cmd.sector_count_exp << 8) |
1418 					(uint16_t)ccb->ataio.cmd.sector_count;
1419 		if (ccb->ataio.cmd.flags & CAM_ATAIO_48BIT) {
1420 			request->flags |= ATA_R_48BIT;
1421 			request->u.ata.lba =
1422 				     ((uint64_t)ccb->ataio.cmd.lba_high_exp << 40) |
1423 				     ((uint64_t)ccb->ataio.cmd.lba_mid_exp << 32) |
1424 				     ((uint64_t)ccb->ataio.cmd.lba_low_exp << 24);
1425 		} else {
1426 			request->u.ata.lba =
1427 				     ((uint64_t)(ccb->ataio.cmd.device & 0x0f) << 24);
1428 		}
1429 		request->u.ata.lba |= ((uint64_t)ccb->ataio.cmd.lba_high << 16) |
1430 				      ((uint64_t)ccb->ataio.cmd.lba_mid << 8) |
1431 				       (uint64_t)ccb->ataio.cmd.lba_low;
1432 		if ((ccb->ccb_h.flags & CAM_DIR_MASK) != CAM_DIR_NONE &&
1433 		    ccb->ataio.cmd.flags & CAM_ATAIO_DMA)
1434 			request->flags |= ATA_R_DMA;
1435 		if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN)
1436 			request->flags |= ATA_R_READ;
1437 		if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_OUT)
1438 			request->flags |= ATA_R_WRITE;
1439 	} else {
1440 		request->data = ccb->csio.data_ptr;
1441 		request->bytecount = ccb->csio.dxfer_len;
1442 		bcopy((ccb->ccb_h.flags & CAM_CDB_POINTER) ?
1443 		    ccb->csio.cdb_io.cdb_ptr : ccb->csio.cdb_io.cdb_bytes,
1444 		    request->u.atapi.ccb, ccb->csio.cdb_len);
1445 		request->flags |= ATA_R_ATAPI;
1446 		if (ch->curr[ccb->ccb_h.target_id].atapi == 16)
1447 			request->flags |= ATA_R_ATAPI16;
1448 		if ((ccb->ccb_h.flags & CAM_DIR_MASK) != CAM_DIR_NONE &&
1449 		    ch->curr[ccb->ccb_h.target_id].mode >= ATA_DMA)
1450 			request->flags |= ATA_R_DMA;
1451 		if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN)
1452 			request->flags |= ATA_R_READ;
1453 		if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_OUT)
1454 			request->flags |= ATA_R_WRITE;
1455 	}
1456 	request->transfersize = min(request->bytecount,
1457 	    ch->curr[ccb->ccb_h.target_id].bytecount);
1458 	request->retries = 0;
1459 	request->timeout = (ccb->ccb_h.timeout + 999) / 1000;
1460 	callout_init_mtx(&request->callout, &ch->state_mtx, CALLOUT_RETURNUNLOCKED);
1461 	request->ccb = ccb;
1462 
1463 	ch->running = request;
1464 	ch->state = ATA_ACTIVE;
1465 	if (ch->hw.begin_transaction(request) == ATA_OP_FINISHED) {
1466 	    ch->running = NULL;
1467 	    ch->state = ATA_IDLE;
1468 	    ata_cam_end_transaction(dev, request);
1469 	    return;
1470 	}
1471 }
1472 
1473 void
1474 ata_cam_end_transaction(device_t dev, struct ata_request *request)
1475 {
1476 	struct ata_channel *ch = device_get_softc(dev);
1477 	union ccb *ccb = request->ccb;
1478 	int fatalerr = 0;
1479 
1480 	ccb->ccb_h.status &= ~CAM_STATUS_MASK;
1481 	if (request->flags & ATA_R_TIMEOUT) {
1482 		xpt_freeze_simq(ch->sim, 1);
1483 		ccb->ccb_h.status &= ~CAM_STATUS_MASK;
1484 		ccb->ccb_h.status |= CAM_CMD_TIMEOUT | CAM_RELEASE_SIMQ;
1485 		fatalerr = 1;
1486 	} else if (request->status & ATA_S_ERROR) {
1487 		if (ccb->ccb_h.func_code == XPT_ATA_IO) {
1488 			ccb->ccb_h.status |= CAM_ATA_STATUS_ERROR;
1489 		} else {
1490 			ccb->ccb_h.status |= CAM_SCSI_STATUS_ERROR;
1491 			ccb->csio.scsi_status = SCSI_STATUS_CHECK_COND;
1492 		}
1493 	} else if (request->result == ERESTART)
1494 		ccb->ccb_h.status |= CAM_REQUEUE_REQ;
1495 	else if (request->result != 0)
1496 		ccb->ccb_h.status |= CAM_REQ_CMP_ERR;
1497 	else
1498 		ccb->ccb_h.status |= CAM_REQ_CMP;
1499 	if ((ccb->ccb_h.status & CAM_STATUS_MASK) != CAM_REQ_CMP &&
1500 	    !(ccb->ccb_h.status & CAM_DEV_QFRZN)) {
1501 		xpt_freeze_devq(ccb->ccb_h.path, 1);
1502 		ccb->ccb_h.status |= CAM_DEV_QFRZN;
1503 	}
1504 	if (ccb->ccb_h.func_code == XPT_ATA_IO &&
1505 	    ((request->status & ATA_S_ERROR) ||
1506 	    (ccb->ataio.cmd.flags & CAM_ATAIO_NEEDRESULT))) {
1507 		struct ata_res *res = &ccb->ataio.res;
1508 		res->status = request->status;
1509 		res->error = request->error;
1510 		res->lba_low = request->u.ata.lba;
1511 		res->lba_mid = request->u.ata.lba >> 8;
1512 		res->lba_high = request->u.ata.lba >> 16;
1513 		res->device = request->u.ata.lba >> 24;
1514 		res->lba_low_exp = request->u.ata.lba >> 24;
1515 		res->lba_mid_exp = request->u.ata.lba >> 32;
1516 		res->lba_high_exp = request->u.ata.lba >> 40;
1517 		res->sector_count = request->u.ata.count;
1518 		res->sector_count_exp = request->u.ata.count >> 8;
1519 	}
1520 	if ((ccb->ccb_h.flags & CAM_DIR_MASK) != CAM_DIR_NONE) {
1521 		if (ccb->ccb_h.func_code == XPT_ATA_IO) {
1522 			ccb->ataio.resid =
1523 			    ccb->ataio.dxfer_len - request->donecount;
1524 		} else {
1525 			ccb->csio.resid =
1526 			    ccb->csio.dxfer_len - request->donecount;
1527 		}
1528 	}
1529 	ata_free_request(request);
1530 	xpt_done(ccb);
1531 	/* Do error recovery if needed. */
1532 	if (fatalerr)
1533 		ata_reinit(dev);
1534 }
1535 
1536 static int
1537 ata_check_ids(device_t dev, union ccb *ccb)
1538 {
1539 	struct ata_channel *ch = device_get_softc(dev);
1540 
1541 	if (ccb->ccb_h.target_id > ((ch->flags & ATA_NO_SLAVE) ? 0 : 1)) {
1542 		ccb->ccb_h.status = CAM_TID_INVALID;
1543 		xpt_done(ccb);
1544 		return (-1);
1545 	}
1546 	if (ccb->ccb_h.target_lun != 0) {
1547 		ccb->ccb_h.status = CAM_LUN_INVALID;
1548 		xpt_done(ccb);
1549 		return (-1);
1550 	}
1551 	return (0);
1552 }
1553 
1554 static void
1555 ataaction(struct cam_sim *sim, union ccb *ccb)
1556 {
1557 	device_t dev, parent;
1558 	struct ata_channel *ch;
1559 
1560 	CAM_DEBUG(ccb->ccb_h.path, CAM_DEBUG_TRACE, ("ataaction func_code=%x\n",
1561 	    ccb->ccb_h.func_code));
1562 
1563 	ch = (struct ata_channel *)cam_sim_softc(sim);
1564 	dev = ch->dev;
1565 	switch (ccb->ccb_h.func_code) {
1566 	/* Common cases first */
1567 	case XPT_ATA_IO:	/* Execute the requested I/O operation */
1568 	case XPT_SCSI_IO:
1569 		if (ata_check_ids(dev, ccb))
1570 			return;
1571 		if ((ch->devices & ((ATA_ATA_MASTER | ATA_ATAPI_MASTER)
1572 		    << ccb->ccb_h.target_id)) == 0) {
1573 			ccb->ccb_h.status = CAM_SEL_TIMEOUT;
1574 			break;
1575 		}
1576 		if (ch->running)
1577 			device_printf(dev, "already running!\n");
1578 		if (ccb->ccb_h.func_code == XPT_ATA_IO &&
1579 		    (ccb->ataio.cmd.flags & CAM_ATAIO_CONTROL) &&
1580 		    (ccb->ataio.cmd.control & ATA_A_RESET)) {
1581 			struct ata_res *res = &ccb->ataio.res;
1582 
1583 			bzero(res, sizeof(*res));
1584 			if (ch->devices & (ATA_ATA_MASTER << ccb->ccb_h.target_id)) {
1585 				res->lba_high = 0;
1586 				res->lba_mid = 0;
1587 			} else {
1588 				res->lba_high = 0xeb;
1589 				res->lba_mid = 0x14;
1590 			}
1591 			ccb->ccb_h.status = CAM_REQ_CMP;
1592 			break;
1593 		}
1594 		ata_cam_begin_transaction(dev, ccb);
1595 		return;
1596 	case XPT_EN_LUN:		/* Enable LUN as a target */
1597 	case XPT_TARGET_IO:		/* Execute target I/O request */
1598 	case XPT_ACCEPT_TARGET_IO:	/* Accept Host Target Mode CDB */
1599 	case XPT_CONT_TARGET_IO:	/* Continue Host Target I/O Connection*/
1600 	case XPT_ABORT:			/* Abort the specified CCB */
1601 		/* XXX Implement */
1602 		ccb->ccb_h.status = CAM_REQ_INVALID;
1603 		break;
1604 	case XPT_SET_TRAN_SETTINGS:
1605 	{
1606 		struct	ccb_trans_settings *cts = &ccb->cts;
1607 		struct	ata_cam_device *d;
1608 
1609 		if (ata_check_ids(dev, ccb))
1610 			return;
1611 		if (cts->type == CTS_TYPE_CURRENT_SETTINGS)
1612 			d = &ch->curr[ccb->ccb_h.target_id];
1613 		else
1614 			d = &ch->user[ccb->ccb_h.target_id];
1615 		if (ch->flags & ATA_SATA) {
1616 			if (cts->xport_specific.sata.valid & CTS_SATA_VALID_REVISION)
1617 				d->revision = cts->xport_specific.sata.revision;
1618 			if (cts->xport_specific.sata.valid & CTS_SATA_VALID_MODE) {
1619 				if (cts->type == CTS_TYPE_CURRENT_SETTINGS) {
1620 					d->mode = ATA_SETMODE(ch->dev,
1621 					    ccb->ccb_h.target_id,
1622 					    cts->xport_specific.sata.mode);
1623 				} else
1624 					d->mode = cts->xport_specific.sata.mode;
1625 			}
1626 			if (cts->xport_specific.sata.valid & CTS_SATA_VALID_BYTECOUNT)
1627 				d->bytecount = min(8192, cts->xport_specific.sata.bytecount);
1628 			if (cts->xport_specific.sata.valid & CTS_SATA_VALID_ATAPI)
1629 				d->atapi = cts->xport_specific.sata.atapi;
1630 		} else {
1631 			if (cts->xport_specific.ata.valid & CTS_ATA_VALID_MODE) {
1632 				if (cts->type == CTS_TYPE_CURRENT_SETTINGS) {
1633 					d->mode = ATA_SETMODE(ch->dev,
1634 					    ccb->ccb_h.target_id,
1635 					    cts->xport_specific.ata.mode);
1636 				} else
1637 					d->mode = cts->xport_specific.ata.mode;
1638 			}
1639 			if (cts->xport_specific.ata.valid & CTS_ATA_VALID_BYTECOUNT)
1640 				d->bytecount = cts->xport_specific.ata.bytecount;
1641 			if (cts->xport_specific.ata.valid & CTS_ATA_VALID_ATAPI)
1642 				d->atapi = cts->xport_specific.ata.atapi;
1643 		}
1644 		ccb->ccb_h.status = CAM_REQ_CMP;
1645 		break;
1646 	}
1647 	case XPT_GET_TRAN_SETTINGS:
1648 	{
1649 		struct	ccb_trans_settings *cts = &ccb->cts;
1650 		struct  ata_cam_device *d;
1651 
1652 		if (ata_check_ids(dev, ccb))
1653 			return;
1654 		if (cts->type == CTS_TYPE_CURRENT_SETTINGS)
1655 			d = &ch->curr[ccb->ccb_h.target_id];
1656 		else
1657 			d = &ch->user[ccb->ccb_h.target_id];
1658 		cts->protocol = PROTO_ATA;
1659 		cts->protocol_version = PROTO_VERSION_UNSPECIFIED;
1660 		if (ch->flags & ATA_SATA) {
1661 			cts->transport = XPORT_SATA;
1662 			cts->transport_version = XPORT_VERSION_UNSPECIFIED;
1663 			cts->xport_specific.sata.valid = 0;
1664 			cts->xport_specific.sata.mode = d->mode;
1665 			cts->xport_specific.sata.valid |= CTS_SATA_VALID_MODE;
1666 			cts->xport_specific.sata.bytecount = d->bytecount;
1667 			cts->xport_specific.sata.valid |= CTS_SATA_VALID_BYTECOUNT;
1668 			if (cts->type == CTS_TYPE_CURRENT_SETTINGS) {
1669 				cts->xport_specific.sata.revision =
1670 				    ATA_GETREV(dev, ccb->ccb_h.target_id);
1671 				if (cts->xport_specific.sata.revision != 0xff) {
1672 					cts->xport_specific.sata.valid |=
1673 					    CTS_SATA_VALID_REVISION;
1674 				}
1675 			} else {
1676 				cts->xport_specific.sata.revision = d->revision;
1677 				cts->xport_specific.sata.valid |= CTS_SATA_VALID_REVISION;
1678 			}
1679 			cts->xport_specific.sata.atapi = d->atapi;
1680 			cts->xport_specific.sata.valid |= CTS_SATA_VALID_ATAPI;
1681 		} else {
1682 			cts->transport = XPORT_ATA;
1683 			cts->transport_version = XPORT_VERSION_UNSPECIFIED;
1684 			cts->xport_specific.ata.valid = 0;
1685 			cts->xport_specific.ata.mode = d->mode;
1686 			cts->xport_specific.ata.valid |= CTS_ATA_VALID_MODE;
1687 			cts->xport_specific.ata.bytecount = d->bytecount;
1688 			cts->xport_specific.ata.valid |= CTS_ATA_VALID_BYTECOUNT;
1689 			cts->xport_specific.ata.atapi = d->atapi;
1690 			cts->xport_specific.ata.valid |= CTS_ATA_VALID_ATAPI;
1691 		}
1692 		ccb->ccb_h.status = CAM_REQ_CMP;
1693 		break;
1694 	}
1695 	case XPT_RESET_BUS:		/* Reset the specified SCSI bus */
1696 	case XPT_RESET_DEV:	/* Bus Device Reset the specified SCSI device */
1697 		ata_reinit(dev);
1698 		ccb->ccb_h.status = CAM_REQ_CMP;
1699 		break;
1700 	case XPT_TERM_IO:		/* Terminate the I/O process */
1701 		/* XXX Implement */
1702 		ccb->ccb_h.status = CAM_REQ_INVALID;
1703 		break;
1704 	case XPT_PATH_INQ:		/* Path routing inquiry */
1705 	{
1706 		struct ccb_pathinq *cpi = &ccb->cpi;
1707 
1708 		parent = device_get_parent(dev);
1709 		cpi->version_num = 1; /* XXX??? */
1710 		cpi->hba_inquiry = PI_SDTR_ABLE;
1711 		cpi->target_sprt = 0;
1712 		cpi->hba_misc = PIM_SEQSCAN;
1713 		cpi->hba_eng_cnt = 0;
1714 		if (ch->flags & ATA_NO_SLAVE)
1715 			cpi->max_target = 0;
1716 		else
1717 			cpi->max_target = 1;
1718 		cpi->max_lun = 0;
1719 		cpi->initiator_id = 0;
1720 		cpi->bus_id = cam_sim_bus(sim);
1721 		if (ch->flags & ATA_SATA)
1722 			cpi->base_transfer_speed = 150000;
1723 		else
1724 			cpi->base_transfer_speed = 3300;
1725 		strncpy(cpi->sim_vid, "FreeBSD", SIM_IDLEN);
1726 		strncpy(cpi->hba_vid, "ATA", HBA_IDLEN);
1727 		strncpy(cpi->dev_name, cam_sim_name(sim), DEV_IDLEN);
1728 		cpi->unit_number = cam_sim_unit(sim);
1729 		if (ch->flags & ATA_SATA)
1730 			cpi->transport = XPORT_SATA;
1731 		else
1732 			cpi->transport = XPORT_ATA;
1733 		cpi->transport_version = XPORT_VERSION_UNSPECIFIED;
1734 		cpi->protocol = PROTO_ATA;
1735 		cpi->protocol_version = PROTO_VERSION_UNSPECIFIED;
1736 		cpi->maxio = ch->dma.max_iosize ? ch->dma.max_iosize : DFLTPHYS;
1737 		if (device_get_devclass(device_get_parent(parent)) ==
1738 		    devclass_find("pci")) {
1739 			cpi->hba_vendor = pci_get_vendor(parent);
1740 			cpi->hba_device = pci_get_device(parent);
1741 			cpi->hba_subvendor = pci_get_subvendor(parent);
1742 			cpi->hba_subdevice = pci_get_subdevice(parent);
1743 		}
1744 		cpi->ccb_h.status = CAM_REQ_CMP;
1745 		break;
1746 	}
1747 	default:
1748 		ccb->ccb_h.status = CAM_REQ_INVALID;
1749 		break;
1750 	}
1751 	xpt_done(ccb);
1752 }
1753 
1754 static void
1755 atapoll(struct cam_sim *sim)
1756 {
1757 	struct ata_channel *ch = (struct ata_channel *)cam_sim_softc(sim);
1758 
1759 	ata_interrupt_locked(ch);
1760 }
1761 #endif
1762 
1763 /*
1764  * module handeling
1765  */
1766 static int
1767 ata_module_event_handler(module_t mod, int what, void *arg)
1768 {
1769 #ifndef ATA_CAM
1770     static struct cdev *atacdev;
1771 #endif
1772 
1773     switch (what) {
1774     case MOD_LOAD:
1775 #ifndef ATA_CAM
1776 	/* register controlling device */
1777 	atacdev = make_dev(&ata_cdevsw, 0, UID_ROOT, GID_OPERATOR, 0600, "ata");
1778 
1779 	if (cold) {
1780 	    /* register boot attach to be run when interrupts are enabled */
1781 	    if (!(ata_delayed_attach = (struct intr_config_hook *)
1782 				       malloc(sizeof(struct intr_config_hook),
1783 					      M_TEMP, M_NOWAIT | M_ZERO))) {
1784 		printf("ata: malloc of delayed attach hook failed\n");
1785 		return EIO;
1786 	    }
1787 	    ata_delayed_attach->ich_func = (void*)ata_boot_attach;
1788 	    if (config_intrhook_establish(ata_delayed_attach) != 0) {
1789 		printf("ata: config_intrhook_establish failed\n");
1790 		free(ata_delayed_attach, M_TEMP);
1791 	    }
1792 	}
1793 #endif
1794 	return 0;
1795 
1796     case MOD_UNLOAD:
1797 #ifndef ATA_CAM
1798 	/* deregister controlling device */
1799 	destroy_dev(atacdev);
1800 #endif
1801 	return 0;
1802 
1803     default:
1804 	return EOPNOTSUPP;
1805     }
1806 }
1807 
1808 static moduledata_t ata_moduledata = { "ata", ata_module_event_handler, NULL };
1809 DECLARE_MODULE(ata, ata_moduledata, SI_SUB_CONFIGURE, SI_ORDER_SECOND);
1810 MODULE_VERSION(ata, 1);
1811 #ifdef ATA_CAM
1812 MODULE_DEPEND(ata, cam, 1, 1, 1);
1813 #endif
1814 
1815 static void
1816 ata_init(void)
1817 {
1818     ata_request_zone = uma_zcreate("ata_request", sizeof(struct ata_request),
1819 				   NULL, NULL, NULL, NULL, 0, 0);
1820     ata_composite_zone = uma_zcreate("ata_composite",
1821 				     sizeof(struct ata_composite),
1822 				     NULL, NULL, NULL, NULL, 0, 0);
1823 }
1824 SYSINIT(ata_register, SI_SUB_DRIVERS, SI_ORDER_SECOND, ata_init, NULL);
1825 
1826 static void
1827 ata_uninit(void)
1828 {
1829     uma_zdestroy(ata_composite_zone);
1830     uma_zdestroy(ata_request_zone);
1831 }
1832 SYSUNINIT(ata_unregister, SI_SUB_DRIVERS, SI_ORDER_SECOND, ata_uninit, NULL);
1833