xref: /titanic_51/usr/src/uts/common/io/aac/aac.c (revision dcda19f50b2b80bfc622fff718ac04fb0e1cb670)
1 /*
2  * Copyright 2009 Sun Microsystems, Inc.  All rights reserved.
3  * Use is subject to license terms.
4  */
5 
6 /*
7  * Copyright 2005-08 Adaptec, Inc.
8  * Copyright (c) 2005-08 Adaptec Inc., Achim Leubner
9  * Copyright (c) 2000 Michael Smith
10  * Copyright (c) 2001 Scott Long
11  * Copyright (c) 2000 BSDi
12  * All rights reserved.
13  *
14  * Redistribution and use in source and binary forms, with or without
15  * modification, are permitted provided that the following conditions
16  * are met:
17  * 1. Redistributions of source code must retain the above copyright
18  *    notice, this list of conditions and the following disclaimer.
19  * 2. Redistributions in binary form must reproduce the above copyright
20  *    notice, this list of conditions and the following disclaimer in the
21  *    documentation and/or other materials provided with the distribution.
22  *
23  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
24  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
25  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
26  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
27  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
28  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
29  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
30  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
31  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
32  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
33  * SUCH DAMAGE.
34  */
35 #include <sys/modctl.h>
36 #include <sys/conf.h>
37 #include <sys/cmn_err.h>
38 #include <sys/ddi.h>
39 #include <sys/devops.h>
40 #include <sys/pci.h>
41 #include <sys/types.h>
42 #include <sys/ddidmareq.h>
43 #include <sys/scsi/scsi.h>
44 #include <sys/ksynch.h>
45 #include <sys/sunddi.h>
46 #include <sys/byteorder.h>
47 #include "aac_regs.h"
48 #include "aac.h"
49 
50 /*
51  * FMA header files
52  */
53 #include <sys/ddifm.h>
54 #include <sys/fm/protocol.h>
55 #include <sys/fm/util.h>
56 #include <sys/fm/io/ddi.h>
57 
58 /*
59  * For minor nodes created by the SCSA framework, minor numbers are
60  * formed by left-shifting instance by INST_MINOR_SHIFT and OR in a
61  * number less than 64.
62  *
63  * To support cfgadm, need to confirm the SCSA framework by creating
64  * devctl/scsi and driver specific minor nodes under SCSA format,
65  * and calling scsi_hba_xxx() functions aacordingly.
66  */
67 
68 #define	AAC_MINOR		32
69 #define	INST2AAC(x)		(((x) << INST_MINOR_SHIFT) | AAC_MINOR)
70 #define	AAC_SCSA_MINOR(x)	((x) & TRAN_MINOR_MASK)
71 #define	AAC_IS_SCSA_NODE(x)	((x) == DEVCTL_MINOR || (x) == SCSI_MINOR)
72 
73 #define	SD2TRAN(sd)		((sd)->sd_address.a_hba_tran)
74 #define	AAC_TRAN2SOFTS(tran) ((struct aac_softstate *)(tran)->tran_hba_private)
75 #define	AAC_DIP2TRAN(dip)	((scsi_hba_tran_t *)ddi_get_driver_private(dip))
76 #define	AAC_DIP2SOFTS(dip)	(AAC_TRAN2SOFTS(AAC_DIP2TRAN(dip)))
77 #define	SD2AAC(sd)		(AAC_TRAN2SOFTS(SD2TRAN(sd)))
78 #define	AAC_PD(t)		((t) - AAC_MAX_LD)
79 #define	AAC_DEV(softs, t)	(((t) < AAC_MAX_LD) ? \
80 				&(softs)->containers[(t)].dev : \
81 				((t) < AAC_MAX_DEV(softs)) ? \
82 				&(softs)->nondasds[AAC_PD(t)].dev : NULL)
83 #define	AAC_DEVCFG_BEGIN(softs, tgt) \
84 				aac_devcfg((softs), (tgt), 1)
85 #define	AAC_DEVCFG_END(softs, tgt) \
86 				aac_devcfg((softs), (tgt), 0)
87 #define	PKT2AC(pkt)		((struct aac_cmd *)(pkt)->pkt_ha_private)
88 #define	AAC_BUSYWAIT(cond, timeout /* in millisecond */) { \
89 		if (!(cond)) { \
90 			int count = (timeout) * 10; \
91 			while (count) { \
92 				drv_usecwait(100); \
93 				if (cond) \
94 					break; \
95 				count--; \
96 			} \
97 			(timeout) = (count + 9) / 10; \
98 		} \
99 	}
100 
101 #define	AAC_SENSE_DATA_DESCR_LEN \
102 	(sizeof (struct scsi_descr_sense_hdr) + \
103 	sizeof (struct scsi_information_sense_descr))
104 #define	AAC_ARQ64_LENGTH \
105 	(sizeof (struct scsi_arq_status) + \
106 	AAC_SENSE_DATA_DESCR_LEN - SENSE_LENGTH)
107 
108 /* NOTE: GETG4ADDRTL(cdbp) is int32_t */
109 #define	AAC_GETGXADDR(cmdlen, cdbp) \
110 	((cmdlen == 6) ? GETG0ADDR(cdbp) : \
111 	(cmdlen == 10) ? (uint32_t)GETG1ADDR(cdbp) : \
112 	((uint64_t)GETG4ADDR(cdbp) << 32) | (uint32_t)GETG4ADDRTL(cdbp))
113 
114 #define	AAC_CDB_INQUIRY_CMDDT	0x02
115 #define	AAC_CDB_INQUIRY_EVPD	0x01
116 #define	AAC_VPD_PAGE_CODE	1
117 #define	AAC_VPD_PAGE_LENGTH	3
118 #define	AAC_VPD_PAGE_DATA	4
119 #define	AAC_VPD_ID_CODESET	0
120 #define	AAC_VPD_ID_TYPE		1
121 #define	AAC_VPD_ID_LENGTH	3
122 #define	AAC_VPD_ID_DATA		4
123 
124 #define	AAC_SCSI_RPTLUNS_HEAD_SIZE			0x08
125 #define	AAC_SCSI_RPTLUNS_ADDR_SIZE			0x08
126 #define	AAC_SCSI_RPTLUNS_ADDR_MASK			0xC0
127 /* 00b - peripheral device addressing method */
128 #define	AAC_SCSI_RPTLUNS_ADDR_PERIPHERAL		0x00
129 /* 01b - flat space addressing method */
130 #define	AAC_SCSI_RPTLUNS_ADDR_FLAT_SPACE		0x40
131 /* 10b - logical unit addressing method */
132 #define	AAC_SCSI_RPTLUNS_ADDR_LOGICAL_UNIT		0x80
133 
134 /* Return the size of FIB with data part type data_type */
135 #define	AAC_FIB_SIZEOF(data_type) \
136 	(sizeof (struct aac_fib_header) + sizeof (data_type))
137 /* Return the container size defined in mir */
138 #define	AAC_MIR_SIZE(softs, acc, mir) \
139 	(((softs)->flags & AAC_FLAGS_LBA_64BIT) ? \
140 	(uint64_t)ddi_get32((acc), &(mir)->MntObj.Capacity) + \
141 	((uint64_t)ddi_get32((acc), &(mir)->MntObj.CapacityHigh) << 32) : \
142 	(uint64_t)ddi_get32((acc), &(mir)->MntObj.Capacity))
143 
144 /* The last entry of aac_cards[] is for unknown cards */
145 #define	AAC_UNKNOWN_CARD \
146 	(sizeof (aac_cards) / sizeof (struct aac_card_type) - 1)
147 #define	CARD_IS_UNKNOWN(i)	(i == AAC_UNKNOWN_CARD)
148 #define	BUF_IS_READ(bp)		((bp)->b_flags & B_READ)
149 #define	AAC_IS_Q_EMPTY(q)	((q)->q_head == NULL)
150 #define	AAC_CMDQ(acp)		(!((acp)->flags & AAC_CMD_SYNC))
151 
152 #define	PCI_MEM_GET32(softs, off) \
153 	ddi_get32((softs)->pci_mem_handle, \
154 	    (void *)((softs)->pci_mem_base_vaddr + (off)))
155 #define	PCI_MEM_PUT32(softs, off, val) \
156 	ddi_put32((softs)->pci_mem_handle, \
157 	    (void *)((softs)->pci_mem_base_vaddr + (off)), \
158 	    (uint32_t)(val))
159 #define	PCI_MEM_GET16(softs, off) \
160 	ddi_get16((softs)->pci_mem_handle, \
161 	(void *)((softs)->pci_mem_base_vaddr + (off)))
162 #define	PCI_MEM_PUT16(softs, off, val) \
163 	ddi_put16((softs)->pci_mem_handle, \
164 	(void *)((softs)->pci_mem_base_vaddr + (off)), (uint16_t)(val))
165 /* Write host data at valp to device mem[off] repeatedly count times */
166 #define	PCI_MEM_REP_PUT8(softs, off, valp, count) \
167 	ddi_rep_put8((softs)->pci_mem_handle, (uint8_t *)(valp), \
168 	    (uint8_t *)((softs)->pci_mem_base_vaddr + (off)), \
169 	    count, DDI_DEV_AUTOINCR)
170 /* Read device data at mem[off] to host addr valp repeatedly count times */
171 #define	PCI_MEM_REP_GET8(softs, off, valp, count) \
172 	ddi_rep_get8((softs)->pci_mem_handle, (uint8_t *)(valp), \
173 	    (uint8_t *)((softs)->pci_mem_base_vaddr + (off)), \
174 	    count, DDI_DEV_AUTOINCR)
175 #define	AAC_GET_FIELD8(acc, d, s, field) \
176 	(d)->field = ddi_get8(acc, (uint8_t *)&(s)->field)
177 #define	AAC_GET_FIELD32(acc, d, s, field) \
178 	(d)->field = ddi_get32(acc, (uint32_t *)&(s)->field)
179 #define	AAC_GET_FIELD64(acc, d, s, field) \
180 	(d)->field = ddi_get64(acc, (uint64_t *)&(s)->field)
181 #define	AAC_REP_GET_FIELD8(acc, d, s, field, r) \
182 	ddi_rep_get8((acc), (uint8_t *)&(d)->field, \
183 	    (uint8_t *)&(s)->field, (r), DDI_DEV_AUTOINCR)
184 #define	AAC_REP_GET_FIELD32(acc, d, s, field, r) \
185 	ddi_rep_get32((acc), (uint32_t *)&(d)->field, \
186 	    (uint32_t *)&(s)->field, (r), DDI_DEV_AUTOINCR)
187 
188 #define	AAC_ENABLE_INTR(softs) { \
189 		if (softs->flags & AAC_FLAGS_NEW_COMM) \
190 			PCI_MEM_PUT32(softs, AAC_OIMR, ~AAC_DB_INTR_NEW); \
191 		else \
192 			PCI_MEM_PUT32(softs, AAC_OIMR, ~AAC_DB_INTR_BITS); \
193 	}
194 
195 #define	AAC_DISABLE_INTR(softs)		PCI_MEM_PUT32(softs, AAC_OIMR, ~0)
196 #define	AAC_STATUS_CLR(softs, mask)	PCI_MEM_PUT32(softs, AAC_ODBR, mask)
197 #define	AAC_STATUS_GET(softs)		PCI_MEM_GET32(softs, AAC_ODBR)
198 #define	AAC_NOTIFY(softs, val)		PCI_MEM_PUT32(softs, AAC_IDBR, val)
199 #define	AAC_OUTB_GET(softs)		PCI_MEM_GET32(softs, AAC_OQUE)
200 #define	AAC_OUTB_SET(softs, val)	PCI_MEM_PUT32(softs, AAC_OQUE, val)
201 #define	AAC_FWSTATUS_GET(softs)	\
202 	((softs)->aac_if.aif_get_fwstatus(softs))
203 #define	AAC_MAILBOX_GET(softs, mb) \
204 	((softs)->aac_if.aif_get_mailbox((softs), (mb)))
205 #define	AAC_MAILBOX_SET(softs, cmd, arg0, arg1, arg2, arg3) \
206 	((softs)->aac_if.aif_set_mailbox((softs), (cmd), \
207 	    (arg0), (arg1), (arg2), (arg3)))
208 
209 #define	AAC_THROTTLE_DRAIN	-1
210 
211 #define	AAC_QUIESCE_TICK	1	/* 1 second */
212 #define	AAC_QUIESCE_TIMEOUT	180	/* 180 seconds */
213 #define	AAC_DEFAULT_TICK	10	/* 10 seconds */
214 #define	AAC_SYNC_TICK		(30*60)	/* 30 minutes */
215 
216 /* Poll time for aac_do_poll_io() */
217 #define	AAC_POLL_TIME		60	/* 60 seconds */
218 
219 /* IOP reset */
220 #define	AAC_IOP_RESET_SUCCEED		0	/* IOP reset succeed */
221 #define	AAC_IOP_RESET_FAILED		-1	/* IOP reset failed */
222 #define	AAC_IOP_RESET_ABNORMAL		-2	/* Reset operation abnormal */
223 
224 /*
225  * Hardware access functions
226  */
227 static int aac_rx_get_fwstatus(struct aac_softstate *);
228 static int aac_rx_get_mailbox(struct aac_softstate *, int);
229 static void aac_rx_set_mailbox(struct aac_softstate *, uint32_t, uint32_t,
230     uint32_t, uint32_t, uint32_t);
231 static int aac_rkt_get_fwstatus(struct aac_softstate *);
232 static int aac_rkt_get_mailbox(struct aac_softstate *, int);
233 static void aac_rkt_set_mailbox(struct aac_softstate *, uint32_t, uint32_t,
234     uint32_t, uint32_t, uint32_t);
235 
236 /*
237  * SCSA function prototypes
238  */
239 static int aac_attach(dev_info_t *, ddi_attach_cmd_t);
240 static int aac_detach(dev_info_t *, ddi_detach_cmd_t);
241 static int aac_reset(dev_info_t *, ddi_reset_cmd_t);
242 static int aac_quiesce(dev_info_t *);
243 
244 /*
245  * Interrupt handler functions
246  */
247 static int aac_query_intrs(struct aac_softstate *, int);
248 static int aac_add_intrs(struct aac_softstate *);
249 static void aac_remove_intrs(struct aac_softstate *);
250 static uint_t aac_intr_old(caddr_t);
251 static uint_t aac_intr_new(caddr_t);
252 static uint_t aac_softintr(caddr_t);
253 
254 /*
255  * Internal functions in attach
256  */
257 static int aac_check_card_type(struct aac_softstate *);
258 static int aac_check_firmware(struct aac_softstate *);
259 static int aac_common_attach(struct aac_softstate *);
260 static void aac_common_detach(struct aac_softstate *);
261 static int aac_probe_containers(struct aac_softstate *);
262 static int aac_alloc_comm_space(struct aac_softstate *);
263 static int aac_setup_comm_space(struct aac_softstate *);
264 static void aac_free_comm_space(struct aac_softstate *);
265 static int aac_hba_setup(struct aac_softstate *);
266 
267 /*
268  * Sync FIB operation functions
269  */
270 int aac_sync_mbcommand(struct aac_softstate *, uint32_t, uint32_t,
271     uint32_t, uint32_t, uint32_t, uint32_t *);
272 static int aac_sync_fib(struct aac_softstate *, uint16_t, uint16_t);
273 
274 /*
275  * Command queue operation functions
276  */
277 static void aac_cmd_initq(struct aac_cmd_queue *);
278 static void aac_cmd_enqueue(struct aac_cmd_queue *, struct aac_cmd *);
279 static struct aac_cmd *aac_cmd_dequeue(struct aac_cmd_queue *);
280 static void aac_cmd_delete(struct aac_cmd_queue *, struct aac_cmd *);
281 
282 /*
283  * FIB queue operation functions
284  */
285 static int aac_fib_enqueue(struct aac_softstate *, int, uint32_t, uint32_t);
286 static int aac_fib_dequeue(struct aac_softstate *, int, int *);
287 
288 /*
289  * Slot operation functions
290  */
291 static int aac_create_slots(struct aac_softstate *);
292 static void aac_destroy_slots(struct aac_softstate *);
293 static void aac_alloc_fibs(struct aac_softstate *);
294 static void aac_destroy_fibs(struct aac_softstate *);
295 static struct aac_slot *aac_get_slot(struct aac_softstate *);
296 static void aac_release_slot(struct aac_softstate *, struct aac_slot *);
297 static int aac_alloc_fib(struct aac_softstate *, struct aac_slot *);
298 static void aac_free_fib(struct aac_slot *);
299 
300 /*
301  * Internal functions
302  */
303 static void aac_cmd_fib_header(struct aac_softstate *, struct aac_slot *,
304     uint16_t, uint16_t);
305 static void aac_cmd_fib_rawio(struct aac_softstate *, struct aac_cmd *);
306 static void aac_cmd_fib_brw64(struct aac_softstate *, struct aac_cmd *);
307 static void aac_cmd_fib_brw(struct aac_softstate *, struct aac_cmd *);
308 static void aac_cmd_fib_sync(struct aac_softstate *, struct aac_cmd *);
309 static void aac_cmd_fib_scsi32(struct aac_softstate *, struct aac_cmd *);
310 static void aac_cmd_fib_scsi64(struct aac_softstate *, struct aac_cmd *);
311 static void aac_start_waiting_io(struct aac_softstate *);
312 static void aac_drain_comp_q(struct aac_softstate *);
313 int aac_do_io(struct aac_softstate *, struct aac_cmd *);
314 static int aac_do_poll_io(struct aac_softstate *, struct aac_cmd *);
315 static int aac_do_sync_io(struct aac_softstate *, struct aac_cmd *);
316 static int aac_send_command(struct aac_softstate *, struct aac_slot *);
317 static void aac_cmd_timeout(struct aac_softstate *, struct aac_cmd *);
318 static int aac_dma_sync_ac(struct aac_cmd *);
319 static int aac_shutdown(struct aac_softstate *);
320 static int aac_reset_adapter(struct aac_softstate *);
321 static int aac_do_quiesce(struct aac_softstate *softs);
322 static int aac_do_unquiesce(struct aac_softstate *softs);
323 static void aac_unhold_bus(struct aac_softstate *, int);
324 static void aac_set_throttle(struct aac_softstate *, struct aac_device *,
325     int, int);
326 
327 /*
328  * Adapter Initiated FIB handling function
329  */
330 static int aac_handle_aif(struct aac_softstate *, struct aac_fib *);
331 
332 /*
333  * Timeout handling thread function
334  */
335 static void aac_daemon(void *);
336 
337 /*
338  * IOCTL interface related functions
339  */
340 static int aac_open(dev_t *, int, int, cred_t *);
341 static int aac_close(dev_t, int, int, cred_t *);
342 static int aac_ioctl(dev_t, int, intptr_t, int, cred_t *, int *);
343 extern int aac_do_ioctl(struct aac_softstate *, dev_t, int, intptr_t, int);
344 
345 /*
346  * FMA Prototypes
347  */
348 static void aac_fm_init(struct aac_softstate *);
349 static void aac_fm_fini(struct aac_softstate *);
350 static int aac_fm_error_cb(dev_info_t *, ddi_fm_error_t *, const void *);
351 int aac_check_acc_handle(ddi_acc_handle_t);
352 int aac_check_dma_handle(ddi_dma_handle_t);
353 void aac_fm_ereport(struct aac_softstate *, char *);
354 
355 /*
356  * Auto enumeration functions
357  */
358 static dev_info_t *aac_find_child(struct aac_softstate *, uint16_t, uint8_t);
359 static int aac_tran_bus_config(dev_info_t *, uint_t, ddi_bus_config_op_t,
360     void *, dev_info_t **);
361 static int aac_dr_event(struct aac_softstate *, int, int, int);
362 
363 #ifdef DEBUG
364 /*
365  * UART	debug output support
366  */
367 
368 #define	AAC_PRINT_BUFFER_SIZE		512
369 #define	AAC_PRINT_TIMEOUT		250	/* 1/4 sec. = 250 msec. */
370 
371 #define	AAC_FW_DBG_STRLEN_OFFSET	0x00
372 #define	AAC_FW_DBG_FLAGS_OFFSET		0x04
373 #define	AAC_FW_DBG_BLED_OFFSET		0x08
374 
375 static int aac_get_fw_debug_buffer(struct aac_softstate *);
376 static void aac_print_scmd(struct aac_softstate *, struct aac_cmd *);
377 static void aac_print_aif(struct aac_softstate *, struct aac_aif_command *);
378 
379 static char aac_prt_buf[AAC_PRINT_BUFFER_SIZE];
380 static char aac_fmt[] = " %s";
381 static char aac_fmt_header[] = " %s.%d: %s";
382 static kmutex_t aac_prt_mutex;
383 
384 /*
385  * Debug flags to be put into the softstate flags field
386  * when initialized
387  */
388 uint32_t aac_debug_flags =
389 /*    AACDB_FLAGS_KERNEL_PRINT | */
390 /*    AACDB_FLAGS_FW_PRINT |	*/
391 /*    AACDB_FLAGS_MISC |	*/
392 /*    AACDB_FLAGS_FUNC1 |	*/
393 /*    AACDB_FLAGS_FUNC2 |	*/
394 /*    AACDB_FLAGS_SCMD |	*/
395 /*    AACDB_FLAGS_AIF |		*/
396 /*    AACDB_FLAGS_FIB |		*/
397 /*    AACDB_FLAGS_IOCTL |	*/
398 0;
399 uint32_t aac_debug_fib_flags =
400 /*    AACDB_FLAGS_FIB_RW |	*/
401 /*    AACDB_FLAGS_FIB_IOCTL |	*/
402 /*    AACDB_FLAGS_FIB_SRB |	*/
403 /*    AACDB_FLAGS_FIB_SYNC |	*/
404 /*    AACDB_FLAGS_FIB_HEADER |	*/
405 /*    AACDB_FLAGS_FIB_TIMEOUT |	*/
406 0;
407 
408 #endif /* DEBUG */
409 
410 static struct cb_ops aac_cb_ops = {
411 	aac_open,	/* open */
412 	aac_close,	/* close */
413 	nodev,		/* strategy */
414 	nodev,		/* print */
415 	nodev,		/* dump */
416 	nodev,		/* read */
417 	nodev,		/* write */
418 	aac_ioctl,	/* ioctl */
419 	nodev,		/* devmap */
420 	nodev,		/* mmap */
421 	nodev,		/* segmap */
422 	nochpoll,	/* poll */
423 	ddi_prop_op,	/* cb_prop_op */
424 	NULL,		/* streamtab */
425 	D_64BIT | D_NEW | D_MP | D_HOTPLUG,	/* cb_flag */
426 	CB_REV,		/* cb_rev */
427 	nodev,		/* async I/O read entry point */
428 	nodev		/* async I/O write entry point */
429 };
430 
431 static struct dev_ops aac_dev_ops = {
432 	DEVO_REV,
433 	0,
434 	nodev,
435 	nulldev,
436 	nulldev,
437 	aac_attach,
438 	aac_detach,
439 	aac_reset,
440 	&aac_cb_ops,
441 	NULL,
442 	NULL,
443 	aac_quiesce,
444 };
445 
446 static struct modldrv aac_modldrv = {
447 	&mod_driverops,
448 	"AAC Driver " AAC_DRIVER_VERSION,
449 	&aac_dev_ops,
450 };
451 
452 static struct modlinkage aac_modlinkage = {
453 	MODREV_1,
454 	&aac_modldrv,
455 	NULL
456 };
457 
458 static struct aac_softstate  *aac_softstatep;
459 
460 /*
461  * Supported card list
462  * ordered in vendor id, subvendor id, subdevice id, and device id
463  */
464 static struct aac_card_type aac_cards[] = {
465 	{0x1028, 0x1, 0x1028, 0x1, AAC_HWIF_I960RX,
466 	    AAC_FLAGS_NO4GB | AAC_FLAGS_34SG, AAC_TYPE_SCSI,
467 	    "Dell", "PERC 3/Di"},
468 	{0x1028, 0x2, 0x1028, 0x2, AAC_HWIF_I960RX,
469 	    AAC_FLAGS_NO4GB | AAC_FLAGS_34SG, AAC_TYPE_SCSI,
470 	    "Dell", "PERC 3/Di"},
471 	{0x1028, 0x3, 0x1028, 0x3, AAC_HWIF_I960RX,
472 	    AAC_FLAGS_NO4GB | AAC_FLAGS_34SG, AAC_TYPE_SCSI,
473 	    "Dell", "PERC 3/Si"},
474 	{0x1028, 0x8, 0x1028, 0xcf, AAC_HWIF_I960RX,
475 	    AAC_FLAGS_NO4GB | AAC_FLAGS_34SG, AAC_TYPE_SCSI,
476 	    "Dell", "PERC 3/Di"},
477 	{0x1028, 0x4, 0x1028, 0xd0, AAC_HWIF_I960RX,
478 	    AAC_FLAGS_NO4GB | AAC_FLAGS_34SG, AAC_TYPE_SCSI,
479 	    "Dell", "PERC 3/Si"},
480 	{0x1028, 0x2, 0x1028, 0xd1, AAC_HWIF_I960RX,
481 	    AAC_FLAGS_NO4GB | AAC_FLAGS_34SG, AAC_TYPE_SCSI,
482 	    "Dell", "PERC 3/Di"},
483 	{0x1028, 0x2, 0x1028, 0xd9, AAC_HWIF_I960RX,
484 	    AAC_FLAGS_NO4GB | AAC_FLAGS_34SG, AAC_TYPE_SCSI,
485 	    "Dell", "PERC 3/Di"},
486 	{0x1028, 0xa, 0x1028, 0x106, AAC_HWIF_I960RX,
487 	    AAC_FLAGS_NO4GB | AAC_FLAGS_34SG, AAC_TYPE_SCSI,
488 	    "Dell", "PERC 3/Di"},
489 	{0x1028, 0xa, 0x1028, 0x11b, AAC_HWIF_I960RX,
490 	    AAC_FLAGS_NO4GB | AAC_FLAGS_34SG, AAC_TYPE_SCSI,
491 	    "Dell", "PERC 3/Di"},
492 	{0x1028, 0xa, 0x1028, 0x121, AAC_HWIF_I960RX,
493 	    AAC_FLAGS_NO4GB | AAC_FLAGS_34SG, AAC_TYPE_SCSI,
494 	    "Dell", "PERC 3/Di"},
495 	{0x9005, 0x285, 0x1028, 0x287, AAC_HWIF_I960RX,
496 	    AAC_FLAGS_NO4GB | AAC_FLAGS_34SG | AAC_FLAGS_256FIBS, AAC_TYPE_SCSI,
497 	    "Dell", "PERC 320/DC"},
498 	{0x9005, 0x285, 0x1028, 0x291, AAC_HWIF_I960RX,
499 	    AAC_FLAGS_17SG, AAC_TYPE_SATA, "Dell", "CERC SR2"},
500 
501 	{0x9005, 0x285, 0x1014, 0x2f2, AAC_HWIF_I960RX,
502 	    0, AAC_TYPE_SCSI, "IBM", "ServeRAID 8i"},
503 	{0x9005, 0x285, 0x1014, 0x34d, AAC_HWIF_I960RX,
504 	    0, AAC_TYPE_SAS, "IBM", "ServeRAID 8s"},
505 	{0x9005, 0x286, 0x1014, 0x9580, AAC_HWIF_RKT,
506 	    0, AAC_TYPE_SAS, "IBM", "ServeRAID 8k"},
507 
508 	{0x9005, 0x285, 0x103c, 0x3227, AAC_HWIF_I960RX,
509 	    AAC_FLAGS_17SG, AAC_TYPE_SATA, "Adaptec", "2610SA"},
510 	{0x9005, 0x285, 0xe11, 0x295, AAC_HWIF_I960RX,
511 	    AAC_FLAGS_17SG, AAC_TYPE_SATA, "Adaptec", "2610SA"},
512 
513 	{0x9005, 0x285, 0x9005, 0x285, AAC_HWIF_I960RX,
514 	    AAC_FLAGS_NO4GB | AAC_FLAGS_34SG | AAC_FLAGS_256FIBS, AAC_TYPE_SCSI,
515 	    "Adaptec", "2200S"},
516 	{0x9005, 0x285, 0x9005, 0x286, AAC_HWIF_I960RX,
517 	    AAC_FLAGS_NO4GB | AAC_FLAGS_34SG | AAC_FLAGS_256FIBS, AAC_TYPE_SCSI,
518 	    "Adaptec", "2120S"},
519 	{0x9005, 0x285, 0x9005, 0x287, AAC_HWIF_I960RX,
520 	    AAC_FLAGS_NO4GB | AAC_FLAGS_34SG | AAC_FLAGS_256FIBS, AAC_TYPE_SCSI,
521 	    "Adaptec", "2200S"},
522 	{0x9005, 0x285, 0x9005, 0x288, AAC_HWIF_I960RX,
523 	    0, AAC_TYPE_SCSI, "Adaptec", "3230S"},
524 	{0x9005, 0x285, 0x9005, 0x289, AAC_HWIF_I960RX,
525 	    0, AAC_TYPE_SCSI, "Adaptec", "3240S"},
526 	{0x9005, 0x285, 0x9005, 0x28a, AAC_HWIF_I960RX,
527 	    0, AAC_TYPE_SCSI, "Adaptec", "2020ZCR"},
528 	{0x9005, 0x285, 0x9005, 0x28b, AAC_HWIF_I960RX,
529 	    0, AAC_TYPE_SCSI, "Adaptec", "2025ZCR"},
530 	{0x9005, 0x286, 0x9005, 0x28c, AAC_HWIF_RKT,
531 	    0, AAC_TYPE_SCSI, "Adaptec", "2230S"},
532 	{0x9005, 0x286, 0x9005, 0x28d, AAC_HWIF_RKT,
533 	    0, AAC_TYPE_SCSI, "Adaptec", "2130S"},
534 	{0x9005, 0x285, 0x9005, 0x28e, AAC_HWIF_I960RX,
535 	    0, AAC_TYPE_SATA, "Adaptec", "2020SA"},
536 	{0x9005, 0x285, 0x9005, 0x28f, AAC_HWIF_I960RX,
537 	    0, AAC_TYPE_SATA, "Adaptec", "2025SA"},
538 	{0x9005, 0x285, 0x9005, 0x290, AAC_HWIF_I960RX,
539 	    AAC_FLAGS_17SG, AAC_TYPE_SATA, "Adaptec", "2410SA"},
540 	{0x9005, 0x285, 0x9005, 0x292, AAC_HWIF_I960RX,
541 	    AAC_FLAGS_17SG, AAC_TYPE_SATA, "Adaptec", "2810SA"},
542 	{0x9005, 0x285, 0x9005, 0x293, AAC_HWIF_I960RX,
543 	    AAC_FLAGS_17SG, AAC_TYPE_SATA, "Adaptec", "21610SA"},
544 	{0x9005, 0x285, 0x9005, 0x294, AAC_HWIF_I960RX,
545 	    0, AAC_TYPE_SATA, "Adaptec", "2026ZCR"},
546 	{0x9005, 0x285, 0x9005, 0x296, AAC_HWIF_I960RX,
547 	    0, AAC_TYPE_SCSI, "Adaptec", "2240S"},
548 	{0x9005, 0x285, 0x9005, 0x297, AAC_HWIF_I960RX,
549 	    0, AAC_TYPE_SAS, "Adaptec", "4005SAS"},
550 	{0x9005, 0x285, 0x9005, 0x298, AAC_HWIF_I960RX,
551 	    0, AAC_TYPE_SAS, "Adaptec", "RAID 4000"},
552 	{0x9005, 0x285, 0x9005, 0x299, AAC_HWIF_I960RX,
553 	    0, AAC_TYPE_SAS, "Adaptec", "4800SAS"},
554 	{0x9005, 0x285, 0x9005, 0x29a, AAC_HWIF_I960RX,
555 	    0, AAC_TYPE_SAS, "Adaptec", "4805SAS"},
556 	{0x9005, 0x286, 0x9005, 0x29b, AAC_HWIF_RKT,
557 	    0, AAC_TYPE_SATA, "Adaptec", "2820SA"},
558 	{0x9005, 0x286, 0x9005, 0x29c, AAC_HWIF_RKT,
559 	    0, AAC_TYPE_SATA, "Adaptec", "2620SA"},
560 	{0x9005, 0x286, 0x9005, 0x29d, AAC_HWIF_RKT,
561 	    0, AAC_TYPE_SATA, "Adaptec", "2420SA"},
562 	{0x9005, 0x286, 0x9005, 0x29e, AAC_HWIF_RKT,
563 	    0, AAC_TYPE_SATA, "ICP", "9024RO"},
564 	{0x9005, 0x286, 0x9005, 0x29f, AAC_HWIF_RKT,
565 	    0, AAC_TYPE_SATA, "ICP", "9014RO"},
566 	{0x9005, 0x286, 0x9005, 0x2a0, AAC_HWIF_RKT,
567 	    0, AAC_TYPE_SATA, "ICP", "9047MA"},
568 	{0x9005, 0x286, 0x9005, 0x2a1, AAC_HWIF_RKT,
569 	    0, AAC_TYPE_SATA, "ICP", "9087MA"},
570 	{0x9005, 0x285, 0x9005, 0x2a4, AAC_HWIF_I960RX,
571 	    0, AAC_TYPE_SAS, "ICP", "9085LI"},
572 	{0x9005, 0x285, 0x9005, 0x2a5, AAC_HWIF_I960RX,
573 	    0, AAC_TYPE_SAS, "ICP", "5085BR"},
574 	{0x9005, 0x286, 0x9005, 0x2a6, AAC_HWIF_RKT,
575 	    0, AAC_TYPE_SATA, "ICP", "9067MA"},
576 	{0x9005, 0x285, 0x9005, 0x2b5, AAC_HWIF_I960RX,
577 	    0, AAC_TYPE_SAS, "Adaptec", "RAID 5445"},
578 	{0x9005, 0x285, 0x9005, 0x2b6, AAC_HWIF_I960RX,
579 	    0, AAC_TYPE_SAS, "Adaptec", "RAID 5805"},
580 	{0x9005, 0x285, 0x9005, 0x2b7, AAC_HWIF_I960RX,
581 	    0, AAC_TYPE_SAS, "Adaptec", "RAID 5085"},
582 	{0x9005, 0x285, 0x9005, 0x2b8, AAC_HWIF_I960RX,
583 	    0, AAC_TYPE_SAS, "ICP", "RAID ICP5445SL"},
584 	{0x9005, 0x285, 0x9005, 0x2b9, AAC_HWIF_I960RX,
585 	    0, AAC_TYPE_SAS, "ICP", "RAID ICP5085SL"},
586 	{0x9005, 0x285, 0x9005, 0x2ba, AAC_HWIF_I960RX,
587 	    0, AAC_TYPE_SAS, "ICP", "RAID ICP5805SL"},
588 
589 	{0, 0, 0, 0, AAC_HWIF_UNKNOWN,
590 	    0, AAC_TYPE_UNKNOWN, "Unknown", "AAC card"},
591 };
592 
593 /*
594  * Hardware access functions for i960 based cards
595  */
596 static struct aac_interface aac_rx_interface = {
597 	aac_rx_get_fwstatus,
598 	aac_rx_get_mailbox,
599 	aac_rx_set_mailbox
600 };
601 
602 /*
603  * Hardware access functions for Rocket based cards
604  */
605 static struct aac_interface aac_rkt_interface = {
606 	aac_rkt_get_fwstatus,
607 	aac_rkt_get_mailbox,
608 	aac_rkt_set_mailbox
609 };
610 
611 ddi_device_acc_attr_t aac_acc_attr = {
612 	DDI_DEVICE_ATTR_V0,
613 	DDI_STRUCTURE_LE_ACC,
614 	DDI_STRICTORDER_ACC
615 };
616 
617 static struct {
618 	int	size;
619 	int	notify;
620 } aac_qinfo[] = {
621 	{AAC_HOST_NORM_CMD_ENTRIES, AAC_DB_COMMAND_NOT_FULL},
622 	{AAC_HOST_HIGH_CMD_ENTRIES, 0},
623 	{AAC_ADAP_NORM_CMD_ENTRIES, AAC_DB_COMMAND_READY},
624 	{AAC_ADAP_HIGH_CMD_ENTRIES, 0},
625 	{AAC_HOST_NORM_RESP_ENTRIES, AAC_DB_RESPONSE_NOT_FULL},
626 	{AAC_HOST_HIGH_RESP_ENTRIES, 0},
627 	{AAC_ADAP_NORM_RESP_ENTRIES, AAC_DB_RESPONSE_READY},
628 	{AAC_ADAP_HIGH_RESP_ENTRIES, 0}
629 };
630 
631 /*
632  * Default aac dma attributes
633  */
634 static ddi_dma_attr_t aac_dma_attr = {
635 	DMA_ATTR_V0,
636 	0,		/* lowest usable address */
637 	0xffffffffull,	/* high DMA address range */
638 	0xffffffffull,	/* DMA counter register */
639 	AAC_DMA_ALIGN,	/* DMA address alignment */
640 	1,		/* DMA burstsizes */
641 	1,		/* min effective DMA size */
642 	0xffffffffull,	/* max DMA xfer size */
643 	0xffffffffull,	/* segment boundary */
644 	1,		/* s/g list length */
645 	AAC_BLK_SIZE,	/* granularity of device */
646 	0		/* DMA transfer flags */
647 };
648 
649 struct aac_drinfo {
650 	struct aac_softstate *softs;
651 	int tgt;
652 	int lun;
653 	int event;
654 };
655 
656 static int aac_tick = AAC_DEFAULT_TICK;	/* tick for the internal timer */
657 static uint32_t aac_timebase = 0;	/* internal timer in seconds */
658 static uint32_t aac_sync_time = 0;	/* next time to sync. with firmware */
659 
660 /*
661  * Warlock directives
662  *
663  * Different variables with the same types have to be protected by the
664  * same mutex; otherwise, warlock will complain with "variables don't
665  * seem to be protected consistently". For example,
666  * aac_softstate::{q_wait, q_comp} are type of aac_cmd_queue, and protected
667  * by aac_softstate::{io_lock, q_comp_mutex} respectively. We have to
668  * declare them as protected explictly at aac_cmd_dequeue().
669  */
670 _NOTE(SCHEME_PROTECTS_DATA("unique per pkt", scsi_pkt scsi_cdb scsi_status \
671     scsi_arq_status scsi_descr_sense_hdr scsi_information_sense_descr \
672     mode_format mode_geometry mode_header aac_cmd))
673 _NOTE(SCHEME_PROTECTS_DATA("unique per aac_cmd", aac_fib ddi_dma_cookie_t \
674     aac_sge))
675 _NOTE(SCHEME_PROTECTS_DATA("unique per aac_fib", aac_blockread aac_blockwrite \
676     aac_blockread64 aac_raw_io aac_sg_entry aac_sg_entry64 aac_sg_entryraw \
677     aac_sg_table aac_srb))
678 _NOTE(SCHEME_PROTECTS_DATA("unique to sync fib and cdb", scsi_inquiry))
679 _NOTE(SCHEME_PROTECTS_DATA("stable data", scsi_device scsi_address))
680 _NOTE(SCHEME_PROTECTS_DATA("unique to dr event", aac_drinfo))
681 _NOTE(SCHEME_PROTECTS_DATA("unique to scsi_transport", buf))
682 
683 int
684 _init(void)
685 {
686 	int rval = 0;
687 
688 #ifdef DEBUG
689 	mutex_init(&aac_prt_mutex, NULL, MUTEX_DRIVER, NULL);
690 #endif
691 	DBCALLED(NULL, 1);
692 
693 	if ((rval = ddi_soft_state_init((void *)&aac_softstatep,
694 	    sizeof (struct aac_softstate), 0)) != 0)
695 		goto error;
696 
697 	if ((rval = scsi_hba_init(&aac_modlinkage)) != 0) {
698 		ddi_soft_state_fini((void *)&aac_softstatep);
699 		goto error;
700 	}
701 
702 	if ((rval = mod_install(&aac_modlinkage)) != 0) {
703 		ddi_soft_state_fini((void *)&aac_softstatep);
704 		scsi_hba_fini(&aac_modlinkage);
705 		goto error;
706 	}
707 	return (rval);
708 
709 error:
710 	AACDB_PRINT(NULL, CE_WARN, "Mod init error!");
711 #ifdef DEBUG
712 	mutex_destroy(&aac_prt_mutex);
713 #endif
714 	return (rval);
715 }
716 
717 int
718 _info(struct modinfo *modinfop)
719 {
720 	DBCALLED(NULL, 1);
721 	return (mod_info(&aac_modlinkage, modinfop));
722 }
723 
724 /*
725  * An HBA driver cannot be unload unless you reboot,
726  * so this function will be of no use.
727  */
728 int
729 _fini(void)
730 {
731 	int rval;
732 
733 	DBCALLED(NULL, 1);
734 
735 	if ((rval = mod_remove(&aac_modlinkage)) != 0)
736 		goto error;
737 
738 	scsi_hba_fini(&aac_modlinkage);
739 	ddi_soft_state_fini((void *)&aac_softstatep);
740 #ifdef DEBUG
741 	mutex_destroy(&aac_prt_mutex);
742 #endif
743 	return (0);
744 
745 error:
746 	AACDB_PRINT(NULL, CE_WARN, "AAC is busy, cannot unload!");
747 	return (rval);
748 }
749 
750 static int
751 aac_attach(dev_info_t *dip, ddi_attach_cmd_t cmd)
752 {
753 	int instance, i;
754 	struct aac_softstate *softs = NULL;
755 	int attach_state = 0;
756 	char *data;
757 	int intr_types;
758 
759 	DBCALLED(NULL, 1);
760 
761 	switch (cmd) {
762 	case DDI_ATTACH:
763 		break;
764 	case DDI_RESUME:
765 		return (DDI_FAILURE);
766 	default:
767 		return (DDI_FAILURE);
768 	}
769 
770 	instance = ddi_get_instance(dip);
771 
772 	/* Get soft state */
773 	if (ddi_soft_state_zalloc(aac_softstatep, instance) != DDI_SUCCESS) {
774 		AACDB_PRINT(softs, CE_WARN, "Cannot alloc soft state");
775 		goto error;
776 	}
777 	softs = ddi_get_soft_state(aac_softstatep, instance);
778 	attach_state |= AAC_ATTACH_SOFTSTATE_ALLOCED;
779 
780 	softs->instance = instance;
781 	softs->devinfo_p = dip;
782 	softs->buf_dma_attr = softs->addr_dma_attr = aac_dma_attr;
783 	softs->addr_dma_attr.dma_attr_granular = 1;
784 	softs->acc_attr = aac_acc_attr;
785 	softs->card = AAC_UNKNOWN_CARD;
786 #ifdef DEBUG
787 	softs->debug_flags = aac_debug_flags;
788 	softs->debug_fib_flags = aac_debug_fib_flags;
789 #endif
790 
791 	/* Initialize FMA */
792 	aac_fm_init(softs);
793 
794 	/* Check the card type */
795 	if (aac_check_card_type(softs) == AACERR) {
796 		AACDB_PRINT(softs, CE_WARN, "Card not supported");
797 		goto error;
798 	}
799 	/* We have found the right card and everything is OK */
800 	attach_state |= AAC_ATTACH_CARD_DETECTED;
801 
802 	/* Map PCI mem space */
803 	if (ddi_regs_map_setup(dip, 1,
804 	    (caddr_t *)&softs->pci_mem_base_vaddr, 0,
805 	    softs->map_size_min, &softs->acc_attr,
806 	    &softs->pci_mem_handle) != DDI_SUCCESS)
807 		goto error;
808 
809 	softs->map_size = softs->map_size_min;
810 	attach_state |= AAC_ATTACH_PCI_MEM_MAPPED;
811 
812 	AAC_DISABLE_INTR(softs);
813 
814 	/* Get the type of device intrrupts */
815 	if (ddi_intr_get_supported_types(dip, &intr_types) != DDI_SUCCESS) {
816 		AACDB_PRINT(softs, CE_WARN,
817 		    "ddi_intr_get_supported_types() failed");
818 		goto error;
819 	}
820 	AACDB_PRINT(softs, CE_NOTE,
821 	    "ddi_intr_get_supported_types() ret: 0x%x", intr_types);
822 
823 	/* Query interrupt, and alloc/init all needed struct */
824 	if (intr_types & DDI_INTR_TYPE_MSI) {
825 		if (aac_query_intrs(softs, DDI_INTR_TYPE_MSI)
826 		    != DDI_SUCCESS) {
827 			AACDB_PRINT(softs, CE_WARN,
828 			    "MSI interrupt query failed");
829 			goto error;
830 		}
831 		softs->intr_type = DDI_INTR_TYPE_MSI;
832 	} else if (intr_types & DDI_INTR_TYPE_FIXED) {
833 		if (aac_query_intrs(softs, DDI_INTR_TYPE_FIXED)
834 		    != DDI_SUCCESS) {
835 			AACDB_PRINT(softs, CE_WARN,
836 			    "FIXED interrupt query failed");
837 			goto error;
838 		}
839 		softs->intr_type = DDI_INTR_TYPE_FIXED;
840 	} else {
841 		AACDB_PRINT(softs, CE_WARN,
842 		    "Device cannot suppport both FIXED and MSI interrupts");
843 		goto error;
844 	}
845 
846 	/* Init mutexes */
847 	mutex_init(&softs->q_comp_mutex, NULL,
848 	    MUTEX_DRIVER, DDI_INTR_PRI(softs->intr_pri));
849 	cv_init(&softs->event, NULL, CV_DRIVER, NULL);
850 	mutex_init(&softs->aifq_mutex, NULL,
851 	    MUTEX_DRIVER, DDI_INTR_PRI(softs->intr_pri));
852 	cv_init(&softs->aifv, NULL, CV_DRIVER, NULL);
853 	cv_init(&softs->drain_cv, NULL, CV_DRIVER, NULL);
854 	mutex_init(&softs->io_lock, NULL, MUTEX_DRIVER,
855 	    DDI_INTR_PRI(softs->intr_pri));
856 	attach_state |= AAC_ATTACH_KMUTEX_INITED;
857 
858 	/* Check for legacy device naming support */
859 	softs->legacy = 1; /* default to use legacy name */
860 	if ((ddi_prop_lookup_string(DDI_DEV_T_ANY, dip, 0,
861 	    "legacy-name-enable", &data) == DDI_SUCCESS)) {
862 		if (strcmp(data, "no") == 0) {
863 			AACDB_PRINT(softs, CE_NOTE, "legacy-name disabled");
864 			softs->legacy = 0;
865 		}
866 		ddi_prop_free(data);
867 	}
868 
869 	/*
870 	 * Everything has been set up till now,
871 	 * we will do some common attach.
872 	 */
873 	if (aac_common_attach(softs) == AACERR)
874 		goto error;
875 	attach_state |= AAC_ATTACH_COMM_SPACE_SETUP;
876 
877 	/* Check for buf breakup support */
878 	if ((ddi_prop_lookup_string(DDI_DEV_T_ANY, dip, 0,
879 	    "breakup-enable", &data) == DDI_SUCCESS)) {
880 		if (strcmp(data, "yes") == 0) {
881 			AACDB_PRINT(softs, CE_NOTE, "buf breakup enabled");
882 			softs->flags |= AAC_FLAGS_BRKUP;
883 		}
884 		ddi_prop_free(data);
885 	}
886 	softs->dma_max = softs->buf_dma_attr.dma_attr_maxxfer;
887 	if (softs->flags & AAC_FLAGS_BRKUP) {
888 		softs->dma_max = ddi_prop_get_int(DDI_DEV_T_ANY, dip,
889 		    DDI_PROP_DONTPASS, "dma-max", softs->dma_max);
890 	}
891 
892 	/* Init the cmd queues */
893 	for (i = 0; i < AAC_CMDQ_NUM; i++)
894 		aac_cmd_initq(&softs->q_wait[i]);
895 	aac_cmd_initq(&softs->q_busy);
896 	aac_cmd_initq(&softs->q_comp);
897 
898 	if (aac_hba_setup(softs) != AACOK)
899 		goto error;
900 	attach_state |= AAC_ATTACH_SCSI_TRAN_SETUP;
901 
902 	/* Connect interrupt handlers */
903 	if (ddi_add_softintr(dip, DDI_SOFTINT_LOW, &softs->softint_id,
904 	    NULL, NULL, aac_softintr, (caddr_t)softs) != DDI_SUCCESS) {
905 		AACDB_PRINT(softs, CE_WARN,
906 		    "Can not setup soft interrupt handler!");
907 		goto error;
908 	}
909 	attach_state |= AAC_ATTACH_SOFT_INTR_SETUP;
910 
911 	if (aac_add_intrs(softs) != DDI_SUCCESS) {
912 		AACDB_PRINT(softs, CE_WARN,
913 		    "Interrupt registration failed, intr type: %s",
914 		    softs->intr_type == DDI_INTR_TYPE_MSI ? "MSI" : "FIXED");
915 		goto error;
916 	}
917 	attach_state |= AAC_ATTACH_HARD_INTR_SETUP;
918 
919 	/* Create devctl/scsi nodes for cfgadm */
920 	if (ddi_create_minor_node(dip, "devctl", S_IFCHR,
921 	    INST2DEVCTL(instance), DDI_NT_SCSI_NEXUS, 0) != DDI_SUCCESS) {
922 		AACDB_PRINT(softs, CE_WARN, "failed to create devctl node");
923 		goto error;
924 	}
925 	attach_state |= AAC_ATTACH_CREATE_DEVCTL;
926 
927 	if (ddi_create_minor_node(dip, "scsi", S_IFCHR, INST2SCSI(instance),
928 	    DDI_NT_SCSI_ATTACHMENT_POINT, 0) != DDI_SUCCESS) {
929 		AACDB_PRINT(softs, CE_WARN, "failed to create scsi node");
930 		goto error;
931 	}
932 	attach_state |= AAC_ATTACH_CREATE_SCSI;
933 
934 	/* Create aac node for app. to issue ioctls */
935 	if (ddi_create_minor_node(dip, "aac", S_IFCHR, INST2AAC(instance),
936 	    DDI_PSEUDO, 0) != DDI_SUCCESS) {
937 		AACDB_PRINT(softs, CE_WARN, "failed to create aac node");
938 		goto error;
939 	}
940 
941 	/* Create a taskq for dealing with dr events */
942 	if ((softs->taskq = ddi_taskq_create(dip, "aac_dr_taskq", 1,
943 	    TASKQ_DEFAULTPRI, 0)) == NULL) {
944 		AACDB_PRINT(softs, CE_WARN, "ddi_taskq_create failed");
945 		goto error;
946 	}
947 
948 	aac_unhold_bus(softs, AAC_IOCMD_SYNC | AAC_IOCMD_ASYNC);
949 	softs->state = AAC_STATE_RUN;
950 
951 	/* Create a thread for command timeout */
952 	softs->timeout_id = timeout(aac_daemon, (void *)softs,
953 	    (60 * drv_usectohz(1000000)));
954 
955 	/* Common attach is OK, so we are attached! */
956 	AAC_ENABLE_INTR(softs);
957 	ddi_report_dev(dip);
958 	AACDB_PRINT(softs, CE_NOTE, "aac attached ok");
959 	return (DDI_SUCCESS);
960 
961 error:
962 	if (softs && softs->taskq)
963 		ddi_taskq_destroy(softs->taskq);
964 	if (attach_state & AAC_ATTACH_CREATE_SCSI)
965 		ddi_remove_minor_node(dip, "scsi");
966 	if (attach_state & AAC_ATTACH_CREATE_DEVCTL)
967 		ddi_remove_minor_node(dip, "devctl");
968 	if (attach_state & AAC_ATTACH_COMM_SPACE_SETUP)
969 		aac_common_detach(softs);
970 	if (attach_state & AAC_ATTACH_SCSI_TRAN_SETUP) {
971 		(void) scsi_hba_detach(dip);
972 		scsi_hba_tran_free(AAC_DIP2TRAN(dip));
973 	}
974 	if (attach_state & AAC_ATTACH_HARD_INTR_SETUP)
975 		aac_remove_intrs(softs);
976 	if (attach_state & AAC_ATTACH_SOFT_INTR_SETUP)
977 		ddi_remove_softintr(softs->softint_id);
978 	if (attach_state & AAC_ATTACH_KMUTEX_INITED) {
979 		mutex_destroy(&softs->q_comp_mutex);
980 		cv_destroy(&softs->event);
981 		mutex_destroy(&softs->aifq_mutex);
982 		cv_destroy(&softs->aifv);
983 		cv_destroy(&softs->drain_cv);
984 		mutex_destroy(&softs->io_lock);
985 	}
986 	if (attach_state & AAC_ATTACH_PCI_MEM_MAPPED)
987 		ddi_regs_map_free(&softs->pci_mem_handle);
988 	aac_fm_fini(softs);
989 	if (attach_state & AAC_ATTACH_CARD_DETECTED)
990 		softs->card = AACERR;
991 	if (attach_state & AAC_ATTACH_SOFTSTATE_ALLOCED)
992 		ddi_soft_state_free(aac_softstatep, instance);
993 	return (DDI_FAILURE);
994 }
995 
996 static int
997 aac_detach(dev_info_t *dip, ddi_detach_cmd_t cmd)
998 {
999 	scsi_hba_tran_t *tran = AAC_DIP2TRAN(dip);
1000 	struct aac_softstate *softs = AAC_TRAN2SOFTS(tran);
1001 
1002 	DBCALLED(softs, 1);
1003 
1004 	switch (cmd) {
1005 	case DDI_DETACH:
1006 		break;
1007 	case DDI_SUSPEND:
1008 		return (DDI_FAILURE);
1009 	default:
1010 		return (DDI_FAILURE);
1011 	}
1012 
1013 	mutex_enter(&softs->io_lock);
1014 	AAC_DISABLE_INTR(softs);
1015 	softs->state = AAC_STATE_STOPPED;
1016 
1017 	mutex_exit(&softs->io_lock);
1018 	(void) untimeout(softs->timeout_id);
1019 	mutex_enter(&softs->io_lock);
1020 	softs->timeout_id = 0;
1021 
1022 	ddi_taskq_destroy(softs->taskq);
1023 
1024 	ddi_remove_minor_node(dip, "aac");
1025 	ddi_remove_minor_node(dip, "scsi");
1026 	ddi_remove_minor_node(dip, "devctl");
1027 
1028 	mutex_exit(&softs->io_lock);
1029 	aac_remove_intrs(softs);
1030 	ddi_remove_softintr(softs->softint_id);
1031 
1032 	aac_common_detach(softs);
1033 
1034 	(void) scsi_hba_detach(dip);
1035 	scsi_hba_tran_free(tran);
1036 
1037 	mutex_destroy(&softs->q_comp_mutex);
1038 	cv_destroy(&softs->event);
1039 	mutex_destroy(&softs->aifq_mutex);
1040 	cv_destroy(&softs->aifv);
1041 	cv_destroy(&softs->drain_cv);
1042 	mutex_destroy(&softs->io_lock);
1043 
1044 	ddi_regs_map_free(&softs->pci_mem_handle);
1045 	aac_fm_fini(softs);
1046 	softs->hwif = AAC_HWIF_UNKNOWN;
1047 	softs->card = AAC_UNKNOWN_CARD;
1048 	ddi_soft_state_free(aac_softstatep, ddi_get_instance(dip));
1049 
1050 	return (DDI_SUCCESS);
1051 }
1052 
1053 /*ARGSUSED*/
1054 static int
1055 aac_reset(dev_info_t *dip, ddi_reset_cmd_t cmd)
1056 {
1057 	struct aac_softstate *softs = AAC_DIP2SOFTS(dip);
1058 
1059 	DBCALLED(softs, 1);
1060 
1061 	mutex_enter(&softs->io_lock);
1062 	(void) aac_shutdown(softs);
1063 	mutex_exit(&softs->io_lock);
1064 
1065 	return (DDI_SUCCESS);
1066 }
1067 
1068 /*
1069  * quiesce(9E) entry point.
1070  *
1071  * This function is called when the system is single-threaded at high
1072  * PIL with preemption disabled. Therefore, this function must not be
1073  * blocked.
1074  *
1075  * This function returns DDI_SUCCESS on success, or DDI_FAILURE on failure.
1076  * DDI_FAILURE indicates an error condition and should almost never happen.
1077  */
1078 static int
1079 aac_quiesce(dev_info_t *dip)
1080 {
1081 	struct aac_softstate *softs = AAC_DIP2SOFTS(dip);
1082 
1083 	if (softs == NULL)
1084 		return (DDI_FAILURE);
1085 
1086 	AAC_DISABLE_INTR(softs);
1087 
1088 	return (DDI_SUCCESS);
1089 }
1090 
1091 /*
1092  * Bring the controller down to a dormant state and detach all child devices.
1093  * This function is called before detach or system shutdown.
1094  * Note: we can assume that the q_wait on the controller is empty, as we
1095  * won't allow shutdown if any device is open.
1096  */
1097 static int
1098 aac_shutdown(struct aac_softstate *softs)
1099 {
1100 	ddi_acc_handle_t acc = softs->sync_slot.fib_acc_handle;
1101 	struct aac_close_command *cc = (struct aac_close_command *) \
1102 	    &softs->sync_slot.fibp->data[0];
1103 	int rval;
1104 
1105 	ddi_put32(acc, &cc->Command, VM_CloseAll);
1106 	ddi_put32(acc, &cc->ContainerId, 0xfffffffful);
1107 
1108 	/* Flush all caches, set FW to write through mode */
1109 	rval = aac_sync_fib(softs, ContainerCommand,
1110 	    AAC_FIB_SIZEOF(struct aac_close_command));
1111 
1112 	AACDB_PRINT(softs, CE_NOTE,
1113 	    "shutting down aac %s", (rval == AACOK) ? "ok" : "fail");
1114 	return (rval);
1115 }
1116 
1117 static uint_t
1118 aac_softintr(caddr_t arg)
1119 {
1120 	struct aac_softstate *softs = (void *)arg;
1121 
1122 	if (!AAC_IS_Q_EMPTY(&softs->q_comp)) {
1123 		aac_drain_comp_q(softs);
1124 		return (DDI_INTR_CLAIMED);
1125 	} else {
1126 		return (DDI_INTR_UNCLAIMED);
1127 	}
1128 }
1129 
1130 /*
1131  * Setup auto sense data for pkt
1132  */
1133 static void
1134 aac_set_arq_data(struct scsi_pkt *pkt, uchar_t key,
1135     uchar_t add_code, uchar_t qual_code, uint64_t info)
1136 {
1137 	struct scsi_arq_status *arqstat = (void *)(pkt->pkt_scbp);
1138 
1139 	*pkt->pkt_scbp = STATUS_CHECK; /* CHECK CONDITION */
1140 	pkt->pkt_state |= STATE_ARQ_DONE;
1141 
1142 	*(uint8_t *)&arqstat->sts_rqpkt_status = STATUS_GOOD;
1143 	arqstat->sts_rqpkt_reason = CMD_CMPLT;
1144 	arqstat->sts_rqpkt_resid = 0;
1145 	arqstat->sts_rqpkt_state =
1146 	    STATE_GOT_BUS |
1147 	    STATE_GOT_TARGET |
1148 	    STATE_SENT_CMD |
1149 	    STATE_XFERRED_DATA;
1150 	arqstat->sts_rqpkt_statistics = 0;
1151 
1152 	if (info <= 0xfffffffful) {
1153 		arqstat->sts_sensedata.es_valid = 1;
1154 		arqstat->sts_sensedata.es_class = CLASS_EXTENDED_SENSE;
1155 		arqstat->sts_sensedata.es_code = CODE_FMT_FIXED_CURRENT;
1156 		arqstat->sts_sensedata.es_key = key;
1157 		arqstat->sts_sensedata.es_add_code = add_code;
1158 		arqstat->sts_sensedata.es_qual_code = qual_code;
1159 
1160 		arqstat->sts_sensedata.es_info_1 = (info >> 24) & 0xFF;
1161 		arqstat->sts_sensedata.es_info_2 = (info >> 16) & 0xFF;
1162 		arqstat->sts_sensedata.es_info_3 = (info >>  8) & 0xFF;
1163 		arqstat->sts_sensedata.es_info_4 = info & 0xFF;
1164 	} else { /* 64-bit LBA */
1165 		struct scsi_descr_sense_hdr *dsp;
1166 		struct scsi_information_sense_descr *isd;
1167 
1168 		dsp = (struct scsi_descr_sense_hdr *)&arqstat->sts_sensedata;
1169 		dsp->ds_class = CLASS_EXTENDED_SENSE;
1170 		dsp->ds_code = CODE_FMT_DESCR_CURRENT;
1171 		dsp->ds_key = key;
1172 		dsp->ds_add_code = add_code;
1173 		dsp->ds_qual_code = qual_code;
1174 		dsp->ds_addl_sense_length =
1175 		    sizeof (struct scsi_information_sense_descr);
1176 
1177 		isd = (struct scsi_information_sense_descr *)(dsp+1);
1178 		isd->isd_descr_type = DESCR_INFORMATION;
1179 		isd->isd_valid = 1;
1180 		isd->isd_information[0] = (info >> 56) & 0xFF;
1181 		isd->isd_information[1] = (info >> 48) & 0xFF;
1182 		isd->isd_information[2] = (info >> 40) & 0xFF;
1183 		isd->isd_information[3] = (info >> 32) & 0xFF;
1184 		isd->isd_information[4] = (info >> 24) & 0xFF;
1185 		isd->isd_information[5] = (info >> 16) & 0xFF;
1186 		isd->isd_information[6] = (info >>  8) & 0xFF;
1187 		isd->isd_information[7] = (info) & 0xFF;
1188 	}
1189 }
1190 
1191 /*
1192  * Setup auto sense data for HARDWARE ERROR
1193  */
1194 static void
1195 aac_set_arq_data_hwerr(struct aac_cmd *acp)
1196 {
1197 	union scsi_cdb *cdbp;
1198 	uint64_t err_blkno;
1199 
1200 	cdbp = (void *)acp->pkt->pkt_cdbp;
1201 	err_blkno = AAC_GETGXADDR(acp->cmdlen, cdbp);
1202 	aac_set_arq_data(acp->pkt, KEY_HARDWARE_ERROR, 0x00, 0x00, err_blkno);
1203 }
1204 
1205 /*
1206  * Setup auto sense data for UNIT ATTENTION
1207  */
1208 /*ARGSUSED*/
1209 static void
1210 aac_set_arq_data_reset(struct aac_softstate *softs, struct aac_cmd *acp)
1211 {
1212 	struct aac_container *dvp = (struct aac_container *)acp->dvp;
1213 
1214 	ASSERT(dvp->dev.type == AAC_DEV_LD);
1215 
1216 	if (dvp->reset) {
1217 		dvp->reset = 0;
1218 		aac_set_arq_data(acp->pkt, KEY_UNIT_ATTENTION, 0x29, 0x02, 0);
1219 	}
1220 }
1221 
1222 /*
1223  * Send a command to the adapter in New Comm. interface
1224  */
1225 static int
1226 aac_send_command(struct aac_softstate *softs, struct aac_slot *slotp)
1227 {
1228 	uint32_t index, device;
1229 
1230 	index = PCI_MEM_GET32(softs, AAC_IQUE);
1231 	if (index == 0xffffffffUL) {
1232 		index = PCI_MEM_GET32(softs, AAC_IQUE);
1233 		if (index == 0xffffffffUL)
1234 			return (AACERR);
1235 	}
1236 
1237 	device = index;
1238 	PCI_MEM_PUT32(softs, device,
1239 	    (uint32_t)(slotp->fib_phyaddr & 0xfffffffful));
1240 	device += 4;
1241 	PCI_MEM_PUT32(softs, device, (uint32_t)(slotp->fib_phyaddr >> 32));
1242 	device += 4;
1243 	PCI_MEM_PUT32(softs, device, slotp->acp->fib_size);
1244 	PCI_MEM_PUT32(softs, AAC_IQUE, index);
1245 	return (AACOK);
1246 }
1247 
1248 static void
1249 aac_end_io(struct aac_softstate *softs, struct aac_cmd *acp)
1250 {
1251 	struct aac_device *dvp = acp->dvp;
1252 	int q = AAC_CMDQ(acp);
1253 
1254 	if (acp->slotp) { /* outstanding cmd */
1255 		aac_release_slot(softs, acp->slotp);
1256 		acp->slotp = NULL;
1257 		if (dvp) {
1258 			dvp->ncmds[q]--;
1259 			if (dvp->throttle[q] == AAC_THROTTLE_DRAIN &&
1260 			    dvp->ncmds[q] == 0 && q == AAC_CMDQ_ASYNC)
1261 				aac_set_throttle(softs, dvp, q,
1262 				    softs->total_slots);
1263 		}
1264 		softs->bus_ncmds[q]--;
1265 		(void) aac_cmd_delete(&softs->q_busy, acp);
1266 	} else { /* cmd in waiting queue */
1267 		aac_cmd_delete(&softs->q_wait[q], acp);
1268 	}
1269 
1270 	if (!(acp->flags & (AAC_CMD_NO_CB | AAC_CMD_NO_INTR))) { /* async IO */
1271 		mutex_enter(&softs->q_comp_mutex);
1272 		aac_cmd_enqueue(&softs->q_comp, acp);
1273 		mutex_exit(&softs->q_comp_mutex);
1274 	} else if (acp->flags & AAC_CMD_NO_CB) { /* sync IO */
1275 		cv_broadcast(&softs->event);
1276 	}
1277 }
1278 
1279 static void
1280 aac_handle_io(struct aac_softstate *softs, int index)
1281 {
1282 	struct aac_slot *slotp;
1283 	struct aac_cmd *acp;
1284 	uint32_t fast;
1285 
1286 	fast = index & AAC_SENDERADDR_MASK_FAST_RESPONSE;
1287 	index >>= 2;
1288 
1289 	/* Make sure firmware reported index is valid */
1290 	ASSERT(index >= 0 && index < softs->total_slots);
1291 	slotp = &softs->io_slot[index];
1292 	ASSERT(slotp->index == index);
1293 	acp = slotp->acp;
1294 
1295 	if (acp == NULL || acp->slotp != slotp) {
1296 		cmn_err(CE_WARN,
1297 		    "Firmware error: invalid slot index received from FW");
1298 		return;
1299 	}
1300 
1301 	acp->flags |= AAC_CMD_CMPLT;
1302 	(void) ddi_dma_sync(slotp->fib_dma_handle, 0, 0, DDI_DMA_SYNC_FORCPU);
1303 
1304 	if (aac_check_dma_handle(slotp->fib_dma_handle) == DDI_SUCCESS) {
1305 		/*
1306 		 * For fast response IO, the firmware do not return any FIB
1307 		 * data, so we need to fill in the FIB status and state so that
1308 		 * FIB users can handle it correctly.
1309 		 */
1310 		if (fast) {
1311 			uint32_t state;
1312 
1313 			state = ddi_get32(slotp->fib_acc_handle,
1314 			    &slotp->fibp->Header.XferState);
1315 			/*
1316 			 * Update state for CPU not for device, no DMA sync
1317 			 * needed
1318 			 */
1319 			ddi_put32(slotp->fib_acc_handle,
1320 			    &slotp->fibp->Header.XferState,
1321 			    state | AAC_FIBSTATE_DONEADAP);
1322 			ddi_put32(slotp->fib_acc_handle,
1323 			    (void *)&slotp->fibp->data[0], ST_OK);
1324 		}
1325 
1326 		/* Handle completed ac */
1327 		acp->ac_comp(softs, acp);
1328 	} else {
1329 		ddi_fm_service_impact(softs->devinfo_p, DDI_SERVICE_UNAFFECTED);
1330 		acp->flags |= AAC_CMD_ERR;
1331 		if (acp->pkt) {
1332 			acp->pkt->pkt_reason = CMD_TRAN_ERR;
1333 			acp->pkt->pkt_statistics = 0;
1334 		}
1335 	}
1336 	aac_end_io(softs, acp);
1337 }
1338 
1339 /*
1340  * Interrupt handler for New Comm. interface
1341  * New Comm. interface use a different mechanism for interrupt. No explict
1342  * message queues, and driver need only accesses the mapped PCI mem space to
1343  * find the completed FIB or AIF.
1344  */
1345 static int
1346 aac_process_intr_new(struct aac_softstate *softs)
1347 {
1348 	uint32_t index;
1349 
1350 	index = AAC_OUTB_GET(softs);
1351 	if (index == 0xfffffffful)
1352 		index = AAC_OUTB_GET(softs);
1353 	if (aac_check_acc_handle(softs->pci_mem_handle) != DDI_SUCCESS) {
1354 		ddi_fm_service_impact(softs->devinfo_p, DDI_SERVICE_UNAFFECTED);
1355 		return (0);
1356 	}
1357 	if (index != 0xfffffffful) {
1358 		do {
1359 			if ((index & AAC_SENDERADDR_MASK_AIF) == 0) {
1360 				aac_handle_io(softs, index);
1361 			} else if (index != 0xfffffffeul) {
1362 				struct aac_fib *fibp;	/* FIB in AIF queue */
1363 				uint16_t fib_size, fib_size0;
1364 
1365 				/*
1366 				 * 0xfffffffe means that the controller wants
1367 				 * more work, ignore it for now. Otherwise,
1368 				 * AIF received.
1369 				 */
1370 				index &= ~2;
1371 
1372 				mutex_enter(&softs->aifq_mutex);
1373 				/*
1374 				 * Copy AIF from adapter to the empty AIF slot
1375 				 */
1376 				fibp = &softs->aifq[softs->aifq_idx].d;
1377 				fib_size0 = PCI_MEM_GET16(softs, index + \
1378 				    offsetof(struct aac_fib, Header.Size));
1379 				fib_size = (fib_size0 > AAC_FIB_SIZE) ?
1380 				    AAC_FIB_SIZE : fib_size0;
1381 				PCI_MEM_REP_GET8(softs, index, fibp,
1382 				    fib_size);
1383 
1384 				if (aac_check_acc_handle(softs-> \
1385 				    pci_mem_handle) == DDI_SUCCESS)
1386 					(void) aac_handle_aif(softs, fibp);
1387 				else
1388 					ddi_fm_service_impact(softs->devinfo_p,
1389 					    DDI_SERVICE_UNAFFECTED);
1390 				mutex_exit(&softs->aifq_mutex);
1391 
1392 				/*
1393 				 * AIF memory is owned by the adapter, so let it
1394 				 * know that we are done with it.
1395 				 */
1396 				AAC_OUTB_SET(softs, index);
1397 				AAC_STATUS_CLR(softs, AAC_DB_RESPONSE_READY);
1398 			}
1399 
1400 			index = AAC_OUTB_GET(softs);
1401 		} while (index != 0xfffffffful);
1402 
1403 		/*
1404 		 * Process waiting cmds before start new ones to
1405 		 * ensure first IOs are serviced first.
1406 		 */
1407 		aac_start_waiting_io(softs);
1408 		return (AAC_DB_COMMAND_READY);
1409 	} else {
1410 		return (0);
1411 	}
1412 }
1413 
1414 static uint_t
1415 aac_intr_new(caddr_t arg)
1416 {
1417 	struct aac_softstate *softs = (void *)arg;
1418 	uint_t rval;
1419 
1420 	mutex_enter(&softs->io_lock);
1421 	if (aac_process_intr_new(softs))
1422 		rval = DDI_INTR_CLAIMED;
1423 	else
1424 		rval = DDI_INTR_UNCLAIMED;
1425 	mutex_exit(&softs->io_lock);
1426 
1427 	aac_drain_comp_q(softs);
1428 	return (rval);
1429 }
1430 
1431 /*
1432  * Interrupt handler for old interface
1433  * Explicit message queues are used to send FIB to and get completed FIB from
1434  * the adapter. Driver and adapter maitain the queues in the producer/consumer
1435  * manner. The driver has to query the queues to find the completed FIB.
1436  */
1437 static int
1438 aac_process_intr_old(struct aac_softstate *softs)
1439 {
1440 	uint16_t status;
1441 
1442 	status = AAC_STATUS_GET(softs);
1443 	if (aac_check_acc_handle(softs->pci_mem_handle) != DDI_SUCCESS) {
1444 		ddi_fm_service_impact(softs->devinfo_p, DDI_SERVICE_UNAFFECTED);
1445 		return (DDI_INTR_UNCLAIMED);
1446 	}
1447 	if (status & AAC_DB_RESPONSE_READY) {
1448 		int slot_idx;
1449 
1450 		/* ACK the intr */
1451 		AAC_STATUS_CLR(softs, AAC_DB_RESPONSE_READY);
1452 		(void) AAC_STATUS_GET(softs);
1453 		while (aac_fib_dequeue(softs, AAC_HOST_NORM_RESP_Q,
1454 		    &slot_idx) == AACOK)
1455 			aac_handle_io(softs, slot_idx);
1456 
1457 		/*
1458 		 * Process waiting cmds before start new ones to
1459 		 * ensure first IOs are serviced first.
1460 		 */
1461 		aac_start_waiting_io(softs);
1462 		return (AAC_DB_RESPONSE_READY);
1463 	} else if (status & AAC_DB_COMMAND_READY) {
1464 		int aif_idx;
1465 
1466 		AAC_STATUS_CLR(softs, AAC_DB_COMMAND_READY);
1467 		(void) AAC_STATUS_GET(softs);
1468 		if (aac_fib_dequeue(softs, AAC_HOST_NORM_CMD_Q, &aif_idx) ==
1469 		    AACOK) {
1470 			ddi_acc_handle_t acc = softs->comm_space_acc_handle;
1471 			struct aac_fib *fibp;	/* FIB in AIF queue */
1472 			struct aac_fib *fibp0;	/* FIB in communication space */
1473 			uint16_t fib_size, fib_size0;
1474 			uint32_t fib_xfer_state;
1475 			uint32_t addr, size;
1476 
1477 			ASSERT((aif_idx >= 0) && (aif_idx < AAC_ADAPTER_FIBS));
1478 
1479 #define	AAC_SYNC_AIF(softs, aif_idx, type) \
1480 	{ (void) ddi_dma_sync((softs)->comm_space_dma_handle, \
1481 	    offsetof(struct aac_comm_space, \
1482 	    adapter_fibs[(aif_idx)]), AAC_FIB_SIZE, \
1483 	    (type)); }
1484 
1485 			mutex_enter(&softs->aifq_mutex);
1486 			/* Copy AIF from adapter to the empty AIF slot */
1487 			fibp = &softs->aifq[softs->aifq_idx].d;
1488 			AAC_SYNC_AIF(softs, aif_idx, DDI_DMA_SYNC_FORCPU);
1489 			fibp0 = &softs->comm_space->adapter_fibs[aif_idx];
1490 			fib_size0 = ddi_get16(acc, &fibp0->Header.Size);
1491 			fib_size = (fib_size0 > AAC_FIB_SIZE) ?
1492 			    AAC_FIB_SIZE : fib_size0;
1493 			ddi_rep_get8(acc, (uint8_t *)fibp, (uint8_t *)fibp0,
1494 			    fib_size, DDI_DEV_AUTOINCR);
1495 
1496 			(void) aac_handle_aif(softs, fibp);
1497 			mutex_exit(&softs->aifq_mutex);
1498 
1499 			/* Complete AIF back to adapter with good status */
1500 			fib_xfer_state = LE_32(fibp->Header.XferState);
1501 			if (fib_xfer_state & AAC_FIBSTATE_FROMADAP) {
1502 				ddi_put32(acc, &fibp0->Header.XferState,
1503 				    fib_xfer_state | AAC_FIBSTATE_DONEHOST);
1504 				ddi_put32(acc, (void *)&fibp0->data[0], ST_OK);
1505 				if (fib_size0 > AAC_FIB_SIZE)
1506 					ddi_put16(acc, &fibp0->Header.Size,
1507 					    AAC_FIB_SIZE);
1508 				AAC_SYNC_AIF(softs, aif_idx,
1509 				    DDI_DMA_SYNC_FORDEV);
1510 			}
1511 
1512 			/* Put the AIF response on the response queue */
1513 			addr = ddi_get32(acc,
1514 			    &softs->comm_space->adapter_fibs[aif_idx]. \
1515 			    Header.SenderFibAddress);
1516 			size = (uint32_t)ddi_get16(acc,
1517 			    &softs->comm_space->adapter_fibs[aif_idx]. \
1518 			    Header.Size);
1519 			ddi_put32(acc,
1520 			    &softs->comm_space->adapter_fibs[aif_idx]. \
1521 			    Header.ReceiverFibAddress, addr);
1522 			if (aac_fib_enqueue(softs, AAC_ADAP_NORM_RESP_Q,
1523 			    addr, size) == AACERR)
1524 				cmn_err(CE_NOTE, "!AIF ack failed");
1525 		}
1526 		return (AAC_DB_COMMAND_READY);
1527 	} else if (status & AAC_DB_PRINTF_READY) {
1528 		/* ACK the intr */
1529 		AAC_STATUS_CLR(softs, AAC_DB_PRINTF_READY);
1530 		(void) AAC_STATUS_GET(softs);
1531 		(void) ddi_dma_sync(softs->comm_space_dma_handle,
1532 		    offsetof(struct aac_comm_space, adapter_print_buf),
1533 		    AAC_ADAPTER_PRINT_BUFSIZE, DDI_DMA_SYNC_FORCPU);
1534 		if (aac_check_dma_handle(softs->comm_space_dma_handle) ==
1535 		    DDI_SUCCESS)
1536 			cmn_err(CE_NOTE, "MSG From Adapter: %s",
1537 			    softs->comm_space->adapter_print_buf);
1538 		else
1539 			ddi_fm_service_impact(softs->devinfo_p,
1540 			    DDI_SERVICE_UNAFFECTED);
1541 		AAC_NOTIFY(softs, AAC_DB_PRINTF_READY);
1542 		return (AAC_DB_PRINTF_READY);
1543 	} else if (status & AAC_DB_COMMAND_NOT_FULL) {
1544 		/*
1545 		 * Without these two condition statements, the OS could hang
1546 		 * after a while, especially if there are a lot of AIF's to
1547 		 * handle, for instance if a drive is pulled from an array
1548 		 * under heavy load.
1549 		 */
1550 		AAC_STATUS_CLR(softs, AAC_DB_COMMAND_NOT_FULL);
1551 		return (AAC_DB_COMMAND_NOT_FULL);
1552 	} else if (status & AAC_DB_RESPONSE_NOT_FULL) {
1553 		AAC_STATUS_CLR(softs, AAC_DB_COMMAND_NOT_FULL);
1554 		AAC_STATUS_CLR(softs, AAC_DB_RESPONSE_NOT_FULL);
1555 		return (AAC_DB_RESPONSE_NOT_FULL);
1556 	} else {
1557 		return (0);
1558 	}
1559 }
1560 
1561 static uint_t
1562 aac_intr_old(caddr_t arg)
1563 {
1564 	struct aac_softstate *softs = (void *)arg;
1565 	int rval;
1566 
1567 	mutex_enter(&softs->io_lock);
1568 	if (aac_process_intr_old(softs))
1569 		rval = DDI_INTR_CLAIMED;
1570 	else
1571 		rval = DDI_INTR_UNCLAIMED;
1572 	mutex_exit(&softs->io_lock);
1573 
1574 	aac_drain_comp_q(softs);
1575 	return (rval);
1576 }
1577 
1578 /*
1579  * Query FIXED or MSI interrupts
1580  */
1581 static int
1582 aac_query_intrs(struct aac_softstate *softs, int intr_type)
1583 {
1584 	dev_info_t *dip = softs->devinfo_p;
1585 	int avail, actual, intr_size, count;
1586 	int i, flag, ret;
1587 
1588 	AACDB_PRINT(softs, CE_NOTE,
1589 	    "aac_query_intrs:interrupt type 0x%x", intr_type);
1590 
1591 	/* Get number of interrupts */
1592 	ret = ddi_intr_get_nintrs(dip, intr_type, &count);
1593 	if ((ret != DDI_SUCCESS) || (count == 0)) {
1594 		AACDB_PRINT(softs, CE_WARN,
1595 		    "ddi_intr_get_nintrs() failed, ret %d count %d",
1596 		    ret, count);
1597 		return (DDI_FAILURE);
1598 	}
1599 
1600 	/* Get number of available interrupts */
1601 	ret = ddi_intr_get_navail(dip, intr_type, &avail);
1602 	if ((ret != DDI_SUCCESS) || (avail == 0)) {
1603 		AACDB_PRINT(softs, CE_WARN,
1604 		    "ddi_intr_get_navail() failed, ret %d avail %d",
1605 		    ret, avail);
1606 		return (DDI_FAILURE);
1607 	}
1608 
1609 	AACDB_PRINT(softs, CE_NOTE,
1610 	    "ddi_intr_get_nvail returned %d, navail() returned %d",
1611 	    count, avail);
1612 
1613 	/* Allocate an array of interrupt handles */
1614 	intr_size = count * sizeof (ddi_intr_handle_t);
1615 	softs->htable = kmem_alloc(intr_size, KM_SLEEP);
1616 
1617 	if (intr_type == DDI_INTR_TYPE_MSI) {
1618 		count = 1; /* only one vector needed by now */
1619 		flag = DDI_INTR_ALLOC_STRICT;
1620 	} else { /* must be DDI_INTR_TYPE_FIXED */
1621 		flag = DDI_INTR_ALLOC_NORMAL;
1622 	}
1623 
1624 	/* Call ddi_intr_alloc() */
1625 	ret = ddi_intr_alloc(dip, softs->htable, intr_type, 0,
1626 	    count, &actual, flag);
1627 
1628 	if ((ret != DDI_SUCCESS) || (actual == 0)) {
1629 		AACDB_PRINT(softs, CE_WARN,
1630 		    "ddi_intr_alloc() failed, ret = %d", ret);
1631 		actual = 0;
1632 		goto error;
1633 	}
1634 
1635 	if (actual < count) {
1636 		AACDB_PRINT(softs, CE_NOTE,
1637 		    "Requested: %d, Received: %d", count, actual);
1638 		goto error;
1639 	}
1640 
1641 	softs->intr_cnt = actual;
1642 
1643 	/* Get priority for first msi, assume remaining are all the same */
1644 	if ((ret = ddi_intr_get_pri(softs->htable[0],
1645 	    &softs->intr_pri)) != DDI_SUCCESS) {
1646 		AACDB_PRINT(softs, CE_WARN,
1647 		    "ddi_intr_get_pri() failed, ret = %d", ret);
1648 		goto error;
1649 	}
1650 
1651 	/* Test for high level mutex */
1652 	if (softs->intr_pri >= ddi_intr_get_hilevel_pri()) {
1653 		AACDB_PRINT(softs, CE_WARN,
1654 		    "aac_query_intrs: Hi level interrupt not supported");
1655 		goto error;
1656 	}
1657 
1658 	return (DDI_SUCCESS);
1659 
1660 error:
1661 	/* Free already allocated intr */
1662 	for (i = 0; i < actual; i++)
1663 		(void) ddi_intr_free(softs->htable[i]);
1664 
1665 	kmem_free(softs->htable, intr_size);
1666 	return (DDI_FAILURE);
1667 }
1668 
1669 /*
1670  * Register FIXED or MSI interrupts, and enable them
1671  */
1672 static int
1673 aac_add_intrs(struct aac_softstate *softs)
1674 {
1675 	int i, ret;
1676 	int intr_size, actual;
1677 	ddi_intr_handler_t *aac_intr;
1678 
1679 	actual = softs->intr_cnt;
1680 	intr_size = actual * sizeof (ddi_intr_handle_t);
1681 	aac_intr = (ddi_intr_handler_t *)((softs->flags & AAC_FLAGS_NEW_COMM) ?
1682 	    aac_intr_new : aac_intr_old);
1683 
1684 	/* Call ddi_intr_add_handler() */
1685 	for (i = 0; i < actual; i++) {
1686 		if ((ret = ddi_intr_add_handler(softs->htable[i],
1687 		    aac_intr, (caddr_t)softs, NULL)) != DDI_SUCCESS) {
1688 			cmn_err(CE_WARN,
1689 			    "ddi_intr_add_handler() failed ret = %d", ret);
1690 
1691 			/* Free already allocated intr */
1692 			for (i = 0; i < actual; i++)
1693 				(void) ddi_intr_free(softs->htable[i]);
1694 
1695 			kmem_free(softs->htable, intr_size);
1696 			return (DDI_FAILURE);
1697 		}
1698 	}
1699 
1700 	if ((ret = ddi_intr_get_cap(softs->htable[0], &softs->intr_cap))
1701 	    != DDI_SUCCESS) {
1702 		cmn_err(CE_WARN, "ddi_intr_get_cap() failed, ret = %d", ret);
1703 
1704 		/* Free already allocated intr */
1705 		for (i = 0; i < actual; i++)
1706 			(void) ddi_intr_free(softs->htable[i]);
1707 
1708 		kmem_free(softs->htable, intr_size);
1709 		return (DDI_FAILURE);
1710 	}
1711 
1712 	/* Enable interrupts */
1713 	if (softs->intr_cap & DDI_INTR_FLAG_BLOCK) {
1714 		/* for MSI block enable */
1715 		(void) ddi_intr_block_enable(softs->htable, softs->intr_cnt);
1716 	} else {
1717 		/* Call ddi_intr_enable() for legacy/MSI non block enable */
1718 		for (i = 0; i < softs->intr_cnt; i++)
1719 			(void) ddi_intr_enable(softs->htable[i]);
1720 	}
1721 
1722 	return (DDI_SUCCESS);
1723 }
1724 
1725 /*
1726  * Unregister FIXED or MSI interrupts
1727  */
1728 static void
1729 aac_remove_intrs(struct aac_softstate *softs)
1730 {
1731 	int i;
1732 
1733 	/* Disable all interrupts */
1734 	if (softs->intr_cap & DDI_INTR_FLAG_BLOCK) {
1735 		/* Call ddi_intr_block_disable() */
1736 		(void) ddi_intr_block_disable(softs->htable, softs->intr_cnt);
1737 	} else {
1738 		for (i = 0; i < softs->intr_cnt; i++)
1739 			(void) ddi_intr_disable(softs->htable[i]);
1740 	}
1741 
1742 	/* Call ddi_intr_remove_handler() */
1743 	for (i = 0; i < softs->intr_cnt; i++) {
1744 		(void) ddi_intr_remove_handler(softs->htable[i]);
1745 		(void) ddi_intr_free(softs->htable[i]);
1746 	}
1747 
1748 	kmem_free(softs->htable, softs->intr_cnt * sizeof (ddi_intr_handle_t));
1749 }
1750 
1751 /*
1752  * Set pkt_reason and OR in pkt_statistics flag
1753  */
1754 static void
1755 aac_set_pkt_reason(struct aac_softstate *softs, struct aac_cmd *acp,
1756     uchar_t reason, uint_t stat)
1757 {
1758 #ifndef __lock_lint
1759 	_NOTE(ARGUNUSED(softs))
1760 #endif
1761 	if (acp->pkt->pkt_reason == CMD_CMPLT)
1762 		acp->pkt->pkt_reason = reason;
1763 	acp->pkt->pkt_statistics |= stat;
1764 }
1765 
1766 /*
1767  * Handle a finished pkt of soft SCMD
1768  */
1769 static void
1770 aac_soft_callback(struct aac_softstate *softs, struct aac_cmd *acp)
1771 {
1772 	ASSERT(acp->pkt);
1773 
1774 	acp->flags |= AAC_CMD_CMPLT;
1775 
1776 	acp->pkt->pkt_state |= STATE_GOT_BUS | STATE_GOT_TARGET | \
1777 	    STATE_SENT_CMD | STATE_GOT_STATUS;
1778 	if (acp->pkt->pkt_state & STATE_XFERRED_DATA)
1779 		acp->pkt->pkt_resid = 0;
1780 
1781 	/* AAC_CMD_NO_INTR means no complete callback */
1782 	if (!(acp->flags & AAC_CMD_NO_INTR)) {
1783 		mutex_enter(&softs->q_comp_mutex);
1784 		aac_cmd_enqueue(&softs->q_comp, acp);
1785 		mutex_exit(&softs->q_comp_mutex);
1786 		ddi_trigger_softintr(softs->softint_id);
1787 	}
1788 }
1789 
1790 /*
1791  * Handlers for completed IOs, common to aac_intr_new() and aac_intr_old()
1792  */
1793 
1794 /*
1795  * Handle completed logical device IO command
1796  */
1797 /*ARGSUSED*/
1798 static void
1799 aac_ld_complete(struct aac_softstate *softs, struct aac_cmd *acp)
1800 {
1801 	struct aac_slot *slotp = acp->slotp;
1802 	struct aac_blockread_response *resp;
1803 	uint32_t status;
1804 
1805 	ASSERT(!(acp->flags & AAC_CMD_SYNC));
1806 	ASSERT(!(acp->flags & AAC_CMD_NO_CB));
1807 
1808 	acp->pkt->pkt_state |= STATE_GOT_STATUS;
1809 
1810 	/*
1811 	 * block_read/write has a similar response header, use blockread
1812 	 * response for both.
1813 	 */
1814 	resp = (struct aac_blockread_response *)&slotp->fibp->data[0];
1815 	status = ddi_get32(slotp->fib_acc_handle, &resp->Status);
1816 	if (status == ST_OK) {
1817 		acp->pkt->pkt_resid = 0;
1818 		acp->pkt->pkt_state |= STATE_XFERRED_DATA;
1819 	} else {
1820 		aac_set_arq_data_hwerr(acp);
1821 	}
1822 }
1823 
1824 /*
1825  * Handle completed phys. device IO command
1826  */
1827 static void
1828 aac_pd_complete(struct aac_softstate *softs, struct aac_cmd *acp)
1829 {
1830 	ddi_acc_handle_t acc = acp->slotp->fib_acc_handle;
1831 	struct aac_fib *fibp = acp->slotp->fibp;
1832 	struct scsi_pkt *pkt = acp->pkt;
1833 	struct aac_srb_reply *resp;
1834 	uint32_t resp_status;
1835 
1836 	ASSERT(!(acp->flags & AAC_CMD_SYNC));
1837 	ASSERT(!(acp->flags & AAC_CMD_NO_CB));
1838 
1839 	resp = (struct aac_srb_reply *)&fibp->data[0];
1840 	resp_status = ddi_get32(acc, &resp->status);
1841 
1842 	/* First check FIB status */
1843 	if (resp_status == ST_OK) {
1844 		uint32_t scsi_status;
1845 		uint32_t srb_status;
1846 		uint32_t data_xfer_length;
1847 
1848 		scsi_status = ddi_get32(acc, &resp->scsi_status);
1849 		srb_status = ddi_get32(acc, &resp->srb_status);
1850 		data_xfer_length = ddi_get32(acc, &resp->data_xfer_length);
1851 
1852 		*pkt->pkt_scbp = (uint8_t)scsi_status;
1853 		pkt->pkt_state |= STATE_GOT_STATUS;
1854 		if (scsi_status == STATUS_GOOD) {
1855 			uchar_t cmd = ((union scsi_cdb *)(void *)
1856 			    (pkt->pkt_cdbp))->scc_cmd;
1857 
1858 			/* Next check SRB status */
1859 			switch (srb_status & 0x3f) {
1860 			case SRB_STATUS_DATA_OVERRUN:
1861 				AACDB_PRINT(softs, CE_NOTE, "DATA_OVERRUN: " \
1862 				    "scmd=%d, xfer=%d, buflen=%d",
1863 				    (uint32_t)cmd, data_xfer_length,
1864 				    acp->bcount);
1865 
1866 				switch (cmd) {
1867 				case SCMD_READ:
1868 				case SCMD_WRITE:
1869 				case SCMD_READ_G1:
1870 				case SCMD_WRITE_G1:
1871 				case SCMD_READ_G4:
1872 				case SCMD_WRITE_G4:
1873 				case SCMD_READ_G5:
1874 				case SCMD_WRITE_G5:
1875 					aac_set_pkt_reason(softs, acp,
1876 					    CMD_DATA_OVR, 0);
1877 					break;
1878 				}
1879 				/*FALLTHRU*/
1880 			case SRB_STATUS_ERROR_RECOVERY:
1881 			case SRB_STATUS_PENDING:
1882 			case SRB_STATUS_SUCCESS:
1883 				/*
1884 				 * pkt_resid should only be calculated if the
1885 				 * status is ERROR_RECOVERY/PENDING/SUCCESS/
1886 				 * OVERRUN/UNDERRUN
1887 				 */
1888 				if (data_xfer_length) {
1889 					pkt->pkt_state |= STATE_XFERRED_DATA;
1890 					pkt->pkt_resid = acp->bcount - \
1891 					    data_xfer_length;
1892 					ASSERT(pkt->pkt_resid >= 0);
1893 				}
1894 				break;
1895 			case SRB_STATUS_ABORTED:
1896 				AACDB_PRINT(softs, CE_NOTE,
1897 				    "SRB_STATUS_ABORTED, xfer=%d, resid=%d",
1898 				    data_xfer_length, pkt->pkt_resid);
1899 				aac_set_pkt_reason(softs, acp, CMD_ABORTED,
1900 				    STAT_ABORTED);
1901 				break;
1902 			case SRB_STATUS_ABORT_FAILED:
1903 				AACDB_PRINT(softs, CE_NOTE,
1904 				    "SRB_STATUS_ABORT_FAILED, xfer=%d, " \
1905 				    "resid=%d", data_xfer_length,
1906 				    pkt->pkt_resid);
1907 				aac_set_pkt_reason(softs, acp, CMD_ABORT_FAIL,
1908 				    0);
1909 				break;
1910 			case SRB_STATUS_PARITY_ERROR:
1911 				AACDB_PRINT(softs, CE_NOTE,
1912 				    "SRB_STATUS_PARITY_ERROR, xfer=%d, " \
1913 				    "resid=%d", data_xfer_length,
1914 				    pkt->pkt_resid);
1915 				aac_set_pkt_reason(softs, acp, CMD_PER_FAIL, 0);
1916 				break;
1917 			case SRB_STATUS_NO_DEVICE:
1918 			case SRB_STATUS_INVALID_PATH_ID:
1919 			case SRB_STATUS_INVALID_TARGET_ID:
1920 			case SRB_STATUS_INVALID_LUN:
1921 			case SRB_STATUS_SELECTION_TIMEOUT:
1922 #ifdef DEBUG
1923 				if (AAC_DEV_IS_VALID(acp->dvp)) {
1924 					AACDB_PRINT(softs, CE_NOTE,
1925 					    "SRB_STATUS_NO_DEVICE(%d), " \
1926 					    "xfer=%d, resid=%d ",
1927 					    srb_status & 0x3f,
1928 					    data_xfer_length, pkt->pkt_resid);
1929 				}
1930 #endif
1931 				aac_set_pkt_reason(softs, acp, CMD_DEV_GONE, 0);
1932 				break;
1933 			case SRB_STATUS_COMMAND_TIMEOUT:
1934 			case SRB_STATUS_TIMEOUT:
1935 				AACDB_PRINT(softs, CE_NOTE,
1936 				    "SRB_STATUS_COMMAND_TIMEOUT, xfer=%d, " \
1937 				    "resid=%d", data_xfer_length,
1938 				    pkt->pkt_resid);
1939 				aac_set_pkt_reason(softs, acp, CMD_TIMEOUT,
1940 				    STAT_TIMEOUT);
1941 				break;
1942 			case SRB_STATUS_BUS_RESET:
1943 				AACDB_PRINT(softs, CE_NOTE,
1944 				    "SRB_STATUS_BUS_RESET, xfer=%d, " \
1945 				    "resid=%d", data_xfer_length,
1946 				    pkt->pkt_resid);
1947 				aac_set_pkt_reason(softs, acp, CMD_RESET,
1948 				    STAT_BUS_RESET);
1949 				break;
1950 			default:
1951 				AACDB_PRINT(softs, CE_NOTE, "srb_status=%d, " \
1952 				    "xfer=%d, resid=%d", srb_status & 0x3f,
1953 				    data_xfer_length, pkt->pkt_resid);
1954 				aac_set_pkt_reason(softs, acp, CMD_TRAN_ERR, 0);
1955 				break;
1956 			}
1957 		} else if (scsi_status == STATUS_CHECK) {
1958 			/* CHECK CONDITION */
1959 			struct scsi_arq_status *arqstat =
1960 			    (void *)(pkt->pkt_scbp);
1961 			uint32_t sense_data_size;
1962 
1963 			pkt->pkt_state |= STATE_ARQ_DONE;
1964 
1965 			*(uint8_t *)&arqstat->sts_rqpkt_status = STATUS_GOOD;
1966 			arqstat->sts_rqpkt_reason = CMD_CMPLT;
1967 			arqstat->sts_rqpkt_resid = 0;
1968 			arqstat->sts_rqpkt_state =
1969 			    STATE_GOT_BUS |
1970 			    STATE_GOT_TARGET |
1971 			    STATE_SENT_CMD |
1972 			    STATE_XFERRED_DATA;
1973 			arqstat->sts_rqpkt_statistics = 0;
1974 
1975 			sense_data_size = ddi_get32(acc,
1976 			    &resp->sense_data_size);
1977 			ASSERT(sense_data_size <= AAC_SENSE_BUFFERSIZE);
1978 			AACDB_PRINT(softs, CE_NOTE,
1979 			    "CHECK CONDITION: sense len=%d, xfer len=%d",
1980 			    sense_data_size, data_xfer_length);
1981 
1982 			if (sense_data_size > SENSE_LENGTH)
1983 				sense_data_size = SENSE_LENGTH;
1984 			ddi_rep_get8(acc, (uint8_t *)&arqstat->sts_sensedata,
1985 			    (uint8_t *)resp->sense_data, sense_data_size,
1986 			    DDI_DEV_AUTOINCR);
1987 		} else {
1988 			AACDB_PRINT(softs, CE_WARN, "invaild scsi status: " \
1989 			    "scsi_status=%d, srb_status=%d",
1990 			    scsi_status, srb_status);
1991 			aac_set_pkt_reason(softs, acp, CMD_TRAN_ERR, 0);
1992 		}
1993 	} else {
1994 		AACDB_PRINT(softs, CE_NOTE, "SRB failed: fib status %d",
1995 		    resp_status);
1996 		aac_set_pkt_reason(softs, acp, CMD_TRAN_ERR, 0);
1997 	}
1998 }
1999 
2000 /*
2001  * Handle completed IOCTL command
2002  */
2003 /*ARGSUSED*/
2004 void
2005 aac_ioctl_complete(struct aac_softstate *softs, struct aac_cmd *acp)
2006 {
2007 	struct aac_slot *slotp = acp->slotp;
2008 
2009 	/*
2010 	 * NOTE: Both aac_ioctl_send_fib() and aac_send_raw_srb()
2011 	 * may wait on softs->event, so use cv_broadcast() instead
2012 	 * of cv_signal().
2013 	 */
2014 	ASSERT(acp->flags & AAC_CMD_SYNC);
2015 	ASSERT(acp->flags & AAC_CMD_NO_CB);
2016 
2017 	/* Get the size of the response FIB from its FIB.Header.Size field */
2018 	acp->fib_size = ddi_get16(slotp->fib_acc_handle,
2019 	    &slotp->fibp->Header.Size);
2020 
2021 	ASSERT(acp->fib_size <= softs->aac_max_fib_size);
2022 	ddi_rep_get8(slotp->fib_acc_handle, (uint8_t *)acp->fibp,
2023 	    (uint8_t *)slotp->fibp, acp->fib_size, DDI_DEV_AUTOINCR);
2024 }
2025 
2026 /*
2027  * Handle completed Flush command
2028  */
2029 /*ARGSUSED*/
2030 static void
2031 aac_synccache_complete(struct aac_softstate *softs, struct aac_cmd *acp)
2032 {
2033 	struct aac_slot *slotp = acp->slotp;
2034 	ddi_acc_handle_t acc = slotp->fib_acc_handle;
2035 	struct aac_synchronize_reply *resp;
2036 	uint32_t status;
2037 
2038 	ASSERT(!(acp->flags & AAC_CMD_SYNC));
2039 
2040 	acp->pkt->pkt_state |= STATE_GOT_STATUS;
2041 
2042 	resp = (struct aac_synchronize_reply *)&slotp->fibp->data[0];
2043 	status = ddi_get32(acc, &resp->Status);
2044 	if (status != CT_OK)
2045 		aac_set_arq_data_hwerr(acp);
2046 }
2047 
2048 /*
2049  * Access PCI space to see if the driver can support the card
2050  */
2051 static int
2052 aac_check_card_type(struct aac_softstate *softs)
2053 {
2054 	ddi_acc_handle_t pci_config_handle;
2055 	int card_index;
2056 	uint32_t pci_cmd;
2057 
2058 	/* Map pci configuration space */
2059 	if ((pci_config_setup(softs->devinfo_p, &pci_config_handle)) !=
2060 	    DDI_SUCCESS) {
2061 		AACDB_PRINT(softs, CE_WARN, "Cannot setup pci config space");
2062 		return (AACERR);
2063 	}
2064 
2065 	softs->vendid = pci_config_get16(pci_config_handle, PCI_CONF_VENID);
2066 	softs->devid = pci_config_get16(pci_config_handle, PCI_CONF_DEVID);
2067 	softs->subvendid = pci_config_get16(pci_config_handle,
2068 	    PCI_CONF_SUBVENID);
2069 	softs->subsysid = pci_config_get16(pci_config_handle,
2070 	    PCI_CONF_SUBSYSID);
2071 
2072 	card_index = 0;
2073 	while (!CARD_IS_UNKNOWN(card_index)) {
2074 		if ((aac_cards[card_index].vendor == softs->vendid) &&
2075 		    (aac_cards[card_index].device == softs->devid) &&
2076 		    (aac_cards[card_index].subvendor == softs->subvendid) &&
2077 		    (aac_cards[card_index].subsys == softs->subsysid)) {
2078 			break;
2079 		}
2080 		card_index++;
2081 	}
2082 
2083 	softs->card = card_index;
2084 	softs->hwif = aac_cards[card_index].hwif;
2085 
2086 	/*
2087 	 * Unknown aac card
2088 	 * do a generic match based on the VendorID and DeviceID to
2089 	 * support the new cards in the aac family
2090 	 */
2091 	if (CARD_IS_UNKNOWN(card_index)) {
2092 		if (softs->vendid != 0x9005) {
2093 			AACDB_PRINT(softs, CE_WARN,
2094 			    "Unknown vendor 0x%x", softs->vendid);
2095 			goto error;
2096 		}
2097 		switch (softs->devid) {
2098 		case 0x285:
2099 			softs->hwif = AAC_HWIF_I960RX;
2100 			break;
2101 		case 0x286:
2102 			softs->hwif = AAC_HWIF_RKT;
2103 			break;
2104 		default:
2105 			AACDB_PRINT(softs, CE_WARN,
2106 			    "Unknown device \"pci9005,%x\"", softs->devid);
2107 			goto error;
2108 		}
2109 	}
2110 
2111 	/* Set hardware dependent interface */
2112 	switch (softs->hwif) {
2113 	case AAC_HWIF_I960RX:
2114 		softs->aac_if = aac_rx_interface;
2115 		softs->map_size_min = AAC_MAP_SIZE_MIN_RX;
2116 		break;
2117 	case AAC_HWIF_RKT:
2118 		softs->aac_if = aac_rkt_interface;
2119 		softs->map_size_min = AAC_MAP_SIZE_MIN_RKT;
2120 		break;
2121 	default:
2122 		AACDB_PRINT(softs, CE_WARN,
2123 		    "Unknown hardware interface %d", softs->hwif);
2124 		goto error;
2125 	}
2126 
2127 	/* Set card names */
2128 	(void *)strncpy(softs->vendor_name, aac_cards[card_index].vid,
2129 	    AAC_VENDOR_LEN);
2130 	(void *)strncpy(softs->product_name, aac_cards[card_index].desc,
2131 	    AAC_PRODUCT_LEN);
2132 
2133 	/* Set up quirks */
2134 	softs->flags = aac_cards[card_index].quirks;
2135 
2136 	/* Force the busmaster enable bit on */
2137 	pci_cmd = pci_config_get16(pci_config_handle, PCI_CONF_COMM);
2138 	if ((pci_cmd & PCI_COMM_ME) == 0) {
2139 		pci_cmd |= PCI_COMM_ME;
2140 		pci_config_put16(pci_config_handle, PCI_CONF_COMM, pci_cmd);
2141 		pci_cmd = pci_config_get16(pci_config_handle, PCI_CONF_COMM);
2142 		if ((pci_cmd & PCI_COMM_ME) == 0) {
2143 			cmn_err(CE_CONT, "?Cannot enable busmaster bit");
2144 			goto error;
2145 		}
2146 	}
2147 
2148 	/* Set memory base to map */
2149 	softs->pci_mem_base_paddr = 0xfffffff0UL & \
2150 	    pci_config_get32(pci_config_handle, PCI_CONF_BASE0);
2151 
2152 	pci_config_teardown(&pci_config_handle);
2153 
2154 	return (AACOK); /* card type detected */
2155 error:
2156 	pci_config_teardown(&pci_config_handle);
2157 	return (AACERR); /* no matched card found */
2158 }
2159 
2160 /*
2161  * Check the firmware to determine the features to support and the FIB
2162  * parameters to use.
2163  */
2164 static int
2165 aac_check_firmware(struct aac_softstate *softs)
2166 {
2167 	uint32_t options;
2168 	uint32_t atu_size;
2169 	ddi_acc_handle_t pci_handle;
2170 	uint8_t *data;
2171 	uint32_t max_fibs;
2172 	uint32_t max_fib_size;
2173 	uint32_t sg_tablesize;
2174 	uint32_t max_sectors;
2175 	uint32_t status;
2176 
2177 	/* Get supported options */
2178 	if ((aac_sync_mbcommand(softs, AAC_MONKER_GETINFO, 0, 0, 0, 0,
2179 	    &status)) != AACOK) {
2180 		if (status != SRB_STATUS_INVALID_REQUEST) {
2181 			cmn_err(CE_CONT,
2182 			    "?Fatal error: request adapter info error");
2183 			return (AACERR);
2184 		}
2185 		options = 0;
2186 		atu_size = 0;
2187 	} else {
2188 		options = AAC_MAILBOX_GET(softs, 1);
2189 		atu_size = AAC_MAILBOX_GET(softs, 2);
2190 	}
2191 
2192 	if (softs->state & AAC_STATE_RESET) {
2193 		if ((softs->support_opt == options) &&
2194 		    (softs->atu_size == atu_size))
2195 			return (AACOK);
2196 
2197 		cmn_err(CE_WARN,
2198 		    "?Fatal error: firmware changed, system needs reboot");
2199 		return (AACERR);
2200 	}
2201 
2202 	/*
2203 	 * The following critical settings are initialized only once during
2204 	 * driver attachment.
2205 	 */
2206 	softs->support_opt = options;
2207 	softs->atu_size = atu_size;
2208 
2209 	/* Process supported options */
2210 	if ((options & AAC_SUPPORTED_4GB_WINDOW) != 0 &&
2211 	    (softs->flags & AAC_FLAGS_NO4GB) == 0) {
2212 		AACDB_PRINT(softs, CE_NOTE, "!Enable FIB map 4GB window");
2213 		softs->flags |= AAC_FLAGS_4GB_WINDOW;
2214 	} else {
2215 		/*
2216 		 * Quirk AAC_FLAGS_NO4GB is for FIB address and thus comm space
2217 		 * only. IO is handled by the DMA engine which does not suffer
2218 		 * from the ATU window programming workarounds necessary for
2219 		 * CPU copy operations.
2220 		 */
2221 		softs->addr_dma_attr.dma_attr_addr_lo = 0x2000ull;
2222 		softs->addr_dma_attr.dma_attr_addr_hi = 0x7fffffffull;
2223 	}
2224 
2225 	if ((options & AAC_SUPPORTED_SGMAP_HOST64) != 0) {
2226 		AACDB_PRINT(softs, CE_NOTE, "!Enable SG map 64-bit address");
2227 		softs->buf_dma_attr.dma_attr_addr_hi = 0xffffffffffffffffull;
2228 		softs->buf_dma_attr.dma_attr_seg = 0xffffffffffffffffull;
2229 		softs->flags |= AAC_FLAGS_SG_64BIT;
2230 	}
2231 
2232 	if (options & AAC_SUPPORTED_64BIT_ARRAYSIZE) {
2233 		softs->flags |= AAC_FLAGS_ARRAY_64BIT;
2234 		AACDB_PRINT(softs, CE_NOTE, "!Enable 64-bit array size");
2235 	}
2236 
2237 	if (options & AAC_SUPPORTED_NONDASD) {
2238 		if ((ddi_prop_lookup_string(DDI_DEV_T_ANY, softs->devinfo_p, 0,
2239 		    "nondasd-enable", (char **)&data) == DDI_SUCCESS)) {
2240 			if (strcmp((char *)data, "yes") == 0) {
2241 				AACDB_PRINT(softs, CE_NOTE,
2242 				    "!Enable Non-DASD access");
2243 				softs->flags |= AAC_FLAGS_NONDASD;
2244 			}
2245 			ddi_prop_free(data);
2246 		}
2247 	}
2248 
2249 	/* Read preferred settings */
2250 	max_fib_size = 0;
2251 	if ((aac_sync_mbcommand(softs, AAC_MONKER_GETCOMMPREF,
2252 	    0, 0, 0, 0, NULL)) == AACOK) {
2253 		options = AAC_MAILBOX_GET(softs, 1);
2254 		max_fib_size = (options & 0xffff);
2255 		max_sectors = (options >> 16) << 1;
2256 		options = AAC_MAILBOX_GET(softs, 2);
2257 		sg_tablesize = (options >> 16);
2258 		options = AAC_MAILBOX_GET(softs, 3);
2259 		max_fibs = (options & 0xffff);
2260 	}
2261 
2262 	/* Enable new comm. and rawio at the same time */
2263 	if ((softs->support_opt & AAC_SUPPORTED_NEW_COMM) &&
2264 	    (max_fib_size != 0)) {
2265 		/* read out and save PCI MBR */
2266 		if ((atu_size > softs->map_size) &&
2267 		    (ddi_regs_map_setup(softs->devinfo_p, 1,
2268 		    (caddr_t *)&data, 0, atu_size, &softs->acc_attr,
2269 		    &pci_handle) == DDI_SUCCESS)) {
2270 			ddi_regs_map_free(&softs->pci_mem_handle);
2271 			softs->pci_mem_handle = pci_handle;
2272 			softs->pci_mem_base_vaddr = data;
2273 			softs->map_size = atu_size;
2274 		}
2275 		if (atu_size == softs->map_size) {
2276 			softs->flags |= AAC_FLAGS_NEW_COMM;
2277 			AACDB_PRINT(softs, CE_NOTE,
2278 			    "!Enable New Comm. interface");
2279 		}
2280 	}
2281 
2282 	/* Set FIB parameters */
2283 	if (softs->flags & AAC_FLAGS_NEW_COMM) {
2284 		softs->aac_max_fibs = max_fibs;
2285 		softs->aac_max_fib_size = max_fib_size;
2286 		softs->aac_max_sectors = max_sectors;
2287 		softs->aac_sg_tablesize = sg_tablesize;
2288 
2289 		softs->flags |= AAC_FLAGS_RAW_IO;
2290 		AACDB_PRINT(softs, CE_NOTE, "!Enable RawIO");
2291 	} else {
2292 		softs->aac_max_fibs =
2293 		    (softs->flags & AAC_FLAGS_256FIBS) ? 256 : 512;
2294 		softs->aac_max_fib_size = AAC_FIB_SIZE;
2295 		softs->aac_max_sectors = 128;	/* 64K */
2296 		if (softs->flags & AAC_FLAGS_17SG)
2297 			softs->aac_sg_tablesize = 17;
2298 		else if (softs->flags & AAC_FLAGS_34SG)
2299 			softs->aac_sg_tablesize = 34;
2300 		else if (softs->flags & AAC_FLAGS_SG_64BIT)
2301 			softs->aac_sg_tablesize = (AAC_FIB_DATASIZE -
2302 			    sizeof (struct aac_blockwrite64) +
2303 			    sizeof (struct aac_sg_entry64)) /
2304 			    sizeof (struct aac_sg_entry64);
2305 		else
2306 			softs->aac_sg_tablesize = (AAC_FIB_DATASIZE -
2307 			    sizeof (struct aac_blockwrite) +
2308 			    sizeof (struct aac_sg_entry)) /
2309 			    sizeof (struct aac_sg_entry);
2310 	}
2311 
2312 	if ((softs->flags & AAC_FLAGS_RAW_IO) &&
2313 	    (softs->flags & AAC_FLAGS_ARRAY_64BIT)) {
2314 		softs->flags |= AAC_FLAGS_LBA_64BIT;
2315 		AACDB_PRINT(softs, CE_NOTE, "!Enable 64-bit array");
2316 	}
2317 	softs->buf_dma_attr.dma_attr_sgllen = softs->aac_sg_tablesize;
2318 	softs->buf_dma_attr.dma_attr_maxxfer = softs->aac_max_sectors << 9;
2319 	/*
2320 	 * 64K maximum segment size in scatter gather list is controlled by
2321 	 * the NEW_COMM bit in the adapter information. If not set, the card
2322 	 * can only accept a maximum of 64K. It is not recommended to permit
2323 	 * more than 128KB of total transfer size to the adapters because
2324 	 * performance is negatively impacted.
2325 	 *
2326 	 * For new comm, segment size equals max xfer size. For old comm,
2327 	 * we use 64K for both.
2328 	 */
2329 	softs->buf_dma_attr.dma_attr_count_max =
2330 	    softs->buf_dma_attr.dma_attr_maxxfer - 1;
2331 
2332 	/* Setup FIB operations */
2333 	if (softs->flags & AAC_FLAGS_RAW_IO)
2334 		softs->aac_cmd_fib = aac_cmd_fib_rawio;
2335 	else if (softs->flags & AAC_FLAGS_SG_64BIT)
2336 		softs->aac_cmd_fib = aac_cmd_fib_brw64;
2337 	else
2338 		softs->aac_cmd_fib = aac_cmd_fib_brw;
2339 	softs->aac_cmd_fib_scsi = (softs->flags & AAC_FLAGS_SG_64BIT) ? \
2340 	    aac_cmd_fib_scsi64 : aac_cmd_fib_scsi32;
2341 
2342 	/* 64-bit LBA needs descriptor format sense data */
2343 	softs->slen = sizeof (struct scsi_arq_status);
2344 	if ((softs->flags & AAC_FLAGS_LBA_64BIT) &&
2345 	    softs->slen < AAC_ARQ64_LENGTH)
2346 		softs->slen = AAC_ARQ64_LENGTH;
2347 
2348 	AACDB_PRINT(softs, CE_NOTE,
2349 	    "!max_fibs %d max_fibsize 0x%x max_sectors %d max_sg %d",
2350 	    softs->aac_max_fibs, softs->aac_max_fib_size,
2351 	    softs->aac_max_sectors, softs->aac_sg_tablesize);
2352 
2353 	return (AACOK);
2354 }
2355 
2356 static void
2357 aac_fsa_rev(struct aac_softstate *softs, struct FsaRev *fsarev0,
2358     struct FsaRev *fsarev1)
2359 {
2360 	ddi_acc_handle_t acc = softs->comm_space_acc_handle;
2361 
2362 	AAC_GET_FIELD8(acc, fsarev1, fsarev0, external.comp.dash);
2363 	AAC_GET_FIELD8(acc, fsarev1, fsarev0, external.comp.type);
2364 	AAC_GET_FIELD8(acc, fsarev1, fsarev0, external.comp.minor);
2365 	AAC_GET_FIELD8(acc, fsarev1, fsarev0, external.comp.major);
2366 	AAC_GET_FIELD32(acc, fsarev1, fsarev0, buildNumber);
2367 }
2368 
2369 /*
2370  * The following function comes from Adaptec:
2371  *
2372  * Query adapter information and supplement adapter information
2373  */
2374 static int
2375 aac_get_adapter_info(struct aac_softstate *softs,
2376     struct aac_adapter_info *ainfr, struct aac_supplement_adapter_info *sinfr)
2377 {
2378 	ddi_acc_handle_t acc = softs->sync_slot.fib_acc_handle;
2379 	struct aac_fib *fibp = softs->sync_slot.fibp;
2380 	struct aac_adapter_info *ainfp;
2381 	struct aac_supplement_adapter_info *sinfp;
2382 
2383 	ddi_put8(acc, &fibp->data[0], 0);
2384 	if (aac_sync_fib(softs, RequestAdapterInfo,
2385 	    sizeof (struct aac_fib_header)) != AACOK) {
2386 		AACDB_PRINT(softs, CE_WARN, "RequestAdapterInfo failed");
2387 		return (AACERR);
2388 	}
2389 	ainfp = (struct aac_adapter_info *)fibp->data;
2390 	if (ainfr) {
2391 		AAC_GET_FIELD32(acc, ainfr, ainfp, SupportedOptions);
2392 		AAC_GET_FIELD32(acc, ainfr, ainfp, PlatformBase);
2393 		AAC_GET_FIELD32(acc, ainfr, ainfp, CpuArchitecture);
2394 		AAC_GET_FIELD32(acc, ainfr, ainfp, CpuVariant);
2395 		AAC_GET_FIELD32(acc, ainfr, ainfp, ClockSpeed);
2396 		AAC_GET_FIELD32(acc, ainfr, ainfp, ExecutionMem);
2397 		AAC_GET_FIELD32(acc, ainfr, ainfp, BufferMem);
2398 		AAC_GET_FIELD32(acc, ainfr, ainfp, TotalMem);
2399 		aac_fsa_rev(softs, &ainfp->KernelRevision,
2400 		    &ainfr->KernelRevision);
2401 		aac_fsa_rev(softs, &ainfp->MonitorRevision,
2402 		    &ainfr->MonitorRevision);
2403 		aac_fsa_rev(softs, &ainfp->HardwareRevision,
2404 		    &ainfr->HardwareRevision);
2405 		aac_fsa_rev(softs, &ainfp->BIOSRevision,
2406 		    &ainfr->BIOSRevision);
2407 		AAC_GET_FIELD32(acc, ainfr, ainfp, ClusteringEnabled);
2408 		AAC_GET_FIELD32(acc, ainfr, ainfp, ClusterChannelMask);
2409 		AAC_GET_FIELD64(acc, ainfr, ainfp, SerialNumber);
2410 		AAC_GET_FIELD32(acc, ainfr, ainfp, batteryPlatform);
2411 		AAC_GET_FIELD32(acc, ainfr, ainfp, SupportedOptions);
2412 		AAC_GET_FIELD32(acc, ainfr, ainfp, OemVariant);
2413 	}
2414 	if (sinfr) {
2415 		if (!(softs->support_opt &
2416 		    AAC_SUPPORTED_SUPPLEMENT_ADAPTER_INFO)) {
2417 			AACDB_PRINT(softs, CE_WARN,
2418 			    "SupplementAdapterInfo not supported");
2419 			return (AACERR);
2420 		}
2421 		ddi_put8(acc, &fibp->data[0], 0);
2422 		if (aac_sync_fib(softs, RequestSupplementAdapterInfo,
2423 		    sizeof (struct aac_fib_header)) != AACOK) {
2424 			AACDB_PRINT(softs, CE_WARN,
2425 			    "RequestSupplementAdapterInfo failed");
2426 			return (AACERR);
2427 		}
2428 		sinfp = (struct aac_supplement_adapter_info *)fibp->data;
2429 		AAC_REP_GET_FIELD8(acc, sinfr, sinfp, AdapterTypeText[0], 17+1);
2430 		AAC_REP_GET_FIELD8(acc, sinfr, sinfp, Pad[0], 2);
2431 		AAC_GET_FIELD32(acc, sinfr, sinfp, FlashMemoryByteSize);
2432 		AAC_GET_FIELD32(acc, sinfr, sinfp, FlashImageId);
2433 		AAC_GET_FIELD32(acc, sinfr, sinfp, MaxNumberPorts);
2434 		AAC_GET_FIELD32(acc, sinfr, sinfp, Version);
2435 		AAC_GET_FIELD32(acc, sinfr, sinfp, FeatureBits);
2436 		AAC_GET_FIELD8(acc, sinfr, sinfp, SlotNumber);
2437 		AAC_REP_GET_FIELD8(acc, sinfr, sinfp, ReservedPad0[0], 3);
2438 		AAC_REP_GET_FIELD8(acc, sinfr, sinfp, BuildDate[0], 12);
2439 		AAC_GET_FIELD32(acc, sinfr, sinfp, CurrentNumberPorts);
2440 		AAC_REP_GET_FIELD8(acc, sinfr, sinfp, VpdInfo,
2441 		    sizeof (struct vpd_info));
2442 		aac_fsa_rev(softs, &sinfp->FlashFirmwareRevision,
2443 		    &sinfr->FlashFirmwareRevision);
2444 		AAC_GET_FIELD32(acc, sinfr, sinfp, RaidTypeMorphOptions);
2445 		aac_fsa_rev(softs, &sinfp->FlashFirmwareBootRevision,
2446 		    &sinfr->FlashFirmwareBootRevision);
2447 		AAC_REP_GET_FIELD8(acc, sinfr, sinfp, MfgPcbaSerialNo,
2448 		    MFG_PCBA_SERIAL_NUMBER_WIDTH);
2449 		AAC_REP_GET_FIELD8(acc, sinfr, sinfp, MfgWWNName[0],
2450 		    MFG_WWN_WIDTH);
2451 		AAC_REP_GET_FIELD32(acc, sinfr, sinfp, ReservedGrowth[0], 2);
2452 	}
2453 	return (AACOK);
2454 }
2455 
2456 static int
2457 aac_get_bus_info(struct aac_softstate *softs, uint32_t *bus_max,
2458     uint32_t *tgt_max)
2459 {
2460 	ddi_acc_handle_t acc = softs->sync_slot.fib_acc_handle;
2461 	struct aac_fib *fibp = softs->sync_slot.fibp;
2462 	struct aac_ctcfg *c_cmd;
2463 	struct aac_ctcfg_resp *c_resp;
2464 	uint32_t scsi_method_id;
2465 	struct aac_bus_info *cmd;
2466 	struct aac_bus_info_response *resp;
2467 	int rval;
2468 
2469 	/* Detect MethodId */
2470 	c_cmd = (struct aac_ctcfg *)&fibp->data[0];
2471 	ddi_put32(acc, &c_cmd->Command, VM_ContainerConfig);
2472 	ddi_put32(acc, &c_cmd->cmd, CT_GET_SCSI_METHOD);
2473 	ddi_put32(acc, &c_cmd->param, 0);
2474 	rval = aac_sync_fib(softs, ContainerCommand,
2475 	    AAC_FIB_SIZEOF(struct aac_ctcfg));
2476 	c_resp = (struct aac_ctcfg_resp *)&fibp->data[0];
2477 	if (rval != AACOK || ddi_get32(acc, &c_resp->Status) != 0) {
2478 		AACDB_PRINT(softs, CE_WARN,
2479 		    "VM_ContainerConfig command fail");
2480 		return (AACERR);
2481 	}
2482 	scsi_method_id = ddi_get32(acc, &c_resp->param);
2483 
2484 	/* Detect phys. bus count and max. target id first */
2485 	cmd = (struct aac_bus_info *)&fibp->data[0];
2486 	ddi_put32(acc, &cmd->Command, VM_Ioctl);
2487 	ddi_put32(acc, &cmd->ObjType, FT_DRIVE); /* physical drive */
2488 	ddi_put32(acc, &cmd->MethodId, scsi_method_id);
2489 	ddi_put32(acc, &cmd->ObjectId, 0);
2490 	ddi_put32(acc, &cmd->CtlCmd, GetBusInfo);
2491 	/*
2492 	 * For VM_Ioctl, the firmware uses the Header.Size filled from the
2493 	 * driver as the size to be returned. Therefore the driver has to use
2494 	 * sizeof (struct aac_bus_info_response) because it is greater than
2495 	 * sizeof (struct aac_bus_info).
2496 	 */
2497 	rval = aac_sync_fib(softs, ContainerCommand,
2498 	    AAC_FIB_SIZEOF(struct aac_bus_info_response));
2499 	resp = (struct aac_bus_info_response *)cmd;
2500 
2501 	/* Scan all coordinates with INQUIRY */
2502 	if ((rval != AACOK) || (ddi_get32(acc, &resp->Status) != 0)) {
2503 		AACDB_PRINT(softs, CE_WARN, "GetBusInfo command fail");
2504 		return (AACERR);
2505 	}
2506 	*bus_max = ddi_get32(acc, &resp->BusCount);
2507 	*tgt_max = ddi_get32(acc, &resp->TargetsPerBus);
2508 	return (AACOK);
2509 }
2510 
2511 /*
2512  * The following function comes from Adaptec:
2513  *
2514  * Routine to be called during initialization of communications with
2515  * the adapter to handle possible adapter configuration issues. When
2516  * the adapter first boots up, it examines attached drives, etc, and
2517  * potentially comes up with a new or revised configuration (relative to
2518  * what's stored in it's NVRAM). Additionally it may discover problems
2519  * that make the current physical configuration unworkable (currently
2520  * applicable only to cluster configuration issues).
2521  *
2522  * If there are no configuration issues or the issues are considered
2523  * trival by the adapter, it will set it's configuration status to
2524  * "FSACT_CONTINUE" and execute the "commit confiuguration" action
2525  * automatically on it's own.
2526  *
2527  * However, if there are non-trivial issues, the adapter will set it's
2528  * internal configuration status to "FSACT_PAUSE" or "FASCT_ABORT"
2529  * and wait for some agent on the host to issue the "\ContainerCommand
2530  * \VM_ContainerConfig\CT_COMMIT_CONFIG" FIB command to cause the
2531  * adapter to commit the new/updated configuration and enable
2532  * un-inhibited operation.  The host agent should first issue the
2533  * "\ContainerCommand\VM_ContainerConfig\CT_GET_CONFIG_STATUS" FIB
2534  * command to obtain information about config issues detected by
2535  * the adapter.
2536  *
2537  * Normally the adapter's PC BIOS will execute on the host following
2538  * adapter poweron and reset and will be responsible for querring the
2539  * adapter with CT_GET_CONFIG_STATUS and issuing the CT_COMMIT_CONFIG
2540  * command if appropriate.
2541  *
2542  * However, with the introduction of IOP reset support, the adapter may
2543  * boot up without the benefit of the adapter's PC BIOS host agent.
2544  * This routine is intended to take care of these issues in situations
2545  * where BIOS doesn't execute following adapter poweron or reset.  The
2546  * CT_COMMIT_CONFIG command is a no-op if it's already been issued, so
2547  * there is no harm in doing this when it's already been done.
2548  */
2549 static int
2550 aac_handle_adapter_config_issues(struct aac_softstate *softs)
2551 {
2552 	ddi_acc_handle_t acc = softs->sync_slot.fib_acc_handle;
2553 	struct aac_fib *fibp = softs->sync_slot.fibp;
2554 	struct aac_Container *cmd;
2555 	struct aac_Container_resp *resp;
2556 	struct aac_cf_status_header *cfg_sts_hdr;
2557 	uint32_t resp_status;
2558 	uint32_t ct_status;
2559 	uint32_t cfg_stat_action;
2560 	int rval;
2561 
2562 	/* Get adapter config status */
2563 	cmd = (struct aac_Container *)&fibp->data[0];
2564 
2565 	bzero(cmd, sizeof (*cmd) - CT_PACKET_SIZE);
2566 	ddi_put32(acc, &cmd->Command, VM_ContainerConfig);
2567 	ddi_put32(acc, &cmd->CTCommand.command, CT_GET_CONFIG_STATUS);
2568 	ddi_put32(acc, &cmd->CTCommand.param[CNT_SIZE],
2569 	    sizeof (struct aac_cf_status_header));
2570 	rval = aac_sync_fib(softs, ContainerCommand,
2571 	    AAC_FIB_SIZEOF(struct aac_Container));
2572 	resp = (struct aac_Container_resp *)cmd;
2573 	cfg_sts_hdr = (struct aac_cf_status_header *)resp->CTResponse.data;
2574 
2575 	resp_status = ddi_get32(acc, &resp->Status);
2576 	ct_status = ddi_get32(acc, &resp->CTResponse.param[0]);
2577 	if ((rval == AACOK) && (resp_status == 0) && (ct_status == CT_OK)) {
2578 		cfg_stat_action = ddi_get32(acc, &cfg_sts_hdr->action);
2579 
2580 		/* Commit configuration if it's reasonable to do so. */
2581 		if (cfg_stat_action <= CFACT_PAUSE) {
2582 			bzero(cmd, sizeof (*cmd) - CT_PACKET_SIZE);
2583 			ddi_put32(acc, &cmd->Command, VM_ContainerConfig);
2584 			ddi_put32(acc, &cmd->CTCommand.command,
2585 			    CT_COMMIT_CONFIG);
2586 			rval = aac_sync_fib(softs, ContainerCommand,
2587 			    AAC_FIB_SIZEOF(struct aac_Container));
2588 
2589 			resp_status = ddi_get32(acc, &resp->Status);
2590 			ct_status = ddi_get32(acc, &resp->CTResponse.param[0]);
2591 			if ((rval == AACOK) && (resp_status == 0) &&
2592 			    (ct_status == CT_OK))
2593 				/* Successful completion */
2594 				rval = AACMPE_OK;
2595 			else
2596 				/* Auto-commit aborted due to error(s). */
2597 				rval = AACMPE_COMMIT_CONFIG;
2598 		} else {
2599 			/*
2600 			 * Auto-commit aborted due to adapter indicating
2601 			 * configuration issue(s) too dangerous to auto-commit.
2602 			 */
2603 			rval = AACMPE_CONFIG_STATUS;
2604 		}
2605 	} else {
2606 		cmn_err(CE_WARN, "!Configuration issue, auto-commit aborted");
2607 		rval = AACMPE_CONFIG_STATUS;
2608 	}
2609 	return (rval);
2610 }
2611 
2612 /*
2613  * Hardware initialization and resource allocation
2614  */
2615 static int
2616 aac_common_attach(struct aac_softstate *softs)
2617 {
2618 	uint32_t status;
2619 	int i;
2620 
2621 	DBCALLED(softs, 1);
2622 
2623 	/*
2624 	 * Do a little check here to make sure there aren't any outstanding
2625 	 * FIBs in the message queue. At this point there should not be and
2626 	 * if there are they are probably left over from another instance of
2627 	 * the driver like when the system crashes and the crash dump driver
2628 	 * gets loaded.
2629 	 */
2630 	while (AAC_OUTB_GET(softs) != 0xfffffffful)
2631 		;
2632 
2633 	/*
2634 	 * Wait the card to complete booting up before do anything that
2635 	 * attempts to communicate with it.
2636 	 */
2637 	status = AAC_FWSTATUS_GET(softs);
2638 	if (status == AAC_SELF_TEST_FAILED || status == AAC_KERNEL_PANIC)
2639 		goto error;
2640 	i = AAC_FWUP_TIMEOUT * 1000; /* set timeout */
2641 	AAC_BUSYWAIT(AAC_FWSTATUS_GET(softs) & AAC_KERNEL_UP_AND_RUNNING, i);
2642 	if (i == 0) {
2643 		cmn_err(CE_CONT, "?Fatal error: controller not ready");
2644 		aac_fm_ereport(softs, DDI_FM_DEVICE_NO_RESPONSE);
2645 		ddi_fm_service_impact(softs->devinfo_p, DDI_SERVICE_LOST);
2646 		goto error;
2647 	}
2648 
2649 	/* Read and set card supported options and settings */
2650 	if (aac_check_firmware(softs) == AACERR) {
2651 		aac_fm_ereport(softs, DDI_FM_DEVICE_NO_RESPONSE);
2652 		ddi_fm_service_impact(softs->devinfo_p, DDI_SERVICE_LOST);
2653 		goto error;
2654 	}
2655 
2656 	/* Clear out all interrupts */
2657 	AAC_STATUS_CLR(softs, ~0);
2658 
2659 	/* Setup communication space with the card */
2660 	if (softs->comm_space_dma_handle == NULL) {
2661 		if (aac_alloc_comm_space(softs) != AACOK)
2662 			goto error;
2663 	}
2664 	if (aac_setup_comm_space(softs) != AACOK) {
2665 		cmn_err(CE_CONT, "?Setup communication space failed");
2666 		aac_fm_ereport(softs, DDI_FM_DEVICE_NO_RESPONSE);
2667 		ddi_fm_service_impact(softs->devinfo_p, DDI_SERVICE_LOST);
2668 		goto error;
2669 	}
2670 
2671 #ifdef DEBUG
2672 	if (aac_get_fw_debug_buffer(softs) != AACOK)
2673 		cmn_err(CE_CONT, "?firmware UART trace not supported");
2674 #endif
2675 
2676 	/* Allocate slots */
2677 	if ((softs->total_slots == 0) && (aac_create_slots(softs) != AACOK)) {
2678 		cmn_err(CE_CONT, "?Fatal error: slots allocate failed");
2679 		goto error;
2680 	}
2681 	AACDB_PRINT(softs, CE_NOTE, "%d slots allocated", softs->total_slots);
2682 
2683 	/* Allocate FIBs */
2684 	if (softs->total_fibs < softs->total_slots) {
2685 		aac_alloc_fibs(softs);
2686 		if (softs->total_fibs == 0)
2687 			goto error;
2688 		AACDB_PRINT(softs, CE_NOTE, "%d fibs allocated",
2689 		    softs->total_fibs);
2690 	}
2691 
2692 	/* Get adapter names */
2693 	if (CARD_IS_UNKNOWN(softs->card)) {
2694 		struct aac_supplement_adapter_info sinf;
2695 
2696 		if (aac_get_adapter_info(softs, NULL, &sinf) != AACOK) {
2697 			cmn_err(CE_CONT, "?Query adapter information failed");
2698 		} else {
2699 			char *p, *p0, *p1;
2700 
2701 			/*
2702 			 * Now find the controller name in supp_adapter_info->
2703 			 * AdapterTypeText. Use the first word as the vendor
2704 			 * and the other words as the product name.
2705 			 */
2706 			AACDB_PRINT(softs, CE_NOTE, "sinf.AdapterTypeText = "
2707 			    "\"%s\"", sinf.AdapterTypeText);
2708 			p = sinf.AdapterTypeText;
2709 			p0 = p1 = NULL;
2710 			/* Skip heading spaces */
2711 			while (*p && (*p == ' ' || *p == '\t'))
2712 				p++;
2713 			p0 = p;
2714 			while (*p && (*p != ' ' && *p != '\t'))
2715 				p++;
2716 			/* Remove middle spaces */
2717 			while (*p && (*p == ' ' || *p == '\t'))
2718 				*p++ = 0;
2719 			p1 = p;
2720 			/* Remove trailing spaces */
2721 			p = p1 + strlen(p1) - 1;
2722 			while (p > p1 && (*p == ' ' || *p == '\t'))
2723 				*p-- = 0;
2724 			if (*p0 && *p1) {
2725 				(void *)strncpy(softs->vendor_name, p0,
2726 				    AAC_VENDOR_LEN);
2727 				(void *)strncpy(softs->product_name, p1,
2728 				    AAC_PRODUCT_LEN);
2729 			} else {
2730 				cmn_err(CE_WARN,
2731 				    "?adapter name mis-formatted\n");
2732 				if (*p0)
2733 					(void *)strncpy(softs->product_name,
2734 					    p0, AAC_PRODUCT_LEN);
2735 			}
2736 		}
2737 	}
2738 
2739 	cmn_err(CE_NOTE,
2740 	    "!aac driver %d.%02d.%02d-%d, found card: " \
2741 	    "%s %s(pci0x%x.%x.%x.%x) at 0x%x",
2742 	    AAC_DRIVER_MAJOR_VERSION,
2743 	    AAC_DRIVER_MINOR_VERSION,
2744 	    AAC_DRIVER_BUGFIX_LEVEL,
2745 	    AAC_DRIVER_BUILD,
2746 	    softs->vendor_name, softs->product_name,
2747 	    softs->vendid, softs->devid, softs->subvendid, softs->subsysid,
2748 	    softs->pci_mem_base_paddr);
2749 
2750 	/* Perform acceptance of adapter-detected config changes if possible */
2751 	if (aac_handle_adapter_config_issues(softs) != AACMPE_OK) {
2752 		cmn_err(CE_CONT, "?Handle adapter config issues failed");
2753 		aac_fm_ereport(softs, DDI_FM_DEVICE_NO_RESPONSE);
2754 		ddi_fm_service_impact(softs->devinfo_p, DDI_SERVICE_LOST);
2755 		goto error;
2756 	}
2757 
2758 	/* Setup containers (logical devices) */
2759 	if (aac_probe_containers(softs) != AACOK) {
2760 		cmn_err(CE_CONT, "?Fatal error: get container info error");
2761 		goto error;
2762 	}
2763 
2764 	/* Setup phys. devices */
2765 	if (softs->flags & AAC_FLAGS_NONDASD) {
2766 		uint32_t bus_max, tgt_max;
2767 		uint32_t bus, tgt;
2768 		int index;
2769 
2770 		if (aac_get_bus_info(softs, &bus_max, &tgt_max) != AACOK) {
2771 			cmn_err(CE_CONT, "?Fatal error: get bus info error");
2772 			goto error;
2773 		}
2774 		AACDB_PRINT(softs, CE_NOTE, "bus_max=%d, tgt_max=%d",
2775 		    bus_max, tgt_max);
2776 		if (bus_max != softs->bus_max || tgt_max != softs->tgt_max) {
2777 			if (softs->state & AAC_STATE_RESET) {
2778 				cmn_err(CE_WARN,
2779 				    "?Fatal error: bus map changed");
2780 				goto error;
2781 			}
2782 			softs->bus_max = bus_max;
2783 			softs->tgt_max = tgt_max;
2784 			if (softs->nondasds) {
2785 				kmem_free(softs->nondasds, AAC_MAX_PD(softs) * \
2786 				    sizeof (struct aac_nondasd));
2787 			}
2788 			softs->nondasds = kmem_zalloc(AAC_MAX_PD(softs) * \
2789 			    sizeof (struct aac_nondasd), KM_SLEEP);
2790 
2791 			index = 0;
2792 			for (bus = 0; bus < softs->bus_max; bus++) {
2793 				for (tgt = 0; tgt < softs->tgt_max; tgt++) {
2794 					struct aac_nondasd *dvp =
2795 					    &softs->nondasds[index++];
2796 					dvp->dev.type = AAC_DEV_PD;
2797 					dvp->bus = bus;
2798 					dvp->tid = tgt;
2799 				}
2800 			}
2801 		}
2802 	}
2803 
2804 	/* Check dma & acc handles allocated in attach */
2805 	if (aac_check_dma_handle(softs->comm_space_dma_handle) != DDI_SUCCESS) {
2806 		ddi_fm_service_impact(softs->devinfo_p, DDI_SERVICE_LOST);
2807 		goto error;
2808 	}
2809 
2810 	if (aac_check_acc_handle(softs->pci_mem_handle) != DDI_SUCCESS) {
2811 		ddi_fm_service_impact(softs->devinfo_p, DDI_SERVICE_LOST);
2812 		goto error;
2813 	}
2814 
2815 	for (i = 0; i < softs->total_slots; i++) {
2816 		if (aac_check_dma_handle(softs->io_slot[i].fib_dma_handle) !=
2817 		    DDI_SUCCESS) {
2818 			ddi_fm_service_impact(softs->devinfo_p,
2819 			    DDI_SERVICE_LOST);
2820 			goto error;
2821 		}
2822 	}
2823 
2824 	return (AACOK);
2825 error:
2826 	if (softs->state & AAC_STATE_RESET)
2827 		return (AACERR);
2828 	if (softs->nondasds) {
2829 		kmem_free(softs->nondasds, AAC_MAX_PD(softs) * \
2830 		    sizeof (struct aac_nondasd));
2831 		softs->nondasds = NULL;
2832 	}
2833 	if (softs->total_fibs > 0)
2834 		aac_destroy_fibs(softs);
2835 	if (softs->total_slots > 0)
2836 		aac_destroy_slots(softs);
2837 	if (softs->comm_space_dma_handle)
2838 		aac_free_comm_space(softs);
2839 	return (AACERR);
2840 }
2841 
2842 /*
2843  * Hardware shutdown and resource release
2844  */
2845 static void
2846 aac_common_detach(struct aac_softstate *softs)
2847 {
2848 	DBCALLED(softs, 1);
2849 
2850 	(void) aac_shutdown(softs);
2851 
2852 	if (softs->nondasds) {
2853 		kmem_free(softs->nondasds, AAC_MAX_PD(softs) * \
2854 		    sizeof (struct aac_nondasd));
2855 		softs->nondasds = NULL;
2856 	}
2857 	aac_destroy_fibs(softs);
2858 	aac_destroy_slots(softs);
2859 	aac_free_comm_space(softs);
2860 }
2861 
2862 /*
2863  * Send a synchronous command to the controller and wait for a result.
2864  * Indicate if the controller completed the command with an error status.
2865  */
2866 int
2867 aac_sync_mbcommand(struct aac_softstate *softs, uint32_t cmd,
2868     uint32_t arg0, uint32_t arg1, uint32_t arg2, uint32_t arg3,
2869     uint32_t *statusp)
2870 {
2871 	int timeout;
2872 	uint32_t status;
2873 
2874 	if (statusp != NULL)
2875 		*statusp = SRB_STATUS_SUCCESS;
2876 
2877 	/* Fill in mailbox */
2878 	AAC_MAILBOX_SET(softs, cmd, arg0, arg1, arg2, arg3);
2879 
2880 	/* Ensure the sync command doorbell flag is cleared */
2881 	AAC_STATUS_CLR(softs, AAC_DB_SYNC_COMMAND);
2882 
2883 	/* Then set it to signal the adapter */
2884 	AAC_NOTIFY(softs, AAC_DB_SYNC_COMMAND);
2885 
2886 	/* Spin waiting for the command to complete */
2887 	timeout = AAC_IMMEDIATE_TIMEOUT * 1000;
2888 	AAC_BUSYWAIT(AAC_STATUS_GET(softs) & AAC_DB_SYNC_COMMAND, timeout);
2889 	if (!timeout) {
2890 		AACDB_PRINT(softs, CE_WARN,
2891 		    "Sync command timed out after %d seconds (0x%x)!",
2892 		    AAC_IMMEDIATE_TIMEOUT, AAC_FWSTATUS_GET(softs));
2893 		return (AACERR);
2894 	}
2895 
2896 	/* Clear the completion flag */
2897 	AAC_STATUS_CLR(softs, AAC_DB_SYNC_COMMAND);
2898 
2899 	/* Get the command status */
2900 	status = AAC_MAILBOX_GET(softs, 0);
2901 	if (statusp != NULL)
2902 		*statusp = status;
2903 	if (status != SRB_STATUS_SUCCESS) {
2904 		AACDB_PRINT(softs, CE_WARN,
2905 		    "Sync command fail: status = 0x%x", status);
2906 		return (AACERR);
2907 	}
2908 
2909 	return (AACOK);
2910 }
2911 
2912 /*
2913  * Send a synchronous FIB to the adapter and wait for its completion
2914  */
2915 static int
2916 aac_sync_fib(struct aac_softstate *softs, uint16_t cmd, uint16_t fibsize)
2917 {
2918 	struct aac_slot *slotp = &softs->sync_slot;
2919 	ddi_dma_handle_t dma = slotp->fib_dma_handle;
2920 	uint32_t status;
2921 	int rval;
2922 
2923 	/* Sync fib only supports 512 bytes */
2924 	if (fibsize > AAC_FIB_SIZE)
2925 		return (AACERR);
2926 
2927 	/*
2928 	 * Setup sync fib
2929 	 * Need not reinitialize FIB header if it's already been filled
2930 	 * by others like aac_cmd_fib_scsi as aac_cmd.
2931 	 */
2932 	if (slotp->acp == NULL)
2933 		aac_cmd_fib_header(softs, slotp, cmd, fibsize);
2934 
2935 	AACDB_PRINT_FIB(softs, &softs->sync_slot);
2936 
2937 	(void) ddi_dma_sync(dma, offsetof(struct aac_comm_space, sync_fib),
2938 	    fibsize, DDI_DMA_SYNC_FORDEV);
2939 
2940 	/* Give the FIB to the controller, wait for a response. */
2941 	rval = aac_sync_mbcommand(softs, AAC_MONKER_SYNCFIB,
2942 	    slotp->fib_phyaddr, 0, 0, 0, &status);
2943 	if (rval == AACERR) {
2944 		AACDB_PRINT(softs, CE_WARN,
2945 		    "Send sync fib to controller failed");
2946 		return (AACERR);
2947 	}
2948 
2949 	(void) ddi_dma_sync(dma, offsetof(struct aac_comm_space, sync_fib),
2950 	    AAC_FIB_SIZE, DDI_DMA_SYNC_FORCPU);
2951 
2952 	if ((aac_check_acc_handle(softs->pci_mem_handle) != DDI_SUCCESS) ||
2953 	    (aac_check_dma_handle(dma) != DDI_SUCCESS)) {
2954 		ddi_fm_service_impact(softs->devinfo_p, DDI_SERVICE_UNAFFECTED);
2955 		return (AACERR);
2956 	}
2957 
2958 	return (AACOK);
2959 }
2960 
2961 static void
2962 aac_cmd_initq(struct aac_cmd_queue *q)
2963 {
2964 	q->q_head = NULL;
2965 	q->q_tail = (struct aac_cmd *)&q->q_head;
2966 }
2967 
2968 /*
2969  * Remove a cmd from the head of q
2970  */
2971 static struct aac_cmd *
2972 aac_cmd_dequeue(struct aac_cmd_queue *q)
2973 {
2974 	struct aac_cmd *acp;
2975 
2976 	_NOTE(ASSUMING_PROTECTED(*q))
2977 
2978 	if ((acp = q->q_head) != NULL) {
2979 		if ((q->q_head = acp->next) != NULL)
2980 			acp->next = NULL;
2981 		else
2982 			q->q_tail = (struct aac_cmd *)&q->q_head;
2983 		acp->prev = NULL;
2984 	}
2985 	return (acp);
2986 }
2987 
2988 /*
2989  * Add a cmd to the tail of q
2990  */
2991 static void
2992 aac_cmd_enqueue(struct aac_cmd_queue *q, struct aac_cmd *acp)
2993 {
2994 	ASSERT(acp->next == NULL);
2995 	acp->prev = q->q_tail;
2996 	q->q_tail->next = acp;
2997 	q->q_tail = acp;
2998 }
2999 
3000 /*
3001  * Remove the cmd ac from q
3002  */
3003 static void
3004 aac_cmd_delete(struct aac_cmd_queue *q, struct aac_cmd *acp)
3005 {
3006 	if (acp->prev) {
3007 		if ((acp->prev->next = acp->next) != NULL) {
3008 			acp->next->prev = acp->prev;
3009 			acp->next = NULL;
3010 		} else {
3011 			q->q_tail = acp->prev;
3012 		}
3013 		acp->prev = NULL;
3014 	}
3015 	/* ac is not in the queue */
3016 }
3017 
3018 /*
3019  * Atomically insert an entry into the nominated queue, returns 0 on success or
3020  * AACERR if the queue is full.
3021  *
3022  * Note: it would be more efficient to defer notifying the controller in
3023  *	 the case where we may be inserting several entries in rapid succession,
3024  *	 but implementing this usefully may be difficult (it would involve a
3025  *	 separate queue/notify interface).
3026  */
3027 static int
3028 aac_fib_enqueue(struct aac_softstate *softs, int queue, uint32_t fib_addr,
3029     uint32_t fib_size)
3030 {
3031 	ddi_dma_handle_t dma = softs->comm_space_dma_handle;
3032 	ddi_acc_handle_t acc = softs->comm_space_acc_handle;
3033 	uint32_t pi, ci;
3034 
3035 	DBCALLED(softs, 2);
3036 
3037 	ASSERT(queue == AAC_ADAP_NORM_CMD_Q || queue == AAC_ADAP_NORM_RESP_Q);
3038 
3039 	/* Get the producer/consumer indices */
3040 	(void) ddi_dma_sync(dma, (uintptr_t)softs->qtablep->qt_qindex[queue] - \
3041 	    (uintptr_t)softs->comm_space, sizeof (uint32_t) * 2,
3042 	    DDI_DMA_SYNC_FORCPU);
3043 	if (aac_check_dma_handle(dma) != DDI_SUCCESS) {
3044 		ddi_fm_service_impact(softs->devinfo_p, DDI_SERVICE_UNAFFECTED);
3045 		return (AACERR);
3046 	}
3047 
3048 	pi = ddi_get32(acc,
3049 	    &softs->qtablep->qt_qindex[queue][AAC_PRODUCER_INDEX]);
3050 	ci = ddi_get32(acc,
3051 	    &softs->qtablep->qt_qindex[queue][AAC_CONSUMER_INDEX]);
3052 
3053 	/*
3054 	 * Wrap the queue first before we check the queue to see
3055 	 * if it is full
3056 	 */
3057 	if (pi >= aac_qinfo[queue].size)
3058 		pi = 0;
3059 
3060 	/* XXX queue full */
3061 	if ((pi + 1) == ci)
3062 		return (AACERR);
3063 
3064 	/* Fill in queue entry */
3065 	ddi_put32(acc, &((softs->qentries[queue] + pi)->aq_fib_size), fib_size);
3066 	ddi_put32(acc, &((softs->qentries[queue] + pi)->aq_fib_addr), fib_addr);
3067 	(void) ddi_dma_sync(dma, (uintptr_t)(softs->qentries[queue] + pi) - \
3068 	    (uintptr_t)softs->comm_space, sizeof (struct aac_queue_entry),
3069 	    DDI_DMA_SYNC_FORDEV);
3070 
3071 	/* Update producer index */
3072 	ddi_put32(acc, &softs->qtablep->qt_qindex[queue][AAC_PRODUCER_INDEX],
3073 	    pi + 1);
3074 	(void) ddi_dma_sync(dma,
3075 	    (uintptr_t)&softs->qtablep->qt_qindex[queue][AAC_PRODUCER_INDEX] - \
3076 	    (uintptr_t)softs->comm_space, sizeof (uint32_t),
3077 	    DDI_DMA_SYNC_FORDEV);
3078 
3079 	if (aac_qinfo[queue].notify != 0)
3080 		AAC_NOTIFY(softs, aac_qinfo[queue].notify);
3081 	return (AACOK);
3082 }
3083 
3084 /*
3085  * Atomically remove one entry from the nominated queue, returns 0 on
3086  * success or AACERR if the queue is empty.
3087  */
3088 static int
3089 aac_fib_dequeue(struct aac_softstate *softs, int queue, int *idxp)
3090 {
3091 	ddi_acc_handle_t acc = softs->comm_space_acc_handle;
3092 	ddi_dma_handle_t dma = softs->comm_space_dma_handle;
3093 	uint32_t pi, ci;
3094 	int unfull = 0;
3095 
3096 	DBCALLED(softs, 2);
3097 
3098 	ASSERT(idxp);
3099 
3100 	/* Get the producer/consumer indices */
3101 	(void) ddi_dma_sync(dma, (uintptr_t)softs->qtablep->qt_qindex[queue] - \
3102 	    (uintptr_t)softs->comm_space, sizeof (uint32_t) * 2,
3103 	    DDI_DMA_SYNC_FORCPU);
3104 	pi = ddi_get32(acc,
3105 	    &softs->qtablep->qt_qindex[queue][AAC_PRODUCER_INDEX]);
3106 	ci = ddi_get32(acc,
3107 	    &softs->qtablep->qt_qindex[queue][AAC_CONSUMER_INDEX]);
3108 
3109 	/* Check for queue empty */
3110 	if (ci == pi)
3111 		return (AACERR);
3112 
3113 	if (pi >= aac_qinfo[queue].size)
3114 		pi = 0;
3115 
3116 	/* Check for queue full */
3117 	if (ci == pi + 1)
3118 		unfull = 1;
3119 
3120 	/*
3121 	 * The controller does not wrap the queue,
3122 	 * so we have to do it by ourselves
3123 	 */
3124 	if (ci >= aac_qinfo[queue].size)
3125 		ci = 0;
3126 
3127 	/* Fetch the entry */
3128 	(void) ddi_dma_sync(dma, (uintptr_t)(softs->qentries[queue] + pi) - \
3129 	    (uintptr_t)softs->comm_space, sizeof (struct aac_queue_entry),
3130 	    DDI_DMA_SYNC_FORCPU);
3131 	if (aac_check_dma_handle(dma) != DDI_SUCCESS) {
3132 		ddi_fm_service_impact(softs->devinfo_p, DDI_SERVICE_UNAFFECTED);
3133 		return (AACERR);
3134 	}
3135 
3136 	switch (queue) {
3137 	case AAC_HOST_NORM_RESP_Q:
3138 	case AAC_HOST_HIGH_RESP_Q:
3139 		*idxp = ddi_get32(acc,
3140 		    &(softs->qentries[queue] + ci)->aq_fib_addr);
3141 		break;
3142 
3143 	case AAC_HOST_NORM_CMD_Q:
3144 	case AAC_HOST_HIGH_CMD_Q:
3145 		*idxp = ddi_get32(acc,
3146 		    &(softs->qentries[queue] + ci)->aq_fib_addr) / AAC_FIB_SIZE;
3147 		break;
3148 
3149 	default:
3150 		cmn_err(CE_NOTE, "!Invalid queue in aac_fib_dequeue()");
3151 		return (AACERR);
3152 	}
3153 
3154 	/* Update consumer index */
3155 	ddi_put32(acc, &softs->qtablep->qt_qindex[queue][AAC_CONSUMER_INDEX],
3156 	    ci + 1);
3157 	(void) ddi_dma_sync(dma,
3158 	    (uintptr_t)&softs->qtablep->qt_qindex[queue][AAC_CONSUMER_INDEX] - \
3159 	    (uintptr_t)softs->comm_space, sizeof (uint32_t),
3160 	    DDI_DMA_SYNC_FORDEV);
3161 
3162 	if (unfull && aac_qinfo[queue].notify != 0)
3163 		AAC_NOTIFY(softs, aac_qinfo[queue].notify);
3164 	return (AACOK);
3165 }
3166 
3167 /*
3168  * Request information of the container cid
3169  */
3170 static struct aac_mntinforesp *
3171 aac_get_container_info(struct aac_softstate *softs, int cid)
3172 {
3173 	ddi_acc_handle_t acc = softs->sync_slot.fib_acc_handle;
3174 	struct aac_fib *fibp = softs->sync_slot.fibp;
3175 	struct aac_mntinfo *mi = (struct aac_mntinfo *)&fibp->data[0];
3176 	struct aac_mntinforesp *mir;
3177 
3178 	ddi_put32(acc, &mi->Command, /* Use 64-bit LBA if enabled */
3179 	    (softs->flags & AAC_FLAGS_LBA_64BIT) ?
3180 	    VM_NameServe64 : VM_NameServe);
3181 	ddi_put32(acc, &mi->MntType, FT_FILESYS);
3182 	ddi_put32(acc, &mi->MntCount, cid);
3183 
3184 	if (aac_sync_fib(softs, ContainerCommand,
3185 	    AAC_FIB_SIZEOF(struct aac_mntinfo)) == AACERR) {
3186 		AACDB_PRINT(softs, CE_WARN, "Error probe container %d", cid);
3187 		return (NULL);
3188 	}
3189 
3190 	mir = (struct aac_mntinforesp *)&fibp->data[0];
3191 	if (ddi_get32(acc, &mir->Status) == ST_OK)
3192 		return (mir);
3193 	return (NULL);
3194 }
3195 
3196 static int
3197 aac_get_container_count(struct aac_softstate *softs, int *count)
3198 {
3199 	ddi_acc_handle_t acc = softs->sync_slot.fib_acc_handle;
3200 	struct aac_mntinforesp *mir;
3201 
3202 	if ((mir = aac_get_container_info(softs, 0)) == NULL)
3203 		return (AACERR);
3204 	*count = ddi_get32(acc, &mir->MntRespCount);
3205 	if (*count > AAC_MAX_LD) {
3206 		AACDB_PRINT(softs, CE_CONT,
3207 		    "container count(%d) > AAC_MAX_LD", *count);
3208 		return (AACERR);
3209 	}
3210 	return (AACOK);
3211 }
3212 
3213 static int
3214 aac_get_container_uid(struct aac_softstate *softs, uint32_t cid, uint32_t *uid)
3215 {
3216 	ddi_acc_handle_t acc = softs->sync_slot.fib_acc_handle;
3217 	struct aac_Container *ct = (struct aac_Container *) \
3218 	    &softs->sync_slot.fibp->data[0];
3219 
3220 	bzero(ct, sizeof (*ct) - CT_PACKET_SIZE);
3221 	ddi_put32(acc, &ct->Command, VM_ContainerConfig);
3222 	ddi_put32(acc, &ct->CTCommand.command, CT_CID_TO_32BITS_UID);
3223 	ddi_put32(acc, &ct->CTCommand.param[0], cid);
3224 
3225 	if (aac_sync_fib(softs, ContainerCommand,
3226 	    AAC_FIB_SIZEOF(struct aac_Container)) == AACERR)
3227 		return (AACERR);
3228 	if (ddi_get32(acc, &ct->CTCommand.param[0]) != CT_OK)
3229 		return (AACERR);
3230 
3231 	*uid = ddi_get32(acc, &ct->CTCommand.param[1]);
3232 	return (AACOK);
3233 }
3234 
3235 static int
3236 aac_probe_container(struct aac_softstate *softs, uint32_t cid)
3237 {
3238 	struct aac_container *dvp = &softs->containers[cid];
3239 	ddi_acc_handle_t acc = softs->sync_slot.fib_acc_handle;
3240 	struct aac_mntinforesp *mir;
3241 	uint64_t size;
3242 	uint32_t uid;
3243 
3244 	/* Get container basic info */
3245 	if ((mir = aac_get_container_info(softs, cid)) == NULL)
3246 		return (AACERR);
3247 
3248 	if (ddi_get32(acc, &mir->MntObj.VolType) == CT_NONE) {
3249 		if (AAC_DEV_IS_VALID(&dvp->dev)) {
3250 			AACDB_PRINT(softs, CE_NOTE,
3251 			    ">>> Container %d deleted", cid);
3252 			dvp->dev.flags &= ~AAC_DFLAG_VALID;
3253 			(void) aac_dr_event(softs, dvp->cid, -1,
3254 			    AAC_EVT_OFFLINE);
3255 		}
3256 	} else {
3257 		size = AAC_MIR_SIZE(softs, acc, mir);
3258 
3259 		/* Get container UID */
3260 		if (aac_get_container_uid(softs, cid, &uid) == AACERR) {
3261 			AACDB_PRINT(softs, CE_CONT,
3262 			    "query container %d uid failed", cid);
3263 			return (AACERR);
3264 		}
3265 		AACDB_PRINT(softs, CE_CONT, "uid=0x%08x", uid);
3266 
3267 		if (AAC_DEV_IS_VALID(&dvp->dev)) {
3268 			if (dvp->uid != uid) {
3269 				AACDB_PRINT(softs, CE_WARN,
3270 				    ">>> Container %u uid changed to %d",
3271 				    cid, uid);
3272 				dvp->uid = uid;
3273 			}
3274 			if (dvp->size != size) {
3275 				AACDB_PRINT(softs, CE_NOTE,
3276 				    ">>> Container %u size changed to %"PRIu64,
3277 				    cid, size);
3278 				dvp->size = size;
3279 			}
3280 		} else { /* Init new container */
3281 			AACDB_PRINT(softs, CE_NOTE,
3282 			    ">>> Container %d added: " \
3283 			    "size=0x%x.%08x, type=%d, name=%s",
3284 			    cid,
3285 			    ddi_get32(acc, &mir->MntObj.CapacityHigh),
3286 			    ddi_get32(acc, &mir->MntObj.Capacity),
3287 			    ddi_get32(acc, &mir->MntObj.VolType),
3288 			    mir->MntObj.FileSystemName);
3289 			dvp->dev.flags |= AAC_DFLAG_VALID;
3290 			dvp->dev.type = AAC_DEV_LD;
3291 
3292 			dvp->cid = cid;
3293 			dvp->uid = uid;
3294 			dvp->size = size;
3295 			dvp->locked = 0;
3296 			dvp->deleted = 0;
3297 			(void) aac_dr_event(softs, dvp->cid, -1,
3298 			    AAC_EVT_ONLINE);
3299 		}
3300 	}
3301 	return (AACOK);
3302 }
3303 
3304 /*
3305  * Do a rescan of all the possible containers and update the container list
3306  * with newly online/offline containers, and prepare for autoconfiguration.
3307  */
3308 static int
3309 aac_probe_containers(struct aac_softstate *softs)
3310 {
3311 	int i, count, total;
3312 
3313 	/* Loop over possible containers */
3314 	count = softs->container_count;
3315 	if (aac_get_container_count(softs, &count) == AACERR)
3316 		return (AACERR);
3317 	for (i = total = 0; i < count; i++) {
3318 		if (aac_probe_container(softs, i) == AACOK)
3319 			total++;
3320 	}
3321 	if (count < softs->container_count) {
3322 		struct aac_container *dvp;
3323 
3324 		for (dvp = &softs->containers[count];
3325 		    dvp < &softs->containers[softs->container_count]; dvp++) {
3326 			if (!AAC_DEV_IS_VALID(&dvp->dev))
3327 				continue;
3328 			AACDB_PRINT(softs, CE_NOTE, ">>> Container %d deleted",
3329 			    dvp->cid);
3330 			dvp->dev.flags &= ~AAC_DFLAG_VALID;
3331 			(void) aac_dr_event(softs, dvp->cid, -1,
3332 			    AAC_EVT_OFFLINE);
3333 		}
3334 	}
3335 	softs->container_count = count;
3336 	AACDB_PRINT(softs, CE_CONT, "?Total %d container(s) found", total);
3337 	return (AACOK);
3338 }
3339 
3340 static int
3341 aac_alloc_comm_space(struct aac_softstate *softs)
3342 {
3343 	size_t rlen;
3344 	ddi_dma_cookie_t cookie;
3345 	uint_t cookien;
3346 
3347 	/* Allocate DMA for comm. space */
3348 	if (ddi_dma_alloc_handle(
3349 	    softs->devinfo_p,
3350 	    &softs->addr_dma_attr,
3351 	    DDI_DMA_SLEEP,
3352 	    NULL,
3353 	    &softs->comm_space_dma_handle) != DDI_SUCCESS) {
3354 		AACDB_PRINT(softs, CE_WARN,
3355 		    "Cannot alloc dma handle for communication area");
3356 		goto error;
3357 	}
3358 	if (ddi_dma_mem_alloc(
3359 	    softs->comm_space_dma_handle,
3360 	    sizeof (struct aac_comm_space),
3361 	    &softs->acc_attr,
3362 	    DDI_DMA_RDWR | DDI_DMA_CONSISTENT,
3363 	    DDI_DMA_SLEEP,
3364 	    NULL,
3365 	    (caddr_t *)&softs->comm_space,
3366 	    &rlen,
3367 	    &softs->comm_space_acc_handle) != DDI_SUCCESS) {
3368 		AACDB_PRINT(softs, CE_WARN,
3369 		    "Cannot alloc mem for communication area");
3370 		goto error;
3371 	}
3372 	if (ddi_dma_addr_bind_handle(
3373 	    softs->comm_space_dma_handle,
3374 	    NULL,
3375 	    (caddr_t)softs->comm_space,
3376 	    sizeof (struct aac_comm_space),
3377 	    DDI_DMA_RDWR | DDI_DMA_CONSISTENT,
3378 	    DDI_DMA_SLEEP,
3379 	    NULL,
3380 	    &cookie,
3381 	    &cookien) != DDI_DMA_MAPPED) {
3382 		AACDB_PRINT(softs, CE_WARN,
3383 		    "DMA bind failed for communication area");
3384 		goto error;
3385 	}
3386 	softs->comm_space_phyaddr = cookie.dmac_address;
3387 
3388 	/* Setup sync FIB space */
3389 	softs->sync_slot.fibp = &softs->comm_space->sync_fib;
3390 	softs->sync_slot.fib_phyaddr = softs->comm_space_phyaddr + \
3391 	    offsetof(struct aac_comm_space, sync_fib);
3392 	softs->sync_slot.fib_acc_handle = softs->comm_space_acc_handle;
3393 	softs->sync_slot.fib_dma_handle = softs->comm_space_dma_handle;
3394 
3395 	return (AACOK);
3396 error:
3397 	if (softs->comm_space_acc_handle) {
3398 		ddi_dma_mem_free(&softs->comm_space_acc_handle);
3399 		softs->comm_space_acc_handle = NULL;
3400 	}
3401 	if (softs->comm_space_dma_handle) {
3402 		ddi_dma_free_handle(&softs->comm_space_dma_handle);
3403 		softs->comm_space_dma_handle = NULL;
3404 	}
3405 	return (AACERR);
3406 }
3407 
3408 static void
3409 aac_free_comm_space(struct aac_softstate *softs)
3410 {
3411 	softs->sync_slot.fibp = NULL;
3412 	softs->sync_slot.fib_phyaddr = NULL;
3413 	softs->sync_slot.fib_acc_handle = NULL;
3414 	softs->sync_slot.fib_dma_handle = NULL;
3415 
3416 	(void) ddi_dma_unbind_handle(softs->comm_space_dma_handle);
3417 	ddi_dma_mem_free(&softs->comm_space_acc_handle);
3418 	softs->comm_space_acc_handle = NULL;
3419 	ddi_dma_free_handle(&softs->comm_space_dma_handle);
3420 	softs->comm_space_dma_handle = NULL;
3421 	softs->comm_space_phyaddr = NULL;
3422 }
3423 
3424 /*
3425  * Initialize the data structures that are required for the communication
3426  * interface to operate
3427  */
3428 static int
3429 aac_setup_comm_space(struct aac_softstate *softs)
3430 {
3431 	ddi_dma_handle_t dma = softs->comm_space_dma_handle;
3432 	ddi_acc_handle_t acc = softs->comm_space_acc_handle;
3433 	uint32_t comm_space_phyaddr;
3434 	struct aac_adapter_init *initp;
3435 	int qoffset;
3436 
3437 	comm_space_phyaddr = softs->comm_space_phyaddr;
3438 
3439 	/* Setup adapter init struct */
3440 	initp = &softs->comm_space->init_data;
3441 	bzero(initp, sizeof (struct aac_adapter_init));
3442 
3443 	ddi_put32(acc, &initp->InitStructRevision, AAC_INIT_STRUCT_REVISION);
3444 	ddi_put32(acc, &initp->HostElapsedSeconds, ddi_get_time());
3445 
3446 	/* Setup new/old comm. specific data */
3447 	if (softs->flags & AAC_FLAGS_RAW_IO) {
3448 		ddi_put32(acc, &initp->InitStructRevision,
3449 		    AAC_INIT_STRUCT_REVISION_4);
3450 		ddi_put32(acc, &initp->InitFlags,
3451 		    (softs->flags & AAC_FLAGS_NEW_COMM) ?
3452 		    AAC_INIT_FLAGS_NEW_COMM_SUPPORTED : 0);
3453 		/* Setup the preferred settings */
3454 		ddi_put32(acc, &initp->MaxIoCommands, softs->aac_max_fibs);
3455 		ddi_put32(acc, &initp->MaxIoSize,
3456 		    (softs->aac_max_sectors << 9));
3457 		ddi_put32(acc, &initp->MaxFibSize, softs->aac_max_fib_size);
3458 	} else {
3459 		/*
3460 		 * Tells the adapter about the physical location of various
3461 		 * important shared data structures
3462 		 */
3463 		ddi_put32(acc, &initp->AdapterFibsPhysicalAddress,
3464 		    comm_space_phyaddr + \
3465 		    offsetof(struct aac_comm_space, adapter_fibs));
3466 		ddi_put32(acc, &initp->AdapterFibsVirtualAddress, 0);
3467 		ddi_put32(acc, &initp->AdapterFibAlign, AAC_FIB_SIZE);
3468 		ddi_put32(acc, &initp->AdapterFibsSize,
3469 		    AAC_ADAPTER_FIBS * AAC_FIB_SIZE);
3470 		ddi_put32(acc, &initp->PrintfBufferAddress,
3471 		    comm_space_phyaddr + \
3472 		    offsetof(struct aac_comm_space, adapter_print_buf));
3473 		ddi_put32(acc, &initp->PrintfBufferSize,
3474 		    AAC_ADAPTER_PRINT_BUFSIZE);
3475 		ddi_put32(acc, &initp->MiniPortRevision,
3476 		    AAC_INIT_STRUCT_MINIPORT_REVISION);
3477 		ddi_put32(acc, &initp->HostPhysMemPages, AAC_MAX_PFN);
3478 
3479 		qoffset = (comm_space_phyaddr + \
3480 		    offsetof(struct aac_comm_space, qtable)) % \
3481 		    AAC_QUEUE_ALIGN;
3482 		if (qoffset)
3483 			qoffset = AAC_QUEUE_ALIGN - qoffset;
3484 		softs->qtablep = (struct aac_queue_table *) \
3485 		    ((char *)&softs->comm_space->qtable + qoffset);
3486 		ddi_put32(acc, &initp->CommHeaderAddress, comm_space_phyaddr + \
3487 		    offsetof(struct aac_comm_space, qtable) + qoffset);
3488 
3489 		/* Init queue table */
3490 		ddi_put32(acc, &softs->qtablep-> \
3491 		    qt_qindex[AAC_HOST_NORM_CMD_Q][AAC_PRODUCER_INDEX],
3492 		    AAC_HOST_NORM_CMD_ENTRIES);
3493 		ddi_put32(acc, &softs->qtablep-> \
3494 		    qt_qindex[AAC_HOST_NORM_CMD_Q][AAC_CONSUMER_INDEX],
3495 		    AAC_HOST_NORM_CMD_ENTRIES);
3496 		ddi_put32(acc, &softs->qtablep-> \
3497 		    qt_qindex[AAC_HOST_HIGH_CMD_Q][AAC_PRODUCER_INDEX],
3498 		    AAC_HOST_HIGH_CMD_ENTRIES);
3499 		ddi_put32(acc, &softs->qtablep-> \
3500 		    qt_qindex[AAC_HOST_HIGH_CMD_Q][AAC_CONSUMER_INDEX],
3501 		    AAC_HOST_HIGH_CMD_ENTRIES);
3502 		ddi_put32(acc, &softs->qtablep-> \
3503 		    qt_qindex[AAC_ADAP_NORM_CMD_Q][AAC_PRODUCER_INDEX],
3504 		    AAC_ADAP_NORM_CMD_ENTRIES);
3505 		ddi_put32(acc, &softs->qtablep-> \
3506 		    qt_qindex[AAC_ADAP_NORM_CMD_Q][AAC_CONSUMER_INDEX],
3507 		    AAC_ADAP_NORM_CMD_ENTRIES);
3508 		ddi_put32(acc, &softs->qtablep-> \
3509 		    qt_qindex[AAC_ADAP_HIGH_CMD_Q][AAC_PRODUCER_INDEX],
3510 		    AAC_ADAP_HIGH_CMD_ENTRIES);
3511 		ddi_put32(acc, &softs->qtablep-> \
3512 		    qt_qindex[AAC_ADAP_HIGH_CMD_Q][AAC_CONSUMER_INDEX],
3513 		    AAC_ADAP_HIGH_CMD_ENTRIES);
3514 		ddi_put32(acc, &softs->qtablep-> \
3515 		    qt_qindex[AAC_HOST_NORM_RESP_Q][AAC_PRODUCER_INDEX],
3516 		    AAC_HOST_NORM_RESP_ENTRIES);
3517 		ddi_put32(acc, &softs->qtablep-> \
3518 		    qt_qindex[AAC_HOST_NORM_RESP_Q][AAC_CONSUMER_INDEX],
3519 		    AAC_HOST_NORM_RESP_ENTRIES);
3520 		ddi_put32(acc, &softs->qtablep-> \
3521 		    qt_qindex[AAC_HOST_HIGH_RESP_Q][AAC_PRODUCER_INDEX],
3522 		    AAC_HOST_HIGH_RESP_ENTRIES);
3523 		ddi_put32(acc, &softs->qtablep-> \
3524 		    qt_qindex[AAC_HOST_HIGH_RESP_Q][AAC_CONSUMER_INDEX],
3525 		    AAC_HOST_HIGH_RESP_ENTRIES);
3526 		ddi_put32(acc, &softs->qtablep-> \
3527 		    qt_qindex[AAC_ADAP_NORM_RESP_Q][AAC_PRODUCER_INDEX],
3528 		    AAC_ADAP_NORM_RESP_ENTRIES);
3529 		ddi_put32(acc, &softs->qtablep-> \
3530 		    qt_qindex[AAC_ADAP_NORM_RESP_Q][AAC_CONSUMER_INDEX],
3531 		    AAC_ADAP_NORM_RESP_ENTRIES);
3532 		ddi_put32(acc, &softs->qtablep-> \
3533 		    qt_qindex[AAC_ADAP_HIGH_RESP_Q][AAC_PRODUCER_INDEX],
3534 		    AAC_ADAP_HIGH_RESP_ENTRIES);
3535 		ddi_put32(acc, &softs->qtablep-> \
3536 		    qt_qindex[AAC_ADAP_HIGH_RESP_Q][AAC_CONSUMER_INDEX],
3537 		    AAC_ADAP_HIGH_RESP_ENTRIES);
3538 
3539 		/* Init queue entries */
3540 		softs->qentries[AAC_HOST_NORM_CMD_Q] =
3541 		    &softs->qtablep->qt_HostNormCmdQueue[0];
3542 		softs->qentries[AAC_HOST_HIGH_CMD_Q] =
3543 		    &softs->qtablep->qt_HostHighCmdQueue[0];
3544 		softs->qentries[AAC_ADAP_NORM_CMD_Q] =
3545 		    &softs->qtablep->qt_AdapNormCmdQueue[0];
3546 		softs->qentries[AAC_ADAP_HIGH_CMD_Q] =
3547 		    &softs->qtablep->qt_AdapHighCmdQueue[0];
3548 		softs->qentries[AAC_HOST_NORM_RESP_Q] =
3549 		    &softs->qtablep->qt_HostNormRespQueue[0];
3550 		softs->qentries[AAC_HOST_HIGH_RESP_Q] =
3551 		    &softs->qtablep->qt_HostHighRespQueue[0];
3552 		softs->qentries[AAC_ADAP_NORM_RESP_Q] =
3553 		    &softs->qtablep->qt_AdapNormRespQueue[0];
3554 		softs->qentries[AAC_ADAP_HIGH_RESP_Q] =
3555 		    &softs->qtablep->qt_AdapHighRespQueue[0];
3556 	}
3557 	(void) ddi_dma_sync(dma, 0, 0, DDI_DMA_SYNC_FORDEV);
3558 
3559 	/* Send init structure to the card */
3560 	if (aac_sync_mbcommand(softs, AAC_MONKER_INITSTRUCT,
3561 	    comm_space_phyaddr + \
3562 	    offsetof(struct aac_comm_space, init_data),
3563 	    0, 0, 0, NULL) == AACERR) {
3564 		AACDB_PRINT(softs, CE_WARN,
3565 		    "Cannot send init structure to adapter");
3566 		return (AACERR);
3567 	}
3568 
3569 	return (AACOK);
3570 }
3571 
3572 static uchar_t *
3573 aac_vendor_id(struct aac_softstate *softs, uchar_t *buf)
3574 {
3575 	(void) memset(buf, ' ', AAC_VENDOR_LEN);
3576 	bcopy(softs->vendor_name, buf, strlen(softs->vendor_name));
3577 	return (buf + AAC_VENDOR_LEN);
3578 }
3579 
3580 static uchar_t *
3581 aac_product_id(struct aac_softstate *softs, uchar_t *buf)
3582 {
3583 	(void) memset(buf, ' ', AAC_PRODUCT_LEN);
3584 	bcopy(softs->product_name, buf, strlen(softs->product_name));
3585 	return (buf + AAC_PRODUCT_LEN);
3586 }
3587 
3588 /*
3589  * Construct unit serial number from container uid
3590  */
3591 static uchar_t *
3592 aac_lun_serialno(struct aac_softstate *softs, int tgt, uchar_t *buf)
3593 {
3594 	int i, d;
3595 	uint32_t uid;
3596 
3597 	ASSERT(tgt >= 0 && tgt < AAC_MAX_LD);
3598 
3599 	uid = softs->containers[tgt].uid;
3600 	for (i = 7; i >= 0; i--) {
3601 		d = uid & 0xf;
3602 		buf[i] = d > 9 ? 'A' + (d - 0xa) : '0' + d;
3603 		uid >>= 4;
3604 	}
3605 	return (buf + 8);
3606 }
3607 
3608 /*
3609  * SPC-3 7.5 INQUIRY command implementation
3610  */
3611 static void
3612 aac_inquiry(struct aac_softstate *softs, struct scsi_pkt *pkt,
3613     union scsi_cdb *cdbp, struct buf *bp)
3614 {
3615 	int tgt = pkt->pkt_address.a_target;
3616 	char *b_addr = NULL;
3617 	uchar_t page = cdbp->cdb_opaque[2];
3618 
3619 	if (cdbp->cdb_opaque[1] & AAC_CDB_INQUIRY_CMDDT) {
3620 		/* Command Support Data is not supported */
3621 		aac_set_arq_data(pkt, KEY_ILLEGAL_REQUEST, 0x24, 0x00, 0);
3622 		return;
3623 	}
3624 
3625 	if (bp && bp->b_un.b_addr && bp->b_bcount) {
3626 		if (bp->b_flags & (B_PHYS | B_PAGEIO))
3627 			bp_mapin(bp);
3628 		b_addr = bp->b_un.b_addr;
3629 	}
3630 
3631 	if (cdbp->cdb_opaque[1] & AAC_CDB_INQUIRY_EVPD) {
3632 		uchar_t *vpdp = (uchar_t *)b_addr;
3633 		uchar_t *idp, *sp;
3634 
3635 		/* SPC-3 8.4 Vital product data parameters */
3636 		switch (page) {
3637 		case 0x00:
3638 			/* Supported VPD pages */
3639 			if (vpdp == NULL ||
3640 			    bp->b_bcount < (AAC_VPD_PAGE_DATA + 3))
3641 				return;
3642 			bzero(vpdp, AAC_VPD_PAGE_LENGTH);
3643 			vpdp[AAC_VPD_PAGE_CODE] = 0x00;
3644 			vpdp[AAC_VPD_PAGE_LENGTH] = 3;
3645 
3646 			vpdp[AAC_VPD_PAGE_DATA] = 0x00;
3647 			vpdp[AAC_VPD_PAGE_DATA + 1] = 0x80;
3648 			vpdp[AAC_VPD_PAGE_DATA + 2] = 0x83;
3649 
3650 			pkt->pkt_state |= STATE_XFERRED_DATA;
3651 			break;
3652 
3653 		case 0x80:
3654 			/* Unit serial number page */
3655 			if (vpdp == NULL ||
3656 			    bp->b_bcount < (AAC_VPD_PAGE_DATA + 8))
3657 				return;
3658 			bzero(vpdp, AAC_VPD_PAGE_LENGTH);
3659 			vpdp[AAC_VPD_PAGE_CODE] = 0x80;
3660 			vpdp[AAC_VPD_PAGE_LENGTH] = 8;
3661 
3662 			sp = &vpdp[AAC_VPD_PAGE_DATA];
3663 			(void) aac_lun_serialno(softs, tgt, sp);
3664 
3665 			pkt->pkt_state |= STATE_XFERRED_DATA;
3666 			break;
3667 
3668 		case 0x83:
3669 			/* Device identification page */
3670 			if (vpdp == NULL ||
3671 			    bp->b_bcount < (AAC_VPD_PAGE_DATA + 32))
3672 				return;
3673 			bzero(vpdp, AAC_VPD_PAGE_LENGTH);
3674 			vpdp[AAC_VPD_PAGE_CODE] = 0x83;
3675 
3676 			idp = &vpdp[AAC_VPD_PAGE_DATA];
3677 			bzero(idp, AAC_VPD_ID_LENGTH);
3678 			idp[AAC_VPD_ID_CODESET] = 0x02;
3679 			idp[AAC_VPD_ID_TYPE] = 0x01;
3680 
3681 			/*
3682 			 * SPC-3 Table 111 - Identifier type
3683 			 * One recommanded method of constructing the remainder
3684 			 * of identifier field is to concatenate the product
3685 			 * identification field from the standard INQUIRY data
3686 			 * field and the product serial number field from the
3687 			 * unit serial number page.
3688 			 */
3689 			sp = &idp[AAC_VPD_ID_DATA];
3690 			sp = aac_vendor_id(softs, sp);
3691 			sp = aac_product_id(softs, sp);
3692 			sp = aac_lun_serialno(softs, tgt, sp);
3693 			idp[AAC_VPD_ID_LENGTH] = (uintptr_t)sp - \
3694 			    (uintptr_t)&idp[AAC_VPD_ID_DATA];
3695 
3696 			vpdp[AAC_VPD_PAGE_LENGTH] = (uintptr_t)sp - \
3697 			    (uintptr_t)&vpdp[AAC_VPD_PAGE_DATA];
3698 			pkt->pkt_state |= STATE_XFERRED_DATA;
3699 			break;
3700 
3701 		default:
3702 			aac_set_arq_data(pkt, KEY_ILLEGAL_REQUEST,
3703 			    0x24, 0x00, 0);
3704 			break;
3705 		}
3706 	} else {
3707 		struct scsi_inquiry *inqp = (struct scsi_inquiry *)b_addr;
3708 		size_t len = sizeof (struct scsi_inquiry);
3709 
3710 		if (page != 0) {
3711 			aac_set_arq_data(pkt, KEY_ILLEGAL_REQUEST,
3712 			    0x24, 0x00, 0);
3713 			return;
3714 		}
3715 		if (inqp == NULL || bp->b_bcount < len)
3716 			return;
3717 
3718 		bzero(inqp, len);
3719 		inqp->inq_len = AAC_ADDITIONAL_LEN;
3720 		inqp->inq_ansi = AAC_ANSI_VER;
3721 		inqp->inq_rdf = AAC_RESP_DATA_FORMAT;
3722 		(void) aac_vendor_id(softs, (uchar_t *)inqp->inq_vid);
3723 		(void) aac_product_id(softs, (uchar_t *)inqp->inq_pid);
3724 		bcopy("V1.0", inqp->inq_revision, 4);
3725 		inqp->inq_cmdque = 1; /* enable tagged-queuing */
3726 		/*
3727 		 * For "sd-max-xfer-size" property which may impact performance
3728 		 * when IO threads increase.
3729 		 */
3730 		inqp->inq_wbus32 = 1;
3731 
3732 		pkt->pkt_state |= STATE_XFERRED_DATA;
3733 	}
3734 }
3735 
3736 /*
3737  * SPC-3 7.10 MODE SENSE command implementation
3738  */
3739 static void
3740 aac_mode_sense(struct aac_softstate *softs, struct scsi_pkt *pkt,
3741     union scsi_cdb *cdbp, struct buf *bp, int capacity)
3742 {
3743 	uchar_t pagecode;
3744 	struct mode_header *headerp;
3745 	struct mode_header_g1 *g1_headerp;
3746 	unsigned int ncyl;
3747 	caddr_t sense_data;
3748 	caddr_t next_page;
3749 	size_t sdata_size;
3750 	size_t pages_size;
3751 	int unsupport_page = 0;
3752 
3753 	ASSERT(cdbp->scc_cmd == SCMD_MODE_SENSE ||
3754 	    cdbp->scc_cmd == SCMD_MODE_SENSE_G1);
3755 
3756 	if (!(bp && bp->b_un.b_addr && bp->b_bcount))
3757 		return;
3758 
3759 	if (bp->b_flags & (B_PHYS | B_PAGEIO))
3760 		bp_mapin(bp);
3761 	pkt->pkt_state |= STATE_XFERRED_DATA;
3762 	pagecode = cdbp->cdb_un.sg.scsi[0] & 0x3F;
3763 
3764 	/* calculate the size of needed buffer */
3765 	if (cdbp->scc_cmd == SCMD_MODE_SENSE)
3766 		sdata_size = MODE_HEADER_LENGTH;
3767 	else /* must be SCMD_MODE_SENSE_G1 */
3768 		sdata_size = MODE_HEADER_LENGTH_G1;
3769 
3770 	pages_size = 0;
3771 	switch (pagecode) {
3772 	case SD_MODE_SENSE_PAGE3_CODE:
3773 		pages_size += sizeof (struct mode_format);
3774 		break;
3775 
3776 	case SD_MODE_SENSE_PAGE4_CODE:
3777 		pages_size += sizeof (struct mode_geometry);
3778 		break;
3779 
3780 	case MODEPAGE_CTRL_MODE:
3781 		if (softs->flags & AAC_FLAGS_LBA_64BIT) {
3782 			pages_size += sizeof (struct mode_control_scsi3);
3783 		} else {
3784 			unsupport_page = 1;
3785 		}
3786 		break;
3787 
3788 	case MODEPAGE_ALLPAGES:
3789 		if (softs->flags & AAC_FLAGS_LBA_64BIT) {
3790 			pages_size += sizeof (struct mode_format) +
3791 			    sizeof (struct mode_geometry) +
3792 			    sizeof (struct mode_control_scsi3);
3793 		} else {
3794 			pages_size += sizeof (struct mode_format) +
3795 			    sizeof (struct mode_geometry);
3796 		}
3797 		break;
3798 
3799 	default:
3800 		/* unsupported pages */
3801 		unsupport_page = 1;
3802 	}
3803 
3804 	/* allocate buffer to fill the send data */
3805 	sdata_size += pages_size;
3806 	sense_data = kmem_zalloc(sdata_size, KM_SLEEP);
3807 
3808 	if (cdbp->scc_cmd == SCMD_MODE_SENSE) {
3809 		headerp = (struct mode_header *)sense_data;
3810 		headerp->length = MODE_HEADER_LENGTH + pages_size -
3811 		    sizeof (headerp->length);
3812 		headerp->bdesc_length = 0;
3813 		next_page = sense_data + sizeof (struct mode_header);
3814 	} else {
3815 		g1_headerp = (void *)sense_data;
3816 		g1_headerp->length = BE_16(MODE_HEADER_LENGTH_G1 + pages_size -
3817 		    sizeof (g1_headerp->length));
3818 		g1_headerp->bdesc_length = 0;
3819 		next_page = sense_data + sizeof (struct mode_header_g1);
3820 	}
3821 
3822 	if (unsupport_page)
3823 		goto finish;
3824 
3825 	if (pagecode == SD_MODE_SENSE_PAGE3_CODE ||
3826 	    pagecode == MODEPAGE_ALLPAGES) {
3827 		/* SBC-3 7.1.3.3 Format device page */
3828 		struct mode_format *page3p;
3829 
3830 		page3p = (void *)next_page;
3831 		page3p->mode_page.code = SD_MODE_SENSE_PAGE3_CODE;
3832 		page3p->mode_page.length = sizeof (struct mode_format);
3833 		page3p->data_bytes_sect = BE_16(AAC_SECTOR_SIZE);
3834 		page3p->sect_track = BE_16(AAC_SECTORS_PER_TRACK);
3835 
3836 		next_page += sizeof (struct mode_format);
3837 	}
3838 
3839 	if (pagecode == SD_MODE_SENSE_PAGE4_CODE ||
3840 	    pagecode == MODEPAGE_ALLPAGES) {
3841 		/* SBC-3 7.1.3.8 Rigid disk device geometry page */
3842 		struct mode_geometry *page4p;
3843 
3844 		page4p = (void *)next_page;
3845 		page4p->mode_page.code = SD_MODE_SENSE_PAGE4_CODE;
3846 		page4p->mode_page.length = sizeof (struct mode_geometry);
3847 		page4p->heads = AAC_NUMBER_OF_HEADS;
3848 		page4p->rpm = BE_16(AAC_ROTATION_SPEED);
3849 		ncyl = capacity / (AAC_NUMBER_OF_HEADS * AAC_SECTORS_PER_TRACK);
3850 		page4p->cyl_lb = ncyl & 0xff;
3851 		page4p->cyl_mb = (ncyl >> 8) & 0xff;
3852 		page4p->cyl_ub = (ncyl >> 16) & 0xff;
3853 
3854 		next_page += sizeof (struct mode_geometry);
3855 	}
3856 
3857 	if ((pagecode == MODEPAGE_CTRL_MODE || pagecode == MODEPAGE_ALLPAGES) &&
3858 	    softs->flags & AAC_FLAGS_LBA_64BIT) {
3859 		/* 64-bit LBA need large sense data */
3860 		struct mode_control_scsi3 *mctl;
3861 
3862 		mctl = (void *)next_page;
3863 		mctl->mode_page.code = MODEPAGE_CTRL_MODE;
3864 		mctl->mode_page.length =
3865 		    sizeof (struct mode_control_scsi3) -
3866 		    sizeof (struct mode_page);
3867 		mctl->d_sense = 1;
3868 	}
3869 
3870 finish:
3871 	/* copyout the valid data. */
3872 	bcopy(sense_data, bp->b_un.b_addr, min(sdata_size, bp->b_bcount));
3873 	kmem_free(sense_data, sdata_size);
3874 }
3875 
3876 static int
3877 aac_name_node(dev_info_t *dip, char *name, int len)
3878 {
3879 	int tgt, lun;
3880 
3881 	tgt = ddi_prop_get_int(DDI_DEV_T_ANY, dip,
3882 	    DDI_PROP_DONTPASS, "target", -1);
3883 	if (tgt == -1)
3884 		return (DDI_FAILURE);
3885 	lun = ddi_prop_get_int(DDI_DEV_T_ANY, dip,
3886 	    DDI_PROP_DONTPASS, "lun", -1);
3887 	if (lun == -1)
3888 		return (DDI_FAILURE);
3889 
3890 	(void) snprintf(name, len, "%x,%x", tgt, lun);
3891 	return (DDI_SUCCESS);
3892 }
3893 
3894 /*ARGSUSED*/
3895 static int
3896 aac_tran_tgt_init(dev_info_t *hba_dip, dev_info_t *tgt_dip,
3897     scsi_hba_tran_t *tran, struct scsi_device *sd)
3898 {
3899 	struct aac_softstate *softs = AAC_TRAN2SOFTS(tran);
3900 #if defined(DEBUG) || defined(__lock_lint)
3901 	int ctl = ddi_get_instance(softs->devinfo_p);
3902 #endif
3903 	uint16_t tgt = sd->sd_address.a_target;
3904 	uint8_t lun = sd->sd_address.a_lun;
3905 	struct aac_device *dvp;
3906 
3907 	DBCALLED(softs, 2);
3908 
3909 	if (ndi_dev_is_persistent_node(tgt_dip) == 0) {
3910 		/*
3911 		 * If no persistent node exist, we don't allow .conf node
3912 		 * to be created.
3913 		 */
3914 		if (aac_find_child(softs, tgt, lun) != NULL) {
3915 			if (ndi_merge_node(tgt_dip, aac_name_node) !=
3916 			    DDI_SUCCESS)
3917 				/* Create this .conf node */
3918 				return (DDI_SUCCESS);
3919 		}
3920 		return (DDI_FAILURE);
3921 	}
3922 
3923 	/*
3924 	 * Only support container/phys. device that has been
3925 	 * detected and valid
3926 	 */
3927 	mutex_enter(&softs->io_lock);
3928 	if (tgt >= AAC_MAX_DEV(softs)) {
3929 		AACDB_PRINT_TRAN(softs,
3930 		    "aac_tran_tgt_init: c%dt%dL%d out", ctl, tgt, lun);
3931 		mutex_exit(&softs->io_lock);
3932 		return (DDI_FAILURE);
3933 	}
3934 
3935 	if (tgt < AAC_MAX_LD) {
3936 		dvp = (struct aac_device *)&softs->containers[tgt];
3937 		if (lun != 0 || !AAC_DEV_IS_VALID(dvp)) {
3938 			AACDB_PRINT_TRAN(softs, "aac_tran_tgt_init: c%dt%dL%d",
3939 			    ctl, tgt, lun);
3940 			mutex_exit(&softs->io_lock);
3941 			return (DDI_FAILURE);
3942 		}
3943 		/*
3944 		 * Save the tgt_dip for the given target if one doesn't exist
3945 		 * already. Dip's for non-existance tgt's will be cleared in
3946 		 * tgt_free.
3947 		 */
3948 		if (softs->containers[tgt].dev.dip == NULL &&
3949 		    strcmp(ddi_driver_name(sd->sd_dev), "sd") == 0)
3950 			softs->containers[tgt].dev.dip = tgt_dip;
3951 	} else {
3952 		dvp = (struct aac_device *)&softs->nondasds[AAC_PD(tgt)];
3953 	}
3954 
3955 	if (softs->flags & AAC_FLAGS_BRKUP) {
3956 		if (ndi_prop_update_int(DDI_DEV_T_NONE, tgt_dip,
3957 		    "buf_break", 1) != DDI_PROP_SUCCESS) {
3958 			cmn_err(CE_CONT, "unable to create "
3959 			    "property for t%dL%d (buf_break)", tgt, lun);
3960 		}
3961 	}
3962 
3963 	AACDB_PRINT(softs, CE_NOTE,
3964 	    "aac_tran_tgt_init: c%dt%dL%d ok (%s)", ctl, tgt, lun,
3965 	    (dvp->type == AAC_DEV_PD) ? "pd" : "ld");
3966 	mutex_exit(&softs->io_lock);
3967 	return (DDI_SUCCESS);
3968 }
3969 
3970 static void
3971 aac_tran_tgt_free(dev_info_t *hba_dip, dev_info_t *tgt_dip,
3972     scsi_hba_tran_t *hba_tran, struct scsi_device *sd)
3973 {
3974 #ifndef __lock_lint
3975 	_NOTE(ARGUNUSED(hba_dip, tgt_dip, hba_tran))
3976 #endif
3977 
3978 	struct aac_softstate *softs = SD2AAC(sd);
3979 	int tgt = sd->sd_address.a_target;
3980 
3981 	mutex_enter(&softs->io_lock);
3982 	if (tgt < AAC_MAX_LD) {
3983 		if (softs->containers[tgt].dev.dip == tgt_dip)
3984 			softs->containers[tgt].dev.dip = NULL;
3985 	} else {
3986 		softs->nondasds[AAC_PD(tgt)].dev.flags &= ~AAC_DFLAG_VALID;
3987 	}
3988 	mutex_exit(&softs->io_lock);
3989 }
3990 
3991 /*
3992  * Check if the firmware is Up And Running. If it is in the Kernel Panic
3993  * state, (BlinkLED code + 1) is returned.
3994  *    0 -- firmware up and running
3995  *   -1 -- firmware dead
3996  *   >0 -- firmware kernel panic
3997  */
3998 static int
3999 aac_check_adapter_health(struct aac_softstate *softs)
4000 {
4001 	int rval;
4002 
4003 	rval = PCI_MEM_GET32(softs, AAC_OMR0);
4004 
4005 	if (rval & AAC_KERNEL_UP_AND_RUNNING) {
4006 		rval = 0;
4007 	} else if (rval & AAC_KERNEL_PANIC) {
4008 		cmn_err(CE_WARN, "firmware panic");
4009 		rval = ((rval >> 16) & 0xff) + 1; /* avoid 0 as return value */
4010 	} else {
4011 		cmn_err(CE_WARN, "firmware dead");
4012 		rval = -1;
4013 	}
4014 	return (rval);
4015 }
4016 
4017 static void
4018 aac_abort_iocmd(struct aac_softstate *softs, struct aac_cmd *acp,
4019     uchar_t reason)
4020 {
4021 	acp->flags |= AAC_CMD_ABORT;
4022 
4023 	if (acp->pkt) {
4024 		/*
4025 		 * Each lun should generate a unit attention
4026 		 * condition when reset.
4027 		 * Phys. drives are treated as logical ones
4028 		 * during error recovery.
4029 		 */
4030 		if (acp->slotp) { /* outstanding cmd */
4031 			acp->pkt->pkt_state |= STATE_GOT_STATUS;
4032 			aac_set_arq_data_reset(softs, acp);
4033 		}
4034 
4035 		switch (reason) {
4036 		case CMD_TIMEOUT:
4037 			AACDB_PRINT(softs, CE_NOTE, "CMD_TIMEOUT: acp=0x%p",
4038 			    acp);
4039 			aac_set_pkt_reason(softs, acp, CMD_TIMEOUT,
4040 			    STAT_TIMEOUT | STAT_BUS_RESET);
4041 			break;
4042 		case CMD_RESET:
4043 			/* aac support only RESET_ALL */
4044 			AACDB_PRINT(softs, CE_NOTE, "CMD_RESET: acp=0x%p", acp);
4045 			aac_set_pkt_reason(softs, acp, CMD_RESET,
4046 			    STAT_BUS_RESET);
4047 			break;
4048 		case CMD_ABORTED:
4049 			AACDB_PRINT(softs, CE_NOTE, "CMD_ABORTED: acp=0x%p",
4050 			    acp);
4051 			aac_set_pkt_reason(softs, acp, CMD_ABORTED,
4052 			    STAT_ABORTED);
4053 			break;
4054 		}
4055 	}
4056 	aac_end_io(softs, acp);
4057 }
4058 
4059 /*
4060  * Abort all the pending commands of type iocmd or just the command pkt
4061  * corresponding to pkt
4062  */
4063 static void
4064 aac_abort_iocmds(struct aac_softstate *softs, int iocmd, struct scsi_pkt *pkt,
4065     int reason)
4066 {
4067 	struct aac_cmd *ac_arg, *acp;
4068 	int i;
4069 
4070 	if (pkt == NULL) {
4071 		ac_arg = NULL;
4072 	} else {
4073 		ac_arg = PKT2AC(pkt);
4074 		iocmd = (ac_arg->flags & AAC_CMD_SYNC) ?
4075 		    AAC_IOCMD_SYNC : AAC_IOCMD_ASYNC;
4076 	}
4077 
4078 	/*
4079 	 * a) outstanding commands on the controller
4080 	 * Note: should abort outstanding commands only after one
4081 	 * IOP reset has been done.
4082 	 */
4083 	if (iocmd & AAC_IOCMD_OUTSTANDING) {
4084 		struct aac_cmd *acp;
4085 
4086 		for (i = 0; i < AAC_MAX_LD; i++) {
4087 			if (AAC_DEV_IS_VALID(&softs->containers[i].dev))
4088 				softs->containers[i].reset = 1;
4089 		}
4090 		while ((acp = softs->q_busy.q_head) != NULL)
4091 			aac_abort_iocmd(softs, acp, reason);
4092 	}
4093 
4094 	/* b) commands in the waiting queues */
4095 	for (i = 0; i < AAC_CMDQ_NUM; i++) {
4096 		if (iocmd & (1 << i)) {
4097 			if (ac_arg) {
4098 				aac_abort_iocmd(softs, ac_arg, reason);
4099 			} else {
4100 				while ((acp = softs->q_wait[i].q_head) != NULL)
4101 					aac_abort_iocmd(softs, acp, reason);
4102 			}
4103 		}
4104 	}
4105 }
4106 
4107 /*
4108  * The draining thread is shared among quiesce threads. It terminates
4109  * when the adapter is quiesced or stopped by aac_stop_drain().
4110  */
4111 static void
4112 aac_check_drain(void *arg)
4113 {
4114 	struct aac_softstate *softs = arg;
4115 
4116 	mutex_enter(&softs->io_lock);
4117 	if (softs->ndrains) {
4118 		softs->drain_timeid = 0;
4119 		/*
4120 		 * If both ASYNC and SYNC bus throttle are held,
4121 		 * wake up threads only when both are drained out.
4122 		 */
4123 		if ((softs->bus_throttle[AAC_CMDQ_ASYNC] > 0 ||
4124 		    softs->bus_ncmds[AAC_CMDQ_ASYNC] == 0) &&
4125 		    (softs->bus_throttle[AAC_CMDQ_SYNC] > 0 ||
4126 		    softs->bus_ncmds[AAC_CMDQ_SYNC] == 0))
4127 			cv_broadcast(&softs->drain_cv);
4128 		else
4129 			softs->drain_timeid = timeout(aac_check_drain, softs,
4130 			    AAC_QUIESCE_TICK * drv_usectohz(1000000));
4131 	}
4132 	mutex_exit(&softs->io_lock);
4133 }
4134 
4135 /*
4136  * If not draining the outstanding cmds, drain them. Otherwise,
4137  * only update ndrains.
4138  */
4139 static void
4140 aac_start_drain(struct aac_softstate *softs)
4141 {
4142 	if (softs->ndrains == 0) {
4143 		ASSERT(softs->drain_timeid == 0);
4144 		softs->drain_timeid = timeout(aac_check_drain, softs,
4145 		    AAC_QUIESCE_TICK * drv_usectohz(1000000));
4146 	}
4147 	softs->ndrains++;
4148 }
4149 
4150 /*
4151  * Stop the draining thread when no other threads use it any longer.
4152  * Side effect: io_lock may be released in the middle.
4153  */
4154 static void
4155 aac_stop_drain(struct aac_softstate *softs)
4156 {
4157 	softs->ndrains--;
4158 	if (softs->ndrains == 0) {
4159 		if (softs->drain_timeid != 0) {
4160 			timeout_id_t tid = softs->drain_timeid;
4161 
4162 			softs->drain_timeid = 0;
4163 			mutex_exit(&softs->io_lock);
4164 			(void) untimeout(tid);
4165 			mutex_enter(&softs->io_lock);
4166 		}
4167 	}
4168 }
4169 
4170 /*
4171  * The following function comes from Adaptec:
4172  *
4173  * Once do an IOP reset, basically the driver have to re-initialize the card
4174  * as if up from a cold boot, and the driver is responsible for any IO that
4175  * is outstanding to the adapter at the time of the IOP RESET. And prepare
4176  * for IOP RESET by making the init code modular with the ability to call it
4177  * from multiple places.
4178  */
4179 static int
4180 aac_reset_adapter(struct aac_softstate *softs)
4181 {
4182 	int health;
4183 	uint32_t status;
4184 	int rval = AAC_IOP_RESET_FAILED;
4185 
4186 	DBCALLED(softs, 1);
4187 
4188 	ASSERT(softs->state & AAC_STATE_RESET);
4189 
4190 	ddi_fm_acc_err_clear(softs->pci_mem_handle, DDI_FME_VER0);
4191 	/* Disable interrupt */
4192 	AAC_DISABLE_INTR(softs);
4193 
4194 	health = aac_check_adapter_health(softs);
4195 	if (health == -1) {
4196 		ddi_fm_service_impact(softs->devinfo_p, DDI_SERVICE_LOST);
4197 		goto finish;
4198 	}
4199 	if (health == 0) /* flush drives if possible */
4200 		(void) aac_shutdown(softs);
4201 
4202 	/* Execute IOP reset */
4203 	if ((aac_sync_mbcommand(softs, AAC_IOP_RESET, 0, 0, 0, 0,
4204 	    &status)) != AACOK) {
4205 		ddi_acc_handle_t acc = softs->comm_space_acc_handle;
4206 		struct aac_fib *fibp;
4207 		struct aac_pause_command *pc;
4208 
4209 		if ((status & 0xf) == 0xf) {
4210 			uint32_t wait_count;
4211 
4212 			/*
4213 			 * Sunrise Lake has dual cores and we must drag the
4214 			 * other core with us to reset simultaneously. There
4215 			 * are 2 bits in the Inbound Reset Control and Status
4216 			 * Register (offset 0x38) of the Sunrise Lake to reset
4217 			 * the chip without clearing out the PCI configuration
4218 			 * info (COMMAND & BARS).
4219 			 */
4220 			PCI_MEM_PUT32(softs, AAC_IRCSR, AAC_IRCSR_CORES_RST);
4221 
4222 			/*
4223 			 * We need to wait for 5 seconds before accessing the MU
4224 			 * again 10000 * 100us = 1000,000us = 1000ms = 1s
4225 			 */
4226 			wait_count = 5 * 10000;
4227 			while (wait_count) {
4228 				drv_usecwait(100); /* delay 100 microseconds */
4229 				wait_count--;
4230 			}
4231 		} else {
4232 			if (status == SRB_STATUS_INVALID_REQUEST)
4233 				cmn_err(CE_WARN, "!IOP_RESET not supported");
4234 			else /* probably timeout */
4235 				cmn_err(CE_WARN, "!IOP_RESET failed");
4236 
4237 			/* Unwind aac_shutdown() */
4238 			fibp = softs->sync_slot.fibp;
4239 			pc = (struct aac_pause_command *)&fibp->data[0];
4240 
4241 			bzero(pc, sizeof (*pc));
4242 			ddi_put32(acc, &pc->Command, VM_ContainerConfig);
4243 			ddi_put32(acc, &pc->Type, CT_PAUSE_IO);
4244 			ddi_put32(acc, &pc->Timeout, 1);
4245 			ddi_put32(acc, &pc->Min, 1);
4246 			ddi_put32(acc, &pc->NoRescan, 1);
4247 
4248 			(void) aac_sync_fib(softs, ContainerCommand,
4249 			    AAC_FIB_SIZEOF(struct aac_pause_command));
4250 
4251 			if (aac_check_adapter_health(softs) != 0)
4252 				ddi_fm_service_impact(softs->devinfo_p,
4253 				    DDI_SERVICE_LOST);
4254 			else
4255 				/*
4256 				 * IOP reset not supported or IOP not reseted
4257 				 */
4258 				rval = AAC_IOP_RESET_ABNORMAL;
4259 			goto finish;
4260 		}
4261 	}
4262 
4263 	/*
4264 	 * Re-read and renegotiate the FIB parameters, as one of the actions
4265 	 * that can result from an IOP reset is the running of a new firmware
4266 	 * image.
4267 	 */
4268 	if (aac_common_attach(softs) != AACOK)
4269 		goto finish;
4270 
4271 	rval = AAC_IOP_RESET_SUCCEED;
4272 
4273 finish:
4274 	AAC_ENABLE_INTR(softs);
4275 	return (rval);
4276 }
4277 
4278 static void
4279 aac_set_throttle(struct aac_softstate *softs, struct aac_device *dvp, int q,
4280     int throttle)
4281 {
4282 	/*
4283 	 * If the bus is draining/quiesced, no changes to the throttles
4284 	 * are allowed. All throttles should have been set to 0.
4285 	 */
4286 	if ((softs->state & AAC_STATE_QUIESCED) || softs->ndrains)
4287 		return;
4288 	dvp->throttle[q] = throttle;
4289 }
4290 
4291 static void
4292 aac_hold_bus(struct aac_softstate *softs, int iocmds)
4293 {
4294 	int i, q;
4295 
4296 	/* Hold bus by holding every device on the bus */
4297 	for (q = 0; q < AAC_CMDQ_NUM; q++) {
4298 		if (iocmds & (1 << q)) {
4299 			softs->bus_throttle[q] = 0;
4300 			for (i = 0; i < AAC_MAX_LD; i++)
4301 				aac_set_throttle(softs,
4302 				    &softs->containers[i].dev, q, 0);
4303 			for (i = 0; i < AAC_MAX_PD(softs); i++)
4304 				aac_set_throttle(softs,
4305 				    &softs->nondasds[i].dev, q, 0);
4306 		}
4307 	}
4308 }
4309 
4310 static void
4311 aac_unhold_bus(struct aac_softstate *softs, int iocmds)
4312 {
4313 	int i, q;
4314 
4315 	for (q = 0; q < AAC_CMDQ_NUM; q++) {
4316 		if (iocmds & (1 << q)) {
4317 			/*
4318 			 * Should not unhold AAC_IOCMD_ASYNC bus, if it has been
4319 			 * quiesced or being drained by possibly some quiesce
4320 			 * threads.
4321 			 */
4322 			if (q == AAC_CMDQ_ASYNC && ((softs->state &
4323 			    AAC_STATE_QUIESCED) || softs->ndrains))
4324 				continue;
4325 			softs->bus_throttle[q] = softs->total_slots;
4326 			for (i = 0; i < AAC_MAX_LD; i++)
4327 				aac_set_throttle(softs,
4328 				    &softs->containers[i].dev,
4329 				    q, softs->total_slots);
4330 			for (i = 0; i < AAC_MAX_PD(softs); i++)
4331 				aac_set_throttle(softs, &softs->nondasds[i].dev,
4332 				    q, softs->total_slots);
4333 		}
4334 	}
4335 }
4336 
4337 static int
4338 aac_do_reset(struct aac_softstate *softs)
4339 {
4340 	int health;
4341 	int rval;
4342 
4343 	softs->state |= AAC_STATE_RESET;
4344 	health = aac_check_adapter_health(softs);
4345 
4346 	/*
4347 	 * Hold off new io commands and wait all outstanding io
4348 	 * commands to complete.
4349 	 */
4350 	if (health == 0) {
4351 		int sync_cmds = softs->bus_ncmds[AAC_CMDQ_SYNC];
4352 		int async_cmds = softs->bus_ncmds[AAC_CMDQ_ASYNC];
4353 
4354 		if (sync_cmds == 0 && async_cmds == 0) {
4355 			rval = AAC_IOP_RESET_SUCCEED;
4356 			goto finish;
4357 		}
4358 		/*
4359 		 * Give the adapter up to AAC_QUIESCE_TIMEOUT more seconds
4360 		 * to complete the outstanding io commands
4361 		 */
4362 		int timeout = AAC_QUIESCE_TIMEOUT * 1000 * 10;
4363 		int (*intr_handler)(struct aac_softstate *);
4364 
4365 		aac_hold_bus(softs, AAC_IOCMD_SYNC | AAC_IOCMD_ASYNC);
4366 		/*
4367 		 * Poll the adapter by ourselves in case interrupt is disabled
4368 		 * and to avoid releasing the io_lock.
4369 		 */
4370 		intr_handler = (softs->flags & AAC_FLAGS_NEW_COMM) ?
4371 		    aac_process_intr_new : aac_process_intr_old;
4372 		while ((softs->bus_ncmds[AAC_CMDQ_SYNC] ||
4373 		    softs->bus_ncmds[AAC_CMDQ_ASYNC]) && timeout) {
4374 			drv_usecwait(100);
4375 			(void) intr_handler(softs);
4376 			timeout--;
4377 		}
4378 		aac_unhold_bus(softs, AAC_IOCMD_SYNC | AAC_IOCMD_ASYNC);
4379 
4380 		if (softs->bus_ncmds[AAC_CMDQ_SYNC] == 0 &&
4381 		    softs->bus_ncmds[AAC_CMDQ_ASYNC] == 0) {
4382 			/* Cmds drained out */
4383 			rval = AAC_IOP_RESET_SUCCEED;
4384 			goto finish;
4385 		} else if (softs->bus_ncmds[AAC_CMDQ_SYNC] < sync_cmds ||
4386 		    softs->bus_ncmds[AAC_CMDQ_ASYNC] < async_cmds) {
4387 			/* Cmds not drained out, adapter overloaded */
4388 			rval = AAC_IOP_RESET_ABNORMAL;
4389 			goto finish;
4390 		}
4391 	}
4392 
4393 	/*
4394 	 * If a longer waiting time still can't drain any outstanding io
4395 	 * commands, do IOP reset.
4396 	 */
4397 	if ((rval = aac_reset_adapter(softs)) == AAC_IOP_RESET_FAILED)
4398 		softs->state |= AAC_STATE_DEAD;
4399 
4400 finish:
4401 	softs->state &= ~AAC_STATE_RESET;
4402 	return (rval);
4403 }
4404 
4405 static int
4406 aac_tran_reset(struct scsi_address *ap, int level)
4407 {
4408 	struct aac_softstate *softs = AAC_TRAN2SOFTS(ap->a_hba_tran);
4409 	int rval;
4410 
4411 	DBCALLED(softs, 1);
4412 
4413 	if (level != RESET_ALL) {
4414 		cmn_err(CE_NOTE, "!reset target/lun not supported");
4415 		return (0);
4416 	}
4417 
4418 	mutex_enter(&softs->io_lock);
4419 	switch (rval = aac_do_reset(softs)) {
4420 	case AAC_IOP_RESET_SUCCEED:
4421 		aac_abort_iocmds(softs, AAC_IOCMD_OUTSTANDING | AAC_IOCMD_ASYNC,
4422 		    NULL, CMD_RESET);
4423 		aac_start_waiting_io(softs);
4424 		break;
4425 	case AAC_IOP_RESET_FAILED:
4426 		/* Abort IOCTL cmds when adapter is dead */
4427 		aac_abort_iocmds(softs, AAC_IOCMD_ALL, NULL, CMD_RESET);
4428 		break;
4429 	case AAC_IOP_RESET_ABNORMAL:
4430 		aac_start_waiting_io(softs);
4431 	}
4432 	mutex_exit(&softs->io_lock);
4433 
4434 	aac_drain_comp_q(softs);
4435 	return (rval == 0);
4436 }
4437 
4438 static int
4439 aac_tran_abort(struct scsi_address *ap, struct scsi_pkt *pkt)
4440 {
4441 	struct aac_softstate *softs = AAC_TRAN2SOFTS(ap->a_hba_tran);
4442 
4443 	DBCALLED(softs, 1);
4444 
4445 	mutex_enter(&softs->io_lock);
4446 	aac_abort_iocmds(softs, 0, pkt, CMD_ABORTED);
4447 	mutex_exit(&softs->io_lock);
4448 
4449 	aac_drain_comp_q(softs);
4450 	return (1);
4451 }
4452 
4453 void
4454 aac_free_dmamap(struct aac_cmd *acp)
4455 {
4456 	/* Free dma mapping */
4457 	if (acp->flags & AAC_CMD_DMA_VALID) {
4458 		ASSERT(acp->buf_dma_handle);
4459 		(void) ddi_dma_unbind_handle(acp->buf_dma_handle);
4460 		acp->flags &= ~AAC_CMD_DMA_VALID;
4461 	}
4462 
4463 	if (acp->abp != NULL) { /* free non-aligned buf DMA */
4464 		ASSERT(acp->buf_dma_handle);
4465 		if ((acp->flags & AAC_CMD_BUF_WRITE) == 0 && acp->bp)
4466 			ddi_rep_get8(acp->abh, (uint8_t *)acp->bp->b_un.b_addr,
4467 			    (uint8_t *)acp->abp, acp->bp->b_bcount,
4468 			    DDI_DEV_AUTOINCR);
4469 		ddi_dma_mem_free(&acp->abh);
4470 		acp->abp = NULL;
4471 	}
4472 
4473 	if (acp->buf_dma_handle) {
4474 		ddi_dma_free_handle(&acp->buf_dma_handle);
4475 		acp->buf_dma_handle = NULL;
4476 	}
4477 }
4478 
4479 static void
4480 aac_unknown_scmd(struct aac_softstate *softs, struct aac_cmd *acp)
4481 {
4482 	AACDB_PRINT(softs, CE_CONT, "SCMD 0x%x not supported",
4483 	    ((union scsi_cdb *)(void *)acp->pkt->pkt_cdbp)->scc_cmd);
4484 	aac_free_dmamap(acp);
4485 	aac_set_arq_data(acp->pkt, KEY_ILLEGAL_REQUEST, 0x20, 0x00, 0);
4486 	aac_soft_callback(softs, acp);
4487 }
4488 
4489 /*
4490  * Handle command to logical device
4491  */
4492 static int
4493 aac_tran_start_ld(struct aac_softstate *softs, struct aac_cmd *acp)
4494 {
4495 	struct aac_container *dvp;
4496 	struct scsi_pkt *pkt;
4497 	union scsi_cdb *cdbp;
4498 	struct buf *bp;
4499 	int rval;
4500 
4501 	dvp = (struct aac_container *)acp->dvp;
4502 	pkt = acp->pkt;
4503 	cdbp = (void *)pkt->pkt_cdbp;
4504 	bp = acp->bp;
4505 
4506 	switch (cdbp->scc_cmd) {
4507 	case SCMD_INQUIRY: /* inquiry */
4508 		aac_free_dmamap(acp);
4509 		aac_inquiry(softs, pkt, cdbp, bp);
4510 		aac_soft_callback(softs, acp);
4511 		rval = TRAN_ACCEPT;
4512 		break;
4513 
4514 	case SCMD_READ_CAPACITY: /* read capacity */
4515 		if (bp && bp->b_un.b_addr && bp->b_bcount) {
4516 			struct scsi_capacity cap;
4517 			uint64_t last_lba;
4518 
4519 			/* check 64-bit LBA */
4520 			last_lba = dvp->size - 1;
4521 			if (last_lba > 0xffffffffull) {
4522 				cap.capacity = 0xfffffffful;
4523 			} else {
4524 				cap.capacity = BE_32(last_lba);
4525 			}
4526 			cap.lbasize = BE_32(AAC_SECTOR_SIZE);
4527 
4528 			aac_free_dmamap(acp);
4529 			if (bp->b_flags & (B_PHYS|B_PAGEIO))
4530 				bp_mapin(bp);
4531 			bcopy(&cap, bp->b_un.b_addr, min(bp->b_bcount, 8));
4532 			pkt->pkt_state |= STATE_XFERRED_DATA;
4533 		}
4534 		aac_soft_callback(softs, acp);
4535 		rval = TRAN_ACCEPT;
4536 		break;
4537 
4538 	case SCMD_SVC_ACTION_IN_G4: /* read capacity 16 */
4539 		/* Check if containers need 64-bit LBA support */
4540 		if (cdbp->cdb_opaque[1] == SSVC_ACTION_READ_CAPACITY_G4) {
4541 			if (bp && bp->b_un.b_addr && bp->b_bcount) {
4542 				struct scsi_capacity_16 cap16;
4543 				int cap_len = sizeof (struct scsi_capacity_16);
4544 
4545 				bzero(&cap16, cap_len);
4546 				cap16.sc_capacity = BE_64(dvp->size - 1);
4547 				cap16.sc_lbasize = BE_32(AAC_SECTOR_SIZE);
4548 
4549 				aac_free_dmamap(acp);
4550 				if (bp->b_flags & (B_PHYS | B_PAGEIO))
4551 					bp_mapin(bp);
4552 				bcopy(&cap16, bp->b_un.b_addr,
4553 				    min(bp->b_bcount, cap_len));
4554 				pkt->pkt_state |= STATE_XFERRED_DATA;
4555 			}
4556 			aac_soft_callback(softs, acp);
4557 		} else {
4558 			aac_unknown_scmd(softs, acp);
4559 		}
4560 		rval = TRAN_ACCEPT;
4561 		break;
4562 
4563 	case SCMD_READ_G4: /* read_16 */
4564 	case SCMD_WRITE_G4: /* write_16 */
4565 		if (softs->flags & AAC_FLAGS_RAW_IO) {
4566 			/* NOTE: GETG4ADDRTL(cdbp) is int32_t */
4567 			acp->blkno = ((uint64_t) \
4568 			    GETG4ADDR(cdbp) << 32) | \
4569 			    (uint32_t)GETG4ADDRTL(cdbp);
4570 			goto do_io;
4571 		}
4572 		AACDB_PRINT(softs, CE_WARN, "64-bit LBA not supported");
4573 		aac_unknown_scmd(softs, acp);
4574 		rval = TRAN_ACCEPT;
4575 		break;
4576 
4577 	case SCMD_READ: /* read_6 */
4578 	case SCMD_WRITE: /* write_6 */
4579 		acp->blkno = GETG0ADDR(cdbp);
4580 		goto do_io;
4581 
4582 	case SCMD_READ_G5: /* read_12 */
4583 	case SCMD_WRITE_G5: /* write_12 */
4584 		acp->blkno = GETG5ADDR(cdbp);
4585 		goto do_io;
4586 
4587 	case SCMD_READ_G1: /* read_10 */
4588 	case SCMD_WRITE_G1: /* write_10 */
4589 		acp->blkno = (uint32_t)GETG1ADDR(cdbp);
4590 do_io:
4591 		if (acp->flags & AAC_CMD_DMA_VALID) {
4592 			uint64_t cnt_size = dvp->size;
4593 
4594 			/*
4595 			 * If LBA > array size AND rawio, the
4596 			 * adapter may hang. So check it before
4597 			 * sending.
4598 			 * NOTE: (blkno + blkcnt) may overflow
4599 			 */
4600 			if ((acp->blkno < cnt_size) &&
4601 			    ((acp->blkno + acp->bcount /
4602 			    AAC_BLK_SIZE) <= cnt_size)) {
4603 				rval = aac_do_io(softs, acp);
4604 			} else {
4605 			/*
4606 			 * Request exceeds the capacity of disk,
4607 			 * set error block number to last LBA
4608 			 * + 1.
4609 			 */
4610 				aac_set_arq_data(pkt,
4611 				    KEY_ILLEGAL_REQUEST, 0x21,
4612 				    0x00, cnt_size);
4613 				aac_soft_callback(softs, acp);
4614 				rval = TRAN_ACCEPT;
4615 			}
4616 		} else if (acp->bcount == 0) {
4617 			/* For 0 length IO, just return ok */
4618 			aac_soft_callback(softs, acp);
4619 			rval = TRAN_ACCEPT;
4620 		} else {
4621 			rval = TRAN_BADPKT;
4622 		}
4623 		break;
4624 
4625 	case SCMD_MODE_SENSE: /* mode_sense_6 */
4626 	case SCMD_MODE_SENSE_G1: { /* mode_sense_10 */
4627 		int capacity;
4628 
4629 		aac_free_dmamap(acp);
4630 		if (dvp->size > 0xffffffffull)
4631 			capacity = 0xfffffffful; /* 64-bit LBA */
4632 		else
4633 			capacity = dvp->size;
4634 		aac_mode_sense(softs, pkt, cdbp, bp, capacity);
4635 		aac_soft_callback(softs, acp);
4636 		rval = TRAN_ACCEPT;
4637 		break;
4638 	}
4639 
4640 	case SCMD_TEST_UNIT_READY:
4641 	case SCMD_REQUEST_SENSE:
4642 	case SCMD_FORMAT:
4643 	case SCMD_START_STOP:
4644 		aac_free_dmamap(acp);
4645 		if (bp && bp->b_un.b_addr && bp->b_bcount) {
4646 			if (acp->flags & AAC_CMD_BUF_READ) {
4647 				if (bp->b_flags & (B_PHYS|B_PAGEIO))
4648 					bp_mapin(bp);
4649 				bzero(bp->b_un.b_addr, bp->b_bcount);
4650 			}
4651 			pkt->pkt_state |= STATE_XFERRED_DATA;
4652 		}
4653 		aac_soft_callback(softs, acp);
4654 		rval = TRAN_ACCEPT;
4655 		break;
4656 
4657 	case SCMD_SYNCHRONIZE_CACHE:
4658 		acp->flags |= AAC_CMD_NTAG;
4659 		acp->aac_cmd_fib = aac_cmd_fib_sync;
4660 		acp->ac_comp = aac_synccache_complete;
4661 		rval = aac_do_io(softs, acp);
4662 		break;
4663 
4664 	case SCMD_DOORLOCK:
4665 		aac_free_dmamap(acp);
4666 		dvp->locked = (pkt->pkt_cdbp[4] & 0x01) ? 1 : 0;
4667 		aac_soft_callback(softs, acp);
4668 		rval = TRAN_ACCEPT;
4669 		break;
4670 
4671 	default: /* unknown command */
4672 		aac_unknown_scmd(softs, acp);
4673 		rval = TRAN_ACCEPT;
4674 		break;
4675 	}
4676 
4677 	return (rval);
4678 }
4679 
4680 static int
4681 aac_tran_start(struct scsi_address *ap, struct scsi_pkt *pkt)
4682 {
4683 	struct aac_softstate *softs = AAC_TRAN2SOFTS(ap->a_hba_tran);
4684 	struct aac_cmd *acp = PKT2AC(pkt);
4685 	struct aac_device *dvp = acp->dvp;
4686 	int rval;
4687 
4688 	DBCALLED(softs, 2);
4689 
4690 	/*
4691 	 * Reinitialize some fields of ac and pkt; the packet may
4692 	 * have been resubmitted
4693 	 */
4694 	acp->flags &= AAC_CMD_CONSISTENT | AAC_CMD_DMA_PARTIAL | \
4695 	    AAC_CMD_BUF_READ | AAC_CMD_BUF_WRITE | AAC_CMD_DMA_VALID;
4696 	acp->timeout = acp->pkt->pkt_time;
4697 	if (pkt->pkt_flags & FLAG_NOINTR)
4698 		acp->flags |= AAC_CMD_NO_INTR;
4699 #ifdef DEBUG
4700 	acp->fib_flags = AACDB_FLAGS_FIB_SCMD;
4701 #endif
4702 	pkt->pkt_reason = CMD_CMPLT;
4703 	pkt->pkt_state = 0;
4704 	pkt->pkt_statistics = 0;
4705 	*pkt->pkt_scbp = STATUS_GOOD; /* clear arq scsi_status */
4706 
4707 	if (acp->flags & AAC_CMD_DMA_VALID) {
4708 		pkt->pkt_resid = acp->bcount;
4709 		/* Consistent packets need to be sync'ed first */
4710 		if ((acp->flags & AAC_CMD_CONSISTENT) &&
4711 		    (acp->flags & AAC_CMD_BUF_WRITE))
4712 			if (aac_dma_sync_ac(acp) != AACOK) {
4713 				ddi_fm_service_impact(softs->devinfo_p,
4714 				    DDI_SERVICE_UNAFFECTED);
4715 				return (TRAN_BADPKT);
4716 			}
4717 	} else {
4718 		pkt->pkt_resid = 0;
4719 	}
4720 
4721 	mutex_enter(&softs->io_lock);
4722 	AACDB_PRINT_SCMD(softs, acp);
4723 	if ((dvp->flags & (AAC_DFLAG_VALID | AAC_DFLAG_CONFIGURING)) &&
4724 	    !(softs->state & AAC_STATE_DEAD)) {
4725 		if (dvp->type == AAC_DEV_LD) {
4726 			if (ap->a_lun == 0)
4727 				rval = aac_tran_start_ld(softs, acp);
4728 			else
4729 				goto error;
4730 		} else {
4731 			rval = aac_do_io(softs, acp);
4732 		}
4733 	} else {
4734 error:
4735 #ifdef DEBUG
4736 		if (!(softs->state & AAC_STATE_DEAD)) {
4737 			AACDB_PRINT_TRAN(softs,
4738 			    "Cannot send cmd to target t%dL%d: %s",
4739 			    ap->a_target, ap->a_lun,
4740 			    "target invalid");
4741 		} else {
4742 			AACDB_PRINT(softs, CE_WARN,
4743 			    "Cannot send cmd to target t%dL%d: %s",
4744 			    ap->a_target, ap->a_lun,
4745 			    "adapter dead");
4746 		}
4747 #endif
4748 		rval = TRAN_FATAL_ERROR;
4749 	}
4750 	mutex_exit(&softs->io_lock);
4751 	return (rval);
4752 }
4753 
4754 static int
4755 aac_tran_getcap(struct scsi_address *ap, char *cap, int whom)
4756 {
4757 	struct aac_softstate *softs = AAC_TRAN2SOFTS(ap->a_hba_tran);
4758 	struct aac_device *dvp;
4759 	int rval;
4760 
4761 	DBCALLED(softs, 2);
4762 
4763 	/* We don't allow inquiring about capabilities for other targets */
4764 	if (cap == NULL || whom == 0) {
4765 		AACDB_PRINT(softs, CE_WARN,
4766 		    "GetCap> %s not supported: whom=%d", cap, whom);
4767 		return (-1);
4768 	}
4769 
4770 	mutex_enter(&softs->io_lock);
4771 	dvp = AAC_DEV(softs, ap->a_target);
4772 	if (dvp == NULL || !AAC_DEV_IS_VALID(dvp)) {
4773 		mutex_exit(&softs->io_lock);
4774 		AACDB_PRINT_TRAN(softs, "Bad target t%dL%d to getcap",
4775 		    ap->a_target, ap->a_lun);
4776 		return (-1);
4777 	}
4778 
4779 	switch (scsi_hba_lookup_capstr(cap)) {
4780 	case SCSI_CAP_ARQ: /* auto request sense */
4781 		rval = 1;
4782 		break;
4783 	case SCSI_CAP_UNTAGGED_QING:
4784 	case SCSI_CAP_TAGGED_QING:
4785 		rval = 1;
4786 		break;
4787 	case SCSI_CAP_DMA_MAX:
4788 		rval = softs->dma_max;
4789 		break;
4790 	default:
4791 		rval = -1;
4792 		break;
4793 	}
4794 	mutex_exit(&softs->io_lock);
4795 
4796 	AACDB_PRINT_TRAN(softs, "GetCap> %s t%dL%d: rval=%d",
4797 	    cap, ap->a_target, ap->a_lun, rval);
4798 	return (rval);
4799 }
4800 
4801 /*ARGSUSED*/
4802 static int
4803 aac_tran_setcap(struct scsi_address *ap, char *cap, int value, int whom)
4804 {
4805 	struct aac_softstate *softs = AAC_TRAN2SOFTS(ap->a_hba_tran);
4806 	struct aac_device *dvp;
4807 	int rval;
4808 
4809 	DBCALLED(softs, 2);
4810 
4811 	/* We don't allow inquiring about capabilities for other targets */
4812 	if (cap == NULL || whom == 0) {
4813 		AACDB_PRINT(softs, CE_WARN,
4814 		    "SetCap> %s not supported: whom=%d", cap, whom);
4815 		return (-1);
4816 	}
4817 
4818 	mutex_enter(&softs->io_lock);
4819 	dvp = AAC_DEV(softs, ap->a_target);
4820 	if (dvp == NULL || !AAC_DEV_IS_VALID(dvp)) {
4821 		mutex_exit(&softs->io_lock);
4822 		AACDB_PRINT_TRAN(softs, "Bad target t%dL%d to setcap",
4823 		    ap->a_target, ap->a_lun);
4824 		return (-1);
4825 	}
4826 
4827 	switch (scsi_hba_lookup_capstr(cap)) {
4828 	case SCSI_CAP_ARQ:
4829 		/* Force auto request sense */
4830 		rval = (value == 1) ? 1 : 0;
4831 		break;
4832 	case SCSI_CAP_UNTAGGED_QING:
4833 	case SCSI_CAP_TAGGED_QING:
4834 		rval = (value == 1) ? 1 : 0;
4835 		break;
4836 	default:
4837 		rval = -1;
4838 		break;
4839 	}
4840 	mutex_exit(&softs->io_lock);
4841 
4842 	AACDB_PRINT_TRAN(softs, "SetCap> %s t%dL%d val=%d: rval=%d",
4843 	    cap, ap->a_target, ap->a_lun, value, rval);
4844 	return (rval);
4845 }
4846 
4847 static void
4848 aac_tran_destroy_pkt(struct scsi_address *ap, struct scsi_pkt *pkt)
4849 {
4850 	struct aac_cmd *acp = PKT2AC(pkt);
4851 
4852 	DBCALLED(NULL, 2);
4853 
4854 	if (acp->sgt) {
4855 		kmem_free(acp->sgt, sizeof (struct aac_sge) * \
4856 		    acp->left_cookien);
4857 	}
4858 	aac_free_dmamap(acp);
4859 	ASSERT(acp->slotp == NULL);
4860 	scsi_hba_pkt_free(ap, pkt);
4861 }
4862 
4863 int
4864 aac_cmd_dma_alloc(struct aac_softstate *softs, struct aac_cmd *acp,
4865     struct buf *bp, int flags, int (*cb)(), caddr_t arg)
4866 {
4867 	int kf = (cb == SLEEP_FUNC) ? KM_SLEEP : KM_NOSLEEP;
4868 	uint_t oldcookiec;
4869 	int bioerr;
4870 	int rval;
4871 
4872 	oldcookiec = acp->left_cookien;
4873 
4874 	/* Move window to build s/g map */
4875 	if (acp->total_nwin > 0) {
4876 		if (++acp->cur_win < acp->total_nwin) {
4877 			off_t off;
4878 			size_t len;
4879 
4880 			rval = ddi_dma_getwin(acp->buf_dma_handle, acp->cur_win,
4881 			    &off, &len, &acp->cookie, &acp->left_cookien);
4882 			if (rval == DDI_SUCCESS)
4883 				goto get_dma_cookies;
4884 			AACDB_PRINT(softs, CE_WARN,
4885 			    "ddi_dma_getwin() fail %d", rval);
4886 			return (AACERR);
4887 		}
4888 		AACDB_PRINT(softs, CE_WARN, "Nothing to transfer");
4889 		return (AACERR);
4890 	}
4891 
4892 	/* We need to transfer data, so we alloc DMA resources for this pkt */
4893 	if (bp && bp->b_bcount != 0 && !(acp->flags & AAC_CMD_DMA_VALID)) {
4894 		uint_t dma_flags = 0;
4895 		struct aac_sge *sge;
4896 
4897 		/*
4898 		 * We will still use this point to fake some
4899 		 * infomation in tran_start
4900 		 */
4901 		acp->bp = bp;
4902 
4903 		/* Set dma flags */
4904 		if (BUF_IS_READ(bp)) {
4905 			dma_flags |= DDI_DMA_READ;
4906 			acp->flags |= AAC_CMD_BUF_READ;
4907 		} else {
4908 			dma_flags |= DDI_DMA_WRITE;
4909 			acp->flags |= AAC_CMD_BUF_WRITE;
4910 		}
4911 		if (flags & PKT_CONSISTENT)
4912 			dma_flags |= DDI_DMA_CONSISTENT;
4913 		if (flags & PKT_DMA_PARTIAL)
4914 			dma_flags |= DDI_DMA_PARTIAL;
4915 
4916 		/* Alloc buf dma handle */
4917 		if (!acp->buf_dma_handle) {
4918 			rval = ddi_dma_alloc_handle(softs->devinfo_p,
4919 			    &softs->buf_dma_attr, cb, arg,
4920 			    &acp->buf_dma_handle);
4921 			if (rval != DDI_SUCCESS) {
4922 				AACDB_PRINT(softs, CE_WARN,
4923 				    "Can't allocate DMA handle, errno=%d",
4924 				    rval);
4925 				goto error_out;
4926 			}
4927 		}
4928 
4929 		/* Bind buf */
4930 		if (((uintptr_t)bp->b_un.b_addr & AAC_DMA_ALIGN_MASK) == 0) {
4931 			rval = ddi_dma_buf_bind_handle(acp->buf_dma_handle,
4932 			    bp, dma_flags, cb, arg, &acp->cookie,
4933 			    &acp->left_cookien);
4934 		} else {
4935 			size_t bufsz;
4936 
4937 			AACDB_PRINT_TRAN(softs,
4938 			    "non-aligned buffer: addr=0x%p, cnt=%lu",
4939 			    (void *)bp->b_un.b_addr, bp->b_bcount);
4940 			if (bp->b_flags & (B_PAGEIO|B_PHYS))
4941 				bp_mapin(bp);
4942 
4943 			rval = ddi_dma_mem_alloc(acp->buf_dma_handle,
4944 			    AAC_ROUNDUP(bp->b_bcount, AAC_DMA_ALIGN),
4945 			    &softs->acc_attr, DDI_DMA_STREAMING,
4946 			    cb, arg, &acp->abp, &bufsz, &acp->abh);
4947 
4948 			if (rval != DDI_SUCCESS) {
4949 				AACDB_PRINT(softs, CE_NOTE,
4950 				    "Cannot alloc DMA to non-aligned buf");
4951 				bioerr = 0;
4952 				goto error_out;
4953 			}
4954 
4955 			if (acp->flags & AAC_CMD_BUF_WRITE)
4956 				ddi_rep_put8(acp->abh,
4957 				    (uint8_t *)bp->b_un.b_addr,
4958 				    (uint8_t *)acp->abp, bp->b_bcount,
4959 				    DDI_DEV_AUTOINCR);
4960 
4961 			rval = ddi_dma_addr_bind_handle(acp->buf_dma_handle,
4962 			    NULL, acp->abp, bufsz, dma_flags, cb, arg,
4963 			    &acp->cookie, &acp->left_cookien);
4964 		}
4965 
4966 		switch (rval) {
4967 		case DDI_DMA_PARTIAL_MAP:
4968 			if (ddi_dma_numwin(acp->buf_dma_handle,
4969 			    &acp->total_nwin) == DDI_FAILURE) {
4970 				AACDB_PRINT(softs, CE_WARN,
4971 				    "Cannot get number of DMA windows");
4972 				bioerr = 0;
4973 				goto error_out;
4974 			}
4975 			AACDB_PRINT_TRAN(softs, "buf bind, %d seg(s)",
4976 			    acp->left_cookien);
4977 			acp->cur_win = 0;
4978 			break;
4979 
4980 		case DDI_DMA_MAPPED:
4981 			AACDB_PRINT_TRAN(softs, "buf bind, %d seg(s)",
4982 			    acp->left_cookien);
4983 			acp->cur_win = 0;
4984 			acp->total_nwin = 1;
4985 			break;
4986 
4987 		case DDI_DMA_NORESOURCES:
4988 			bioerr = 0;
4989 			AACDB_PRINT(softs, CE_WARN,
4990 			    "Cannot bind buf for DMA: DDI_DMA_NORESOURCES");
4991 			goto error_out;
4992 		case DDI_DMA_BADATTR:
4993 		case DDI_DMA_NOMAPPING:
4994 			bioerr = EFAULT;
4995 			AACDB_PRINT(softs, CE_WARN,
4996 			    "Cannot bind buf for DMA: DDI_DMA_NOMAPPING");
4997 			goto error_out;
4998 		case DDI_DMA_TOOBIG:
4999 			bioerr = EINVAL;
5000 			AACDB_PRINT(softs, CE_WARN,
5001 			    "Cannot bind buf for DMA: DDI_DMA_TOOBIG(%d)",
5002 			    bp->b_bcount);
5003 			goto error_out;
5004 		default:
5005 			bioerr = EINVAL;
5006 			AACDB_PRINT(softs, CE_WARN,
5007 			    "Cannot bind buf for DMA: %d", rval);
5008 			goto error_out;
5009 		}
5010 		acp->flags |= AAC_CMD_DMA_VALID;
5011 
5012 get_dma_cookies:
5013 		ASSERT(acp->left_cookien > 0);
5014 		if (acp->left_cookien > softs->aac_sg_tablesize) {
5015 			AACDB_PRINT(softs, CE_NOTE, "large cookiec received %d",
5016 			    acp->left_cookien);
5017 			bioerr = EINVAL;
5018 			goto error_out;
5019 		}
5020 		if (oldcookiec != acp->left_cookien && acp->sgt != NULL) {
5021 			kmem_free(acp->sgt, sizeof (struct aac_sge) * \
5022 			    oldcookiec);
5023 			acp->sgt = NULL;
5024 		}
5025 		if (acp->sgt == NULL) {
5026 			acp->sgt = kmem_alloc(sizeof (struct aac_sge) * \
5027 			    acp->left_cookien, kf);
5028 			if (acp->sgt == NULL) {
5029 				AACDB_PRINT(softs, CE_WARN,
5030 				    "sgt kmem_alloc fail");
5031 				bioerr = ENOMEM;
5032 				goto error_out;
5033 			}
5034 		}
5035 
5036 		sge = &acp->sgt[0];
5037 		sge->bcount = acp->cookie.dmac_size;
5038 		sge->addr.ad64.lo = AAC_LS32(acp->cookie.dmac_laddress);
5039 		sge->addr.ad64.hi = AAC_MS32(acp->cookie.dmac_laddress);
5040 		acp->bcount = acp->cookie.dmac_size;
5041 		for (sge++; sge < &acp->sgt[acp->left_cookien]; sge++) {
5042 			ddi_dma_nextcookie(acp->buf_dma_handle, &acp->cookie);
5043 			sge->bcount = acp->cookie.dmac_size;
5044 			sge->addr.ad64.lo = AAC_LS32(acp->cookie.dmac_laddress);
5045 			sge->addr.ad64.hi = AAC_MS32(acp->cookie.dmac_laddress);
5046 			acp->bcount += acp->cookie.dmac_size;
5047 		}
5048 
5049 		/*
5050 		 * Note: The old DMA engine do not correctly handle
5051 		 * dma_attr_maxxfer attribute. So we have to ensure
5052 		 * it by ourself.
5053 		 */
5054 		if (acp->bcount > softs->buf_dma_attr.dma_attr_maxxfer) {
5055 			AACDB_PRINT(softs, CE_NOTE,
5056 			    "large xfer size received %d\n", acp->bcount);
5057 			bioerr = EINVAL;
5058 			goto error_out;
5059 		}
5060 
5061 		acp->total_xfer += acp->bcount;
5062 
5063 		if (acp->pkt) {
5064 			/* Return remaining byte count */
5065 			if (acp->total_xfer <= bp->b_bcount) {
5066 				acp->pkt->pkt_resid = bp->b_bcount - \
5067 				    acp->total_xfer;
5068 			} else {
5069 				/*
5070 				 * Allocated DMA size is greater than the buf
5071 				 * size of bp. This is caused by devices like
5072 				 * tape. we have extra bytes allocated, but
5073 				 * the packet residual has to stay correct.
5074 				 */
5075 				acp->pkt->pkt_resid = 0;
5076 			}
5077 			AACDB_PRINT_TRAN(softs,
5078 			    "bp=0x%p, xfered=%d/%d, resid=%d",
5079 			    (void *)bp->b_un.b_addr, (int)acp->total_xfer,
5080 			    (int)bp->b_bcount, (int)acp->pkt->pkt_resid);
5081 		}
5082 	}
5083 	return (AACOK);
5084 
5085 error_out:
5086 	bioerror(bp, bioerr);
5087 	return (AACERR);
5088 }
5089 
5090 static struct scsi_pkt *
5091 aac_tran_init_pkt(struct scsi_address *ap, struct scsi_pkt *pkt,
5092     struct buf *bp, int cmdlen, int statuslen, int tgtlen, int flags,
5093     int (*callback)(), caddr_t arg)
5094 {
5095 	struct aac_softstate *softs = AAC_TRAN2SOFTS(ap->a_hba_tran);
5096 	struct aac_cmd *acp, *new_acp;
5097 
5098 	DBCALLED(softs, 2);
5099 
5100 	/* Allocate pkt */
5101 	if (pkt == NULL) {
5102 		int slen;
5103 
5104 		/* Force auto request sense */
5105 		slen = (statuslen > softs->slen) ? statuslen : softs->slen;
5106 		pkt = scsi_hba_pkt_alloc(softs->devinfo_p, ap, cmdlen,
5107 		    slen, tgtlen, sizeof (struct aac_cmd), callback, arg);
5108 		if (pkt == NULL) {
5109 			AACDB_PRINT(softs, CE_WARN, "Alloc scsi pkt failed");
5110 			return (NULL);
5111 		}
5112 		acp = new_acp = PKT2AC(pkt);
5113 		acp->pkt = pkt;
5114 		acp->cmdlen = cmdlen;
5115 
5116 		if (ap->a_target < AAC_MAX_LD) {
5117 			acp->dvp = &softs->containers[ap->a_target].dev;
5118 			acp->aac_cmd_fib = softs->aac_cmd_fib;
5119 			acp->ac_comp = aac_ld_complete;
5120 		} else {
5121 			_NOTE(ASSUMING_PROTECTED(softs->nondasds))
5122 
5123 			acp->dvp = &softs->nondasds[AAC_PD(ap->a_target)].dev;
5124 			acp->aac_cmd_fib = softs->aac_cmd_fib_scsi;
5125 			acp->ac_comp = aac_pd_complete;
5126 		}
5127 	} else {
5128 		acp = PKT2AC(pkt);
5129 		new_acp = NULL;
5130 	}
5131 
5132 	if (aac_cmd_dma_alloc(softs, acp, bp, flags, callback, arg) == AACOK)
5133 		return (pkt);
5134 
5135 	if (new_acp)
5136 		aac_tran_destroy_pkt(ap, pkt);
5137 	return (NULL);
5138 }
5139 
5140 /*
5141  * tran_sync_pkt(9E) - explicit DMA synchronization
5142  */
5143 /*ARGSUSED*/
5144 static void
5145 aac_tran_sync_pkt(struct scsi_address *ap, struct scsi_pkt *pkt)
5146 {
5147 	struct aac_cmd *acp = PKT2AC(pkt);
5148 
5149 	DBCALLED(NULL, 2);
5150 
5151 	if (aac_dma_sync_ac(acp) != AACOK)
5152 		ddi_fm_service_impact(
5153 		    (AAC_TRAN2SOFTS(ap->a_hba_tran))->devinfo_p,
5154 		    DDI_SERVICE_UNAFFECTED);
5155 }
5156 
5157 /*
5158  * tran_dmafree(9E) - deallocate DMA resources allocated for command
5159  */
5160 /*ARGSUSED*/
5161 static void
5162 aac_tran_dmafree(struct scsi_address *ap, struct scsi_pkt *pkt)
5163 {
5164 	struct aac_cmd *acp = PKT2AC(pkt);
5165 
5166 	DBCALLED(NULL, 2);
5167 
5168 	aac_free_dmamap(acp);
5169 }
5170 
5171 static int
5172 aac_do_quiesce(struct aac_softstate *softs)
5173 {
5174 	aac_hold_bus(softs, AAC_IOCMD_ASYNC);
5175 	if (softs->bus_ncmds[AAC_CMDQ_ASYNC]) {
5176 		aac_start_drain(softs);
5177 		do {
5178 			if (cv_wait_sig(&softs->drain_cv,
5179 			    &softs->io_lock) == 0) {
5180 				/* Quiesce has been interrupted */
5181 				aac_stop_drain(softs);
5182 				aac_unhold_bus(softs, AAC_IOCMD_ASYNC);
5183 				aac_start_waiting_io(softs);
5184 				return (AACERR);
5185 			}
5186 		} while (softs->bus_ncmds[AAC_CMDQ_ASYNC]);
5187 		aac_stop_drain(softs);
5188 	}
5189 
5190 	softs->state |= AAC_STATE_QUIESCED;
5191 	return (AACOK);
5192 }
5193 
5194 static int
5195 aac_tran_quiesce(dev_info_t *dip)
5196 {
5197 	struct aac_softstate *softs = AAC_DIP2SOFTS(dip);
5198 	int rval;
5199 
5200 	DBCALLED(softs, 1);
5201 
5202 	mutex_enter(&softs->io_lock);
5203 	if (aac_do_quiesce(softs) == AACOK)
5204 		rval = 0;
5205 	else
5206 		rval = 1;
5207 	mutex_exit(&softs->io_lock);
5208 	return (rval);
5209 }
5210 
5211 static int
5212 aac_do_unquiesce(struct aac_softstate *softs)
5213 {
5214 	softs->state &= ~AAC_STATE_QUIESCED;
5215 	aac_unhold_bus(softs, AAC_IOCMD_ASYNC);
5216 
5217 	aac_start_waiting_io(softs);
5218 	return (AACOK);
5219 }
5220 
5221 static int
5222 aac_tran_unquiesce(dev_info_t *dip)
5223 {
5224 	struct aac_softstate *softs = AAC_DIP2SOFTS(dip);
5225 	int rval;
5226 
5227 	DBCALLED(softs, 1);
5228 
5229 	mutex_enter(&softs->io_lock);
5230 	if (aac_do_unquiesce(softs) == AACOK)
5231 		rval = 0;
5232 	else
5233 		rval = 1;
5234 	mutex_exit(&softs->io_lock);
5235 	return (rval);
5236 }
5237 
5238 static int
5239 aac_hba_setup(struct aac_softstate *softs)
5240 {
5241 	scsi_hba_tran_t *hba_tran;
5242 	int rval;
5243 
5244 	hba_tran = scsi_hba_tran_alloc(softs->devinfo_p, SCSI_HBA_CANSLEEP);
5245 	if (hba_tran == NULL)
5246 		return (AACERR);
5247 	hba_tran->tran_hba_private = softs;
5248 	hba_tran->tran_tgt_init = aac_tran_tgt_init;
5249 	hba_tran->tran_tgt_free = aac_tran_tgt_free;
5250 	hba_tran->tran_tgt_probe = scsi_hba_probe;
5251 	hba_tran->tran_start = aac_tran_start;
5252 	hba_tran->tran_getcap = aac_tran_getcap;
5253 	hba_tran->tran_setcap = aac_tran_setcap;
5254 	hba_tran->tran_init_pkt = aac_tran_init_pkt;
5255 	hba_tran->tran_destroy_pkt = aac_tran_destroy_pkt;
5256 	hba_tran->tran_reset = aac_tran_reset;
5257 	hba_tran->tran_abort = aac_tran_abort;
5258 	hba_tran->tran_sync_pkt = aac_tran_sync_pkt;
5259 	hba_tran->tran_dmafree = aac_tran_dmafree;
5260 	hba_tran->tran_quiesce = aac_tran_quiesce;
5261 	hba_tran->tran_unquiesce = aac_tran_unquiesce;
5262 	hba_tran->tran_bus_config = aac_tran_bus_config;
5263 	rval = scsi_hba_attach_setup(softs->devinfo_p, &softs->buf_dma_attr,
5264 	    hba_tran, 0);
5265 	if (rval != DDI_SUCCESS) {
5266 		scsi_hba_tran_free(hba_tran);
5267 		AACDB_PRINT(softs, CE_WARN, "aac_hba_setup failed");
5268 		return (AACERR);
5269 	}
5270 
5271 	softs->hba_tran = hba_tran;
5272 	return (AACOK);
5273 }
5274 
5275 /*
5276  * FIB setup operations
5277  */
5278 
5279 /*
5280  * Init FIB header
5281  */
5282 static void
5283 aac_cmd_fib_header(struct aac_softstate *softs, struct aac_slot *slotp,
5284     uint16_t cmd, uint16_t fib_size)
5285 {
5286 	ddi_acc_handle_t acc = slotp->fib_acc_handle;
5287 	struct aac_fib *fibp = slotp->fibp;
5288 	uint32_t xfer_state;
5289 
5290 	xfer_state =
5291 	    AAC_FIBSTATE_HOSTOWNED |
5292 	    AAC_FIBSTATE_INITIALISED |
5293 	    AAC_FIBSTATE_EMPTY |
5294 	    AAC_FIBSTATE_FROMHOST |
5295 	    AAC_FIBSTATE_REXPECTED |
5296 	    AAC_FIBSTATE_NORM;
5297 	if (slotp->acp && !(slotp->acp->flags & AAC_CMD_SYNC)) {
5298 		xfer_state |=
5299 		    AAC_FIBSTATE_ASYNC |
5300 		    AAC_FIBSTATE_FAST_RESPONSE /* enable fast io */;
5301 		ddi_put16(acc, &fibp->Header.SenderSize,
5302 		    softs->aac_max_fib_size);
5303 	} else {
5304 		ddi_put16(acc, &fibp->Header.SenderSize, AAC_FIB_SIZE);
5305 	}
5306 
5307 	ddi_put32(acc, &fibp->Header.XferState, xfer_state);
5308 	ddi_put16(acc, &fibp->Header.Command, cmd);
5309 	ddi_put8(acc, &fibp->Header.StructType, AAC_FIBTYPE_TFIB);
5310 	ddi_put8(acc, &fibp->Header.Flags, 0); /* don't care */
5311 	ddi_put16(acc, &fibp->Header.Size, fib_size);
5312 	ddi_put32(acc, &fibp->Header.SenderFibAddress, (slotp->index << 2));
5313 	ddi_put32(acc, &fibp->Header.ReceiverFibAddress, slotp->fib_phyaddr);
5314 	ddi_put32(acc, &fibp->Header.SenderData, 0); /* don't care */
5315 }
5316 
5317 /*
5318  * Init FIB for raw IO command
5319  */
5320 static void
5321 aac_cmd_fib_rawio(struct aac_softstate *softs, struct aac_cmd *acp)
5322 {
5323 	ddi_acc_handle_t acc = acp->slotp->fib_acc_handle;
5324 	struct aac_raw_io *io = (struct aac_raw_io *)&acp->slotp->fibp->data[0];
5325 	struct aac_sg_entryraw *sgp;
5326 	struct aac_sge *sge;
5327 
5328 	/* Calculate FIB size */
5329 	acp->fib_size = sizeof (struct aac_fib_header) + \
5330 	    sizeof (struct aac_raw_io) + (acp->left_cookien - 1) * \
5331 	    sizeof (struct aac_sg_entryraw);
5332 
5333 	aac_cmd_fib_header(softs, acp->slotp, RawIo, acp->fib_size);
5334 
5335 	ddi_put16(acc, &io->Flags, (acp->flags & AAC_CMD_BUF_READ) ? 1 : 0);
5336 	ddi_put16(acc, &io->BpTotal, 0);
5337 	ddi_put16(acc, &io->BpComplete, 0);
5338 
5339 	ddi_put32(acc, AAC_LO32(&io->BlockNumber), AAC_LS32(acp->blkno));
5340 	ddi_put32(acc, AAC_HI32(&io->BlockNumber), AAC_MS32(acp->blkno));
5341 	ddi_put16(acc, &io->ContainerId,
5342 	    ((struct aac_container *)acp->dvp)->cid);
5343 
5344 	/* Fill SG table */
5345 	ddi_put32(acc, &io->SgMapRaw.SgCount, acp->left_cookien);
5346 	ddi_put32(acc, &io->ByteCount, acp->bcount);
5347 
5348 	for (sge = &acp->sgt[0], sgp = &io->SgMapRaw.SgEntryRaw[0];
5349 	    sge < &acp->sgt[acp->left_cookien]; sge++, sgp++) {
5350 		ddi_put32(acc, AAC_LO32(&sgp->SgAddress), sge->addr.ad64.lo);
5351 		ddi_put32(acc, AAC_HI32(&sgp->SgAddress), sge->addr.ad64.hi);
5352 		ddi_put32(acc, &sgp->SgByteCount, sge->bcount);
5353 		sgp->Next = 0;
5354 		sgp->Prev = 0;
5355 		sgp->Flags = 0;
5356 	}
5357 }
5358 
5359 /* Init FIB for 64-bit block IO command */
5360 static void
5361 aac_cmd_fib_brw64(struct aac_softstate *softs, struct aac_cmd *acp)
5362 {
5363 	ddi_acc_handle_t acc = acp->slotp->fib_acc_handle;
5364 	struct aac_blockread64 *br = (struct aac_blockread64 *) \
5365 	    &acp->slotp->fibp->data[0];
5366 	struct aac_sg_entry64 *sgp;
5367 	struct aac_sge *sge;
5368 
5369 	acp->fib_size = sizeof (struct aac_fib_header) + \
5370 	    sizeof (struct aac_blockread64) + (acp->left_cookien - 1) * \
5371 	    sizeof (struct aac_sg_entry64);
5372 
5373 	aac_cmd_fib_header(softs, acp->slotp, ContainerCommand64,
5374 	    acp->fib_size);
5375 
5376 	/*
5377 	 * The definitions for aac_blockread64 and aac_blockwrite64
5378 	 * are the same.
5379 	 */
5380 	ddi_put32(acc, &br->BlockNumber, (uint32_t)acp->blkno);
5381 	ddi_put16(acc, &br->ContainerId,
5382 	    ((struct aac_container *)acp->dvp)->cid);
5383 	ddi_put32(acc, &br->Command, (acp->flags & AAC_CMD_BUF_READ) ?
5384 	    VM_CtHostRead64 : VM_CtHostWrite64);
5385 	ddi_put16(acc, &br->Pad, 0);
5386 	ddi_put16(acc, &br->Flags, 0);
5387 
5388 	/* Fill SG table */
5389 	ddi_put32(acc, &br->SgMap64.SgCount, acp->left_cookien);
5390 	ddi_put16(acc, &br->SectorCount, acp->bcount / AAC_BLK_SIZE);
5391 
5392 	for (sge = &acp->sgt[0], sgp = &br->SgMap64.SgEntry64[0];
5393 	    sge < &acp->sgt[acp->left_cookien]; sge++, sgp++) {
5394 		ddi_put32(acc, AAC_LO32(&sgp->SgAddress), sge->addr.ad64.lo);
5395 		ddi_put32(acc, AAC_HI32(&sgp->SgAddress), sge->addr.ad64.hi);
5396 		ddi_put32(acc, &sgp->SgByteCount, sge->bcount);
5397 	}
5398 }
5399 
5400 /* Init FIB for block IO command */
5401 static void
5402 aac_cmd_fib_brw(struct aac_softstate *softs, struct aac_cmd *acp)
5403 {
5404 	ddi_acc_handle_t acc = acp->slotp->fib_acc_handle;
5405 	struct aac_blockread *br = (struct aac_blockread *) \
5406 	    &acp->slotp->fibp->data[0];
5407 	struct aac_sg_entry *sgp;
5408 	struct aac_sge *sge = &acp->sgt[0];
5409 
5410 	if (acp->flags & AAC_CMD_BUF_READ) {
5411 		acp->fib_size = sizeof (struct aac_fib_header) + \
5412 		    sizeof (struct aac_blockread) + (acp->left_cookien - 1) * \
5413 		    sizeof (struct aac_sg_entry);
5414 
5415 		ddi_put32(acc, &br->Command, VM_CtBlockRead);
5416 		ddi_put32(acc, &br->SgMap.SgCount, acp->left_cookien);
5417 		sgp = &br->SgMap.SgEntry[0];
5418 	} else {
5419 		struct aac_blockwrite *bw = (struct aac_blockwrite *)br;
5420 
5421 		acp->fib_size = sizeof (struct aac_fib_header) + \
5422 		    sizeof (struct aac_blockwrite) + (acp->left_cookien - 1) * \
5423 		    sizeof (struct aac_sg_entry);
5424 
5425 		ddi_put32(acc, &bw->Command, VM_CtBlockWrite);
5426 		ddi_put32(acc, &bw->Stable, CUNSTABLE);
5427 		ddi_put32(acc, &bw->SgMap.SgCount, acp->left_cookien);
5428 		sgp = &bw->SgMap.SgEntry[0];
5429 	}
5430 	aac_cmd_fib_header(softs, acp->slotp, ContainerCommand, acp->fib_size);
5431 
5432 	/*
5433 	 * aac_blockread and aac_blockwrite have the similar
5434 	 * structure head, so use br for bw here
5435 	 */
5436 	ddi_put32(acc, &br->BlockNumber, (uint32_t)acp->blkno);
5437 	ddi_put32(acc, &br->ContainerId,
5438 	    ((struct aac_container *)acp->dvp)->cid);
5439 	ddi_put32(acc, &br->ByteCount, acp->bcount);
5440 
5441 	/* Fill SG table */
5442 	for (sge = &acp->sgt[0];
5443 	    sge < &acp->sgt[acp->left_cookien]; sge++, sgp++) {
5444 		ddi_put32(acc, &sgp->SgAddress, sge->addr.ad32);
5445 		ddi_put32(acc, &sgp->SgByteCount, sge->bcount);
5446 	}
5447 }
5448 
5449 /*ARGSUSED*/
5450 void
5451 aac_cmd_fib_copy(struct aac_softstate *softs, struct aac_cmd *acp)
5452 {
5453 	struct aac_slot *slotp = acp->slotp;
5454 	struct aac_fib *fibp = slotp->fibp;
5455 	ddi_acc_handle_t acc = slotp->fib_acc_handle;
5456 
5457 	ddi_rep_put8(acc, (uint8_t *)acp->fibp, (uint8_t *)fibp,
5458 	    acp->fib_size,   /* only copy data of needed length */
5459 	    DDI_DEV_AUTOINCR);
5460 	ddi_put32(acc, &fibp->Header.ReceiverFibAddress, slotp->fib_phyaddr);
5461 	ddi_put32(acc, &fibp->Header.SenderFibAddress, slotp->index << 2);
5462 }
5463 
5464 static void
5465 aac_cmd_fib_sync(struct aac_softstate *softs, struct aac_cmd *acp)
5466 {
5467 	struct aac_slot *slotp = acp->slotp;
5468 	ddi_acc_handle_t acc = slotp->fib_acc_handle;
5469 	struct aac_synchronize_command *sync =
5470 	    (struct aac_synchronize_command *)&slotp->fibp->data[0];
5471 
5472 	acp->fib_size = sizeof (struct aac_fib_header) + \
5473 	    sizeof (struct aac_synchronize_command);
5474 
5475 	aac_cmd_fib_header(softs, slotp, ContainerCommand, acp->fib_size);
5476 	ddi_put32(acc, &sync->Command, VM_ContainerConfig);
5477 	ddi_put32(acc, &sync->Type, (uint32_t)CT_FLUSH_CACHE);
5478 	ddi_put32(acc, &sync->Cid, ((struct aac_container *)acp->dvp)->cid);
5479 	ddi_put32(acc, &sync->Count,
5480 	    sizeof (((struct aac_synchronize_reply *)0)->Data));
5481 }
5482 
5483 /*
5484  * Init FIB for pass-through SCMD
5485  */
5486 static void
5487 aac_cmd_fib_srb(struct aac_cmd *acp)
5488 {
5489 	struct aac_slot *slotp = acp->slotp;
5490 	ddi_acc_handle_t acc = slotp->fib_acc_handle;
5491 	struct aac_srb *srb = (struct aac_srb *)&slotp->fibp->data[0];
5492 	uint8_t *cdb;
5493 
5494 	ddi_put32(acc, &srb->function, SRBF_ExecuteScsi);
5495 	ddi_put32(acc, &srb->retry_limit, 0);
5496 	ddi_put32(acc, &srb->cdb_size, acp->cmdlen);
5497 	ddi_put32(acc, &srb->timeout, 0); /* use driver timeout */
5498 	if (acp->fibp == NULL) {
5499 		if (acp->flags & AAC_CMD_BUF_READ)
5500 			ddi_put32(acc, &srb->flags, SRB_DataIn);
5501 		else if (acp->flags & AAC_CMD_BUF_WRITE)
5502 			ddi_put32(acc, &srb->flags, SRB_DataOut);
5503 		ddi_put32(acc, &srb->channel,
5504 		    ((struct aac_nondasd *)acp->dvp)->bus);
5505 		ddi_put32(acc, &srb->id, ((struct aac_nondasd *)acp->dvp)->tid);
5506 		ddi_put32(acc, &srb->lun, 0);
5507 		cdb = acp->pkt->pkt_cdbp;
5508 	} else {
5509 		struct aac_srb *srb0 = (struct aac_srb *)&acp->fibp->data[0];
5510 
5511 		ddi_put32(acc, &srb->flags, srb0->flags);
5512 		ddi_put32(acc, &srb->channel, srb0->channel);
5513 		ddi_put32(acc, &srb->id, srb0->id);
5514 		ddi_put32(acc, &srb->lun, srb0->lun);
5515 		cdb = srb0->cdb;
5516 	}
5517 	ddi_rep_put8(acc, cdb, srb->cdb, acp->cmdlen, DDI_DEV_AUTOINCR);
5518 }
5519 
5520 static void
5521 aac_cmd_fib_scsi32(struct aac_softstate *softs, struct aac_cmd *acp)
5522 {
5523 	ddi_acc_handle_t acc = acp->slotp->fib_acc_handle;
5524 	struct aac_srb *srb = (struct aac_srb *)&acp->slotp->fibp->data[0];
5525 	struct aac_sg_entry *sgp;
5526 	struct aac_sge *sge;
5527 
5528 	acp->fib_size = sizeof (struct aac_fib_header) + \
5529 	    sizeof (struct aac_srb) - sizeof (struct aac_sg_entry) + \
5530 	    acp->left_cookien * sizeof (struct aac_sg_entry);
5531 
5532 	/* Fill FIB and SRB headers, and copy cdb */
5533 	aac_cmd_fib_header(softs, acp->slotp, ScsiPortCommand, acp->fib_size);
5534 	aac_cmd_fib_srb(acp);
5535 
5536 	/* Fill SG table */
5537 	ddi_put32(acc, &srb->sg.SgCount, acp->left_cookien);
5538 	ddi_put32(acc, &srb->count, acp->bcount);
5539 
5540 	for (sge = &acp->sgt[0], sgp = &srb->sg.SgEntry[0];
5541 	    sge < &acp->sgt[acp->left_cookien]; sge++, sgp++) {
5542 		ddi_put32(acc, &sgp->SgAddress, sge->addr.ad32);
5543 		ddi_put32(acc, &sgp->SgByteCount, sge->bcount);
5544 	}
5545 }
5546 
5547 static void
5548 aac_cmd_fib_scsi64(struct aac_softstate *softs, struct aac_cmd *acp)
5549 {
5550 	ddi_acc_handle_t acc = acp->slotp->fib_acc_handle;
5551 	struct aac_srb *srb = (struct aac_srb *)&acp->slotp->fibp->data[0];
5552 	struct aac_sg_entry64 *sgp;
5553 	struct aac_sge *sge;
5554 
5555 	acp->fib_size = sizeof (struct aac_fib_header) + \
5556 	    sizeof (struct aac_srb) - sizeof (struct aac_sg_entry) + \
5557 	    acp->left_cookien * sizeof (struct aac_sg_entry64);
5558 
5559 	/* Fill FIB and SRB headers, and copy cdb */
5560 	aac_cmd_fib_header(softs, acp->slotp, ScsiPortCommandU64,
5561 	    acp->fib_size);
5562 	aac_cmd_fib_srb(acp);
5563 
5564 	/* Fill SG table */
5565 	ddi_put32(acc, &srb->sg.SgCount, acp->left_cookien);
5566 	ddi_put32(acc, &srb->count, acp->bcount);
5567 
5568 	for (sge = &acp->sgt[0],
5569 	    sgp = &((struct aac_sg_table64 *)&srb->sg)->SgEntry64[0];
5570 	    sge < &acp->sgt[acp->left_cookien]; sge++, sgp++) {
5571 		ddi_put32(acc, AAC_LO32(&sgp->SgAddress), sge->addr.ad64.lo);
5572 		ddi_put32(acc, AAC_HI32(&sgp->SgAddress), sge->addr.ad64.hi);
5573 		ddi_put32(acc, &sgp->SgByteCount, sge->bcount);
5574 	}
5575 }
5576 
5577 static int
5578 aac_cmd_slot_bind(struct aac_softstate *softs, struct aac_cmd *acp)
5579 {
5580 	struct aac_slot *slotp;
5581 
5582 	if (slotp = aac_get_slot(softs)) {
5583 		acp->slotp = slotp;
5584 		slotp->acp = acp;
5585 		acp->aac_cmd_fib(softs, acp);
5586 		(void) ddi_dma_sync(slotp->fib_dma_handle, 0, 0,
5587 		    DDI_DMA_SYNC_FORDEV);
5588 		return (AACOK);
5589 	}
5590 	return (AACERR);
5591 }
5592 
5593 static int
5594 aac_bind_io(struct aac_softstate *softs, struct aac_cmd *acp)
5595 {
5596 	struct aac_device *dvp = acp->dvp;
5597 	int q = AAC_CMDQ(acp);
5598 
5599 	if (dvp) {
5600 		if (dvp->ncmds[q] < dvp->throttle[q]) {
5601 			if (!(acp->flags & AAC_CMD_NTAG) ||
5602 			    dvp->ncmds[q] == 0) {
5603 do_bind:
5604 				return (aac_cmd_slot_bind(softs, acp));
5605 			}
5606 			ASSERT(q == AAC_CMDQ_ASYNC);
5607 			aac_set_throttle(softs, dvp, AAC_CMDQ_ASYNC,
5608 			    AAC_THROTTLE_DRAIN);
5609 		}
5610 	} else {
5611 		if (softs->bus_ncmds[q] < softs->bus_throttle[q])
5612 			goto do_bind;
5613 	}
5614 	return (AACERR);
5615 }
5616 
5617 static void
5618 aac_start_io(struct aac_softstate *softs, struct aac_cmd *acp)
5619 {
5620 	struct aac_slot *slotp = acp->slotp;
5621 	int q = AAC_CMDQ(acp);
5622 	int rval;
5623 
5624 	/* Set ac and pkt */
5625 	if (acp->pkt) { /* ac from ioctl has no pkt */
5626 		acp->pkt->pkt_state |=
5627 		    STATE_GOT_BUS | STATE_GOT_TARGET | STATE_SENT_CMD;
5628 	}
5629 	if (acp->timeout) /* 0 indicates no timeout */
5630 		acp->timeout += aac_timebase + aac_tick;
5631 
5632 	if (acp->dvp)
5633 		acp->dvp->ncmds[q]++;
5634 	softs->bus_ncmds[q]++;
5635 	aac_cmd_enqueue(&softs->q_busy, acp);
5636 
5637 	AACDB_PRINT_FIB(softs, slotp);
5638 
5639 	if (softs->flags & AAC_FLAGS_NEW_COMM) {
5640 		rval = aac_send_command(softs, slotp);
5641 	} else {
5642 		/*
5643 		 * If fib can not be enqueued, the adapter is in an abnormal
5644 		 * state, there will be no interrupt to us.
5645 		 */
5646 		rval = aac_fib_enqueue(softs, AAC_ADAP_NORM_CMD_Q,
5647 		    slotp->fib_phyaddr, acp->fib_size);
5648 	}
5649 
5650 	if (aac_check_dma_handle(slotp->fib_dma_handle) != DDI_SUCCESS)
5651 		ddi_fm_service_impact(softs->devinfo_p, DDI_SERVICE_UNAFFECTED);
5652 
5653 	/*
5654 	 * NOTE: We send command only when slots availabe, so should never
5655 	 * reach here.
5656 	 */
5657 	if (rval != AACOK) {
5658 		AACDB_PRINT(softs, CE_NOTE, "SCMD send failed");
5659 		if (acp->pkt) {
5660 			acp->pkt->pkt_state &= ~STATE_SENT_CMD;
5661 			aac_set_pkt_reason(softs, acp, CMD_INCOMPLETE, 0);
5662 		}
5663 		aac_end_io(softs, acp);
5664 		if (!(acp->flags & (AAC_CMD_NO_INTR | AAC_CMD_NO_CB)))
5665 			ddi_trigger_softintr(softs->softint_id);
5666 	}
5667 }
5668 
5669 static void
5670 aac_start_waitq(struct aac_softstate *softs, struct aac_cmd_queue *q)
5671 {
5672 	struct aac_cmd *acp, *next_acp;
5673 
5674 	/* Serve as many waiting io's as possible */
5675 	for (acp = q->q_head; acp; acp = next_acp) {
5676 		next_acp = acp->next;
5677 		if (aac_bind_io(softs, acp) == AACOK) {
5678 			aac_cmd_delete(q, acp);
5679 			aac_start_io(softs, acp);
5680 		}
5681 		if (softs->free_io_slot_head == NULL)
5682 			break;
5683 	}
5684 }
5685 
5686 static void
5687 aac_start_waiting_io(struct aac_softstate *softs)
5688 {
5689 	/*
5690 	 * Sync FIB io is served before async FIB io so that io requests
5691 	 * sent by interactive userland commands get responded asap.
5692 	 */
5693 	if (softs->q_wait[AAC_CMDQ_SYNC].q_head)
5694 		aac_start_waitq(softs, &softs->q_wait[AAC_CMDQ_SYNC]);
5695 	if (softs->q_wait[AAC_CMDQ_ASYNC].q_head)
5696 		aac_start_waitq(softs, &softs->q_wait[AAC_CMDQ_ASYNC]);
5697 }
5698 
5699 static void
5700 aac_drain_comp_q(struct aac_softstate *softs)
5701 {
5702 	struct aac_cmd *acp;
5703 	struct scsi_pkt *pkt;
5704 
5705 	/*CONSTCOND*/
5706 	while (1) {
5707 		mutex_enter(&softs->q_comp_mutex);
5708 		acp = aac_cmd_dequeue(&softs->q_comp);
5709 		mutex_exit(&softs->q_comp_mutex);
5710 		if (acp != NULL) {
5711 			ASSERT(acp->pkt != NULL);
5712 			pkt = acp->pkt;
5713 
5714 			if (pkt->pkt_reason == CMD_CMPLT) {
5715 				/*
5716 				 * Consistent packets need to be sync'ed first
5717 				 */
5718 				if ((acp->flags & AAC_CMD_CONSISTENT) &&
5719 				    (acp->flags & AAC_CMD_BUF_READ)) {
5720 					if (aac_dma_sync_ac(acp) != AACOK) {
5721 						ddi_fm_service_impact(
5722 						    softs->devinfo_p,
5723 						    DDI_SERVICE_UNAFFECTED);
5724 						pkt->pkt_reason = CMD_TRAN_ERR;
5725 						pkt->pkt_statistics = 0;
5726 					}
5727 				}
5728 				if ((aac_check_acc_handle(softs-> \
5729 				    comm_space_acc_handle) != DDI_SUCCESS) ||
5730 				    (aac_check_acc_handle(softs-> \
5731 				    pci_mem_handle) != DDI_SUCCESS)) {
5732 					ddi_fm_service_impact(softs->devinfo_p,
5733 					    DDI_SERVICE_UNAFFECTED);
5734 					ddi_fm_acc_err_clear(softs-> \
5735 					    pci_mem_handle, DDI_FME_VER0);
5736 					pkt->pkt_reason = CMD_TRAN_ERR;
5737 					pkt->pkt_statistics = 0;
5738 				}
5739 				if (aac_check_dma_handle(softs-> \
5740 				    comm_space_dma_handle) != DDI_SUCCESS) {
5741 					ddi_fm_service_impact(softs->devinfo_p,
5742 					    DDI_SERVICE_UNAFFECTED);
5743 					pkt->pkt_reason = CMD_TRAN_ERR;
5744 					pkt->pkt_statistics = 0;
5745 				}
5746 			}
5747 			scsi_hba_pkt_comp(pkt);
5748 		} else {
5749 			break;
5750 		}
5751 	}
5752 }
5753 
5754 static int
5755 aac_alloc_fib(struct aac_softstate *softs, struct aac_slot *slotp)
5756 {
5757 	size_t rlen;
5758 	ddi_dma_cookie_t cookie;
5759 	uint_t cookien;
5760 
5761 	/* Allocate FIB dma resource */
5762 	if (ddi_dma_alloc_handle(
5763 	    softs->devinfo_p,
5764 	    &softs->addr_dma_attr,
5765 	    DDI_DMA_SLEEP,
5766 	    NULL,
5767 	    &slotp->fib_dma_handle) != DDI_SUCCESS) {
5768 		AACDB_PRINT(softs, CE_WARN,
5769 		    "Cannot alloc dma handle for slot fib area");
5770 		goto error;
5771 	}
5772 	if (ddi_dma_mem_alloc(
5773 	    slotp->fib_dma_handle,
5774 	    softs->aac_max_fib_size,
5775 	    &softs->acc_attr,
5776 	    DDI_DMA_RDWR | DDI_DMA_CONSISTENT,
5777 	    DDI_DMA_SLEEP,
5778 	    NULL,
5779 	    (caddr_t *)&slotp->fibp,
5780 	    &rlen,
5781 	    &slotp->fib_acc_handle) != DDI_SUCCESS) {
5782 		AACDB_PRINT(softs, CE_WARN,
5783 		    "Cannot alloc mem for slot fib area");
5784 		goto error;
5785 	}
5786 	if (ddi_dma_addr_bind_handle(
5787 	    slotp->fib_dma_handle,
5788 	    NULL,
5789 	    (caddr_t)slotp->fibp,
5790 	    softs->aac_max_fib_size,
5791 	    DDI_DMA_RDWR | DDI_DMA_CONSISTENT,
5792 	    DDI_DMA_SLEEP,
5793 	    NULL,
5794 	    &cookie,
5795 	    &cookien) != DDI_DMA_MAPPED) {
5796 		AACDB_PRINT(softs, CE_WARN,
5797 		    "dma bind failed for slot fib area");
5798 		goto error;
5799 	}
5800 
5801 	/* Check dma handles allocated in fib attach */
5802 	if (aac_check_dma_handle(slotp->fib_dma_handle) != DDI_SUCCESS) {
5803 		ddi_fm_service_impact(softs->devinfo_p, DDI_SERVICE_LOST);
5804 		goto error;
5805 	}
5806 
5807 	/* Check acc handles allocated in fib attach */
5808 	if (aac_check_acc_handle(slotp->fib_acc_handle) != DDI_SUCCESS) {
5809 		ddi_fm_service_impact(softs->devinfo_p, DDI_SERVICE_LOST);
5810 		goto error;
5811 	}
5812 
5813 	slotp->fib_phyaddr = cookie.dmac_laddress;
5814 	return (AACOK);
5815 
5816 error:
5817 	if (slotp->fib_acc_handle) {
5818 		ddi_dma_mem_free(&slotp->fib_acc_handle);
5819 		slotp->fib_acc_handle = NULL;
5820 	}
5821 	if (slotp->fib_dma_handle) {
5822 		ddi_dma_free_handle(&slotp->fib_dma_handle);
5823 		slotp->fib_dma_handle = NULL;
5824 	}
5825 	return (AACERR);
5826 }
5827 
5828 static void
5829 aac_free_fib(struct aac_slot *slotp)
5830 {
5831 	(void) ddi_dma_unbind_handle(slotp->fib_dma_handle);
5832 	ddi_dma_mem_free(&slotp->fib_acc_handle);
5833 	slotp->fib_acc_handle = NULL;
5834 	ddi_dma_free_handle(&slotp->fib_dma_handle);
5835 	slotp->fib_dma_handle = NULL;
5836 	slotp->fib_phyaddr = 0;
5837 }
5838 
5839 static void
5840 aac_alloc_fibs(struct aac_softstate *softs)
5841 {
5842 	int i;
5843 	struct aac_slot *slotp;
5844 
5845 	for (i = 0; i < softs->total_slots &&
5846 	    softs->total_fibs < softs->total_slots; i++) {
5847 		slotp = &(softs->io_slot[i]);
5848 		if (slotp->fib_phyaddr)
5849 			continue;
5850 		if (aac_alloc_fib(softs, slotp) != AACOK)
5851 			break;
5852 
5853 		/* Insert the slot to the free slot list */
5854 		aac_release_slot(softs, slotp);
5855 		softs->total_fibs++;
5856 	}
5857 }
5858 
5859 static void
5860 aac_destroy_fibs(struct aac_softstate *softs)
5861 {
5862 	struct aac_slot *slotp;
5863 
5864 	while ((slotp = softs->free_io_slot_head) != NULL) {
5865 		ASSERT(slotp->fib_phyaddr);
5866 		softs->free_io_slot_head = slotp->next;
5867 		aac_free_fib(slotp);
5868 		ASSERT(slotp->index == (slotp - softs->io_slot));
5869 		softs->total_fibs--;
5870 	}
5871 	ASSERT(softs->total_fibs == 0);
5872 }
5873 
5874 static int
5875 aac_create_slots(struct aac_softstate *softs)
5876 {
5877 	int i;
5878 
5879 	softs->total_slots = softs->aac_max_fibs;
5880 	softs->io_slot = kmem_zalloc(sizeof (struct aac_slot) * \
5881 	    softs->total_slots, KM_SLEEP);
5882 	if (softs->io_slot == NULL) {
5883 		AACDB_PRINT(softs, CE_WARN, "Cannot allocate slot");
5884 		return (AACERR);
5885 	}
5886 	for (i = 0; i < softs->total_slots; i++)
5887 		softs->io_slot[i].index = i;
5888 	softs->free_io_slot_head = NULL;
5889 	softs->total_fibs = 0;
5890 	return (AACOK);
5891 }
5892 
5893 static void
5894 aac_destroy_slots(struct aac_softstate *softs)
5895 {
5896 	ASSERT(softs->free_io_slot_head == NULL);
5897 
5898 	kmem_free(softs->io_slot, sizeof (struct aac_slot) * \
5899 	    softs->total_slots);
5900 	softs->io_slot = NULL;
5901 	softs->total_slots = 0;
5902 }
5903 
5904 struct aac_slot *
5905 aac_get_slot(struct aac_softstate *softs)
5906 {
5907 	struct aac_slot *slotp;
5908 
5909 	if ((slotp = softs->free_io_slot_head) != NULL) {
5910 		softs->free_io_slot_head = slotp->next;
5911 		slotp->next = NULL;
5912 	}
5913 	return (slotp);
5914 }
5915 
5916 static void
5917 aac_release_slot(struct aac_softstate *softs, struct aac_slot *slotp)
5918 {
5919 	ASSERT((slotp->index >= 0) && (slotp->index < softs->total_slots));
5920 	ASSERT(slotp == &softs->io_slot[slotp->index]);
5921 
5922 	slotp->acp = NULL;
5923 	slotp->next = softs->free_io_slot_head;
5924 	softs->free_io_slot_head = slotp;
5925 }
5926 
5927 int
5928 aac_do_io(struct aac_softstate *softs, struct aac_cmd *acp)
5929 {
5930 	if (aac_bind_io(softs, acp) == AACOK)
5931 		aac_start_io(softs, acp);
5932 	else
5933 		aac_cmd_enqueue(&softs->q_wait[AAC_CMDQ(acp)], acp);
5934 
5935 	if (!(acp->flags & (AAC_CMD_NO_CB | AAC_CMD_NO_INTR)))
5936 		return (TRAN_ACCEPT);
5937 	/*
5938 	 * Because sync FIB is always 512 bytes and used for critical
5939 	 * functions, async FIB is used for poll IO.
5940 	 */
5941 	if (acp->flags & AAC_CMD_NO_INTR) {
5942 		if (aac_do_poll_io(softs, acp) == AACOK)
5943 			return (TRAN_ACCEPT);
5944 	} else {
5945 		if (aac_do_sync_io(softs, acp) == AACOK)
5946 			return (TRAN_ACCEPT);
5947 	}
5948 	return (TRAN_BADPKT);
5949 }
5950 
5951 static int
5952 aac_do_poll_io(struct aac_softstate *softs, struct aac_cmd *acp)
5953 {
5954 	int (*intr_handler)(struct aac_softstate *);
5955 
5956 	/*
5957 	 * Interrupt is disabled, we have to poll the adapter by ourselves.
5958 	 */
5959 	intr_handler = (softs->flags & AAC_FLAGS_NEW_COMM) ?
5960 	    aac_process_intr_new : aac_process_intr_old;
5961 	while (!(acp->flags & (AAC_CMD_CMPLT | AAC_CMD_ABORT))) {
5962 		int i = AAC_POLL_TIME * 1000;
5963 
5964 		AAC_BUSYWAIT((intr_handler(softs) != AAC_DB_RESPONSE_READY), i);
5965 		if (i == 0)
5966 			aac_cmd_timeout(softs, acp);
5967 	}
5968 
5969 	ddi_trigger_softintr(softs->softint_id);
5970 
5971 	if ((acp->flags & AAC_CMD_CMPLT) && !(acp->flags & AAC_CMD_ERR))
5972 		return (AACOK);
5973 	return (AACERR);
5974 }
5975 
5976 static int
5977 aac_do_sync_io(struct aac_softstate *softs, struct aac_cmd *acp)
5978 {
5979 	ASSERT(softs && acp);
5980 
5981 	while (!(acp->flags & (AAC_CMD_CMPLT | AAC_CMD_ABORT)))
5982 		cv_wait(&softs->event, &softs->io_lock);
5983 
5984 	if (acp->flags & AAC_CMD_CMPLT)
5985 		return (AACOK);
5986 	return (AACERR);
5987 }
5988 
5989 static int
5990 aac_dma_sync_ac(struct aac_cmd *acp)
5991 {
5992 	if (acp->buf_dma_handle) {
5993 		if (acp->flags & AAC_CMD_BUF_WRITE) {
5994 			if (acp->abp != NULL)
5995 				ddi_rep_put8(acp->abh,
5996 				    (uint8_t *)acp->bp->b_un.b_addr,
5997 				    (uint8_t *)acp->abp, acp->bp->b_bcount,
5998 				    DDI_DEV_AUTOINCR);
5999 			(void) ddi_dma_sync(acp->buf_dma_handle, 0, 0,
6000 			    DDI_DMA_SYNC_FORDEV);
6001 		} else {
6002 			(void) ddi_dma_sync(acp->buf_dma_handle, 0, 0,
6003 			    DDI_DMA_SYNC_FORCPU);
6004 			if (aac_check_dma_handle(acp->buf_dma_handle) !=
6005 			    DDI_SUCCESS)
6006 				return (AACERR);
6007 			if (acp->abp != NULL)
6008 				ddi_rep_get8(acp->abh,
6009 				    (uint8_t *)acp->bp->b_un.b_addr,
6010 				    (uint8_t *)acp->abp, acp->bp->b_bcount,
6011 				    DDI_DEV_AUTOINCR);
6012 		}
6013 	}
6014 	return (AACOK);
6015 }
6016 
6017 /*
6018  * The following function comes from Adaptec:
6019  *
6020  * When driver sees a particular event that means containers are changed, it
6021  * will rescan containers. However a change may not be complete until some
6022  * other event is received. For example, creating or deleting an array will
6023  * incur as many as six AifEnConfigChange events which would generate six
6024  * container rescans. To diminish rescans, driver set a flag to wait for
6025  * another particular event. When sees that events come in, it will do rescan.
6026  */
6027 static int
6028 aac_handle_aif(struct aac_softstate *softs, struct aac_fib *fibp)
6029 {
6030 	ddi_acc_handle_t acc = softs->comm_space_acc_handle;
6031 	uint16_t fib_command;
6032 	struct aac_aif_command *aif;
6033 	int en_type;
6034 	int devcfg_needed;
6035 	int current, next;
6036 
6037 	fib_command = LE_16(fibp->Header.Command);
6038 	if (fib_command != AifRequest) {
6039 		cmn_err(CE_NOTE, "!Unknown command from controller: 0x%x",
6040 		    fib_command);
6041 		return (AACERR);
6042 	}
6043 
6044 	/* Update internal container state */
6045 	aif = (struct aac_aif_command *)&fibp->data[0];
6046 
6047 	AACDB_PRINT_AIF(softs, aif);
6048 	devcfg_needed = 0;
6049 	en_type = LE_32((uint32_t)aif->data.EN.type);
6050 
6051 	switch (LE_32((uint32_t)aif->command)) {
6052 	case AifCmdDriverNotify: {
6053 		int cid = LE_32(aif->data.EN.data.ECC.container[0]);
6054 
6055 		switch (en_type) {
6056 		case AifDenMorphComplete:
6057 		case AifDenVolumeExtendComplete:
6058 			if (AAC_DEV_IS_VALID(&softs->containers[cid].dev))
6059 				softs->devcfg_wait_on = AifEnConfigChange;
6060 			break;
6061 		}
6062 		if (softs->devcfg_wait_on == en_type)
6063 			devcfg_needed = 1;
6064 		break;
6065 	}
6066 
6067 	case AifCmdEventNotify:
6068 		switch (en_type) {
6069 		case AifEnAddContainer:
6070 		case AifEnDeleteContainer:
6071 			softs->devcfg_wait_on = AifEnConfigChange;
6072 			break;
6073 		case AifEnContainerChange:
6074 			if (!softs->devcfg_wait_on)
6075 				softs->devcfg_wait_on = AifEnConfigChange;
6076 			break;
6077 		case AifEnContainerEvent:
6078 			if (ddi_get32(acc, &aif-> \
6079 			    data.EN.data.ECE.eventType) == CT_PUP_MISSING_DRIVE)
6080 				devcfg_needed = 1;
6081 			break;
6082 		}
6083 		if (softs->devcfg_wait_on == en_type)
6084 			devcfg_needed = 1;
6085 		break;
6086 
6087 	case AifCmdJobProgress:
6088 		if (LE_32((uint32_t)aif->data.PR[0].jd.type) == AifJobCtrZero) {
6089 			int pr_status;
6090 			uint32_t pr_ftick, pr_ctick;
6091 
6092 			pr_status = LE_32((uint32_t)aif->data.PR[0].status);
6093 			pr_ctick = LE_32(aif->data.PR[0].currentTick);
6094 			pr_ftick = LE_32(aif->data.PR[0].finalTick);
6095 
6096 			if ((pr_ctick == pr_ftick) ||
6097 			    (pr_status == AifJobStsSuccess))
6098 				softs->devcfg_wait_on = AifEnContainerChange;
6099 			else if ((pr_ctick == 0) &&
6100 			    (pr_status == AifJobStsRunning))
6101 				softs->devcfg_wait_on = AifEnContainerChange;
6102 		}
6103 		break;
6104 	}
6105 
6106 	if (devcfg_needed) {
6107 		softs->devcfg_wait_on = 0;
6108 		(void) aac_probe_containers(softs);
6109 	}
6110 
6111 	/* Modify AIF contexts */
6112 	current = softs->aifq_idx;
6113 	next = (current + 1) % AAC_AIFQ_LENGTH;
6114 	if (next == 0) {
6115 		struct aac_fib_context *ctx;
6116 
6117 		softs->aifq_wrap = 1;
6118 		for (ctx = softs->fibctx; ctx; ctx = ctx->next) {
6119 			if (next == ctx->ctx_idx) {
6120 				ctx->ctx_filled = 1;
6121 			} else if (current == ctx->ctx_idx && ctx->ctx_filled) {
6122 				ctx->ctx_idx = next;
6123 				AACDB_PRINT(softs, CE_NOTE,
6124 				    "-- AIF queue(%x) overrun", ctx->unique);
6125 			}
6126 		}
6127 	}
6128 	softs->aifq_idx = next;
6129 
6130 	/* Wakeup applications */
6131 	cv_broadcast(&softs->aifv);
6132 	return (AACOK);
6133 }
6134 
6135 /*
6136  * Timeout recovery
6137  */
6138 /*ARGSUSED*/
6139 static void
6140 aac_cmd_timeout(struct aac_softstate *softs, struct aac_cmd *acp)
6141 {
6142 #ifdef DEBUG
6143 	acp->fib_flags |= AACDB_FLAGS_FIB_TIMEOUT;
6144 	AACDB_PRINT(softs, CE_WARN, "acp %p timed out", acp);
6145 	AACDB_PRINT_FIB(softs, acp->slotp);
6146 #endif
6147 
6148 	/*
6149 	 * Besides the firmware in unhealthy state, an overloaded
6150 	 * adapter may also incur pkt timeout.
6151 	 * There is a chance for an adapter with a slower IOP to take
6152 	 * longer than 60 seconds to process the commands, such as when
6153 	 * to perform IOs. So the adapter is doing a build on a RAID-5
6154 	 * while being required longer completion times should be
6155 	 * tolerated.
6156 	 */
6157 	switch (aac_do_reset(softs)) {
6158 	case AAC_IOP_RESET_SUCCEED:
6159 		aac_abort_iocmds(softs, AAC_IOCMD_OUTSTANDING, NULL, CMD_RESET);
6160 		aac_start_waiting_io(softs);
6161 		break;
6162 	case AAC_IOP_RESET_FAILED:
6163 		/* Abort all waiting cmds when adapter is dead */
6164 		aac_abort_iocmds(softs, AAC_IOCMD_ALL, NULL, CMD_TIMEOUT);
6165 		break;
6166 	case AAC_IOP_RESET_ABNORMAL:
6167 		aac_start_waiting_io(softs);
6168 	}
6169 }
6170 
6171 /*
6172  * The following function comes from Adaptec:
6173  *
6174  * Time sync. command added to synchronize time with firmware every 30
6175  * minutes (required for correct AIF timestamps etc.)
6176  */
6177 static int
6178 aac_sync_tick(struct aac_softstate *softs)
6179 {
6180 	ddi_acc_handle_t acc = softs->sync_slot.fib_acc_handle;
6181 	struct aac_fib *fibp = softs->sync_slot.fibp;
6182 
6183 	ddi_put32(acc, (void *)&fibp->data[0], ddi_get_time());
6184 	return (aac_sync_fib(softs, SendHostTime, AAC_FIB_SIZEOF(uint32_t)));
6185 }
6186 
6187 static void
6188 aac_daemon(void *arg)
6189 {
6190 	struct aac_softstate *softs = (struct aac_softstate *)arg;
6191 	struct aac_cmd *acp;
6192 
6193 	DBCALLED(softs, 2);
6194 
6195 	mutex_enter(&softs->io_lock);
6196 	/* Check slot for timeout pkts */
6197 	aac_timebase += aac_tick;
6198 	for (acp = softs->q_busy.q_head; acp; acp = acp->next) {
6199 		if (acp->timeout) {
6200 			if (acp->timeout <= aac_timebase) {
6201 				aac_cmd_timeout(softs, acp);
6202 				ddi_trigger_softintr(softs->softint_id);
6203 			}
6204 			break;
6205 		}
6206 	}
6207 
6208 	/* Time sync. with firmware every AAC_SYNC_TICK */
6209 	if (aac_sync_time <= aac_timebase) {
6210 		aac_sync_time = aac_timebase;
6211 		if (aac_sync_tick(softs) != AACOK)
6212 			aac_sync_time += aac_tick << 1; /* retry shortly */
6213 		else
6214 			aac_sync_time += AAC_SYNC_TICK;
6215 	}
6216 
6217 	if ((softs->state & AAC_STATE_RUN) && (softs->timeout_id != 0))
6218 		softs->timeout_id = timeout(aac_daemon, (void *)softs,
6219 		    (aac_tick * drv_usectohz(1000000)));
6220 	mutex_exit(&softs->io_lock);
6221 }
6222 
6223 /*
6224  * Architecture dependent functions
6225  */
6226 static int
6227 aac_rx_get_fwstatus(struct aac_softstate *softs)
6228 {
6229 	return (PCI_MEM_GET32(softs, AAC_OMR0));
6230 }
6231 
6232 static int
6233 aac_rx_get_mailbox(struct aac_softstate *softs, int mb)
6234 {
6235 	return (PCI_MEM_GET32(softs, AAC_RX_MAILBOX + mb * 4));
6236 }
6237 
6238 static void
6239 aac_rx_set_mailbox(struct aac_softstate *softs, uint32_t cmd,
6240     uint32_t arg0, uint32_t arg1, uint32_t arg2, uint32_t arg3)
6241 {
6242 	PCI_MEM_PUT32(softs, AAC_RX_MAILBOX, cmd);
6243 	PCI_MEM_PUT32(softs, AAC_RX_MAILBOX + 4, arg0);
6244 	PCI_MEM_PUT32(softs, AAC_RX_MAILBOX + 8, arg1);
6245 	PCI_MEM_PUT32(softs, AAC_RX_MAILBOX + 12, arg2);
6246 	PCI_MEM_PUT32(softs, AAC_RX_MAILBOX + 16, arg3);
6247 }
6248 
6249 static int
6250 aac_rkt_get_fwstatus(struct aac_softstate *softs)
6251 {
6252 	return (PCI_MEM_GET32(softs, AAC_OMR0));
6253 }
6254 
6255 static int
6256 aac_rkt_get_mailbox(struct aac_softstate *softs, int mb)
6257 {
6258 	return (PCI_MEM_GET32(softs, AAC_RKT_MAILBOX + mb *4));
6259 }
6260 
6261 static void
6262 aac_rkt_set_mailbox(struct aac_softstate *softs, uint32_t cmd,
6263     uint32_t arg0, uint32_t arg1, uint32_t arg2, uint32_t arg3)
6264 {
6265 	PCI_MEM_PUT32(softs, AAC_RKT_MAILBOX, cmd);
6266 	PCI_MEM_PUT32(softs, AAC_RKT_MAILBOX + 4, arg0);
6267 	PCI_MEM_PUT32(softs, AAC_RKT_MAILBOX + 8, arg1);
6268 	PCI_MEM_PUT32(softs, AAC_RKT_MAILBOX + 12, arg2);
6269 	PCI_MEM_PUT32(softs, AAC_RKT_MAILBOX + 16, arg3);
6270 }
6271 
6272 /*
6273  * cb_ops functions
6274  */
6275 static int
6276 aac_open(dev_t *devp, int flag, int otyp, cred_t *cred)
6277 {
6278 	struct aac_softstate *softs;
6279 	int minor0, minor;
6280 	int instance;
6281 
6282 	DBCALLED(NULL, 2);
6283 
6284 	if (otyp != OTYP_BLK && otyp != OTYP_CHR)
6285 		return (EINVAL);
6286 
6287 	minor0 = getminor(*devp);
6288 	minor = AAC_SCSA_MINOR(minor0);
6289 
6290 	if (AAC_IS_SCSA_NODE(minor))
6291 		return (scsi_hba_open(devp, flag, otyp, cred));
6292 
6293 	instance = MINOR2INST(minor0);
6294 	if (instance >= AAC_MAX_ADAPTERS)
6295 		return (ENXIO);
6296 
6297 	softs = ddi_get_soft_state(aac_softstatep, instance);
6298 	if (softs == NULL)
6299 		return (ENXIO);
6300 
6301 	return (0);
6302 }
6303 
6304 /*ARGSUSED*/
6305 static int
6306 aac_close(dev_t dev, int flag, int otyp, cred_t *cred)
6307 {
6308 	int minor0, minor;
6309 	int instance;
6310 
6311 	DBCALLED(NULL, 2);
6312 
6313 	if (otyp != OTYP_BLK && otyp != OTYP_CHR)
6314 		return (EINVAL);
6315 
6316 	minor0 = getminor(dev);
6317 	minor = AAC_SCSA_MINOR(minor0);
6318 
6319 	if (AAC_IS_SCSA_NODE(minor))
6320 		return (scsi_hba_close(dev, flag, otyp, cred));
6321 
6322 	instance = MINOR2INST(minor0);
6323 	if (instance >= AAC_MAX_ADAPTERS)
6324 		return (ENXIO);
6325 
6326 	return (0);
6327 }
6328 
6329 static int
6330 aac_ioctl(dev_t dev, int cmd, intptr_t arg, int flag, cred_t *cred_p,
6331     int *rval_p)
6332 {
6333 	struct aac_softstate *softs;
6334 	int minor0, minor;
6335 	int instance;
6336 
6337 	DBCALLED(NULL, 2);
6338 
6339 	if (drv_priv(cred_p) != 0)
6340 		return (EPERM);
6341 
6342 	minor0 = getminor(dev);
6343 	minor = AAC_SCSA_MINOR(minor0);
6344 
6345 	if (AAC_IS_SCSA_NODE(minor))
6346 		return (scsi_hba_ioctl(dev, cmd, arg, flag, cred_p, rval_p));
6347 
6348 	instance = MINOR2INST(minor0);
6349 	if (instance < AAC_MAX_ADAPTERS) {
6350 		softs = ddi_get_soft_state(aac_softstatep, instance);
6351 		return (aac_do_ioctl(softs, dev, cmd, arg, flag));
6352 	}
6353 	return (ENXIO);
6354 }
6355 
6356 /*
6357  * The IO fault service error handling callback function
6358  */
6359 /*ARGSUSED*/
6360 static int
6361 aac_fm_error_cb(dev_info_t *dip, ddi_fm_error_t *err, const void *impl_data)
6362 {
6363 	/*
6364 	 * as the driver can always deal with an error in any dma or
6365 	 * access handle, we can just return the fme_status value.
6366 	 */
6367 	pci_ereport_post(dip, err, NULL);
6368 	return (err->fme_status);
6369 }
6370 
6371 /*
6372  * aac_fm_init - initialize fma capabilities and register with IO
6373  *               fault services.
6374  */
6375 static void
6376 aac_fm_init(struct aac_softstate *softs)
6377 {
6378 	/*
6379 	 * Need to change iblock to priority for new MSI intr
6380 	 */
6381 	ddi_iblock_cookie_t fm_ibc;
6382 
6383 	softs->fm_capabilities = ddi_getprop(DDI_DEV_T_ANY, softs->devinfo_p,
6384 	    DDI_PROP_CANSLEEP | DDI_PROP_DONTPASS, "fm-capable",
6385 	    DDI_FM_EREPORT_CAPABLE | DDI_FM_ACCCHK_CAPABLE |
6386 	    DDI_FM_DMACHK_CAPABLE | DDI_FM_ERRCB_CAPABLE);
6387 
6388 	/* Only register with IO Fault Services if we have some capability */
6389 	if (softs->fm_capabilities) {
6390 		/* Adjust access and dma attributes for FMA */
6391 		softs->acc_attr.devacc_attr_access |= DDI_FLAGERR_ACC;
6392 		softs->addr_dma_attr.dma_attr_flags |= DDI_DMA_FLAGERR;
6393 		softs->buf_dma_attr.dma_attr_flags |= DDI_DMA_FLAGERR;
6394 
6395 		/*
6396 		 * Register capabilities with IO Fault Services.
6397 		 * fm_capabilities will be updated to indicate
6398 		 * capabilities actually supported (not requested.)
6399 		 */
6400 		ddi_fm_init(softs->devinfo_p, &softs->fm_capabilities, &fm_ibc);
6401 
6402 		/*
6403 		 * Initialize pci ereport capabilities if ereport
6404 		 * capable (should always be.)
6405 		 */
6406 		if (DDI_FM_EREPORT_CAP(softs->fm_capabilities) ||
6407 		    DDI_FM_ERRCB_CAP(softs->fm_capabilities)) {
6408 			pci_ereport_setup(softs->devinfo_p);
6409 		}
6410 
6411 		/*
6412 		 * Register error callback if error callback capable.
6413 		 */
6414 		if (DDI_FM_ERRCB_CAP(softs->fm_capabilities)) {
6415 			ddi_fm_handler_register(softs->devinfo_p,
6416 			    aac_fm_error_cb, (void *) softs);
6417 		}
6418 	}
6419 }
6420 
6421 /*
6422  * aac_fm_fini - Releases fma capabilities and un-registers with IO
6423  *               fault services.
6424  */
6425 static void
6426 aac_fm_fini(struct aac_softstate *softs)
6427 {
6428 	/* Only unregister FMA capabilities if registered */
6429 	if (softs->fm_capabilities) {
6430 		/*
6431 		 * Un-register error callback if error callback capable.
6432 		 */
6433 		if (DDI_FM_ERRCB_CAP(softs->fm_capabilities)) {
6434 			ddi_fm_handler_unregister(softs->devinfo_p);
6435 		}
6436 
6437 		/*
6438 		 * Release any resources allocated by pci_ereport_setup()
6439 		 */
6440 		if (DDI_FM_EREPORT_CAP(softs->fm_capabilities) ||
6441 		    DDI_FM_ERRCB_CAP(softs->fm_capabilities)) {
6442 			pci_ereport_teardown(softs->devinfo_p);
6443 		}
6444 
6445 		/* Unregister from IO Fault Services */
6446 		ddi_fm_fini(softs->devinfo_p);
6447 
6448 		/* Adjust access and dma attributes for FMA */
6449 		softs->acc_attr.devacc_attr_access &= ~DDI_FLAGERR_ACC;
6450 		softs->addr_dma_attr.dma_attr_flags &= ~DDI_DMA_FLAGERR;
6451 		softs->buf_dma_attr.dma_attr_flags &= ~DDI_DMA_FLAGERR;
6452 	}
6453 }
6454 
6455 int
6456 aac_check_acc_handle(ddi_acc_handle_t handle)
6457 {
6458 	ddi_fm_error_t de;
6459 
6460 	ddi_fm_acc_err_get(handle, &de, DDI_FME_VERSION);
6461 	return (de.fme_status);
6462 }
6463 
6464 int
6465 aac_check_dma_handle(ddi_dma_handle_t handle)
6466 {
6467 	ddi_fm_error_t de;
6468 
6469 	ddi_fm_dma_err_get(handle, &de, DDI_FME_VERSION);
6470 	return (de.fme_status);
6471 }
6472 
6473 void
6474 aac_fm_ereport(struct aac_softstate *softs, char *detail)
6475 {
6476 	uint64_t ena;
6477 	char buf[FM_MAX_CLASS];
6478 
6479 	(void) snprintf(buf, FM_MAX_CLASS, "%s.%s", DDI_FM_DEVICE, detail);
6480 	ena = fm_ena_generate(0, FM_ENA_FMT1);
6481 	if (DDI_FM_EREPORT_CAP(softs->fm_capabilities)) {
6482 		ddi_fm_ereport_post(softs->devinfo_p, buf, ena, DDI_NOSLEEP,
6483 		    FM_VERSION, DATA_TYPE_UINT8, FM_EREPORT_VERSION, NULL);
6484 	}
6485 }
6486 
6487 /*
6488  * Autoconfiguration support
6489  */
6490 static int
6491 aac_parse_devname(char *devnm, int *tgt, int *lun)
6492 {
6493 	char devbuf[SCSI_MAXNAMELEN];
6494 	char *addr;
6495 	char *p,  *tp, *lp;
6496 	long num;
6497 
6498 	/* Parse dev name and address */
6499 	(void) strcpy(devbuf, devnm);
6500 	addr = "";
6501 	for (p = devbuf; *p != '\0'; p++) {
6502 		if (*p == '@') {
6503 			addr = p + 1;
6504 			*p = '\0';
6505 		} else if (*p == ':') {
6506 			*p = '\0';
6507 			break;
6508 		}
6509 	}
6510 
6511 	/* Parse taget and lun */
6512 	for (p = tp = addr, lp = NULL; *p != '\0'; p++) {
6513 		if (*p == ',') {
6514 			lp = p + 1;
6515 			*p = '\0';
6516 			break;
6517 		}
6518 	}
6519 	if (tgt && tp) {
6520 		if (ddi_strtol(tp, NULL, 0x10, &num))
6521 			return (AACERR);
6522 		*tgt = (int)num;
6523 	}
6524 	if (lun && lp) {
6525 		if (ddi_strtol(lp, NULL, 0x10, &num))
6526 			return (AACERR);
6527 		*lun = (int)num;
6528 	}
6529 	return (AACOK);
6530 }
6531 
6532 static dev_info_t *
6533 aac_find_child(struct aac_softstate *softs, uint16_t tgt, uint8_t lun)
6534 {
6535 	dev_info_t *child = NULL;
6536 	char addr[SCSI_MAXNAMELEN];
6537 	char tmp[MAXNAMELEN];
6538 
6539 	if (tgt < AAC_MAX_LD) {
6540 		if (lun == 0) {
6541 			struct aac_device *dvp = &softs->containers[tgt].dev;
6542 
6543 			child = dvp->dip;
6544 		}
6545 	} else {
6546 		(void) sprintf(addr, "%x,%x", tgt, lun);
6547 		for (child = ddi_get_child(softs->devinfo_p);
6548 		    child; child = ddi_get_next_sibling(child)) {
6549 			/* We don't care about non-persistent node */
6550 			if (ndi_dev_is_persistent_node(child) == 0)
6551 				continue;
6552 
6553 			if (aac_name_node(child, tmp, MAXNAMELEN) !=
6554 			    DDI_SUCCESS)
6555 				continue;
6556 			if (strcmp(addr, tmp) == 0)
6557 				break;
6558 		}
6559 	}
6560 	return (child);
6561 }
6562 
6563 static int
6564 aac_config_child(struct aac_softstate *softs, struct scsi_device *sd,
6565     dev_info_t **dipp)
6566 {
6567 	char *nodename = NULL;
6568 	char **compatible = NULL;
6569 	int ncompatible = 0;
6570 	char *childname;
6571 	dev_info_t *ldip = NULL;
6572 	int tgt = sd->sd_address.a_target;
6573 	int lun = sd->sd_address.a_lun;
6574 	int dtype = sd->sd_inq->inq_dtype & DTYPE_MASK;
6575 	int rval;
6576 
6577 	DBCALLED(softs, 2);
6578 
6579 	scsi_hba_nodename_compatible_get(sd->sd_inq, NULL, dtype,
6580 	    NULL, &nodename, &compatible, &ncompatible);
6581 	if (nodename == NULL) {
6582 		AACDB_PRINT(softs, CE_WARN,
6583 		    "found no comptible driver for t%dL%d", tgt, lun);
6584 		rval = NDI_FAILURE;
6585 		goto finish;
6586 	}
6587 	childname = (softs->legacy && dtype == DTYPE_DIRECT) ? "sd" : nodename;
6588 
6589 	/* Create dev node */
6590 	rval = ndi_devi_alloc(softs->devinfo_p, childname, DEVI_SID_NODEID,
6591 	    &ldip);
6592 	if (rval == NDI_SUCCESS) {
6593 		if (ndi_prop_update_int(DDI_DEV_T_NONE, ldip, "target", tgt)
6594 		    != DDI_PROP_SUCCESS) {
6595 			AACDB_PRINT(softs, CE_WARN, "unable to create "
6596 			    "property for t%dL%d (target)", tgt, lun);
6597 			rval = NDI_FAILURE;
6598 			goto finish;
6599 		}
6600 		if (ndi_prop_update_int(DDI_DEV_T_NONE, ldip, "lun", lun)
6601 		    != DDI_PROP_SUCCESS) {
6602 			AACDB_PRINT(softs, CE_WARN, "unable to create "
6603 			    "property for t%dL%d (lun)", tgt, lun);
6604 			rval = NDI_FAILURE;
6605 			goto finish;
6606 		}
6607 		if (ndi_prop_update_string_array(DDI_DEV_T_NONE, ldip,
6608 		    "compatible", compatible, ncompatible)
6609 		    != DDI_PROP_SUCCESS) {
6610 			AACDB_PRINT(softs, CE_WARN, "unable to create "
6611 			    "property for t%dL%d (compatible)", tgt, lun);
6612 			rval = NDI_FAILURE;
6613 			goto finish;
6614 		}
6615 
6616 		rval = ndi_devi_online(ldip, NDI_ONLINE_ATTACH);
6617 		if (rval != NDI_SUCCESS) {
6618 			AACDB_PRINT(softs, CE_WARN, "unable to online t%dL%d",
6619 			    tgt, lun);
6620 			ndi_prop_remove_all(ldip);
6621 			(void) ndi_devi_free(ldip);
6622 		}
6623 	}
6624 finish:
6625 	if (dipp)
6626 		*dipp = ldip;
6627 
6628 	scsi_hba_nodename_compatible_free(nodename, compatible);
6629 	return (rval);
6630 }
6631 
6632 /*ARGSUSED*/
6633 static int
6634 aac_probe_lun(struct aac_softstate *softs, struct scsi_device *sd)
6635 {
6636 	int tgt = sd->sd_address.a_target;
6637 	int lun = sd->sd_address.a_lun;
6638 
6639 	DBCALLED(softs, 2);
6640 
6641 	if (tgt < AAC_MAX_LD) {
6642 		int rval;
6643 
6644 		if (lun == 0) {
6645 			mutex_enter(&softs->io_lock);
6646 			rval = aac_probe_container(softs, tgt);
6647 			mutex_exit(&softs->io_lock);
6648 			if (rval == AACOK) {
6649 				if (scsi_hba_probe(sd, NULL) ==
6650 				    SCSIPROBE_EXISTS)
6651 					return (NDI_SUCCESS);
6652 			}
6653 		}
6654 		return (NDI_FAILURE);
6655 	} else {
6656 		int dtype;
6657 
6658 		if (scsi_hba_probe(sd, NULL) != SCSIPROBE_EXISTS)
6659 			return (NDI_FAILURE);
6660 
6661 		dtype = sd->sd_inq->inq_dtype & DTYPE_MASK;
6662 
6663 		AACDB_PRINT(softs, CE_NOTE,
6664 		    "Phys. device found: tgt %d dtype %d: %s",
6665 		    tgt, dtype, sd->sd_inq->inq_vid);
6666 
6667 		/* Only non-DASD exposed */
6668 		if (dtype != DTYPE_RODIRECT /* CDROM */ &&
6669 		    dtype != DTYPE_SEQUENTIAL /* TAPE */ &&
6670 		    dtype != DTYPE_ESI /* SES */)
6671 			return (NDI_FAILURE);
6672 
6673 		AACDB_PRINT(softs, CE_NOTE, "non-DASD %d found", tgt);
6674 		mutex_enter(&softs->io_lock);
6675 		softs->nondasds[AAC_PD(tgt)].dev.flags |= AAC_DFLAG_VALID;
6676 		mutex_exit(&softs->io_lock);
6677 		return (NDI_SUCCESS);
6678 	}
6679 }
6680 
6681 static int
6682 aac_config_lun(struct aac_softstate *softs, uint16_t tgt, uint8_t lun,
6683     dev_info_t **ldip)
6684 {
6685 	struct scsi_device sd;
6686 	dev_info_t *child;
6687 	int rval;
6688 
6689 	DBCALLED(softs, 2);
6690 
6691 	if ((child = aac_find_child(softs, tgt, lun)) != NULL) {
6692 		if (ldip)
6693 			*ldip = child;
6694 		return (NDI_SUCCESS);
6695 	}
6696 
6697 	bzero(&sd, sizeof (struct scsi_device));
6698 	sd.sd_address.a_hba_tran = softs->hba_tran;
6699 	sd.sd_address.a_target = (uint16_t)tgt;
6700 	sd.sd_address.a_lun = (uint8_t)lun;
6701 	if ((rval = aac_probe_lun(softs, &sd)) == NDI_SUCCESS)
6702 		rval = aac_config_child(softs, &sd, ldip);
6703 	scsi_unprobe(&sd);
6704 	return (rval);
6705 }
6706 
6707 static int
6708 aac_config_tgt(struct aac_softstate *softs, int tgt)
6709 {
6710 	struct scsi_address ap;
6711 	struct buf *bp = NULL;
6712 	int buf_len = AAC_SCSI_RPTLUNS_HEAD_SIZE + AAC_SCSI_RPTLUNS_ADDR_SIZE;
6713 	int list_len = 0;
6714 	int lun_total = 0;
6715 	dev_info_t *ldip;
6716 	int i;
6717 
6718 	ap.a_hba_tran = softs->hba_tran;
6719 	ap.a_target = (uint16_t)tgt;
6720 	ap.a_lun = 0;
6721 
6722 	for (i = 0; i < 2; i++) {
6723 		struct scsi_pkt *pkt;
6724 		uchar_t *cdb;
6725 		uchar_t *p;
6726 		uint32_t data;
6727 
6728 		if (bp == NULL) {
6729 			if ((bp = scsi_alloc_consistent_buf(&ap, NULL,
6730 			    buf_len, B_READ, NULL_FUNC, NULL)) == NULL)
6731 			return (AACERR);
6732 		}
6733 		if ((pkt = scsi_init_pkt(&ap, NULL, bp, CDB_GROUP5,
6734 		    sizeof (struct scsi_arq_status), 0, PKT_CONSISTENT,
6735 		    NULL, NULL)) == NULL) {
6736 			scsi_free_consistent_buf(bp);
6737 			return (AACERR);
6738 		}
6739 		cdb = pkt->pkt_cdbp;
6740 		bzero(cdb, CDB_GROUP5);
6741 		cdb[0] = SCMD_REPORT_LUNS;
6742 
6743 		/* Convert buffer len from local to LE_32 */
6744 		data = buf_len;
6745 		for (p = &cdb[9]; p > &cdb[5]; p--) {
6746 			*p = data & 0xff;
6747 			data >>= 8;
6748 		}
6749 
6750 		if (scsi_poll(pkt) < 0 ||
6751 		    ((struct scsi_status *)pkt->pkt_scbp)->sts_chk) {
6752 			scsi_destroy_pkt(pkt);
6753 			break;
6754 		}
6755 
6756 		/* Convert list_len from LE_32 to local */
6757 		for (p = (uchar_t *)bp->b_un.b_addr;
6758 		    p < (uchar_t *)bp->b_un.b_addr + 4; p++) {
6759 			data <<= 8;
6760 			data |= *p;
6761 		}
6762 		list_len = data;
6763 		if (buf_len < list_len + AAC_SCSI_RPTLUNS_HEAD_SIZE) {
6764 			scsi_free_consistent_buf(bp);
6765 			bp = NULL;
6766 			buf_len = list_len + AAC_SCSI_RPTLUNS_HEAD_SIZE;
6767 		}
6768 		scsi_destroy_pkt(pkt);
6769 	}
6770 	if (i >= 2) {
6771 		uint8_t *buf = (uint8_t *)(bp->b_un.b_addr +
6772 		    AAC_SCSI_RPTLUNS_HEAD_SIZE);
6773 
6774 		for (i = 0; i < (list_len / AAC_SCSI_RPTLUNS_ADDR_SIZE); i++) {
6775 			uint16_t lun;
6776 
6777 			/* Determine report luns addressing type */
6778 			switch (buf[0] & AAC_SCSI_RPTLUNS_ADDR_MASK) {
6779 			/*
6780 			 * Vendors in the field have been found to be
6781 			 * concatenating bus/target/lun to equal the
6782 			 * complete lun value instead of switching to
6783 			 * flat space addressing
6784 			 */
6785 			case AAC_SCSI_RPTLUNS_ADDR_PERIPHERAL:
6786 			case AAC_SCSI_RPTLUNS_ADDR_LOGICAL_UNIT:
6787 			case AAC_SCSI_RPTLUNS_ADDR_FLAT_SPACE:
6788 				lun = ((buf[0] & 0x3f) << 8) | buf[1];
6789 				if (lun > UINT8_MAX) {
6790 					AACDB_PRINT(softs, CE_WARN,
6791 					    "abnormal lun number: %d", lun);
6792 					break;
6793 				}
6794 				if (aac_config_lun(softs, tgt, lun, &ldip) ==
6795 				    NDI_SUCCESS)
6796 					lun_total++;
6797 				break;
6798 			}
6799 
6800 			buf += AAC_SCSI_RPTLUNS_ADDR_SIZE;
6801 		}
6802 	} else {
6803 		/* The target may do not support SCMD_REPORT_LUNS. */
6804 		if (aac_config_lun(softs, tgt, 0, &ldip) == NDI_SUCCESS)
6805 			lun_total++;
6806 	}
6807 	scsi_free_consistent_buf(bp);
6808 	return (lun_total);
6809 }
6810 
6811 static void
6812 aac_devcfg(struct aac_softstate *softs, int tgt, int en)
6813 {
6814 	struct aac_device *dvp;
6815 
6816 	mutex_enter(&softs->io_lock);
6817 	dvp = AAC_DEV(softs, tgt);
6818 	if (en)
6819 		dvp->flags |= AAC_DFLAG_CONFIGURING;
6820 	else
6821 		dvp->flags &= ~AAC_DFLAG_CONFIGURING;
6822 	mutex_exit(&softs->io_lock);
6823 }
6824 
6825 static int
6826 aac_tran_bus_config(dev_info_t *parent, uint_t flags, ddi_bus_config_op_t op,
6827     void *arg, dev_info_t **childp)
6828 {
6829 	struct aac_softstate *softs;
6830 	int circ = 0;
6831 	int rval;
6832 
6833 	if ((softs = ddi_get_soft_state(aac_softstatep,
6834 	    ddi_get_instance(parent))) == NULL)
6835 		return (NDI_FAILURE);
6836 
6837 	/* Commands for bus config should be blocked as the bus is quiesced */
6838 	mutex_enter(&softs->io_lock);
6839 	if (softs->state & AAC_STATE_QUIESCED) {
6840 		AACDB_PRINT(softs, CE_NOTE,
6841 		    "bus_config abroted because bus is quiesced");
6842 		mutex_exit(&softs->io_lock);
6843 		return (NDI_FAILURE);
6844 	}
6845 	mutex_exit(&softs->io_lock);
6846 
6847 	DBCALLED(softs, 1);
6848 
6849 	/* Hold the nexus across the bus_config */
6850 	ndi_devi_enter(parent, &circ);
6851 	switch (op) {
6852 	case BUS_CONFIG_ONE: {
6853 		int tgt, lun;
6854 
6855 		if (aac_parse_devname(arg, &tgt, &lun) != AACOK) {
6856 			rval = NDI_FAILURE;
6857 			break;
6858 		}
6859 
6860 		AAC_DEVCFG_BEGIN(softs, tgt);
6861 		rval = aac_config_lun(softs, tgt, lun, childp);
6862 		AAC_DEVCFG_END(softs, tgt);
6863 		break;
6864 	}
6865 
6866 	case BUS_CONFIG_DRIVER:
6867 	case BUS_CONFIG_ALL: {
6868 		uint32_t bus, tgt;
6869 		int index, total;
6870 
6871 		for (tgt = 0; tgt < AAC_MAX_LD; tgt++) {
6872 			AAC_DEVCFG_BEGIN(softs, tgt);
6873 			(void) aac_config_lun(softs, tgt, 0, NULL);
6874 			AAC_DEVCFG_END(softs, tgt);
6875 		}
6876 
6877 		/* Config the non-DASD devices connected to the card */
6878 		total = 0;
6879 		index = AAC_MAX_LD;
6880 		for (bus = 0; bus < softs->bus_max; bus++) {
6881 			AACDB_PRINT(softs, CE_NOTE, "bus %d:", bus);
6882 			for (tgt = 0; tgt < softs->tgt_max; tgt++, index++) {
6883 				AAC_DEVCFG_BEGIN(softs, index);
6884 				if (aac_config_tgt(softs, index))
6885 					total++;
6886 				AAC_DEVCFG_END(softs, index);
6887 			}
6888 		}
6889 		AACDB_PRINT(softs, CE_CONT,
6890 		    "?Total %d phys. device(s) found", total);
6891 		rval = NDI_SUCCESS;
6892 		break;
6893 	}
6894 	}
6895 
6896 	if (rval == NDI_SUCCESS)
6897 		rval = ndi_busop_bus_config(parent, flags, op, arg, childp, 0);
6898 	ndi_devi_exit(parent, circ);
6899 	return (rval);
6900 }
6901 
6902 static void
6903 aac_handle_dr(struct aac_drinfo *drp)
6904 {
6905 	struct aac_softstate *softs = drp->softs;
6906 	struct aac_device *dvp;
6907 	dev_info_t *dip;
6908 	int valid;
6909 	int circ1 = 0;
6910 
6911 	DBCALLED(softs, 1);
6912 
6913 	/* Hold the nexus across the bus_config */
6914 	mutex_enter(&softs->io_lock);
6915 	dvp = AAC_DEV(softs, drp->tgt);
6916 	valid = AAC_DEV_IS_VALID(dvp);
6917 	dip = dvp->dip;
6918 	mutex_exit(&softs->io_lock);
6919 
6920 	switch (drp->event) {
6921 	case AAC_EVT_ONLINE:
6922 	case AAC_EVT_OFFLINE:
6923 		/* Device onlined */
6924 		if (dip == NULL && valid) {
6925 			ndi_devi_enter(softs->devinfo_p, &circ1);
6926 			(void) aac_config_lun(softs, drp->tgt, 0, NULL);
6927 			AACDB_PRINT(softs, CE_NOTE, "c%dt%dL%d onlined",
6928 			    softs->instance, drp->tgt, drp->lun);
6929 			ndi_devi_exit(softs->devinfo_p, circ1);
6930 		}
6931 		/* Device offlined */
6932 		if (dip && !valid) {
6933 			mutex_enter(&softs->io_lock);
6934 			(void) aac_do_reset(softs);
6935 			mutex_exit(&softs->io_lock);
6936 
6937 			(void) ndi_devi_offline(dip, NDI_DEVI_REMOVE);
6938 			AACDB_PRINT(softs, CE_NOTE, "c%dt%dL%d offlined",
6939 			    softs->instance, drp->tgt, drp->lun);
6940 		}
6941 		break;
6942 	}
6943 	kmem_free(drp, sizeof (struct aac_drinfo));
6944 }
6945 
6946 static int
6947 aac_dr_event(struct aac_softstate *softs, int tgt, int lun, int event)
6948 {
6949 	struct aac_drinfo *drp;
6950 
6951 	DBCALLED(softs, 1);
6952 
6953 	if (softs->taskq == NULL ||
6954 	    (drp = kmem_zalloc(sizeof (struct aac_drinfo), KM_NOSLEEP)) == NULL)
6955 		return (AACERR);
6956 
6957 	drp->softs = softs;
6958 	drp->tgt = tgt;
6959 	drp->lun = lun;
6960 	drp->event = event;
6961 	if ((ddi_taskq_dispatch(softs->taskq, (void (*)(void *))aac_handle_dr,
6962 	    drp, DDI_NOSLEEP)) != DDI_SUCCESS) {
6963 		AACDB_PRINT(softs, CE_WARN, "DR task start failed");
6964 		kmem_free(drp, sizeof (struct aac_drinfo));
6965 		return (AACERR);
6966 	}
6967 	return (AACOK);
6968 }
6969 
6970 #ifdef DEBUG
6971 
6972 /* -------------------------debug aid functions-------------------------- */
6973 
6974 #define	AAC_FIB_CMD_KEY_STRINGS \
6975 	TestCommandResponse, "TestCommandResponse", \
6976 	TestAdapterCommand, "TestAdapterCommand", \
6977 	LastTestCommand, "LastTestCommand", \
6978 	ReinitHostNormCommandQueue, "ReinitHostNormCommandQueue", \
6979 	ReinitHostHighCommandQueue, "ReinitHostHighCommandQueue", \
6980 	ReinitHostHighRespQueue, "ReinitHostHighRespQueue", \
6981 	ReinitHostNormRespQueue, "ReinitHostNormRespQueue", \
6982 	ReinitAdapNormCommandQueue, "ReinitAdapNormCommandQueue", \
6983 	ReinitAdapHighCommandQueue, "ReinitAdapHighCommandQueue", \
6984 	ReinitAdapHighRespQueue, "ReinitAdapHighRespQueue", \
6985 	ReinitAdapNormRespQueue, "ReinitAdapNormRespQueue", \
6986 	InterfaceShutdown, "InterfaceShutdown", \
6987 	DmaCommandFib, "DmaCommandFib", \
6988 	StartProfile, "StartProfile", \
6989 	TermProfile, "TermProfile", \
6990 	SpeedTest, "SpeedTest", \
6991 	TakeABreakPt, "TakeABreakPt", \
6992 	RequestPerfData, "RequestPerfData", \
6993 	SetInterruptDefTimer, "SetInterruptDefTimer", \
6994 	SetInterruptDefCount, "SetInterruptDefCount", \
6995 	GetInterruptDefStatus, "GetInterruptDefStatus", \
6996 	LastCommCommand, "LastCommCommand", \
6997 	NuFileSystem, "NuFileSystem", \
6998 	UFS, "UFS", \
6999 	HostFileSystem, "HostFileSystem", \
7000 	LastFileSystemCommand, "LastFileSystemCommand", \
7001 	ContainerCommand, "ContainerCommand", \
7002 	ContainerCommand64, "ContainerCommand64", \
7003 	ClusterCommand, "ClusterCommand", \
7004 	ScsiPortCommand, "ScsiPortCommand", \
7005 	ScsiPortCommandU64, "ScsiPortCommandU64", \
7006 	AifRequest, "AifRequest", \
7007 	CheckRevision, "CheckRevision", \
7008 	FsaHostShutdown, "FsaHostShutdown", \
7009 	RequestAdapterInfo, "RequestAdapterInfo", \
7010 	IsAdapterPaused, "IsAdapterPaused", \
7011 	SendHostTime, "SendHostTime", \
7012 	LastMiscCommand, "LastMiscCommand"
7013 
7014 #define	AAC_CTVM_SUBCMD_KEY_STRINGS \
7015 	VM_Null, "VM_Null", \
7016 	VM_NameServe, "VM_NameServe", \
7017 	VM_ContainerConfig, "VM_ContainerConfig", \
7018 	VM_Ioctl, "VM_Ioctl", \
7019 	VM_FilesystemIoctl, "VM_FilesystemIoctl", \
7020 	VM_CloseAll, "VM_CloseAll", \
7021 	VM_CtBlockRead, "VM_CtBlockRead", \
7022 	VM_CtBlockWrite, "VM_CtBlockWrite", \
7023 	VM_SliceBlockRead, "VM_SliceBlockRead", \
7024 	VM_SliceBlockWrite, "VM_SliceBlockWrite", \
7025 	VM_DriveBlockRead, "VM_DriveBlockRead", \
7026 	VM_DriveBlockWrite, "VM_DriveBlockWrite", \
7027 	VM_EnclosureMgt, "VM_EnclosureMgt", \
7028 	VM_Unused, "VM_Unused", \
7029 	VM_CtBlockVerify, "VM_CtBlockVerify", \
7030 	VM_CtPerf, "VM_CtPerf", \
7031 	VM_CtBlockRead64, "VM_CtBlockRead64", \
7032 	VM_CtBlockWrite64, "VM_CtBlockWrite64", \
7033 	VM_CtBlockVerify64, "VM_CtBlockVerify64", \
7034 	VM_CtHostRead64, "VM_CtHostRead64", \
7035 	VM_CtHostWrite64, "VM_CtHostWrite64", \
7036 	VM_NameServe64, "VM_NameServe64"
7037 
7038 #define	AAC_CT_SUBCMD_KEY_STRINGS \
7039 	CT_Null, "CT_Null", \
7040 	CT_GET_SLICE_COUNT, "CT_GET_SLICE_COUNT", \
7041 	CT_GET_PARTITION_COUNT, "CT_GET_PARTITION_COUNT", \
7042 	CT_GET_PARTITION_INFO, "CT_GET_PARTITION_INFO", \
7043 	CT_GET_CONTAINER_COUNT, "CT_GET_CONTAINER_COUNT", \
7044 	CT_GET_CONTAINER_INFO_OLD, "CT_GET_CONTAINER_INFO_OLD", \
7045 	CT_WRITE_MBR, "CT_WRITE_MBR", \
7046 	CT_WRITE_PARTITION, "CT_WRITE_PARTITION", \
7047 	CT_UPDATE_PARTITION, "CT_UPDATE_PARTITION", \
7048 	CT_UNLOAD_CONTAINER, "CT_UNLOAD_CONTAINER", \
7049 	CT_CONFIG_SINGLE_PRIMARY, "CT_CONFIG_SINGLE_PRIMARY", \
7050 	CT_READ_CONFIG_AGE, "CT_READ_CONFIG_AGE", \
7051 	CT_WRITE_CONFIG_AGE, "CT_WRITE_CONFIG_AGE", \
7052 	CT_READ_SERIAL_NUMBER, "CT_READ_SERIAL_NUMBER", \
7053 	CT_ZERO_PAR_ENTRY, "CT_ZERO_PAR_ENTRY", \
7054 	CT_READ_MBR, "CT_READ_MBR", \
7055 	CT_READ_PARTITION, "CT_READ_PARTITION", \
7056 	CT_DESTROY_CONTAINER, "CT_DESTROY_CONTAINER", \
7057 	CT_DESTROY2_CONTAINER, "CT_DESTROY2_CONTAINER", \
7058 	CT_SLICE_SIZE, "CT_SLICE_SIZE", \
7059 	CT_CHECK_CONFLICTS, "CT_CHECK_CONFLICTS", \
7060 	CT_MOVE_CONTAINER, "CT_MOVE_CONTAINER", \
7061 	CT_READ_LAST_DRIVE, "CT_READ_LAST_DRIVE", \
7062 	CT_WRITE_LAST_DRIVE, "CT_WRITE_LAST_DRIVE", \
7063 	CT_UNMIRROR, "CT_UNMIRROR", \
7064 	CT_MIRROR_DELAY, "CT_MIRROR_DELAY", \
7065 	CT_GEN_MIRROR, "CT_GEN_MIRROR", \
7066 	CT_GEN_MIRROR2, "CT_GEN_MIRROR2", \
7067 	CT_TEST_CONTAINER, "CT_TEST_CONTAINER", \
7068 	CT_MOVE2, "CT_MOVE2", \
7069 	CT_SPLIT, "CT_SPLIT", \
7070 	CT_SPLIT2, "CT_SPLIT2", \
7071 	CT_SPLIT_BROKEN, "CT_SPLIT_BROKEN", \
7072 	CT_SPLIT_BROKEN2, "CT_SPLIT_BROKEN2", \
7073 	CT_RECONFIG, "CT_RECONFIG", \
7074 	CT_BREAK2, "CT_BREAK2", \
7075 	CT_BREAK, "CT_BREAK", \
7076 	CT_MERGE2, "CT_MERGE2", \
7077 	CT_MERGE, "CT_MERGE", \
7078 	CT_FORCE_ERROR, "CT_FORCE_ERROR", \
7079 	CT_CLEAR_ERROR, "CT_CLEAR_ERROR", \
7080 	CT_ASSIGN_FAILOVER, "CT_ASSIGN_FAILOVER", \
7081 	CT_CLEAR_FAILOVER, "CT_CLEAR_FAILOVER", \
7082 	CT_GET_FAILOVER_DATA, "CT_GET_FAILOVER_DATA", \
7083 	CT_VOLUME_ADD, "CT_VOLUME_ADD", \
7084 	CT_VOLUME_ADD2, "CT_VOLUME_ADD2", \
7085 	CT_MIRROR_STATUS, "CT_MIRROR_STATUS", \
7086 	CT_COPY_STATUS, "CT_COPY_STATUS", \
7087 	CT_COPY, "CT_COPY", \
7088 	CT_UNLOCK_CONTAINER, "CT_UNLOCK_CONTAINER", \
7089 	CT_LOCK_CONTAINER, "CT_LOCK_CONTAINER", \
7090 	CT_MAKE_READ_ONLY, "CT_MAKE_READ_ONLY", \
7091 	CT_MAKE_READ_WRITE, "CT_MAKE_READ_WRITE", \
7092 	CT_CLEAN_DEAD, "CT_CLEAN_DEAD", \
7093 	CT_ABORT_MIRROR_COMMAND, "CT_ABORT_MIRROR_COMMAND", \
7094 	CT_SET, "CT_SET", \
7095 	CT_GET, "CT_GET", \
7096 	CT_GET_NVLOG_ENTRY, "CT_GET_NVLOG_ENTRY", \
7097 	CT_GET_DELAY, "CT_GET_DELAY", \
7098 	CT_ZERO_CONTAINER_SPACE, "CT_ZERO_CONTAINER_SPACE", \
7099 	CT_GET_ZERO_STATUS, "CT_GET_ZERO_STATUS", \
7100 	CT_SCRUB, "CT_SCRUB", \
7101 	CT_GET_SCRUB_STATUS, "CT_GET_SCRUB_STATUS", \
7102 	CT_GET_SLICE_INFO, "CT_GET_SLICE_INFO", \
7103 	CT_GET_SCSI_METHOD, "CT_GET_SCSI_METHOD", \
7104 	CT_PAUSE_IO, "CT_PAUSE_IO", \
7105 	CT_RELEASE_IO, "CT_RELEASE_IO", \
7106 	CT_SCRUB2, "CT_SCRUB2", \
7107 	CT_MCHECK, "CT_MCHECK", \
7108 	CT_CORRUPT, "CT_CORRUPT", \
7109 	CT_GET_TASK_COUNT, "CT_GET_TASK_COUNT", \
7110 	CT_PROMOTE, "CT_PROMOTE", \
7111 	CT_SET_DEAD, "CT_SET_DEAD", \
7112 	CT_CONTAINER_OPTIONS, "CT_CONTAINER_OPTIONS", \
7113 	CT_GET_NV_PARAM, "CT_GET_NV_PARAM", \
7114 	CT_GET_PARAM, "CT_GET_PARAM", \
7115 	CT_NV_PARAM_SIZE, "CT_NV_PARAM_SIZE", \
7116 	CT_COMMON_PARAM_SIZE, "CT_COMMON_PARAM_SIZE", \
7117 	CT_PLATFORM_PARAM_SIZE, "CT_PLATFORM_PARAM_SIZE", \
7118 	CT_SET_NV_PARAM, "CT_SET_NV_PARAM", \
7119 	CT_ABORT_SCRUB, "CT_ABORT_SCRUB", \
7120 	CT_GET_SCRUB_ERROR, "CT_GET_SCRUB_ERROR", \
7121 	CT_LABEL_CONTAINER, "CT_LABEL_CONTAINER", \
7122 	CT_CONTINUE_DATA, "CT_CONTINUE_DATA", \
7123 	CT_STOP_DATA, "CT_STOP_DATA", \
7124 	CT_GET_PARTITION_TABLE, "CT_GET_PARTITION_TABLE", \
7125 	CT_GET_DISK_PARTITIONS, "CT_GET_DISK_PARTITIONS", \
7126 	CT_GET_MISC_STATUS, "CT_GET_MISC_STATUS", \
7127 	CT_GET_CONTAINER_PERF_INFO, "CT_GET_CONTAINER_PERF_INFO", \
7128 	CT_GET_TIME, "CT_GET_TIME", \
7129 	CT_READ_DATA, "CT_READ_DATA", \
7130 	CT_CTR, "CT_CTR", \
7131 	CT_CTL, "CT_CTL", \
7132 	CT_DRAINIO, "CT_DRAINIO", \
7133 	CT_RELEASEIO, "CT_RELEASEIO", \
7134 	CT_GET_NVRAM, "CT_GET_NVRAM", \
7135 	CT_GET_MEMORY, "CT_GET_MEMORY", \
7136 	CT_PRINT_CT_LOG, "CT_PRINT_CT_LOG", \
7137 	CT_ADD_LEVEL, "CT_ADD_LEVEL", \
7138 	CT_NV_ZERO, "CT_NV_ZERO", \
7139 	CT_READ_SIGNATURE, "CT_READ_SIGNATURE", \
7140 	CT_THROTTLE_ON, "CT_THROTTLE_ON", \
7141 	CT_THROTTLE_OFF, "CT_THROTTLE_OFF", \
7142 	CT_GET_THROTTLE_STATS, "CT_GET_THROTTLE_STATS", \
7143 	CT_MAKE_SNAPSHOT, "CT_MAKE_SNAPSHOT", \
7144 	CT_REMOVE_SNAPSHOT, "CT_REMOVE_SNAPSHOT", \
7145 	CT_WRITE_USER_FLAGS, "CT_WRITE_USER_FLAGS", \
7146 	CT_READ_USER_FLAGS, "CT_READ_USER_FLAGS", \
7147 	CT_MONITOR, "CT_MONITOR", \
7148 	CT_GEN_MORPH, "CT_GEN_MORPH", \
7149 	CT_GET_SNAPSHOT_INFO, "CT_GET_SNAPSHOT_INFO", \
7150 	CT_CACHE_SET, "CT_CACHE_SET", \
7151 	CT_CACHE_STAT, "CT_CACHE_STAT", \
7152 	CT_TRACE_START, "CT_TRACE_START", \
7153 	CT_TRACE_STOP, "CT_TRACE_STOP", \
7154 	CT_TRACE_ENABLE, "CT_TRACE_ENABLE", \
7155 	CT_TRACE_DISABLE, "CT_TRACE_DISABLE", \
7156 	CT_FORCE_CORE_DUMP, "CT_FORCE_CORE_DUMP", \
7157 	CT_SET_SERIAL_NUMBER, "CT_SET_SERIAL_NUMBER", \
7158 	CT_RESET_SERIAL_NUMBER, "CT_RESET_SERIAL_NUMBER", \
7159 	CT_ENABLE_RAID5, "CT_ENABLE_RAID5", \
7160 	CT_CLEAR_VALID_DUMP_FLAG, "CT_CLEAR_VALID_DUMP_FLAG", \
7161 	CT_GET_MEM_STATS, "CT_GET_MEM_STATS", \
7162 	CT_GET_CORE_SIZE, "CT_GET_CORE_SIZE", \
7163 	CT_CREATE_CONTAINER_OLD, "CT_CREATE_CONTAINER_OLD", \
7164 	CT_STOP_DUMPS, "CT_STOP_DUMPS", \
7165 	CT_PANIC_ON_TAKE_A_BREAK, "CT_PANIC_ON_TAKE_A_BREAK", \
7166 	CT_GET_CACHE_STATS, "CT_GET_CACHE_STATS", \
7167 	CT_MOVE_PARTITION, "CT_MOVE_PARTITION", \
7168 	CT_FLUSH_CACHE, "CT_FLUSH_CACHE", \
7169 	CT_READ_NAME, "CT_READ_NAME", \
7170 	CT_WRITE_NAME, "CT_WRITE_NAME", \
7171 	CT_TOSS_CACHE, "CT_TOSS_CACHE", \
7172 	CT_LOCK_DRAINIO, "CT_LOCK_DRAINIO", \
7173 	CT_CONTAINER_OFFLINE, "CT_CONTAINER_OFFLINE", \
7174 	CT_SET_CACHE_SIZE, "CT_SET_CACHE_SIZE", \
7175 	CT_CLEAN_SHUTDOWN_STATUS, "CT_CLEAN_SHUTDOWN_STATUS", \
7176 	CT_CLEAR_DISKLOG_ON_DISK, "CT_CLEAR_DISKLOG_ON_DISK", \
7177 	CT_CLEAR_ALL_DISKLOG, "CT_CLEAR_ALL_DISKLOG", \
7178 	CT_CACHE_FAVOR, "CT_CACHE_FAVOR", \
7179 	CT_READ_PASSTHRU_MBR, "CT_READ_PASSTHRU_MBR", \
7180 	CT_SCRUB_NOFIX, "CT_SCRUB_NOFIX", \
7181 	CT_SCRUB2_NOFIX, "CT_SCRUB2_NOFIX", \
7182 	CT_FLUSH, "CT_FLUSH", \
7183 	CT_REBUILD, "CT_REBUILD", \
7184 	CT_FLUSH_CONTAINER, "CT_FLUSH_CONTAINER", \
7185 	CT_RESTART, "CT_RESTART", \
7186 	CT_GET_CONFIG_STATUS, "CT_GET_CONFIG_STATUS", \
7187 	CT_TRACE_FLAG, "CT_TRACE_FLAG", \
7188 	CT_RESTART_MORPH, "CT_RESTART_MORPH", \
7189 	CT_GET_TRACE_INFO, "CT_GET_TRACE_INFO", \
7190 	CT_GET_TRACE_ITEM, "CT_GET_TRACE_ITEM", \
7191 	CT_COMMIT_CONFIG, "CT_COMMIT_CONFIG", \
7192 	CT_CONTAINER_EXISTS, "CT_CONTAINER_EXISTS", \
7193 	CT_GET_SLICE_FROM_DEVT, "CT_GET_SLICE_FROM_DEVT", \
7194 	CT_OPEN_READ_WRITE, "CT_OPEN_READ_WRITE", \
7195 	CT_WRITE_MEMORY_BLOCK, "CT_WRITE_MEMORY_BLOCK", \
7196 	CT_GET_CACHE_PARAMS, "CT_GET_CACHE_PARAMS", \
7197 	CT_CRAZY_CACHE, "CT_CRAZY_CACHE", \
7198 	CT_GET_PROFILE_STRUCT, "CT_GET_PROFILE_STRUCT", \
7199 	CT_SET_IO_TRACE_FLAG, "CT_SET_IO_TRACE_FLAG", \
7200 	CT_GET_IO_TRACE_STRUCT, "CT_GET_IO_TRACE_STRUCT", \
7201 	CT_CID_TO_64BITS_UID, "CT_CID_TO_64BITS_UID", \
7202 	CT_64BITS_UID_TO_CID, "CT_64BITS_UID_TO_CID", \
7203 	CT_PAR_TO_64BITS_UID, "CT_PAR_TO_64BITS_UID", \
7204 	CT_CID_TO_32BITS_UID, "CT_CID_TO_32BITS_UID", \
7205 	CT_32BITS_UID_TO_CID, "CT_32BITS_UID_TO_CID", \
7206 	CT_PAR_TO_32BITS_UID, "CT_PAR_TO_32BITS_UID", \
7207 	CT_SET_FAILOVER_OPTION, "CT_SET_FAILOVER_OPTION", \
7208 	CT_GET_FAILOVER_OPTION, "CT_GET_FAILOVER_OPTION", \
7209 	CT_STRIPE_ADD2, "CT_STRIPE_ADD2", \
7210 	CT_CREATE_VOLUME_SET, "CT_CREATE_VOLUME_SET", \
7211 	CT_CREATE_STRIPE_SET, "CT_CREATE_STRIPE_SET", \
7212 	CT_VERIFY_CONTAINER, "CT_VERIFY_CONTAINER", \
7213 	CT_IS_CONTAINER_DEAD, "CT_IS_CONTAINER_DEAD", \
7214 	CT_GET_CONTAINER_OPTION, "CT_GET_CONTAINER_OPTION", \
7215 	CT_GET_SNAPSHOT_UNUSED_STRUCT, "CT_GET_SNAPSHOT_UNUSED_STRUCT", \
7216 	CT_CLEAR_SNAPSHOT_UNUSED_STRUCT, "CT_CLEAR_SNAPSHOT_UNUSED_STRUCT", \
7217 	CT_GET_CONTAINER_INFO, "CT_GET_CONTAINER_INFO", \
7218 	CT_CREATE_CONTAINER, "CT_CREATE_CONTAINER", \
7219 	CT_CHANGE_CREATIONINFO, "CT_CHANGE_CREATIONINFO", \
7220 	CT_CHECK_CONFLICT_UID, "CT_CHECK_CONFLICT_UID", \
7221 	CT_CONTAINER_UID_CHECK, "CT_CONTAINER_UID_CHECK", \
7222 	CT_IS_CONTAINER_MEATADATA_STANDARD, \
7223 	    "CT_IS_CONTAINER_MEATADATA_STANDARD", \
7224 	CT_IS_SLICE_METADATA_STANDARD, "CT_IS_SLICE_METADATA_STANDARD", \
7225 	CT_GET_IMPORT_COUNT, "CT_GET_IMPORT_COUNT", \
7226 	CT_CANCEL_ALL_IMPORTS, "CT_CANCEL_ALL_IMPORTS", \
7227 	CT_GET_IMPORT_INFO, "CT_GET_IMPORT_INFO", \
7228 	CT_IMPORT_ARRAY, "CT_IMPORT_ARRAY", \
7229 	CT_GET_LOG_SIZE, "CT_GET_LOG_SIZE", \
7230 	CT_ALARM_GET_STATE, "CT_ALARM_GET_STATE", \
7231 	CT_ALARM_SET_STATE, "CT_ALARM_SET_STATE", \
7232 	CT_ALARM_ON_OFF, "CT_ALARM_ON_OFF", \
7233 	CT_GET_EE_OEM_ID, "CT_GET_EE_OEM_ID", \
7234 	CT_GET_PPI_HEADERS, "CT_GET_PPI_HEADERS", \
7235 	CT_GET_PPI_DATA, "CT_GET_PPI_DATA", \
7236 	CT_GET_PPI_ENTRIES, "CT_GET_PPI_ENTRIES", \
7237 	CT_DELETE_PPI_BUNDLE, "CT_DELETE_PPI_BUNDLE", \
7238 	CT_GET_PARTITION_TABLE_2, "CT_GET_PARTITION_TABLE_2", \
7239 	CT_GET_PARTITION_INFO_2, "CT_GET_PARTITION_INFO_2", \
7240 	CT_GET_DISK_PARTITIONS_2, "CT_GET_DISK_PARTITIONS_2", \
7241 	CT_QUIESCE_ADAPTER, "CT_QUIESCE_ADAPTER", \
7242 	CT_CLEAR_PPI_TABLE, "CT_CLEAR_PPI_TABLE"
7243 
7244 #define	AAC_CL_SUBCMD_KEY_STRINGS \
7245 	CL_NULL, "CL_NULL", \
7246 	DS_INIT, "DS_INIT", \
7247 	DS_RESCAN, "DS_RESCAN", \
7248 	DS_CREATE, "DS_CREATE", \
7249 	DS_DELETE, "DS_DELETE", \
7250 	DS_ADD_DISK, "DS_ADD_DISK", \
7251 	DS_REMOVE_DISK, "DS_REMOVE_DISK", \
7252 	DS_MOVE_DISK, "DS_MOVE_DISK", \
7253 	DS_TAKE_OWNERSHIP, "DS_TAKE_OWNERSHIP", \
7254 	DS_RELEASE_OWNERSHIP, "DS_RELEASE_OWNERSHIP", \
7255 	DS_FORCE_OWNERSHIP, "DS_FORCE_OWNERSHIP", \
7256 	DS_GET_DISK_SET_PARAM, "DS_GET_DISK_SET_PARAM", \
7257 	DS_GET_DRIVE_PARAM, "DS_GET_DRIVE_PARAM", \
7258 	DS_GET_SLICE_PARAM, "DS_GET_SLICE_PARAM", \
7259 	DS_GET_DISK_SETS, "DS_GET_DISK_SETS", \
7260 	DS_GET_DRIVES, "DS_GET_DRIVES", \
7261 	DS_SET_DISK_SET_PARAM, "DS_SET_DISK_SET_PARAM", \
7262 	DS_ONLINE, "DS_ONLINE", \
7263 	DS_OFFLINE, "DS_OFFLINE", \
7264 	DS_ONLINE_CONTAINERS, "DS_ONLINE_CONTAINERS", \
7265 	DS_FSAPRINT, "DS_FSAPRINT", \
7266 	CL_CFG_SET_HOST_IDS, "CL_CFG_SET_HOST_IDS", \
7267 	CL_CFG_SET_PARTNER_HOST_IDS, "CL_CFG_SET_PARTNER_HOST_IDS", \
7268 	CL_CFG_GET_CLUSTER_CONFIG, "CL_CFG_GET_CLUSTER_CONFIG", \
7269 	CC_CLI_CLEAR_MESSAGE_BUFFER, "CC_CLI_CLEAR_MESSAGE_BUFFER", \
7270 	CC_SRV_CLEAR_MESSAGE_BUFFER, "CC_SRV_CLEAR_MESSAGE_BUFFER", \
7271 	CC_CLI_SHOW_MESSAGE_BUFFER, "CC_CLI_SHOW_MESSAGE_BUFFER", \
7272 	CC_SRV_SHOW_MESSAGE_BUFFER, "CC_SRV_SHOW_MESSAGE_BUFFER", \
7273 	CC_CLI_SEND_MESSAGE, "CC_CLI_SEND_MESSAGE", \
7274 	CC_SRV_SEND_MESSAGE, "CC_SRV_SEND_MESSAGE", \
7275 	CC_CLI_GET_MESSAGE, "CC_CLI_GET_MESSAGE", \
7276 	CC_SRV_GET_MESSAGE, "CC_SRV_GET_MESSAGE", \
7277 	CC_SEND_TEST_MESSAGE, "CC_SEND_TEST_MESSAGE", \
7278 	CC_GET_BUSINFO, "CC_GET_BUSINFO", \
7279 	CC_GET_PORTINFO, "CC_GET_PORTINFO", \
7280 	CC_GET_NAMEINFO, "CC_GET_NAMEINFO", \
7281 	CC_GET_CONFIGINFO, "CC_GET_CONFIGINFO", \
7282 	CQ_QUORUM_OP, "CQ_QUORUM_OP"
7283 
7284 #define	AAC_AIF_SUBCMD_KEY_STRINGS \
7285 	AifCmdEventNotify, "AifCmdEventNotify", \
7286 	AifCmdJobProgress, "AifCmdJobProgress", \
7287 	AifCmdAPIReport, "AifCmdAPIReport", \
7288 	AifCmdDriverNotify, "AifCmdDriverNotify", \
7289 	AifReqJobList, "AifReqJobList", \
7290 	AifReqJobsForCtr, "AifReqJobsForCtr", \
7291 	AifReqJobsForScsi, "AifReqJobsForScsi", \
7292 	AifReqJobReport, "AifReqJobReport", \
7293 	AifReqTerminateJob, "AifReqTerminateJob", \
7294 	AifReqSuspendJob, "AifReqSuspendJob", \
7295 	AifReqResumeJob, "AifReqResumeJob", \
7296 	AifReqSendAPIReport, "AifReqSendAPIReport", \
7297 	AifReqAPIJobStart, "AifReqAPIJobStart", \
7298 	AifReqAPIJobUpdate, "AifReqAPIJobUpdate", \
7299 	AifReqAPIJobFinish, "AifReqAPIJobFinish"
7300 
7301 #define	AAC_IOCTL_SUBCMD_KEY_STRINGS \
7302 	Reserved_IOCTL, "Reserved_IOCTL", \
7303 	GetDeviceHandle, "GetDeviceHandle", \
7304 	BusTargetLun_to_DeviceHandle, "BusTargetLun_to_DeviceHandle", \
7305 	DeviceHandle_to_BusTargetLun, "DeviceHandle_to_BusTargetLun", \
7306 	RescanBus, "RescanBus", \
7307 	GetDeviceProbeInfo, "GetDeviceProbeInfo", \
7308 	GetDeviceCapacity, "GetDeviceCapacity", \
7309 	GetContainerProbeInfo, "GetContainerProbeInfo", \
7310 	GetRequestedMemorySize, "GetRequestedMemorySize", \
7311 	GetBusInfo, "GetBusInfo", \
7312 	GetVendorSpecific, "GetVendorSpecific", \
7313 	EnhancedGetDeviceProbeInfo, "EnhancedGetDeviceProbeInfo", \
7314 	EnhancedGetBusInfo, "EnhancedGetBusInfo", \
7315 	SetupExtendedCounters, "SetupExtendedCounters", \
7316 	GetPerformanceCounters, "GetPerformanceCounters", \
7317 	ResetPerformanceCounters, "ResetPerformanceCounters", \
7318 	ReadModePage, "ReadModePage", \
7319 	WriteModePage, "WriteModePage", \
7320 	ReadDriveParameter, "ReadDriveParameter", \
7321 	WriteDriveParameter, "WriteDriveParameter", \
7322 	ResetAdapter, "ResetAdapter", \
7323 	ResetBus, "ResetBus", \
7324 	ResetBusDevice, "ResetBusDevice", \
7325 	ExecuteSrb, "ExecuteSrb", \
7326 	Create_IO_Task, "Create_IO_Task", \
7327 	Delete_IO_Task, "Delete_IO_Task", \
7328 	Get_IO_Task_Info, "Get_IO_Task_Info", \
7329 	Check_Task_Progress, "Check_Task_Progress", \
7330 	InjectError, "InjectError", \
7331 	GetDeviceDefectCounts, "GetDeviceDefectCounts", \
7332 	GetDeviceDefectInfo, "GetDeviceDefectInfo", \
7333 	GetDeviceStatus, "GetDeviceStatus", \
7334 	ClearDeviceStatus, "ClearDeviceStatus", \
7335 	DiskSpinControl, "DiskSpinControl", \
7336 	DiskSmartControl, "DiskSmartControl", \
7337 	WriteSame, "WriteSame", \
7338 	ReadWriteLong, "ReadWriteLong", \
7339 	FormatUnit, "FormatUnit", \
7340 	TargetDeviceControl, "TargetDeviceControl", \
7341 	TargetChannelControl, "TargetChannelControl", \
7342 	FlashNewCode, "FlashNewCode", \
7343 	DiskCheck, "DiskCheck", \
7344 	RequestSense, "RequestSense", \
7345 	DiskPERControl, "DiskPERControl", \
7346 	Read10, "Read10", \
7347 	Write10, "Write10"
7348 
7349 #define	AAC_AIFEN_KEY_STRINGS \
7350 	AifEnGeneric, "Generic", \
7351 	AifEnTaskComplete, "TaskComplete", \
7352 	AifEnConfigChange, "Config change", \
7353 	AifEnContainerChange, "Container change", \
7354 	AifEnDeviceFailure, "device failed", \
7355 	AifEnMirrorFailover, "Mirror failover", \
7356 	AifEnContainerEvent, "container event", \
7357 	AifEnFileSystemChange, "File system changed", \
7358 	AifEnConfigPause, "Container pause event", \
7359 	AifEnConfigResume, "Container resume event", \
7360 	AifEnFailoverChange, "Failover space assignment changed", \
7361 	AifEnRAID5RebuildDone, "RAID5 rebuild finished", \
7362 	AifEnEnclosureManagement, "Enclosure management event", \
7363 	AifEnBatteryEvent, "battery event", \
7364 	AifEnAddContainer, "Add container", \
7365 	AifEnDeleteContainer, "Delete container", \
7366 	AifEnSMARTEvent, "SMART Event", \
7367 	AifEnBatteryNeedsRecond, "battery needs reconditioning", \
7368 	AifEnClusterEvent, "cluster event", \
7369 	AifEnDiskSetEvent, "disk set event occured", \
7370 	AifDenMorphComplete, "morph operation completed", \
7371 	AifDenVolumeExtendComplete, "VolumeExtendComplete"
7372 
7373 struct aac_key_strings {
7374 	int key;
7375 	char *message;
7376 };
7377 
7378 extern struct scsi_key_strings scsi_cmds[];
7379 
7380 static struct aac_key_strings aac_fib_cmds[] = {
7381 	AAC_FIB_CMD_KEY_STRINGS,
7382 	-1,			NULL
7383 };
7384 
7385 static struct aac_key_strings aac_ctvm_subcmds[] = {
7386 	AAC_CTVM_SUBCMD_KEY_STRINGS,
7387 	-1,			NULL
7388 };
7389 
7390 static struct aac_key_strings aac_ct_subcmds[] = {
7391 	AAC_CT_SUBCMD_KEY_STRINGS,
7392 	-1,			NULL
7393 };
7394 
7395 static struct aac_key_strings aac_cl_subcmds[] = {
7396 	AAC_CL_SUBCMD_KEY_STRINGS,
7397 	-1,			NULL
7398 };
7399 
7400 static struct aac_key_strings aac_aif_subcmds[] = {
7401 	AAC_AIF_SUBCMD_KEY_STRINGS,
7402 	-1,			NULL
7403 };
7404 
7405 static struct aac_key_strings aac_ioctl_subcmds[] = {
7406 	AAC_IOCTL_SUBCMD_KEY_STRINGS,
7407 	-1,			NULL
7408 };
7409 
7410 static struct aac_key_strings aac_aifens[] = {
7411 	AAC_AIFEN_KEY_STRINGS,
7412 	-1,			NULL
7413 };
7414 
7415 /*
7416  * The following function comes from Adaptec:
7417  *
7418  * Get the firmware print buffer parameters from the firmware,
7419  * if the command was successful map in the address.
7420  */
7421 static int
7422 aac_get_fw_debug_buffer(struct aac_softstate *softs)
7423 {
7424 	if (aac_sync_mbcommand(softs, AAC_MONKER_GETDRVPROP,
7425 	    0, 0, 0, 0, NULL) == AACOK) {
7426 		uint32_t mondrv_buf_paddrl = AAC_MAILBOX_GET(softs, 1);
7427 		uint32_t mondrv_buf_paddrh = AAC_MAILBOX_GET(softs, 2);
7428 		uint32_t mondrv_buf_size = AAC_MAILBOX_GET(softs, 3);
7429 		uint32_t mondrv_hdr_size = AAC_MAILBOX_GET(softs, 4);
7430 
7431 		if (mondrv_buf_size) {
7432 			uint32_t offset = mondrv_buf_paddrl - \
7433 			    softs->pci_mem_base_paddr;
7434 
7435 			/*
7436 			 * See if the address is already mapped in, and
7437 			 * if so set it up from the base address
7438 			 */
7439 			if ((mondrv_buf_paddrh == 0) &&
7440 			    (offset + mondrv_buf_size < softs->map_size)) {
7441 				mutex_enter(&aac_prt_mutex);
7442 				softs->debug_buf_offset = offset;
7443 				softs->debug_header_size = mondrv_hdr_size;
7444 				softs->debug_buf_size = mondrv_buf_size;
7445 				softs->debug_fw_flags = 0;
7446 				softs->debug_flags &= ~AACDB_FLAGS_FW_PRINT;
7447 				mutex_exit(&aac_prt_mutex);
7448 
7449 				return (AACOK);
7450 			}
7451 		}
7452 	}
7453 	return (AACERR);
7454 }
7455 
7456 int
7457 aac_dbflag_on(struct aac_softstate *softs, int flag)
7458 {
7459 	int debug_flags = softs ? softs->debug_flags : aac_debug_flags;
7460 
7461 	return ((debug_flags & (AACDB_FLAGS_FW_PRINT | \
7462 	    AACDB_FLAGS_KERNEL_PRINT)) && (debug_flags & flag));
7463 }
7464 
7465 static void
7466 aac_cmn_err(struct aac_softstate *softs, uint_t lev, char sl, int noheader)
7467 {
7468 	if (noheader) {
7469 		if (sl) {
7470 			aac_fmt[0] = sl;
7471 			cmn_err(lev, aac_fmt, aac_prt_buf);
7472 		} else {
7473 			cmn_err(lev, &aac_fmt[1], aac_prt_buf);
7474 		}
7475 	} else {
7476 		if (sl) {
7477 			aac_fmt_header[0] = sl;
7478 			cmn_err(lev, aac_fmt_header,
7479 			    softs->vendor_name, softs->instance,
7480 			    aac_prt_buf);
7481 		} else {
7482 			cmn_err(lev, &aac_fmt_header[1],
7483 			    softs->vendor_name, softs->instance,
7484 			    aac_prt_buf);
7485 		}
7486 	}
7487 }
7488 
7489 /*
7490  * The following function comes from Adaptec:
7491  *
7492  * Format and print out the data passed in to UART or console
7493  * as specified by debug flags.
7494  */
7495 void
7496 aac_printf(struct aac_softstate *softs, uint_t lev, const char *fmt, ...)
7497 {
7498 	va_list args;
7499 	char sl; /* system log character */
7500 
7501 	mutex_enter(&aac_prt_mutex);
7502 	/* Set up parameters and call sprintf function to format the data */
7503 	if (strchr("^!?", fmt[0]) == NULL) {
7504 		sl = 0;
7505 	} else {
7506 		sl = fmt[0];
7507 		fmt++;
7508 	}
7509 	va_start(args, fmt);
7510 	(void) vsprintf(aac_prt_buf, fmt, args);
7511 	va_end(args);
7512 
7513 	/* Make sure the softs structure has been passed in for this section */
7514 	if (softs) {
7515 		if ((softs->debug_flags & AACDB_FLAGS_FW_PRINT) &&
7516 		    /* If we are set up for a Firmware print */
7517 		    (softs->debug_buf_size)) {
7518 			uint32_t count, i;
7519 
7520 			/* Make sure the string size is within boundaries */
7521 			count = strlen(aac_prt_buf);
7522 			if (count > softs->debug_buf_size)
7523 				count = (uint16_t)softs->debug_buf_size;
7524 
7525 			/*
7526 			 * Wait for no more than AAC_PRINT_TIMEOUT for the
7527 			 * previous message length to clear (the handshake).
7528 			 */
7529 			for (i = 0; i < AAC_PRINT_TIMEOUT; i++) {
7530 				if (!PCI_MEM_GET32(softs,
7531 				    softs->debug_buf_offset + \
7532 				    AAC_FW_DBG_STRLEN_OFFSET))
7533 					break;
7534 
7535 				drv_usecwait(1000);
7536 			}
7537 
7538 			/*
7539 			 * If the length is clear, copy over the message, the
7540 			 * flags, and the length. Make sure the length is the
7541 			 * last because that is the signal for the Firmware to
7542 			 * pick it up.
7543 			 */
7544 			if (!PCI_MEM_GET32(softs, softs->debug_buf_offset + \
7545 			    AAC_FW_DBG_STRLEN_OFFSET)) {
7546 				PCI_MEM_REP_PUT8(softs,
7547 				    softs->debug_buf_offset + \
7548 				    softs->debug_header_size,
7549 				    aac_prt_buf, count);
7550 				PCI_MEM_PUT32(softs,
7551 				    softs->debug_buf_offset + \
7552 				    AAC_FW_DBG_FLAGS_OFFSET,
7553 				    softs->debug_fw_flags);
7554 				PCI_MEM_PUT32(softs,
7555 				    softs->debug_buf_offset + \
7556 				    AAC_FW_DBG_STRLEN_OFFSET, count);
7557 			} else {
7558 				cmn_err(CE_WARN, "UART output fail");
7559 				softs->debug_flags &= ~AACDB_FLAGS_FW_PRINT;
7560 			}
7561 		}
7562 
7563 		/*
7564 		 * If the Kernel Debug Print flag is set, send it off
7565 		 * to the Kernel Debugger
7566 		 */
7567 		if (softs->debug_flags & AACDB_FLAGS_KERNEL_PRINT)
7568 			aac_cmn_err(softs, lev, sl,
7569 			    (softs->debug_flags & AACDB_FLAGS_NO_HEADERS));
7570 	} else {
7571 		/* Driver not initialized yet, no firmware or header output */
7572 		if (aac_debug_flags & AACDB_FLAGS_KERNEL_PRINT)
7573 			aac_cmn_err(softs, lev, sl, 1);
7574 	}
7575 	mutex_exit(&aac_prt_mutex);
7576 }
7577 
7578 /*
7579  * Translate command number to description string
7580  */
7581 static char *
7582 aac_cmd_name(int cmd, struct aac_key_strings *cmdlist)
7583 {
7584 	int i;
7585 
7586 	for (i = 0; cmdlist[i].key != -1; i++) {
7587 		if (cmd == cmdlist[i].key)
7588 			return (cmdlist[i].message);
7589 	}
7590 	return (NULL);
7591 }
7592 
7593 static void
7594 aac_print_scmd(struct aac_softstate *softs, struct aac_cmd *acp)
7595 {
7596 	struct scsi_pkt *pkt = acp->pkt;
7597 	struct scsi_address *ap = &pkt->pkt_address;
7598 	int is_pd = 0;
7599 	int ctl = ddi_get_instance(softs->devinfo_p);
7600 	int tgt = ap->a_target;
7601 	int lun = ap->a_lun;
7602 	union scsi_cdb *cdbp = (void *)pkt->pkt_cdbp;
7603 	uchar_t cmd = cdbp->scc_cmd;
7604 	char *desc;
7605 
7606 	if (tgt >= AAC_MAX_LD) {
7607 		is_pd = 1;
7608 		ctl = ((struct aac_nondasd *)acp->dvp)->bus;
7609 		tgt = ((struct aac_nondasd *)acp->dvp)->tid;
7610 		lun = 0;
7611 	}
7612 
7613 	if ((desc = aac_cmd_name(cmd,
7614 	    (struct aac_key_strings *)scsi_cmds)) == NULL) {
7615 		aac_printf(softs, CE_NOTE,
7616 		    "SCMD> Unknown(0x%2x) --> c%dt%dL%d %s",
7617 		    cmd, ctl, tgt, lun, is_pd ? "(pd)" : "");
7618 		return;
7619 	}
7620 
7621 	switch (cmd) {
7622 	case SCMD_READ:
7623 	case SCMD_WRITE:
7624 		aac_printf(softs, CE_NOTE,
7625 		    "SCMD> %s 0x%x[%d] %s --> c%dt%dL%d %s",
7626 		    desc, GETG0ADDR(cdbp), GETG0COUNT(cdbp),
7627 		    (acp->flags & AAC_CMD_NO_INTR) ? "poll" : "intr",
7628 		    ctl, tgt, lun, is_pd ? "(pd)" : "");
7629 		break;
7630 	case SCMD_READ_G1:
7631 	case SCMD_WRITE_G1:
7632 		aac_printf(softs, CE_NOTE,
7633 		    "SCMD> %s 0x%x[%d] %s --> c%dt%dL%d %s",
7634 		    desc, GETG1ADDR(cdbp), GETG1COUNT(cdbp),
7635 		    (acp->flags & AAC_CMD_NO_INTR) ? "poll" : "intr",
7636 		    ctl, tgt, lun, is_pd ? "(pd)" : "");
7637 		break;
7638 	case SCMD_READ_G4:
7639 	case SCMD_WRITE_G4:
7640 		aac_printf(softs, CE_NOTE,
7641 		    "SCMD> %s 0x%x.%08x[%d] %s --> c%dt%dL%d %s",
7642 		    desc, GETG4ADDR(cdbp), GETG4ADDRTL(cdbp),
7643 		    GETG4COUNT(cdbp),
7644 		    (acp->flags & AAC_CMD_NO_INTR) ? "poll" : "intr",
7645 		    ctl, tgt, lun, is_pd ? "(pd)" : "");
7646 		break;
7647 	case SCMD_READ_G5:
7648 	case SCMD_WRITE_G5:
7649 		aac_printf(softs, CE_NOTE,
7650 		    "SCMD> %s 0x%x[%d] %s --> c%dt%dL%d %s",
7651 		    desc, GETG5ADDR(cdbp), GETG5COUNT(cdbp),
7652 		    (acp->flags & AAC_CMD_NO_INTR) ? "poll" : "intr",
7653 		    ctl, tgt, lun, is_pd ? "(pd)" : "");
7654 		break;
7655 	default:
7656 		aac_printf(softs, CE_NOTE, "SCMD> %s --> c%dt%dL%d %s",
7657 		    desc, ctl, tgt, lun, is_pd ? "(pd)" : "");
7658 	}
7659 }
7660 
7661 void
7662 aac_print_fib(struct aac_softstate *softs, struct aac_slot *slotp)
7663 {
7664 	struct aac_cmd *acp = slotp->acp;
7665 	struct aac_fib *fibp = slotp->fibp;
7666 	ddi_acc_handle_t acc = slotp->fib_acc_handle;
7667 	uint16_t fib_size;
7668 	uint32_t fib_cmd, sub_cmd;
7669 	char *cmdstr, *subcmdstr;
7670 	char *caller;
7671 	int i;
7672 
7673 	if (acp) {
7674 		if (!(softs->debug_fib_flags & acp->fib_flags))
7675 			return;
7676 		if (acp->fib_flags & AACDB_FLAGS_FIB_SCMD)
7677 			caller = "SCMD";
7678 		else if (acp->fib_flags & AACDB_FLAGS_FIB_IOCTL)
7679 			caller = "IOCTL";
7680 		else if (acp->fib_flags & AACDB_FLAGS_FIB_SRB)
7681 			caller = "SRB";
7682 		else
7683 			return;
7684 	} else {
7685 		if (!(softs->debug_fib_flags & AACDB_FLAGS_FIB_SYNC))
7686 			return;
7687 		caller = "SYNC";
7688 	}
7689 
7690 	fib_cmd = ddi_get16(acc, &fibp->Header.Command);
7691 	cmdstr = aac_cmd_name(fib_cmd, aac_fib_cmds);
7692 	sub_cmd = (uint32_t)-1;
7693 	subcmdstr = NULL;
7694 
7695 	/* Print FIB header */
7696 	if (softs->debug_fib_flags & AACDB_FLAGS_FIB_HEADER) {
7697 		aac_printf(softs, CE_NOTE, "FIB> from %s", caller);
7698 		aac_printf(softs, CE_NOTE, "     XferState  %d",
7699 		    ddi_get32(acc, &fibp->Header.XferState));
7700 		aac_printf(softs, CE_NOTE, "     Command    %d",
7701 		    ddi_get16(acc, &fibp->Header.Command));
7702 		aac_printf(softs, CE_NOTE, "     StructType %d",
7703 		    ddi_get8(acc, &fibp->Header.StructType));
7704 		aac_printf(softs, CE_NOTE, "     Flags      0x%x",
7705 		    ddi_get8(acc, &fibp->Header.Flags));
7706 		aac_printf(softs, CE_NOTE, "     Size       %d",
7707 		    ddi_get16(acc, &fibp->Header.Size));
7708 		aac_printf(softs, CE_NOTE, "     SenderSize %d",
7709 		    ddi_get16(acc, &fibp->Header.SenderSize));
7710 		aac_printf(softs, CE_NOTE, "     SenderAddr 0x%x",
7711 		    ddi_get32(acc, &fibp->Header.SenderFibAddress));
7712 		aac_printf(softs, CE_NOTE, "     RcvrAddr   0x%x",
7713 		    ddi_get32(acc, &fibp->Header.ReceiverFibAddress));
7714 		aac_printf(softs, CE_NOTE, "     SenderData 0x%x",
7715 		    ddi_get32(acc, &fibp->Header.SenderData));
7716 	}
7717 
7718 	/* Print FIB data */
7719 	switch (fib_cmd) {
7720 	case ContainerCommand:
7721 		sub_cmd = ddi_get32(acc,
7722 		    (void *)&(((uint32_t *)(void *)&fibp->data[0])[0]));
7723 		subcmdstr = aac_cmd_name(sub_cmd, aac_ctvm_subcmds);
7724 		if (subcmdstr == NULL)
7725 			break;
7726 
7727 		switch (sub_cmd) {
7728 		case VM_ContainerConfig: {
7729 			struct aac_Container *pContainer =
7730 			    (struct aac_Container *)fibp->data;
7731 
7732 			fib_cmd = sub_cmd;
7733 			cmdstr = subcmdstr;
7734 			sub_cmd = (uint32_t)-1;
7735 			subcmdstr = NULL;
7736 
7737 			sub_cmd = ddi_get32(acc,
7738 			    &pContainer->CTCommand.command);
7739 			subcmdstr = aac_cmd_name(sub_cmd, aac_ct_subcmds);
7740 			if (subcmdstr == NULL)
7741 				break;
7742 			aac_printf(softs, CE_NOTE, "FIB> %s (0x%x, 0x%x, 0x%x)",
7743 			    subcmdstr,
7744 			    ddi_get32(acc, &pContainer->CTCommand.param[0]),
7745 			    ddi_get32(acc, &pContainer->CTCommand.param[1]),
7746 			    ddi_get32(acc, &pContainer->CTCommand.param[2]));
7747 			return;
7748 		}
7749 
7750 		case VM_Ioctl:
7751 			fib_cmd = sub_cmd;
7752 			cmdstr = subcmdstr;
7753 			sub_cmd = (uint32_t)-1;
7754 			subcmdstr = NULL;
7755 
7756 			sub_cmd = ddi_get32(acc,
7757 			    (void *)&(((uint32_t *)(void *)&fibp->data[0])[4]));
7758 			subcmdstr = aac_cmd_name(sub_cmd, aac_ioctl_subcmds);
7759 			break;
7760 
7761 		case VM_CtBlockRead:
7762 		case VM_CtBlockWrite: {
7763 			struct aac_blockread *br =
7764 			    (struct aac_blockread *)fibp->data;
7765 			struct aac_sg_table *sg = &br->SgMap;
7766 			uint32_t sgcount = ddi_get32(acc, &sg->SgCount);
7767 
7768 			aac_printf(softs, CE_NOTE,
7769 			    "FIB> %s Container %d  0x%x/%d", subcmdstr,
7770 			    ddi_get32(acc, &br->ContainerId),
7771 			    ddi_get32(acc, &br->BlockNumber),
7772 			    ddi_get32(acc, &br->ByteCount));
7773 			for (i = 0; i < sgcount; i++)
7774 				aac_printf(softs, CE_NOTE,
7775 				    "     %d: 0x%08x/%d", i,
7776 				    ddi_get32(acc, &sg->SgEntry[i].SgAddress),
7777 				    ddi_get32(acc, &sg->SgEntry[i]. \
7778 				    SgByteCount));
7779 			return;
7780 		}
7781 		}
7782 		break;
7783 
7784 	case ContainerCommand64: {
7785 		struct aac_blockread64 *br =
7786 		    (struct aac_blockread64 *)fibp->data;
7787 		struct aac_sg_table64 *sg = &br->SgMap64;
7788 		uint32_t sgcount = ddi_get32(acc, &sg->SgCount);
7789 		uint64_t sgaddr;
7790 
7791 		sub_cmd = br->Command;
7792 		subcmdstr = NULL;
7793 		if (sub_cmd == VM_CtHostRead64)
7794 			subcmdstr = "VM_CtHostRead64";
7795 		else if (sub_cmd == VM_CtHostWrite64)
7796 			subcmdstr = "VM_CtHostWrite64";
7797 		else
7798 			break;
7799 
7800 		aac_printf(softs, CE_NOTE,
7801 		    "FIB> %s Container %d  0x%x/%d", subcmdstr,
7802 		    ddi_get16(acc, &br->ContainerId),
7803 		    ddi_get32(acc, &br->BlockNumber),
7804 		    ddi_get16(acc, &br->SectorCount));
7805 		for (i = 0; i < sgcount; i++) {
7806 			sgaddr = ddi_get64(acc,
7807 			    &sg->SgEntry64[i].SgAddress);
7808 			aac_printf(softs, CE_NOTE,
7809 			    "     %d: 0x%08x.%08x/%d", i,
7810 			    AAC_MS32(sgaddr), AAC_LS32(sgaddr),
7811 			    ddi_get32(acc, &sg->SgEntry64[i]. \
7812 			    SgByteCount));
7813 		}
7814 		return;
7815 	}
7816 
7817 	case RawIo: {
7818 		struct aac_raw_io *io = (struct aac_raw_io *)fibp->data;
7819 		struct aac_sg_tableraw *sg = &io->SgMapRaw;
7820 		uint32_t sgcount = ddi_get32(acc, &sg->SgCount);
7821 		uint64_t sgaddr;
7822 
7823 		aac_printf(softs, CE_NOTE,
7824 		    "FIB> RawIo Container %d  0x%llx/%d 0x%x",
7825 		    ddi_get16(acc, &io->ContainerId),
7826 		    ddi_get64(acc, &io->BlockNumber),
7827 		    ddi_get32(acc, &io->ByteCount),
7828 		    ddi_get16(acc, &io->Flags));
7829 		for (i = 0; i < sgcount; i++) {
7830 			sgaddr = ddi_get64(acc, &sg->SgEntryRaw[i].SgAddress);
7831 			aac_printf(softs, CE_NOTE, "     %d: 0x%08x.%08x/%d", i,
7832 			    AAC_MS32(sgaddr), AAC_LS32(sgaddr),
7833 			    ddi_get32(acc, &sg->SgEntryRaw[i].SgByteCount));
7834 		}
7835 		return;
7836 	}
7837 
7838 	case ClusterCommand:
7839 		sub_cmd = ddi_get32(acc,
7840 		    (void *)&(((uint32_t *)(void *)fibp->data)[0]));
7841 		subcmdstr = aac_cmd_name(sub_cmd, aac_cl_subcmds);
7842 		break;
7843 
7844 	case AifRequest:
7845 		sub_cmd = ddi_get32(acc,
7846 		    (void *)&(((uint32_t *)(void *)fibp->data)[0]));
7847 		subcmdstr = aac_cmd_name(sub_cmd, aac_aif_subcmds);
7848 		break;
7849 
7850 	default:
7851 		break;
7852 	}
7853 
7854 	fib_size = ddi_get16(acc, &(fibp->Header.Size));
7855 	if (subcmdstr)
7856 		aac_printf(softs, CE_NOTE, "FIB> %s, sz=%d",
7857 		    subcmdstr, fib_size);
7858 	else if (cmdstr && sub_cmd == (uint32_t)-1)
7859 		aac_printf(softs, CE_NOTE, "FIB> %s, sz=%d",
7860 		    cmdstr, fib_size);
7861 	else if (cmdstr)
7862 		aac_printf(softs, CE_NOTE, "FIB> %s: Unknown(0x%x), sz=%d",
7863 		    cmdstr, sub_cmd, fib_size);
7864 	else
7865 		aac_printf(softs, CE_NOTE, "FIB> Unknown(0x%x), sz=%d",
7866 		    fib_cmd, fib_size);
7867 }
7868 
7869 static void
7870 aac_print_aif(struct aac_softstate *softs, struct aac_aif_command *aif)
7871 {
7872 	int aif_command;
7873 	uint32_t aif_seqnumber;
7874 	int aif_en_type;
7875 	char *str;
7876 
7877 	aif_command = LE_32(aif->command);
7878 	aif_seqnumber = LE_32(aif->seqNumber);
7879 	aif_en_type = LE_32(aif->data.EN.type);
7880 
7881 	switch (aif_command) {
7882 	case AifCmdEventNotify:
7883 		str = aac_cmd_name(aif_en_type, aac_aifens);
7884 		if (str)
7885 			aac_printf(softs, CE_NOTE, "AIF! %s", str);
7886 		else
7887 			aac_printf(softs, CE_NOTE, "AIF! Unknown(0x%x)",
7888 			    aif_en_type);
7889 		break;
7890 
7891 	case AifCmdJobProgress:
7892 		switch (LE_32(aif->data.PR[0].status)) {
7893 		case AifJobStsSuccess:
7894 			str = "success"; break;
7895 		case AifJobStsFinished:
7896 			str = "finished"; break;
7897 		case AifJobStsAborted:
7898 			str = "aborted"; break;
7899 		case AifJobStsFailed:
7900 			str = "failed"; break;
7901 		case AifJobStsSuspended:
7902 			str = "suspended"; break;
7903 		case AifJobStsRunning:
7904 			str = "running"; break;
7905 		default:
7906 			str = "unknown"; break;
7907 		}
7908 		aac_printf(softs, CE_NOTE,
7909 		    "AIF! JobProgress (%d) - %s (%d, %d)",
7910 		    aif_seqnumber, str,
7911 		    LE_32(aif->data.PR[0].currentTick),
7912 		    LE_32(aif->data.PR[0].finalTick));
7913 		break;
7914 
7915 	case AifCmdAPIReport:
7916 		aac_printf(softs, CE_NOTE, "AIF! APIReport (%d)",
7917 		    aif_seqnumber);
7918 		break;
7919 
7920 	case AifCmdDriverNotify:
7921 		aac_printf(softs, CE_NOTE, "AIF! DriverNotify (%d)",
7922 		    aif_seqnumber);
7923 		break;
7924 
7925 	default:
7926 		aac_printf(softs, CE_NOTE, "AIF! AIF %d (%d)",
7927 		    aif_command, aif_seqnumber);
7928 		break;
7929 	}
7930 }
7931 
7932 #endif /* DEBUG */
7933