xref: /linux/drivers/scsi/aic7xxx/aic7xxx_osm.c (revision 13abf8130139c2ccd4962a7e5a8902be5e6cb5a7)
1 /*
2  * Adaptec AIC7xxx device driver for Linux.
3  *
4  * $Id: //depot/aic7xxx/linux/drivers/scsi/aic7xxx/aic7xxx_osm.c#235 $
5  *
6  * Copyright (c) 1994 John Aycock
7  *   The University of Calgary Department of Computer Science.
8  *
9  * This program is free software; you can redistribute it and/or modify
10  * it under the terms of the GNU General Public License as published by
11  * the Free Software Foundation; either version 2, or (at your option)
12  * any later version.
13  *
14  * This program is distributed in the hope that it will be useful,
15  * but WITHOUT ANY WARRANTY; without even the implied warranty of
16  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
17  * GNU General Public License for more details.
18  *
19  * You should have received a copy of the GNU General Public License
20  * along with this program; see the file COPYING.  If not, write to
21  * the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139, USA.
22  *
23  * Sources include the Adaptec 1740 driver (aha1740.c), the Ultrastor 24F
24  * driver (ultrastor.c), various Linux kernel source, the Adaptec EISA
25  * config file (!adp7771.cfg), the Adaptec AHA-2740A Series User's Guide,
26  * the Linux Kernel Hacker's Guide, Writing a SCSI Device Driver for Linux,
27  * the Adaptec 1542 driver (aha1542.c), the Adaptec EISA overlay file
28  * (adp7770.ovl), the Adaptec AHA-2740 Series Technical Reference Manual,
29  * the Adaptec AIC-7770 Data Book, the ANSI SCSI specification, the
30  * ANSI SCSI-2 specification (draft 10c), ...
31  *
32  * --------------------------------------------------------------------------
33  *
34  *  Modifications by Daniel M. Eischen (deischen@iworks.InterWorks.org):
35  *
36  *  Substantially modified to include support for wide and twin bus
37  *  adapters, DMAing of SCBs, tagged queueing, IRQ sharing, bug fixes,
38  *  SCB paging, and other rework of the code.
39  *
40  * --------------------------------------------------------------------------
41  * Copyright (c) 1994-2000 Justin T. Gibbs.
42  * Copyright (c) 2000-2001 Adaptec Inc.
43  * All rights reserved.
44  *
45  * Redistribution and use in source and binary forms, with or without
46  * modification, are permitted provided that the following conditions
47  * are met:
48  * 1. Redistributions of source code must retain the above copyright
49  *    notice, this list of conditions, and the following disclaimer,
50  *    without modification.
51  * 2. Redistributions in binary form must reproduce at minimum a disclaimer
52  *    substantially similar to the "NO WARRANTY" disclaimer below
53  *    ("Disclaimer") and any redistribution must be conditioned upon
54  *    including a substantially similar Disclaimer requirement for further
55  *    binary redistribution.
56  * 3. Neither the names of the above-listed copyright holders nor the names
57  *    of any contributors may be used to endorse or promote products derived
58  *    from this software without specific prior written permission.
59  *
60  * Alternatively, this software may be distributed under the terms of the
61  * GNU General Public License ("GPL") version 2 as published by the Free
62  * Software Foundation.
63  *
64  * NO WARRANTY
65  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
66  * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
67  * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTIBILITY AND FITNESS FOR
68  * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
69  * HOLDERS OR CONTRIBUTORS BE LIABLE FOR SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
70  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
71  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
72  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,
73  * STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING
74  * IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
75  * POSSIBILITY OF SUCH DAMAGES.
76  *
77  *---------------------------------------------------------------------------
78  *
79  *  Thanks also go to (in alphabetical order) the following:
80  *
81  *    Rory Bolt     - Sequencer bug fixes
82  *    Jay Estabrook - Initial DEC Alpha support
83  *    Doug Ledford  - Much needed abort/reset bug fixes
84  *    Kai Makisara  - DMAing of SCBs
85  *
86  *  A Boot time option was also added for not resetting the scsi bus.
87  *
88  *    Form:  aic7xxx=extended
89  *           aic7xxx=no_reset
90  *           aic7xxx=verbose
91  *
92  *  Daniel M. Eischen, deischen@iworks.InterWorks.org, 1/23/97
93  *
94  *  Id: aic7xxx.c,v 4.1 1997/06/12 08:23:42 deang Exp
95  */
96 
97 /*
98  * Further driver modifications made by Doug Ledford <dledford@redhat.com>
99  *
100  * Copyright (c) 1997-1999 Doug Ledford
101  *
102  * These changes are released under the same licensing terms as the FreeBSD
103  * driver written by Justin Gibbs.  Please see his Copyright notice above
104  * for the exact terms and conditions covering my changes as well as the
105  * warranty statement.
106  *
107  * Modifications made to the aic7xxx.c,v 4.1 driver from Dan Eischen include
108  * but are not limited to:
109  *
110  *  1: Import of the latest FreeBSD sequencer code for this driver
111  *  2: Modification of kernel code to accommodate different sequencer semantics
112  *  3: Extensive changes throughout kernel portion of driver to improve
113  *     abort/reset processing and error hanndling
114  *  4: Other work contributed by various people on the Internet
115  *  5: Changes to printk information and verbosity selection code
116  *  6: General reliability related changes, especially in IRQ management
117  *  7: Modifications to the default probe/attach order for supported cards
118  *  8: SMP friendliness has been improved
119  *
120  */
121 
122 #include "aic7xxx_osm.h"
123 #include "aic7xxx_inline.h"
124 #include <scsi/scsicam.h>
125 
126 static struct scsi_transport_template *ahc_linux_transport_template = NULL;
127 
128 /*
129  * Include aiclib.c as part of our
130  * "module dependencies are hard" work around.
131  */
132 #include "aiclib.c"
133 
134 #include <linux/init.h>		/* __setup */
135 #include <linux/mm.h>		/* For fetching system memory size */
136 #include <linux/blkdev.h>		/* For block_size() */
137 #include <linux/delay.h>	/* For ssleep/msleep */
138 
139 
140 /*
141  * Set this to the delay in seconds after SCSI bus reset.
142  * Note, we honor this only for the initial bus reset.
143  * The scsi error recovery code performs its own bus settle
144  * delay handling for error recovery actions.
145  */
146 #ifdef CONFIG_AIC7XXX_RESET_DELAY_MS
147 #define AIC7XXX_RESET_DELAY CONFIG_AIC7XXX_RESET_DELAY_MS
148 #else
149 #define AIC7XXX_RESET_DELAY 5000
150 #endif
151 
152 /*
153  * Control collection of SCSI transfer statistics for the /proc filesystem.
154  *
155  * NOTE: Do NOT enable this when running on kernels version 1.2.x and below.
156  * NOTE: This does affect performance since it has to maintain statistics.
157  */
158 #ifdef CONFIG_AIC7XXX_PROC_STATS
159 #define AIC7XXX_PROC_STATS
160 #endif
161 
162 /*
163  * To change the default number of tagged transactions allowed per-device,
164  * add a line to the lilo.conf file like:
165  * append="aic7xxx=verbose,tag_info:{{32,32,32,32},{32,32,32,32}}"
166  * which will result in the first four devices on the first two
167  * controllers being set to a tagged queue depth of 32.
168  *
169  * The tag_commands is an array of 16 to allow for wide and twin adapters.
170  * Twin adapters will use indexes 0-7 for channel 0, and indexes 8-15
171  * for channel 1.
172  */
173 typedef struct {
174 	uint8_t tag_commands[16];	/* Allow for wide/twin adapters. */
175 } adapter_tag_info_t;
176 
177 /*
178  * Modify this as you see fit for your system.
179  *
180  * 0			tagged queuing disabled
181  * 1 <= n <= 253	n == max tags ever dispatched.
182  *
183  * The driver will throttle the number of commands dispatched to a
184  * device if it returns queue full.  For devices with a fixed maximum
185  * queue depth, the driver will eventually determine this depth and
186  * lock it in (a console message is printed to indicate that a lock
187  * has occurred).  On some devices, queue full is returned for a temporary
188  * resource shortage.  These devices will return queue full at varying
189  * depths.  The driver will throttle back when the queue fulls occur and
190  * attempt to slowly increase the depth over time as the device recovers
191  * from the resource shortage.
192  *
193  * In this example, the first line will disable tagged queueing for all
194  * the devices on the first probed aic7xxx adapter.
195  *
196  * The second line enables tagged queueing with 4 commands/LUN for IDs
197  * (0, 2-11, 13-15), disables tagged queueing for ID 12, and tells the
198  * driver to attempt to use up to 64 tags for ID 1.
199  *
200  * The third line is the same as the first line.
201  *
202  * The fourth line disables tagged queueing for devices 0 and 3.  It
203  * enables tagged queueing for the other IDs, with 16 commands/LUN
204  * for IDs 1 and 4, 127 commands/LUN for ID 8, and 4 commands/LUN for
205  * IDs 2, 5-7, and 9-15.
206  */
207 
208 /*
209  * NOTE: The below structure is for reference only, the actual structure
210  *       to modify in order to change things is just below this comment block.
211 adapter_tag_info_t aic7xxx_tag_info[] =
212 {
213 	{{0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0}},
214 	{{4, 64, 4, 4, 4, 4, 4, 4, 4, 4, 4, 4, 0, 4, 4, 4}},
215 	{{0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0}},
216 	{{0, 16, 4, 0, 16, 4, 4, 4, 127, 4, 4, 4, 4, 4, 4, 4}}
217 };
218 */
219 
220 #ifdef CONFIG_AIC7XXX_CMDS_PER_DEVICE
221 #define AIC7XXX_CMDS_PER_DEVICE CONFIG_AIC7XXX_CMDS_PER_DEVICE
222 #else
223 #define AIC7XXX_CMDS_PER_DEVICE AHC_MAX_QUEUE
224 #endif
225 
226 #define AIC7XXX_CONFIGED_TAG_COMMANDS {					\
227 	AIC7XXX_CMDS_PER_DEVICE, AIC7XXX_CMDS_PER_DEVICE,		\
228 	AIC7XXX_CMDS_PER_DEVICE, AIC7XXX_CMDS_PER_DEVICE,		\
229 	AIC7XXX_CMDS_PER_DEVICE, AIC7XXX_CMDS_PER_DEVICE,		\
230 	AIC7XXX_CMDS_PER_DEVICE, AIC7XXX_CMDS_PER_DEVICE,		\
231 	AIC7XXX_CMDS_PER_DEVICE, AIC7XXX_CMDS_PER_DEVICE,		\
232 	AIC7XXX_CMDS_PER_DEVICE, AIC7XXX_CMDS_PER_DEVICE,		\
233 	AIC7XXX_CMDS_PER_DEVICE, AIC7XXX_CMDS_PER_DEVICE,		\
234 	AIC7XXX_CMDS_PER_DEVICE, AIC7XXX_CMDS_PER_DEVICE		\
235 }
236 
237 /*
238  * By default, use the number of commands specified by
239  * the users kernel configuration.
240  */
241 static adapter_tag_info_t aic7xxx_tag_info[] =
242 {
243 	{AIC7XXX_CONFIGED_TAG_COMMANDS},
244 	{AIC7XXX_CONFIGED_TAG_COMMANDS},
245 	{AIC7XXX_CONFIGED_TAG_COMMANDS},
246 	{AIC7XXX_CONFIGED_TAG_COMMANDS},
247 	{AIC7XXX_CONFIGED_TAG_COMMANDS},
248 	{AIC7XXX_CONFIGED_TAG_COMMANDS},
249 	{AIC7XXX_CONFIGED_TAG_COMMANDS},
250 	{AIC7XXX_CONFIGED_TAG_COMMANDS},
251 	{AIC7XXX_CONFIGED_TAG_COMMANDS},
252 	{AIC7XXX_CONFIGED_TAG_COMMANDS},
253 	{AIC7XXX_CONFIGED_TAG_COMMANDS},
254 	{AIC7XXX_CONFIGED_TAG_COMMANDS},
255 	{AIC7XXX_CONFIGED_TAG_COMMANDS},
256 	{AIC7XXX_CONFIGED_TAG_COMMANDS},
257 	{AIC7XXX_CONFIGED_TAG_COMMANDS},
258 	{AIC7XXX_CONFIGED_TAG_COMMANDS}
259 };
260 
261 /*
262  * There should be a specific return value for this in scsi.h, but
263  * it seems that most drivers ignore it.
264  */
265 #define DID_UNDERFLOW   DID_ERROR
266 
267 void
268 ahc_print_path(struct ahc_softc *ahc, struct scb *scb)
269 {
270 	printk("(scsi%d:%c:%d:%d): ",
271 	       ahc->platform_data->host->host_no,
272 	       scb != NULL ? SCB_GET_CHANNEL(ahc, scb) : 'X',
273 	       scb != NULL ? SCB_GET_TARGET(ahc, scb) : -1,
274 	       scb != NULL ? SCB_GET_LUN(scb) : -1);
275 }
276 
277 /*
278  * XXX - these options apply unilaterally to _all_ 274x/284x/294x
279  *       cards in the system.  This should be fixed.  Exceptions to this
280  *       rule are noted in the comments.
281  */
282 
283 /*
284  * Skip the scsi bus reset.  Non 0 make us skip the reset at startup.  This
285  * has no effect on any later resets that might occur due to things like
286  * SCSI bus timeouts.
287  */
288 static uint32_t aic7xxx_no_reset;
289 
290 /*
291  * Should we force EXTENDED translation on a controller.
292  *     0 == Use whatever is in the SEEPROM or default to off
293  *     1 == Use whatever is in the SEEPROM or default to on
294  */
295 static uint32_t aic7xxx_extended;
296 
297 /*
298  * PCI bus parity checking of the Adaptec controllers.  This is somewhat
299  * dubious at best.  To my knowledge, this option has never actually
300  * solved a PCI parity problem, but on certain machines with broken PCI
301  * chipset configurations where stray PCI transactions with bad parity are
302  * the norm rather than the exception, the error messages can be overwelming.
303  * It's included in the driver for completeness.
304  *   0	   = Shut off PCI parity check
305  *   non-0 = reverse polarity pci parity checking
306  */
307 static uint32_t aic7xxx_pci_parity = ~0;
308 
309 /*
310  * There are lots of broken chipsets in the world.  Some of them will
311  * violate the PCI spec when we issue byte sized memory writes to our
312  * controller.  I/O mapped register access, if allowed by the given
313  * platform, will work in almost all cases.
314  */
315 uint32_t aic7xxx_allow_memio = ~0;
316 
317 /*
318  * So that we can set how long each device is given as a selection timeout.
319  * The table of values goes like this:
320  *   0 - 256ms
321  *   1 - 128ms
322  *   2 - 64ms
323  *   3 - 32ms
324  * We default to 256ms because some older devices need a longer time
325  * to respond to initial selection.
326  */
327 static uint32_t aic7xxx_seltime;
328 
329 /*
330  * Certain devices do not perform any aging on commands.  Should the
331  * device be saturated by commands in one portion of the disk, it is
332  * possible for transactions on far away sectors to never be serviced.
333  * To handle these devices, we can periodically send an ordered tag to
334  * force all outstanding transactions to be serviced prior to a new
335  * transaction.
336  */
337 uint32_t aic7xxx_periodic_otag;
338 
339 /*
340  * Module information and settable options.
341  */
342 static char *aic7xxx = NULL;
343 
344 MODULE_AUTHOR("Maintainer: Justin T. Gibbs <gibbs@scsiguy.com>");
345 MODULE_DESCRIPTION("Adaptec Aic77XX/78XX SCSI Host Bus Adapter driver");
346 MODULE_LICENSE("Dual BSD/GPL");
347 MODULE_VERSION(AIC7XXX_DRIVER_VERSION);
348 module_param(aic7xxx, charp, 0444);
349 MODULE_PARM_DESC(aic7xxx,
350 "period delimited, options string.\n"
351 "	verbose			Enable verbose/diagnostic logging\n"
352 "	allow_memio		Allow device registers to be memory mapped\n"
353 "	debug			Bitmask of debug values to enable\n"
354 "	no_probe		Toggle EISA/VLB controller probing\n"
355 "	probe_eisa_vl		Toggle EISA/VLB controller probing\n"
356 "	no_reset		Supress initial bus resets\n"
357 "	extended		Enable extended geometry on all controllers\n"
358 "	periodic_otag		Send an ordered tagged transaction\n"
359 "				periodically to prevent tag starvation.\n"
360 "				This may be required by some older disk\n"
361 "				drives or RAID arrays.\n"
362 "	reverse_scan		Sort PCI devices highest Bus/Slot to lowest\n"
363 "	tag_info:<tag_str>	Set per-target tag depth\n"
364 "	global_tag_depth:<int>	Global tag depth for every target\n"
365 "				on every bus\n"
366 "	seltime:<int>		Selection Timeout\n"
367 "				(0/256ms,1/128ms,2/64ms,3/32ms)\n"
368 "\n"
369 "	Sample /etc/modprobe.conf line:\n"
370 "		Toggle EISA/VLB probing\n"
371 "		Set tag depth on Controller 1/Target 1 to 10 tags\n"
372 "		Shorten the selection timeout to 128ms\n"
373 "\n"
374 "	options aic7xxx 'aic7xxx=probe_eisa_vl.tag_info:{{}.{.10}}.seltime:1'\n"
375 );
376 
377 static void ahc_linux_handle_scsi_status(struct ahc_softc *,
378 					 struct scsi_device *,
379 					 struct scb *);
380 static void ahc_linux_queue_cmd_complete(struct ahc_softc *ahc,
381 					 struct scsi_cmnd *cmd);
382 static void ahc_linux_sem_timeout(u_long arg);
383 static void ahc_linux_freeze_simq(struct ahc_softc *ahc);
384 static void ahc_linux_release_simq(u_long arg);
385 static int  ahc_linux_queue_recovery_cmd(struct scsi_cmnd *cmd, scb_flag flag);
386 static void ahc_linux_initialize_scsi_bus(struct ahc_softc *ahc);
387 static u_int ahc_linux_user_tagdepth(struct ahc_softc *ahc,
388 				     struct ahc_devinfo *devinfo);
389 static void ahc_linux_device_queue_depth(struct scsi_device *);
390 static int ahc_linux_run_command(struct ahc_softc*,
391 				 struct ahc_linux_device *,
392 				 struct scsi_cmnd *);
393 static void ahc_linux_setup_tag_info_global(char *p);
394 static aic_option_callback_t ahc_linux_setup_tag_info;
395 static int  aic7xxx_setup(char *s);
396 
397 static int ahc_linux_unit;
398 
399 
400 /********************************* Inlines ************************************/
401 static __inline void ahc_linux_unmap_scb(struct ahc_softc*, struct scb*);
402 
403 static __inline int ahc_linux_map_seg(struct ahc_softc *ahc, struct scb *scb,
404 		 		      struct ahc_dma_seg *sg,
405 				      dma_addr_t addr, bus_size_t len);
406 
407 static __inline void
408 ahc_linux_unmap_scb(struct ahc_softc *ahc, struct scb *scb)
409 {
410 	struct scsi_cmnd *cmd;
411 
412 	cmd = scb->io_ctx;
413 	ahc_sync_sglist(ahc, scb, BUS_DMASYNC_POSTWRITE);
414 	if (cmd->use_sg != 0) {
415 		struct scatterlist *sg;
416 
417 		sg = (struct scatterlist *)cmd->request_buffer;
418 		pci_unmap_sg(ahc->dev_softc, sg, cmd->use_sg,
419 			     cmd->sc_data_direction);
420 	} else if (cmd->request_bufflen != 0) {
421 		pci_unmap_single(ahc->dev_softc,
422 				 scb->platform_data->buf_busaddr,
423 				 cmd->request_bufflen,
424 				 cmd->sc_data_direction);
425 	}
426 }
427 
428 static __inline int
429 ahc_linux_map_seg(struct ahc_softc *ahc, struct scb *scb,
430 		  struct ahc_dma_seg *sg, dma_addr_t addr, bus_size_t len)
431 {
432 	int	 consumed;
433 
434 	if ((scb->sg_count + 1) > AHC_NSEG)
435 		panic("Too few segs for dma mapping.  "
436 		      "Increase AHC_NSEG\n");
437 
438 	consumed = 1;
439 	sg->addr = ahc_htole32(addr & 0xFFFFFFFF);
440 	scb->platform_data->xfer_len += len;
441 
442 	if (sizeof(dma_addr_t) > 4
443 	 && (ahc->flags & AHC_39BIT_ADDRESSING) != 0)
444 		len |= (addr >> 8) & AHC_SG_HIGH_ADDR_MASK;
445 
446 	sg->len = ahc_htole32(len);
447 	return (consumed);
448 }
449 
450 /*
451  * Return a string describing the driver.
452  */
453 static const char *
454 ahc_linux_info(struct Scsi_Host *host)
455 {
456 	static char buffer[512];
457 	char	ahc_info[256];
458 	char   *bp;
459 	struct ahc_softc *ahc;
460 
461 	bp = &buffer[0];
462 	ahc = *(struct ahc_softc **)host->hostdata;
463 	memset(bp, 0, sizeof(buffer));
464 	strcpy(bp, "Adaptec AIC7XXX EISA/VLB/PCI SCSI HBA DRIVER, Rev ");
465 	strcat(bp, AIC7XXX_DRIVER_VERSION);
466 	strcat(bp, "\n");
467 	strcat(bp, "        <");
468 	strcat(bp, ahc->description);
469 	strcat(bp, ">\n");
470 	strcat(bp, "        ");
471 	ahc_controller_info(ahc, ahc_info);
472 	strcat(bp, ahc_info);
473 	strcat(bp, "\n");
474 
475 	return (bp);
476 }
477 
478 /*
479  * Queue an SCB to the controller.
480  */
481 static int
482 ahc_linux_queue(struct scsi_cmnd * cmd, void (*scsi_done) (struct scsi_cmnd *))
483 {
484 	struct	 ahc_softc *ahc;
485 	struct	 ahc_linux_device *dev = scsi_transport_device_data(cmd->device);
486 
487 	ahc = *(struct ahc_softc **)cmd->device->host->hostdata;
488 
489 	/*
490 	 * Save the callback on completion function.
491 	 */
492 	cmd->scsi_done = scsi_done;
493 
494 	/*
495 	 * Close the race of a command that was in the process of
496 	 * being queued to us just as our simq was frozen.  Let
497 	 * DV commands through so long as we are only frozen to
498 	 * perform DV.
499 	 */
500 	if (ahc->platform_data->qfrozen != 0)
501 		return SCSI_MLQUEUE_HOST_BUSY;
502 
503 	cmd->result = CAM_REQ_INPROG << 16;
504 
505 	return ahc_linux_run_command(ahc, dev, cmd);
506 }
507 
508 static inline struct scsi_target **
509 ahc_linux_target_in_softc(struct scsi_target *starget)
510 {
511 	struct	ahc_softc *ahc =
512 		*((struct ahc_softc **)dev_to_shost(&starget->dev)->hostdata);
513 	unsigned int target_offset;
514 
515 	target_offset = starget->id;
516 	if (starget->channel != 0)
517 		target_offset += 8;
518 
519 	return &ahc->platform_data->starget[target_offset];
520 }
521 
522 static int
523 ahc_linux_target_alloc(struct scsi_target *starget)
524 {
525 	struct	ahc_softc *ahc =
526 		*((struct ahc_softc **)dev_to_shost(&starget->dev)->hostdata);
527 	struct seeprom_config *sc = ahc->seep_config;
528 	unsigned long flags;
529 	struct scsi_target **ahc_targp = ahc_linux_target_in_softc(starget);
530 	struct ahc_linux_target *targ = scsi_transport_target_data(starget);
531 	unsigned short scsirate;
532 	struct ahc_devinfo devinfo;
533 	struct ahc_initiator_tinfo *tinfo;
534 	struct ahc_tmode_tstate *tstate;
535 	char channel = starget->channel + 'A';
536 	unsigned int our_id = ahc->our_id;
537 	unsigned int target_offset;
538 
539 	target_offset = starget->id;
540 	if (starget->channel != 0)
541 		target_offset += 8;
542 
543 	if (starget->channel)
544 		our_id = ahc->our_id_b;
545 
546 	ahc_lock(ahc, &flags);
547 
548 	BUG_ON(*ahc_targp != NULL);
549 
550 	*ahc_targp = starget;
551 	memset(targ, 0, sizeof(*targ));
552 
553 	if (sc) {
554 		int maxsync = AHC_SYNCRATE_DT;
555 		int ultra = 0;
556 		int flags = sc->device_flags[target_offset];
557 
558 		if (ahc->flags & AHC_NEWEEPROM_FMT) {
559 		    if (flags & CFSYNCHISULTRA)
560 			ultra = 1;
561 		} else if (flags & CFULTRAEN)
562 			ultra = 1;
563 		/* AIC nutcase; 10MHz appears as ultra = 1, CFXFER = 0x04
564 		 * change it to ultra=0, CFXFER = 0 */
565 		if(ultra && (flags & CFXFER) == 0x04) {
566 			ultra = 0;
567 			flags &= ~CFXFER;
568 		}
569 
570 		if ((ahc->features & AHC_ULTRA2) != 0) {
571 			scsirate = (flags & CFXFER) | (ultra ? 0x8 : 0);
572 		} else {
573 			scsirate = (flags & CFXFER) << 4;
574 			maxsync = ultra ? AHC_SYNCRATE_ULTRA :
575 				AHC_SYNCRATE_FAST;
576 		}
577 		spi_max_width(starget) = (flags & CFWIDEB) ? 1 : 0;
578 		if (!(flags & CFSYNCH))
579 			spi_max_offset(starget) = 0;
580 		spi_min_period(starget) =
581 			ahc_find_period(ahc, scsirate, maxsync);
582 
583 		tinfo = ahc_fetch_transinfo(ahc, channel, ahc->our_id,
584 					    starget->id, &tstate);
585 	}
586 	ahc_compile_devinfo(&devinfo, our_id, starget->id,
587 			    CAM_LUN_WILDCARD, channel,
588 			    ROLE_INITIATOR);
589 	ahc_set_syncrate(ahc, &devinfo, NULL, 0, 0, 0,
590 			 AHC_TRANS_GOAL, /*paused*/FALSE);
591 	ahc_set_width(ahc, &devinfo, MSG_EXT_WDTR_BUS_8_BIT,
592 		      AHC_TRANS_GOAL, /*paused*/FALSE);
593 	ahc_unlock(ahc, &flags);
594 
595 	return 0;
596 }
597 
598 static void
599 ahc_linux_target_destroy(struct scsi_target *starget)
600 {
601 	struct scsi_target **ahc_targp = ahc_linux_target_in_softc(starget);
602 
603 	*ahc_targp = NULL;
604 }
605 
606 static int
607 ahc_linux_slave_alloc(struct scsi_device *sdev)
608 {
609 	struct	ahc_softc *ahc =
610 		*((struct ahc_softc **)sdev->host->hostdata);
611 	struct scsi_target *starget = sdev->sdev_target;
612 	struct ahc_linux_target *targ = scsi_transport_target_data(starget);
613 	struct ahc_linux_device *dev;
614 
615 	if (bootverbose)
616 		printf("%s: Slave Alloc %d\n", ahc_name(ahc), sdev->id);
617 
618 	BUG_ON(targ->sdev[sdev->lun] != NULL);
619 
620 	dev = scsi_transport_device_data(sdev);
621 	memset(dev, 0, sizeof(*dev));
622 
623 	/*
624 	 * We start out life using untagged
625 	 * transactions of which we allow one.
626 	 */
627 	dev->openings = 1;
628 
629 	/*
630 	 * Set maxtags to 0.  This will be changed if we
631 	 * later determine that we are dealing with
632 	 * a tagged queuing capable device.
633 	 */
634 	dev->maxtags = 0;
635 
636 	targ->sdev[sdev->lun] = sdev;
637 
638 	return 0;
639 }
640 
641 static int
642 ahc_linux_slave_configure(struct scsi_device *sdev)
643 {
644 	struct	ahc_softc *ahc;
645 
646 	ahc = *((struct ahc_softc **)sdev->host->hostdata);
647 
648 	if (bootverbose)
649 		printf("%s: Slave Configure %d\n", ahc_name(ahc), sdev->id);
650 
651 	ahc_linux_device_queue_depth(sdev);
652 
653 	/* Initial Domain Validation */
654 	if (!spi_initial_dv(sdev->sdev_target))
655 		spi_dv_device(sdev);
656 
657 	return 0;
658 }
659 
660 static void
661 ahc_linux_slave_destroy(struct scsi_device *sdev)
662 {
663 	struct	ahc_softc *ahc;
664 	struct	ahc_linux_device *dev = scsi_transport_device_data(sdev);
665 	struct	ahc_linux_target *targ = scsi_transport_target_data(sdev->sdev_target);
666 
667 	ahc = *((struct ahc_softc **)sdev->host->hostdata);
668 	if (bootverbose)
669 		printf("%s: Slave Destroy %d\n", ahc_name(ahc), sdev->id);
670 
671 	BUG_ON(dev->active);
672 
673 	targ->sdev[sdev->lun] = NULL;
674 }
675 
676 #if defined(__i386__)
677 /*
678  * Return the disk geometry for the given SCSI device.
679  */
680 static int
681 ahc_linux_biosparam(struct scsi_device *sdev, struct block_device *bdev,
682 		    sector_t capacity, int geom[])
683 {
684 	uint8_t *bh;
685 	int	 heads;
686 	int	 sectors;
687 	int	 cylinders;
688 	int	 ret;
689 	int	 extended;
690 	struct	 ahc_softc *ahc;
691 	u_int	 channel;
692 
693 	ahc = *((struct ahc_softc **)sdev->host->hostdata);
694 	channel = sdev->channel;
695 
696 	bh = scsi_bios_ptable(bdev);
697 	if (bh) {
698 		ret = scsi_partsize(bh, capacity,
699 				    &geom[2], &geom[0], &geom[1]);
700 		kfree(bh);
701 		if (ret != -1)
702 			return (ret);
703 	}
704 	heads = 64;
705 	sectors = 32;
706 	cylinders = aic_sector_div(capacity, heads, sectors);
707 
708 	if (aic7xxx_extended != 0)
709 		extended = 1;
710 	else if (channel == 0)
711 		extended = (ahc->flags & AHC_EXTENDED_TRANS_A) != 0;
712 	else
713 		extended = (ahc->flags & AHC_EXTENDED_TRANS_B) != 0;
714 	if (extended && cylinders >= 1024) {
715 		heads = 255;
716 		sectors = 63;
717 		cylinders = aic_sector_div(capacity, heads, sectors);
718 	}
719 	geom[0] = heads;
720 	geom[1] = sectors;
721 	geom[2] = cylinders;
722 	return (0);
723 }
724 #endif
725 
726 /*
727  * Abort the current SCSI command(s).
728  */
729 static int
730 ahc_linux_abort(struct scsi_cmnd *cmd)
731 {
732 	int error;
733 
734 	error = ahc_linux_queue_recovery_cmd(cmd, SCB_ABORT);
735 	if (error != 0)
736 		printf("aic7xxx_abort returns 0x%x\n", error);
737 	return (error);
738 }
739 
740 /*
741  * Attempt to send a target reset message to the device that timed out.
742  */
743 static int
744 ahc_linux_dev_reset(struct scsi_cmnd *cmd)
745 {
746 	int error;
747 
748 	error = ahc_linux_queue_recovery_cmd(cmd, SCB_DEVICE_RESET);
749 	if (error != 0)
750 		printf("aic7xxx_dev_reset returns 0x%x\n", error);
751 	return (error);
752 }
753 
754 /*
755  * Reset the SCSI bus.
756  */
757 static int
758 ahc_linux_bus_reset(struct scsi_cmnd *cmd)
759 {
760 	struct ahc_softc *ahc;
761 	int    found;
762 	unsigned long flags;
763 
764 	ahc = *(struct ahc_softc **)cmd->device->host->hostdata;
765 
766 	ahc_lock(ahc, &flags);
767 	found = ahc_reset_channel(ahc, cmd->device->channel + 'A',
768 				  /*initiate reset*/TRUE);
769 	ahc_unlock(ahc, &flags);
770 
771 	if (bootverbose)
772 		printf("%s: SCSI bus reset delivered. "
773 		       "%d SCBs aborted.\n", ahc_name(ahc), found);
774 
775 	return SUCCESS;
776 }
777 
778 struct scsi_host_template aic7xxx_driver_template = {
779 	.module			= THIS_MODULE,
780 	.name			= "aic7xxx",
781 	.proc_name		= "aic7xxx",
782 	.proc_info		= ahc_linux_proc_info,
783 	.info			= ahc_linux_info,
784 	.queuecommand		= ahc_linux_queue,
785 	.eh_abort_handler	= ahc_linux_abort,
786 	.eh_device_reset_handler = ahc_linux_dev_reset,
787 	.eh_bus_reset_handler	= ahc_linux_bus_reset,
788 #if defined(__i386__)
789 	.bios_param		= ahc_linux_biosparam,
790 #endif
791 	.can_queue		= AHC_MAX_QUEUE,
792 	.this_id		= -1,
793 	.cmd_per_lun		= 2,
794 	.use_clustering		= ENABLE_CLUSTERING,
795 	.slave_alloc		= ahc_linux_slave_alloc,
796 	.slave_configure	= ahc_linux_slave_configure,
797 	.slave_destroy		= ahc_linux_slave_destroy,
798 	.target_alloc		= ahc_linux_target_alloc,
799 	.target_destroy		= ahc_linux_target_destroy,
800 };
801 
802 /**************************** Tasklet Handler *********************************/
803 
804 /******************************** Macros **************************************/
805 #define BUILD_SCSIID(ahc, cmd)						    \
806 	((((cmd)->device->id << TID_SHIFT) & TID)			    \
807 	| (((cmd)->device->channel == 0) ? (ahc)->our_id : (ahc)->our_id_b) \
808 	| (((cmd)->device->channel == 0) ? 0 : TWIN_CHNLB))
809 
810 /******************************** Bus DMA *************************************/
811 int
812 ahc_dma_tag_create(struct ahc_softc *ahc, bus_dma_tag_t parent,
813 		   bus_size_t alignment, bus_size_t boundary,
814 		   dma_addr_t lowaddr, dma_addr_t highaddr,
815 		   bus_dma_filter_t *filter, void *filterarg,
816 		   bus_size_t maxsize, int nsegments,
817 		   bus_size_t maxsegsz, int flags, bus_dma_tag_t *ret_tag)
818 {
819 	bus_dma_tag_t dmat;
820 
821 	dmat = malloc(sizeof(*dmat), M_DEVBUF, M_NOWAIT);
822 	if (dmat == NULL)
823 		return (ENOMEM);
824 
825 	/*
826 	 * Linux is very simplistic about DMA memory.  For now don't
827 	 * maintain all specification information.  Once Linux supplies
828 	 * better facilities for doing these operations, or the
829 	 * needs of this particular driver change, we might need to do
830 	 * more here.
831 	 */
832 	dmat->alignment = alignment;
833 	dmat->boundary = boundary;
834 	dmat->maxsize = maxsize;
835 	*ret_tag = dmat;
836 	return (0);
837 }
838 
839 void
840 ahc_dma_tag_destroy(struct ahc_softc *ahc, bus_dma_tag_t dmat)
841 {
842 	free(dmat, M_DEVBUF);
843 }
844 
845 int
846 ahc_dmamem_alloc(struct ahc_softc *ahc, bus_dma_tag_t dmat, void** vaddr,
847 		 int flags, bus_dmamap_t *mapp)
848 {
849 	*vaddr = pci_alloc_consistent(ahc->dev_softc,
850 				      dmat->maxsize, mapp);
851 	if (*vaddr == NULL)
852 		return ENOMEM;
853 	return 0;
854 }
855 
856 void
857 ahc_dmamem_free(struct ahc_softc *ahc, bus_dma_tag_t dmat,
858 		void* vaddr, bus_dmamap_t map)
859 {
860 	pci_free_consistent(ahc->dev_softc, dmat->maxsize,
861 			    vaddr, map);
862 }
863 
864 int
865 ahc_dmamap_load(struct ahc_softc *ahc, bus_dma_tag_t dmat, bus_dmamap_t map,
866 		void *buf, bus_size_t buflen, bus_dmamap_callback_t *cb,
867 		void *cb_arg, int flags)
868 {
869 	/*
870 	 * Assume for now that this will only be used during
871 	 * initialization and not for per-transaction buffer mapping.
872 	 */
873 	bus_dma_segment_t stack_sg;
874 
875 	stack_sg.ds_addr = map;
876 	stack_sg.ds_len = dmat->maxsize;
877 	cb(cb_arg, &stack_sg, /*nseg*/1, /*error*/0);
878 	return (0);
879 }
880 
881 void
882 ahc_dmamap_destroy(struct ahc_softc *ahc, bus_dma_tag_t dmat, bus_dmamap_t map)
883 {
884 }
885 
886 int
887 ahc_dmamap_unload(struct ahc_softc *ahc, bus_dma_tag_t dmat, bus_dmamap_t map)
888 {
889 	/* Nothing to do */
890 	return (0);
891 }
892 
893 static void
894 ahc_linux_setup_tag_info_global(char *p)
895 {
896 	int tags, i, j;
897 
898 	tags = simple_strtoul(p + 1, NULL, 0) & 0xff;
899 	printf("Setting Global Tags= %d\n", tags);
900 
901 	for (i = 0; i < NUM_ELEMENTS(aic7xxx_tag_info); i++) {
902 		for (j = 0; j < AHC_NUM_TARGETS; j++) {
903 			aic7xxx_tag_info[i].tag_commands[j] = tags;
904 		}
905 	}
906 }
907 
908 static void
909 ahc_linux_setup_tag_info(u_long arg, int instance, int targ, int32_t value)
910 {
911 
912 	if ((instance >= 0) && (targ >= 0)
913 	 && (instance < NUM_ELEMENTS(aic7xxx_tag_info))
914 	 && (targ < AHC_NUM_TARGETS)) {
915 		aic7xxx_tag_info[instance].tag_commands[targ] = value & 0xff;
916 		if (bootverbose)
917 			printf("tag_info[%d:%d] = %d\n", instance, targ, value);
918 	}
919 }
920 
921 /*
922  * Handle Linux boot parameters. This routine allows for assigning a value
923  * to a parameter with a ':' between the parameter and the value.
924  * ie. aic7xxx=stpwlev:1,extended
925  */
926 static int
927 aic7xxx_setup(char *s)
928 {
929 	int	i, n;
930 	char   *p;
931 	char   *end;
932 
933 	static struct {
934 		const char *name;
935 		uint32_t *flag;
936 	} options[] = {
937 		{ "extended", &aic7xxx_extended },
938 		{ "no_reset", &aic7xxx_no_reset },
939 		{ "verbose", &aic7xxx_verbose },
940 		{ "allow_memio", &aic7xxx_allow_memio},
941 #ifdef AHC_DEBUG
942 		{ "debug", &ahc_debug },
943 #endif
944 		{ "periodic_otag", &aic7xxx_periodic_otag },
945 		{ "pci_parity", &aic7xxx_pci_parity },
946 		{ "seltime", &aic7xxx_seltime },
947 		{ "tag_info", NULL },
948 		{ "global_tag_depth", NULL },
949 		{ "dv", NULL }
950 	};
951 
952 	end = strchr(s, '\0');
953 
954 	/*
955 	 * XXX ia64 gcc isn't smart enough to know that NUM_ELEMENTS
956 	 * will never be 0 in this case.
957 	 */
958 	n = 0;
959 
960 	while ((p = strsep(&s, ",.")) != NULL) {
961 		if (*p == '\0')
962 			continue;
963 		for (i = 0; i < NUM_ELEMENTS(options); i++) {
964 
965 			n = strlen(options[i].name);
966 			if (strncmp(options[i].name, p, n) == 0)
967 				break;
968 		}
969 		if (i == NUM_ELEMENTS(options))
970 			continue;
971 
972 		if (strncmp(p, "global_tag_depth", n) == 0) {
973 			ahc_linux_setup_tag_info_global(p + n);
974 		} else if (strncmp(p, "tag_info", n) == 0) {
975 			s = aic_parse_brace_option("tag_info", p + n, end,
976 			    2, ahc_linux_setup_tag_info, 0);
977 		} else if (p[n] == ':') {
978 			*(options[i].flag) = simple_strtoul(p + n + 1, NULL, 0);
979 		} else if (strncmp(p, "verbose", n) == 0) {
980 			*(options[i].flag) = 1;
981 		} else {
982 			*(options[i].flag) ^= 0xFFFFFFFF;
983 		}
984 	}
985 	return 1;
986 }
987 
988 __setup("aic7xxx=", aic7xxx_setup);
989 
990 uint32_t aic7xxx_verbose;
991 
992 int
993 ahc_linux_register_host(struct ahc_softc *ahc, struct scsi_host_template *template)
994 {
995 	char	 buf[80];
996 	struct	 Scsi_Host *host;
997 	char	*new_name;
998 	u_long	 s;
999 
1000 	template->name = ahc->description;
1001 	host = scsi_host_alloc(template, sizeof(struct ahc_softc *));
1002 	if (host == NULL)
1003 		return (ENOMEM);
1004 
1005 	*((struct ahc_softc **)host->hostdata) = ahc;
1006 	ahc_lock(ahc, &s);
1007 	scsi_assign_lock(host, &ahc->platform_data->spin_lock);
1008 	ahc->platform_data->host = host;
1009 	host->can_queue = AHC_MAX_QUEUE;
1010 	host->cmd_per_lun = 2;
1011 	/* XXX No way to communicate the ID for multiple channels */
1012 	host->this_id = ahc->our_id;
1013 	host->irq = ahc->platform_data->irq;
1014 	host->max_id = (ahc->features & AHC_WIDE) ? 16 : 8;
1015 	host->max_lun = AHC_NUM_LUNS;
1016 	host->max_channel = (ahc->features & AHC_TWIN) ? 1 : 0;
1017 	host->sg_tablesize = AHC_NSEG;
1018 	ahc_set_unit(ahc, ahc_linux_unit++);
1019 	sprintf(buf, "scsi%d", host->host_no);
1020 	new_name = malloc(strlen(buf) + 1, M_DEVBUF, M_NOWAIT);
1021 	if (new_name != NULL) {
1022 		strcpy(new_name, buf);
1023 		ahc_set_name(ahc, new_name);
1024 	}
1025 	host->unique_id = ahc->unit;
1026 	ahc_linux_initialize_scsi_bus(ahc);
1027 	ahc_intr_enable(ahc, TRUE);
1028 	ahc_unlock(ahc, &s);
1029 
1030 	host->transportt = ahc_linux_transport_template;
1031 
1032 	scsi_add_host(host, (ahc->dev_softc ? &ahc->dev_softc->dev : NULL)); /* XXX handle failure */
1033 	scsi_scan_host(host);
1034 	return (0);
1035 }
1036 
1037 uint64_t
1038 ahc_linux_get_memsize(void)
1039 {
1040 	struct sysinfo si;
1041 
1042 	si_meminfo(&si);
1043 	return ((uint64_t)si.totalram << PAGE_SHIFT);
1044 }
1045 
1046 /*
1047  * Place the SCSI bus into a known state by either resetting it,
1048  * or forcing transfer negotiations on the next command to any
1049  * target.
1050  */
1051 void
1052 ahc_linux_initialize_scsi_bus(struct ahc_softc *ahc)
1053 {
1054 	int i;
1055 	int numtarg;
1056 
1057 	i = 0;
1058 	numtarg = 0;
1059 
1060 	if (aic7xxx_no_reset != 0)
1061 		ahc->flags &= ~(AHC_RESET_BUS_A|AHC_RESET_BUS_B);
1062 
1063 	if ((ahc->flags & AHC_RESET_BUS_A) != 0)
1064 		ahc_reset_channel(ahc, 'A', /*initiate_reset*/TRUE);
1065 	else
1066 		numtarg = (ahc->features & AHC_WIDE) ? 16 : 8;
1067 
1068 	if ((ahc->features & AHC_TWIN) != 0) {
1069 
1070 		if ((ahc->flags & AHC_RESET_BUS_B) != 0) {
1071 			ahc_reset_channel(ahc, 'B', /*initiate_reset*/TRUE);
1072 		} else {
1073 			if (numtarg == 0)
1074 				i = 8;
1075 			numtarg += 8;
1076 		}
1077 	}
1078 
1079 	/*
1080 	 * Force negotiation to async for all targets that
1081 	 * will not see an initial bus reset.
1082 	 */
1083 	for (; i < numtarg; i++) {
1084 		struct ahc_devinfo devinfo;
1085 		struct ahc_initiator_tinfo *tinfo;
1086 		struct ahc_tmode_tstate *tstate;
1087 		u_int our_id;
1088 		u_int target_id;
1089 		char channel;
1090 
1091 		channel = 'A';
1092 		our_id = ahc->our_id;
1093 		target_id = i;
1094 		if (i > 7 && (ahc->features & AHC_TWIN) != 0) {
1095 			channel = 'B';
1096 			our_id = ahc->our_id_b;
1097 			target_id = i % 8;
1098 		}
1099 		tinfo = ahc_fetch_transinfo(ahc, channel, our_id,
1100 					    target_id, &tstate);
1101 		ahc_compile_devinfo(&devinfo, our_id, target_id,
1102 				    CAM_LUN_WILDCARD, channel, ROLE_INITIATOR);
1103 		ahc_update_neg_request(ahc, &devinfo, tstate,
1104 				       tinfo, AHC_NEG_ALWAYS);
1105 	}
1106 	/* Give the bus some time to recover */
1107 	if ((ahc->flags & (AHC_RESET_BUS_A|AHC_RESET_BUS_B)) != 0) {
1108 		ahc_linux_freeze_simq(ahc);
1109 		init_timer(&ahc->platform_data->reset_timer);
1110 		ahc->platform_data->reset_timer.data = (u_long)ahc;
1111 		ahc->platform_data->reset_timer.expires =
1112 		    jiffies + (AIC7XXX_RESET_DELAY * HZ)/1000;
1113 		ahc->platform_data->reset_timer.function =
1114 		    ahc_linux_release_simq;
1115 		add_timer(&ahc->platform_data->reset_timer);
1116 	}
1117 }
1118 
1119 int
1120 ahc_platform_alloc(struct ahc_softc *ahc, void *platform_arg)
1121 {
1122 
1123 	ahc->platform_data =
1124 	    malloc(sizeof(struct ahc_platform_data), M_DEVBUF, M_NOWAIT);
1125 	if (ahc->platform_data == NULL)
1126 		return (ENOMEM);
1127 	memset(ahc->platform_data, 0, sizeof(struct ahc_platform_data));
1128 	ahc->platform_data->irq = AHC_LINUX_NOIRQ;
1129 	ahc_lockinit(ahc);
1130 	init_MUTEX_LOCKED(&ahc->platform_data->eh_sem);
1131 	ahc->seltime = (aic7xxx_seltime & 0x3) << 4;
1132 	ahc->seltime_b = (aic7xxx_seltime & 0x3) << 4;
1133 	if (aic7xxx_pci_parity == 0)
1134 		ahc->flags |= AHC_DISABLE_PCI_PERR;
1135 
1136 	return (0);
1137 }
1138 
1139 void
1140 ahc_platform_free(struct ahc_softc *ahc)
1141 {
1142 	struct scsi_target *starget;
1143 	int i, j;
1144 
1145 	if (ahc->platform_data != NULL) {
1146 		if (ahc->platform_data->host != NULL) {
1147 			scsi_remove_host(ahc->platform_data->host);
1148 			scsi_host_put(ahc->platform_data->host);
1149 		}
1150 
1151 		/* destroy all of the device and target objects */
1152 		for (i = 0; i < AHC_NUM_TARGETS; i++) {
1153 			starget = ahc->platform_data->starget[i];
1154 			if (starget != NULL) {
1155 				for (j = 0; j < AHC_NUM_LUNS; j++) {
1156 					struct ahc_linux_target *targ =
1157 						scsi_transport_target_data(starget);
1158 
1159 					if (targ->sdev[j] == NULL)
1160 						continue;
1161 					targ->sdev[j] = NULL;
1162 				}
1163 				ahc->platform_data->starget[i] = NULL;
1164  			}
1165  		}
1166 
1167 		if (ahc->platform_data->irq != AHC_LINUX_NOIRQ)
1168 			free_irq(ahc->platform_data->irq, ahc);
1169 		if (ahc->tag == BUS_SPACE_PIO
1170 		 && ahc->bsh.ioport != 0)
1171 			release_region(ahc->bsh.ioport, 256);
1172 		if (ahc->tag == BUS_SPACE_MEMIO
1173 		 && ahc->bsh.maddr != NULL) {
1174 			iounmap(ahc->bsh.maddr);
1175 			release_mem_region(ahc->platform_data->mem_busaddr,
1176 					   0x1000);
1177 		}
1178 
1179 		free(ahc->platform_data, M_DEVBUF);
1180 	}
1181 }
1182 
1183 void
1184 ahc_platform_freeze_devq(struct ahc_softc *ahc, struct scb *scb)
1185 {
1186 	ahc_platform_abort_scbs(ahc, SCB_GET_TARGET(ahc, scb),
1187 				SCB_GET_CHANNEL(ahc, scb),
1188 				SCB_GET_LUN(scb), SCB_LIST_NULL,
1189 				ROLE_UNKNOWN, CAM_REQUEUE_REQ);
1190 }
1191 
1192 void
1193 ahc_platform_set_tags(struct ahc_softc *ahc, struct ahc_devinfo *devinfo,
1194 		      ahc_queue_alg alg)
1195 {
1196 	struct scsi_target *starget;
1197 	struct ahc_linux_target *targ;
1198 	struct ahc_linux_device *dev;
1199 	struct scsi_device *sdev;
1200 	u_int target_offset;
1201 	int was_queuing;
1202 	int now_queuing;
1203 
1204 	target_offset = devinfo->target;
1205 	if (devinfo->channel != 'A')
1206 		target_offset += 8;
1207 	starget = ahc->platform_data->starget[target_offset];
1208 	targ = scsi_transport_target_data(starget);
1209 	BUG_ON(targ == NULL);
1210 	sdev = targ->sdev[devinfo->lun];
1211 	if (sdev == NULL)
1212 		return;
1213 	dev = scsi_transport_device_data(sdev);
1214 
1215 	was_queuing = dev->flags & (AHC_DEV_Q_BASIC|AHC_DEV_Q_TAGGED);
1216 	switch (alg) {
1217 	default:
1218 	case AHC_QUEUE_NONE:
1219 		now_queuing = 0;
1220 		break;
1221 	case AHC_QUEUE_BASIC:
1222 		now_queuing = AHC_DEV_Q_BASIC;
1223 		break;
1224 	case AHC_QUEUE_TAGGED:
1225 		now_queuing = AHC_DEV_Q_TAGGED;
1226 		break;
1227 	}
1228 	if ((dev->flags & AHC_DEV_FREEZE_TIL_EMPTY) == 0
1229 	 && (was_queuing != now_queuing)
1230 	 && (dev->active != 0)) {
1231 		dev->flags |= AHC_DEV_FREEZE_TIL_EMPTY;
1232 		dev->qfrozen++;
1233 	}
1234 
1235 	dev->flags &= ~(AHC_DEV_Q_BASIC|AHC_DEV_Q_TAGGED|AHC_DEV_PERIODIC_OTAG);
1236 	if (now_queuing) {
1237 		u_int usertags;
1238 
1239 		usertags = ahc_linux_user_tagdepth(ahc, devinfo);
1240 		if (!was_queuing) {
1241 			/*
1242 			 * Start out agressively and allow our
1243 			 * dynamic queue depth algorithm to take
1244 			 * care of the rest.
1245 			 */
1246 			dev->maxtags = usertags;
1247 			dev->openings = dev->maxtags - dev->active;
1248 		}
1249 		if (dev->maxtags == 0) {
1250 			/*
1251 			 * Queueing is disabled by the user.
1252 			 */
1253 			dev->openings = 1;
1254 		} else if (alg == AHC_QUEUE_TAGGED) {
1255 			dev->flags |= AHC_DEV_Q_TAGGED;
1256 			if (aic7xxx_periodic_otag != 0)
1257 				dev->flags |= AHC_DEV_PERIODIC_OTAG;
1258 		} else
1259 			dev->flags |= AHC_DEV_Q_BASIC;
1260 	} else {
1261 		/* We can only have one opening. */
1262 		dev->maxtags = 0;
1263 		dev->openings =  1 - dev->active;
1264 	}
1265 	switch ((dev->flags & (AHC_DEV_Q_BASIC|AHC_DEV_Q_TAGGED))) {
1266 	case AHC_DEV_Q_BASIC:
1267 		scsi_set_tag_type(sdev, MSG_SIMPLE_TAG);
1268 		scsi_activate_tcq(sdev, dev->openings + dev->active);
1269 		break;
1270 	case AHC_DEV_Q_TAGGED:
1271 		scsi_set_tag_type(sdev, MSG_ORDERED_TAG);
1272 		scsi_activate_tcq(sdev, dev->openings + dev->active);
1273 		break;
1274 	default:
1275 		/*
1276 		 * We allow the OS to queue 2 untagged transactions to
1277 		 * us at any time even though we can only execute them
1278 		 * serially on the controller/device.  This should
1279 		 * remove some latency.
1280 		 */
1281 		scsi_deactivate_tcq(sdev, 2);
1282 		break;
1283 	}
1284 }
1285 
1286 int
1287 ahc_platform_abort_scbs(struct ahc_softc *ahc, int target, char channel,
1288 			int lun, u_int tag, role_t role, uint32_t status)
1289 {
1290 	return 0;
1291 }
1292 
1293 static u_int
1294 ahc_linux_user_tagdepth(struct ahc_softc *ahc, struct ahc_devinfo *devinfo)
1295 {
1296 	static int warned_user;
1297 	u_int tags;
1298 
1299 	tags = 0;
1300 	if ((ahc->user_discenable & devinfo->target_mask) != 0) {
1301 		if (ahc->unit >= NUM_ELEMENTS(aic7xxx_tag_info)) {
1302 			if (warned_user == 0) {
1303 
1304 				printf(KERN_WARNING
1305 "aic7xxx: WARNING: Insufficient tag_info instances\n"
1306 "aic7xxx: for installed controllers. Using defaults\n"
1307 "aic7xxx: Please update the aic7xxx_tag_info array in\n"
1308 "aic7xxx: the aic7xxx_osm..c source file.\n");
1309 				warned_user++;
1310 			}
1311 			tags = AHC_MAX_QUEUE;
1312 		} else {
1313 			adapter_tag_info_t *tag_info;
1314 
1315 			tag_info = &aic7xxx_tag_info[ahc->unit];
1316 			tags = tag_info->tag_commands[devinfo->target_offset];
1317 			if (tags > AHC_MAX_QUEUE)
1318 				tags = AHC_MAX_QUEUE;
1319 		}
1320 	}
1321 	return (tags);
1322 }
1323 
1324 /*
1325  * Determines the queue depth for a given device.
1326  */
1327 static void
1328 ahc_linux_device_queue_depth(struct scsi_device *sdev)
1329 {
1330 	struct	ahc_devinfo devinfo;
1331 	u_int	tags;
1332 	struct ahc_softc *ahc = *((struct ahc_softc **)sdev->host->hostdata);
1333 
1334 	ahc_compile_devinfo(&devinfo,
1335 			    sdev->sdev_target->channel == 0
1336 			  ? ahc->our_id : ahc->our_id_b,
1337 			    sdev->sdev_target->id, sdev->lun,
1338 			    sdev->sdev_target->channel == 0 ? 'A' : 'B',
1339 			    ROLE_INITIATOR);
1340 	tags = ahc_linux_user_tagdepth(ahc, &devinfo);
1341 	if (tags != 0 && sdev->tagged_supported != 0) {
1342 
1343 		ahc_set_tags(ahc, &devinfo, AHC_QUEUE_TAGGED);
1344 		ahc_print_devinfo(ahc, &devinfo);
1345 		printf("Tagged Queuing enabled.  Depth %d\n", tags);
1346 	} else {
1347 		ahc_set_tags(ahc, &devinfo, AHC_QUEUE_NONE);
1348 	}
1349 }
1350 
1351 static int
1352 ahc_linux_run_command(struct ahc_softc *ahc, struct ahc_linux_device *dev,
1353 		      struct scsi_cmnd *cmd)
1354 {
1355 	struct	 scb *scb;
1356 	struct	 hardware_scb *hscb;
1357 	struct	 ahc_initiator_tinfo *tinfo;
1358 	struct	 ahc_tmode_tstate *tstate;
1359 	uint16_t mask;
1360 	struct scb_tailq *untagged_q = NULL;
1361 
1362 	/*
1363 	 * Schedule us to run later.  The only reason we are not
1364 	 * running is because the whole controller Q is frozen.
1365 	 */
1366 	if (ahc->platform_data->qfrozen != 0)
1367 		return SCSI_MLQUEUE_HOST_BUSY;
1368 
1369 	/*
1370 	 * We only allow one untagged transaction
1371 	 * per target in the initiator role unless
1372 	 * we are storing a full busy target *lun*
1373 	 * table in SCB space.
1374 	 */
1375 	if (!blk_rq_tagged(cmd->request)
1376 	    && (ahc->features & AHC_SCB_BTT) == 0) {
1377 		int target_offset;
1378 
1379 		target_offset = cmd->device->id + cmd->device->channel * 8;
1380 		untagged_q = &(ahc->untagged_queues[target_offset]);
1381 		if (!TAILQ_EMPTY(untagged_q))
1382 			/* if we're already executing an untagged command
1383 			 * we're busy to another */
1384 			return SCSI_MLQUEUE_DEVICE_BUSY;
1385 	}
1386 
1387 	/*
1388 	 * Get an scb to use.
1389 	 */
1390 	scb = ahc_get_scb(ahc);
1391 	if (!scb)
1392 		return SCSI_MLQUEUE_HOST_BUSY;
1393 
1394 	scb->io_ctx = cmd;
1395 	scb->platform_data->dev = dev;
1396 	hscb = scb->hscb;
1397 	cmd->host_scribble = (char *)scb;
1398 
1399 	/*
1400 	 * Fill out basics of the HSCB.
1401 	 */
1402 	hscb->control = 0;
1403 	hscb->scsiid = BUILD_SCSIID(ahc, cmd);
1404 	hscb->lun = cmd->device->lun;
1405 	mask = SCB_GET_TARGET_MASK(ahc, scb);
1406 	tinfo = ahc_fetch_transinfo(ahc, SCB_GET_CHANNEL(ahc, scb),
1407 				    SCB_GET_OUR_ID(scb),
1408 				    SCB_GET_TARGET(ahc, scb), &tstate);
1409 	hscb->scsirate = tinfo->scsirate;
1410 	hscb->scsioffset = tinfo->curr.offset;
1411 	if ((tstate->ultraenb & mask) != 0)
1412 		hscb->control |= ULTRAENB;
1413 
1414 	if ((ahc->user_discenable & mask) != 0)
1415 		hscb->control |= DISCENB;
1416 
1417 	if ((tstate->auto_negotiate & mask) != 0) {
1418 		scb->flags |= SCB_AUTO_NEGOTIATE;
1419 		scb->hscb->control |= MK_MESSAGE;
1420 	}
1421 
1422 	if ((dev->flags & (AHC_DEV_Q_TAGGED|AHC_DEV_Q_BASIC)) != 0) {
1423 		int	msg_bytes;
1424 		uint8_t tag_msgs[2];
1425 
1426 		msg_bytes = scsi_populate_tag_msg(cmd, tag_msgs);
1427 		if (msg_bytes && tag_msgs[0] != MSG_SIMPLE_TASK) {
1428 			hscb->control |= tag_msgs[0];
1429 			if (tag_msgs[0] == MSG_ORDERED_TASK)
1430 				dev->commands_since_idle_or_otag = 0;
1431 		} else if (dev->commands_since_idle_or_otag == AHC_OTAG_THRESH
1432 				&& (dev->flags & AHC_DEV_Q_TAGGED) != 0) {
1433 			hscb->control |= MSG_ORDERED_TASK;
1434 			dev->commands_since_idle_or_otag = 0;
1435 		} else {
1436 			hscb->control |= MSG_SIMPLE_TASK;
1437 		}
1438 	}
1439 
1440 	hscb->cdb_len = cmd->cmd_len;
1441 	if (hscb->cdb_len <= 12) {
1442 		memcpy(hscb->shared_data.cdb, cmd->cmnd, hscb->cdb_len);
1443 	} else {
1444 		memcpy(hscb->cdb32, cmd->cmnd, hscb->cdb_len);
1445 		scb->flags |= SCB_CDB32_PTR;
1446 	}
1447 
1448 	scb->platform_data->xfer_len = 0;
1449 	ahc_set_residual(scb, 0);
1450 	ahc_set_sense_residual(scb, 0);
1451 	scb->sg_count = 0;
1452 	if (cmd->use_sg != 0) {
1453 		struct	ahc_dma_seg *sg;
1454 		struct	scatterlist *cur_seg;
1455 		struct	scatterlist *end_seg;
1456 		int	nseg;
1457 
1458 		cur_seg = (struct scatterlist *)cmd->request_buffer;
1459 		nseg = pci_map_sg(ahc->dev_softc, cur_seg, cmd->use_sg,
1460 				  cmd->sc_data_direction);
1461 		end_seg = cur_seg + nseg;
1462 		/* Copy the segments into the SG list. */
1463 		sg = scb->sg_list;
1464 		/*
1465 		 * The sg_count may be larger than nseg if
1466 		 * a transfer crosses a 32bit page.
1467 		 */
1468 		while (cur_seg < end_seg) {
1469 			dma_addr_t addr;
1470 			bus_size_t len;
1471 			int consumed;
1472 
1473 			addr = sg_dma_address(cur_seg);
1474 			len = sg_dma_len(cur_seg);
1475 			consumed = ahc_linux_map_seg(ahc, scb,
1476 						     sg, addr, len);
1477 			sg += consumed;
1478 			scb->sg_count += consumed;
1479 			cur_seg++;
1480 		}
1481 		sg--;
1482 		sg->len |= ahc_htole32(AHC_DMA_LAST_SEG);
1483 
1484 		/*
1485 		 * Reset the sg list pointer.
1486 		 */
1487 		scb->hscb->sgptr =
1488 			ahc_htole32(scb->sg_list_phys | SG_FULL_RESID);
1489 
1490 		/*
1491 		 * Copy the first SG into the "current"
1492 		 * data pointer area.
1493 		 */
1494 		scb->hscb->dataptr = scb->sg_list->addr;
1495 		scb->hscb->datacnt = scb->sg_list->len;
1496 	} else if (cmd->request_bufflen != 0) {
1497 		struct	 ahc_dma_seg *sg;
1498 		dma_addr_t addr;
1499 
1500 		sg = scb->sg_list;
1501 		addr = pci_map_single(ahc->dev_softc,
1502 				      cmd->request_buffer,
1503 				      cmd->request_bufflen,
1504 				      cmd->sc_data_direction);
1505 		scb->platform_data->buf_busaddr = addr;
1506 		scb->sg_count = ahc_linux_map_seg(ahc, scb,
1507 						  sg, addr,
1508 						  cmd->request_bufflen);
1509 		sg->len |= ahc_htole32(AHC_DMA_LAST_SEG);
1510 
1511 		/*
1512 		 * Reset the sg list pointer.
1513 		 */
1514 		scb->hscb->sgptr =
1515 			ahc_htole32(scb->sg_list_phys | SG_FULL_RESID);
1516 
1517 		/*
1518 		 * Copy the first SG into the "current"
1519 		 * data pointer area.
1520 		 */
1521 		scb->hscb->dataptr = sg->addr;
1522 		scb->hscb->datacnt = sg->len;
1523 	} else {
1524 		scb->hscb->sgptr = ahc_htole32(SG_LIST_NULL);
1525 		scb->hscb->dataptr = 0;
1526 		scb->hscb->datacnt = 0;
1527 		scb->sg_count = 0;
1528 	}
1529 
1530 	LIST_INSERT_HEAD(&ahc->pending_scbs, scb, pending_links);
1531 	dev->openings--;
1532 	dev->active++;
1533 	dev->commands_issued++;
1534 	if ((dev->flags & AHC_DEV_PERIODIC_OTAG) != 0)
1535 		dev->commands_since_idle_or_otag++;
1536 
1537 	scb->flags |= SCB_ACTIVE;
1538 	if (untagged_q) {
1539 		TAILQ_INSERT_TAIL(untagged_q, scb, links.tqe);
1540 		scb->flags |= SCB_UNTAGGEDQ;
1541 	}
1542 	ahc_queue_scb(ahc, scb);
1543 	return 0;
1544 }
1545 
1546 /*
1547  * SCSI controller interrupt handler.
1548  */
1549 irqreturn_t
1550 ahc_linux_isr(int irq, void *dev_id, struct pt_regs * regs)
1551 {
1552 	struct	ahc_softc *ahc;
1553 	u_long	flags;
1554 	int	ours;
1555 
1556 	ahc = (struct ahc_softc *) dev_id;
1557 	ahc_lock(ahc, &flags);
1558 	ours = ahc_intr(ahc);
1559 	ahc_unlock(ahc, &flags);
1560 	return IRQ_RETVAL(ours);
1561 }
1562 
1563 void
1564 ahc_platform_flushwork(struct ahc_softc *ahc)
1565 {
1566 
1567 }
1568 
1569 void
1570 ahc_send_async(struct ahc_softc *ahc, char channel,
1571 	       u_int target, u_int lun, ac_code code, void *arg)
1572 {
1573 	switch (code) {
1574 	case AC_TRANSFER_NEG:
1575 	{
1576 		char	buf[80];
1577 		struct	scsi_target *starget;
1578 		struct	ahc_linux_target *targ;
1579 		struct	info_str info;
1580 		struct	ahc_initiator_tinfo *tinfo;
1581 		struct	ahc_tmode_tstate *tstate;
1582 		int	target_offset;
1583 		unsigned int target_ppr_options;
1584 
1585 		BUG_ON(target == CAM_TARGET_WILDCARD);
1586 
1587 		info.buffer = buf;
1588 		info.length = sizeof(buf);
1589 		info.offset = 0;
1590 		info.pos = 0;
1591 		tinfo = ahc_fetch_transinfo(ahc, channel,
1592 						channel == 'A' ? ahc->our_id
1593 							       : ahc->our_id_b,
1594 						target, &tstate);
1595 
1596 		/*
1597 		 * Don't bother reporting results while
1598 		 * negotiations are still pending.
1599 		 */
1600 		if (tinfo->curr.period != tinfo->goal.period
1601 		 || tinfo->curr.width != tinfo->goal.width
1602 		 || tinfo->curr.offset != tinfo->goal.offset
1603 		 || tinfo->curr.ppr_options != tinfo->goal.ppr_options)
1604 			if (bootverbose == 0)
1605 				break;
1606 
1607 		/*
1608 		 * Don't bother reporting results that
1609 		 * are identical to those last reported.
1610 		 */
1611 		target_offset = target;
1612 		if (channel == 'B')
1613 			target_offset += 8;
1614 		starget = ahc->platform_data->starget[target_offset];
1615 		targ = scsi_transport_target_data(starget);
1616 		if (targ == NULL)
1617 			break;
1618 
1619 		target_ppr_options =
1620 			(spi_dt(starget) ? MSG_EXT_PPR_DT_REQ : 0)
1621 			+ (spi_qas(starget) ? MSG_EXT_PPR_QAS_REQ : 0)
1622 			+ (spi_iu(starget) ?  MSG_EXT_PPR_IU_REQ : 0);
1623 
1624 		if (tinfo->curr.period == spi_period(starget)
1625 		    && tinfo->curr.width == spi_width(starget)
1626 		    && tinfo->curr.offset == spi_offset(starget)
1627 		 && tinfo->curr.ppr_options == target_ppr_options)
1628 			if (bootverbose == 0)
1629 				break;
1630 
1631 		spi_period(starget) = tinfo->curr.period;
1632 		spi_width(starget) = tinfo->curr.width;
1633 		spi_offset(starget) = tinfo->curr.offset;
1634 		spi_dt(starget) = tinfo->curr.ppr_options & MSG_EXT_PPR_DT_REQ ? 1 : 0;
1635 		spi_qas(starget) = tinfo->curr.ppr_options & MSG_EXT_PPR_QAS_REQ ? 1 : 0;
1636 		spi_iu(starget) = tinfo->curr.ppr_options & MSG_EXT_PPR_IU_REQ ? 1 : 0;
1637 		spi_display_xfer_agreement(starget);
1638 		break;
1639 	}
1640         case AC_SENT_BDR:
1641 	{
1642 		WARN_ON(lun != CAM_LUN_WILDCARD);
1643 		scsi_report_device_reset(ahc->platform_data->host,
1644 					 channel - 'A', target);
1645 		break;
1646 	}
1647         case AC_BUS_RESET:
1648 		if (ahc->platform_data->host != NULL) {
1649 			scsi_report_bus_reset(ahc->platform_data->host,
1650 					      channel - 'A');
1651 		}
1652                 break;
1653         default:
1654                 panic("ahc_send_async: Unexpected async event");
1655         }
1656 }
1657 
1658 /*
1659  * Calls the higher level scsi done function and frees the scb.
1660  */
1661 void
1662 ahc_done(struct ahc_softc *ahc, struct scb *scb)
1663 {
1664 	struct scsi_cmnd *cmd;
1665 	struct	   ahc_linux_device *dev;
1666 
1667 	LIST_REMOVE(scb, pending_links);
1668 	if ((scb->flags & SCB_UNTAGGEDQ) != 0) {
1669 		struct scb_tailq *untagged_q;
1670 		int target_offset;
1671 
1672 		target_offset = SCB_GET_TARGET_OFFSET(ahc, scb);
1673 		untagged_q = &(ahc->untagged_queues[target_offset]);
1674 		TAILQ_REMOVE(untagged_q, scb, links.tqe);
1675 		BUG_ON(!TAILQ_EMPTY(untagged_q));
1676 	}
1677 
1678 	if ((scb->flags & SCB_ACTIVE) == 0) {
1679 		printf("SCB %d done'd twice\n", scb->hscb->tag);
1680 		ahc_dump_card_state(ahc);
1681 		panic("Stopping for safety");
1682 	}
1683 	cmd = scb->io_ctx;
1684 	dev = scb->platform_data->dev;
1685 	dev->active--;
1686 	dev->openings++;
1687 	if ((cmd->result & (CAM_DEV_QFRZN << 16)) != 0) {
1688 		cmd->result &= ~(CAM_DEV_QFRZN << 16);
1689 		dev->qfrozen--;
1690 	}
1691 	ahc_linux_unmap_scb(ahc, scb);
1692 
1693 	/*
1694 	 * Guard against stale sense data.
1695 	 * The Linux mid-layer assumes that sense
1696 	 * was retrieved anytime the first byte of
1697 	 * the sense buffer looks "sane".
1698 	 */
1699 	cmd->sense_buffer[0] = 0;
1700 	if (ahc_get_transaction_status(scb) == CAM_REQ_INPROG) {
1701 		uint32_t amount_xferred;
1702 
1703 		amount_xferred =
1704 		    ahc_get_transfer_length(scb) - ahc_get_residual(scb);
1705 		if ((scb->flags & SCB_TRANSMISSION_ERROR) != 0) {
1706 #ifdef AHC_DEBUG
1707 			if ((ahc_debug & AHC_SHOW_MISC) != 0) {
1708 				ahc_print_path(ahc, scb);
1709 				printf("Set CAM_UNCOR_PARITY\n");
1710 			}
1711 #endif
1712 			ahc_set_transaction_status(scb, CAM_UNCOR_PARITY);
1713 #ifdef AHC_REPORT_UNDERFLOWS
1714 		/*
1715 		 * This code is disabled by default as some
1716 		 * clients of the SCSI system do not properly
1717 		 * initialize the underflow parameter.  This
1718 		 * results in spurious termination of commands
1719 		 * that complete as expected (e.g. underflow is
1720 		 * allowed as command can return variable amounts
1721 		 * of data.
1722 		 */
1723 		} else if (amount_xferred < scb->io_ctx->underflow) {
1724 			u_int i;
1725 
1726 			ahc_print_path(ahc, scb);
1727 			printf("CDB:");
1728 			for (i = 0; i < scb->io_ctx->cmd_len; i++)
1729 				printf(" 0x%x", scb->io_ctx->cmnd[i]);
1730 			printf("\n");
1731 			ahc_print_path(ahc, scb);
1732 			printf("Saw underflow (%ld of %ld bytes). "
1733 			       "Treated as error\n",
1734 				ahc_get_residual(scb),
1735 				ahc_get_transfer_length(scb));
1736 			ahc_set_transaction_status(scb, CAM_DATA_RUN_ERR);
1737 #endif
1738 		} else {
1739 			ahc_set_transaction_status(scb, CAM_REQ_CMP);
1740 		}
1741 	} else if (ahc_get_transaction_status(scb) == CAM_SCSI_STATUS_ERROR) {
1742 		ahc_linux_handle_scsi_status(ahc, cmd->device, scb);
1743 	}
1744 
1745 	if (dev->openings == 1
1746 	 && ahc_get_transaction_status(scb) == CAM_REQ_CMP
1747 	 && ahc_get_scsi_status(scb) != SCSI_STATUS_QUEUE_FULL)
1748 		dev->tag_success_count++;
1749 	/*
1750 	 * Some devices deal with temporary internal resource
1751 	 * shortages by returning queue full.  When the queue
1752 	 * full occurrs, we throttle back.  Slowly try to get
1753 	 * back to our previous queue depth.
1754 	 */
1755 	if ((dev->openings + dev->active) < dev->maxtags
1756 	 && dev->tag_success_count > AHC_TAG_SUCCESS_INTERVAL) {
1757 		dev->tag_success_count = 0;
1758 		dev->openings++;
1759 	}
1760 
1761 	if (dev->active == 0)
1762 		dev->commands_since_idle_or_otag = 0;
1763 
1764 	if ((scb->flags & SCB_RECOVERY_SCB) != 0) {
1765 		printf("Recovery SCB completes\n");
1766 		if (ahc_get_transaction_status(scb) == CAM_BDR_SENT
1767 		 || ahc_get_transaction_status(scb) == CAM_REQ_ABORTED)
1768 			ahc_set_transaction_status(scb, CAM_CMD_TIMEOUT);
1769 		if ((ahc->platform_data->flags & AHC_UP_EH_SEMAPHORE) != 0) {
1770 			ahc->platform_data->flags &= ~AHC_UP_EH_SEMAPHORE;
1771 			up(&ahc->platform_data->eh_sem);
1772 		}
1773 	}
1774 
1775 	ahc_free_scb(ahc, scb);
1776 	ahc_linux_queue_cmd_complete(ahc, cmd);
1777 }
1778 
1779 static void
1780 ahc_linux_handle_scsi_status(struct ahc_softc *ahc,
1781 			     struct scsi_device *sdev, struct scb *scb)
1782 {
1783 	struct	ahc_devinfo devinfo;
1784 	struct ahc_linux_device *dev = scsi_transport_device_data(sdev);
1785 
1786 	ahc_compile_devinfo(&devinfo,
1787 			    ahc->our_id,
1788 			    sdev->sdev_target->id, sdev->lun,
1789 			    sdev->sdev_target->channel == 0 ? 'A' : 'B',
1790 			    ROLE_INITIATOR);
1791 
1792 	/*
1793 	 * We don't currently trust the mid-layer to
1794 	 * properly deal with queue full or busy.  So,
1795 	 * when one occurs, we tell the mid-layer to
1796 	 * unconditionally requeue the command to us
1797 	 * so that we can retry it ourselves.  We also
1798 	 * implement our own throttling mechanism so
1799 	 * we don't clobber the device with too many
1800 	 * commands.
1801 	 */
1802 	switch (ahc_get_scsi_status(scb)) {
1803 	default:
1804 		break;
1805 	case SCSI_STATUS_CHECK_COND:
1806 	case SCSI_STATUS_CMD_TERMINATED:
1807 	{
1808 		struct scsi_cmnd *cmd;
1809 
1810 		/*
1811 		 * Copy sense information to the OS's cmd
1812 		 * structure if it is available.
1813 		 */
1814 		cmd = scb->io_ctx;
1815 		if (scb->flags & SCB_SENSE) {
1816 			u_int sense_size;
1817 
1818 			sense_size = MIN(sizeof(struct scsi_sense_data)
1819 				       - ahc_get_sense_residual(scb),
1820 					 sizeof(cmd->sense_buffer));
1821 			memcpy(cmd->sense_buffer,
1822 			       ahc_get_sense_buf(ahc, scb), sense_size);
1823 			if (sense_size < sizeof(cmd->sense_buffer))
1824 				memset(&cmd->sense_buffer[sense_size], 0,
1825 				       sizeof(cmd->sense_buffer) - sense_size);
1826 			cmd->result |= (DRIVER_SENSE << 24);
1827 #ifdef AHC_DEBUG
1828 			if (ahc_debug & AHC_SHOW_SENSE) {
1829 				int i;
1830 
1831 				printf("Copied %d bytes of sense data:",
1832 				       sense_size);
1833 				for (i = 0; i < sense_size; i++) {
1834 					if ((i & 0xF) == 0)
1835 						printf("\n");
1836 					printf("0x%x ", cmd->sense_buffer[i]);
1837 				}
1838 				printf("\n");
1839 			}
1840 #endif
1841 		}
1842 		break;
1843 	}
1844 	case SCSI_STATUS_QUEUE_FULL:
1845 	{
1846 		/*
1847 		 * By the time the core driver has returned this
1848 		 * command, all other commands that were queued
1849 		 * to us but not the device have been returned.
1850 		 * This ensures that dev->active is equal to
1851 		 * the number of commands actually queued to
1852 		 * the device.
1853 		 */
1854 		dev->tag_success_count = 0;
1855 		if (dev->active != 0) {
1856 			/*
1857 			 * Drop our opening count to the number
1858 			 * of commands currently outstanding.
1859 			 */
1860 			dev->openings = 0;
1861 /*
1862 			ahc_print_path(ahc, scb);
1863 			printf("Dropping tag count to %d\n", dev->active);
1864  */
1865 			if (dev->active == dev->tags_on_last_queuefull) {
1866 
1867 				dev->last_queuefull_same_count++;
1868 				/*
1869 				 * If we repeatedly see a queue full
1870 				 * at the same queue depth, this
1871 				 * device has a fixed number of tag
1872 				 * slots.  Lock in this tag depth
1873 				 * so we stop seeing queue fulls from
1874 				 * this device.
1875 				 */
1876 				if (dev->last_queuefull_same_count
1877 				 == AHC_LOCK_TAGS_COUNT) {
1878 					dev->maxtags = dev->active;
1879 					ahc_print_path(ahc, scb);
1880 					printf("Locking max tag count at %d\n",
1881 					       dev->active);
1882 				}
1883 			} else {
1884 				dev->tags_on_last_queuefull = dev->active;
1885 				dev->last_queuefull_same_count = 0;
1886 			}
1887 			ahc_set_transaction_status(scb, CAM_REQUEUE_REQ);
1888 			ahc_set_scsi_status(scb, SCSI_STATUS_OK);
1889 			ahc_platform_set_tags(ahc, &devinfo,
1890 				     (dev->flags & AHC_DEV_Q_BASIC)
1891 				   ? AHC_QUEUE_BASIC : AHC_QUEUE_TAGGED);
1892 			break;
1893 		}
1894 		/*
1895 		 * Drop down to a single opening, and treat this
1896 		 * as if the target returned BUSY SCSI status.
1897 		 */
1898 		dev->openings = 1;
1899 		ahc_set_scsi_status(scb, SCSI_STATUS_BUSY);
1900 		ahc_platform_set_tags(ahc, &devinfo,
1901 			     (dev->flags & AHC_DEV_Q_BASIC)
1902 			   ? AHC_QUEUE_BASIC : AHC_QUEUE_TAGGED);
1903 		break;
1904 	}
1905 	}
1906 }
1907 
1908 static void
1909 ahc_linux_queue_cmd_complete(struct ahc_softc *ahc, struct scsi_cmnd *cmd)
1910 {
1911 	/*
1912 	 * Map CAM error codes into Linux Error codes.  We
1913 	 * avoid the conversion so that the DV code has the
1914 	 * full error information available when making
1915 	 * state change decisions.
1916 	 */
1917 	{
1918 		u_int new_status;
1919 
1920 		switch (ahc_cmd_get_transaction_status(cmd)) {
1921 		case CAM_REQ_INPROG:
1922 		case CAM_REQ_CMP:
1923 		case CAM_SCSI_STATUS_ERROR:
1924 			new_status = DID_OK;
1925 			break;
1926 		case CAM_REQ_ABORTED:
1927 			new_status = DID_ABORT;
1928 			break;
1929 		case CAM_BUSY:
1930 			new_status = DID_BUS_BUSY;
1931 			break;
1932 		case CAM_REQ_INVALID:
1933 		case CAM_PATH_INVALID:
1934 			new_status = DID_BAD_TARGET;
1935 			break;
1936 		case CAM_SEL_TIMEOUT:
1937 			new_status = DID_NO_CONNECT;
1938 			break;
1939 		case CAM_SCSI_BUS_RESET:
1940 		case CAM_BDR_SENT:
1941 			new_status = DID_RESET;
1942 			break;
1943 		case CAM_UNCOR_PARITY:
1944 			new_status = DID_PARITY;
1945 			break;
1946 		case CAM_CMD_TIMEOUT:
1947 			new_status = DID_TIME_OUT;
1948 			break;
1949 		case CAM_UA_ABORT:
1950 		case CAM_REQ_CMP_ERR:
1951 		case CAM_AUTOSENSE_FAIL:
1952 		case CAM_NO_HBA:
1953 		case CAM_DATA_RUN_ERR:
1954 		case CAM_UNEXP_BUSFREE:
1955 		case CAM_SEQUENCE_FAIL:
1956 		case CAM_CCB_LEN_ERR:
1957 		case CAM_PROVIDE_FAIL:
1958 		case CAM_REQ_TERMIO:
1959 		case CAM_UNREC_HBA_ERROR:
1960 		case CAM_REQ_TOO_BIG:
1961 			new_status = DID_ERROR;
1962 			break;
1963 		case CAM_REQUEUE_REQ:
1964 			new_status = DID_REQUEUE;
1965 			break;
1966 		default:
1967 			/* We should never get here */
1968 			new_status = DID_ERROR;
1969 			break;
1970 		}
1971 
1972 		ahc_cmd_set_transaction_status(cmd, new_status);
1973 	}
1974 
1975 	cmd->scsi_done(cmd);
1976 }
1977 
1978 static void
1979 ahc_linux_sem_timeout(u_long arg)
1980 {
1981 	struct	ahc_softc *ahc;
1982 	u_long	s;
1983 
1984 	ahc = (struct ahc_softc *)arg;
1985 
1986 	ahc_lock(ahc, &s);
1987 	if ((ahc->platform_data->flags & AHC_UP_EH_SEMAPHORE) != 0) {
1988 		ahc->platform_data->flags &= ~AHC_UP_EH_SEMAPHORE;
1989 		up(&ahc->platform_data->eh_sem);
1990 	}
1991 	ahc_unlock(ahc, &s);
1992 }
1993 
1994 static void
1995 ahc_linux_freeze_simq(struct ahc_softc *ahc)
1996 {
1997 	ahc->platform_data->qfrozen++;
1998 	if (ahc->platform_data->qfrozen == 1) {
1999 		scsi_block_requests(ahc->platform_data->host);
2000 
2001 		/* XXX What about Twin channels? */
2002 		ahc_platform_abort_scbs(ahc, CAM_TARGET_WILDCARD, ALL_CHANNELS,
2003 					CAM_LUN_WILDCARD, SCB_LIST_NULL,
2004 					ROLE_INITIATOR, CAM_REQUEUE_REQ);
2005 	}
2006 }
2007 
2008 static void
2009 ahc_linux_release_simq(u_long arg)
2010 {
2011 	struct ahc_softc *ahc;
2012 	u_long s;
2013 	int    unblock_reqs;
2014 
2015 	ahc = (struct ahc_softc *)arg;
2016 
2017 	unblock_reqs = 0;
2018 	ahc_lock(ahc, &s);
2019 	if (ahc->platform_data->qfrozen > 0)
2020 		ahc->platform_data->qfrozen--;
2021 	if (ahc->platform_data->qfrozen == 0)
2022 		unblock_reqs = 1;
2023 	ahc_unlock(ahc, &s);
2024 	/*
2025 	 * There is still a race here.  The mid-layer
2026 	 * should keep its own freeze count and use
2027 	 * a bottom half handler to run the queues
2028 	 * so we can unblock with our own lock held.
2029 	 */
2030 	if (unblock_reqs)
2031 		scsi_unblock_requests(ahc->platform_data->host);
2032 }
2033 
2034 static int
2035 ahc_linux_queue_recovery_cmd(struct scsi_cmnd *cmd, scb_flag flag)
2036 {
2037 	struct ahc_softc *ahc;
2038 	struct ahc_linux_device *dev;
2039 	struct scb *pending_scb;
2040 	u_int  saved_scbptr;
2041 	u_int  active_scb_index;
2042 	u_int  last_phase;
2043 	u_int  saved_scsiid;
2044 	u_int  cdb_byte;
2045 	int    retval;
2046 	int    was_paused;
2047 	int    paused;
2048 	int    wait;
2049 	int    disconnected;
2050 
2051 	pending_scb = NULL;
2052 	paused = FALSE;
2053 	wait = FALSE;
2054 	ahc = *(struct ahc_softc **)cmd->device->host->hostdata;
2055 
2056 	printf("%s:%d:%d:%d: Attempting to queue a%s message\n",
2057 	       ahc_name(ahc), cmd->device->channel,
2058 	       cmd->device->id, cmd->device->lun,
2059 	       flag == SCB_ABORT ? "n ABORT" : " TARGET RESET");
2060 
2061 	printf("CDB:");
2062 	for (cdb_byte = 0; cdb_byte < cmd->cmd_len; cdb_byte++)
2063 		printf(" 0x%x", cmd->cmnd[cdb_byte]);
2064 	printf("\n");
2065 
2066 	spin_lock_irq(&ahc->platform_data->spin_lock);
2067 
2068 	/*
2069 	 * First determine if we currently own this command.
2070 	 * Start by searching the device queue.  If not found
2071 	 * there, check the pending_scb list.  If not found
2072 	 * at all, and the system wanted us to just abort the
2073 	 * command, return success.
2074 	 */
2075 	dev = scsi_transport_device_data(cmd->device);
2076 
2077 	if (dev == NULL) {
2078 		/*
2079 		 * No target device for this command exists,
2080 		 * so we must not still own the command.
2081 		 */
2082 		printf("%s:%d:%d:%d: Is not an active device\n",
2083 		       ahc_name(ahc), cmd->device->channel, cmd->device->id,
2084 		       cmd->device->lun);
2085 		retval = SUCCESS;
2086 		goto no_cmd;
2087 	}
2088 
2089 	if ((dev->flags & (AHC_DEV_Q_BASIC|AHC_DEV_Q_TAGGED)) == 0
2090 	 && ahc_search_untagged_queues(ahc, cmd, cmd->device->id,
2091 				       cmd->device->channel + 'A',
2092 				       cmd->device->lun,
2093 				       CAM_REQ_ABORTED, SEARCH_COMPLETE) != 0) {
2094 		printf("%s:%d:%d:%d: Command found on untagged queue\n",
2095 		       ahc_name(ahc), cmd->device->channel, cmd->device->id,
2096 		       cmd->device->lun);
2097 		retval = SUCCESS;
2098 		goto done;
2099 	}
2100 
2101 	/*
2102 	 * See if we can find a matching cmd in the pending list.
2103 	 */
2104 	LIST_FOREACH(pending_scb, &ahc->pending_scbs, pending_links) {
2105 		if (pending_scb->io_ctx == cmd)
2106 			break;
2107 	}
2108 
2109 	if (pending_scb == NULL && flag == SCB_DEVICE_RESET) {
2110 
2111 		/* Any SCB for this device will do for a target reset */
2112 		LIST_FOREACH(pending_scb, &ahc->pending_scbs, pending_links) {
2113 		  	if (ahc_match_scb(ahc, pending_scb, cmd->device->id,
2114 					  cmd->device->channel + 'A',
2115 					  CAM_LUN_WILDCARD,
2116 					  SCB_LIST_NULL, ROLE_INITIATOR) == 0)
2117 				break;
2118 		}
2119 	}
2120 
2121 	if (pending_scb == NULL) {
2122 		printf("%s:%d:%d:%d: Command not found\n",
2123 		       ahc_name(ahc), cmd->device->channel, cmd->device->id,
2124 		       cmd->device->lun);
2125 		goto no_cmd;
2126 	}
2127 
2128 	if ((pending_scb->flags & SCB_RECOVERY_SCB) != 0) {
2129 		/*
2130 		 * We can't queue two recovery actions using the same SCB
2131 		 */
2132 		retval = FAILED;
2133 		goto  done;
2134 	}
2135 
2136 	/*
2137 	 * Ensure that the card doesn't do anything
2138 	 * behind our back and that we didn't "just" miss
2139 	 * an interrupt that would affect this cmd.
2140 	 */
2141 	was_paused = ahc_is_paused(ahc);
2142 	ahc_pause_and_flushwork(ahc);
2143 	paused = TRUE;
2144 
2145 	if ((pending_scb->flags & SCB_ACTIVE) == 0) {
2146 		printf("%s:%d:%d:%d: Command already completed\n",
2147 		       ahc_name(ahc), cmd->device->channel, cmd->device->id,
2148 		       cmd->device->lun);
2149 		goto no_cmd;
2150 	}
2151 
2152 	printf("%s: At time of recovery, card was %spaused\n",
2153 	       ahc_name(ahc), was_paused ? "" : "not ");
2154 	ahc_dump_card_state(ahc);
2155 
2156 	disconnected = TRUE;
2157 	if (flag == SCB_ABORT) {
2158 		if (ahc_search_qinfifo(ahc, cmd->device->id,
2159 				       cmd->device->channel + 'A',
2160 				       cmd->device->lun,
2161 				       pending_scb->hscb->tag,
2162 				       ROLE_INITIATOR, CAM_REQ_ABORTED,
2163 				       SEARCH_COMPLETE) > 0) {
2164 			printf("%s:%d:%d:%d: Cmd aborted from QINFIFO\n",
2165 			       ahc_name(ahc), cmd->device->channel,
2166 					cmd->device->id, cmd->device->lun);
2167 			retval = SUCCESS;
2168 			goto done;
2169 		}
2170 	} else if (ahc_search_qinfifo(ahc, cmd->device->id,
2171 				      cmd->device->channel + 'A',
2172 				      cmd->device->lun, pending_scb->hscb->tag,
2173 				      ROLE_INITIATOR, /*status*/0,
2174 				      SEARCH_COUNT) > 0) {
2175 		disconnected = FALSE;
2176 	}
2177 
2178 	if (disconnected && (ahc_inb(ahc, SEQ_FLAGS) & NOT_IDENTIFIED) == 0) {
2179 		struct scb *bus_scb;
2180 
2181 		bus_scb = ahc_lookup_scb(ahc, ahc_inb(ahc, SCB_TAG));
2182 		if (bus_scb == pending_scb)
2183 			disconnected = FALSE;
2184 		else if (flag != SCB_ABORT
2185 		      && ahc_inb(ahc, SAVED_SCSIID) == pending_scb->hscb->scsiid
2186 		      && ahc_inb(ahc, SAVED_LUN) == SCB_GET_LUN(pending_scb))
2187 			disconnected = FALSE;
2188 	}
2189 
2190 	/*
2191 	 * At this point, pending_scb is the scb associated with the
2192 	 * passed in command.  That command is currently active on the
2193 	 * bus, is in the disconnected state, or we're hoping to find
2194 	 * a command for the same target active on the bus to abuse to
2195 	 * send a BDR.  Queue the appropriate message based on which of
2196 	 * these states we are in.
2197 	 */
2198 	last_phase = ahc_inb(ahc, LASTPHASE);
2199 	saved_scbptr = ahc_inb(ahc, SCBPTR);
2200 	active_scb_index = ahc_inb(ahc, SCB_TAG);
2201 	saved_scsiid = ahc_inb(ahc, SAVED_SCSIID);
2202 	if (last_phase != P_BUSFREE
2203 	 && (pending_scb->hscb->tag == active_scb_index
2204 	  || (flag == SCB_DEVICE_RESET
2205 	   && SCSIID_TARGET(ahc, saved_scsiid) == cmd->device->id))) {
2206 
2207 		/*
2208 		 * We're active on the bus, so assert ATN
2209 		 * and hope that the target responds.
2210 		 */
2211 		pending_scb = ahc_lookup_scb(ahc, active_scb_index);
2212 		pending_scb->flags |= SCB_RECOVERY_SCB|flag;
2213 		ahc_outb(ahc, MSG_OUT, HOST_MSG);
2214 		ahc_outb(ahc, SCSISIGO, last_phase|ATNO);
2215 		printf("%s:%d:%d:%d: Device is active, asserting ATN\n",
2216 		       ahc_name(ahc), cmd->device->channel, cmd->device->id,
2217 		       cmd->device->lun);
2218 		wait = TRUE;
2219 	} else if (disconnected) {
2220 
2221 		/*
2222 		 * Actually re-queue this SCB in an attempt
2223 		 * to select the device before it reconnects.
2224 		 * In either case (selection or reselection),
2225 		 * we will now issue the approprate message
2226 		 * to the timed-out device.
2227 		 *
2228 		 * Set the MK_MESSAGE control bit indicating
2229 		 * that we desire to send a message.  We
2230 		 * also set the disconnected flag since
2231 		 * in the paging case there is no guarantee
2232 		 * that our SCB control byte matches the
2233 		 * version on the card.  We don't want the
2234 		 * sequencer to abort the command thinking
2235 		 * an unsolicited reselection occurred.
2236 		 */
2237 		pending_scb->hscb->control |= MK_MESSAGE|DISCONNECTED;
2238 		pending_scb->flags |= SCB_RECOVERY_SCB|flag;
2239 
2240 		/*
2241 		 * Remove any cached copy of this SCB in the
2242 		 * disconnected list in preparation for the
2243 		 * queuing of our abort SCB.  We use the
2244 		 * same element in the SCB, SCB_NEXT, for
2245 		 * both the qinfifo and the disconnected list.
2246 		 */
2247 		ahc_search_disc_list(ahc, cmd->device->id,
2248 				     cmd->device->channel + 'A',
2249 				     cmd->device->lun, pending_scb->hscb->tag,
2250 				     /*stop_on_first*/TRUE,
2251 				     /*remove*/TRUE,
2252 				     /*save_state*/FALSE);
2253 
2254 		/*
2255 		 * In the non-paging case, the sequencer will
2256 		 * never re-reference the in-core SCB.
2257 		 * To make sure we are notified during
2258 		 * reslection, set the MK_MESSAGE flag in
2259 		 * the card's copy of the SCB.
2260 		 */
2261 		if ((ahc->flags & AHC_PAGESCBS) == 0) {
2262 			ahc_outb(ahc, SCBPTR, pending_scb->hscb->tag);
2263 			ahc_outb(ahc, SCB_CONTROL,
2264 				 ahc_inb(ahc, SCB_CONTROL)|MK_MESSAGE);
2265 		}
2266 
2267 		/*
2268 		 * Clear out any entries in the QINFIFO first
2269 		 * so we are the next SCB for this target
2270 		 * to run.
2271 		 */
2272 		ahc_search_qinfifo(ahc, cmd->device->id,
2273 				   cmd->device->channel + 'A',
2274 				   cmd->device->lun, SCB_LIST_NULL,
2275 				   ROLE_INITIATOR, CAM_REQUEUE_REQ,
2276 				   SEARCH_COMPLETE);
2277 		ahc_qinfifo_requeue_tail(ahc, pending_scb);
2278 		ahc_outb(ahc, SCBPTR, saved_scbptr);
2279 		ahc_print_path(ahc, pending_scb);
2280 		printf("Device is disconnected, re-queuing SCB\n");
2281 		wait = TRUE;
2282 	} else {
2283 		printf("%s:%d:%d:%d: Unable to deliver message\n",
2284 		       ahc_name(ahc), cmd->device->channel, cmd->device->id,
2285 		       cmd->device->lun);
2286 		retval = FAILED;
2287 		goto done;
2288 	}
2289 
2290 no_cmd:
2291 	/*
2292 	 * Our assumption is that if we don't have the command, no
2293 	 * recovery action was required, so we return success.  Again,
2294 	 * the semantics of the mid-layer recovery engine are not
2295 	 * well defined, so this may change in time.
2296 	 */
2297 	retval = SUCCESS;
2298 done:
2299 	if (paused)
2300 		ahc_unpause(ahc);
2301 	if (wait) {
2302 		struct timer_list timer;
2303 		int ret;
2304 
2305 		ahc->platform_data->flags |= AHC_UP_EH_SEMAPHORE;
2306 		spin_unlock_irq(&ahc->platform_data->spin_lock);
2307 		init_timer(&timer);
2308 		timer.data = (u_long)ahc;
2309 		timer.expires = jiffies + (5 * HZ);
2310 		timer.function = ahc_linux_sem_timeout;
2311 		add_timer(&timer);
2312 		printf("Recovery code sleeping\n");
2313 		down(&ahc->platform_data->eh_sem);
2314 		printf("Recovery code awake\n");
2315         	ret = del_timer_sync(&timer);
2316 		if (ret == 0) {
2317 			printf("Timer Expired\n");
2318 			retval = FAILED;
2319 		}
2320 		spin_lock_irq(&ahc->platform_data->spin_lock);
2321 	}
2322 
2323 	spin_unlock_irq(&ahc->platform_data->spin_lock);
2324 	return (retval);
2325 }
2326 
2327 void
2328 ahc_platform_dump_card_state(struct ahc_softc *ahc)
2329 {
2330 }
2331 
2332 static void ahc_linux_exit(void);
2333 
2334 static void ahc_linux_set_width(struct scsi_target *starget, int width)
2335 {
2336 	struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
2337 	struct ahc_softc *ahc = *((struct ahc_softc **)shost->hostdata);
2338 	struct ahc_devinfo devinfo;
2339 	unsigned long flags;
2340 
2341 	ahc_compile_devinfo(&devinfo, shost->this_id, starget->id, 0,
2342 			    starget->channel + 'A', ROLE_INITIATOR);
2343 	ahc_lock(ahc, &flags);
2344 	ahc_set_width(ahc, &devinfo, width, AHC_TRANS_GOAL, FALSE);
2345 	ahc_unlock(ahc, &flags);
2346 }
2347 
2348 static void ahc_linux_set_period(struct scsi_target *starget, int period)
2349 {
2350 	struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
2351 	struct ahc_softc *ahc = *((struct ahc_softc **)shost->hostdata);
2352 	struct ahc_tmode_tstate *tstate;
2353 	struct ahc_initiator_tinfo *tinfo
2354 		= ahc_fetch_transinfo(ahc,
2355 				      starget->channel + 'A',
2356 				      shost->this_id, starget->id, &tstate);
2357 	struct ahc_devinfo devinfo;
2358 	unsigned int ppr_options = tinfo->goal.ppr_options;
2359 	unsigned long flags;
2360 	unsigned long offset = tinfo->goal.offset;
2361 	struct ahc_syncrate *syncrate;
2362 
2363 	if (offset == 0)
2364 		offset = MAX_OFFSET;
2365 
2366 	if (period < 9)
2367 		period = 9;	/* 12.5ns is our minimum */
2368 	if (period == 9)
2369 		ppr_options |= MSG_EXT_PPR_DT_REQ;
2370 
2371 	ahc_compile_devinfo(&devinfo, shost->this_id, starget->id, 0,
2372 			    starget->channel + 'A', ROLE_INITIATOR);
2373 
2374 	/* all PPR requests apart from QAS require wide transfers */
2375 	if (ppr_options & ~MSG_EXT_PPR_QAS_REQ) {
2376 		if (spi_width(starget) == 0)
2377 			ppr_options &= MSG_EXT_PPR_QAS_REQ;
2378 	}
2379 
2380 	syncrate = ahc_find_syncrate(ahc, &period, &ppr_options, AHC_SYNCRATE_DT);
2381 	ahc_lock(ahc, &flags);
2382 	ahc_set_syncrate(ahc, &devinfo, syncrate, period, offset,
2383 			 ppr_options, AHC_TRANS_GOAL, FALSE);
2384 	ahc_unlock(ahc, &flags);
2385 }
2386 
2387 static void ahc_linux_set_offset(struct scsi_target *starget, int offset)
2388 {
2389 	struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
2390 	struct ahc_softc *ahc = *((struct ahc_softc **)shost->hostdata);
2391 	struct ahc_tmode_tstate *tstate;
2392 	struct ahc_initiator_tinfo *tinfo
2393 		= ahc_fetch_transinfo(ahc,
2394 				      starget->channel + 'A',
2395 				      shost->this_id, starget->id, &tstate);
2396 	struct ahc_devinfo devinfo;
2397 	unsigned int ppr_options = 0;
2398 	unsigned int period = 0;
2399 	unsigned long flags;
2400 	struct ahc_syncrate *syncrate = NULL;
2401 
2402 	ahc_compile_devinfo(&devinfo, shost->this_id, starget->id, 0,
2403 			    starget->channel + 'A', ROLE_INITIATOR);
2404 	if (offset != 0) {
2405 		syncrate = ahc_find_syncrate(ahc, &period, &ppr_options, AHC_SYNCRATE_DT);
2406 		period = tinfo->goal.period;
2407 		ppr_options = tinfo->goal.ppr_options;
2408 	}
2409 	ahc_lock(ahc, &flags);
2410 	ahc_set_syncrate(ahc, &devinfo, syncrate, period, offset,
2411 			 ppr_options, AHC_TRANS_GOAL, FALSE);
2412 	ahc_unlock(ahc, &flags);
2413 }
2414 
2415 static void ahc_linux_set_dt(struct scsi_target *starget, int dt)
2416 {
2417 	struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
2418 	struct ahc_softc *ahc = *((struct ahc_softc **)shost->hostdata);
2419 	struct ahc_tmode_tstate *tstate;
2420 	struct ahc_initiator_tinfo *tinfo
2421 		= ahc_fetch_transinfo(ahc,
2422 				      starget->channel + 'A',
2423 				      shost->this_id, starget->id, &tstate);
2424 	struct ahc_devinfo devinfo;
2425 	unsigned int ppr_options = tinfo->goal.ppr_options
2426 		& ~MSG_EXT_PPR_DT_REQ;
2427 	unsigned int period = tinfo->goal.period;
2428 	unsigned int width = tinfo->goal.width;
2429 	unsigned long flags;
2430 	struct ahc_syncrate *syncrate;
2431 
2432 	if (dt) {
2433 		ppr_options |= MSG_EXT_PPR_DT_REQ;
2434 		if (!width)
2435 			ahc_linux_set_width(starget, 1);
2436 	} else if (period == 9)
2437 		period = 10;	/* if resetting DT, period must be >= 25ns */
2438 
2439 	ahc_compile_devinfo(&devinfo, shost->this_id, starget->id, 0,
2440 			    starget->channel + 'A', ROLE_INITIATOR);
2441 	syncrate = ahc_find_syncrate(ahc, &period, &ppr_options,AHC_SYNCRATE_DT);
2442 	ahc_lock(ahc, &flags);
2443 	ahc_set_syncrate(ahc, &devinfo, syncrate, period, tinfo->goal.offset,
2444 			 ppr_options, AHC_TRANS_GOAL, FALSE);
2445 	ahc_unlock(ahc, &flags);
2446 }
2447 
2448 #if 0
2449 /* FIXME: This code claims to support IU and QAS.  However, the actual
2450  * sequencer code and aic7xxx_core have no support for these parameters and
2451  * will get into a bad state if they're negotiated.  Do not enable this
2452  * unless you know what you're doing */
2453 static void ahc_linux_set_qas(struct scsi_target *starget, int qas)
2454 {
2455 	struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
2456 	struct ahc_softc *ahc = *((struct ahc_softc **)shost->hostdata);
2457 	struct ahc_tmode_tstate *tstate;
2458 	struct ahc_initiator_tinfo *tinfo
2459 		= ahc_fetch_transinfo(ahc,
2460 				      starget->channel + 'A',
2461 				      shost->this_id, starget->id, &tstate);
2462 	struct ahc_devinfo devinfo;
2463 	unsigned int ppr_options = tinfo->goal.ppr_options
2464 		& ~MSG_EXT_PPR_QAS_REQ;
2465 	unsigned int period = tinfo->goal.period;
2466 	unsigned long flags;
2467 	struct ahc_syncrate *syncrate;
2468 
2469 	if (qas)
2470 		ppr_options |= MSG_EXT_PPR_QAS_REQ;
2471 
2472 	ahc_compile_devinfo(&devinfo, shost->this_id, starget->id, 0,
2473 			    starget->channel + 'A', ROLE_INITIATOR);
2474 	syncrate = ahc_find_syncrate(ahc, &period, &ppr_options, AHC_SYNCRATE_DT);
2475 	ahc_lock(ahc, &flags);
2476 	ahc_set_syncrate(ahc, &devinfo, syncrate, period, tinfo->goal.offset,
2477 			 ppr_options, AHC_TRANS_GOAL, FALSE);
2478 	ahc_unlock(ahc, &flags);
2479 }
2480 
2481 static void ahc_linux_set_iu(struct scsi_target *starget, int iu)
2482 {
2483 	struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
2484 	struct ahc_softc *ahc = *((struct ahc_softc **)shost->hostdata);
2485 	struct ahc_tmode_tstate *tstate;
2486 	struct ahc_initiator_tinfo *tinfo
2487 		= ahc_fetch_transinfo(ahc,
2488 				      starget->channel + 'A',
2489 				      shost->this_id, starget->id, &tstate);
2490 	struct ahc_devinfo devinfo;
2491 	unsigned int ppr_options = tinfo->goal.ppr_options
2492 		& ~MSG_EXT_PPR_IU_REQ;
2493 	unsigned int period = tinfo->goal.period;
2494 	unsigned long flags;
2495 	struct ahc_syncrate *syncrate;
2496 
2497 	if (iu)
2498 		ppr_options |= MSG_EXT_PPR_IU_REQ;
2499 
2500 	ahc_compile_devinfo(&devinfo, shost->this_id, starget->id, 0,
2501 			    starget->channel + 'A', ROLE_INITIATOR);
2502 	syncrate = ahc_find_syncrate(ahc, &period, &ppr_options, AHC_SYNCRATE_DT);
2503 	ahc_lock(ahc, &flags);
2504 	ahc_set_syncrate(ahc, &devinfo, syncrate, period, tinfo->goal.offset,
2505 			 ppr_options, AHC_TRANS_GOAL, FALSE);
2506 	ahc_unlock(ahc, &flags);
2507 }
2508 #endif
2509 
2510 static struct spi_function_template ahc_linux_transport_functions = {
2511 	.set_offset	= ahc_linux_set_offset,
2512 	.show_offset	= 1,
2513 	.set_period	= ahc_linux_set_period,
2514 	.show_period	= 1,
2515 	.set_width	= ahc_linux_set_width,
2516 	.show_width	= 1,
2517 	.set_dt		= ahc_linux_set_dt,
2518 	.show_dt	= 1,
2519 #if 0
2520 	.set_iu		= ahc_linux_set_iu,
2521 	.show_iu	= 1,
2522 	.set_qas	= ahc_linux_set_qas,
2523 	.show_qas	= 1,
2524 #endif
2525 };
2526 
2527 
2528 
2529 static int __init
2530 ahc_linux_init(void)
2531 {
2532 	/*
2533 	 * If we've been passed any parameters, process them now.
2534 	 */
2535 	if (aic7xxx)
2536 		aic7xxx_setup(aic7xxx);
2537 
2538 	ahc_linux_transport_template =
2539 		spi_attach_transport(&ahc_linux_transport_functions);
2540 	if (!ahc_linux_transport_template)
2541 		return -ENODEV;
2542 
2543 	scsi_transport_reserve_target(ahc_linux_transport_template,
2544 				      sizeof(struct ahc_linux_target));
2545 	scsi_transport_reserve_device(ahc_linux_transport_template,
2546 				      sizeof(struct ahc_linux_device));
2547 
2548 	ahc_linux_pci_init();
2549 	ahc_linux_eisa_init();
2550 	return 0;
2551 }
2552 
2553 static void
2554 ahc_linux_exit(void)
2555 {
2556 	ahc_linux_pci_exit();
2557 	ahc_linux_eisa_exit();
2558 	spi_release_transport(ahc_linux_transport_template);
2559 }
2560 
2561 module_init(ahc_linux_init);
2562 module_exit(ahc_linux_exit);
2563