xref: /titanic_50/usr/src/uts/common/io/fibre-channel/fca/qlc/ql_api.c (revision db5b5f1e7e8c59712dae4a0b5c55ecee0e4c1bfc)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 
22 /* Copyright 2010 QLogic Corporation */
23 
24 /*
25  * Copyright (c) 2008, 2010, Oracle and/or its affiliates. All rights reserved.
26  */
27 /*
28  * Copyright 2011 Nexenta Systems, Inc.  All rights reserved.
29  */
30 
31 #pragma ident	"Copyright 2010 QLogic Corporation; ql_api.c"
32 
33 /*
34  * ISP2xxx Solaris Fibre Channel Adapter (FCA) driver source file.
35  *
36  * ***********************************************************************
37  * *									**
38  * *				NOTICE					**
39  * *		COPYRIGHT (C) 1996-2010 QLOGIC CORPORATION		**
40  * *			ALL RIGHTS RESERVED				**
41  * *									**
42  * ***********************************************************************
43  *
44  */
45 
46 #include <ql_apps.h>
47 #include <ql_api.h>
48 #include <ql_debug.h>
49 #include <ql_init.h>
50 #include <ql_iocb.h>
51 #include <ql_ioctl.h>
52 #include <ql_isr.h>
53 #include <ql_mbx.h>
54 #include <ql_nx.h>
55 #include <ql_xioctl.h>
56 
57 /*
58  * Solaris external defines.
59  */
60 extern pri_t minclsyspri;
61 extern pri_t maxclsyspri;
62 
63 /*
64  * dev_ops functions prototypes
65  */
66 static int ql_getinfo(dev_info_t *, ddi_info_cmd_t, void *, void **);
67 static int ql_attach(dev_info_t *, ddi_attach_cmd_t);
68 static int ql_detach(dev_info_t *, ddi_detach_cmd_t);
69 static int ql_power(dev_info_t *, int, int);
70 static int ql_quiesce(dev_info_t *);
71 
72 /*
73  * FCA functions prototypes exported by means of the transport table
74  */
75 static opaque_t ql_bind_port(dev_info_t *, fc_fca_port_info_t *,
76     fc_fca_bind_info_t *);
77 static void ql_unbind_port(opaque_t);
78 static int ql_init_pkt(opaque_t, fc_packet_t *, int);
79 static int ql_un_init_pkt(opaque_t, fc_packet_t *);
80 static int ql_els_send(opaque_t, fc_packet_t *);
81 static int ql_get_cap(opaque_t, char *, void *);
82 static int ql_set_cap(opaque_t, char *, void *);
83 static int ql_getmap(opaque_t, fc_lilpmap_t *);
84 static int ql_transport(opaque_t, fc_packet_t *);
85 static int ql_ub_alloc(opaque_t, uint64_t *, uint32_t, uint32_t *, uint32_t);
86 static int ql_ub_free(opaque_t, uint32_t, uint64_t *);
87 static int ql_ub_release(opaque_t, uint32_t, uint64_t *);
88 static int ql_abort(opaque_t, fc_packet_t *, int);
89 static int ql_reset(opaque_t, uint32_t);
90 static int ql_port_manage(opaque_t, fc_fca_pm_t *);
91 static opaque_t ql_get_device(opaque_t, fc_portid_t);
92 
93 /*
94  * FCA Driver Support Function Prototypes.
95  */
96 static uint16_t	ql_wait_outstanding(ql_adapter_state_t *);
97 static void ql_task_mgmt(ql_adapter_state_t *, ql_tgt_t *, fc_packet_t *,
98     ql_srb_t *);
99 static void ql_task_daemon(void *);
100 static void ql_task_thread(ql_adapter_state_t *);
101 static void ql_unsol_callback(ql_srb_t *);
102 static void ql_free_unsolicited_buffer(ql_adapter_state_t *,
103     fc_unsol_buf_t *);
104 static void ql_timer(void *);
105 static void ql_watchdog(ql_adapter_state_t *, uint32_t *, uint32_t *);
106 static void ql_cmd_timeout(ql_adapter_state_t *, ql_tgt_t *q, ql_srb_t *,
107     uint32_t *, uint32_t *);
108 static void ql_halt(ql_adapter_state_t *, int);
109 static int ql_els_plogi(ql_adapter_state_t *, fc_packet_t *);
110 static int ql_els_flogi(ql_adapter_state_t *, fc_packet_t *);
111 static int ql_els_logo(ql_adapter_state_t *, fc_packet_t *);
112 static int ql_els_prli(ql_adapter_state_t *, fc_packet_t *);
113 static int ql_els_prlo(ql_adapter_state_t *, fc_packet_t *);
114 static int ql_els_adisc(ql_adapter_state_t *, fc_packet_t *);
115 static int ql_els_linit(ql_adapter_state_t *, fc_packet_t *);
116 static int ql_els_lpc(ql_adapter_state_t *, fc_packet_t *);
117 static int ql_els_lsts(ql_adapter_state_t *, fc_packet_t *);
118 static int ql_els_scr(ql_adapter_state_t *, fc_packet_t *);
119 static int ql_els_rscn(ql_adapter_state_t *, fc_packet_t *);
120 static int ql_els_farp_req(ql_adapter_state_t *, fc_packet_t *);
121 static int ql_els_farp_reply(ql_adapter_state_t *, fc_packet_t *);
122 static int ql_els_rls(ql_adapter_state_t *, fc_packet_t *);
123 static int ql_els_rnid(ql_adapter_state_t *, fc_packet_t *);
124 static int ql_login_port(ql_adapter_state_t *, port_id_t);
125 static int ql_login_fabric_port(ql_adapter_state_t *, ql_tgt_t *, uint16_t);
126 static int ql_logout_port(ql_adapter_state_t *, port_id_t);
127 static ql_lun_t *ql_lun_queue(ql_adapter_state_t *, ql_tgt_t *, uint16_t);
128 static int ql_fcp_scsi_cmd(ql_adapter_state_t *, fc_packet_t *, ql_srb_t *);
129 static int ql_fcp_ip_cmd(ql_adapter_state_t *, fc_packet_t *, ql_srb_t *);
130 static int ql_fc_services(ql_adapter_state_t *, fc_packet_t *);
131 static int ql_poll_cmd(ql_adapter_state_t *, ql_srb_t *, time_t);
132 static int ql_start_cmd(ql_adapter_state_t *, ql_tgt_t *, fc_packet_t *,
133     ql_srb_t *);
134 static int ql_kstat_update(kstat_t *, int);
135 static ql_adapter_state_t *ql_fca_handle_to_state(opaque_t);
136 static ql_adapter_state_t *ql_cmd_setup(opaque_t, fc_packet_t *, int *);
137 static int ql_program_flash_address(ql_adapter_state_t *, uint32_t, uint8_t);
138 static void ql_rst_aen(ql_adapter_state_t *);
139 static void ql_restart_queues(ql_adapter_state_t *);
140 static void ql_abort_queues(ql_adapter_state_t *);
141 static void ql_abort_device_queues(ql_adapter_state_t *ha, ql_tgt_t *tq);
142 static void ql_idle_check(ql_adapter_state_t *);
143 static int ql_loop_resync(ql_adapter_state_t *);
144 static size_t ql_24xx_ascii_fw_dump(ql_adapter_state_t *, caddr_t);
145 static size_t ql_2581_ascii_fw_dump(ql_adapter_state_t *, caddr_t);
146 static int ql_save_config_regs(dev_info_t *);
147 static int ql_restore_config_regs(dev_info_t *);
148 static int ql_process_rscn(ql_adapter_state_t *, fc_affected_id_t *);
149 static int ql_handle_rscn_update(ql_adapter_state_t *);
150 static int ql_send_plogi(ql_adapter_state_t *, ql_tgt_t *, ql_head_t *);
151 static int ql_process_rscn_for_device(ql_adapter_state_t *, ql_tgt_t *);
152 static int ql_dump_firmware(ql_adapter_state_t *);
153 static int ql_process_logo_for_device(ql_adapter_state_t *, ql_tgt_t *);
154 static int ql_2200_binary_fw_dump(ql_adapter_state_t *, ql_fw_dump_t *);
155 static int ql_2300_binary_fw_dump(ql_adapter_state_t *, ql_fw_dump_t *);
156 static int ql_24xx_binary_fw_dump(ql_adapter_state_t *, ql_24xx_fw_dump_t *);
157 static int ql_25xx_binary_fw_dump(ql_adapter_state_t *, ql_25xx_fw_dump_t *);
158 static int ql_81xx_binary_fw_dump(ql_adapter_state_t *, ql_81xx_fw_dump_t *);
159 static int ql_read_risc_ram(ql_adapter_state_t *, uint32_t, uint32_t,
160     void *);
161 static void *ql_read_regs(ql_adapter_state_t *, void *, void *, uint32_t,
162     uint8_t);
163 static int ql_busy_plogi(ql_adapter_state_t *, fc_packet_t *, ql_tgt_t *);
164 static int ql_suspend_adapter(ql_adapter_state_t *);
165 static int ql_bstr_to_dec(char *, uint32_t *, uint32_t);
166 static void ql_update_rscn(ql_adapter_state_t *, fc_affected_id_t *);
167 int ql_alloc_dma_resouce(ql_adapter_state_t *, dma_mem_t *, int);
168 static int ql_bind_dma_buffer(ql_adapter_state_t *, dma_mem_t *, int);
169 static void ql_unbind_dma_buffer(ql_adapter_state_t *, dma_mem_t *);
170 static void ql_timeout_insert(ql_adapter_state_t *, ql_tgt_t *, ql_srb_t *);
171 static int ql_setup_interrupts(ql_adapter_state_t *);
172 static int ql_setup_msi(ql_adapter_state_t *);
173 static int ql_setup_msix(ql_adapter_state_t *);
174 static int ql_setup_fixed(ql_adapter_state_t *);
175 static void ql_release_intr(ql_adapter_state_t *);
176 static void ql_disable_intr(ql_adapter_state_t *);
177 static int ql_legacy_intr(ql_adapter_state_t *);
178 static int ql_init_mutex(ql_adapter_state_t *);
179 static void ql_destroy_mutex(ql_adapter_state_t *);
180 static void ql_iidma(ql_adapter_state_t *);
181 
182 static int ql_n_port_plogi(ql_adapter_state_t *);
183 static void ql_fca_isp_els_request(ql_adapter_state_t *, fc_packet_t *,
184     els_descriptor_t *);
185 static void ql_isp_els_request_ctor(els_descriptor_t *,
186     els_passthru_entry_t *);
187 static int ql_p2p_plogi(ql_adapter_state_t *, fc_packet_t *);
188 static int ql_wait_for_td_stop(ql_adapter_state_t *);
189 static void ql_process_idc_event(ql_adapter_state_t *);
190 
191 /*
192  * Global data
193  */
194 static uint8_t	ql_enable_pm = 1;
195 static int	ql_flash_sbus_fpga = 0;
196 uint32_t	ql_os_release_level;
197 uint32_t	ql_disable_aif = 0;
198 uint32_t	ql_disable_msi = 0;
199 uint32_t	ql_disable_msix = 0;
200 uint32_t	ql_enable_ets = 0;
201 uint16_t	ql_osc_wait_count = 1000;
202 
203 /* Timer routine variables. */
204 static timeout_id_t	ql_timer_timeout_id = NULL;
205 static clock_t		ql_timer_ticks;
206 
207 /* Soft state head pointer. */
208 void *ql_state = NULL;
209 
210 /* Head adapter link. */
211 ql_head_t ql_hba = {
212 	NULL,
213 	NULL
214 };
215 
216 /* Global hba index */
217 uint32_t ql_gfru_hba_index = 1;
218 
219 /*
220  * Some IP defines and globals
221  */
222 uint32_t	ql_ip_buffer_count = 128;
223 uint32_t	ql_ip_low_water = 10;
224 uint8_t		ql_ip_fast_post_count = 5;
225 static int	ql_ip_mtu = 65280;		/* equivalent to FCIPMTU */
226 
227 /* Device AL_PA to Device Head Queue index array. */
228 uint8_t ql_alpa_to_index[] = {
229 	0x7e, 0x7d, 0x7c, 0x00, 0x7b, 0x01, 0x02, 0x03, 0x7a, 0x04,
230 	0x05, 0x06, 0x07, 0x08, 0x09, 0x79, 0x78, 0x0a, 0x0b, 0x0c,
231 	0x0d, 0x0e, 0x0f, 0x77, 0x76, 0x10, 0x11, 0x75, 0x12, 0x74,
232 	0x73, 0x72, 0x13, 0x14, 0x15, 0x71, 0x16, 0x70, 0x6f, 0x6e,
233 	0x17, 0x6d, 0x6c, 0x6b, 0x6a, 0x69, 0x68, 0x18, 0x19, 0x67,
234 	0x66, 0x65, 0x64, 0x63, 0x62, 0x20, 0x21, 0x61, 0x60, 0x23,
235 	0x5f, 0x24, 0x25, 0x26, 0x27, 0x28, 0x29, 0x5e, 0x2a, 0x5d,
236 	0x5c, 0x5b, 0x2b, 0x5a, 0x59, 0x58, 0x57, 0x56, 0x55, 0x2c,
237 	0x2d, 0x54, 0x53, 0x52, 0x51, 0x50, 0x4f, 0x2e, 0x2f, 0x4e,
238 	0x4d, 0x30, 0x4c, 0x31, 0x32, 0x33, 0x34, 0x35, 0x36, 0x4b,
239 	0x37, 0x4a, 0x49, 0x48, 0x38, 0x47, 0x46, 0x45, 0x44, 0x43,
240 	0x42, 0x39, 0x3a, 0x41, 0x40, 0x3f, 0x3e, 0x3d, 0x3c, 0x3b,
241 	0x3c, 0x3b, 0x3a, 0x3d, 0x39, 0x3e, 0x3f, 0x40, 0x38, 0x37,
242 	0x36, 0x41, 0x35, 0x42, 0x43, 0x44, 0x34, 0x45, 0x46, 0x47,
243 	0x48, 0x49, 0x4a, 0x33, 0x32, 0x4b, 0x4c, 0x4d, 0x4e, 0x4f,
244 	0x50, 0x31, 0x30, 0x51, 0x52, 0x2f, 0x53, 0x2e, 0x2d, 0x2c,
245 	0x54, 0x55, 0x56, 0x2b, 0x57, 0x2a, 0x29, 0x28, 0x58, 0x27,
246 	0x26, 0x25, 0x24, 0x23, 0x22, 0x59, 0x5a, 0x21, 0x20, 0x1f,
247 	0x1e, 0x1d, 0x1c, 0x5b, 0x5c, 0x1b, 0x1a, 0x5d, 0x19, 0x5e,
248 	0x5f, 0x60, 0x61, 0x62, 0x63, 0x18, 0x64, 0x17, 0x16, 0x15,
249 	0x65, 0x14, 0x13, 0x12, 0x11, 0x10, 0x0f, 0x66, 0x67, 0x0e,
250 	0x0d, 0x0c, 0x0b, 0x0a, 0x09, 0x68, 0x69, 0x08, 0x07, 0x6a,
251 	0x06, 0x6b, 0x6c, 0x6d, 0x05, 0x04, 0x03, 0x6e, 0x02, 0x6f,
252 	0x70, 0x71, 0x01, 0x72, 0x73, 0x74, 0x75, 0x76, 0x77, 0x00,
253 	0x78, 0x79, 0x7a, 0x7b, 0x7c, 0x7d, 0x7f, 0x80, 0x00, 0x01,
254 	0x02, 0x03, 0x80, 0x7f, 0x7e, 0x04
255 };
256 
257 /* Device loop_id to ALPA array. */
258 static uint8_t ql_index_to_alpa[] = {
259 	0xef, 0xe8, 0xe4, 0xe2, 0xe1, 0xe0, 0xdc, 0xda, 0xd9, 0xd6,
260 	0xd5, 0xd4, 0xd3, 0xd2, 0xd1, 0xce, 0xcd, 0xcc, 0xcb, 0xca,
261 	0xc9, 0xc7, 0xc6, 0xc5, 0xc3, 0xbc, 0xba, 0xb9, 0xb6, 0xb5,
262 	0xb4, 0xb3, 0xb2, 0xb1, 0xae, 0xad, 0xac, 0xab, 0xaa, 0xa9,
263 	0xa7, 0xa6, 0xa5, 0xa3, 0x9f, 0x9e, 0x9d, 0x9b, 0x98, 0x97,
264 	0x90, 0x8f, 0x88, 0x84, 0x82, 0x81, 0x80, 0x7c, 0x7a, 0x79,
265 	0x76, 0x75, 0x74, 0x73, 0x72, 0x71, 0x6e, 0x6d, 0x6c, 0x6b,
266 	0x6a, 0x69, 0x67, 0x66, 0x65, 0x63, 0x5c, 0x5a, 0x59, 0x56,
267 	0x55, 0x54, 0x53, 0x52, 0x51, 0x4e, 0x4d, 0x4c, 0x4b, 0x4a,
268 	0x49, 0x47, 0x46, 0x45, 0x43, 0x3c, 0x3a, 0x39, 0x36, 0x35,
269 	0x34, 0x33, 0x32, 0x31, 0x2e, 0x2d, 0x2c, 0x2b, 0x2a, 0x29,
270 	0x27, 0x26, 0x25, 0x23, 0x1f, 0x1e, 0x1d, 0x1b, 0x18, 0x17,
271 	0x10, 0x0f, 0x08, 0x04, 0x02, 0x01
272 };
273 
274 /* 2200 register offsets */
275 static reg_off_t reg_off_2200 = {
276 	0x00,	/* flash_address */
277 	0x02,	/* flash_data */
278 	0x06,	/* ctrl_status */
279 	0x08,	/* ictrl */
280 	0x0a,	/* istatus */
281 	0x0c,	/* semaphore */
282 	0x0e,	/* nvram */
283 	0x18,	/* req_in */
284 	0x18,	/* req_out */
285 	0x1a,	/* resp_in */
286 	0x1a,	/* resp_out */
287 	0xff,	/* risc2host - n/a */
288 	24,	/* Number of mailboxes */
289 
290 	/* Mailbox in register offsets 0 - 23 */
291 	0x10, 0x12, 0x14, 0x16, 0x18, 0x1a, 0x1c, 0x1e,
292 	0xe0, 0xe2, 0xe4, 0xe6, 0xe8, 0xea, 0xec, 0xee,
293 	0xf0, 0xf2, 0xf4, 0xf6, 0xf8, 0xfa, 0xfc, 0xfe,
294 	/* 2200 does not have mailbox 24-31 - n/a */
295 	0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff,
296 
297 	/* Mailbox out register offsets 0 - 23 */
298 	0x10, 0x12, 0x14, 0x16, 0x18, 0x1a, 0x1c, 0x1e,
299 	0xe0, 0xe2, 0xe4, 0xe6, 0xe8, 0xea, 0xec, 0xee,
300 	0xf0, 0xf2, 0xf4, 0xf6, 0xf8, 0xfa, 0xfc, 0xfe,
301 	/* 2200 does not have mailbox 24-31 - n/a */
302 	0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff,
303 
304 	0x96,	/* fpm_diag_config */
305 	0xa4,	/* pcr */
306 	0xb0,	/* mctr */
307 	0xb8,	/* fb_cmd */
308 	0xc0,	/* hccr */
309 	0xcc,	/* gpiod */
310 	0xce,	/* gpioe */
311 	0xff,	/* host_to_host_sema - n/a */
312 	0xff,	/* pri_req_in - n/a */
313 	0xff,	/* pri_req_out - n/a */
314 	0xff,	/* atio_req_in - n/a */
315 	0xff,	/* atio_req_out - n/a */
316 	0xff,	/* io_base_addr - n/a */
317 	0xff,	/* nx_host_int - n/a */
318 	0xff	/* nx_risc_int - n/a */
319 };
320 
321 /* 2300 register offsets */
322 static reg_off_t reg_off_2300 = {
323 	0x00,	/* flash_address */
324 	0x02,	/* flash_data */
325 	0x06,	/* ctrl_status */
326 	0x08,	/* ictrl */
327 	0x0a,	/* istatus */
328 	0x0c,	/* semaphore */
329 	0x0e,	/* nvram */
330 	0x10,	/* req_in */
331 	0x12,	/* req_out */
332 	0x14,	/* resp_in */
333 	0x16,	/* resp_out */
334 	0x18,	/* risc2host */
335 	32,	/* Number of mailboxes */
336 
337 	/* Mailbox in register offsets 0 - 31 */
338 	0x40, 0x42, 0x44, 0x46, 0x48, 0x4a, 0x4c, 0x4e,
339 	0x50, 0x52, 0x54, 0x56, 0x58, 0x5a, 0x5c, 0x5e,
340 	0x60, 0x62, 0x64, 0x66, 0x68, 0x6a, 0x6c, 0x6e,
341 	0x70, 0x72, 0x74, 0x76, 0x78, 0x7a, 0x7c, 0x7e,
342 
343 	/* Mailbox out register offsets 0 - 31 */
344 	0x40, 0x42, 0x44, 0x46, 0x48, 0x4a, 0x4c, 0x4e,
345 	0x50, 0x52, 0x54, 0x56, 0x58, 0x5a, 0x5c, 0x5e,
346 	0x60, 0x62, 0x64, 0x66, 0x68, 0x6a, 0x6c, 0x6e,
347 	0x70, 0x72, 0x74, 0x76, 0x78, 0x7a, 0x7c, 0x7e,
348 
349 	0x96,	/* fpm_diag_config */
350 	0xa4,	/* pcr */
351 	0xb0,	/* mctr */
352 	0x80,	/* fb_cmd */
353 	0xc0,	/* hccr */
354 	0xcc,	/* gpiod */
355 	0xce,	/* gpioe */
356 	0x1c,	/* host_to_host_sema */
357 	0xff,	/* pri_req_in - n/a */
358 	0xff,	/* pri_req_out - n/a */
359 	0xff,	/* atio_req_in - n/a */
360 	0xff,	/* atio_req_out - n/a */
361 	0xff,	/* io_base_addr - n/a */
362 	0xff,	/* nx_host_int - n/a */
363 	0xff	/* nx_risc_int - n/a */
364 };
365 
366 /* 2400/2500 register offsets */
367 reg_off_t reg_off_2400_2500 = {
368 	0x00,	/* flash_address */
369 	0x04,	/* flash_data */
370 	0x08,	/* ctrl_status */
371 	0x0c,	/* ictrl */
372 	0x10,	/* istatus */
373 	0xff,	/* semaphore - n/a */
374 	0xff,	/* nvram - n/a */
375 	0x1c,	/* req_in */
376 	0x20,	/* req_out */
377 	0x24,	/* resp_in */
378 	0x28,	/* resp_out */
379 	0x44,	/* risc2host */
380 	32,	/* Number of mailboxes */
381 
382 	/* Mailbox in register offsets 0 - 31 */
383 	0x80, 0x82, 0x84, 0x86, 0x88, 0x8a, 0x8c, 0x8e,
384 	0x90, 0x92, 0x94, 0x96, 0x98, 0x9a, 0x9c, 0x9e,
385 	0xa0, 0xa2, 0xa4, 0xa6, 0xa8, 0xaa, 0xac, 0xae,
386 	0xb0, 0xb2, 0xb4, 0xb6, 0xb8, 0xba, 0xbc, 0xbe,
387 
388 	/* Mailbox out register offsets 0 - 31 */
389 	0x80, 0x82, 0x84, 0x86, 0x88, 0x8a, 0x8c, 0x8e,
390 	0x90, 0x92, 0x94, 0x96, 0x98, 0x9a, 0x9c, 0x9e,
391 	0xa0, 0xa2, 0xa4, 0xa6, 0xa8, 0xaa, 0xac, 0xae,
392 	0xb0, 0xb2, 0xb4, 0xb6, 0xb8, 0xba, 0xbc, 0xbe,
393 
394 	0xff,	/* fpm_diag_config  - n/a */
395 	0xff,	/* pcr - n/a */
396 	0xff,	/* mctr - n/a */
397 	0xff,	/* fb_cmd - n/a */
398 	0x48,	/* hccr */
399 	0x4c,	/* gpiod */
400 	0x50,	/* gpioe */
401 	0xff,	/* host_to_host_sema - n/a */
402 	0x2c,	/* pri_req_in */
403 	0x30,	/* pri_req_out */
404 	0x3c,	/* atio_req_in */
405 	0x40,	/* atio_req_out */
406 	0x54,	/* io_base_addr */
407 	0xff,	/* nx_host_int - n/a */
408 	0xff	/* nx_risc_int - n/a */
409 };
410 
411 /* P3 register offsets */
412 static reg_off_t reg_off_8021 = {
413 	0x00,	/* flash_address */
414 	0x04,	/* flash_data */
415 	0x08,	/* ctrl_status */
416 	0x0c,	/* ictrl */
417 	0x10,	/* istatus */
418 	0xff,	/* semaphore - n/a */
419 	0xff,	/* nvram - n/a */
420 	0xff,	/* req_in - n/a */
421 	0x0,	/* req_out */
422 	0x100,	/* resp_in */
423 	0x200,	/* resp_out */
424 	0x500,	/* risc2host */
425 	32,	/* Number of mailboxes */
426 
427 	/* Mailbox in register offsets 0 - 31 */
428 	0x300, 0x302, 0x304, 0x306, 0x308, 0x30a, 0x30c, 0x30e,
429 	0x310, 0x312, 0x314, 0x316, 0x318, 0x31a, 0x31c, 0x31e,
430 	0x320, 0x322, 0x324, 0x326, 0x328, 0x32a, 0x32c, 0x32e,
431 	0x330, 0x332, 0x334, 0x336, 0x338, 0x33a, 0x33c, 0x33e,
432 
433 	/* Mailbox out register offsets 0 - 31 */
434 	0x400, 0x402, 0x404, 0x406, 0x408, 0x40a, 0x40c, 0x40e,
435 	0x410, 0x412, 0x414, 0x416, 0x418, 0x41a, 0x41c, 0x41e,
436 	0x420, 0x422, 0x424, 0x426, 0x428, 0x42a, 0x42c, 0x42e,
437 	0x430, 0x432, 0x434, 0x436, 0x438, 0x43a, 0x43c, 0x43e,
438 
439 	0xff,	/* fpm_diag_config  - n/a */
440 	0xff,	/* pcr - n/a */
441 	0xff,	/* mctr - n/a */
442 	0xff,	/* fb_cmd - n/a */
443 	0x48,	/* hccr */
444 	0x4c,	/* gpiod */
445 	0x50,	/* gpioe */
446 	0xff,	/* host_to_host_sema - n/a */
447 	0x2c,	/* pri_req_in */
448 	0x30,	/* pri_req_out */
449 	0x3c,	/* atio_req_in */
450 	0x40,	/* atio_req_out */
451 	0x54,	/* io_base_addr */
452 	0x380,	/* nx_host_int */
453 	0x504	/* nx_risc_int */
454 };
455 
456 /* mutex for protecting variables shared by all instances of the driver */
457 kmutex_t ql_global_mutex;
458 kmutex_t ql_global_hw_mutex;
459 kmutex_t ql_global_el_mutex;
460 
461 /* DMA access attribute structure. */
462 static ddi_device_acc_attr_t ql_dev_acc_attr = {
463 	DDI_DEVICE_ATTR_V0,
464 	DDI_STRUCTURE_LE_ACC,
465 	DDI_STRICTORDER_ACC
466 };
467 
468 /* I/O DMA attributes structures. */
469 static ddi_dma_attr_t ql_64bit_io_dma_attr = {
470 	DMA_ATTR_V0,			/* dma_attr_version */
471 	QL_DMA_LOW_ADDRESS,		/* low DMA address range */
472 	QL_DMA_HIGH_64BIT_ADDRESS,	/* high DMA address range */
473 	QL_DMA_XFER_COUNTER,		/* DMA counter register */
474 	QL_DMA_ADDRESS_ALIGNMENT,	/* DMA address alignment */
475 	QL_DMA_BURSTSIZES,		/* DMA burstsizes */
476 	QL_DMA_MIN_XFER_SIZE,		/* min effective DMA size */
477 	QL_DMA_MAX_XFER_SIZE,		/* max DMA xfer size */
478 	QL_DMA_SEGMENT_BOUNDARY,	/* segment boundary */
479 	QL_DMA_SG_LIST_LENGTH,		/* s/g list length */
480 	QL_DMA_GRANULARITY,		/* granularity of device */
481 	QL_DMA_XFER_FLAGS		/* DMA transfer flags */
482 };
483 
484 static ddi_dma_attr_t ql_32bit_io_dma_attr = {
485 	DMA_ATTR_V0,			/* dma_attr_version */
486 	QL_DMA_LOW_ADDRESS,		/* low DMA address range */
487 	QL_DMA_HIGH_32BIT_ADDRESS,	/* high DMA address range */
488 	QL_DMA_XFER_COUNTER,		/* DMA counter register */
489 	QL_DMA_ADDRESS_ALIGNMENT,	/* DMA address alignment */
490 	QL_DMA_BURSTSIZES,		/* DMA burstsizes */
491 	QL_DMA_MIN_XFER_SIZE,		/* min effective DMA size */
492 	QL_DMA_MAX_XFER_SIZE,		/* max DMA xfer size */
493 	QL_DMA_SEGMENT_BOUNDARY,	/* segment boundary */
494 	QL_DMA_SG_LIST_LENGTH,		/* s/g list length */
495 	QL_DMA_GRANULARITY,		/* granularity of device */
496 	QL_DMA_XFER_FLAGS		/* DMA transfer flags */
497 };
498 
499 /* Load the default dma attributes */
500 static	ddi_dma_attr_t	ql_32fcsm_cmd_dma_attr;
501 static	ddi_dma_attr_t	ql_64fcsm_cmd_dma_attr;
502 static	ddi_dma_attr_t	ql_32fcsm_rsp_dma_attr;
503 static	ddi_dma_attr_t	ql_64fcsm_rsp_dma_attr;
504 static	ddi_dma_attr_t	ql_32fcip_cmd_dma_attr;
505 static	ddi_dma_attr_t	ql_64fcip_cmd_dma_attr;
506 static	ddi_dma_attr_t	ql_32fcip_rsp_dma_attr;
507 static	ddi_dma_attr_t	ql_64fcip_rsp_dma_attr;
508 static	ddi_dma_attr_t	ql_32fcp_cmd_dma_attr;
509 static	ddi_dma_attr_t	ql_64fcp_cmd_dma_attr;
510 static	ddi_dma_attr_t	ql_32fcp_rsp_dma_attr;
511 static	ddi_dma_attr_t	ql_64fcp_rsp_dma_attr;
512 static	ddi_dma_attr_t	ql_32fcp_data_dma_attr;
513 static	ddi_dma_attr_t	ql_64fcp_data_dma_attr;
514 
515 /* Static declarations of cb_ops entry point functions... */
516 static struct cb_ops ql_cb_ops = {
517 	ql_open,			/* b/c open */
518 	ql_close,			/* b/c close */
519 	nodev,				/* b strategy */
520 	nodev,				/* b print */
521 	nodev,				/* b dump */
522 	nodev,				/* c read */
523 	nodev,				/* c write */
524 	ql_ioctl,			/* c ioctl */
525 	nodev,				/* c devmap */
526 	nodev,				/* c mmap */
527 	nodev,				/* c segmap */
528 	nochpoll,			/* c poll */
529 	nodev,				/* cb_prop_op */
530 	NULL,				/* streamtab  */
531 	D_MP | D_NEW | D_HOTPLUG,	/* Driver compatibility flag */
532 	CB_REV,				/* cb_ops revision */
533 	nodev,				/* c aread */
534 	nodev				/* c awrite */
535 };
536 
537 /* Static declarations of dev_ops entry point functions... */
538 static struct dev_ops ql_devops = {
539 	DEVO_REV,			/* devo_rev */
540 	0,				/* refcnt */
541 	ql_getinfo,			/* getinfo */
542 	nulldev,			/* identify */
543 	nulldev,			/* probe */
544 	ql_attach,			/* attach */
545 	ql_detach,			/* detach */
546 	nodev,				/* reset */
547 	&ql_cb_ops,			/* char/block ops */
548 	NULL,				/* bus operations */
549 	ql_power,			/* power management */
550 	ql_quiesce			/* quiesce device */
551 };
552 
553 /* ELS command code to text converter */
554 cmd_table_t els_cmd_tbl[] = ELS_CMD_TABLE();
555 /* Mailbox command code to text converter */
556 cmd_table_t mbox_cmd_tbl[] = MBOX_CMD_TABLE();
557 
558 char qlc_driver_version[] = QL_VERSION;
559 
560 /*
561  * Loadable Driver Interface Structures.
562  * Declare and initialize the module configuration section...
563  */
564 static struct modldrv modldrv = {
565 	&mod_driverops,				/* type of module: driver */
566 	"SunFC Qlogic FCA v" QL_VERSION,	/* name of module */
567 	&ql_devops				/* driver dev_ops */
568 };
569 
570 static struct modlinkage modlinkage = {
571 	MODREV_1,
572 	&modldrv,
573 	NULL
574 };
575 
576 /* ************************************************************************ */
577 /*				Loadable Module Routines.		    */
578 /* ************************************************************************ */
579 
580 /*
581  * _init
582  *	Initializes a loadable module. It is called before any other
583  *	routine in a loadable module.
584  *
585  * Returns:
586  *	0 = success
587  *
588  * Context:
589  *	Kernel context.
590  */
591 int
_init(void)592 _init(void)
593 {
594 	uint16_t	w16;
595 	int		rval = 0;
596 
597 	/* Get OS major release level. */
598 	for (w16 = 0; w16 < sizeof (utsname.release); w16++) {
599 		if (utsname.release[w16] == '.') {
600 			w16++;
601 			break;
602 		}
603 	}
604 	if (w16 < sizeof (utsname.release)) {
605 		(void) ql_bstr_to_dec(&utsname.release[w16],
606 		    &ql_os_release_level, 0);
607 	} else {
608 		ql_os_release_level = 0;
609 	}
610 	if (ql_os_release_level < 6) {
611 		cmn_err(CE_WARN, "%s Unsupported OS release level = %d",
612 		    QL_NAME, ql_os_release_level);
613 		rval = EINVAL;
614 	}
615 	if (ql_os_release_level == 6) {
616 		ql_32bit_io_dma_attr.dma_attr_count_max = 0x00ffffff;
617 		ql_64bit_io_dma_attr.dma_attr_count_max = 0x00ffffff;
618 	}
619 
620 	if (rval == 0) {
621 		rval = ddi_soft_state_init(&ql_state,
622 		    sizeof (ql_adapter_state_t), 0);
623 	}
624 	if (rval == 0) {
625 		/* allow the FC Transport to tweak the dev_ops */
626 		fc_fca_init(&ql_devops);
627 
628 		mutex_init(&ql_global_mutex, NULL, MUTEX_DRIVER, NULL);
629 		mutex_init(&ql_global_hw_mutex, NULL, MUTEX_DRIVER, NULL);
630 		mutex_init(&ql_global_el_mutex, NULL, MUTEX_DRIVER, NULL);
631 		rval = mod_install(&modlinkage);
632 		if (rval != 0) {
633 			mutex_destroy(&ql_global_hw_mutex);
634 			mutex_destroy(&ql_global_mutex);
635 			mutex_destroy(&ql_global_el_mutex);
636 			ddi_soft_state_fini(&ql_state);
637 		} else {
638 			/*EMPTY*/
639 			ql_32fcsm_cmd_dma_attr = ql_32bit_io_dma_attr;
640 			ql_64fcsm_cmd_dma_attr = ql_64bit_io_dma_attr;
641 			ql_32fcsm_rsp_dma_attr = ql_32bit_io_dma_attr;
642 			ql_64fcsm_rsp_dma_attr = ql_64bit_io_dma_attr;
643 			ql_32fcip_cmd_dma_attr = ql_32bit_io_dma_attr;
644 			ql_64fcip_cmd_dma_attr = ql_64bit_io_dma_attr;
645 			ql_32fcip_rsp_dma_attr = ql_32bit_io_dma_attr;
646 			ql_64fcip_rsp_dma_attr = ql_64bit_io_dma_attr;
647 			ql_32fcp_cmd_dma_attr = ql_32bit_io_dma_attr;
648 			ql_64fcp_cmd_dma_attr = ql_64bit_io_dma_attr;
649 			ql_32fcp_rsp_dma_attr = ql_32bit_io_dma_attr;
650 			ql_64fcp_rsp_dma_attr = ql_64bit_io_dma_attr;
651 			ql_32fcp_data_dma_attr = ql_32bit_io_dma_attr;
652 			ql_64fcp_data_dma_attr = ql_64bit_io_dma_attr;
653 			ql_32fcsm_cmd_dma_attr.dma_attr_sgllen =
654 			    ql_64fcsm_cmd_dma_attr.dma_attr_sgllen =
655 			    QL_FCSM_CMD_SGLLEN;
656 			ql_32fcsm_rsp_dma_attr.dma_attr_sgllen =
657 			    ql_64fcsm_rsp_dma_attr.dma_attr_sgllen =
658 			    QL_FCSM_RSP_SGLLEN;
659 			ql_32fcip_cmd_dma_attr.dma_attr_sgllen =
660 			    ql_64fcip_cmd_dma_attr.dma_attr_sgllen =
661 			    QL_FCIP_CMD_SGLLEN;
662 			ql_32fcip_rsp_dma_attr.dma_attr_sgllen =
663 			    ql_64fcip_rsp_dma_attr.dma_attr_sgllen =
664 			    QL_FCIP_RSP_SGLLEN;
665 			ql_32fcp_cmd_dma_attr.dma_attr_sgllen =
666 			    ql_64fcp_cmd_dma_attr.dma_attr_sgllen =
667 			    QL_FCP_CMD_SGLLEN;
668 			ql_32fcp_rsp_dma_attr.dma_attr_sgllen =
669 			    ql_64fcp_rsp_dma_attr.dma_attr_sgllen =
670 			    QL_FCP_RSP_SGLLEN;
671 		}
672 	}
673 
674 	if (rval != 0) {
675 		cmn_err(CE_CONT, "?Unable to install/attach driver '%s'",
676 		    QL_NAME);
677 	}
678 
679 	return (rval);
680 }
681 
682 /*
683  * _fini
684  *	Prepares a module for unloading. It is called when the system
685  *	wants to unload a module. If the module determines that it can
686  *	be unloaded, then _fini() returns the value returned by
687  *	mod_remove(). Upon successful return from _fini() no other
688  *	routine in the module will be called before _init() is called.
689  *
690  * Returns:
691  *	0 = success
692  *
693  * Context:
694  *	Kernel context.
695  */
696 int
_fini(void)697 _fini(void)
698 {
699 	int	rval;
700 
701 	rval = mod_remove(&modlinkage);
702 	if (rval == 0) {
703 		mutex_destroy(&ql_global_hw_mutex);
704 		mutex_destroy(&ql_global_mutex);
705 		mutex_destroy(&ql_global_el_mutex);
706 		ddi_soft_state_fini(&ql_state);
707 	}
708 
709 	return (rval);
710 }
711 
712 /*
713  * _info
714  *	Returns information about loadable module.
715  *
716  * Input:
717  *	modinfo = pointer to module information structure.
718  *
719  * Returns:
720  *	Value returned by mod_info().
721  *
722  * Context:
723  *	Kernel context.
724  */
725 int
_info(struct modinfo * modinfop)726 _info(struct modinfo *modinfop)
727 {
728 	return (mod_info(&modlinkage, modinfop));
729 }
730 
731 /* ************************************************************************ */
732 /*			dev_ops functions				    */
733 /* ************************************************************************ */
734 
735 /*
736  * ql_getinfo
737  *	Returns the pointer associated with arg when cmd is
738  *	set to DDI_INFO_DEVT2DEVINFO, or it should return the
739  *	instance number associated with arg when cmd is set
740  *	to DDI_INFO_DEV2INSTANCE.
741  *
742  * Input:
743  *	dip = Do not use.
744  *	cmd = command argument.
745  *	arg = command specific argument.
746  *	resultp = pointer to where request information is stored.
747  *
748  * Returns:
749  *	DDI_SUCCESS or DDI_FAILURE.
750  *
751  * Context:
752  *	Kernel context.
753  */
754 /* ARGSUSED */
755 static int
ql_getinfo(dev_info_t * dip,ddi_info_cmd_t cmd,void * arg,void ** resultp)756 ql_getinfo(dev_info_t *dip, ddi_info_cmd_t cmd, void *arg, void **resultp)
757 {
758 	ql_adapter_state_t	*ha;
759 	int			minor;
760 	int			rval = DDI_FAILURE;
761 
762 	minor = (int)(getminor((dev_t)arg));
763 	ha = ddi_get_soft_state(ql_state, minor);
764 	if (ha == NULL) {
765 		QL_PRINT_2(CE_CONT, "failed, unknown minor=%d\n",
766 		    getminor((dev_t)arg));
767 		*resultp = NULL;
768 		return (rval);
769 	}
770 
771 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
772 
773 	switch (cmd) {
774 	case DDI_INFO_DEVT2DEVINFO:
775 		*resultp = ha->dip;
776 		rval = DDI_SUCCESS;
777 		break;
778 	case DDI_INFO_DEVT2INSTANCE:
779 		*resultp = (void *)(uintptr_t)(ha->instance);
780 		rval = DDI_SUCCESS;
781 		break;
782 	default:
783 		EL(ha, "failed, unsupported cmd=%d\n", cmd);
784 		rval = DDI_FAILURE;
785 		break;
786 	}
787 
788 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
789 
790 	return (rval);
791 }
792 
793 /*
794  * ql_attach
795  *	Configure and attach an instance of the driver
796  *	for a port.
797  *
798  * Input:
799  *	dip = pointer to device information structure.
800  *	cmd = attach type.
801  *
802  * Returns:
803  *	DDI_SUCCESS or DDI_FAILURE.
804  *
805  * Context:
806  *	Kernel context.
807  */
808 static int
ql_attach(dev_info_t * dip,ddi_attach_cmd_t cmd)809 ql_attach(dev_info_t *dip, ddi_attach_cmd_t cmd)
810 {
811 	off_t			regsize;
812 	uint32_t		size;
813 	int			rval, *ptr;
814 	int			instance;
815 	uint_t			progress = 0;
816 	char			*buf;
817 	ushort_t		caps_ptr, cap;
818 	fc_fca_tran_t		*tran;
819 	ql_adapter_state_t	*ha = NULL;
820 
821 	static char *pmcomps[] = {
822 		NULL,
823 		PM_LEVEL_D3_STR,		/* Device OFF */
824 		PM_LEVEL_D0_STR,		/* Device ON */
825 	};
826 
827 	QL_PRINT_3(CE_CONT, "(%d): started, cmd=%xh\n",
828 	    ddi_get_instance(dip), cmd);
829 
830 	buf = (char *)(kmem_zalloc(MAXPATHLEN, KM_SLEEP));
831 
832 	switch (cmd) {
833 	case DDI_ATTACH:
834 		/* first get the instance */
835 		instance = ddi_get_instance(dip);
836 
837 		cmn_err(CE_CONT, "!Qlogic %s(%d) FCA Driver v%s\n",
838 		    QL_NAME, instance, QL_VERSION);
839 
840 		/* Correct OS version? */
841 		if (ql_os_release_level != 11) {
842 			cmn_err(CE_WARN, "%s(%d): This driver is for Solaris "
843 			    "11", QL_NAME, instance);
844 			goto attach_failed;
845 		}
846 
847 		/* Hardware is installed in a DMA-capable slot? */
848 		if (ddi_slaveonly(dip) == DDI_SUCCESS) {
849 			cmn_err(CE_WARN, "%s(%d): slave only", QL_NAME,
850 			    instance);
851 			goto attach_failed;
852 		}
853 
854 		/* No support for high-level interrupts */
855 		if (ddi_intr_hilevel(dip, 0) != 0) {
856 			cmn_err(CE_WARN, "%s(%d): High level interrupt"
857 			    " not supported", QL_NAME, instance);
858 			goto attach_failed;
859 		}
860 
861 		/* Allocate our per-device-instance structure */
862 		if (ddi_soft_state_zalloc(ql_state,
863 		    instance) != DDI_SUCCESS) {
864 			cmn_err(CE_WARN, "%s(%d): soft state alloc failed",
865 			    QL_NAME, instance);
866 			goto attach_failed;
867 		}
868 		progress |= QL_SOFT_STATE_ALLOCED;
869 
870 		ha = ddi_get_soft_state(ql_state, instance);
871 		if (ha == NULL) {
872 			cmn_err(CE_WARN, "%s(%d): can't get soft state",
873 			    QL_NAME, instance);
874 			goto attach_failed;
875 		}
876 		ha->dip = dip;
877 		ha->instance = instance;
878 		ha->hba.base_address = ha;
879 		ha->pha = ha;
880 
881 		if (ql_el_trace_desc_ctor(ha) != DDI_SUCCESS) {
882 			cmn_err(CE_WARN, "%s(%d): can't setup el tracing",
883 			    QL_NAME, instance);
884 			goto attach_failed;
885 		}
886 
887 		/* Get extended logging and dump flags. */
888 		ql_common_properties(ha);
889 
890 		if (strcmp(ddi_driver_name(ddi_get_parent(dip)),
891 		    "sbus") == 0) {
892 			EL(ha, "%s SBUS card detected", QL_NAME);
893 			ha->cfg_flags |= CFG_SBUS_CARD;
894 		}
895 
896 		ha->dev = kmem_zalloc(sizeof (*ha->dev) *
897 		    DEVICE_HEAD_LIST_SIZE, KM_SLEEP);
898 
899 		ha->outstanding_cmds = kmem_zalloc(
900 		    sizeof (*ha->outstanding_cmds) * MAX_OUTSTANDING_COMMANDS,
901 		    KM_SLEEP);
902 
903 		ha->ub_array = kmem_zalloc(sizeof (*ha->ub_array) *
904 		    QL_UB_LIMIT, KM_SLEEP);
905 
906 		ha->adapter_stats = kmem_zalloc(sizeof (*ha->adapter_stats),
907 		    KM_SLEEP);
908 
909 		(void) ddi_pathname(dip, buf);
910 		ha->devpath = kmem_zalloc(strlen(buf)+1, KM_SLEEP);
911 		if (ha->devpath == NULL) {
912 			EL(ha, "devpath mem alloc failed\n");
913 		} else {
914 			(void) strcpy(ha->devpath, buf);
915 			EL(ha, "devpath is: %s\n", ha->devpath);
916 		}
917 
918 		if (CFG_IST(ha, CFG_SBUS_CARD)) {
919 			/*
920 			 * For cards where PCI is mapped to sbus e.g. Ivory.
921 			 *
922 			 * 0x00	: 0x000 - 0x0FF PCI Config Space for 2200
923 			 *	: 0x100 - 0x3FF PCI IO space for 2200
924 			 * 0x01	: 0x000 - 0x0FF PCI Config Space for fpga
925 			 *	: 0x100 - 0x3FF PCI IO Space for fpga
926 			 */
927 			if (ddi_regs_map_setup(dip, 0, (caddr_t *)&ha->iobase,
928 			    0x100, 0x300, &ql_dev_acc_attr, &ha->dev_handle) !=
929 			    DDI_SUCCESS) {
930 				cmn_err(CE_WARN, "%s(%d): Unable to map device"
931 				    " registers", QL_NAME, instance);
932 				goto attach_failed;
933 			}
934 			if (ddi_regs_map_setup(dip, 1,
935 			    (caddr_t *)&ha->sbus_fpga_iobase, 0, 0x400,
936 			    &ql_dev_acc_attr, &ha->sbus_fpga_dev_handle) !=
937 			    DDI_SUCCESS) {
938 				/* We should not fail attach here */
939 				cmn_err(CE_WARN, "%s(%d): Unable to map FPGA",
940 				    QL_NAME, instance);
941 				ha->sbus_fpga_iobase = NULL;
942 			}
943 			progress |= QL_REGS_MAPPED;
944 
945 			/*
946 			 * We should map config space before adding interrupt
947 			 * So that the chip type (2200 or 2300) can be
948 			 * determined before the interrupt routine gets a
949 			 * chance to execute.
950 			 */
951 			if (ddi_regs_map_setup(dip, 0,
952 			    (caddr_t *)&ha->sbus_config_base, 0, 0x100,
953 			    &ql_dev_acc_attr, &ha->sbus_config_handle) !=
954 			    DDI_SUCCESS) {
955 				cmn_err(CE_WARN, "%s(%d): Unable to map sbus "
956 				    "config registers", QL_NAME, instance);
957 				goto attach_failed;
958 			}
959 			progress |= QL_CONFIG_SPACE_SETUP;
960 		} else {
961 			/*LINTED [Solaris DDI_DEV_T_ANY Lint error]*/
962 			rval = ddi_prop_lookup_int_array(DDI_DEV_T_ANY, dip,
963 			    DDI_PROP_DONTPASS, "reg", &ptr, &size);
964 			if (rval != DDI_PROP_SUCCESS) {
965 				cmn_err(CE_WARN, "%s(%d): Unable to get PCI "
966 				    "address registers", QL_NAME, instance);
967 				goto attach_failed;
968 			} else {
969 				ha->pci_bus_addr = ptr[0];
970 				ha->function_number = (uint8_t)
971 				    (ha->pci_bus_addr >> 8 & 7);
972 				ddi_prop_free(ptr);
973 			}
974 
975 			/*
976 			 * We should map config space before adding interrupt
977 			 * So that the chip type (2200 or 2300) can be
978 			 * determined before the interrupt routine gets a
979 			 * chance to execute.
980 			 */
981 			if (pci_config_setup(ha->dip, &ha->pci_handle) !=
982 			    DDI_SUCCESS) {
983 				cmn_err(CE_WARN, "%s(%d): can't setup PCI "
984 				    "config space", QL_NAME, instance);
985 				goto attach_failed;
986 			}
987 			progress |= QL_CONFIG_SPACE_SETUP;
988 
989 			/*
990 			 * Setup the ISP2200 registers address mapping to be
991 			 * accessed by this particular driver.
992 			 * 0x0   Configuration Space
993 			 * 0x1   I/O Space
994 			 * 0x2   32-bit Memory Space address
995 			 * 0x3   64-bit Memory Space address
996 			 */
997 			size = ql_pci_config_get32(ha, PCI_CONF_BASE0) & BIT_0 ?
998 			    2 : 1;
999 			if (ddi_dev_regsize(dip, size, &regsize) !=
1000 			    DDI_SUCCESS ||
1001 			    ddi_regs_map_setup(dip, size, &ha->iobase,
1002 			    0, regsize, &ql_dev_acc_attr, &ha->dev_handle) !=
1003 			    DDI_SUCCESS) {
1004 				cmn_err(CE_WARN, "%s(%d): regs_map_setup(mem) "
1005 				    "failed", QL_NAME, instance);
1006 				goto attach_failed;
1007 			}
1008 			progress |= QL_REGS_MAPPED;
1009 
1010 			/*
1011 			 * We need I/O space mappings for 23xx HBAs for
1012 			 * loading flash (FCode). The chip has a bug due to
1013 			 * which loading flash fails through mem space
1014 			 * mappings in PCI-X mode.
1015 			 */
1016 			if (size == 1) {
1017 				ha->iomap_iobase = ha->iobase;
1018 				ha->iomap_dev_handle = ha->dev_handle;
1019 			} else {
1020 				if (ddi_dev_regsize(dip, 1, &regsize) !=
1021 				    DDI_SUCCESS ||
1022 				    ddi_regs_map_setup(dip, 1,
1023 				    &ha->iomap_iobase, 0, regsize,
1024 				    &ql_dev_acc_attr, &ha->iomap_dev_handle) !=
1025 				    DDI_SUCCESS) {
1026 					cmn_err(CE_WARN, "%s(%d): regs_map_"
1027 					    "setup(I/O) failed", QL_NAME,
1028 					    instance);
1029 					goto attach_failed;
1030 				}
1031 				progress |= QL_IOMAP_IOBASE_MAPPED;
1032 			}
1033 		}
1034 
1035 		ha->subsys_id = (uint16_t)ql_pci_config_get16(ha,
1036 		    PCI_CONF_SUBSYSID);
1037 		ha->subven_id = (uint16_t)ql_pci_config_get16(ha,
1038 		    PCI_CONF_SUBVENID);
1039 		ha->ven_id = (uint16_t)ql_pci_config_get16(ha,
1040 		    PCI_CONF_VENID);
1041 		ha->device_id = (uint16_t)ql_pci_config_get16(ha,
1042 		    PCI_CONF_DEVID);
1043 		ha->rev_id = (uint8_t)ql_pci_config_get8(ha,
1044 		    PCI_CONF_REVID);
1045 
1046 		EL(ha, "ISP%x chip detected (RevID=%x, VenID=%x, SVenID=%x, "
1047 		    "SSysID=%x)\n", ha->device_id, ha->rev_id, ha->ven_id,
1048 		    ha->subven_id, ha->subsys_id);
1049 
1050 		switch (ha->device_id) {
1051 		case 0x2300:
1052 		case 0x2312:
1053 		case 0x2322:
1054 		case 0x6312:
1055 		case 0x6322:
1056 			if (ql_pci_config_get8(ha, PCI_CONF_IPIN) == 2) {
1057 				ha->flags |= FUNCTION_1;
1058 			}
1059 			if ((ha->device_id == 0x6322) ||
1060 			    (ha->device_id == 0x2322)) {
1061 				ha->cfg_flags |= CFG_CTRL_6322;
1062 				ha->fw_class = 0x6322;
1063 				ha->risc_dump_size = QL_6322_FW_DUMP_SIZE;
1064 			} else {
1065 				ha->cfg_flags |= CFG_CTRL_2300;
1066 				ha->fw_class = 0x2300;
1067 				ha->risc_dump_size = QL_2300_FW_DUMP_SIZE;
1068 			}
1069 			ha->reg_off = &reg_off_2300;
1070 			if (ql_fwmodule_resolve(ha) != QL_SUCCESS) {
1071 				goto attach_failed;
1072 			}
1073 			ha->fcp_cmd = ql_command_iocb;
1074 			ha->ip_cmd = ql_ip_iocb;
1075 			ha->ms_cmd = ql_ms_iocb;
1076 			if (CFG_IST(ha, CFG_SBUS_CARD)) {
1077 				ha->cmd_segs = CMD_TYPE_2_DATA_SEGMENTS;
1078 				ha->cmd_cont_segs = CONT_TYPE_0_DATA_SEGMENTS;
1079 			} else {
1080 				ha->cmd_segs = CMD_TYPE_3_DATA_SEGMENTS;
1081 				ha->cmd_cont_segs = CONT_TYPE_1_DATA_SEGMENTS;
1082 			}
1083 			break;
1084 
1085 		case 0x2200:
1086 			ha->cfg_flags |= CFG_CTRL_2200;
1087 			ha->reg_off = &reg_off_2200;
1088 			ha->fw_class = 0x2200;
1089 			if (ql_fwmodule_resolve(ha) != QL_SUCCESS) {
1090 				goto attach_failed;
1091 			}
1092 			ha->risc_dump_size = QL_2200_FW_DUMP_SIZE;
1093 			ha->fcp_cmd = ql_command_iocb;
1094 			ha->ip_cmd = ql_ip_iocb;
1095 			ha->ms_cmd = ql_ms_iocb;
1096 			if (CFG_IST(ha, CFG_SBUS_CARD)) {
1097 				ha->cmd_segs = CMD_TYPE_2_DATA_SEGMENTS;
1098 				ha->cmd_cont_segs = CONT_TYPE_0_DATA_SEGMENTS;
1099 			} else {
1100 				ha->cmd_segs = CMD_TYPE_3_DATA_SEGMENTS;
1101 				ha->cmd_cont_segs = CONT_TYPE_1_DATA_SEGMENTS;
1102 			}
1103 			break;
1104 
1105 		case 0x2422:
1106 		case 0x2432:
1107 		case 0x5422:
1108 		case 0x5432:
1109 		case 0x8432:
1110 			if (ql_pci_config_get8(ha, PCI_CONF_IPIN) == 2) {
1111 				ha->flags |= FUNCTION_1;
1112 			}
1113 			ha->cfg_flags |= CFG_CTRL_2422;
1114 			if (ha->device_id == 0x8432) {
1115 				ha->cfg_flags |= CFG_CTRL_MENLO;
1116 			} else {
1117 				ha->flags |= VP_ENABLED;
1118 			}
1119 
1120 			ha->reg_off = &reg_off_2400_2500;
1121 			ha->fw_class = 0x2400;
1122 			if (ql_fwmodule_resolve(ha) != QL_SUCCESS) {
1123 				goto attach_failed;
1124 			}
1125 			ha->risc_dump_size = QL_24XX_FW_DUMP_SIZE;
1126 			ha->fcp_cmd = ql_command_24xx_iocb;
1127 			ha->ip_cmd = ql_ip_24xx_iocb;
1128 			ha->ms_cmd = ql_ms_24xx_iocb;
1129 			ha->els_cmd = ql_els_24xx_iocb;
1130 			ha->cmd_segs = CMD_TYPE_7_DATA_SEGMENTS;
1131 			ha->cmd_cont_segs = CONT_TYPE_1_DATA_SEGMENTS;
1132 			break;
1133 
1134 		case 0x2522:
1135 		case 0x2532:
1136 			if (ql_pci_config_get8(ha, PCI_CONF_IPIN) == 2) {
1137 				ha->flags |= FUNCTION_1;
1138 			}
1139 			ha->cfg_flags |= CFG_CTRL_25XX;
1140 			ha->flags |= VP_ENABLED;
1141 			ha->fw_class = 0x2500;
1142 			ha->reg_off = &reg_off_2400_2500;
1143 			if (ql_fwmodule_resolve(ha) != QL_SUCCESS) {
1144 				goto attach_failed;
1145 			}
1146 			ha->risc_dump_size = QL_25XX_FW_DUMP_SIZE;
1147 			ha->fcp_cmd = ql_command_24xx_iocb;
1148 			ha->ip_cmd = ql_ip_24xx_iocb;
1149 			ha->ms_cmd = ql_ms_24xx_iocb;
1150 			ha->els_cmd = ql_els_24xx_iocb;
1151 			ha->cmd_segs = CMD_TYPE_7_DATA_SEGMENTS;
1152 			ha->cmd_cont_segs = CONT_TYPE_1_DATA_SEGMENTS;
1153 			break;
1154 
1155 		case 0x8001:
1156 			if (ql_pci_config_get8(ha, PCI_CONF_IPIN) == 4) {
1157 				ha->flags |= FUNCTION_1;
1158 			}
1159 			ha->cfg_flags |= CFG_CTRL_81XX;
1160 			ha->flags |= VP_ENABLED;
1161 			ha->fw_class = 0x8100;
1162 			ha->reg_off = &reg_off_2400_2500;
1163 			if (ql_fwmodule_resolve(ha) != QL_SUCCESS) {
1164 				goto attach_failed;
1165 			}
1166 			ha->risc_dump_size = QL_25XX_FW_DUMP_SIZE;
1167 			ha->fcp_cmd = ql_command_24xx_iocb;
1168 			ha->ip_cmd = ql_ip_24xx_iocb;
1169 			ha->ms_cmd = ql_ms_24xx_iocb;
1170 			ha->cmd_segs = CMD_TYPE_7_DATA_SEGMENTS;
1171 			ha->cmd_cont_segs = CONT_TYPE_1_DATA_SEGMENTS;
1172 			break;
1173 
1174 		case 0x8021:
1175 			if (ha->function_number & BIT_0) {
1176 				ha->flags |= FUNCTION_1;
1177 			}
1178 			ha->cfg_flags |= CFG_CTRL_8021;
1179 			ha->reg_off = &reg_off_8021;
1180 			ha->risc_dump_size = QL_25XX_FW_DUMP_SIZE;
1181 			ha->fcp_cmd = ql_command_24xx_iocb;
1182 			ha->ms_cmd = ql_ms_24xx_iocb;
1183 			ha->cmd_segs = CMD_TYPE_7_DATA_SEGMENTS;
1184 			ha->cmd_cont_segs = CONT_TYPE_1_DATA_SEGMENTS;
1185 
1186 			ha->nx_pcibase = ha->iobase;
1187 			ha->iobase += 0xBC000 + (ha->function_number << 11);
1188 			ha->iomap_iobase += 0xBC000 +
1189 			    (ha->function_number << 11);
1190 
1191 			/* map doorbell */
1192 			if (ddi_dev_regsize(dip, 2, &regsize) != DDI_SUCCESS ||
1193 			    ddi_regs_map_setup(dip, 2, &ha->db_iobase,
1194 			    0, regsize, &ql_dev_acc_attr, &ha->db_dev_handle) !=
1195 			    DDI_SUCCESS) {
1196 				cmn_err(CE_WARN, "%s(%d): regs_map_setup"
1197 				    "(doorbell) failed", QL_NAME, instance);
1198 				goto attach_failed;
1199 			}
1200 			progress |= QL_DB_IOBASE_MAPPED;
1201 
1202 			ha->nx_req_in = (uint32_t *)(ha->db_iobase +
1203 			    (ha->function_number << 12));
1204 			ha->db_read = ha->nx_pcibase + (512 * 1024) +
1205 			    (ha->function_number * 8);
1206 
1207 			ql_8021_update_crb_int_ptr(ha);
1208 			ql_8021_set_drv_active(ha);
1209 			break;
1210 
1211 		default:
1212 			cmn_err(CE_WARN, "%s(%d): Unsupported device id: %x",
1213 			    QL_NAME, instance, ha->device_id);
1214 			goto attach_failed;
1215 		}
1216 
1217 		/* Setup hba buffer. */
1218 
1219 		size = CFG_IST(ha, CFG_CTRL_24258081) ?
1220 		    (REQUEST_QUEUE_SIZE + RESPONSE_QUEUE_SIZE) :
1221 		    (REQUEST_QUEUE_SIZE + RESPONSE_QUEUE_SIZE +
1222 		    RCVBUF_QUEUE_SIZE);
1223 
1224 		if (ql_get_dma_mem(ha, &ha->hba_buf, size, LITTLE_ENDIAN_DMA,
1225 		    QL_DMA_RING_ALIGN) != QL_SUCCESS) {
1226 			cmn_err(CE_WARN, "%s(%d): request queue DMA memory "
1227 			    "alloc failed", QL_NAME, instance);
1228 			goto attach_failed;
1229 		}
1230 		progress |= QL_HBA_BUFFER_SETUP;
1231 
1232 		/* Setup buffer pointers. */
1233 		ha->request_dvma = ha->hba_buf.cookie.dmac_laddress +
1234 		    REQUEST_Q_BUFFER_OFFSET;
1235 		ha->request_ring_bp = (struct cmd_entry *)
1236 		    ((caddr_t)ha->hba_buf.bp + REQUEST_Q_BUFFER_OFFSET);
1237 
1238 		ha->response_dvma = ha->hba_buf.cookie.dmac_laddress +
1239 		    RESPONSE_Q_BUFFER_OFFSET;
1240 		ha->response_ring_bp = (struct sts_entry *)
1241 		    ((caddr_t)ha->hba_buf.bp + RESPONSE_Q_BUFFER_OFFSET);
1242 
1243 		ha->rcvbuf_dvma = ha->hba_buf.cookie.dmac_laddress +
1244 		    RCVBUF_Q_BUFFER_OFFSET;
1245 		ha->rcvbuf_ring_bp = (struct rcvbuf *)
1246 		    ((caddr_t)ha->hba_buf.bp + RCVBUF_Q_BUFFER_OFFSET);
1247 
1248 		/* Allocate resource for QLogic IOCTL */
1249 		(void) ql_alloc_xioctl_resource(ha);
1250 
1251 		/* Setup interrupts */
1252 		if ((rval = ql_setup_interrupts(ha)) != DDI_SUCCESS) {
1253 			cmn_err(CE_WARN, "%s(%d): Failed to add interrupt, "
1254 			    "rval=%xh", QL_NAME, instance, rval);
1255 			goto attach_failed;
1256 		}
1257 
1258 		progress |= (QL_INTR_ADDED | QL_MUTEX_CV_INITED);
1259 
1260 		if (ql_nvram_cache_desc_ctor(ha) != DDI_SUCCESS) {
1261 			cmn_err(CE_WARN, "%s(%d): can't setup nvram cache",
1262 			    QL_NAME, instance);
1263 			goto attach_failed;
1264 		}
1265 
1266 		/*
1267 		 * Allocate an N Port information structure
1268 		 * for use when in P2P topology.
1269 		 */
1270 		ha->n_port = (ql_n_port_info_t *)
1271 		    kmem_zalloc(sizeof (ql_n_port_info_t), KM_SLEEP);
1272 		if (ha->n_port == NULL) {
1273 			cmn_err(CE_WARN, "%s(%d): Failed to create N Port info",
1274 			    QL_NAME, instance);
1275 			goto attach_failed;
1276 		}
1277 
1278 		progress |= QL_N_PORT_INFO_CREATED;
1279 
1280 		/*
1281 		 * Determine support for Power Management
1282 		 */
1283 		caps_ptr = (uint8_t)ql_pci_config_get8(ha, PCI_CONF_CAP_PTR);
1284 
1285 		while (caps_ptr != PCI_CAP_NEXT_PTR_NULL) {
1286 			cap = (uint8_t)ql_pci_config_get8(ha, caps_ptr);
1287 			if (cap == PCI_CAP_ID_PM) {
1288 				ha->pm_capable = 1;
1289 				break;
1290 			}
1291 			caps_ptr = (uint8_t)ql_pci_config_get8(ha, caps_ptr +
1292 			    PCI_CAP_NEXT_PTR);
1293 		}
1294 
1295 		if (ha->pm_capable) {
1296 			/*
1297 			 * Enable PM for 2200 based HBAs only.
1298 			 */
1299 			if (ha->device_id != 0x2200) {
1300 				ha->pm_capable = 0;
1301 			}
1302 		}
1303 
1304 		if (ha->pm_capable) {
1305 			ha->pm_capable = ql_enable_pm;
1306 		}
1307 
1308 		if (ha->pm_capable) {
1309 			/*
1310 			 * Initialize power management bookkeeping;
1311 			 * components are created idle.
1312 			 */
1313 			(void) sprintf(buf, "NAME=%s(%d)", QL_NAME, instance);
1314 			pmcomps[0] = buf;
1315 
1316 			/*LINTED [Solaris DDI_DEV_T_NONE Lint warning]*/
1317 			if (ddi_prop_update_string_array(DDI_DEV_T_NONE,
1318 			    dip, "pm-components", pmcomps,
1319 			    sizeof (pmcomps) / sizeof (pmcomps[0])) !=
1320 			    DDI_PROP_SUCCESS) {
1321 				cmn_err(CE_WARN, "%s(%d): failed to create"
1322 				    " pm-components property", QL_NAME,
1323 				    instance);
1324 
1325 				/* Initialize adapter. */
1326 				ha->power_level = PM_LEVEL_D0;
1327 				if (ql_initialize_adapter(ha) != QL_SUCCESS) {
1328 					cmn_err(CE_WARN, "%s(%d): failed to"
1329 					    " initialize adapter", QL_NAME,
1330 					    instance);
1331 					goto attach_failed;
1332 				}
1333 			} else {
1334 				ha->power_level = PM_LEVEL_D3;
1335 				if (pm_raise_power(dip, QL_POWER_COMPONENT,
1336 				    PM_LEVEL_D0) != DDI_SUCCESS) {
1337 					cmn_err(CE_WARN, "%s(%d): failed to"
1338 					    " raise power or initialize"
1339 					    " adapter", QL_NAME, instance);
1340 				}
1341 			}
1342 		} else {
1343 			/* Initialize adapter. */
1344 			ha->power_level = PM_LEVEL_D0;
1345 			if (ql_initialize_adapter(ha) != QL_SUCCESS) {
1346 				cmn_err(CE_WARN, "%s(%d): failed to initialize"
1347 				    " adapter", QL_NAME, instance);
1348 			}
1349 		}
1350 
1351 		if (ha->fw_major_version == 0 && ha->fw_minor_version == 0 &&
1352 		    ha->fw_subminor_version == 0) {
1353 			cmn_err(CE_NOTE, "!%s(%d): Firmware not loaded",
1354 			    QL_NAME, ha->instance);
1355 		} else {
1356 			int	rval;
1357 			char	ver_fmt[256];
1358 
1359 			rval = (int)snprintf(ver_fmt, (size_t)sizeof (ver_fmt),
1360 			    "Firmware version %d.%d.%d", ha->fw_major_version,
1361 			    ha->fw_minor_version, ha->fw_subminor_version);
1362 
1363 			if (CFG_IST(ha, CFG_CTRL_81XX)) {
1364 				rval = (int)snprintf(ver_fmt + rval,
1365 				    (size_t)sizeof (ver_fmt),
1366 				    ", MPI fw version %d.%d.%d",
1367 				    ha->mpi_fw_major_version,
1368 				    ha->mpi_fw_minor_version,
1369 				    ha->mpi_fw_subminor_version);
1370 
1371 				if (ha->subsys_id == 0x17B ||
1372 				    ha->subsys_id == 0x17D) {
1373 					(void) snprintf(ver_fmt + rval,
1374 					    (size_t)sizeof (ver_fmt),
1375 					    ", PHY fw version %d.%d.%d",
1376 					    ha->phy_fw_major_version,
1377 					    ha->phy_fw_minor_version,
1378 					    ha->phy_fw_subminor_version);
1379 				}
1380 			}
1381 			cmn_err(CE_NOTE, "!%s(%d): %s",
1382 			    QL_NAME, ha->instance, ver_fmt);
1383 		}
1384 
1385 		ha->k_stats = kstat_create(QL_NAME, instance, "statistics",
1386 		    "controller", KSTAT_TYPE_RAW,
1387 		    (uint32_t)sizeof (ql_adapter_stat_t), KSTAT_FLAG_VIRTUAL);
1388 		if (ha->k_stats == NULL) {
1389 			cmn_err(CE_WARN, "%s(%d): Failed to create kstat",
1390 			    QL_NAME, instance);
1391 			goto attach_failed;
1392 		}
1393 		progress |= QL_KSTAT_CREATED;
1394 
1395 		ha->adapter_stats->version = 1;
1396 		ha->k_stats->ks_data = (void *)ha->adapter_stats;
1397 		ha->k_stats->ks_private = ha;
1398 		ha->k_stats->ks_update = ql_kstat_update;
1399 		ha->k_stats->ks_ndata = 1;
1400 		ha->k_stats->ks_data_size = sizeof (ql_adapter_stat_t);
1401 		kstat_install(ha->k_stats);
1402 
1403 		if (ddi_create_minor_node(dip, "devctl", S_IFCHR,
1404 		    instance, DDI_NT_NEXUS, 0) != DDI_SUCCESS) {
1405 			cmn_err(CE_WARN, "%s(%d): failed to create minor node",
1406 			    QL_NAME, instance);
1407 			goto attach_failed;
1408 		}
1409 		progress |= QL_MINOR_NODE_CREATED;
1410 
1411 		/* Allocate a transport structure for this instance */
1412 		tran = kmem_zalloc(sizeof (fc_fca_tran_t), KM_SLEEP);
1413 		if (tran == NULL) {
1414 			cmn_err(CE_WARN, "%s(%d): failed to allocate transport",
1415 			    QL_NAME, instance);
1416 			goto attach_failed;
1417 		}
1418 
1419 		progress |= QL_FCA_TRAN_ALLOCED;
1420 
1421 		/* fill in the structure */
1422 		tran->fca_numports = 1;
1423 		tran->fca_version = FCTL_FCA_MODREV_5;
1424 		if (CFG_IST(ha, CFG_CTRL_2422)) {
1425 			tran->fca_num_npivports = MAX_24_VIRTUAL_PORTS;
1426 		} else if (CFG_IST(ha, CFG_CTRL_2581)) {
1427 			tran->fca_num_npivports = MAX_25_VIRTUAL_PORTS;
1428 		}
1429 		bcopy(ha->loginparams.node_ww_name.raw_wwn,
1430 		    tran->fca_perm_pwwn.raw_wwn, 8);
1431 
1432 		EL(ha, "FCA version %d\n", tran->fca_version);
1433 
1434 		/* Specify the amount of space needed in each packet */
1435 		tran->fca_pkt_size = sizeof (ql_srb_t);
1436 
1437 		/* command limits are usually dictated by hardware */
1438 		tran->fca_cmd_max = MAX_OUTSTANDING_COMMANDS;
1439 
1440 		/* dmaattr are static, set elsewhere. */
1441 		if (CFG_IST(ha, CFG_ENABLE_64BIT_ADDRESSING)) {
1442 			tran->fca_dma_attr = &ql_64bit_io_dma_attr;
1443 			tran->fca_dma_fcp_cmd_attr = &ql_64fcp_cmd_dma_attr;
1444 			tran->fca_dma_fcp_rsp_attr = &ql_64fcp_rsp_dma_attr;
1445 			tran->fca_dma_fcp_data_attr = &ql_64fcp_data_dma_attr;
1446 			tran->fca_dma_fcsm_cmd_attr = &ql_64fcsm_cmd_dma_attr;
1447 			tran->fca_dma_fcsm_rsp_attr = &ql_64fcsm_rsp_dma_attr;
1448 			tran->fca_dma_fcip_cmd_attr = &ql_64fcip_cmd_dma_attr;
1449 			tran->fca_dma_fcip_rsp_attr = &ql_64fcip_rsp_dma_attr;
1450 		} else {
1451 			tran->fca_dma_attr = &ql_32bit_io_dma_attr;
1452 			tran->fca_dma_fcp_cmd_attr = &ql_32fcp_cmd_dma_attr;
1453 			tran->fca_dma_fcp_rsp_attr = &ql_32fcp_rsp_dma_attr;
1454 			tran->fca_dma_fcp_data_attr = &ql_32fcp_data_dma_attr;
1455 			tran->fca_dma_fcsm_cmd_attr = &ql_32fcsm_cmd_dma_attr;
1456 			tran->fca_dma_fcsm_rsp_attr = &ql_32fcsm_rsp_dma_attr;
1457 			tran->fca_dma_fcip_cmd_attr = &ql_32fcip_cmd_dma_attr;
1458 			tran->fca_dma_fcip_rsp_attr = &ql_32fcip_rsp_dma_attr;
1459 		}
1460 
1461 		tran->fca_acc_attr = &ql_dev_acc_attr;
1462 		tran->fca_iblock = &(ha->iblock_cookie);
1463 
1464 		/* the remaining values are simply function vectors */
1465 		tran->fca_bind_port = ql_bind_port;
1466 		tran->fca_unbind_port = ql_unbind_port;
1467 		tran->fca_init_pkt = ql_init_pkt;
1468 		tran->fca_un_init_pkt = ql_un_init_pkt;
1469 		tran->fca_els_send = ql_els_send;
1470 		tran->fca_get_cap = ql_get_cap;
1471 		tran->fca_set_cap = ql_set_cap;
1472 		tran->fca_getmap = ql_getmap;
1473 		tran->fca_transport = ql_transport;
1474 		tran->fca_ub_alloc = ql_ub_alloc;
1475 		tran->fca_ub_free = ql_ub_free;
1476 		tran->fca_ub_release = ql_ub_release;
1477 		tran->fca_abort = ql_abort;
1478 		tran->fca_reset = ql_reset;
1479 		tran->fca_port_manage = ql_port_manage;
1480 		tran->fca_get_device = ql_get_device;
1481 
1482 		/* give it to the FC transport */
1483 		if (fc_fca_attach(dip, tran) != DDI_SUCCESS) {
1484 			cmn_err(CE_WARN, "%s(%d): FCA attach failed", QL_NAME,
1485 			    instance);
1486 			goto attach_failed;
1487 		}
1488 		progress |= QL_FCA_ATTACH_DONE;
1489 
1490 		/* Stash the structure so it can be freed at detach */
1491 		ha->tran = tran;
1492 
1493 		/* Acquire global state lock. */
1494 		GLOBAL_STATE_LOCK();
1495 
1496 		/* Add adapter structure to link list. */
1497 		ql_add_link_b(&ql_hba, &ha->hba);
1498 
1499 		/* Start one second driver timer. */
1500 		if (ql_timer_timeout_id == NULL) {
1501 			ql_timer_ticks = drv_usectohz(1000000);
1502 			ql_timer_timeout_id = timeout(ql_timer, (void *)0,
1503 			    ql_timer_ticks);
1504 		}
1505 
1506 		/* Release global state lock. */
1507 		GLOBAL_STATE_UNLOCK();
1508 
1509 		/* Determine and populate HBA fru info */
1510 		ql_setup_fruinfo(ha);
1511 
1512 		/* Setup task_daemon thread. */
1513 		(void) thread_create(NULL, 0, (void (*)())ql_task_daemon, ha,
1514 		    0, &p0, TS_RUN, minclsyspri);
1515 
1516 		progress |= QL_TASK_DAEMON_STARTED;
1517 
1518 		ddi_report_dev(dip);
1519 
1520 		/* Disable link reset in panic path */
1521 		ha->lip_on_panic = 1;
1522 
1523 		rval = DDI_SUCCESS;
1524 		break;
1525 
1526 attach_failed:
1527 		if (progress & QL_FCA_ATTACH_DONE) {
1528 			(void) fc_fca_detach(dip);
1529 			progress &= ~QL_FCA_ATTACH_DONE;
1530 		}
1531 
1532 		if (progress & QL_FCA_TRAN_ALLOCED) {
1533 			kmem_free(tran, sizeof (fc_fca_tran_t));
1534 			progress &= ~QL_FCA_TRAN_ALLOCED;
1535 		}
1536 
1537 		if (progress & QL_MINOR_NODE_CREATED) {
1538 			ddi_remove_minor_node(dip, "devctl");
1539 			progress &= ~QL_MINOR_NODE_CREATED;
1540 		}
1541 
1542 		if (progress & QL_KSTAT_CREATED) {
1543 			kstat_delete(ha->k_stats);
1544 			progress &= ~QL_KSTAT_CREATED;
1545 		}
1546 
1547 		if (progress & QL_N_PORT_INFO_CREATED) {
1548 			kmem_free(ha->n_port, sizeof (ql_n_port_info_t));
1549 			progress &= ~QL_N_PORT_INFO_CREATED;
1550 		}
1551 
1552 		if (progress & QL_TASK_DAEMON_STARTED) {
1553 			TASK_DAEMON_LOCK(ha);
1554 
1555 			ha->task_daemon_flags |= TASK_DAEMON_STOP_FLG;
1556 
1557 			cv_signal(&ha->cv_task_daemon);
1558 
1559 			/* Release task daemon lock. */
1560 			TASK_DAEMON_UNLOCK(ha);
1561 
1562 			/* Wait for for task daemon to stop running. */
1563 			while (ha->task_daemon_flags & TASK_DAEMON_STOP_FLG) {
1564 				ql_delay(ha, 10000);
1565 			}
1566 			progress &= ~QL_TASK_DAEMON_STARTED;
1567 		}
1568 
1569 		if (progress & QL_DB_IOBASE_MAPPED) {
1570 			ql_8021_clr_drv_active(ha);
1571 			ddi_regs_map_free(&ha->db_dev_handle);
1572 			progress &= ~QL_DB_IOBASE_MAPPED;
1573 		}
1574 		if (progress & QL_IOMAP_IOBASE_MAPPED) {
1575 			ddi_regs_map_free(&ha->iomap_dev_handle);
1576 			progress &= ~QL_IOMAP_IOBASE_MAPPED;
1577 		}
1578 
1579 		if (progress & QL_CONFIG_SPACE_SETUP) {
1580 			if (CFG_IST(ha, CFG_SBUS_CARD)) {
1581 				ddi_regs_map_free(&ha->sbus_config_handle);
1582 			} else {
1583 				pci_config_teardown(&ha->pci_handle);
1584 			}
1585 			progress &= ~QL_CONFIG_SPACE_SETUP;
1586 		}
1587 
1588 		if (progress & QL_INTR_ADDED) {
1589 			ql_disable_intr(ha);
1590 			ql_release_intr(ha);
1591 			progress &= ~QL_INTR_ADDED;
1592 		}
1593 
1594 		if (progress & QL_MUTEX_CV_INITED) {
1595 			ql_destroy_mutex(ha);
1596 			progress &= ~QL_MUTEX_CV_INITED;
1597 		}
1598 
1599 		if (progress & QL_HBA_BUFFER_SETUP) {
1600 			ql_free_phys(ha, &ha->hba_buf);
1601 			progress &= ~QL_HBA_BUFFER_SETUP;
1602 		}
1603 
1604 		if (progress & QL_REGS_MAPPED) {
1605 			ddi_regs_map_free(&ha->dev_handle);
1606 			if (ha->sbus_fpga_iobase != NULL) {
1607 				ddi_regs_map_free(&ha->sbus_fpga_dev_handle);
1608 			}
1609 			progress &= ~QL_REGS_MAPPED;
1610 		}
1611 
1612 		if (progress & QL_SOFT_STATE_ALLOCED) {
1613 
1614 			ql_fcache_rel(ha->fcache);
1615 
1616 			kmem_free(ha->adapter_stats,
1617 			    sizeof (*ha->adapter_stats));
1618 
1619 			kmem_free(ha->ub_array, sizeof (*ha->ub_array) *
1620 			    QL_UB_LIMIT);
1621 
1622 			kmem_free(ha->outstanding_cmds,
1623 			    sizeof (*ha->outstanding_cmds) *
1624 			    MAX_OUTSTANDING_COMMANDS);
1625 
1626 			if (ha->devpath != NULL) {
1627 				kmem_free(ha->devpath,
1628 				    strlen(ha->devpath) + 1);
1629 			}
1630 
1631 			kmem_free(ha->dev, sizeof (*ha->dev) *
1632 			    DEVICE_HEAD_LIST_SIZE);
1633 
1634 			if (ha->xioctl != NULL) {
1635 				ql_free_xioctl_resource(ha);
1636 			}
1637 
1638 			if (ha->fw_module != NULL) {
1639 				(void) ddi_modclose(ha->fw_module);
1640 			}
1641 			(void) ql_el_trace_desc_dtor(ha);
1642 			(void) ql_nvram_cache_desc_dtor(ha);
1643 
1644 			ddi_soft_state_free(ql_state, instance);
1645 			progress &= ~QL_SOFT_STATE_ALLOCED;
1646 		}
1647 
1648 		ddi_prop_remove_all(dip);
1649 		rval = DDI_FAILURE;
1650 		break;
1651 
1652 	case DDI_RESUME:
1653 		rval = DDI_FAILURE;
1654 
1655 		ha = ddi_get_soft_state(ql_state, ddi_get_instance(dip));
1656 		if (ha == NULL) {
1657 			cmn_err(CE_WARN, "%s(%d): can't get soft state",
1658 			    QL_NAME, instance);
1659 			break;
1660 		}
1661 
1662 		ha->power_level = PM_LEVEL_D3;
1663 		if (ha->pm_capable) {
1664 			/*
1665 			 * Get ql_power to do power on initialization
1666 			 */
1667 			if (pm_raise_power(dip, QL_POWER_COMPONENT,
1668 			    PM_LEVEL_D0) != DDI_SUCCESS) {
1669 				cmn_err(CE_WARN, "%s(%d): can't raise adapter"
1670 				    " power", QL_NAME, instance);
1671 			}
1672 		}
1673 
1674 		/*
1675 		 * There is a bug in DR that prevents PM framework
1676 		 * from calling ql_power.
1677 		 */
1678 		if (ha->power_level == PM_LEVEL_D3) {
1679 			ha->power_level = PM_LEVEL_D0;
1680 
1681 			if (ql_initialize_adapter(ha) != QL_SUCCESS) {
1682 				cmn_err(CE_WARN, "%s(%d): can't initialize the"
1683 				    " adapter", QL_NAME, instance);
1684 			}
1685 
1686 			/* Wake up task_daemon. */
1687 			ql_awaken_task_daemon(ha, NULL, TASK_DAEMON_ALIVE_FLG,
1688 			    0);
1689 		}
1690 
1691 		/* Acquire global state lock. */
1692 		GLOBAL_STATE_LOCK();
1693 
1694 		/* Restart driver timer. */
1695 		if (ql_timer_timeout_id == NULL) {
1696 			ql_timer_timeout_id = timeout(ql_timer, (void *)0,
1697 			    ql_timer_ticks);
1698 		}
1699 
1700 		/* Release global state lock. */
1701 		GLOBAL_STATE_UNLOCK();
1702 
1703 		/* Wake up command start routine. */
1704 		ADAPTER_STATE_LOCK(ha);
1705 		ha->flags &= ~ADAPTER_SUSPENDED;
1706 		ADAPTER_STATE_UNLOCK(ha);
1707 
1708 		/*
1709 		 * Transport doesn't make FC discovery in polled
1710 		 * mode; So we need the daemon thread's services
1711 		 * right here.
1712 		 */
1713 		(void) callb_generic_cpr(&ha->cprinfo, CB_CODE_CPR_RESUME);
1714 
1715 		rval = DDI_SUCCESS;
1716 
1717 		/* Restart IP if it was running. */
1718 		if (ha->flags & IP_ENABLED && !(ha->flags & IP_INITIALIZED)) {
1719 			(void) ql_initialize_ip(ha);
1720 			ql_isp_rcvbuf(ha);
1721 		}
1722 		break;
1723 
1724 	default:
1725 		cmn_err(CE_WARN, "%s(%d): attach, unknown code:"
1726 		    " %x", QL_NAME, ddi_get_instance(dip), cmd);
1727 		rval = DDI_FAILURE;
1728 		break;
1729 	}
1730 
1731 	kmem_free(buf, MAXPATHLEN);
1732 
1733 	if (rval != DDI_SUCCESS) {
1734 		/*EMPTY*/
1735 		QL_PRINT_2(CE_CONT, "(%d): failed, rval = %xh\n",
1736 		    ddi_get_instance(dip), rval);
1737 	} else {
1738 		/*EMPTY*/
1739 		QL_PRINT_3(CE_CONT, "(%d): done\n", ddi_get_instance(dip));
1740 	}
1741 
1742 	return (rval);
1743 }
1744 
1745 /*
1746  * ql_detach
1747  *	Used to remove all the states associated with a given
1748  *	instances of a device node prior to the removal of that
1749  *	instance from the system.
1750  *
1751  * Input:
1752  *	dip = pointer to device information structure.
1753  *	cmd = type of detach.
1754  *
1755  * Returns:
1756  *	DDI_SUCCESS or DDI_FAILURE.
1757  *
1758  * Context:
1759  *	Kernel context.
1760  */
1761 static int
ql_detach(dev_info_t * dip,ddi_detach_cmd_t cmd)1762 ql_detach(dev_info_t *dip, ddi_detach_cmd_t cmd)
1763 {
1764 	ql_adapter_state_t	*ha, *vha;
1765 	ql_tgt_t		*tq;
1766 	int			delay_cnt;
1767 	uint16_t		index;
1768 	ql_link_t		*link;
1769 	char			*buf;
1770 	timeout_id_t		timer_id = NULL;
1771 	int			suspend, rval = DDI_SUCCESS;
1772 
1773 	ha = ddi_get_soft_state(ql_state, ddi_get_instance(dip));
1774 	if (ha == NULL) {
1775 		QL_PRINT_2(CE_CONT, "(%d): no adapter\n",
1776 		    ddi_get_instance(dip));
1777 		return (DDI_FAILURE);
1778 	}
1779 
1780 	QL_PRINT_3(CE_CONT, "(%d): started, cmd=%xh\n", ha->instance, cmd);
1781 
1782 	buf = (char *)(kmem_zalloc(MAXPATHLEN, KM_SLEEP));
1783 
1784 	switch (cmd) {
1785 	case DDI_DETACH:
1786 		ADAPTER_STATE_LOCK(ha);
1787 		ha->flags |= (ADAPTER_SUSPENDED | ABORT_CMDS_LOOP_DOWN_TMO);
1788 		ADAPTER_STATE_UNLOCK(ha);
1789 
1790 		TASK_DAEMON_LOCK(ha);
1791 
1792 		if (ha->task_daemon_flags & TASK_DAEMON_ALIVE_FLG) {
1793 			ha->task_daemon_flags |= TASK_DAEMON_STOP_FLG;
1794 			cv_signal(&ha->cv_task_daemon);
1795 
1796 			TASK_DAEMON_UNLOCK(ha);
1797 
1798 			(void) ql_wait_for_td_stop(ha);
1799 
1800 			TASK_DAEMON_LOCK(ha);
1801 			if (ha->task_daemon_flags & TASK_DAEMON_STOP_FLG) {
1802 				ha->task_daemon_flags &= ~TASK_DAEMON_STOP_FLG;
1803 				EL(ha, "failed, could not stop task daemon\n");
1804 			}
1805 		}
1806 		TASK_DAEMON_UNLOCK(ha);
1807 
1808 		GLOBAL_STATE_LOCK();
1809 
1810 		/* Disable driver timer if no adapters. */
1811 		if (ql_timer_timeout_id && ql_hba.first == &ha->hba &&
1812 		    ql_hba.last == &ha->hba) {
1813 			timer_id = ql_timer_timeout_id;
1814 			ql_timer_timeout_id = NULL;
1815 		}
1816 		ql_remove_link(&ql_hba, &ha->hba);
1817 
1818 		GLOBAL_STATE_UNLOCK();
1819 
1820 		if (timer_id) {
1821 			(void) untimeout(timer_id);
1822 		}
1823 
1824 		if (ha->pm_capable) {
1825 			if (pm_lower_power(dip, QL_POWER_COMPONENT,
1826 			    PM_LEVEL_D3) != DDI_SUCCESS) {
1827 				cmn_err(CE_WARN, "%s(%d): failed to lower the"
1828 				    " power", QL_NAME, ha->instance);
1829 			}
1830 		}
1831 
1832 		/*
1833 		 * If pm_lower_power shutdown the adapter, there
1834 		 * isn't much else to do
1835 		 */
1836 		if (ha->power_level != PM_LEVEL_D3) {
1837 			ql_halt(ha, PM_LEVEL_D3);
1838 		}
1839 
1840 		/* Remove virtual ports. */
1841 		while ((vha = ha->vp_next) != NULL) {
1842 			ql_vport_destroy(vha);
1843 		}
1844 
1845 		/* Free target queues. */
1846 		for (index = 0; index < DEVICE_HEAD_LIST_SIZE; index++) {
1847 			link = ha->dev[index].first;
1848 			while (link != NULL) {
1849 				tq = link->base_address;
1850 				link = link->next;
1851 				ql_dev_free(ha, tq);
1852 			}
1853 		}
1854 
1855 		/*
1856 		 * Free unsolicited buffers.
1857 		 * If we are here then there are no ULPs still
1858 		 * alive that wish to talk to ql so free up
1859 		 * any SRB_IP_UB_UNUSED buffers that are
1860 		 * lingering around
1861 		 */
1862 		QL_UB_LOCK(ha);
1863 		for (index = 0; index < QL_UB_LIMIT; index++) {
1864 			fc_unsol_buf_t *ubp = ha->ub_array[index];
1865 
1866 			if (ubp != NULL) {
1867 				ql_srb_t *sp = ubp->ub_fca_private;
1868 
1869 				sp->flags |= SRB_UB_FREE_REQUESTED;
1870 
1871 				while (!(sp->flags & SRB_UB_IN_FCA) ||
1872 				    (sp->flags & (SRB_UB_CALLBACK |
1873 				    SRB_UB_ACQUIRED))) {
1874 					QL_UB_UNLOCK(ha);
1875 					delay(drv_usectohz(100000));
1876 					QL_UB_LOCK(ha);
1877 				}
1878 				ha->ub_array[index] = NULL;
1879 
1880 				QL_UB_UNLOCK(ha);
1881 				ql_free_unsolicited_buffer(ha, ubp);
1882 				QL_UB_LOCK(ha);
1883 			}
1884 		}
1885 		QL_UB_UNLOCK(ha);
1886 
1887 		/* Free any saved RISC code. */
1888 		if (ha->risc_code != NULL) {
1889 			kmem_free(ha->risc_code, ha->risc_code_size);
1890 			ha->risc_code = NULL;
1891 			ha->risc_code_size = 0;
1892 		}
1893 
1894 		if (ha->fw_module != NULL) {
1895 			(void) ddi_modclose(ha->fw_module);
1896 			ha->fw_module = NULL;
1897 		}
1898 
1899 		/* Free resources. */
1900 		ddi_prop_remove_all(dip);
1901 		(void) fc_fca_detach(dip);
1902 		kmem_free(ha->tran, sizeof (fc_fca_tran_t));
1903 		ddi_remove_minor_node(dip, "devctl");
1904 		if (ha->k_stats != NULL) {
1905 			kstat_delete(ha->k_stats);
1906 		}
1907 
1908 		if (CFG_IST(ha, CFG_SBUS_CARD)) {
1909 			ddi_regs_map_free(&ha->sbus_config_handle);
1910 		} else {
1911 			if (CFG_IST(ha, CFG_CTRL_8021)) {
1912 				ql_8021_clr_drv_active(ha);
1913 				ddi_regs_map_free(&ha->db_dev_handle);
1914 			}
1915 			if (ha->iomap_dev_handle != ha->dev_handle) {
1916 				ddi_regs_map_free(&ha->iomap_dev_handle);
1917 			}
1918 			pci_config_teardown(&ha->pci_handle);
1919 		}
1920 
1921 		ql_disable_intr(ha);
1922 		ql_release_intr(ha);
1923 
1924 		ql_free_xioctl_resource(ha);
1925 
1926 		ql_destroy_mutex(ha);
1927 
1928 		ql_free_phys(ha, &ha->hba_buf);
1929 		ql_free_phys(ha, &ha->fwexttracebuf);
1930 		ql_free_phys(ha, &ha->fwfcetracebuf);
1931 
1932 		ddi_regs_map_free(&ha->dev_handle);
1933 		if (ha->sbus_fpga_iobase != NULL) {
1934 			ddi_regs_map_free(&ha->sbus_fpga_dev_handle);
1935 		}
1936 
1937 		ql_fcache_rel(ha->fcache);
1938 		if (ha->vcache != NULL) {
1939 			kmem_free(ha->vcache, QL_24XX_VPD_SIZE);
1940 		}
1941 
1942 		if (ha->pi_attrs != NULL) {
1943 			kmem_free(ha->pi_attrs, sizeof (fca_port_attrs_t));
1944 		}
1945 
1946 		kmem_free(ha->adapter_stats, sizeof (*ha->adapter_stats));
1947 
1948 		kmem_free(ha->ub_array, sizeof (*ha->ub_array) * QL_UB_LIMIT);
1949 
1950 		kmem_free(ha->outstanding_cmds,
1951 		    sizeof (*ha->outstanding_cmds) * MAX_OUTSTANDING_COMMANDS);
1952 
1953 		if (ha->n_port != NULL) {
1954 			kmem_free(ha->n_port, sizeof (ql_n_port_info_t));
1955 		}
1956 
1957 		if (ha->devpath != NULL) {
1958 			kmem_free(ha->devpath, strlen(ha->devpath) + 1);
1959 		}
1960 
1961 		kmem_free(ha->dev, sizeof (*ha->dev) * DEVICE_HEAD_LIST_SIZE);
1962 
1963 		EL(ha, "detached\n");
1964 
1965 		ddi_soft_state_free(ql_state, (int)ha->instance);
1966 
1967 		break;
1968 
1969 	case DDI_SUSPEND:
1970 		ADAPTER_STATE_LOCK(ha);
1971 
1972 		delay_cnt = 0;
1973 		ha->flags |= ADAPTER_SUSPENDED;
1974 		while (ha->flags & ADAPTER_TIMER_BUSY && delay_cnt++ < 10) {
1975 			ADAPTER_STATE_UNLOCK(ha);
1976 			delay(drv_usectohz(1000000));
1977 			ADAPTER_STATE_LOCK(ha);
1978 		}
1979 		if (ha->busy || ha->flags & ADAPTER_TIMER_BUSY) {
1980 			ha->flags &= ~ADAPTER_SUSPENDED;
1981 			ADAPTER_STATE_UNLOCK(ha);
1982 			rval = DDI_FAILURE;
1983 			cmn_err(CE_WARN, "!%s(%d): Fail suspend"
1984 			    " busy %xh flags %xh", QL_NAME, ha->instance,
1985 			    ha->busy, ha->flags);
1986 			break;
1987 		}
1988 
1989 		ADAPTER_STATE_UNLOCK(ha);
1990 
1991 		if (ha->flags & IP_INITIALIZED) {
1992 			(void) ql_shutdown_ip(ha);
1993 		}
1994 
1995 		if ((suspend = ql_suspend_adapter(ha)) != QL_SUCCESS) {
1996 			ADAPTER_STATE_LOCK(ha);
1997 			ha->flags &= ~ADAPTER_SUSPENDED;
1998 			ADAPTER_STATE_UNLOCK(ha);
1999 			cmn_err(CE_WARN, "%s(%d): Fail suspend rval %xh",
2000 			    QL_NAME, ha->instance, suspend);
2001 
2002 			/* Restart IP if it was running. */
2003 			if (ha->flags & IP_ENABLED &&
2004 			    !(ha->flags & IP_INITIALIZED)) {
2005 				(void) ql_initialize_ip(ha);
2006 				ql_isp_rcvbuf(ha);
2007 			}
2008 			rval = DDI_FAILURE;
2009 			break;
2010 		}
2011 
2012 		/* Acquire global state lock. */
2013 		GLOBAL_STATE_LOCK();
2014 
2015 		/* Disable driver timer if last adapter. */
2016 		if (ql_timer_timeout_id && ql_hba.first == &ha->hba &&
2017 		    ql_hba.last == &ha->hba) {
2018 			timer_id = ql_timer_timeout_id;
2019 			ql_timer_timeout_id = NULL;
2020 		}
2021 		GLOBAL_STATE_UNLOCK();
2022 
2023 		if (timer_id) {
2024 			(void) untimeout(timer_id);
2025 		}
2026 
2027 		EL(ha, "suspended\n");
2028 
2029 		break;
2030 
2031 	default:
2032 		rval = DDI_FAILURE;
2033 		break;
2034 	}
2035 
2036 	kmem_free(buf, MAXPATHLEN);
2037 
2038 	if (rval != DDI_SUCCESS) {
2039 		if (ha != NULL) {
2040 			EL(ha, "failed, rval = %xh\n", rval);
2041 		} else {
2042 			/*EMPTY*/
2043 			QL_PRINT_2(CE_CONT, "(%d): failed, rval = %xh\n",
2044 			    ddi_get_instance(dip), rval);
2045 		}
2046 	} else {
2047 		/*EMPTY*/
2048 		QL_PRINT_3(CE_CONT, "(%d): done\n", ddi_get_instance(dip));
2049 	}
2050 
2051 	return (rval);
2052 }
2053 
2054 
2055 /*
2056  * ql_power
2057  *	Power a device attached to the system.
2058  *
2059  * Input:
2060  *	dip = pointer to device information structure.
2061  *	component = device.
2062  *	level = power level.
2063  *
2064  * Returns:
2065  *	DDI_SUCCESS or DDI_FAILURE.
2066  *
2067  * Context:
2068  *	Kernel context.
2069  */
2070 /* ARGSUSED */
2071 static int
ql_power(dev_info_t * dip,int component,int level)2072 ql_power(dev_info_t *dip, int component, int level)
2073 {
2074 	int			rval = DDI_FAILURE;
2075 	off_t			csr;
2076 	uint8_t			saved_pm_val;
2077 	ql_adapter_state_t	*ha;
2078 	char			*buf;
2079 	char			*path;
2080 
2081 	ha = ddi_get_soft_state(ql_state, ddi_get_instance(dip));
2082 	if (ha == NULL || ha->pm_capable == 0) {
2083 		QL_PRINT_2(CE_CONT, "(%d): no hba or PM not supported\n",
2084 		    ddi_get_instance(dip));
2085 		return (rval);
2086 	}
2087 
2088 	QL_PRINT_10(CE_CONT, "(%d,%d): started\n", ha->instance, ha->vp_index);
2089 
2090 	buf = (char *)(kmem_zalloc(MAXPATHLEN, KM_SLEEP));
2091 	path = (char *)(kmem_zalloc(MAXPATHLEN, KM_SLEEP));
2092 
2093 	if (component != QL_POWER_COMPONENT || (level != PM_LEVEL_D0 &&
2094 	    level != PM_LEVEL_D3)) {
2095 		EL(ha, "invalid, component=%xh or level=%xh\n",
2096 		    component, level);
2097 		return (rval);
2098 	}
2099 
2100 	GLOBAL_HW_LOCK();
2101 	csr = (uint8_t)ql_pci_config_get8(ha, PCI_CONF_CAP_PTR) + PCI_PMCSR;
2102 	GLOBAL_HW_UNLOCK();
2103 
2104 	(void) snprintf(buf, sizeof (buf),
2105 	    "Qlogic %s(%d): %s\n\t", QL_NAME, ddi_get_instance(dip),
2106 	    ddi_pathname(dip, path));
2107 
2108 	switch (level) {
2109 	case PM_LEVEL_D0:	/* power up to D0 state - fully on */
2110 
2111 		QL_PM_LOCK(ha);
2112 		if (ha->power_level == PM_LEVEL_D0) {
2113 			QL_PM_UNLOCK(ha);
2114 			rval = DDI_SUCCESS;
2115 			break;
2116 		}
2117 
2118 		/*
2119 		 * Enable interrupts now
2120 		 */
2121 		saved_pm_val = ha->power_level;
2122 		ha->power_level = PM_LEVEL_D0;
2123 		QL_PM_UNLOCK(ha);
2124 
2125 		GLOBAL_HW_LOCK();
2126 
2127 		ql_pci_config_put16(ha, csr, PCI_PMCSR_D0);
2128 
2129 		/*
2130 		 * Delay after reset, for chip to recover.
2131 		 * Otherwise causes system PANIC
2132 		 */
2133 		drv_usecwait(200000);
2134 
2135 		GLOBAL_HW_UNLOCK();
2136 
2137 		if (ha->config_saved) {
2138 			ha->config_saved = 0;
2139 			if (QL_RESTORE_CONFIG_REGS(dip) != DDI_SUCCESS) {
2140 				QL_PM_LOCK(ha);
2141 				ha->power_level = saved_pm_val;
2142 				QL_PM_UNLOCK(ha);
2143 				cmn_err(CE_WARN, "%s failed to restore "
2144 				    "config regs", buf);
2145 				break;
2146 			}
2147 		}
2148 
2149 		if (ql_initialize_adapter(ha) != QL_SUCCESS) {
2150 			cmn_err(CE_WARN, "%s adapter initialization failed",
2151 			    buf);
2152 		}
2153 
2154 		/* Wake up task_daemon. */
2155 		ql_awaken_task_daemon(ha, NULL, TASK_DAEMON_ALIVE_FLG |
2156 		    TASK_DAEMON_SLEEPING_FLG, 0);
2157 
2158 		/* Restart IP if it was running. */
2159 		if (ha->flags & IP_ENABLED && !(ha->flags & IP_INITIALIZED)) {
2160 			(void) ql_initialize_ip(ha);
2161 			ql_isp_rcvbuf(ha);
2162 		}
2163 
2164 		cmn_err(CE_NOTE, QL_BANG "ql_power(%d): %s is powered ON\n",
2165 		    ha->instance, QL_NAME);
2166 
2167 		rval = DDI_SUCCESS;
2168 		break;
2169 
2170 	case PM_LEVEL_D3:	/* power down to D3 state - off */
2171 
2172 		QL_PM_LOCK(ha);
2173 
2174 		if (ha->busy || ((ha->task_daemon_flags &
2175 		    TASK_DAEMON_SLEEPING_FLG) == 0)) {
2176 			QL_PM_UNLOCK(ha);
2177 			break;
2178 		}
2179 
2180 		if (ha->power_level == PM_LEVEL_D3) {
2181 			rval = DDI_SUCCESS;
2182 			QL_PM_UNLOCK(ha);
2183 			break;
2184 		}
2185 		QL_PM_UNLOCK(ha);
2186 
2187 		if (QL_SAVE_CONFIG_REGS(dip) != DDI_SUCCESS) {
2188 			cmn_err(CE_WARN, "!Qlogic %s(%d): %s failed to save"
2189 			    " config regs", QL_NAME, ha->instance, buf);
2190 			break;
2191 		}
2192 		ha->config_saved = 1;
2193 
2194 		/*
2195 		 * Don't enable interrupts. Running mailbox commands with
2196 		 * interrupts enabled could cause hangs since pm_run_scan()
2197 		 * runs out of a callout thread and on single cpu systems
2198 		 * cv_reltimedwait_sig(), called from ql_mailbox_command(),
2199 		 * would not get to run.
2200 		 */
2201 		TASK_DAEMON_LOCK(ha);
2202 		ha->task_daemon_flags |= TASK_DAEMON_POWERING_DOWN;
2203 		TASK_DAEMON_UNLOCK(ha);
2204 
2205 		ql_halt(ha, PM_LEVEL_D3);
2206 
2207 		/*
2208 		 * Setup ql_intr to ignore interrupts from here on.
2209 		 */
2210 		QL_PM_LOCK(ha);
2211 		ha->power_level = PM_LEVEL_D3;
2212 		QL_PM_UNLOCK(ha);
2213 
2214 		/*
2215 		 * Wait for ISR to complete.
2216 		 */
2217 		INTR_LOCK(ha);
2218 		ql_pci_config_put16(ha, csr, PCI_PMCSR_D3HOT);
2219 		INTR_UNLOCK(ha);
2220 
2221 		cmn_err(CE_NOTE, QL_BANG "ql_power(%d): %s is powered OFF\n",
2222 		    ha->instance, QL_NAME);
2223 
2224 		rval = DDI_SUCCESS;
2225 		break;
2226 	}
2227 
2228 	kmem_free(buf, MAXPATHLEN);
2229 	kmem_free(path, MAXPATHLEN);
2230 
2231 	QL_PRINT_10(CE_CONT, "(%d,%d): done\n", ha->instance, ha->vp_index);
2232 
2233 	return (rval);
2234 }
2235 
2236 /*
2237  * ql_quiesce
2238  *	quiesce a device attached to the system.
2239  *
2240  * Input:
2241  *	dip = pointer to device information structure.
2242  *
2243  * Returns:
2244  *	DDI_SUCCESS
2245  *
2246  * Context:
2247  *	Kernel context.
2248  */
2249 static int
ql_quiesce(dev_info_t * dip)2250 ql_quiesce(dev_info_t *dip)
2251 {
2252 	ql_adapter_state_t	*ha;
2253 	uint32_t		timer;
2254 	uint32_t		stat;
2255 
2256 	ha = ddi_get_soft_state(ql_state, ddi_get_instance(dip));
2257 	if (ha == NULL) {
2258 		/* Oh well.... */
2259 		QL_PRINT_2(CE_CONT, "(%d): no adapter\n",
2260 		    ddi_get_instance(dip));
2261 		return (DDI_SUCCESS);
2262 	}
2263 
2264 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
2265 
2266 	if (CFG_IST(ha, CFG_CTRL_8021)) {
2267 		(void) ql_stop_firmware(ha);
2268 	} else if (CFG_IST(ha, CFG_CTRL_242581)) {
2269 		WRT32_IO_REG(ha, hccr, HC24_CLR_RISC_INT);
2270 		WRT16_IO_REG(ha, mailbox_in[0], MBC_STOP_FIRMWARE);
2271 		WRT32_IO_REG(ha, hccr, HC24_SET_HOST_INT);
2272 		for (timer = 0; timer < 30000; timer++) {
2273 			stat = RD32_IO_REG(ha, risc2host);
2274 			if (stat & BIT_15) {
2275 				if ((stat & 0xff) < 0x12) {
2276 					WRT32_IO_REG(ha, hccr,
2277 					    HC24_CLR_RISC_INT);
2278 					break;
2279 				}
2280 				WRT32_IO_REG(ha, hccr, HC24_CLR_RISC_INT);
2281 			}
2282 			drv_usecwait(100);
2283 		}
2284 		/* Reset the chip. */
2285 		WRT32_IO_REG(ha, ctrl_status, ISP_RESET | DMA_SHUTDOWN |
2286 		    MWB_4096_BYTES);
2287 		drv_usecwait(100);
2288 
2289 	} else {
2290 		/* Disable ISP interrupts. */
2291 		WRT16_IO_REG(ha, ictrl, 0);
2292 		/* Select RISC module registers. */
2293 		WRT16_IO_REG(ha, ctrl_status, 0);
2294 		/* Reset ISP semaphore. */
2295 		WRT16_IO_REG(ha, semaphore, 0);
2296 		/* Reset RISC module. */
2297 		WRT16_IO_REG(ha, hccr, HC_RESET_RISC);
2298 		/* Release RISC module. */
2299 		WRT16_IO_REG(ha, hccr, HC_RELEASE_RISC);
2300 	}
2301 
2302 	ql_disable_intr(ha);
2303 
2304 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
2305 
2306 	return (DDI_SUCCESS);
2307 }
2308 
2309 /* ************************************************************************ */
2310 /*		Fibre Channel Adapter (FCA) Transport Functions.	    */
2311 /* ************************************************************************ */
2312 
2313 /*
2314  * ql_bind_port
2315  *	Handling port binding. The FC Transport attempts to bind an FCA port
2316  *	when it is ready to start transactions on the port. The FC Transport
2317  *	will call the fca_bind_port() function specified in the fca_transport
2318  *	structure it receives. The FCA must fill in the port_info structure
2319  *	passed in the call and also stash the information for future calls.
2320  *
2321  * Input:
2322  *	dip = pointer to FCA information structure.
2323  *	port_info = pointer to port information structure.
2324  *	bind_info = pointer to bind information structure.
2325  *
2326  * Returns:
2327  *	NULL = failure
2328  *
2329  * Context:
2330  *	Kernel context.
2331  */
2332 static opaque_t
ql_bind_port(dev_info_t * dip,fc_fca_port_info_t * port_info,fc_fca_bind_info_t * bind_info)2333 ql_bind_port(dev_info_t *dip, fc_fca_port_info_t *port_info,
2334     fc_fca_bind_info_t *bind_info)
2335 {
2336 	ql_adapter_state_t	*ha, *vha;
2337 	opaque_t		fca_handle = NULL;
2338 	port_id_t		d_id;
2339 	int			port_npiv = bind_info->port_npiv;
2340 	uchar_t			*port_nwwn = bind_info->port_nwwn.raw_wwn;
2341 	uchar_t			*port_pwwn = bind_info->port_pwwn.raw_wwn;
2342 
2343 	/* get state info based on the dip */
2344 	ha = ddi_get_soft_state(ql_state, ddi_get_instance(dip));
2345 	if (ha == NULL) {
2346 		QL_PRINT_2(CE_CONT, "(%d): no adapter\n",
2347 		    ddi_get_instance(dip));
2348 		return (NULL);
2349 	}
2350 	QL_PRINT_10(CE_CONT, "(%d,%d): started\n", ha->instance, ha->vp_index);
2351 
2352 	/* Verify port number is supported. */
2353 	if (port_npiv != 0) {
2354 		if (!(ha->flags & VP_ENABLED)) {
2355 			QL_PRINT_2(CE_CONT, "(%d): FC_NPIV_NOT_SUPPORTED\n",
2356 			    ha->instance);
2357 			port_info->pi_error = FC_NPIV_NOT_SUPPORTED;
2358 			return (NULL);
2359 		}
2360 		if (!(ha->flags & POINT_TO_POINT)) {
2361 			QL_PRINT_2(CE_CONT, "(%d): FC_NPIV_WRONG_TOPOLOGY\n",
2362 			    ha->instance);
2363 			port_info->pi_error = FC_NPIV_WRONG_TOPOLOGY;
2364 			return (NULL);
2365 		}
2366 		if (!(ha->flags & FDISC_ENABLED)) {
2367 			QL_PRINT_2(CE_CONT, "(%d): switch does not support "
2368 			    "FDISC\n", ha->instance);
2369 			port_info->pi_error = FC_NPIV_FDISC_FAILED;
2370 			return (NULL);
2371 		}
2372 		if (bind_info->port_num > (CFG_IST(ha, CFG_CTRL_2422) ?
2373 		    MAX_24_VIRTUAL_PORTS : MAX_25_VIRTUAL_PORTS)) {
2374 			QL_PRINT_2(CE_CONT, "(%d): port number=%d "
2375 			    "FC_OUTOFBOUNDS\n", ha->instance);
2376 			port_info->pi_error = FC_OUTOFBOUNDS;
2377 			return (NULL);
2378 		}
2379 	} else if (bind_info->port_num != 0) {
2380 		QL_PRINT_2(CE_CONT, "(%d): failed, port number=%d is not "
2381 		    "supported\n", ha->instance, bind_info->port_num);
2382 		port_info->pi_error = FC_OUTOFBOUNDS;
2383 		return (NULL);
2384 	}
2385 
2386 	/* Locate port context. */
2387 	for (vha = ha; vha != NULL; vha = vha->vp_next) {
2388 		if (vha->vp_index == bind_info->port_num) {
2389 			break;
2390 		}
2391 	}
2392 
2393 	/* If virtual port does not exist. */
2394 	if (vha == NULL) {
2395 		vha = ql_vport_create(ha, (uint8_t)bind_info->port_num);
2396 	}
2397 
2398 	/* make sure this port isn't already bound */
2399 	if (vha->flags & FCA_BOUND) {
2400 		port_info->pi_error = FC_ALREADY;
2401 	} else {
2402 		if (vha->vp_index != 0) {
2403 			bcopy(port_nwwn,
2404 			    vha->loginparams.node_ww_name.raw_wwn, 8);
2405 			bcopy(port_pwwn,
2406 			    vha->loginparams.nport_ww_name.raw_wwn, 8);
2407 		}
2408 		if (vha->vp_index != 0 && !(vha->flags & VP_ENABLED)) {
2409 			if (ql_vport_enable(vha) != QL_SUCCESS) {
2410 				QL_PRINT_2(CE_CONT, "(%d): failed to enable "
2411 				    "virtual port=%d\n", ha->instance,
2412 				    vha->vp_index);
2413 				port_info->pi_error = FC_NPIV_FDISC_FAILED;
2414 				return (NULL);
2415 			}
2416 			cmn_err(CE_CONT, "!Qlogic %s(%d) NPIV(%d) "
2417 			    "WWPN=%02x%02x%02x%02x%02x%02x%02x%02x : "
2418 			    "WWNN=%02x%02x%02x%02x%02x%02x%02x%02x\n",
2419 			    QL_NAME, ha->instance, vha->vp_index,
2420 			    port_pwwn[0], port_pwwn[1], port_pwwn[2],
2421 			    port_pwwn[3], port_pwwn[4], port_pwwn[5],
2422 			    port_pwwn[6], port_pwwn[7],
2423 			    port_nwwn[0], port_nwwn[1], port_nwwn[2],
2424 			    port_nwwn[3], port_nwwn[4], port_nwwn[5],
2425 			    port_nwwn[6], port_nwwn[7]);
2426 		}
2427 
2428 		/* stash the bind_info supplied by the FC Transport */
2429 		vha->bind_info.port_handle = bind_info->port_handle;
2430 		vha->bind_info.port_statec_cb =
2431 		    bind_info->port_statec_cb;
2432 		vha->bind_info.port_unsol_cb = bind_info->port_unsol_cb;
2433 
2434 		/* Set port's source ID. */
2435 		port_info->pi_s_id.port_id = vha->d_id.b24;
2436 
2437 		/* copy out the default login parameters */
2438 		bcopy((void *)&vha->loginparams,
2439 		    (void *)&port_info->pi_login_params,
2440 		    sizeof (la_els_logi_t));
2441 
2442 		/* Set port's hard address if enabled. */
2443 		port_info->pi_hard_addr.hard_addr = 0;
2444 		if (bind_info->port_num == 0) {
2445 			d_id.b24 = ha->d_id.b24;
2446 			if (CFG_IST(ha, CFG_CTRL_24258081)) {
2447 				if (ha->init_ctrl_blk.cb24.
2448 				    firmware_options_1[0] & BIT_0) {
2449 					d_id.b.al_pa = ql_index_to_alpa[ha->
2450 					    init_ctrl_blk.cb24.
2451 					    hard_address[0]];
2452 					port_info->pi_hard_addr.hard_addr =
2453 					    d_id.b24;
2454 				}
2455 			} else if (ha->init_ctrl_blk.cb.firmware_options[0] &
2456 			    BIT_0) {
2457 				d_id.b.al_pa = ql_index_to_alpa[ha->
2458 				    init_ctrl_blk.cb.hard_address[0]];
2459 				port_info->pi_hard_addr.hard_addr = d_id.b24;
2460 			}
2461 
2462 			/* Set the node id data */
2463 			if (ql_get_rnid_params(ha,
2464 			    sizeof (port_info->pi_rnid_params.params),
2465 			    (caddr_t)&port_info->pi_rnid_params.params) ==
2466 			    QL_SUCCESS) {
2467 				port_info->pi_rnid_params.status = FC_SUCCESS;
2468 			} else {
2469 				port_info->pi_rnid_params.status = FC_FAILURE;
2470 			}
2471 
2472 			/* Populate T11 FC-HBA details */
2473 			ql_populate_hba_fru_details(ha, port_info);
2474 			ha->pi_attrs = kmem_zalloc(sizeof (fca_port_attrs_t),
2475 			    KM_SLEEP);
2476 			if (ha->pi_attrs != NULL) {
2477 				bcopy(&port_info->pi_attrs, ha->pi_attrs,
2478 				    sizeof (fca_port_attrs_t));
2479 			}
2480 		} else {
2481 			port_info->pi_rnid_params.status = FC_FAILURE;
2482 			if (ha->pi_attrs != NULL) {
2483 				bcopy(ha->pi_attrs, &port_info->pi_attrs,
2484 				    sizeof (fca_port_attrs_t));
2485 			}
2486 		}
2487 
2488 		/* Generate handle for this FCA. */
2489 		fca_handle = (opaque_t)vha;
2490 
2491 		ADAPTER_STATE_LOCK(ha);
2492 		vha->flags |= FCA_BOUND;
2493 		ADAPTER_STATE_UNLOCK(ha);
2494 		/* Set port's current state. */
2495 		port_info->pi_port_state = vha->state;
2496 	}
2497 
2498 	QL_PRINT_10(CE_CONT, "(%d,%d): done, pi_port_state=%xh, "
2499 	    "pi_s_id.port_id=%xh\n", ha->instance, ha->vp_index,
2500 	    port_info->pi_port_state, port_info->pi_s_id.port_id);
2501 
2502 	return (fca_handle);
2503 }
2504 
2505 /*
2506  * ql_unbind_port
2507  *	To unbind a Fibre Channel Adapter from an FC Port driver.
2508  *
2509  * Input:
2510  *	fca_handle = handle setup by ql_bind_port().
2511  *
2512  * Context:
2513  *	Kernel context.
2514  */
2515 static void
ql_unbind_port(opaque_t fca_handle)2516 ql_unbind_port(opaque_t fca_handle)
2517 {
2518 	ql_adapter_state_t	*ha;
2519 	ql_tgt_t		*tq;
2520 	uint32_t		flgs;
2521 
2522 	ha = ql_fca_handle_to_state(fca_handle);
2523 	if (ha == NULL) {
2524 		/*EMPTY*/
2525 		QL_PRINT_2(CE_CONT, "failed, no adapter=%ph\n",
2526 		    (void *)fca_handle);
2527 	} else {
2528 		QL_PRINT_10(CE_CONT, "(%d,%d): started\n", ha->instance,
2529 		    ha->vp_index);
2530 
2531 		if (!(ha->flags & FCA_BOUND)) {
2532 			/*EMPTY*/
2533 			QL_PRINT_2(CE_CONT, "(%d): port=%d already unbound\n",
2534 			    ha->instance, ha->vp_index);
2535 		} else {
2536 			if (ha->vp_index != 0 && ha->flags & VP_ENABLED) {
2537 				if ((tq = ql_loop_id_to_queue(ha,
2538 				    FL_PORT_24XX_HDL)) != NULL) {
2539 					(void) ql_logout_fabric_port(ha, tq);
2540 				}
2541 				(void) ql_vport_control(ha, (uint8_t)
2542 				    (CFG_IST(ha, CFG_CTRL_2425) ?
2543 				    VPC_DISABLE_INIT : VPC_DISABLE_LOGOUT));
2544 				flgs = FCA_BOUND | VP_ENABLED;
2545 			} else {
2546 				flgs = FCA_BOUND;
2547 			}
2548 			ADAPTER_STATE_LOCK(ha);
2549 			ha->flags &= ~flgs;
2550 			ADAPTER_STATE_UNLOCK(ha);
2551 		}
2552 
2553 		QL_PRINT_10(CE_CONT, "(%d,%d): done\n", ha->instance,
2554 		    ha->vp_index);
2555 	}
2556 }
2557 
2558 /*
2559  * ql_init_pkt
2560  *	Initialize FCA portion of packet.
2561  *
2562  * Input:
2563  *	fca_handle = handle setup by ql_bind_port().
2564  *	pkt = pointer to fc_packet.
2565  *
2566  * Returns:
2567  *	FC_SUCCESS - the packet has successfully been initialized.
2568  *	FC_UNBOUND - the fca_handle specified is not bound.
2569  *	FC_NOMEM - the FCA failed initialization due to an allocation error.
2570  *	FC_FAILURE - the FCA failed initialization for undisclosed reasons
2571  *
2572  * Context:
2573  *	Kernel context.
2574  */
2575 /* ARGSUSED */
2576 static int
ql_init_pkt(opaque_t fca_handle,fc_packet_t * pkt,int sleep)2577 ql_init_pkt(opaque_t fca_handle, fc_packet_t *pkt, int sleep)
2578 {
2579 	ql_adapter_state_t	*ha;
2580 	ql_srb_t		*sp;
2581 	int			rval = FC_SUCCESS;
2582 
2583 	ha = ql_fca_handle_to_state(fca_handle);
2584 	if (ha == NULL) {
2585 		QL_PRINT_2(CE_CONT, "failed, no adapter=%ph\n",
2586 		    (void *)fca_handle);
2587 		return (FC_UNBOUND);
2588 	}
2589 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
2590 
2591 	sp = (ql_srb_t *)pkt->pkt_fca_private;
2592 	sp->flags = 0;
2593 
2594 	/* init cmd links */
2595 	sp->cmd.base_address = sp;
2596 	sp->cmd.prev = NULL;
2597 	sp->cmd.next = NULL;
2598 	sp->cmd.head = NULL;
2599 
2600 	/* init watchdog links */
2601 	sp->wdg.base_address = sp;
2602 	sp->wdg.prev = NULL;
2603 	sp->wdg.next = NULL;
2604 	sp->wdg.head = NULL;
2605 	sp->pkt = pkt;
2606 	sp->ha = ha;
2607 	sp->magic_number = QL_FCA_BRAND;
2608 	sp->sg_dma.dma_handle = NULL;
2609 #ifndef __sparc
2610 	if (CFG_IST(ha, CFG_CTRL_8021)) {
2611 		/* Setup DMA for scatter gather list. */
2612 		sp->sg_dma.size = sizeof (cmd6_2400_dma_t);
2613 		sp->sg_dma.type = LITTLE_ENDIAN_DMA;
2614 		sp->sg_dma.cookie_count = 1;
2615 		sp->sg_dma.alignment = 64;
2616 		if (ql_alloc_phys(ha, &sp->sg_dma, KM_SLEEP) != QL_SUCCESS) {
2617 			rval = FC_NOMEM;
2618 		}
2619 	}
2620 #endif	/* __sparc */
2621 
2622 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
2623 
2624 	return (rval);
2625 }
2626 
2627 /*
2628  * ql_un_init_pkt
2629  *	Release all local resources bound to packet.
2630  *
2631  * Input:
2632  *	fca_handle = handle setup by ql_bind_port().
2633  *	pkt = pointer to fc_packet.
2634  *
2635  * Returns:
2636  *	FC_SUCCESS - the packet has successfully been invalidated.
2637  *	FC_UNBOUND - the fca_handle specified is not bound.
2638  *	FC_BADPACKET - the packet has not been initialized or has
2639  *			already been freed by this FCA.
2640  *
2641  * Context:
2642  *	Kernel context.
2643  */
2644 static int
ql_un_init_pkt(opaque_t fca_handle,fc_packet_t * pkt)2645 ql_un_init_pkt(opaque_t fca_handle, fc_packet_t *pkt)
2646 {
2647 	ql_adapter_state_t *ha;
2648 	int rval;
2649 	ql_srb_t *sp;
2650 
2651 	ha = ql_fca_handle_to_state(fca_handle);
2652 	if (ha == NULL) {
2653 		QL_PRINT_2(CE_CONT, "failed, no adapter=%ph\n",
2654 		    (void *)fca_handle);
2655 		return (FC_UNBOUND);
2656 	}
2657 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
2658 
2659 	sp = (ql_srb_t *)pkt->pkt_fca_private;
2660 
2661 	if (sp->magic_number != QL_FCA_BRAND) {
2662 		EL(ha, "failed, FC_BADPACKET\n");
2663 		rval = FC_BADPACKET;
2664 	} else {
2665 		sp->magic_number = NULL;
2666 		ql_free_phys(ha, &sp->sg_dma);
2667 		rval = FC_SUCCESS;
2668 	}
2669 
2670 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
2671 
2672 	return (rval);
2673 }
2674 
2675 /*
2676  * ql_els_send
2677  *	Issue a extended link service request.
2678  *
2679  * Input:
2680  *	fca_handle = handle setup by ql_bind_port().
2681  *	pkt = pointer to fc_packet.
2682  *
2683  * Returns:
2684  *	FC_SUCCESS - the command was successful.
2685  *	FC_ELS_FREJECT - the command was rejected by a Fabric.
2686  *	FC_ELS_PREJECT - the command was rejected by an N-port.
2687  *	FC_TRANSPORT_ERROR - a transport error occurred.
2688  *	FC_UNBOUND - the fca_handle specified is not bound.
2689  *	FC_ELS_BAD - the FCA can not issue the requested ELS.
2690  *
2691  * Context:
2692  *	Kernel context.
2693  */
2694 static int
ql_els_send(opaque_t fca_handle,fc_packet_t * pkt)2695 ql_els_send(opaque_t fca_handle, fc_packet_t *pkt)
2696 {
2697 	ql_adapter_state_t	*ha;
2698 	int			rval;
2699 	clock_t			timer = drv_usectohz(30000000);
2700 	ls_code_t		els;
2701 	la_els_rjt_t		rjt;
2702 	ql_srb_t		*sp = (ql_srb_t *)pkt->pkt_fca_private;
2703 
2704 	/* Verify proper command. */
2705 	ha = ql_cmd_setup(fca_handle, pkt, &rval);
2706 	if (ha == NULL) {
2707 		QL_PRINT_2(CE_CONT, "failed, ql_cmd_setup=%xh, fcah=%ph\n",
2708 		    rval, fca_handle);
2709 		return (FC_INVALID_REQUEST);
2710 	}
2711 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
2712 
2713 	/* Wait for suspension to end. */
2714 	TASK_DAEMON_LOCK(ha);
2715 	while (ha->task_daemon_flags & QL_SUSPENDED) {
2716 		ha->task_daemon_flags |= SUSPENDED_WAKEUP_FLG;
2717 
2718 		/* 30 seconds from now */
2719 		if (cv_reltimedwait(&ha->pha->cv_dr_suspended,
2720 		    &ha->pha->task_daemon_mutex, timer, TR_CLOCK_TICK) == -1) {
2721 			/*
2722 			 * The timeout time 'timer' was
2723 			 * reached without the condition
2724 			 * being signaled.
2725 			 */
2726 			pkt->pkt_state = FC_PKT_TRAN_BSY;
2727 			pkt->pkt_reason = FC_REASON_XCHG_BSY;
2728 
2729 			/* Release task daemon lock. */
2730 			TASK_DAEMON_UNLOCK(ha);
2731 
2732 			EL(ha, "QL_SUSPENDED failed=%xh\n",
2733 			    QL_FUNCTION_TIMEOUT);
2734 			return (FC_TRAN_BUSY);
2735 		}
2736 	}
2737 	/* Release task daemon lock. */
2738 	TASK_DAEMON_UNLOCK(ha);
2739 
2740 	/* Setup response header. */
2741 	bcopy((void *)&pkt->pkt_cmd_fhdr, (void *)&pkt->pkt_resp_fhdr,
2742 	    sizeof (fc_frame_hdr_t));
2743 
2744 	if (pkt->pkt_rsplen) {
2745 		bzero((void *)pkt->pkt_resp, pkt->pkt_rsplen);
2746 	}
2747 
2748 	pkt->pkt_resp_fhdr.d_id = ha->d_id.b24;
2749 	pkt->pkt_resp_fhdr.s_id = pkt->pkt_cmd_fhdr.d_id;
2750 	pkt->pkt_resp_fhdr.r_ctl = R_CTL_EXTENDED_SVC |
2751 	    R_CTL_SOLICITED_CONTROL;
2752 	pkt->pkt_resp_fhdr.f_ctl = F_CTL_XCHG_CONTEXT | F_CTL_LAST_SEQ |
2753 	    F_CTL_END_SEQ;
2754 
2755 	sp->flags &= ~(SRB_UB_CALLBACK | SRB_UB_RSCN | SRB_UB_FCP |
2756 	    SRB_FCP_CMD_PKT | SRB_FCP_DATA_PKT | SRB_FCP_RSP_PKT |
2757 	    SRB_IP_PKT | SRB_COMMAND_TIMEOUT | SRB_UB_ACQUIRED | SRB_MS_PKT);
2758 
2759 	sp->flags |= SRB_ELS_PKT;
2760 
2761 	/* map the type of ELS to a function */
2762 	ddi_rep_get8(pkt->pkt_cmd_acc, (uint8_t *)&els,
2763 	    (uint8_t *)pkt->pkt_cmd, sizeof (els), DDI_DEV_AUTOINCR);
2764 
2765 #if 0
2766 	QL_PRINT_3(CE_CONT, "(%d): command fhdr:\n", ha->instance);
2767 	QL_DUMP_3((uint8_t *)&pkt->pkt_cmd_fhdr, 32,
2768 	    sizeof (fc_frame_hdr_t) / 4);
2769 	QL_PRINT_3(CE_CONT, "(%d): command:\n", ha->instance);
2770 	QL_DUMP_3((uint8_t *)&els, 32, sizeof (els) / 4);
2771 #endif
2772 
2773 	sp->iocb = ha->els_cmd;
2774 	sp->req_cnt = 1;
2775 
2776 	switch (els.ls_code) {
2777 	case LA_ELS_RJT:
2778 	case LA_ELS_ACC:
2779 		EL(ha, "LA_ELS_RJT\n");
2780 		pkt->pkt_state = FC_PKT_SUCCESS;
2781 		rval = FC_SUCCESS;
2782 		break;
2783 	case LA_ELS_PLOGI:
2784 	case LA_ELS_PDISC:
2785 		rval = ql_els_plogi(ha, pkt);
2786 		break;
2787 	case LA_ELS_FLOGI:
2788 	case LA_ELS_FDISC:
2789 		rval = ql_els_flogi(ha, pkt);
2790 		break;
2791 	case LA_ELS_LOGO:
2792 		rval = ql_els_logo(ha, pkt);
2793 		break;
2794 	case LA_ELS_PRLI:
2795 		rval = ql_els_prli(ha, pkt);
2796 		break;
2797 	case LA_ELS_PRLO:
2798 		rval = ql_els_prlo(ha, pkt);
2799 		break;
2800 	case LA_ELS_ADISC:
2801 		rval = ql_els_adisc(ha, pkt);
2802 		break;
2803 	case LA_ELS_LINIT:
2804 		rval = ql_els_linit(ha, pkt);
2805 		break;
2806 	case LA_ELS_LPC:
2807 		rval = ql_els_lpc(ha, pkt);
2808 		break;
2809 	case LA_ELS_LSTS:
2810 		rval = ql_els_lsts(ha, pkt);
2811 		break;
2812 	case LA_ELS_SCR:
2813 		rval = ql_els_scr(ha, pkt);
2814 		break;
2815 	case LA_ELS_RSCN:
2816 		rval = ql_els_rscn(ha, pkt);
2817 		break;
2818 	case LA_ELS_FARP_REQ:
2819 		rval = ql_els_farp_req(ha, pkt);
2820 		break;
2821 	case LA_ELS_FARP_REPLY:
2822 		rval = ql_els_farp_reply(ha, pkt);
2823 		break;
2824 	case LA_ELS_RLS:
2825 		rval = ql_els_rls(ha, pkt);
2826 		break;
2827 	case LA_ELS_RNID:
2828 		rval = ql_els_rnid(ha, pkt);
2829 		break;
2830 	default:
2831 		EL(ha, "LA_ELS_RJT, FC_REASON_CMD_UNSUPPORTED=%xh\n",
2832 		    els.ls_code);
2833 		/* Build RJT. */
2834 		bzero(&rjt, sizeof (rjt));
2835 		rjt.ls_code.ls_code = LA_ELS_RJT;
2836 		rjt.reason = FC_REASON_CMD_UNSUPPORTED;
2837 
2838 		ddi_rep_put8(pkt->pkt_resp_acc, (uint8_t *)&rjt,
2839 		    (uint8_t *)pkt->pkt_resp, sizeof (rjt), DDI_DEV_AUTOINCR);
2840 
2841 		pkt->pkt_state = FC_PKT_LOCAL_RJT;
2842 		pkt->pkt_reason = FC_REASON_UNSUPPORTED;
2843 		rval = FC_SUCCESS;
2844 		break;
2845 	}
2846 
2847 #if 0
2848 	QL_PRINT_3(CE_CONT, "(%d): response fhdr:\n", ha->instance);
2849 	QL_DUMP_3((uint8_t *)&pkt->pkt_resp_fhdr, 32,
2850 	    sizeof (fc_frame_hdr_t) / 4);
2851 #endif
2852 	/*
2853 	 * Return success if the srb was consumed by an iocb. The packet
2854 	 * completion callback will be invoked by the response handler.
2855 	 */
2856 	if (rval == QL_CONSUMED) {
2857 		rval = FC_SUCCESS;
2858 	} else if (rval == FC_SUCCESS &&
2859 	    !(pkt->pkt_tran_flags & FC_TRAN_NO_INTR) && pkt->pkt_comp) {
2860 		/* Do command callback only if no error */
2861 		ql_awaken_task_daemon(ha, sp, 0, 0);
2862 	}
2863 
2864 	if (rval != FC_SUCCESS) {
2865 		EL(ha, "failed, rval = %xh\n", rval);
2866 	} else {
2867 		/*EMPTY*/
2868 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
2869 	}
2870 	return (rval);
2871 }
2872 
2873 /*
2874  * ql_get_cap
2875  *	Export FCA hardware and software capabilities.
2876  *
2877  * Input:
2878  *	fca_handle = handle setup by ql_bind_port().
2879  *	cap = pointer to the capabilities string.
2880  *	ptr = buffer pointer for return capability.
2881  *
2882  * Returns:
2883  *	FC_CAP_ERROR - no such capability
2884  *	FC_CAP_FOUND - the capability was returned and cannot be set
2885  *	FC_CAP_SETTABLE - the capability was returned and can be set
2886  *	FC_UNBOUND - the fca_handle specified is not bound.
2887  *
2888  * Context:
2889  *	Kernel context.
2890  */
2891 static int
ql_get_cap(opaque_t fca_handle,char * cap,void * ptr)2892 ql_get_cap(opaque_t fca_handle, char *cap, void *ptr)
2893 {
2894 	ql_adapter_state_t	*ha;
2895 	int			rval;
2896 	uint32_t		*rptr = (uint32_t *)ptr;
2897 
2898 	ha = ql_fca_handle_to_state(fca_handle);
2899 	if (ha == NULL) {
2900 		QL_PRINT_2(CE_CONT, "failed, no adapter=%ph\n",
2901 		    (void *)fca_handle);
2902 		return (FC_UNBOUND);
2903 	}
2904 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
2905 
2906 	if (strcmp(cap, FC_NODE_WWN) == 0) {
2907 		bcopy((void *)&ha->loginparams.node_ww_name.raw_wwn[0],
2908 		    ptr, 8);
2909 		rval = FC_CAP_FOUND;
2910 	} else if (strcmp(cap, FC_LOGIN_PARAMS) == 0) {
2911 		bcopy((void *)&ha->loginparams, ptr,
2912 		    sizeof (la_els_logi_t));
2913 		rval = FC_CAP_FOUND;
2914 	} else if (strcmp(cap, FC_CAP_UNSOL_BUF) == 0) {
2915 		*rptr = (uint32_t)QL_UB_LIMIT;
2916 		rval = FC_CAP_FOUND;
2917 	} else if (strcmp(cap, FC_CAP_NOSTREAM_ON_UNALIGN_BUF) == 0) {
2918 
2919 		dev_info_t	*psydip = NULL;
2920 #ifdef __sparc
2921 		/*
2922 		 * Disable streaming for certain 2 chip adapters
2923 		 * below Psycho to handle Psycho byte hole issue.
2924 		 */
2925 		if ((CFG_IST(ha, CFG_MULTI_CHIP_ADAPTER)) &&
2926 		    (!CFG_IST(ha, CFG_SBUS_CARD))) {
2927 			for (psydip = ddi_get_parent(ha->dip); psydip;
2928 			    psydip = ddi_get_parent(psydip)) {
2929 				if (strcmp(ddi_driver_name(psydip),
2930 				    "pcipsy") == 0) {
2931 					break;
2932 				}
2933 			}
2934 		}
2935 #endif	/* __sparc */
2936 
2937 		if (psydip) {
2938 			*rptr = (uint32_t)FC_NO_STREAMING;
2939 			EL(ha, "No Streaming\n");
2940 		} else {
2941 			*rptr = (uint32_t)FC_ALLOW_STREAMING;
2942 			EL(ha, "Allow Streaming\n");
2943 		}
2944 		rval = FC_CAP_FOUND;
2945 	} else if (strcmp(cap, FC_CAP_PAYLOAD_SIZE) == 0) {
2946 		if (CFG_IST(ha, CFG_CTRL_24258081)) {
2947 			*rptr = (uint32_t)CHAR_TO_SHORT(
2948 			    ha->init_ctrl_blk.cb24.max_frame_length[0],
2949 			    ha->init_ctrl_blk.cb24.max_frame_length[1]);
2950 		} else {
2951 			*rptr = (uint32_t)CHAR_TO_SHORT(
2952 			    ha->init_ctrl_blk.cb.max_frame_length[0],
2953 			    ha->init_ctrl_blk.cb.max_frame_length[1]);
2954 		}
2955 		rval = FC_CAP_FOUND;
2956 	} else if (strcmp(cap, FC_CAP_POST_RESET_BEHAVIOR) == 0) {
2957 		*rptr = FC_RESET_RETURN_ALL;
2958 		rval = FC_CAP_FOUND;
2959 	} else if (strcmp(cap, FC_CAP_FCP_DMA) == 0) {
2960 		*rptr = FC_NO_DVMA_SPACE;
2961 		rval = FC_CAP_FOUND;
2962 	} else {
2963 		EL(ha, "unknown=%s, FC_CAP_ERROR\n", cap);
2964 		rval = FC_CAP_ERROR;
2965 	}
2966 
2967 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
2968 
2969 	return (rval);
2970 }
2971 
2972 /*
2973  * ql_set_cap
2974  *	Allow the FC Transport to set FCA capabilities if possible.
2975  *
2976  * Input:
2977  *	fca_handle = handle setup by ql_bind_port().
2978  *	cap = pointer to the capabilities string.
2979  *	ptr = buffer pointer for capability.
2980  *
2981  * Returns:
2982  *	FC_CAP_ERROR - no such capability
2983  *	FC_CAP_FOUND - the capability cannot be set by the FC Transport.
2984  *	FC_CAP_SETTABLE - the capability was successfully set.
2985  *	FC_UNBOUND - the fca_handle specified is not bound.
2986  *
2987  * Context:
2988  *	Kernel context.
2989  */
2990 /* ARGSUSED */
2991 static int
ql_set_cap(opaque_t fca_handle,char * cap,void * ptr)2992 ql_set_cap(opaque_t fca_handle, char *cap, void *ptr)
2993 {
2994 	ql_adapter_state_t	*ha;
2995 	int			rval;
2996 
2997 	ha = ql_fca_handle_to_state(fca_handle);
2998 	if (ha == NULL) {
2999 		QL_PRINT_2(CE_CONT, "failed, no adapter=%ph\n",
3000 		    (void *)fca_handle);
3001 		return (FC_UNBOUND);
3002 	}
3003 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
3004 
3005 	if (strcmp(cap, FC_NODE_WWN) == 0) {
3006 		rval = FC_CAP_FOUND;
3007 	} else if (strcmp(cap, FC_LOGIN_PARAMS) == 0) {
3008 		rval = FC_CAP_FOUND;
3009 	} else if (strcmp(cap, FC_CAP_UNSOL_BUF) == 0) {
3010 		rval = FC_CAP_FOUND;
3011 	} else if (strcmp(cap, FC_CAP_PAYLOAD_SIZE) == 0) {
3012 		rval = FC_CAP_FOUND;
3013 	} else if (strcmp(cap, FC_CAP_POST_RESET_BEHAVIOR) == 0) {
3014 		rval = FC_CAP_FOUND;
3015 	} else {
3016 		EL(ha, "unknown=%s, FC_CAP_ERROR\n", cap);
3017 		rval = FC_CAP_ERROR;
3018 	}
3019 
3020 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
3021 
3022 	return (rval);
3023 }
3024 
3025 /*
3026  * ql_getmap
3027  *	Request of Arbitrated Loop (AL-PA) map.
3028  *
3029  * Input:
3030  *	fca_handle = handle setup by ql_bind_port().
3031  *	mapbuf= buffer pointer for map.
3032  *
3033  * Returns:
3034  *	FC_OLDPORT - the specified port is not operating in loop mode.
3035  *	FC_OFFLINE - the specified port is not online.
3036  *	FC_NOMAP - there is no loop map available for this port.
3037  *	FC_UNBOUND - the fca_handle specified is not bound.
3038  *	FC_SUCCESS - a valid map has been placed in mapbuf.
3039  *
3040  * Context:
3041  *	Kernel context.
3042  */
3043 static int
ql_getmap(opaque_t fca_handle,fc_lilpmap_t * mapbuf)3044 ql_getmap(opaque_t fca_handle, fc_lilpmap_t *mapbuf)
3045 {
3046 	ql_adapter_state_t	*ha;
3047 	clock_t			timer = drv_usectohz(30000000);
3048 	int			rval = FC_SUCCESS;
3049 
3050 	ha = ql_fca_handle_to_state(fca_handle);
3051 	if (ha == NULL) {
3052 		QL_PRINT_2(CE_CONT, "failed, no adapter=%ph\n",
3053 		    (void *)fca_handle);
3054 		return (FC_UNBOUND);
3055 	}
3056 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
3057 
3058 	mapbuf->lilp_magic = (uint16_t)MAGIC_LIRP;
3059 	mapbuf->lilp_myalpa = ha->d_id.b.al_pa;
3060 
3061 	/* Wait for suspension to end. */
3062 	TASK_DAEMON_LOCK(ha);
3063 	while (ha->task_daemon_flags & QL_SUSPENDED) {
3064 		ha->task_daemon_flags |= SUSPENDED_WAKEUP_FLG;
3065 
3066 		/* 30 seconds from now */
3067 		if (cv_reltimedwait(&ha->pha->cv_dr_suspended,
3068 		    &ha->pha->task_daemon_mutex, timer, TR_CLOCK_TICK) == -1) {
3069 			/*
3070 			 * The timeout time 'timer' was
3071 			 * reached without the condition
3072 			 * being signaled.
3073 			 */
3074 
3075 			/* Release task daemon lock. */
3076 			TASK_DAEMON_UNLOCK(ha);
3077 
3078 			EL(ha, "QL_SUSPENDED failed, FC_TRAN_BUSY\n");
3079 			return (FC_TRAN_BUSY);
3080 		}
3081 	}
3082 	/* Release task daemon lock. */
3083 	TASK_DAEMON_UNLOCK(ha);
3084 
3085 	if (ql_get_loop_position_map(ha, LOOP_POSITION_MAP_SIZE,
3086 	    (caddr_t)&mapbuf->lilp_length) != QL_SUCCESS) {
3087 		/*
3088 		 * Now, since transport drivers cosider this as an
3089 		 * offline condition, let's wait for few seconds
3090 		 * for any loop transitions before we reset the.
3091 		 * chip and restart all over again.
3092 		 */
3093 		ql_delay(ha, 2000000);
3094 		EL(ha, "failed, FC_NOMAP\n");
3095 		rval = FC_NOMAP;
3096 	} else {
3097 		/*EMPTY*/
3098 		QL_PRINT_3(CE_CONT, "(%d): my_alpa %xh len %xh "
3099 		    "data %xh %xh %xh %xh\n", ha->instance,
3100 		    mapbuf->lilp_myalpa, mapbuf->lilp_length,
3101 		    mapbuf->lilp_alpalist[0], mapbuf->lilp_alpalist[1],
3102 		    mapbuf->lilp_alpalist[2], mapbuf->lilp_alpalist[3]);
3103 	}
3104 
3105 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
3106 #if 0
3107 	QL_DUMP_3((uint8_t *)mapbuf, 8, sizeof (fc_lilpmap_t));
3108 #endif
3109 	return (rval);
3110 }
3111 
3112 /*
3113  * ql_transport
3114  *	Issue an I/O request. Handles all regular requests.
3115  *
3116  * Input:
3117  *	fca_handle = handle setup by ql_bind_port().
3118  *	pkt = pointer to fc_packet.
3119  *
3120  * Returns:
3121  *	FC_SUCCESS - the packet was accepted for transport.
3122  *	FC_TRANSPORT_ERROR - a transport error occurred.
3123  *	FC_BADPACKET - the packet to be transported had not been
3124  *			initialized by this FCA.
3125  *	FC_UNBOUND - the fca_handle specified is not bound.
3126  *
3127  * Context:
3128  *	Kernel context.
3129  */
3130 static int
ql_transport(opaque_t fca_handle,fc_packet_t * pkt)3131 ql_transport(opaque_t fca_handle, fc_packet_t *pkt)
3132 {
3133 	ql_adapter_state_t	*ha;
3134 	int			rval = FC_TRANSPORT_ERROR;
3135 	ql_srb_t		*sp = (ql_srb_t *)pkt->pkt_fca_private;
3136 
3137 	/* Verify proper command. */
3138 	ha = ql_cmd_setup(fca_handle, pkt, &rval);
3139 	if (ha == NULL) {
3140 		QL_PRINT_2(CE_CONT, "failed, ql_cmd_setup=%xh, fcah=%ph\n",
3141 		    rval, fca_handle);
3142 		return (rval);
3143 	}
3144 	QL_PRINT_3(CE_CONT, "(%d): started command:\n", ha->instance);
3145 #if 0
3146 	QL_DUMP_3((uint8_t *)&pkt->pkt_cmd_fhdr, 32,
3147 	    sizeof (fc_frame_hdr_t) / 4);
3148 	QL_PRINT_3(CE_CONT, "(%d): command:\n", ha->instance);
3149 	QL_DUMP_3((uint8_t *)pkt->pkt_cmd, 8, pkt->pkt_cmdlen);
3150 #endif
3151 
3152 	/* Reset SRB flags. */
3153 	sp->flags &= ~(SRB_ISP_STARTED | SRB_ISP_COMPLETED | SRB_RETRY |
3154 	    SRB_POLL | SRB_WATCHDOG_ENABLED | SRB_ABORT | SRB_UB_CALLBACK |
3155 	    SRB_UB_RSCN | SRB_UB_FCP | SRB_FCP_CMD_PKT | SRB_FCP_DATA_PKT |
3156 	    SRB_FCP_RSP_PKT | SRB_IP_PKT | SRB_GENERIC_SERVICES_PKT |
3157 	    SRB_COMMAND_TIMEOUT | SRB_ABORTING | SRB_IN_DEVICE_QUEUE |
3158 	    SRB_IN_TOKEN_ARRAY | SRB_UB_FREE_REQUESTED | SRB_UB_ACQUIRED |
3159 	    SRB_MS_PKT | SRB_ELS_PKT);
3160 
3161 	pkt->pkt_resp_fhdr.d_id = ha->d_id.b24;
3162 	pkt->pkt_resp_fhdr.r_ctl = R_CTL_STATUS;
3163 	pkt->pkt_resp_fhdr.s_id = pkt->pkt_cmd_fhdr.d_id;
3164 	pkt->pkt_resp_fhdr.f_ctl = pkt->pkt_cmd_fhdr.f_ctl;
3165 	pkt->pkt_resp_fhdr.type = pkt->pkt_cmd_fhdr.type;
3166 
3167 	switch (pkt->pkt_cmd_fhdr.r_ctl) {
3168 	case R_CTL_COMMAND:
3169 		if (pkt->pkt_cmd_fhdr.type == FC_TYPE_SCSI_FCP) {
3170 			sp->flags |= SRB_FCP_CMD_PKT;
3171 			rval = ql_fcp_scsi_cmd(ha, pkt, sp);
3172 		}
3173 		break;
3174 
3175 	default:
3176 		/* Setup response header and buffer. */
3177 		if (pkt->pkt_rsplen) {
3178 			bzero((void *)pkt->pkt_resp, pkt->pkt_rsplen);
3179 		}
3180 
3181 		switch (pkt->pkt_cmd_fhdr.r_ctl) {
3182 		case R_CTL_UNSOL_DATA:
3183 			if (pkt->pkt_cmd_fhdr.type == FC_TYPE_IS8802_SNAP) {
3184 				sp->flags |= SRB_IP_PKT;
3185 				rval = ql_fcp_ip_cmd(ha, pkt, sp);
3186 			}
3187 			break;
3188 
3189 		case R_CTL_UNSOL_CONTROL:
3190 			if (pkt->pkt_cmd_fhdr.type == FC_TYPE_FC_SERVICES) {
3191 				sp->flags |= SRB_GENERIC_SERVICES_PKT;
3192 				rval = ql_fc_services(ha, pkt);
3193 			}
3194 			break;
3195 
3196 		case R_CTL_SOLICITED_DATA:
3197 		case R_CTL_STATUS:
3198 		default:
3199 			pkt->pkt_state = FC_PKT_LOCAL_RJT;
3200 			pkt->pkt_reason = FC_REASON_UNSUPPORTED;
3201 			rval = FC_TRANSPORT_ERROR;
3202 			EL(ha, "unknown, r_ctl=%xh\n",
3203 			    pkt->pkt_cmd_fhdr.r_ctl);
3204 			break;
3205 		}
3206 	}
3207 
3208 	if (rval != FC_SUCCESS) {
3209 		EL(ha, "failed, rval = %xh\n", rval);
3210 	} else {
3211 		/*EMPTY*/
3212 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
3213 	}
3214 
3215 	return (rval);
3216 }
3217 
3218 /*
3219  * ql_ub_alloc
3220  *	Allocate buffers for unsolicited exchanges.
3221  *
3222  * Input:
3223  *	fca_handle = handle setup by ql_bind_port().
3224  *	tokens = token array for each buffer.
3225  *	size = size of each buffer.
3226  *	count = pointer to number of buffers.
3227  *	type = the FC-4 type the buffers are reserved for.
3228  *		1 = Extended Link Services, 5 = LLC/SNAP
3229  *
3230  * Returns:
3231  *	FC_FAILURE - buffers could not be allocated.
3232  *	FC_TOOMANY - the FCA could not allocate the requested
3233  *			number of buffers.
3234  *	FC_SUCCESS - unsolicited buffers were allocated.
3235  *	FC_UNBOUND - the fca_handle specified is not bound.
3236  *
3237  * Context:
3238  *	Kernel context.
3239  */
3240 static int
ql_ub_alloc(opaque_t fca_handle,uint64_t tokens[],uint32_t size,uint32_t * count,uint32_t type)3241 ql_ub_alloc(opaque_t fca_handle, uint64_t tokens[], uint32_t size,
3242     uint32_t *count, uint32_t type)
3243 {
3244 	ql_adapter_state_t	*ha;
3245 	caddr_t			bufp = NULL;
3246 	fc_unsol_buf_t		*ubp;
3247 	ql_srb_t		*sp;
3248 	uint32_t		index;
3249 	uint32_t		cnt;
3250 	uint32_t		ub_array_index = 0;
3251 	int			rval = FC_SUCCESS;
3252 	int			ub_updated = FALSE;
3253 
3254 	/* Check handle. */
3255 	ha = ql_fca_handle_to_state(fca_handle);
3256 	if (ha == NULL) {
3257 		QL_PRINT_2(CE_CONT, "failed, no adapter=%ph\n",
3258 		    (void *)fca_handle);
3259 		return (FC_UNBOUND);
3260 	}
3261 	QL_PRINT_3(CE_CONT, "(%d,%d): started, count = %xh\n",
3262 	    ha->instance, ha->vp_index, *count);
3263 
3264 	QL_PM_LOCK(ha);
3265 	if (ha->power_level != PM_LEVEL_D0) {
3266 		QL_PM_UNLOCK(ha);
3267 		QL_PRINT_3(CE_CONT, "(%d,%d): down done\n", ha->instance,
3268 		    ha->vp_index);
3269 		return (FC_FAILURE);
3270 	}
3271 	QL_PM_UNLOCK(ha);
3272 
3273 	/* Acquire adapter state lock. */
3274 	ADAPTER_STATE_LOCK(ha);
3275 
3276 	/* Check the count. */
3277 	if ((*count + ha->ub_allocated) > QL_UB_LIMIT) {
3278 		*count = 0;
3279 		EL(ha, "failed, FC_TOOMANY\n");
3280 		rval = FC_TOOMANY;
3281 	}
3282 
3283 	/*
3284 	 * reset ub_array_index
3285 	 */
3286 	ub_array_index = 0;
3287 
3288 	/*
3289 	 * Now proceed to allocate any buffers required
3290 	 */
3291 	for (index = 0; index < *count && rval == FC_SUCCESS; index++) {
3292 		/* Allocate all memory needed. */
3293 		ubp = (fc_unsol_buf_t *)kmem_zalloc(sizeof (fc_unsol_buf_t),
3294 		    KM_SLEEP);
3295 		if (ubp == NULL) {
3296 			EL(ha, "failed, FC_FAILURE\n");
3297 			rval = FC_FAILURE;
3298 		} else {
3299 			sp = kmem_zalloc(sizeof (ql_srb_t), KM_SLEEP);
3300 			if (sp == NULL) {
3301 				kmem_free(ubp, sizeof (fc_unsol_buf_t));
3302 				rval = FC_FAILURE;
3303 			} else {
3304 				if (type == FC_TYPE_IS8802_SNAP) {
3305 #ifdef	__sparc
3306 					if (ql_get_dma_mem(ha,
3307 					    &sp->ub_buffer, size,
3308 					    BIG_ENDIAN_DMA,
3309 					    QL_DMA_DATA_ALIGN) != QL_SUCCESS) {
3310 						rval = FC_FAILURE;
3311 						kmem_free(ubp,
3312 						    sizeof (fc_unsol_buf_t));
3313 						kmem_free(sp,
3314 						    sizeof (ql_srb_t));
3315 					} else {
3316 						bufp = sp->ub_buffer.bp;
3317 						sp->ub_size = size;
3318 					}
3319 #else
3320 					if (ql_get_dma_mem(ha,
3321 					    &sp->ub_buffer, size,
3322 					    LITTLE_ENDIAN_DMA,
3323 					    QL_DMA_DATA_ALIGN) != QL_SUCCESS) {
3324 						rval = FC_FAILURE;
3325 						kmem_free(ubp,
3326 						    sizeof (fc_unsol_buf_t));
3327 						kmem_free(sp,
3328 						    sizeof (ql_srb_t));
3329 					} else {
3330 						bufp = sp->ub_buffer.bp;
3331 						sp->ub_size = size;
3332 					}
3333 #endif
3334 				} else {
3335 					bufp = kmem_zalloc(size, KM_SLEEP);
3336 					if (bufp == NULL) {
3337 						rval = FC_FAILURE;
3338 						kmem_free(ubp,
3339 						    sizeof (fc_unsol_buf_t));
3340 						kmem_free(sp,
3341 						    sizeof (ql_srb_t));
3342 					} else {
3343 						sp->ub_size = size;
3344 					}
3345 				}
3346 			}
3347 		}
3348 
3349 		if (rval == FC_SUCCESS) {
3350 			/* Find next available slot. */
3351 			QL_UB_LOCK(ha);
3352 			while (ha->ub_array[ub_array_index] != NULL) {
3353 				ub_array_index++;
3354 			}
3355 
3356 			ubp->ub_fca_private = (void *)sp;
3357 
3358 			/* init cmd links */
3359 			sp->cmd.base_address = sp;
3360 			sp->cmd.prev = NULL;
3361 			sp->cmd.next = NULL;
3362 			sp->cmd.head = NULL;
3363 
3364 			/* init wdg links */
3365 			sp->wdg.base_address = sp;
3366 			sp->wdg.prev = NULL;
3367 			sp->wdg.next = NULL;
3368 			sp->wdg.head = NULL;
3369 			sp->ha = ha;
3370 
3371 			ubp->ub_buffer = bufp;
3372 			ubp->ub_bufsize = size;
3373 			ubp->ub_port_handle = fca_handle;
3374 			ubp->ub_token = ub_array_index;
3375 
3376 			/* Save the token. */
3377 			tokens[index] = ub_array_index;
3378 
3379 			/* Setup FCA private information. */
3380 			sp->ub_type = type;
3381 			sp->handle = ub_array_index;
3382 			sp->flags |= SRB_UB_IN_FCA;
3383 
3384 			ha->ub_array[ub_array_index] = ubp;
3385 			ha->ub_allocated++;
3386 			ub_updated = TRUE;
3387 			QL_UB_UNLOCK(ha);
3388 		}
3389 	}
3390 
3391 	/* Release adapter state lock. */
3392 	ADAPTER_STATE_UNLOCK(ha);
3393 
3394 	/* IP buffer. */
3395 	if (ub_updated) {
3396 		if ((type == FC_TYPE_IS8802_SNAP) &&
3397 		    (!(CFG_IST(ha, (CFG_CTRL_6322 | CFG_CTRL_2581))))) {
3398 
3399 			ADAPTER_STATE_LOCK(ha);
3400 			ha->flags |= IP_ENABLED;
3401 			ADAPTER_STATE_UNLOCK(ha);
3402 
3403 			if (!(ha->flags & IP_INITIALIZED)) {
3404 				if (CFG_IST(ha, CFG_CTRL_2422)) {
3405 					ha->ip_init_ctrl_blk.cb24.mtu_size[0] =
3406 					    LSB(ql_ip_mtu);
3407 					ha->ip_init_ctrl_blk.cb24.mtu_size[1] =
3408 					    MSB(ql_ip_mtu);
3409 					ha->ip_init_ctrl_blk.cb24.buf_size[0] =
3410 					    LSB(size);
3411 					ha->ip_init_ctrl_blk.cb24.buf_size[1] =
3412 					    MSB(size);
3413 
3414 					cnt = CHAR_TO_SHORT(
3415 					    ha->ip_init_ctrl_blk.cb24.cc[0],
3416 					    ha->ip_init_ctrl_blk.cb24.cc[1]);
3417 
3418 					if (cnt < *count) {
3419 						ha->ip_init_ctrl_blk.cb24.cc[0]
3420 						    = LSB(*count);
3421 						ha->ip_init_ctrl_blk.cb24.cc[1]
3422 						    = MSB(*count);
3423 					}
3424 				} else {
3425 					ha->ip_init_ctrl_blk.cb.mtu_size[0] =
3426 					    LSB(ql_ip_mtu);
3427 					ha->ip_init_ctrl_blk.cb.mtu_size[1] =
3428 					    MSB(ql_ip_mtu);
3429 					ha->ip_init_ctrl_blk.cb.buf_size[0] =
3430 					    LSB(size);
3431 					ha->ip_init_ctrl_blk.cb.buf_size[1] =
3432 					    MSB(size);
3433 
3434 					cnt = CHAR_TO_SHORT(
3435 					    ha->ip_init_ctrl_blk.cb.cc[0],
3436 					    ha->ip_init_ctrl_blk.cb.cc[1]);
3437 
3438 					if (cnt < *count) {
3439 						ha->ip_init_ctrl_blk.cb.cc[0] =
3440 						    LSB(*count);
3441 						ha->ip_init_ctrl_blk.cb.cc[1] =
3442 						    MSB(*count);
3443 					}
3444 				}
3445 
3446 				(void) ql_initialize_ip(ha);
3447 			}
3448 			ql_isp_rcvbuf(ha);
3449 		}
3450 	}
3451 
3452 	if (rval != FC_SUCCESS) {
3453 		EL(ha, "failed=%xh\n", rval);
3454 	} else {
3455 		/*EMPTY*/
3456 		QL_PRINT_3(CE_CONT, "(%d,%d): done\n", ha->instance,
3457 		    ha->vp_index);
3458 	}
3459 	return (rval);
3460 }
3461 
3462 /*
3463  * ql_ub_free
3464  *	Free unsolicited buffers.
3465  *
3466  * Input:
3467  *	fca_handle = handle setup by ql_bind_port().
3468  *	count = number of buffers.
3469  *	tokens = token array for each buffer.
3470  *
3471  * Returns:
3472  *	FC_SUCCESS - the requested buffers have been freed.
3473  *	FC_UNBOUND - the fca_handle specified is not bound.
3474  *	FC_UB_BADTOKEN - an invalid token was encountered.
3475  *			 No buffers have been released.
3476  *
3477  * Context:
3478  *	Kernel context.
3479  */
3480 static int
ql_ub_free(opaque_t fca_handle,uint32_t count,uint64_t tokens[])3481 ql_ub_free(opaque_t fca_handle, uint32_t count, uint64_t tokens[])
3482 {
3483 	ql_adapter_state_t	*ha;
3484 	ql_srb_t		*sp;
3485 	uint32_t		index;
3486 	uint64_t		ub_array_index;
3487 	int			rval = FC_SUCCESS;
3488 
3489 	/* Check handle. */
3490 	ha = ql_fca_handle_to_state(fca_handle);
3491 	if (ha == NULL) {
3492 		QL_PRINT_2(CE_CONT, "failed, no adapter=%ph\n",
3493 		    (void *)fca_handle);
3494 		return (FC_UNBOUND);
3495 	}
3496 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
3497 
3498 	/* Acquire adapter state lock. */
3499 	ADAPTER_STATE_LOCK(ha);
3500 
3501 	/* Check all returned tokens. */
3502 	for (index = 0; index < count; index++) {
3503 		fc_unsol_buf_t	*ubp;
3504 
3505 		/* Check the token range. */
3506 		if ((ub_array_index = tokens[index]) >= QL_UB_LIMIT) {
3507 			EL(ha, "failed, FC_UB_BADTOKEN\n");
3508 			rval = FC_UB_BADTOKEN;
3509 			break;
3510 		}
3511 
3512 		/* Check the unsolicited buffer array. */
3513 		QL_UB_LOCK(ha);
3514 		ubp = ha->ub_array[ub_array_index];
3515 
3516 		if (ubp == NULL) {
3517 			EL(ha, "failed, FC_UB_BADTOKEN-2\n");
3518 			rval = FC_UB_BADTOKEN;
3519 			QL_UB_UNLOCK(ha);
3520 			break;
3521 		}
3522 
3523 		/* Check the state of the unsolicited buffer. */
3524 		sp = ha->ub_array[ub_array_index]->ub_fca_private;
3525 		sp->flags |= SRB_UB_FREE_REQUESTED;
3526 
3527 		while (!(sp->flags & SRB_UB_IN_FCA) ||
3528 		    (sp->flags & (SRB_UB_CALLBACK | SRB_UB_ACQUIRED))) {
3529 			QL_UB_UNLOCK(ha);
3530 			ADAPTER_STATE_UNLOCK(ha);
3531 			delay(drv_usectohz(100000));
3532 			ADAPTER_STATE_LOCK(ha);
3533 			QL_UB_LOCK(ha);
3534 		}
3535 		ha->ub_array[ub_array_index] = NULL;
3536 		QL_UB_UNLOCK(ha);
3537 		ql_free_unsolicited_buffer(ha, ubp);
3538 	}
3539 
3540 	if (rval == FC_SUCCESS) {
3541 		/*
3542 		 * Signal any pending hardware reset when there are
3543 		 * no more unsolicited buffers in use.
3544 		 */
3545 		if (ha->ub_allocated == 0) {
3546 			cv_broadcast(&ha->pha->cv_ub);
3547 		}
3548 	}
3549 
3550 	/* Release adapter state lock. */
3551 	ADAPTER_STATE_UNLOCK(ha);
3552 
3553 	if (rval != FC_SUCCESS) {
3554 		EL(ha, "failed=%xh\n", rval);
3555 	} else {
3556 		/*EMPTY*/
3557 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
3558 	}
3559 	return (rval);
3560 }
3561 
3562 /*
3563  * ql_ub_release
3564  *	Release unsolicited buffers from FC Transport
3565  *	to FCA for future use.
3566  *
3567  * Input:
3568  *	fca_handle = handle setup by ql_bind_port().
3569  *	count = number of buffers.
3570  *	tokens = token array for each buffer.
3571  *
3572  * Returns:
3573  *	FC_SUCCESS - the requested buffers have been released.
3574  *	FC_UNBOUND - the fca_handle specified is not bound.
3575  *	FC_UB_BADTOKEN - an invalid token was encountered.
3576  *		No buffers have been released.
3577  *
3578  * Context:
3579  *	Kernel context.
3580  */
3581 static int
ql_ub_release(opaque_t fca_handle,uint32_t count,uint64_t tokens[])3582 ql_ub_release(opaque_t fca_handle, uint32_t count, uint64_t tokens[])
3583 {
3584 	ql_adapter_state_t	*ha;
3585 	ql_srb_t		*sp;
3586 	uint32_t		index;
3587 	uint64_t		ub_array_index;
3588 	int			rval = FC_SUCCESS;
3589 	int			ub_ip_updated = FALSE;
3590 
3591 	/* Check handle. */
3592 	ha = ql_fca_handle_to_state(fca_handle);
3593 	if (ha == NULL) {
3594 		QL_PRINT_2(CE_CONT, ": failed, no adapter=%ph\n",
3595 		    (void *)fca_handle);
3596 		return (FC_UNBOUND);
3597 	}
3598 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
3599 
3600 	/* Acquire adapter state lock. */
3601 	ADAPTER_STATE_LOCK(ha);
3602 	QL_UB_LOCK(ha);
3603 
3604 	/* Check all returned tokens. */
3605 	for (index = 0; index < count; index++) {
3606 		/* Check the token range. */
3607 		if ((ub_array_index = tokens[index]) >= QL_UB_LIMIT) {
3608 			EL(ha, "failed, FC_UB_BADTOKEN\n");
3609 			rval = FC_UB_BADTOKEN;
3610 			break;
3611 		}
3612 
3613 		/* Check the unsolicited buffer array. */
3614 		if (ha->ub_array[ub_array_index] == NULL) {
3615 			EL(ha, "failed, FC_UB_BADTOKEN-2\n");
3616 			rval = FC_UB_BADTOKEN;
3617 			break;
3618 		}
3619 
3620 		/* Check the state of the unsolicited buffer. */
3621 		sp = ha->ub_array[ub_array_index]->ub_fca_private;
3622 		if (sp->flags & SRB_UB_IN_FCA) {
3623 			EL(ha, "failed, FC_UB_BADTOKEN-3\n");
3624 			rval = FC_UB_BADTOKEN;
3625 			break;
3626 		}
3627 	}
3628 
3629 	/* If all tokens checkout, release the buffers. */
3630 	if (rval == FC_SUCCESS) {
3631 		/* Check all returned tokens. */
3632 		for (index = 0; index < count; index++) {
3633 			fc_unsol_buf_t	*ubp;
3634 
3635 			ub_array_index = tokens[index];
3636 			ubp = ha->ub_array[ub_array_index];
3637 			sp = ubp->ub_fca_private;
3638 
3639 			ubp->ub_resp_flags = 0;
3640 			sp->flags &= ~(SRB_UB_ACQUIRED | SRB_UB_CALLBACK);
3641 			sp->flags |= SRB_UB_IN_FCA;
3642 
3643 			/* IP buffer. */
3644 			if (sp->ub_type == FC_TYPE_IS8802_SNAP) {
3645 				ub_ip_updated = TRUE;
3646 			}
3647 		}
3648 	}
3649 
3650 	QL_UB_UNLOCK(ha);
3651 	/* Release adapter state lock. */
3652 	ADAPTER_STATE_UNLOCK(ha);
3653 
3654 	/*
3655 	 * XXX: We should call ql_isp_rcvbuf() to return a
3656 	 * buffer to ISP only if the number of buffers fall below
3657 	 * the low water mark.
3658 	 */
3659 	if (ub_ip_updated) {
3660 		ql_isp_rcvbuf(ha);
3661 	}
3662 
3663 	if (rval != FC_SUCCESS) {
3664 		EL(ha, "failed, rval = %xh\n", rval);
3665 	} else {
3666 		/*EMPTY*/
3667 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
3668 	}
3669 	return (rval);
3670 }
3671 
3672 /*
3673  * ql_abort
3674  *	Abort a packet.
3675  *
3676  * Input:
3677  *	fca_handle = handle setup by ql_bind_port().
3678  *	pkt = pointer to fc_packet.
3679  *	flags = KM_SLEEP flag.
3680  *
3681  * Returns:
3682  *	FC_SUCCESS - the packet has successfully aborted.
3683  *	FC_ABORTED - the packet has successfully aborted.
3684  *	FC_ABORTING - the packet is being aborted.
3685  *	FC_ABORT_FAILED - the packet could not be aborted.
3686  *	FC_TRANSPORT_ERROR - a transport error occurred while attempting
3687  *		to abort the packet.
3688  *	FC_BADEXCHANGE - no packet found.
3689  *	FC_UNBOUND - the fca_handle specified is not bound.
3690  *
3691  * Context:
3692  *	Kernel context.
3693  */
3694 static int
ql_abort(opaque_t fca_handle,fc_packet_t * pkt,int flags)3695 ql_abort(opaque_t fca_handle, fc_packet_t *pkt, int flags)
3696 {
3697 	port_id_t		d_id;
3698 	ql_link_t		*link;
3699 	ql_adapter_state_t	*ha, *pha;
3700 	ql_srb_t		*sp;
3701 	ql_tgt_t		*tq;
3702 	ql_lun_t		*lq;
3703 	int			rval = FC_ABORTED;
3704 
3705 	ha = ql_fca_handle_to_state(fca_handle);
3706 	if (ha == NULL) {
3707 		QL_PRINT_2(CE_CONT, "failed, no adapter=%ph\n",
3708 		    (void *)fca_handle);
3709 		return (FC_UNBOUND);
3710 	}
3711 
3712 	pha = ha->pha;
3713 
3714 	QL_PRINT_3(CE_CONT, "(%d,%d): started\n", ha->instance, ha->vp_index);
3715 
3716 	/* Get target queue pointer. */
3717 	d_id.b24 = pkt->pkt_cmd_fhdr.d_id;
3718 	tq = ql_d_id_to_queue(ha, d_id);
3719 
3720 	if ((tq == NULL) || (pha->task_daemon_flags & LOOP_DOWN)) {
3721 		if (tq == NULL) {
3722 			EL(ha, "failed, FC_TRANSPORT_ERROR\n");
3723 			rval = FC_TRANSPORT_ERROR;
3724 		} else {
3725 			EL(ha, "failed, FC_OFFLINE\n");
3726 			rval = FC_OFFLINE;
3727 		}
3728 		return (rval);
3729 	}
3730 
3731 	sp = (ql_srb_t *)pkt->pkt_fca_private;
3732 	lq = sp->lun_queue;
3733 
3734 	/* Set poll flag if sleep wanted. */
3735 	if (flags == KM_SLEEP) {
3736 		sp->flags |= SRB_POLL;
3737 	}
3738 
3739 	/* Acquire target queue lock. */
3740 	DEVICE_QUEUE_LOCK(tq);
3741 	REQUEST_RING_LOCK(ha);
3742 
3743 	/* If command not already started. */
3744 	if (!(sp->flags & SRB_ISP_STARTED)) {
3745 		/* Check pending queue for command. */
3746 		sp = NULL;
3747 		for (link = pha->pending_cmds.first; link != NULL;
3748 		    link = link->next) {
3749 			sp = link->base_address;
3750 			if (sp == (ql_srb_t *)pkt->pkt_fca_private) {
3751 				/* Remove srb from q. */
3752 				ql_remove_link(&pha->pending_cmds, &sp->cmd);
3753 				break;
3754 			} else {
3755 				sp = NULL;
3756 			}
3757 		}
3758 		REQUEST_RING_UNLOCK(ha);
3759 
3760 		if (sp == NULL) {
3761 			/* Check for cmd on device queue. */
3762 			for (link = lq->cmd.first; link != NULL;
3763 			    link = link->next) {
3764 				sp = link->base_address;
3765 				if (sp == (ql_srb_t *)pkt->pkt_fca_private) {
3766 					/* Remove srb from q. */
3767 					ql_remove_link(&lq->cmd, &sp->cmd);
3768 					break;
3769 				} else {
3770 					sp = NULL;
3771 				}
3772 			}
3773 		}
3774 		/* Release device lock */
3775 		DEVICE_QUEUE_UNLOCK(tq);
3776 
3777 		/* If command on target queue. */
3778 		if (sp != NULL) {
3779 			sp->flags &= ~SRB_IN_DEVICE_QUEUE;
3780 
3781 			/* Set return status */
3782 			pkt->pkt_reason = CS_ABORTED;
3783 
3784 			sp->cmd.next = NULL;
3785 			ql_done(&sp->cmd);
3786 			rval = FC_ABORTED;
3787 		} else {
3788 			EL(ha, "failed, FC_BADEXCHANGE\n");
3789 			rval = FC_BADEXCHANGE;
3790 		}
3791 	} else if (sp->flags & SRB_ISP_COMPLETED) {
3792 		/* Release device queue lock. */
3793 		REQUEST_RING_UNLOCK(ha);
3794 		DEVICE_QUEUE_UNLOCK(tq);
3795 		EL(ha, "failed, already done, FC_FAILURE\n");
3796 		rval = FC_FAILURE;
3797 	} else if ((sp->pkt->pkt_cmd_fhdr.r_ctl == R_CTL_SOLICITED_DATA) ||
3798 	    (sp->pkt->pkt_cmd_fhdr.r_ctl == R_CTL_STATUS)) {
3799 		/*
3800 		 * If here, target data/resp ctio is with Fw.
3801 		 * Since firmware is supposed to terminate such I/Os
3802 		 * with an error, we need not do any thing. If FW
3803 		 * decides not to terminate those IOs and simply keep
3804 		 * quite then we need to initiate cleanup here by
3805 		 * calling ql_done.
3806 		 */
3807 		REQUEST_RING_UNLOCK(ha);
3808 		DEVICE_QUEUE_UNLOCK(tq);
3809 		rval = FC_ABORTED;
3810 	} else {
3811 		request_t	*ep = pha->request_ring_bp;
3812 		uint16_t	cnt;
3813 
3814 		if (sp->handle != 0) {
3815 			for (cnt = 0; cnt < REQUEST_ENTRY_CNT; cnt++) {
3816 				if (sp->handle == ddi_get32(
3817 				    pha->hba_buf.acc_handle, &ep->handle)) {
3818 					ep->entry_type = INVALID_ENTRY_TYPE;
3819 					break;
3820 				}
3821 				ep++;
3822 			}
3823 		}
3824 
3825 		/* Release device queue lock. */
3826 		REQUEST_RING_UNLOCK(ha);
3827 		DEVICE_QUEUE_UNLOCK(tq);
3828 
3829 		sp->flags |= SRB_ABORTING;
3830 		(void) ql_abort_command(ha, sp);
3831 		pkt->pkt_reason = CS_ABORTED;
3832 		rval = FC_ABORTED;
3833 	}
3834 
3835 	QL_PRINT_3(CE_CONT, "(%d,%d): done\n", ha->instance, ha->vp_index);
3836 
3837 	return (rval);
3838 }
3839 
3840 /*
3841  * ql_reset
3842  *	Reset link or hardware.
3843  *
3844  * Input:
3845  *	fca_handle = handle setup by ql_bind_port().
3846  *	cmd = reset type command.
3847  *
3848  * Returns:
3849  *	FC_SUCCESS - reset has successfully finished.
3850  *	FC_UNBOUND - the fca_handle specified is not bound.
3851  *	FC_FAILURE - reset failed.
3852  *
3853  * Context:
3854  *	Kernel context.
3855  */
3856 static int
ql_reset(opaque_t fca_handle,uint32_t cmd)3857 ql_reset(opaque_t fca_handle, uint32_t cmd)
3858 {
3859 	ql_adapter_state_t	*ha;
3860 	int			rval = FC_SUCCESS, rval2;
3861 
3862 	ha = ql_fca_handle_to_state(fca_handle);
3863 	if (ha == NULL) {
3864 		QL_PRINT_2(CE_CONT, "failed, no adapter=%ph\n",
3865 		    (void *)fca_handle);
3866 		return (FC_UNBOUND);
3867 	}
3868 
3869 	QL_PRINT_3(CE_CONT, "(%d,%d): started, cmd=%d\n", ha->instance,
3870 	    ha->vp_index, cmd);
3871 
3872 	switch (cmd) {
3873 	case FC_FCA_CORE:
3874 		/* dump firmware core if specified. */
3875 		if (ha->vp_index == 0) {
3876 			if (ql_dump_firmware(ha) != QL_SUCCESS) {
3877 				EL(ha, "failed, FC_FAILURE\n");
3878 				rval = FC_FAILURE;
3879 			}
3880 		}
3881 		break;
3882 	case FC_FCA_LINK_RESET:
3883 		if (!(ha->pha->task_daemon_flags & LOOP_DOWN)) {
3884 			if (ql_loop_reset(ha) != QL_SUCCESS) {
3885 				EL(ha, "failed, FC_FAILURE-2\n");
3886 				rval = FC_FAILURE;
3887 			}
3888 		}
3889 		break;
3890 	case FC_FCA_RESET_CORE:
3891 	case FC_FCA_RESET:
3892 		/* if dump firmware core if specified. */
3893 		if (cmd == FC_FCA_RESET_CORE) {
3894 			if (ha->vp_index != 0) {
3895 				rval2 = ha->pha->task_daemon_flags & LOOP_DOWN
3896 				    ? QL_SUCCESS : ql_loop_reset(ha);
3897 			} else {
3898 				rval2 = ql_dump_firmware(ha);
3899 			}
3900 			if (rval2 != QL_SUCCESS) {
3901 				EL(ha, "failed, FC_FAILURE-3\n");
3902 				rval = FC_FAILURE;
3903 			}
3904 		}
3905 
3906 		/* Free up all unsolicited buffers. */
3907 		if (ha->ub_allocated != 0) {
3908 			/* Inform to release buffers. */
3909 			ha->state = FC_PORT_SPEED_MASK(ha->state);
3910 			ha->state |= FC_STATE_RESET_REQUESTED;
3911 			if (ha->flags & FCA_BOUND) {
3912 				(ha->bind_info.port_statec_cb)
3913 				    (ha->bind_info.port_handle,
3914 				    ha->state);
3915 			}
3916 		}
3917 
3918 		ha->state = FC_PORT_SPEED_MASK(ha->state);
3919 
3920 		/* All buffers freed */
3921 		if (ha->ub_allocated == 0) {
3922 			/* Hardware reset. */
3923 			if (cmd == FC_FCA_RESET) {
3924 				if (ha->vp_index == 0) {
3925 					(void) ql_abort_isp(ha);
3926 				} else if (!(ha->pha->task_daemon_flags &
3927 				    LOOP_DOWN)) {
3928 					(void) ql_loop_reset(ha);
3929 				}
3930 			}
3931 
3932 			/* Inform that the hardware has been reset */
3933 			ha->state |= FC_STATE_RESET;
3934 		} else {
3935 			/*
3936 			 * the port driver expects an online if
3937 			 * buffers are not freed.
3938 			 */
3939 			if (ha->topology & QL_LOOP_CONNECTION) {
3940 				ha->state |= FC_STATE_LOOP;
3941 			} else {
3942 				ha->state |= FC_STATE_ONLINE;
3943 			}
3944 		}
3945 
3946 		TASK_DAEMON_LOCK(ha);
3947 		ha->task_daemon_flags |= FC_STATE_CHANGE;
3948 		TASK_DAEMON_UNLOCK(ha);
3949 
3950 		ql_awaken_task_daemon(ha, NULL, FC_STATE_CHANGE, 0);
3951 
3952 		break;
3953 	default:
3954 		EL(ha, "unknown cmd=%xh\n", cmd);
3955 		break;
3956 	}
3957 
3958 	if (rval != FC_SUCCESS) {
3959 		EL(ha, "cmd=%xh, failed=%xh\n", cmd, rval);
3960 	} else {
3961 		/*EMPTY*/
3962 		QL_PRINT_3(CE_CONT, "(%d,%d): done\n", ha->instance,
3963 		    ha->vp_index);
3964 	}
3965 
3966 	return (rval);
3967 }
3968 
3969 /*
3970  * ql_port_manage
3971  *	Perform port management or diagnostics.
3972  *
3973  * Input:
3974  *	fca_handle = handle setup by ql_bind_port().
3975  *	cmd = pointer to command structure.
3976  *
3977  * Returns:
3978  *	FC_SUCCESS - the request completed successfully.
3979  *	FC_FAILURE - the request did not complete successfully.
3980  *	FC_UNBOUND - the fca_handle specified is not bound.
3981  *
3982  * Context:
3983  *	Kernel context.
3984  */
3985 static int
ql_port_manage(opaque_t fca_handle,fc_fca_pm_t * cmd)3986 ql_port_manage(opaque_t fca_handle, fc_fca_pm_t *cmd)
3987 {
3988 	clock_t			timer;
3989 	uint16_t		index;
3990 	uint32_t		*bp;
3991 	port_id_t		d_id;
3992 	ql_link_t		*link;
3993 	ql_adapter_state_t	*ha, *pha;
3994 	ql_tgt_t		*tq;
3995 	dma_mem_t		buffer_xmt, buffer_rcv;
3996 	size_t			length;
3997 	uint32_t		cnt;
3998 	char			buf[80];
3999 	lbp_t			*lb;
4000 	ql_mbx_data_t		mr;
4001 	app_mbx_cmd_t		*mcp;
4002 	int			i0;
4003 	uint8_t			*bptr;
4004 	int			rval2, rval = FC_SUCCESS;
4005 	uint32_t		opcode;
4006 	uint32_t		set_flags = 0;
4007 
4008 	ha = ql_fca_handle_to_state(fca_handle);
4009 	if (ha == NULL) {
4010 		QL_PRINT_2(CE_CONT, ": failed, no adapter=%ph\n",
4011 		    (void *)fca_handle);
4012 		return (FC_UNBOUND);
4013 	}
4014 	pha = ha->pha;
4015 
4016 	QL_PRINT_3(CE_CONT, "(%d): started=%xh\n", ha->instance,
4017 	    cmd->pm_cmd_code);
4018 
4019 	ql_awaken_task_daemon(ha, NULL, DRIVER_STALL, 0);
4020 
4021 	/*
4022 	 * Wait for all outstanding commands to complete
4023 	 */
4024 	index = (uint16_t)ql_wait_outstanding(ha);
4025 
4026 	if (index != MAX_OUTSTANDING_COMMANDS) {
4027 		ql_awaken_task_daemon(ha, NULL, 0, DRIVER_STALL);
4028 		ql_restart_queues(ha);
4029 		EL(ha, "failed, FC_TRAN_BUSY\n");
4030 		return (FC_TRAN_BUSY);
4031 	}
4032 
4033 	switch (cmd->pm_cmd_code) {
4034 	case FC_PORT_BYPASS:
4035 		d_id.b24 = *cmd->pm_cmd_buf;
4036 		tq = ql_d_id_to_queue(ha, d_id);
4037 		if (tq == NULL || ql_loop_port_bypass(ha, tq) != QL_SUCCESS) {
4038 			EL(ha, "failed, FC_PORT_BYPASS FC_FAILURE\n");
4039 			rval = FC_FAILURE;
4040 		}
4041 		break;
4042 	case FC_PORT_UNBYPASS:
4043 		d_id.b24 = *cmd->pm_cmd_buf;
4044 		tq = ql_d_id_to_queue(ha, d_id);
4045 		if (tq == NULL || ql_loop_port_enable(ha, tq) != QL_SUCCESS) {
4046 			EL(ha, "failed, FC_PORT_UNBYPASS FC_FAILURE\n");
4047 			rval = FC_FAILURE;
4048 		}
4049 		break;
4050 	case FC_PORT_GET_FW_REV:
4051 		(void) sprintf(buf, "%d.%d.%d", pha->fw_major_version,
4052 		    pha->fw_minor_version, pha->fw_subminor_version);
4053 		length = strlen(buf) + 1;
4054 		if (cmd->pm_data_len < length) {
4055 			cmd->pm_data_len = length;
4056 			EL(ha, "failed, FC_PORT_GET_FW_REV FC_FAILURE\n");
4057 			rval = FC_FAILURE;
4058 		} else {
4059 			(void) strcpy(cmd->pm_data_buf, buf);
4060 		}
4061 		break;
4062 
4063 	case FC_PORT_GET_FCODE_REV: {
4064 		caddr_t		fcode_ver_buf = NULL;
4065 
4066 		i0 = 0;
4067 		/*LINTED [Solaris DDI_DEV_T_ANY Lint warning]*/
4068 		rval2 = ddi_getlongprop(DDI_DEV_T_ANY, ha->dip,
4069 		    DDI_PROP_DONTPASS | DDI_PROP_CANSLEEP, "version",
4070 		    (caddr_t)&fcode_ver_buf, &i0);
4071 		length = (uint_t)i0;
4072 
4073 		if (rval2 != DDI_PROP_SUCCESS) {
4074 			EL(ha, "failed, getting version = %xh\n", rval2);
4075 			length = 20;
4076 			fcode_ver_buf = kmem_alloc(length, KM_SLEEP);
4077 			if (fcode_ver_buf != NULL) {
4078 				(void) sprintf(fcode_ver_buf,
4079 				    "NO FCODE FOUND");
4080 			}
4081 		}
4082 
4083 		if (cmd->pm_data_len < length) {
4084 			EL(ha, "length error, FC_PORT_GET_FCODE_REV "
4085 			    "dst=%ld, src=%ld\n", cmd->pm_data_len, length);
4086 			cmd->pm_data_len = length;
4087 			rval = FC_FAILURE;
4088 		} else if (fcode_ver_buf != NULL) {
4089 			bcopy((void *)fcode_ver_buf, (void *)cmd->pm_data_buf,
4090 			    length);
4091 		}
4092 
4093 		if (fcode_ver_buf != NULL) {
4094 			kmem_free(fcode_ver_buf, length);
4095 		}
4096 		break;
4097 	}
4098 
4099 	case FC_PORT_GET_DUMP:
4100 		QL_DUMP_LOCK(pha);
4101 		if (cmd->pm_data_len < (size_t)pha->risc_dump_size) {
4102 			EL(ha, "failed, FC_PORT_GET_DUMP incorrect "
4103 			    "length=%lxh\n", cmd->pm_data_len);
4104 			cmd->pm_data_len = pha->risc_dump_size;
4105 			rval = FC_FAILURE;
4106 		} else if (pha->ql_dump_state & QL_DUMPING) {
4107 			EL(ha, "failed, FC_PORT_GET_DUMP FC_TRAN_BUSY\n");
4108 			rval = FC_TRAN_BUSY;
4109 		} else if (pha->ql_dump_state & QL_DUMP_VALID) {
4110 			(void) ql_ascii_fw_dump(ha, cmd->pm_data_buf);
4111 			pha->ql_dump_state |= QL_DUMP_UPLOADED;
4112 		} else {
4113 			EL(ha, "failed, FC_PORT_GET_DUMP no dump file\n");
4114 			rval = FC_FAILURE;
4115 		}
4116 		QL_DUMP_UNLOCK(pha);
4117 		break;
4118 	case FC_PORT_FORCE_DUMP:
4119 		PORTMANAGE_LOCK(ha);
4120 		if (ql_dump_firmware(ha) != QL_SUCCESS) {
4121 			EL(ha, "failed, FC_PORT_FORCE_DUMP FC_FAILURE\n");
4122 			rval = FC_FAILURE;
4123 		}
4124 		PORTMANAGE_UNLOCK(ha);
4125 		break;
4126 	case FC_PORT_DOWNLOAD_FW:
4127 		PORTMANAGE_LOCK(ha);
4128 		if (CFG_IST(ha, CFG_CTRL_24258081)) {
4129 			if (ql_24xx_load_flash(ha, (uint8_t *)cmd->pm_data_buf,
4130 			    (uint32_t)cmd->pm_data_len,
4131 			    ha->flash_fw_addr << 2) != QL_SUCCESS) {
4132 				EL(ha, "failed, FC_PORT_DOWNLOAD_FW\n");
4133 				rval = FC_FAILURE;
4134 			}
4135 			ql_reset_chip(ha);
4136 			set_flags |= ISP_ABORT_NEEDED;
4137 		} else {
4138 			/* Save copy of the firmware. */
4139 			if (pha->risc_code != NULL) {
4140 				kmem_free(pha->risc_code, pha->risc_code_size);
4141 				pha->risc_code = NULL;
4142 				pha->risc_code_size = 0;
4143 			}
4144 
4145 			pha->risc_code = kmem_alloc(cmd->pm_data_len,
4146 			    KM_SLEEP);
4147 			if (pha->risc_code != NULL) {
4148 				pha->risc_code_size =
4149 				    (uint32_t)cmd->pm_data_len;
4150 				bcopy(cmd->pm_data_buf, pha->risc_code,
4151 				    cmd->pm_data_len);
4152 
4153 				/* Do abort to force reload. */
4154 				ql_reset_chip(ha);
4155 				if (ql_abort_isp(ha) != QL_SUCCESS) {
4156 					kmem_free(pha->risc_code,
4157 					    pha->risc_code_size);
4158 					pha->risc_code = NULL;
4159 					pha->risc_code_size = 0;
4160 					ql_reset_chip(ha);
4161 					(void) ql_abort_isp(ha);
4162 					EL(ha, "failed, FC_PORT_DOWNLOAD_FW"
4163 					    " FC_FAILURE\n");
4164 					rval = FC_FAILURE;
4165 				}
4166 			}
4167 		}
4168 		PORTMANAGE_UNLOCK(ha);
4169 		break;
4170 	case FC_PORT_GET_DUMP_SIZE:
4171 		bp = (uint32_t *)cmd->pm_data_buf;
4172 		*bp = pha->risc_dump_size;
4173 		break;
4174 	case FC_PORT_DIAG:
4175 		/*
4176 		 * Prevents concurrent diags
4177 		 */
4178 		PORTMANAGE_LOCK(ha);
4179 
4180 		/* Wait for suspension to end. */
4181 		for (timer = 0; timer < 3000 &&
4182 		    pha->task_daemon_flags & QL_LOOP_TRANSITION; timer++) {
4183 			ql_delay(ha, 10000);
4184 		}
4185 
4186 		if (pha->task_daemon_flags & QL_LOOP_TRANSITION) {
4187 			EL(ha, "failed, FC_TRAN_BUSY-2\n");
4188 			rval = FC_TRAN_BUSY;
4189 			PORTMANAGE_UNLOCK(ha);
4190 			break;
4191 		}
4192 
4193 		switch (cmd->pm_cmd_flags) {
4194 		case QL_DIAG_EXEFMW:
4195 			if (ql_start_firmware(ha) != QL_SUCCESS) {
4196 				EL(ha, "failed, QL_DIAG_EXEFMW FC_FAILURE\n");
4197 				rval = FC_FAILURE;
4198 			}
4199 			break;
4200 		case QL_DIAG_CHKCMDQUE:
4201 			for (i0 = 1, cnt = 0; i0 < MAX_OUTSTANDING_COMMANDS;
4202 			    i0++) {
4203 				cnt += (pha->outstanding_cmds[i0] != NULL);
4204 			}
4205 			if (cnt != 0) {
4206 				EL(ha, "failed, QL_DIAG_CHKCMDQUE "
4207 				    "FC_FAILURE\n");
4208 				rval = FC_FAILURE;
4209 			}
4210 			break;
4211 		case QL_DIAG_FMWCHKSUM:
4212 			if (ql_verify_checksum(ha) != QL_SUCCESS) {
4213 				EL(ha, "failed, QL_DIAG_FMWCHKSUM "
4214 				    "FC_FAILURE\n");
4215 				rval = FC_FAILURE;
4216 			}
4217 			break;
4218 		case QL_DIAG_SLFTST:
4219 			if (ql_online_selftest(ha) != QL_SUCCESS) {
4220 				EL(ha, "failed, QL_DIAG_SLFTST FC_FAILURE\n");
4221 				rval = FC_FAILURE;
4222 			}
4223 			ql_reset_chip(ha);
4224 			set_flags |= ISP_ABORT_NEEDED;
4225 			break;
4226 		case QL_DIAG_REVLVL:
4227 			if (cmd->pm_stat_len <
4228 			    sizeof (ql_adapter_revlvl_t)) {
4229 				EL(ha, "failed, QL_DIAG_REVLVL FC_NOMEM, "
4230 				    "slen=%lxh, rlvllen=%lxh\n",
4231 				    cmd->pm_stat_len,
4232 				    sizeof (ql_adapter_revlvl_t));
4233 				rval = FC_NOMEM;
4234 			} else {
4235 				bcopy((void *)&(pha->adapter_stats->revlvl),
4236 				    cmd->pm_stat_buf,
4237 				    (size_t)cmd->pm_stat_len);
4238 				cmd->pm_stat_len =
4239 				    sizeof (ql_adapter_revlvl_t);
4240 			}
4241 			break;
4242 		case QL_DIAG_LPBMBX:
4243 
4244 			if (cmd->pm_data_len != sizeof (struct app_mbx_cmd)) {
4245 				EL(ha, "failed, QL_DIAG_LPBMBX "
4246 				    "FC_INVALID_REQUEST, pmlen=%lxh, "
4247 				    "reqd=%lxh\n", cmd->pm_data_len,
4248 				    sizeof (struct app_mbx_cmd));
4249 				rval = FC_INVALID_REQUEST;
4250 				break;
4251 			}
4252 			/*
4253 			 * Don't do the wrap test on a 2200 when the
4254 			 * firmware is running.
4255 			 */
4256 			if (!CFG_IST(ha, CFG_CTRL_2200)) {
4257 				mcp = (app_mbx_cmd_t *)cmd->pm_data_buf;
4258 				mr.mb[1] = mcp->mb[1];
4259 				mr.mb[2] = mcp->mb[2];
4260 				mr.mb[3] = mcp->mb[3];
4261 				mr.mb[4] = mcp->mb[4];
4262 				mr.mb[5] = mcp->mb[5];
4263 				mr.mb[6] = mcp->mb[6];
4264 				mr.mb[7] = mcp->mb[7];
4265 
4266 				bcopy(&mr.mb[0], &mr.mb[10],
4267 				    sizeof (uint16_t) * 8);
4268 
4269 				if (ql_mbx_wrap_test(ha, &mr) != QL_SUCCESS) {
4270 					EL(ha, "failed, QL_DIAG_LPBMBX "
4271 					    "FC_FAILURE\n");
4272 					rval = FC_FAILURE;
4273 					break;
4274 				} else {
4275 					for (i0 = 1; i0 < 8; i0++) {
4276 						if (mr.mb[i0] !=
4277 						    mr.mb[i0 + 10]) {
4278 							EL(ha, "failed, "
4279 							    "QL_DIAG_LPBMBX "
4280 							    "FC_FAILURE-2\n");
4281 							rval = FC_FAILURE;
4282 							break;
4283 						}
4284 					}
4285 				}
4286 
4287 				if (rval == FC_FAILURE) {
4288 					(void) ql_flash_errlog(ha,
4289 					    FLASH_ERRLOG_ISP_ERR, 0,
4290 					    RD16_IO_REG(ha, hccr),
4291 					    RD16_IO_REG(ha, istatus));
4292 					set_flags |= ISP_ABORT_NEEDED;
4293 				}
4294 			}
4295 			break;
4296 		case QL_DIAG_LPBDTA:
4297 			/*
4298 			 * For loopback data, we receive the
4299 			 * data back in pm_stat_buf. This provides
4300 			 * the user an opportunity to compare the
4301 			 * transmitted and received data.
4302 			 *
4303 			 * NB: lb->options are:
4304 			 *	0 --> Ten bit loopback
4305 			 *	1 --> One bit loopback
4306 			 *	2 --> External loopback
4307 			 */
4308 			if (cmd->pm_data_len > 65536) {
4309 				rval = FC_TOOMANY;
4310 				EL(ha, "failed, QL_DIAG_LPBDTA "
4311 				    "FC_TOOMANY=%lxh\n", cmd->pm_data_len);
4312 				break;
4313 			}
4314 			if (ql_get_dma_mem(ha, &buffer_xmt,
4315 			    (uint32_t)cmd->pm_data_len, LITTLE_ENDIAN_DMA,
4316 			    QL_DMA_DATA_ALIGN) != QL_SUCCESS) {
4317 				EL(ha, "failed, QL_DIAG_LPBDTA FC_NOMEM\n");
4318 				rval = FC_NOMEM;
4319 				break;
4320 			}
4321 			if (ql_get_dma_mem(ha, &buffer_rcv,
4322 			    (uint32_t)cmd->pm_data_len, LITTLE_ENDIAN_DMA,
4323 			    QL_DMA_DATA_ALIGN) != QL_SUCCESS) {
4324 				EL(ha, "failed, QL_DIAG_LPBDTA FC_NOMEM-2\n");
4325 				rval = FC_NOMEM;
4326 				break;
4327 			}
4328 			ddi_rep_put8(buffer_xmt.acc_handle,
4329 			    (uint8_t *)cmd->pm_data_buf,
4330 			    (uint8_t *)buffer_xmt.bp,
4331 			    cmd->pm_data_len, DDI_DEV_AUTOINCR);
4332 
4333 			/* 22xx's adapter must be in loop mode for test. */
4334 			if (CFG_IST(ha, CFG_CTRL_2200)) {
4335 				bptr = &ha->init_ctrl_blk.cb.add_fw_opt[0];
4336 				if (ha->flags & POINT_TO_POINT ||
4337 				    (ha->task_daemon_flags & LOOP_DOWN &&
4338 				    *bptr & (BIT_6 | BIT_5 | BIT_4))) {
4339 					cnt = *bptr;
4340 					*bptr = (uint8_t)
4341 					    (*bptr & ~(BIT_6|BIT_5|BIT_4));
4342 					(void) ql_abort_isp(ha);
4343 					*bptr = (uint8_t)cnt;
4344 				}
4345 			}
4346 
4347 			/* Shutdown IP. */
4348 			if (pha->flags & IP_INITIALIZED) {
4349 				(void) ql_shutdown_ip(pha);
4350 			}
4351 
4352 			lb = (lbp_t *)cmd->pm_cmd_buf;
4353 			lb->transfer_count =
4354 			    (uint32_t)cmd->pm_data_len;
4355 			lb->transfer_segment_count = 0;
4356 			lb->receive_segment_count = 0;
4357 			lb->transfer_data_address =
4358 			    buffer_xmt.cookie.dmac_address;
4359 			lb->receive_data_address =
4360 			    buffer_rcv.cookie.dmac_address;
4361 
4362 			if (ql_loop_back(ha, 0, lb,
4363 			    buffer_xmt.cookie.dmac_notused,
4364 			    buffer_rcv.cookie.dmac_notused) == QL_SUCCESS) {
4365 				bzero((void *)cmd->pm_stat_buf,
4366 				    cmd->pm_stat_len);
4367 				ddi_rep_get8(buffer_rcv.acc_handle,
4368 				    (uint8_t *)cmd->pm_stat_buf,
4369 				    (uint8_t *)buffer_rcv.bp,
4370 				    cmd->pm_stat_len, DDI_DEV_AUTOINCR);
4371 				rval = FC_SUCCESS;
4372 			} else {
4373 				EL(ha, "failed, QL_DIAG_LPBDTA FC_FAILURE\n");
4374 				rval = FC_FAILURE;
4375 			}
4376 
4377 			ql_free_phys(ha, &buffer_xmt);
4378 			ql_free_phys(ha, &buffer_rcv);
4379 
4380 			/* Needed to recover the f/w */
4381 			set_flags |= ISP_ABORT_NEEDED;
4382 
4383 			/* Restart IP if it was shutdown. */
4384 			if (pha->flags & IP_ENABLED &&
4385 			    !(pha->flags & IP_INITIALIZED)) {
4386 				(void) ql_initialize_ip(pha);
4387 				ql_isp_rcvbuf(pha);
4388 			}
4389 
4390 			break;
4391 		case QL_DIAG_ECHO: {
4392 			/*
4393 			 * issue an echo command with a user supplied
4394 			 * data pattern and destination address
4395 			 */
4396 			echo_t		echo;		/* temp echo struct */
4397 
4398 			/* Setup echo cmd & adjust for platform */
4399 			opcode = QL_ECHO_CMD;
4400 			BIG_ENDIAN_32(&opcode);
4401 
4402 			/*
4403 			 * due to limitations in the ql
4404 			 * firmaware the echo data field is
4405 			 * limited to 220
4406 			 */
4407 			if ((cmd->pm_cmd_len > QL_ECHO_CMD_LENGTH) ||
4408 			    (cmd->pm_stat_len > QL_ECHO_CMD_LENGTH)) {
4409 				EL(ha, "failed, QL_DIAG_ECHO FC_TOOMANY, "
4410 				    "cmdl1=%lxh, statl2=%lxh\n",
4411 				    cmd->pm_cmd_len, cmd->pm_stat_len);
4412 				rval = FC_TOOMANY;
4413 				break;
4414 			}
4415 
4416 			/*
4417 			 * the input data buffer has the user
4418 			 * supplied data pattern.  The "echoed"
4419 			 * data will be DMAed into the output
4420 			 * data buffer.  Therefore the length
4421 			 * of the output buffer must be equal
4422 			 * to or greater then the input buffer
4423 			 * length
4424 			 */
4425 			if (cmd->pm_cmd_len > cmd->pm_stat_len) {
4426 				EL(ha, "failed, QL_DIAG_ECHO FC_TOOMANY-2,"
4427 				    " cmdl1=%lxh, statl2=%lxh\n",
4428 				    cmd->pm_cmd_len, cmd->pm_stat_len);
4429 				rval = FC_TOOMANY;
4430 				break;
4431 			}
4432 			/* add four bytes for the opcode */
4433 			echo.transfer_count = (uint32_t)(cmd->pm_cmd_len + 4);
4434 
4435 			/*
4436 			 * are we 32 or 64 bit addressed???
4437 			 * We need to get the appropriate
4438 			 * DMA and set the command options;
4439 			 * 64 bit (bit 6) or 32 bit
4440 			 * (no bit 6) addressing.
4441 			 * while we are at it lets ask for
4442 			 * real echo (bit 15)
4443 			 */
4444 			echo.options = BIT_15;
4445 			if (CFG_IST(ha, CFG_ENABLE_64BIT_ADDRESSING) &&
4446 			    !(CFG_IST(ha, CFG_CTRL_8081))) {
4447 				echo.options = (uint16_t)
4448 				    (echo.options | BIT_6);
4449 			}
4450 
4451 			/*
4452 			 * Set up the DMA mappings for the
4453 			 * output and input data buffers.
4454 			 * First the output buffer
4455 			 */
4456 			if (ql_get_dma_mem(ha, &buffer_xmt,
4457 			    (uint32_t)(cmd->pm_data_len + 4),
4458 			    LITTLE_ENDIAN_DMA,
4459 			    QL_DMA_DATA_ALIGN) != QL_SUCCESS) {
4460 				EL(ha, "failed, QL_DIAG_ECHO FC_NOMEM\n");
4461 				rval = FC_NOMEM;
4462 				break;
4463 			}
4464 			echo.transfer_data_address = buffer_xmt.cookie;
4465 
4466 			/* Next the input buffer */
4467 			if (ql_get_dma_mem(ha, &buffer_rcv,
4468 			    (uint32_t)(cmd->pm_data_len + 4),
4469 			    LITTLE_ENDIAN_DMA,
4470 			    QL_DMA_DATA_ALIGN) != QL_SUCCESS) {
4471 				/*
4472 				 * since we could not allocate
4473 				 * DMA space for the input
4474 				 * buffer we need to clean up
4475 				 * by freeing the DMA space
4476 				 * we allocated for the output
4477 				 * buffer
4478 				 */
4479 				ql_free_phys(ha, &buffer_xmt);
4480 				EL(ha, "failed, QL_DIAG_ECHO FC_NOMEM-2\n");
4481 				rval = FC_NOMEM;
4482 				break;
4483 			}
4484 			echo.receive_data_address = buffer_rcv.cookie;
4485 
4486 			/*
4487 			 * copy the 4 byte ECHO op code to the
4488 			 * allocated DMA space
4489 			 */
4490 			ddi_rep_put8(buffer_xmt.acc_handle, (uint8_t *)&opcode,
4491 			    (uint8_t *)buffer_xmt.bp, 4, DDI_DEV_AUTOINCR);
4492 
4493 			/*
4494 			 * copy the user supplied data to the
4495 			 * allocated DMA space
4496 			 */
4497 			ddi_rep_put8(buffer_xmt.acc_handle,
4498 			    (uint8_t *)cmd->pm_cmd_buf,
4499 			    (uint8_t *)buffer_xmt.bp + 4, cmd->pm_cmd_len,
4500 			    DDI_DEV_AUTOINCR);
4501 
4502 			/* Shutdown IP. */
4503 			if (pha->flags & IP_INITIALIZED) {
4504 				(void) ql_shutdown_ip(pha);
4505 			}
4506 
4507 			/* send the echo */
4508 			if (ql_echo(ha, 0, &echo) == QL_SUCCESS) {
4509 				ddi_rep_put8(buffer_rcv.acc_handle,
4510 				    (uint8_t *)buffer_rcv.bp + 4,
4511 				    (uint8_t *)cmd->pm_stat_buf,
4512 				    cmd->pm_stat_len, DDI_DEV_AUTOINCR);
4513 			} else {
4514 				EL(ha, "failed, QL_DIAG_ECHO FC_FAILURE\n");
4515 				rval = FC_FAILURE;
4516 			}
4517 
4518 			/* Restart IP if it was shutdown. */
4519 			if (pha->flags & IP_ENABLED &&
4520 			    !(pha->flags & IP_INITIALIZED)) {
4521 				(void) ql_initialize_ip(pha);
4522 				ql_isp_rcvbuf(pha);
4523 			}
4524 			/* free up our DMA buffers */
4525 			ql_free_phys(ha, &buffer_xmt);
4526 			ql_free_phys(ha, &buffer_rcv);
4527 			break;
4528 		}
4529 		default:
4530 			EL(ha, "unknown=%xh, FC_PORT_DIAG "
4531 			    "FC_INVALID_REQUEST\n", cmd->pm_cmd_flags);
4532 			rval = FC_INVALID_REQUEST;
4533 			break;
4534 		}
4535 		PORTMANAGE_UNLOCK(ha);
4536 		break;
4537 	case FC_PORT_LINK_STATE:
4538 		/* Check for name equal to null. */
4539 		for (index = 0; index < 8 && index < cmd->pm_cmd_len;
4540 		    index++) {
4541 			if (cmd->pm_cmd_buf[index] != 0) {
4542 				break;
4543 			}
4544 		}
4545 
4546 		/* If name not null. */
4547 		if (index < 8 && cmd->pm_cmd_len >= 8) {
4548 			/* Locate device queue. */
4549 			tq = NULL;
4550 			for (index = 0; index < DEVICE_HEAD_LIST_SIZE &&
4551 			    tq == NULL; index++) {
4552 				for (link = ha->dev[index].first; link != NULL;
4553 				    link = link->next) {
4554 					tq = link->base_address;
4555 
4556 					if (bcmp((void *)&tq->port_name[0],
4557 					    (void *)cmd->pm_cmd_buf, 8) == 0) {
4558 						break;
4559 					} else {
4560 						tq = NULL;
4561 					}
4562 				}
4563 			}
4564 
4565 			if (tq != NULL && VALID_DEVICE_ID(ha, tq->loop_id)) {
4566 				cmd->pm_stat_buf[0] = (int8_t)LSB(ha->state);
4567 				cmd->pm_stat_buf[1] = (int8_t)MSB(ha->state);
4568 			} else {
4569 				cnt = FC_PORT_SPEED_MASK(ha->state) |
4570 				    FC_STATE_OFFLINE;
4571 				cmd->pm_stat_buf[0] = (int8_t)LSB(cnt);
4572 				cmd->pm_stat_buf[1] = (int8_t)MSB(cnt);
4573 			}
4574 		} else {
4575 			cmd->pm_stat_buf[0] = (int8_t)LSB(ha->state);
4576 			cmd->pm_stat_buf[1] = (int8_t)MSB(ha->state);
4577 		}
4578 		break;
4579 	case FC_PORT_INITIALIZE:
4580 		if (cmd->pm_cmd_len >= 8) {
4581 			tq = NULL;
4582 			for (index = 0; index < DEVICE_HEAD_LIST_SIZE &&
4583 			    tq == NULL; index++) {
4584 				for (link = ha->dev[index].first; link != NULL;
4585 				    link = link->next) {
4586 					tq = link->base_address;
4587 
4588 					if (bcmp((void *)&tq->port_name[0],
4589 					    (void *)cmd->pm_cmd_buf, 8) == 0) {
4590 						if (!VALID_DEVICE_ID(ha,
4591 						    tq->loop_id)) {
4592 							tq = NULL;
4593 						}
4594 						break;
4595 					} else {
4596 						tq = NULL;
4597 					}
4598 				}
4599 			}
4600 
4601 			if (tq == NULL || ql_target_reset(ha, tq,
4602 			    ha->loop_reset_delay) != QL_SUCCESS) {
4603 				EL(ha, "failed, FC_PORT_INITIALIZE "
4604 				    "FC_FAILURE\n");
4605 				rval = FC_FAILURE;
4606 			}
4607 		} else {
4608 			EL(ha, "failed, FC_PORT_INITIALIZE FC_FAILURE-2, "
4609 			    "clen=%lxh\n", cmd->pm_cmd_len);
4610 
4611 			rval = FC_FAILURE;
4612 		}
4613 		break;
4614 	case FC_PORT_RLS:
4615 		if (cmd->pm_data_len < sizeof (fc_rls_acc_t)) {
4616 			EL(ha, "failed, buffer size passed: %lxh, "
4617 			    "req: %lxh\n", cmd->pm_data_len,
4618 			    (sizeof (fc_rls_acc_t)));
4619 			rval = FC_FAILURE;
4620 		} else if (LOOP_NOT_READY(pha)) {
4621 			EL(ha, "loop NOT ready\n");
4622 			bzero(cmd->pm_data_buf, cmd->pm_data_len);
4623 		} else if (ql_get_link_status(ha, ha->loop_id,
4624 		    cmd->pm_data_len, cmd->pm_data_buf, 0) != QL_SUCCESS) {
4625 			EL(ha, "failed, FC_PORT_RLS FC_FAILURE\n");
4626 			rval = FC_FAILURE;
4627 #ifdef _BIG_ENDIAN
4628 		} else {
4629 			fc_rls_acc_t		*rls;
4630 
4631 			rls = (fc_rls_acc_t *)cmd->pm_data_buf;
4632 			LITTLE_ENDIAN_32(&rls->rls_link_fail);
4633 			LITTLE_ENDIAN_32(&rls->rls_sync_loss);
4634 			LITTLE_ENDIAN_32(&rls->rls_sig_loss);
4635 			LITTLE_ENDIAN_32(&rls->rls_invalid_crc);
4636 #endif /* _BIG_ENDIAN */
4637 		}
4638 		break;
4639 	case FC_PORT_GET_NODE_ID:
4640 		if (ql_get_rnid_params(ha, cmd->pm_data_len,
4641 		    cmd->pm_data_buf) != QL_SUCCESS) {
4642 			EL(ha, "failed, FC_PORT_GET_NODE_ID FC_FAILURE\n");
4643 			rval = FC_FAILURE;
4644 		}
4645 		break;
4646 	case FC_PORT_SET_NODE_ID:
4647 		if (ql_set_rnid_params(ha, cmd->pm_data_len,
4648 		    cmd->pm_data_buf) != QL_SUCCESS) {
4649 			EL(ha, "failed, FC_PORT_SET_NODE_ID FC_FAILURE\n");
4650 			rval = FC_FAILURE;
4651 		}
4652 		break;
4653 	case FC_PORT_DOWNLOAD_FCODE:
4654 		PORTMANAGE_LOCK(ha);
4655 		if ((CFG_IST(ha, CFG_CTRL_24258081)) == 0) {
4656 			rval = ql_load_flash(ha, (uint8_t *)cmd->pm_data_buf,
4657 			    (uint32_t)cmd->pm_data_len);
4658 		} else {
4659 			if (cmd->pm_data_buf[0] == 4 &&
4660 			    cmd->pm_data_buf[8] == 0 &&
4661 			    cmd->pm_data_buf[9] == 0x10 &&
4662 			    cmd->pm_data_buf[10] == 0 &&
4663 			    cmd->pm_data_buf[11] == 0) {
4664 				rval = ql_24xx_load_flash(ha,
4665 				    (uint8_t *)cmd->pm_data_buf,
4666 				    (uint32_t)cmd->pm_data_len,
4667 				    ha->flash_fw_addr << 2);
4668 			} else {
4669 				rval = ql_24xx_load_flash(ha,
4670 				    (uint8_t *)cmd->pm_data_buf,
4671 				    (uint32_t)cmd->pm_data_len, 0);
4672 			}
4673 		}
4674 
4675 		if (rval != QL_SUCCESS) {
4676 			EL(ha, "failed, FC_PORT_DOWNLOAD_FCODE FC_FAILURE\n");
4677 			rval = FC_FAILURE;
4678 		} else {
4679 			rval = FC_SUCCESS;
4680 		}
4681 		ql_reset_chip(ha);
4682 		set_flags |= ISP_ABORT_NEEDED;
4683 		PORTMANAGE_UNLOCK(ha);
4684 		break;
4685 	default:
4686 		EL(ha, "unknown=%xh, FC_BADCMD\n", cmd->pm_cmd_code);
4687 		rval = FC_BADCMD;
4688 		break;
4689 	}
4690 
4691 	/* Wait for suspension to end. */
4692 	ql_awaken_task_daemon(ha, NULL, set_flags, DRIVER_STALL);
4693 	timer = 0;
4694 
4695 	while (timer++ < 3000 &&
4696 	    ha->task_daemon_flags & (QL_LOOP_TRANSITION | DRIVER_STALL)) {
4697 		ql_delay(ha, 10000);
4698 	}
4699 
4700 	ql_restart_queues(ha);
4701 
4702 	if (rval != FC_SUCCESS) {
4703 		EL(ha, "failed, rval = %xh\n", rval);
4704 	} else {
4705 		/*EMPTY*/
4706 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
4707 	}
4708 
4709 	return (rval);
4710 }
4711 
4712 static opaque_t
ql_get_device(opaque_t fca_handle,fc_portid_t d_id)4713 ql_get_device(opaque_t fca_handle, fc_portid_t d_id)
4714 {
4715 	port_id_t		id;
4716 	ql_adapter_state_t	*ha;
4717 	ql_tgt_t		*tq;
4718 
4719 	id.r.rsvd_1 = 0;
4720 	id.b24 = d_id.port_id;
4721 
4722 	ha = ql_fca_handle_to_state(fca_handle);
4723 	if (ha == NULL) {
4724 		QL_PRINT_2(CE_CONT, "failed, no adapter=%ph\n",
4725 		    (void *)fca_handle);
4726 		return (NULL);
4727 	}
4728 	QL_PRINT_3(CE_CONT, "(%d): started, d_id=%xh\n", ha->instance, id.b24);
4729 
4730 	tq = ql_d_id_to_queue(ha, id);
4731 
4732 	if (tq == NULL) {
4733 		EL(ha, "failed, tq=NULL\n");
4734 	} else {
4735 		/*EMPTY*/
4736 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
4737 	}
4738 	return (tq);
4739 }
4740 
4741 /* ************************************************************************ */
4742 /*			FCA Driver Local Support Functions.		    */
4743 /* ************************************************************************ */
4744 
4745 /*
4746  * ql_cmd_setup
4747  *	Verifies proper command.
4748  *
4749  * Input:
4750  *	fca_handle = handle setup by ql_bind_port().
4751  *	pkt = pointer to fc_packet.
4752  *	rval = pointer for return value.
4753  *
4754  * Returns:
4755  *	Adapter state pointer, NULL = failure.
4756  *
4757  * Context:
4758  *	Kernel context.
4759  */
4760 static ql_adapter_state_t *
ql_cmd_setup(opaque_t fca_handle,fc_packet_t * pkt,int * rval)4761 ql_cmd_setup(opaque_t fca_handle, fc_packet_t *pkt, int *rval)
4762 {
4763 	ql_adapter_state_t	*ha, *pha;
4764 	ql_srb_t		*sp = (ql_srb_t *)pkt->pkt_fca_private;
4765 	ql_tgt_t		*tq;
4766 	port_id_t		d_id;
4767 
4768 	pkt->pkt_resp_resid = 0;
4769 	pkt->pkt_data_resid = 0;
4770 
4771 	/* check that the handle is assigned by this FCA */
4772 	ha = ql_fca_handle_to_state(fca_handle);
4773 	if (ha == NULL) {
4774 		*rval = FC_UNBOUND;
4775 		QL_PRINT_2(CE_CONT, "failed, no adapter=%ph\n",
4776 		    (void *)fca_handle);
4777 		return (NULL);
4778 	}
4779 	pha = ha->pha;
4780 
4781 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
4782 
4783 	if (ddi_in_panic() || pkt->pkt_tran_flags & FC_TRAN_DUMPING) {
4784 		return (ha);
4785 	}
4786 
4787 	if (!(pha->flags & ONLINE)) {
4788 		pkt->pkt_state = FC_PKT_LOCAL_RJT;
4789 		pkt->pkt_reason = FC_REASON_HW_ERROR;
4790 		*rval = FC_TRANSPORT_ERROR;
4791 		EL(ha, "failed, not online hf=%xh\n", pha->flags);
4792 		return (NULL);
4793 	}
4794 
4795 	/* Exit on loop down. */
4796 	if (CFG_IST(ha, CFG_ENABLE_LINK_DOWN_REPORTING) &&
4797 	    pha->task_daemon_flags & LOOP_DOWN &&
4798 	    pha->loop_down_timer <= pha->loop_down_abort_time) {
4799 		pkt->pkt_state = FC_PKT_PORT_OFFLINE;
4800 		pkt->pkt_reason = FC_REASON_NO_CONNECTION;
4801 		*rval = FC_OFFLINE;
4802 		EL(ha, "failed, loop down tdf=%xh\n", pha->task_daemon_flags);
4803 		return (NULL);
4804 	}
4805 
4806 	if (pkt->pkt_cmd_fhdr.r_ctl == R_CTL_COMMAND &&
4807 	    pkt->pkt_cmd_fhdr.type == FC_TYPE_SCSI_FCP) {
4808 		tq = (ql_tgt_t *)pkt->pkt_fca_device;
4809 		if ((tq == NULL) || (!VALID_DEVICE_ID(ha, tq->loop_id))) {
4810 			d_id.r.rsvd_1 = 0;
4811 			d_id.b24 = pkt->pkt_cmd_fhdr.d_id;
4812 			tq = ql_d_id_to_queue(ha, d_id);
4813 
4814 			pkt->pkt_fca_device = (opaque_t)tq;
4815 		}
4816 
4817 		if (tq != NULL) {
4818 			DEVICE_QUEUE_LOCK(tq);
4819 			if (tq->flags & (TQF_RSCN_RCVD |
4820 			    TQF_NEED_AUTHENTICATION)) {
4821 				*rval = FC_DEVICE_BUSY;
4822 				DEVICE_QUEUE_UNLOCK(tq);
4823 				EL(ha, "failed, busy qf=%xh, d_id=%xh\n",
4824 				    tq->flags, tq->d_id.b24);
4825 				return (NULL);
4826 			}
4827 			DEVICE_QUEUE_UNLOCK(tq);
4828 		}
4829 	}
4830 
4831 	/*
4832 	 * Check DMA pointers.
4833 	 */
4834 	*rval = DDI_SUCCESS;
4835 	if (pkt->pkt_cmd_acc != NULL && pkt->pkt_cmdlen) {
4836 		QL_CLEAR_DMA_HANDLE(pkt->pkt_cmd_dma);
4837 		*rval = ddi_check_dma_handle(pkt->pkt_cmd_dma);
4838 		if (*rval == DDI_SUCCESS) {
4839 			*rval = ddi_check_acc_handle(pkt->pkt_cmd_acc);
4840 		}
4841 	}
4842 
4843 	if (pkt->pkt_resp_acc != NULL && *rval == DDI_SUCCESS &&
4844 	    pkt->pkt_rsplen != 0) {
4845 		QL_CLEAR_DMA_HANDLE(pkt->pkt_resp_dma);
4846 		*rval = ddi_check_dma_handle(pkt->pkt_resp_dma);
4847 		if (*rval == DDI_SUCCESS) {
4848 			*rval = ddi_check_acc_handle(pkt->pkt_resp_acc);
4849 		}
4850 	}
4851 
4852 	/*
4853 	 * Minimum branch conditional; Change it with care.
4854 	 */
4855 	if (((pkt->pkt_data_acc != NULL) & (*rval == DDI_SUCCESS) &
4856 	    (pkt->pkt_datalen != 0)) != 0) {
4857 		QL_CLEAR_DMA_HANDLE(pkt->pkt_data_dma);
4858 		*rval = ddi_check_dma_handle(pkt->pkt_data_dma);
4859 		if (*rval == DDI_SUCCESS) {
4860 			*rval = ddi_check_acc_handle(pkt->pkt_data_acc);
4861 		}
4862 	}
4863 
4864 	if (*rval != DDI_SUCCESS) {
4865 		pkt->pkt_state = FC_PKT_TRAN_ERROR;
4866 		pkt->pkt_reason = FC_REASON_DMA_ERROR;
4867 
4868 		/* Do command callback. */
4869 		if (!(pkt->pkt_tran_flags & FC_TRAN_NO_INTR) && pkt->pkt_comp) {
4870 			ql_awaken_task_daemon(ha, sp, 0, 0);
4871 		}
4872 		*rval = FC_BADPACKET;
4873 		EL(ha, "failed, bad DMA pointers\n");
4874 		return (NULL);
4875 	}
4876 
4877 	if (sp->magic_number != QL_FCA_BRAND) {
4878 		*rval = FC_BADPACKET;
4879 		EL(ha, "failed, magic number=%xh\n", sp->magic_number);
4880 		return (NULL);
4881 	}
4882 	*rval = FC_SUCCESS;
4883 
4884 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
4885 
4886 	return (ha);
4887 }
4888 
4889 /*
4890  * ql_els_plogi
4891  *	Issue a extended link service port login request.
4892  *
4893  * Input:
4894  *	ha = adapter state pointer.
4895  *	pkt = pointer to fc_packet.
4896  *
4897  * Returns:
4898  *	FC_SUCCESS - the packet was accepted for transport.
4899  *	FC_TRANSPORT_ERROR - a transport error occurred.
4900  *
4901  * Context:
4902  *	Kernel context.
4903  */
4904 static int
ql_els_plogi(ql_adapter_state_t * ha,fc_packet_t * pkt)4905 ql_els_plogi(ql_adapter_state_t *ha, fc_packet_t *pkt)
4906 {
4907 	ql_tgt_t		*tq = NULL;
4908 	port_id_t		d_id;
4909 	la_els_logi_t		acc;
4910 	class_svc_param_t	*class3_param;
4911 	int			ret;
4912 	int			rval = FC_SUCCESS;
4913 
4914 	QL_PRINT_3(CE_CONT, "(%d): started, d_id=%xh\n", ha->instance,
4915 	    pkt->pkt_cmd_fhdr.d_id);
4916 
4917 	TASK_DAEMON_LOCK(ha);
4918 	if (!(ha->task_daemon_flags & STATE_ONLINE)) {
4919 		TASK_DAEMON_UNLOCK(ha);
4920 		QL_PRINT_3(CE_CONT, "(%d): offline done\n", ha->instance);
4921 		return (FC_OFFLINE);
4922 	}
4923 	TASK_DAEMON_UNLOCK(ha);
4924 
4925 	bzero(&acc, sizeof (acc));
4926 	d_id.b24 = pkt->pkt_cmd_fhdr.d_id;
4927 
4928 	ret = QL_SUCCESS;
4929 
4930 	if (CFG_IST(ha, CFG_CTRL_2425) && ha->topology & QL_N_PORT) {
4931 		/*
4932 		 * In p2p topology he sends a PLOGI after determining
4933 		 * he has the N_Port login initiative.
4934 		 */
4935 		ret = ql_p2p_plogi(ha, pkt);
4936 	}
4937 	if (ret == QL_CONSUMED) {
4938 		return (ret);
4939 	}
4940 
4941 	switch (ret = ql_login_port(ha, d_id)) {
4942 	case QL_SUCCESS:
4943 		tq = ql_d_id_to_queue(ha, d_id);
4944 		break;
4945 
4946 	case QL_LOOP_ID_USED:
4947 		if ((ret = ql_login_port(ha, d_id)) == QL_SUCCESS) {
4948 			tq = ql_d_id_to_queue(ha, d_id);
4949 		}
4950 		break;
4951 
4952 	default:
4953 		break;
4954 	}
4955 
4956 	if (ret != QL_SUCCESS) {
4957 		/*
4958 		 * Invalidate this entry so as to seek a fresh loop ID
4959 		 * in case firmware reassigns it to something else
4960 		 */
4961 		tq = ql_d_id_to_queue(ha, d_id);
4962 		if (tq && (ret != QL_MEMORY_ALLOC_FAILED)) {
4963 			tq->loop_id = PORT_NO_LOOP_ID;
4964 		}
4965 	} else if (tq) {
4966 		(void) ql_get_port_database(ha, tq, PDF_ADISC);
4967 	}
4968 
4969 	if (tq != NULL && VALID_DEVICE_ID(ha, tq->loop_id) &&
4970 	    (ret != QL_MEMORY_ALLOC_FAILED) && PD_PORT_LOGIN(tq)) {
4971 
4972 		/* Build ACC. */
4973 		acc.ls_code.ls_code = LA_ELS_ACC;
4974 		acc.common_service.fcph_version = 0x2006;
4975 		acc.common_service.cmn_features = 0x8800;
4976 		acc.common_service.rx_bufsize = QL_MAX_FRAME_SIZE(ha);
4977 		acc.common_service.conc_sequences = 0xff;
4978 		acc.common_service.relative_offset = 0x03;
4979 		acc.common_service.e_d_tov = 0x7d0;
4980 
4981 		bcopy((void *)&tq->port_name[0],
4982 		    (void *)&acc.nport_ww_name.raw_wwn[0], 8);
4983 		bcopy((void *)&tq->node_name[0],
4984 		    (void *)&acc.node_ww_name.raw_wwn[0], 8);
4985 
4986 		class3_param = (class_svc_param_t *)&acc.class_3;
4987 		class3_param->class_valid_svc_opt = 0x8000;
4988 		class3_param->recipient_ctl = tq->class3_recipient_ctl;
4989 		class3_param->rcv_data_size = tq->class3_rcv_data_size;
4990 		class3_param->conc_sequences = tq->class3_conc_sequences;
4991 		class3_param->open_sequences_per_exch =
4992 		    tq->class3_open_sequences_per_exch;
4993 
4994 		if ((ql_busy_plogi(ha, pkt, tq) == FC_TRAN_BUSY)) {
4995 			acc.ls_code.ls_code = LA_ELS_RJT;
4996 			pkt->pkt_state = FC_PKT_TRAN_BSY;
4997 			pkt->pkt_reason = FC_REASON_XCHG_BSY;
4998 			EL(ha, "LA_ELS_RJT, FC_REASON_XCHG_BSY\n");
4999 			rval = FC_TRAN_BUSY;
5000 		} else {
5001 			DEVICE_QUEUE_LOCK(tq);
5002 			tq->logout_sent = 0;
5003 			tq->flags &= ~TQF_NEED_AUTHENTICATION;
5004 			if (CFG_IST(ha, CFG_CTRL_242581)) {
5005 				tq->flags |= TQF_IIDMA_NEEDED;
5006 			}
5007 			DEVICE_QUEUE_UNLOCK(tq);
5008 
5009 			if (CFG_IST(ha, CFG_CTRL_242581)) {
5010 				TASK_DAEMON_LOCK(ha);
5011 				ha->task_daemon_flags |= TD_IIDMA_NEEDED;
5012 				TASK_DAEMON_UNLOCK(ha);
5013 			}
5014 
5015 			pkt->pkt_state = FC_PKT_SUCCESS;
5016 		}
5017 	} else {
5018 		/* Build RJT. */
5019 		acc.ls_code.ls_code = LA_ELS_RJT;
5020 
5021 		switch (ret) {
5022 		case QL_FUNCTION_TIMEOUT:
5023 			pkt->pkt_state = FC_PKT_TIMEOUT;
5024 			pkt->pkt_reason = FC_REASON_HW_ERROR;
5025 			break;
5026 
5027 		case QL_MEMORY_ALLOC_FAILED:
5028 			pkt->pkt_state = FC_PKT_LOCAL_BSY;
5029 			pkt->pkt_reason = FC_REASON_NOMEM;
5030 			rval = FC_TRAN_BUSY;
5031 			break;
5032 
5033 		case QL_FABRIC_NOT_INITIALIZED:
5034 			pkt->pkt_state = FC_PKT_FABRIC_BSY;
5035 			pkt->pkt_reason = FC_REASON_NO_CONNECTION;
5036 			rval = FC_TRAN_BUSY;
5037 			break;
5038 
5039 		default:
5040 			pkt->pkt_state = FC_PKT_TRAN_ERROR;
5041 			pkt->pkt_reason = FC_REASON_NO_CONNECTION;
5042 			break;
5043 		}
5044 
5045 		EL(ha, "Plogi unsuccess for %xh state %xh reason %xh "
5046 		    "ret %xh rval %xh\n", d_id.b24, pkt->pkt_state,
5047 		    pkt->pkt_reason, ret, rval);
5048 	}
5049 
5050 	if (tq != NULL) {
5051 		DEVICE_QUEUE_LOCK(tq);
5052 		tq->flags &= ~(TQF_PLOGI_PROGRS | TQF_QUEUE_SUSPENDED);
5053 		if (rval == FC_TRAN_BUSY) {
5054 			if (tq->d_id.b24 != BROADCAST_ADDR) {
5055 				tq->flags |= TQF_NEED_AUTHENTICATION;
5056 			}
5057 		}
5058 		DEVICE_QUEUE_UNLOCK(tq);
5059 	}
5060 
5061 	ddi_rep_put8(pkt->pkt_resp_acc, (uint8_t *)&acc,
5062 	    (uint8_t *)pkt->pkt_resp, sizeof (acc), DDI_DEV_AUTOINCR);
5063 
5064 	if (rval != FC_SUCCESS) {
5065 		EL(ha, "failed, rval = %xh\n", rval);
5066 	} else {
5067 		/*EMPTY*/
5068 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
5069 	}
5070 	return (rval);
5071 }
5072 
5073 /*
5074  * ql_p2p_plogi
5075  *	Start an extended link service port login request using
5076  *	an ELS Passthru iocb.
5077  *
5078  * Input:
5079  *	ha = adapter state pointer.
5080  *	pkt = pointer to fc_packet.
5081  *
5082  * Returns:
5083  *	QL_CONSUMMED - the iocb was queued for transport.
5084  *
5085  * Context:
5086  *	Kernel context.
5087  */
5088 static int
ql_p2p_plogi(ql_adapter_state_t * ha,fc_packet_t * pkt)5089 ql_p2p_plogi(ql_adapter_state_t *ha, fc_packet_t *pkt)
5090 {
5091 	uint16_t	id;
5092 	ql_tgt_t	tmp;
5093 	ql_tgt_t	*tq = &tmp;
5094 	int		rval;
5095 	port_id_t	d_id;
5096 	ql_srb_t	*sp = (ql_srb_t *)pkt->pkt_fca_private;
5097 
5098 	tq->d_id.b.al_pa = 0;
5099 	tq->d_id.b.area = 0;
5100 	tq->d_id.b.domain = 0;
5101 
5102 	/*
5103 	 * Verify that the port database hasn't moved beneath our feet by
5104 	 * switching to the appropriate n_port_handle if necessary.  This is
5105 	 * less unplesant than the error recovery if the wrong one is used.
5106 	 */
5107 	for (id = 0; id <= LAST_LOCAL_LOOP_ID; id++) {
5108 		tq->loop_id = id;
5109 		rval = ql_get_port_database(ha, tq, PDF_NONE);
5110 		EL(ha, "rval=%xh\n", rval);
5111 		/* check all the ones not logged in for possible use */
5112 		if (rval == QL_NOT_LOGGED_IN) {
5113 			if (tq->master_state == PD_STATE_PLOGI_PENDING) {
5114 				ha->n_port->n_port_handle = tq->loop_id;
5115 				EL(ha, "n_port_handle =%xh, master state=%x\n",
5116 				    tq->loop_id, tq->master_state);
5117 				break;
5118 			}
5119 			/*
5120 			 * Use a 'port unavailable' entry only
5121 			 * if we used it before.
5122 			 */
5123 			if (tq->master_state == PD_STATE_PORT_UNAVAILABLE) {
5124 				/* if the port_id matches, reuse it */
5125 				if (pkt->pkt_cmd_fhdr.d_id == tq->d_id.b24) {
5126 					EL(ha, "n_port_handle =%xh,"
5127 					    "master state=%xh\n",
5128 					    tq->loop_id, tq->master_state);
5129 					break;
5130 				} else if (tq->loop_id ==
5131 				    ha->n_port->n_port_handle) {
5132 				    // avoid a lint error
5133 					uint16_t *hndl;
5134 					uint16_t val;
5135 
5136 					hndl = &ha->n_port->n_port_handle;
5137 					val = *hndl;
5138 					val++;
5139 					val++;
5140 					*hndl = val;
5141 				}
5142 			EL(ha, "rval=%xh, id=%d, n_port_handle =%xh, "
5143 			    "master state=%x\n", rval, id, tq->loop_id,
5144 			    tq->master_state);
5145 			}
5146 
5147 		}
5148 		if (rval == QL_SUCCESS) {
5149 			if ((tq->flags & TQF_INITIATOR_DEVICE) == 0) {
5150 				ha->n_port->n_port_handle = tq->loop_id;
5151 				EL(ha, "n_port_handle =%xh, master state=%x\n",
5152 				    tq->loop_id, tq->master_state);
5153 				break;
5154 			}
5155 			EL(ha, "rval=%xh, id=%d, n_port_handle =%xh, "
5156 			    "master state=%x\n", rval, id, tq->loop_id,
5157 			    tq->master_state);
5158 		}
5159 	}
5160 	(void) ddi_dma_sync(pkt->pkt_cmd_dma, 0, 0, DDI_DMA_SYNC_FORDEV);
5161 
5162 	d_id.b24 = pkt->pkt_cmd_fhdr.d_id;
5163 	tq = ql_d_id_to_queue(ha, d_id);
5164 	ql_timeout_insert(ha, tq, sp);
5165 	ql_start_iocb(ha, sp);
5166 
5167 	return (QL_CONSUMED);
5168 }
5169 
5170 
5171 /*
5172  * ql_els_flogi
5173  *	Issue a extended link service fabric login request.
5174  *
5175  * Input:
5176  *	ha = adapter state pointer.
5177  *	pkt = pointer to fc_packet.
5178  *
5179  * Returns:
5180  *	FC_SUCCESS - the packet was accepted for transport.
5181  *	FC_TRANSPORT_ERROR - a transport error occurred.
5182  *
5183  * Context:
5184  *	Kernel context.
5185  */
5186 static int
ql_els_flogi(ql_adapter_state_t * ha,fc_packet_t * pkt)5187 ql_els_flogi(ql_adapter_state_t *ha, fc_packet_t *pkt)
5188 {
5189 	ql_tgt_t		*tq = NULL;
5190 	port_id_t		d_id;
5191 	la_els_logi_t		acc;
5192 	class_svc_param_t	*class3_param;
5193 	int			rval = FC_SUCCESS;
5194 	int			accept = 0;
5195 
5196 	QL_PRINT_3(CE_CONT, "(%d): started, d_id=%xh\n", ha->instance,
5197 	    pkt->pkt_cmd_fhdr.d_id);
5198 
5199 	bzero(&acc, sizeof (acc));
5200 	d_id.b24 = pkt->pkt_cmd_fhdr.d_id;
5201 
5202 	if (CFG_IST(ha, CFG_CTRL_2425) && ha->topology & QL_N_PORT) {
5203 		/*
5204 		 * d_id of zero in a FLOGI accept response in a point to point
5205 		 * topology triggers evaluation of N Port login initiative.
5206 		 */
5207 		pkt->pkt_resp_fhdr.d_id = 0;
5208 		/*
5209 		 * An N_Port already logged in with the firmware
5210 		 * will have the only database entry.
5211 		 */
5212 		if (LOCAL_LOOP_ID(ha->n_port->n_port_handle)) {
5213 			tq = ql_loop_id_to_queue(ha, ha->n_port->n_port_handle);
5214 		}
5215 
5216 		if (tq != NULL) {
5217 			/*
5218 			 * If the target port has initiative send
5219 			 * up a PLOGI about the new device.
5220 			 */
5221 			if ((ql_wwn_cmp(ha, (la_wwn_t *)&tq->port_name[0],
5222 			    (la_wwn_t *)(CFG_IST(ha, CFG_CTRL_2425) ?
5223 			    &ha->init_ctrl_blk.cb24.port_name[0] :
5224 			    &ha->init_ctrl_blk.cb.port_name[0])) == 1)) {
5225 				ha->send_plogi_timer = 3;
5226 			} else {
5227 				ha->send_plogi_timer = 0;
5228 			}
5229 			pkt->pkt_resp_fhdr.s_id = tq->d_id.b24;
5230 		} else {
5231 			/*
5232 			 * An N_Port not logged in with the firmware will not
5233 			 * have a database entry.  We accept anyway and rely
5234 			 * on a PLOGI from the upper layers to set the d_id
5235 			 * and s_id.
5236 			 */
5237 			accept = 1;
5238 		}
5239 	} else {
5240 		tq = ql_d_id_to_queue(ha, d_id);
5241 	}
5242 	if ((tq != NULL) || (accept != NULL)) {
5243 		/* Build ACC. */
5244 		pkt->pkt_state = FC_PKT_SUCCESS;
5245 		class3_param = (class_svc_param_t *)&acc.class_3;
5246 
5247 		acc.ls_code.ls_code = LA_ELS_ACC;
5248 		acc.common_service.fcph_version = 0x2006;
5249 		if (ha->topology & QL_N_PORT) {
5250 			/* clear F_Port indicator */
5251 			acc.common_service.cmn_features = 0x0800;
5252 		} else {
5253 			acc.common_service.cmn_features = 0x1b00;
5254 		}
5255 		CFG_IST(ha, CFG_CTRL_24258081) ?
5256 		    (acc.common_service.rx_bufsize = CHAR_TO_SHORT(
5257 		    ha->init_ctrl_blk.cb24.max_frame_length[0],
5258 		    ha->init_ctrl_blk.cb24.max_frame_length[1])) :
5259 		    (acc.common_service.rx_bufsize = CHAR_TO_SHORT(
5260 		    ha->init_ctrl_blk.cb.max_frame_length[0],
5261 		    ha->init_ctrl_blk.cb.max_frame_length[1]));
5262 		acc.common_service.conc_sequences = 0xff;
5263 		acc.common_service.relative_offset = 0x03;
5264 		acc.common_service.e_d_tov = 0x7d0;
5265 		if (accept) {
5266 			/* Use the saved N_Port WWNN and WWPN */
5267 			if (ha->n_port != NULL) {
5268 				bcopy((void *)&ha->n_port->port_name[0],
5269 				    (void *)&acc.nport_ww_name.raw_wwn[0], 8);
5270 				bcopy((void *)&ha->n_port->node_name[0],
5271 				    (void *)&acc.node_ww_name.raw_wwn[0], 8);
5272 				/* mark service options invalid */
5273 				class3_param->class_valid_svc_opt = 0x0800;
5274 			} else {
5275 				EL(ha, "ha->n_port is NULL\n");
5276 				/* Build RJT. */
5277 				acc.ls_code.ls_code = LA_ELS_RJT;
5278 
5279 				pkt->pkt_state = FC_PKT_TRAN_ERROR;
5280 				pkt->pkt_reason = FC_REASON_NO_CONNECTION;
5281 			}
5282 		} else {
5283 			bcopy((void *)&tq->port_name[0],
5284 			    (void *)&acc.nport_ww_name.raw_wwn[0], 8);
5285 			bcopy((void *)&tq->node_name[0],
5286 			    (void *)&acc.node_ww_name.raw_wwn[0], 8);
5287 
5288 			class3_param = (class_svc_param_t *)&acc.class_3;
5289 			class3_param->class_valid_svc_opt = 0x8800;
5290 			class3_param->recipient_ctl = tq->class3_recipient_ctl;
5291 			class3_param->rcv_data_size = tq->class3_rcv_data_size;
5292 			class3_param->conc_sequences =
5293 			    tq->class3_conc_sequences;
5294 			class3_param->open_sequences_per_exch =
5295 			    tq->class3_open_sequences_per_exch;
5296 		}
5297 	} else {
5298 		/* Build RJT. */
5299 		acc.ls_code.ls_code = LA_ELS_RJT;
5300 
5301 		pkt->pkt_state = FC_PKT_TRAN_ERROR;
5302 		pkt->pkt_reason = FC_REASON_NO_CONNECTION;
5303 		EL(ha, "LA_ELS_RJT, FC_REASON_NO_CONNECTION\n");
5304 	}
5305 
5306 	ddi_rep_put8(pkt->pkt_resp_acc, (uint8_t *)&acc,
5307 	    (uint8_t *)pkt->pkt_resp, sizeof (acc), DDI_DEV_AUTOINCR);
5308 
5309 	if (rval != FC_SUCCESS) {
5310 		EL(ha, "failed, rval = %xh\n", rval);
5311 	} else {
5312 		/*EMPTY*/
5313 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
5314 	}
5315 	return (rval);
5316 }
5317 
5318 /*
5319  * ql_els_logo
5320  *	Issue a extended link service logout request.
5321  *
5322  * Input:
5323  *	ha = adapter state pointer.
5324  *	pkt = pointer to fc_packet.
5325  *
5326  * Returns:
5327  *	FC_SUCCESS - the packet was accepted for transport.
5328  *	FC_TRANSPORT_ERROR - a transport error occurred.
5329  *
5330  * Context:
5331  *	Kernel context.
5332  */
5333 static int
ql_els_logo(ql_adapter_state_t * ha,fc_packet_t * pkt)5334 ql_els_logo(ql_adapter_state_t *ha, fc_packet_t *pkt)
5335 {
5336 	port_id_t	d_id;
5337 	ql_tgt_t	*tq;
5338 	la_els_logo_t	acc;
5339 	int		rval = FC_SUCCESS;
5340 
5341 	QL_PRINT_3(CE_CONT, "(%d): started, d_id=%xh\n", ha->instance,
5342 	    pkt->pkt_cmd_fhdr.d_id);
5343 
5344 	bzero(&acc, sizeof (acc));
5345 	d_id.b24 = pkt->pkt_cmd_fhdr.d_id;
5346 
5347 	tq = ql_d_id_to_queue(ha, d_id);
5348 	if (tq) {
5349 		DEVICE_QUEUE_LOCK(tq);
5350 		if (tq->d_id.b24 == BROADCAST_ADDR) {
5351 			DEVICE_QUEUE_UNLOCK(tq);
5352 			return (FC_SUCCESS);
5353 		}
5354 
5355 		tq->flags |= TQF_NEED_AUTHENTICATION;
5356 
5357 		do {
5358 			DEVICE_QUEUE_UNLOCK(tq);
5359 			(void) ql_abort_device(ha, tq, 1);
5360 
5361 			/*
5362 			 * Wait for commands to drain in F/W (doesn't
5363 			 * take more than a few milliseconds)
5364 			 */
5365 			ql_delay(ha, 10000);
5366 
5367 			DEVICE_QUEUE_LOCK(tq);
5368 		} while (tq->outcnt);
5369 
5370 		DEVICE_QUEUE_UNLOCK(tq);
5371 	}
5372 
5373 	if (ql_logout_port(ha, d_id) == QL_SUCCESS) {
5374 		/* Build ACC. */
5375 		acc.ls_code.ls_code = LA_ELS_ACC;
5376 
5377 		pkt->pkt_state = FC_PKT_SUCCESS;
5378 	} else {
5379 		/* Build RJT. */
5380 		acc.ls_code.ls_code = LA_ELS_RJT;
5381 
5382 		pkt->pkt_state = FC_PKT_TRAN_ERROR;
5383 		pkt->pkt_reason = FC_REASON_NO_CONNECTION;
5384 		EL(ha, "LA_ELS_RJT, FC_REASON_NO_CONNECTION\n");
5385 	}
5386 
5387 	ddi_rep_put8(pkt->pkt_resp_acc, (uint8_t *)&acc,
5388 	    (uint8_t *)pkt->pkt_resp, sizeof (acc), DDI_DEV_AUTOINCR);
5389 
5390 	if (rval != FC_SUCCESS) {
5391 		EL(ha, "failed, rval = %xh\n", rval);
5392 	} else {
5393 		/*EMPTY*/
5394 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
5395 	}
5396 	return (rval);
5397 }
5398 
5399 /*
5400  * ql_els_prli
5401  *	Issue a extended link service process login request.
5402  *
5403  * Input:
5404  *	ha = adapter state pointer.
5405  *	pkt = pointer to fc_packet.
5406  *
5407  * Returns:
5408  *	FC_SUCCESS - the packet was accepted for transport.
5409  *	FC_TRANSPORT_ERROR - a transport error occurred.
5410  *
5411  * Context:
5412  *	Kernel context.
5413  */
5414 static int
ql_els_prli(ql_adapter_state_t * ha,fc_packet_t * pkt)5415 ql_els_prli(ql_adapter_state_t *ha, fc_packet_t *pkt)
5416 {
5417 	ql_tgt_t		*tq;
5418 	port_id_t		d_id;
5419 	la_els_prli_t		acc;
5420 	prli_svc_param_t	*param;
5421 	ql_srb_t		*sp = (ql_srb_t *)pkt->pkt_fca_private;
5422 	int			rval = FC_SUCCESS;
5423 
5424 	QL_PRINT_3(CE_CONT, "(%d): started, d_id=%xh\n", ha->instance,
5425 	    pkt->pkt_cmd_fhdr.d_id);
5426 
5427 	d_id.b24 = pkt->pkt_cmd_fhdr.d_id;
5428 
5429 	tq = ql_d_id_to_queue(ha, d_id);
5430 	if (tq != NULL) {
5431 		(void) ql_get_port_database(ha, tq, PDF_NONE);
5432 
5433 		if ((ha->topology & QL_N_PORT) &&
5434 		    (tq->master_state == PD_STATE_PLOGI_COMPLETED)) {
5435 			ql_timeout_insert(ha, tq, sp);
5436 			ql_start_iocb(ha, sp);
5437 			rval = QL_CONSUMED;
5438 		} else {
5439 			/* Build ACC. */
5440 			bzero(&acc, sizeof (acc));
5441 			acc.ls_code = LA_ELS_ACC;
5442 			acc.page_length = 0x10;
5443 			acc.payload_length = tq->prli_payload_length;
5444 
5445 			param = (prli_svc_param_t *)&acc.service_params[0];
5446 			param->type = 0x08;
5447 			param->rsvd = 0x00;
5448 			param->process_assoc_flags = tq->prli_svc_param_word_0;
5449 			param->process_flags = tq->prli_svc_param_word_3;
5450 
5451 			ddi_rep_put8(pkt->pkt_resp_acc, (uint8_t *)&acc,
5452 			    (uint8_t *)pkt->pkt_resp, sizeof (acc),
5453 			    DDI_DEV_AUTOINCR);
5454 
5455 			pkt->pkt_state = FC_PKT_SUCCESS;
5456 		}
5457 	} else {
5458 		la_els_rjt_t rjt;
5459 
5460 		/* Build RJT. */
5461 		bzero(&rjt, sizeof (rjt));
5462 		rjt.ls_code.ls_code = LA_ELS_RJT;
5463 
5464 		ddi_rep_put8(pkt->pkt_resp_acc, (uint8_t *)&rjt,
5465 		    (uint8_t *)pkt->pkt_resp, sizeof (rjt), DDI_DEV_AUTOINCR);
5466 
5467 		pkt->pkt_state = FC_PKT_TRAN_ERROR;
5468 		pkt->pkt_reason = FC_REASON_NO_CONNECTION;
5469 		EL(ha, "LA_ELS_RJT, FC_REASON_NO_CONNECTION\n");
5470 	}
5471 
5472 	if ((rval != FC_SUCCESS) && (rval != QL_CONSUMED)) {
5473 		EL(ha, "failed, rval = %xh\n", rval);
5474 	} else {
5475 		/*EMPTY*/
5476 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
5477 	}
5478 	return (rval);
5479 }
5480 
5481 /*
5482  * ql_els_prlo
5483  *	Issue a extended link service process logout request.
5484  *
5485  * Input:
5486  *	ha = adapter state pointer.
5487  *	pkt = pointer to fc_packet.
5488  *
5489  * Returns:
5490  *	FC_SUCCESS - the packet was accepted for transport.
5491  *	FC_TRANSPORT_ERROR - a transport error occurred.
5492  *
5493  * Context:
5494  *	Kernel context.
5495  */
5496 /* ARGSUSED */
5497 static int
ql_els_prlo(ql_adapter_state_t * ha,fc_packet_t * pkt)5498 ql_els_prlo(ql_adapter_state_t *ha, fc_packet_t *pkt)
5499 {
5500 	la_els_prli_t	acc;
5501 	int		rval = FC_SUCCESS;
5502 
5503 	QL_PRINT_3(CE_CONT, "(%d): started, d_id=%xh\n", ha->instance,
5504 	    pkt->pkt_cmd_fhdr.d_id);
5505 
5506 	/* Build ACC. */
5507 	ddi_rep_get8(pkt->pkt_cmd_acc, (uint8_t *)&acc,
5508 	    (uint8_t *)pkt->pkt_cmd, sizeof (acc), DDI_DEV_AUTOINCR);
5509 
5510 	acc.ls_code = LA_ELS_ACC;
5511 	acc.service_params[2] = 1;
5512 
5513 	ddi_rep_put8(pkt->pkt_resp_acc, (uint8_t *)&acc,
5514 	    (uint8_t *)pkt->pkt_resp, sizeof (acc), DDI_DEV_AUTOINCR);
5515 
5516 	pkt->pkt_state = FC_PKT_SUCCESS;
5517 
5518 	if (rval != FC_SUCCESS) {
5519 		EL(ha, "failed, rval = %xh\n", rval);
5520 	} else {
5521 		/*EMPTY*/
5522 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
5523 	}
5524 	return (rval);
5525 }
5526 
5527 /*
5528  * ql_els_adisc
5529  *	Issue a extended link service address discovery request.
5530  *
5531  * Input:
5532  *	ha = adapter state pointer.
5533  *	pkt = pointer to fc_packet.
5534  *
5535  * Returns:
5536  *	FC_SUCCESS - the packet was accepted for transport.
5537  *	FC_TRANSPORT_ERROR - a transport error occurred.
5538  *
5539  * Context:
5540  *	Kernel context.
5541  */
5542 static int
ql_els_adisc(ql_adapter_state_t * ha,fc_packet_t * pkt)5543 ql_els_adisc(ql_adapter_state_t *ha, fc_packet_t *pkt)
5544 {
5545 	ql_dev_id_list_t	*list;
5546 	uint32_t		list_size;
5547 	ql_link_t		*link;
5548 	ql_tgt_t		*tq;
5549 	ql_lun_t		*lq;
5550 	port_id_t		d_id;
5551 	la_els_adisc_t		acc;
5552 	uint16_t		index, loop_id;
5553 	ql_mbx_data_t		mr;
5554 	int			rval = FC_SUCCESS;
5555 
5556 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
5557 
5558 	bzero(&acc, sizeof (acc));
5559 	d_id.b24 = pkt->pkt_cmd_fhdr.d_id;
5560 
5561 	/*
5562 	 * MBC_GET_PORT_DATABASE causes ADISC to go out to
5563 	 * the device from the firmware
5564 	 */
5565 	index = ql_alpa_to_index[d_id.b.al_pa];
5566 	tq = NULL;
5567 	for (link = ha->dev[index].first; link != NULL; link = link->next) {
5568 		tq = link->base_address;
5569 		if (tq->d_id.b24 == d_id.b24) {
5570 			break;
5571 		} else {
5572 			tq = NULL;
5573 		}
5574 	}
5575 
5576 	if ((tq != NULL) && (!VALID_DEVICE_ID(ha, tq->loop_id))) {
5577 		list_size = sizeof (ql_dev_id_list_t) * DEVICE_LIST_ENTRIES;
5578 		list = (ql_dev_id_list_t *)kmem_zalloc(list_size, KM_SLEEP);
5579 
5580 		if (list != NULL &&
5581 		    ql_get_id_list(ha, (caddr_t)list, list_size, &mr) ==
5582 		    QL_SUCCESS) {
5583 
5584 			for (index = 0; index < mr.mb[1]; index++) {
5585 				ql_dev_list(ha, list, index, &d_id, &loop_id);
5586 
5587 				if (tq->d_id.b24 == d_id.b24) {
5588 					tq->loop_id = loop_id;
5589 					break;
5590 				}
5591 			}
5592 		} else {
5593 			cmn_err(CE_WARN, "!%s(%d) didn't get list for %xh",
5594 			    QL_NAME, ha->instance, d_id.b24);
5595 			tq = NULL;
5596 		}
5597 		if ((tq != NULL) && (!VALID_DEVICE_ID(ha, tq->loop_id))) {
5598 			cmn_err(CE_WARN, "!%s(%d) no loop_id for adisc %xh",
5599 			    QL_NAME, ha->instance, tq->d_id.b24);
5600 			tq = NULL;
5601 		}
5602 
5603 		if (list != NULL) {
5604 			kmem_free(list, list_size);
5605 		}
5606 	}
5607 
5608 	if ((tq != NULL) && (VALID_DEVICE_ID(ha, tq->loop_id)) &&
5609 	    ql_get_port_database(ha, tq, PDF_ADISC) == QL_SUCCESS) {
5610 
5611 		/* Build ACC. */
5612 
5613 		DEVICE_QUEUE_LOCK(tq);
5614 		tq->flags &= ~TQF_NEED_AUTHENTICATION;
5615 		if (tq->prli_svc_param_word_3 & PRLI_W3_RETRY) {
5616 			for (link = tq->lun_queues.first; link != NULL;
5617 			    link = link->next) {
5618 				lq = link->base_address;
5619 
5620 				if (lq->cmd.first != NULL) {
5621 					ql_next(ha, lq);
5622 					DEVICE_QUEUE_LOCK(tq);
5623 				}
5624 			}
5625 		}
5626 		DEVICE_QUEUE_UNLOCK(tq);
5627 
5628 		acc.ls_code.ls_code = LA_ELS_ACC;
5629 		acc.hard_addr.hard_addr = tq->hard_addr.b24;
5630 
5631 		bcopy((void *)&tq->port_name[0],
5632 		    (void *)&acc.port_wwn.raw_wwn[0], 8);
5633 		bcopy((void *)&tq->node_name[0],
5634 		    (void *)&acc.node_wwn.raw_wwn[0], 8);
5635 
5636 		acc.nport_id.port_id = tq->d_id.b24;
5637 
5638 		pkt->pkt_state = FC_PKT_SUCCESS;
5639 	} else {
5640 		/* Build RJT. */
5641 		acc.ls_code.ls_code = LA_ELS_RJT;
5642 
5643 		pkt->pkt_state = FC_PKT_TRAN_ERROR;
5644 		pkt->pkt_reason = FC_REASON_NO_CONNECTION;
5645 		EL(ha, "LA_ELS_RJT, FC_REASON_NO_CONNECTION\n");
5646 	}
5647 
5648 	ddi_rep_put8(pkt->pkt_resp_acc, (uint8_t *)&acc,
5649 	    (uint8_t *)pkt->pkt_resp, sizeof (acc), DDI_DEV_AUTOINCR);
5650 
5651 	if (rval != FC_SUCCESS) {
5652 		EL(ha, "failed, rval = %xh\n", rval);
5653 	} else {
5654 		/*EMPTY*/
5655 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
5656 	}
5657 	return (rval);
5658 }
5659 
5660 /*
5661  * ql_els_linit
5662  *	Issue a extended link service loop initialize request.
5663  *
5664  * Input:
5665  *	ha = adapter state pointer.
5666  *	pkt = pointer to fc_packet.
5667  *
5668  * Returns:
5669  *	FC_SUCCESS - the packet was accepted for transport.
5670  *	FC_TRANSPORT_ERROR - a transport error occurred.
5671  *
5672  * Context:
5673  *	Kernel context.
5674  */
5675 static int
ql_els_linit(ql_adapter_state_t * ha,fc_packet_t * pkt)5676 ql_els_linit(ql_adapter_state_t *ha, fc_packet_t *pkt)
5677 {
5678 	ddi_dma_cookie_t	*cp;
5679 	uint32_t		cnt;
5680 	conv_num_t		n;
5681 	port_id_t		d_id;
5682 	int			rval = FC_SUCCESS;
5683 
5684 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
5685 
5686 	d_id.b24 = pkt->pkt_cmd_fhdr.d_id;
5687 	if (ha->topology & QL_SNS_CONNECTION) {
5688 		fc_linit_req_t els;
5689 		lfa_cmd_t lfa;
5690 
5691 		ddi_rep_get8(pkt->pkt_cmd_acc, (uint8_t *)&els,
5692 		    (uint8_t *)pkt->pkt_cmd, sizeof (els), DDI_DEV_AUTOINCR);
5693 
5694 		/* Setup LFA mailbox command data. */
5695 		bzero((void *)&lfa, sizeof (lfa_cmd_t));
5696 
5697 		lfa.resp_buffer_length[0] = 4;
5698 
5699 		cp = pkt->pkt_resp_cookie;
5700 		if (CFG_IST(ha, CFG_ENABLE_64BIT_ADDRESSING)) {
5701 			n.size64 = (uint64_t)cp->dmac_laddress;
5702 			LITTLE_ENDIAN_64(&n.size64);
5703 		} else {
5704 			n.size32[0] = LSD(cp->dmac_laddress);
5705 			LITTLE_ENDIAN_32(&n.size32[0]);
5706 			n.size32[1] = MSD(cp->dmac_laddress);
5707 			LITTLE_ENDIAN_32(&n.size32[1]);
5708 		}
5709 
5710 		/* Set buffer address. */
5711 		for (cnt = 0; cnt < 8; cnt++) {
5712 			lfa.resp_buffer_address[cnt] = n.size8[cnt];
5713 		}
5714 
5715 		lfa.subcommand_length[0] = 4;
5716 		n.size32[0] = d_id.b24;
5717 		LITTLE_ENDIAN_32(&n.size32[0]);
5718 		lfa.addr[0] = n.size8[0];
5719 		lfa.addr[1] = n.size8[1];
5720 		lfa.addr[2] = n.size8[2];
5721 		lfa.subcommand[1] = 0x70;
5722 		lfa.payload[2] = els.func;
5723 		lfa.payload[4] = els.lip_b3;
5724 		lfa.payload[5] = els.lip_b4;
5725 
5726 		if (ql_send_lfa(ha, &lfa) != QL_SUCCESS) {
5727 			pkt->pkt_state = FC_PKT_TRAN_ERROR;
5728 		} else {
5729 			pkt->pkt_state = FC_PKT_SUCCESS;
5730 		}
5731 	} else {
5732 		fc_linit_resp_t rjt;
5733 
5734 		/* Build RJT. */
5735 		bzero(&rjt, sizeof (rjt));
5736 		rjt.ls_code.ls_code = LA_ELS_RJT;
5737 
5738 		ddi_rep_put8(pkt->pkt_resp_acc, (uint8_t *)&rjt,
5739 		    (uint8_t *)pkt->pkt_resp, sizeof (rjt), DDI_DEV_AUTOINCR);
5740 
5741 		pkt->pkt_state = FC_PKT_TRAN_ERROR;
5742 		pkt->pkt_reason = FC_REASON_NO_CONNECTION;
5743 		EL(ha, "LA_ELS_RJT, FC_REASON_NO_CONNECTION\n");
5744 	}
5745 
5746 	if (rval != FC_SUCCESS) {
5747 		EL(ha, "failed, rval = %xh\n", rval);
5748 	} else {
5749 		/*EMPTY*/
5750 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
5751 	}
5752 	return (rval);
5753 }
5754 
5755 /*
5756  * ql_els_lpc
5757  *	Issue a extended link service loop control request.
5758  *
5759  * Input:
5760  *	ha = adapter state pointer.
5761  *	pkt = pointer to fc_packet.
5762  *
5763  * Returns:
5764  *	FC_SUCCESS - the packet was accepted for transport.
5765  *	FC_TRANSPORT_ERROR - a transport error occurred.
5766  *
5767  * Context:
5768  *	Kernel context.
5769  */
5770 static int
ql_els_lpc(ql_adapter_state_t * ha,fc_packet_t * pkt)5771 ql_els_lpc(ql_adapter_state_t *ha, fc_packet_t *pkt)
5772 {
5773 	ddi_dma_cookie_t	*cp;
5774 	uint32_t		cnt;
5775 	conv_num_t		n;
5776 	port_id_t		d_id;
5777 	int			rval = FC_SUCCESS;
5778 
5779 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
5780 
5781 	d_id.b24 = pkt->pkt_cmd_fhdr.d_id;
5782 	if (ha->topology & QL_SNS_CONNECTION) {
5783 		ql_lpc_t els;
5784 		lfa_cmd_t lfa;
5785 
5786 		ddi_rep_get8(pkt->pkt_cmd_acc, (uint8_t *)&els,
5787 		    (uint8_t *)pkt->pkt_cmd, sizeof (els), DDI_DEV_AUTOINCR);
5788 
5789 		/* Setup LFA mailbox command data. */
5790 		bzero((void *)&lfa, sizeof (lfa_cmd_t));
5791 
5792 		lfa.resp_buffer_length[0] = 4;
5793 
5794 		cp = pkt->pkt_resp_cookie;
5795 		if (CFG_IST(ha, CFG_ENABLE_64BIT_ADDRESSING)) {
5796 			n.size64 = (uint64_t)(cp->dmac_laddress);
5797 			LITTLE_ENDIAN_64(&n.size64);
5798 		} else {
5799 			n.size32[0] = cp->dmac_address;
5800 			LITTLE_ENDIAN_32(&n.size32[0]);
5801 			n.size32[1] = 0;
5802 		}
5803 
5804 		/* Set buffer address. */
5805 		for (cnt = 0; cnt < 8; cnt++) {
5806 			lfa.resp_buffer_address[cnt] = n.size8[cnt];
5807 		}
5808 
5809 		lfa.subcommand_length[0] = 20;
5810 		n.size32[0] = d_id.b24;
5811 		LITTLE_ENDIAN_32(&n.size32[0]);
5812 		lfa.addr[0] = n.size8[0];
5813 		lfa.addr[1] = n.size8[1];
5814 		lfa.addr[2] = n.size8[2];
5815 		lfa.subcommand[1] = 0x71;
5816 		lfa.payload[4] = els.port_control;
5817 		bcopy((void *)&els.lpb[0], (void *)&lfa.payload[6], 32);
5818 
5819 		if (ql_send_lfa(ha, &lfa) != QL_SUCCESS) {
5820 			pkt->pkt_state = FC_PKT_TRAN_ERROR;
5821 		} else {
5822 			pkt->pkt_state = FC_PKT_SUCCESS;
5823 		}
5824 	} else {
5825 		ql_lpc_resp_t rjt;
5826 
5827 		/* Build RJT. */
5828 		bzero(&rjt, sizeof (rjt));
5829 		rjt.ls_code.ls_code = LA_ELS_RJT;
5830 
5831 		ddi_rep_put8(pkt->pkt_resp_acc, (uint8_t *)&rjt,
5832 		    (uint8_t *)pkt->pkt_resp, sizeof (rjt), DDI_DEV_AUTOINCR);
5833 
5834 		pkt->pkt_state = FC_PKT_TRAN_ERROR;
5835 		pkt->pkt_reason = FC_REASON_NO_CONNECTION;
5836 		EL(ha, "LA_ELS_RJT, FC_REASON_NO_CONNECTION\n");
5837 	}
5838 
5839 	if (rval != FC_SUCCESS) {
5840 		EL(ha, "failed, rval = %xh\n", rval);
5841 	} else {
5842 		/*EMPTY*/
5843 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
5844 	}
5845 	return (rval);
5846 }
5847 
5848 /*
5849  * ql_els_lsts
5850  *	Issue a extended link service loop status request.
5851  *
5852  * Input:
5853  *	ha = adapter state pointer.
5854  *	pkt = pointer to fc_packet.
5855  *
5856  * Returns:
5857  *	FC_SUCCESS - the packet was accepted for transport.
5858  *	FC_TRANSPORT_ERROR - a transport error occurred.
5859  *
5860  * Context:
5861  *	Kernel context.
5862  */
5863 static int
ql_els_lsts(ql_adapter_state_t * ha,fc_packet_t * pkt)5864 ql_els_lsts(ql_adapter_state_t *ha, fc_packet_t *pkt)
5865 {
5866 	ddi_dma_cookie_t	*cp;
5867 	uint32_t		cnt;
5868 	conv_num_t		n;
5869 	port_id_t		d_id;
5870 	int			rval = FC_SUCCESS;
5871 
5872 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
5873 
5874 	d_id.b24 = pkt->pkt_cmd_fhdr.d_id;
5875 	if (ha->topology & QL_SNS_CONNECTION) {
5876 		fc_lsts_req_t els;
5877 		lfa_cmd_t lfa;
5878 
5879 		ddi_rep_get8(pkt->pkt_cmd_acc, (uint8_t *)&els,
5880 		    (uint8_t *)pkt->pkt_cmd, sizeof (els), DDI_DEV_AUTOINCR);
5881 
5882 		/* Setup LFA mailbox command data. */
5883 		bzero((void *)&lfa, sizeof (lfa_cmd_t));
5884 
5885 		lfa.resp_buffer_length[0] = 84;
5886 
5887 		cp = pkt->pkt_resp_cookie;
5888 		if (CFG_IST(ha, CFG_ENABLE_64BIT_ADDRESSING)) {
5889 			n.size64 = cp->dmac_laddress;
5890 			LITTLE_ENDIAN_64(&n.size64);
5891 		} else {
5892 			n.size32[0] = cp->dmac_address;
5893 			LITTLE_ENDIAN_32(&n.size32[0]);
5894 			n.size32[1] = 0;
5895 		}
5896 
5897 		/* Set buffer address. */
5898 		for (cnt = 0; cnt < 8; cnt++) {
5899 			lfa.resp_buffer_address[cnt] = n.size8[cnt];
5900 		}
5901 
5902 		lfa.subcommand_length[0] = 2;
5903 		n.size32[0] = d_id.b24;
5904 		LITTLE_ENDIAN_32(&n.size32[0]);
5905 		lfa.addr[0] = n.size8[0];
5906 		lfa.addr[1] = n.size8[1];
5907 		lfa.addr[2] = n.size8[2];
5908 		lfa.subcommand[1] = 0x72;
5909 
5910 		if (ql_send_lfa(ha, &lfa) != QL_SUCCESS) {
5911 			pkt->pkt_state = FC_PKT_TRAN_ERROR;
5912 		} else {
5913 			pkt->pkt_state = FC_PKT_SUCCESS;
5914 		}
5915 	} else {
5916 		fc_lsts_resp_t rjt;
5917 
5918 		/* Build RJT. */
5919 		bzero(&rjt, sizeof (rjt));
5920 		rjt.lsts_ls_code.ls_code = LA_ELS_RJT;
5921 
5922 		ddi_rep_put8(pkt->pkt_resp_acc, (uint8_t *)&rjt,
5923 		    (uint8_t *)pkt->pkt_resp, sizeof (rjt), DDI_DEV_AUTOINCR);
5924 
5925 		pkt->pkt_state = FC_PKT_TRAN_ERROR;
5926 		pkt->pkt_reason = FC_REASON_NO_CONNECTION;
5927 		EL(ha, "LA_ELS_RJT, FC_REASON_NO_CONNECTION\n");
5928 	}
5929 
5930 	if (rval != FC_SUCCESS) {
5931 		EL(ha, "failed=%xh\n", rval);
5932 	} else {
5933 		/*EMPTY*/
5934 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
5935 	}
5936 	return (rval);
5937 }
5938 
5939 /*
5940  * ql_els_scr
5941  *	Issue a extended link service state change registration request.
5942  *
5943  * Input:
5944  *	ha = adapter state pointer.
5945  *	pkt = pointer to fc_packet.
5946  *
5947  * Returns:
5948  *	FC_SUCCESS - the packet was accepted for transport.
5949  *	FC_TRANSPORT_ERROR - a transport error occurred.
5950  *
5951  * Context:
5952  *	Kernel context.
5953  */
5954 static int
ql_els_scr(ql_adapter_state_t * ha,fc_packet_t * pkt)5955 ql_els_scr(ql_adapter_state_t *ha, fc_packet_t *pkt)
5956 {
5957 	fc_scr_resp_t	acc;
5958 	int		rval = FC_SUCCESS;
5959 
5960 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
5961 
5962 	bzero(&acc, sizeof (acc));
5963 	if (ha->topology & QL_SNS_CONNECTION) {
5964 		fc_scr_req_t els;
5965 
5966 		ddi_rep_get8(pkt->pkt_cmd_acc, (uint8_t *)&els,
5967 		    (uint8_t *)pkt->pkt_cmd, sizeof (els), DDI_DEV_AUTOINCR);
5968 
5969 		if (ql_send_change_request(ha, els.scr_func) ==
5970 		    QL_SUCCESS) {
5971 			/* Build ACC. */
5972 			acc.scr_acc = LA_ELS_ACC;
5973 
5974 			pkt->pkt_state = FC_PKT_SUCCESS;
5975 		} else {
5976 			/* Build RJT. */
5977 			acc.scr_acc = LA_ELS_RJT;
5978 
5979 			pkt->pkt_state = FC_PKT_TRAN_ERROR;
5980 			pkt->pkt_reason = FC_REASON_HW_ERROR;
5981 			EL(ha, "LA_ELS_RJT, FC_REASON_HW_ERROR\n");
5982 		}
5983 	} else {
5984 		/* Build RJT. */
5985 		acc.scr_acc = LA_ELS_RJT;
5986 
5987 		pkt->pkt_state = FC_PKT_TRAN_ERROR;
5988 		pkt->pkt_reason = FC_REASON_NO_CONNECTION;
5989 		EL(ha, "LA_ELS_RJT, FC_REASON_NO_CONNECTION\n");
5990 	}
5991 
5992 	ddi_rep_put8(pkt->pkt_resp_acc, (uint8_t *)&acc,
5993 	    (uint8_t *)pkt->pkt_resp, sizeof (acc), DDI_DEV_AUTOINCR);
5994 
5995 	if (rval != FC_SUCCESS) {
5996 		EL(ha, "failed, rval = %xh\n", rval);
5997 	} else {
5998 		/*EMPTY*/
5999 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
6000 	}
6001 	return (rval);
6002 }
6003 
6004 /*
6005  * ql_els_rscn
6006  *	Issue a extended link service register state
6007  *	change notification request.
6008  *
6009  * Input:
6010  *	ha = adapter state pointer.
6011  *	pkt = pointer to fc_packet.
6012  *
6013  * Returns:
6014  *	FC_SUCCESS - the packet was accepted for transport.
6015  *	FC_TRANSPORT_ERROR - a transport error occurred.
6016  *
6017  * Context:
6018  *	Kernel context.
6019  */
6020 static int
ql_els_rscn(ql_adapter_state_t * ha,fc_packet_t * pkt)6021 ql_els_rscn(ql_adapter_state_t *ha, fc_packet_t *pkt)
6022 {
6023 	ql_rscn_resp_t	acc;
6024 	int		rval = FC_SUCCESS;
6025 
6026 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
6027 
6028 	bzero(&acc, sizeof (acc));
6029 	if (ha->topology & QL_SNS_CONNECTION) {
6030 		/* Build ACC. */
6031 		acc.scr_acc = LA_ELS_ACC;
6032 
6033 		pkt->pkt_state = FC_PKT_SUCCESS;
6034 	} else {
6035 		/* Build RJT. */
6036 		acc.scr_acc = LA_ELS_RJT;
6037 
6038 		pkt->pkt_state = FC_PKT_TRAN_ERROR;
6039 		pkt->pkt_reason = FC_REASON_NO_CONNECTION;
6040 		EL(ha, "LA_ELS_RJT, FC_REASON_NO_CONNECTION\n");
6041 	}
6042 
6043 	ddi_rep_put8(pkt->pkt_resp_acc, (uint8_t *)&acc,
6044 	    (uint8_t *)pkt->pkt_resp, sizeof (acc), DDI_DEV_AUTOINCR);
6045 
6046 	if (rval != FC_SUCCESS) {
6047 		EL(ha, "failed, rval = %xh\n", rval);
6048 	} else {
6049 		/*EMPTY*/
6050 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
6051 	}
6052 	return (rval);
6053 }
6054 
6055 /*
6056  * ql_els_farp_req
6057  *	Issue FC Address Resolution Protocol (FARP)
6058  *	extended link service request.
6059  *
6060  *	Note: not supported.
6061  *
6062  * Input:
6063  *	ha = adapter state pointer.
6064  *	pkt = pointer to fc_packet.
6065  *
6066  * Returns:
6067  *	FC_SUCCESS - the packet was accepted for transport.
6068  *	FC_TRANSPORT_ERROR - a transport error occurred.
6069  *
6070  * Context:
6071  *	Kernel context.
6072  */
6073 static int
ql_els_farp_req(ql_adapter_state_t * ha,fc_packet_t * pkt)6074 ql_els_farp_req(ql_adapter_state_t *ha, fc_packet_t *pkt)
6075 {
6076 	ql_acc_rjt_t	acc;
6077 	int		rval = FC_SUCCESS;
6078 
6079 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
6080 
6081 	bzero(&acc, sizeof (acc));
6082 
6083 	/* Build ACC. */
6084 	acc.ls_code.ls_code = LA_ELS_ACC;
6085 
6086 	pkt->pkt_state = FC_PKT_SUCCESS;
6087 
6088 	ddi_rep_put8(pkt->pkt_resp_acc, (uint8_t *)&acc,
6089 	    (uint8_t *)pkt->pkt_resp, sizeof (acc), DDI_DEV_AUTOINCR);
6090 
6091 	if (rval != FC_SUCCESS) {
6092 		EL(ha, "failed, rval = %xh\n", rval);
6093 	} else {
6094 		/*EMPTY*/
6095 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
6096 	}
6097 	return (rval);
6098 }
6099 
6100 /*
6101  * ql_els_farp_reply
6102  *	Issue FC Address Resolution Protocol (FARP)
6103  *	extended link service reply.
6104  *
6105  *	Note: not supported.
6106  *
6107  * Input:
6108  *	ha = adapter state pointer.
6109  *	pkt = pointer to fc_packet.
6110  *
6111  * Returns:
6112  *	FC_SUCCESS - the packet was accepted for transport.
6113  *	FC_TRANSPORT_ERROR - a transport error occurred.
6114  *
6115  * Context:
6116  *	Kernel context.
6117  */
6118 /* ARGSUSED */
6119 static int
ql_els_farp_reply(ql_adapter_state_t * ha,fc_packet_t * pkt)6120 ql_els_farp_reply(ql_adapter_state_t *ha, fc_packet_t *pkt)
6121 {
6122 	ql_acc_rjt_t	acc;
6123 	int		rval = FC_SUCCESS;
6124 
6125 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
6126 
6127 	bzero(&acc, sizeof (acc));
6128 
6129 	/* Build ACC. */
6130 	acc.ls_code.ls_code = LA_ELS_ACC;
6131 
6132 	pkt->pkt_state = FC_PKT_SUCCESS;
6133 
6134 	ddi_rep_put8(pkt->pkt_resp_acc, (uint8_t *)&acc,
6135 	    (uint8_t *)pkt->pkt_resp, sizeof (acc), DDI_DEV_AUTOINCR);
6136 
6137 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
6138 
6139 	return (rval);
6140 }
6141 
6142 static int
ql_els_rnid(ql_adapter_state_t * ha,fc_packet_t * pkt)6143 ql_els_rnid(ql_adapter_state_t *ha, fc_packet_t *pkt)
6144 {
6145 	uchar_t			*rnid_acc;
6146 	port_id_t		d_id;
6147 	ql_link_t		*link;
6148 	ql_tgt_t		*tq;
6149 	uint16_t		index;
6150 	la_els_rnid_acc_t	acc;
6151 	la_els_rnid_t		*req;
6152 	size_t			req_len;
6153 
6154 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
6155 
6156 	req_len =  FCIO_RNID_MAX_DATA_LEN + sizeof (fc_rnid_hdr_t);
6157 	d_id.b24 = pkt->pkt_cmd_fhdr.d_id;
6158 	index = ql_alpa_to_index[d_id.b.al_pa];
6159 
6160 	tq = NULL;
6161 	for (link = ha->dev[index].first; link != NULL; link = link->next) {
6162 		tq = link->base_address;
6163 		if (tq->d_id.b24 == d_id.b24) {
6164 			break;
6165 		} else {
6166 			tq = NULL;
6167 		}
6168 	}
6169 
6170 	/* Allocate memory for rnid status block */
6171 	rnid_acc = kmem_zalloc(req_len, KM_SLEEP);
6172 
6173 	bzero(&acc, sizeof (acc));
6174 
6175 	req = (la_els_rnid_t *)pkt->pkt_cmd;
6176 	if ((tq == NULL) || (!VALID_DEVICE_ID(ha, tq->loop_id)) ||
6177 	    (ql_send_rnid_els(ha, tq->loop_id, req->data_format, req_len,
6178 	    (caddr_t)rnid_acc) != QL_SUCCESS)) {
6179 
6180 		kmem_free(rnid_acc, req_len);
6181 		acc.ls_code.ls_code = LA_ELS_RJT;
6182 
6183 		ddi_rep_put8(pkt->pkt_resp_acc, (uint8_t *)&acc,
6184 		    (uint8_t *)pkt->pkt_resp, sizeof (acc), DDI_DEV_AUTOINCR);
6185 
6186 		pkt->pkt_state = FC_PKT_TRAN_ERROR;
6187 		pkt->pkt_reason = FC_REASON_NO_CONNECTION;
6188 		EL(ha, "LA_ELS_RJT, FC_REASON_NO_CONNECTION\n");
6189 
6190 		return (FC_FAILURE);
6191 	}
6192 
6193 	acc.ls_code.ls_code = LA_ELS_ACC;
6194 	bcopy(rnid_acc, &acc.hdr, req_len);
6195 	ddi_rep_put8(pkt->pkt_resp_acc, (uint8_t *)&acc,
6196 	    (uint8_t *)pkt->pkt_resp, sizeof (acc), DDI_DEV_AUTOINCR);
6197 
6198 	kmem_free(rnid_acc, req_len);
6199 	pkt->pkt_state = FC_PKT_SUCCESS;
6200 
6201 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
6202 
6203 	return (FC_SUCCESS);
6204 }
6205 
6206 static int
ql_els_rls(ql_adapter_state_t * ha,fc_packet_t * pkt)6207 ql_els_rls(ql_adapter_state_t *ha, fc_packet_t *pkt)
6208 {
6209 	fc_rls_acc_t		*rls_acc;
6210 	port_id_t		d_id;
6211 	ql_link_t		*link;
6212 	ql_tgt_t		*tq;
6213 	uint16_t		index;
6214 	la_els_rls_acc_t	acc;
6215 
6216 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
6217 
6218 	d_id.b24 = pkt->pkt_cmd_fhdr.d_id;
6219 	index = ql_alpa_to_index[d_id.b.al_pa];
6220 
6221 	tq = NULL;
6222 	for (link = ha->dev[index].first; link != NULL; link = link->next) {
6223 		tq = link->base_address;
6224 		if (tq->d_id.b24 == d_id.b24) {
6225 			break;
6226 		} else {
6227 			tq = NULL;
6228 		}
6229 	}
6230 
6231 	/* Allocate memory for link error status block */
6232 	rls_acc = kmem_zalloc(sizeof (*rls_acc), KM_SLEEP);
6233 
6234 	bzero(&acc, sizeof (la_els_rls_acc_t));
6235 
6236 	if ((tq == NULL) || (!VALID_DEVICE_ID(ha, tq->loop_id)) ||
6237 	    (ql_get_link_status(ha, tq->loop_id, sizeof (*rls_acc),
6238 	    (caddr_t)rls_acc, 0) != QL_SUCCESS)) {
6239 
6240 		kmem_free(rls_acc, sizeof (*rls_acc));
6241 		acc.ls_code.ls_code = LA_ELS_RJT;
6242 
6243 		ddi_rep_put8(pkt->pkt_resp_acc, (uint8_t *)&acc,
6244 		    (uint8_t *)pkt->pkt_resp, sizeof (acc), DDI_DEV_AUTOINCR);
6245 
6246 		pkt->pkt_state = FC_PKT_TRAN_ERROR;
6247 		pkt->pkt_reason = FC_REASON_NO_CONNECTION;
6248 		EL(ha, "LA_ELS_RJT, FC_REASON_NO_CONNECTION\n");
6249 
6250 		return (FC_FAILURE);
6251 	}
6252 
6253 	LITTLE_ENDIAN_32(&rls_acc->rls_link_fail);
6254 	LITTLE_ENDIAN_32(&rls_acc->rls_sync_loss);
6255 	LITTLE_ENDIAN_32(&rls_acc->rls_sig_loss);
6256 	LITTLE_ENDIAN_32(&rls_acc->rls_invalid_word);
6257 	LITTLE_ENDIAN_32(&rls_acc->rls_invalid_crc);
6258 
6259 	acc.ls_code.ls_code = LA_ELS_ACC;
6260 	acc.rls_link_params.rls_link_fail = rls_acc->rls_link_fail;
6261 	acc.rls_link_params.rls_sync_loss = rls_acc->rls_sync_loss;
6262 	acc.rls_link_params.rls_sig_loss  = rls_acc->rls_sig_loss;
6263 	acc.rls_link_params.rls_invalid_word = rls_acc->rls_invalid_word;
6264 	acc.rls_link_params.rls_invalid_crc = rls_acc->rls_invalid_crc;
6265 	ddi_rep_put8(pkt->pkt_resp_acc, (uint8_t *)&acc,
6266 	    (uint8_t *)pkt->pkt_resp, sizeof (acc), DDI_DEV_AUTOINCR);
6267 
6268 	kmem_free(rls_acc, sizeof (*rls_acc));
6269 	pkt->pkt_state = FC_PKT_SUCCESS;
6270 
6271 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
6272 
6273 	return (FC_SUCCESS);
6274 }
6275 
6276 static int
ql_busy_plogi(ql_adapter_state_t * ha,fc_packet_t * pkt,ql_tgt_t * tq)6277 ql_busy_plogi(ql_adapter_state_t *ha, fc_packet_t *pkt, ql_tgt_t *tq)
6278 {
6279 	port_id_t	d_id;
6280 	ql_srb_t	*sp;
6281 	fc_unsol_buf_t  *ubp;
6282 	ql_link_t	*link, *next_link;
6283 	int		rval = FC_SUCCESS;
6284 	int		cnt = 5;
6285 
6286 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
6287 
6288 	/*
6289 	 * we need to ensure that q->outcnt == 0, otherwise
6290 	 * any cmd completed with PKT_PORT_OFFLINE after PLOGI
6291 	 * will confuse ulps.
6292 	 */
6293 
6294 	DEVICE_QUEUE_LOCK(tq);
6295 	do {
6296 		/*
6297 		 * wait for the cmds to get drained. If they
6298 		 * don't get drained then the transport will
6299 		 * retry PLOGI after few secs.
6300 		 */
6301 		if (tq->outcnt != 0) {
6302 			rval = FC_TRAN_BUSY;
6303 			DEVICE_QUEUE_UNLOCK(tq);
6304 			ql_delay(ha, 10000);
6305 			DEVICE_QUEUE_LOCK(tq);
6306 			cnt--;
6307 			if (!cnt) {
6308 				cmn_err(CE_NOTE, "!%s(%d) Plogi busy"
6309 				    " for %xh outcount %xh", QL_NAME,
6310 				    ha->instance, tq->d_id.b24, tq->outcnt);
6311 			}
6312 		} else {
6313 			rval = FC_SUCCESS;
6314 			break;
6315 		}
6316 	} while (cnt > 0);
6317 	DEVICE_QUEUE_UNLOCK(tq);
6318 
6319 	/*
6320 	 * return, if busy or if the plogi was asynchronous.
6321 	 */
6322 	if ((rval != FC_SUCCESS) ||
6323 	    (!(pkt->pkt_tran_flags & FC_TRAN_NO_INTR) &&
6324 	    pkt->pkt_comp)) {
6325 		QL_PRINT_3(CE_CONT, "(%d): done, busy or async\n",
6326 		    ha->instance);
6327 		return (rval);
6328 	}
6329 
6330 	/*
6331 	 * Let us give daemon sufficient time and hopefully
6332 	 * when transport retries PLOGI, it would have flushed
6333 	 * callback queue.
6334 	 */
6335 	TASK_DAEMON_LOCK(ha);
6336 	for (link = ha->callback_queue.first; link != NULL;
6337 	    link = next_link) {
6338 		next_link = link->next;
6339 		sp = link->base_address;
6340 		if (sp->flags & SRB_UB_CALLBACK) {
6341 			ubp = ha->ub_array[sp->handle];
6342 			d_id.b24 = ubp->ub_frame.s_id;
6343 		} else {
6344 			d_id.b24 = sp->pkt->pkt_cmd_fhdr.d_id;
6345 		}
6346 		if (tq->d_id.b24 == d_id.b24) {
6347 			cmn_err(CE_NOTE, "!%s(%d) Plogi busy for %xh", QL_NAME,
6348 			    ha->instance, tq->d_id.b24);
6349 			rval = FC_TRAN_BUSY;
6350 			break;
6351 		}
6352 	}
6353 	TASK_DAEMON_UNLOCK(ha);
6354 
6355 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
6356 
6357 	return (rval);
6358 }
6359 
6360 /*
6361  * ql_login_port
6362  *	Logs in a device if not already logged in.
6363  *
6364  * Input:
6365  *	ha = adapter state pointer.
6366  *	d_id = 24 bit port ID.
6367  *	DEVICE_QUEUE_LOCK must be released.
6368  *
6369  * Returns:
6370  *	QL local function return status code.
6371  *
6372  * Context:
6373  *	Kernel context.
6374  */
6375 static int
ql_login_port(ql_adapter_state_t * ha,port_id_t d_id)6376 ql_login_port(ql_adapter_state_t *ha, port_id_t d_id)
6377 {
6378 	ql_adapter_state_t	*vha;
6379 	ql_link_t		*link;
6380 	uint16_t		index;
6381 	ql_tgt_t		*tq, *tq2;
6382 	uint16_t		loop_id, first_loop_id, last_loop_id;
6383 	int			rval = QL_SUCCESS;
6384 
6385 	QL_PRINT_3(CE_CONT, "(%d): started, d_id=%xh\n", ha->instance,
6386 	    d_id.b24);
6387 
6388 	/* Get head queue index. */
6389 	index = ql_alpa_to_index[d_id.b.al_pa];
6390 
6391 	/* Check for device already has a queue. */
6392 	tq = NULL;
6393 	for (link = ha->dev[index].first; link != NULL; link = link->next) {
6394 		tq = link->base_address;
6395 		if (tq->d_id.b24 == d_id.b24) {
6396 			loop_id = tq->loop_id;
6397 			break;
6398 		} else {
6399 			tq = NULL;
6400 		}
6401 	}
6402 
6403 	/* Let's stop issuing any IO and unsolicited logo */
6404 	if ((tq != NULL) && (!(ddi_in_panic()))) {
6405 		DEVICE_QUEUE_LOCK(tq);
6406 		tq->flags |= (TQF_QUEUE_SUSPENDED | TQF_PLOGI_PROGRS);
6407 		tq->flags &= ~TQF_RSCN_RCVD;
6408 		DEVICE_QUEUE_UNLOCK(tq);
6409 	}
6410 	if ((tq != NULL) && (tq->loop_id & PORT_LOST_ID) &&
6411 	    !(tq->flags & TQF_FABRIC_DEVICE)) {
6412 		loop_id = (uint16_t)(tq->loop_id & ~PORT_LOST_ID);
6413 	}
6414 
6415 	/* Special case for Nameserver */
6416 	if (d_id.b24 == 0xFFFFFC) {
6417 		loop_id = (uint16_t)(CFG_IST(ha, CFG_CTRL_24258081) ?
6418 		    SNS_24XX_HDL : SIMPLE_NAME_SERVER_LOOP_ID);
6419 		if (tq == NULL) {
6420 			ADAPTER_STATE_LOCK(ha);
6421 			tq = ql_dev_init(ha, d_id, loop_id);
6422 			ADAPTER_STATE_UNLOCK(ha);
6423 			if (tq == NULL) {
6424 				EL(ha, "failed=%xh, d_id=%xh\n",
6425 				    QL_FUNCTION_FAILED, d_id.b24);
6426 				return (QL_FUNCTION_FAILED);
6427 			}
6428 		}
6429 		if (!(CFG_IST(ha, CFG_CTRL_8021))) {
6430 			rval = ql_login_fabric_port(ha, tq, loop_id);
6431 			if (rval == QL_SUCCESS) {
6432 				tq->loop_id = loop_id;
6433 				tq->flags |= TQF_FABRIC_DEVICE;
6434 				(void) ql_get_port_database(ha, tq, PDF_NONE);
6435 			}
6436 		} else {
6437 			ha->topology = (uint8_t)
6438 			    (ha->topology | QL_SNS_CONNECTION);
6439 		}
6440 	/* Check for device already logged in. */
6441 	} else if (tq != NULL && VALID_DEVICE_ID(ha, loop_id)) {
6442 		if (tq->flags & TQF_FABRIC_DEVICE) {
6443 			rval = ql_login_fabric_port(ha, tq, loop_id);
6444 			if (rval == QL_PORT_ID_USED) {
6445 				rval = QL_SUCCESS;
6446 			}
6447 		} else if (LOCAL_LOOP_ID(loop_id)) {
6448 			rval = ql_login_lport(ha, tq, loop_id, (uint16_t)
6449 			    (tq->flags & TQF_INITIATOR_DEVICE ?
6450 			    LLF_NONE : LLF_PLOGI));
6451 			if (rval == QL_SUCCESS) {
6452 				DEVICE_QUEUE_LOCK(tq);
6453 				tq->loop_id = loop_id;
6454 				DEVICE_QUEUE_UNLOCK(tq);
6455 			}
6456 		}
6457 	} else if (ha->topology & QL_SNS_CONNECTION) {
6458 		/* Locate unused loop ID. */
6459 		if (CFG_IST(ha, CFG_CTRL_24258081)) {
6460 			first_loop_id = 0;
6461 			last_loop_id = LAST_N_PORT_HDL;
6462 		} else if (ha->topology & QL_F_PORT) {
6463 			first_loop_id = 0;
6464 			last_loop_id = SNS_LAST_LOOP_ID;
6465 		} else {
6466 			first_loop_id = SNS_FIRST_LOOP_ID;
6467 			last_loop_id = SNS_LAST_LOOP_ID;
6468 		}
6469 
6470 		/* Acquire adapter state lock. */
6471 		ADAPTER_STATE_LOCK(ha);
6472 
6473 		tq = ql_dev_init(ha, d_id, PORT_NO_LOOP_ID);
6474 		if (tq == NULL) {
6475 			EL(ha, "failed=%xh, d_id=%xh\n", QL_FUNCTION_FAILED,
6476 			    d_id.b24);
6477 
6478 			ADAPTER_STATE_UNLOCK(ha);
6479 
6480 			return (QL_FUNCTION_FAILED);
6481 		}
6482 
6483 		rval = QL_FUNCTION_FAILED;
6484 		loop_id = ha->pha->free_loop_id++;
6485 		for (index = (uint16_t)(last_loop_id - first_loop_id); index;
6486 		    index--) {
6487 			if (loop_id < first_loop_id ||
6488 			    loop_id > last_loop_id) {
6489 				loop_id = first_loop_id;
6490 				ha->pha->free_loop_id = (uint16_t)
6491 				    (loop_id + 1);
6492 			}
6493 
6494 			/* Bypass if loop ID used. */
6495 			for (vha = ha->pha; vha != NULL; vha = vha->vp_next) {
6496 				tq2 = ql_loop_id_to_queue(vha, loop_id);
6497 				if (tq2 != NULL && tq2 != tq) {
6498 					break;
6499 				}
6500 			}
6501 			if (vha != NULL || RESERVED_LOOP_ID(ha, loop_id) ||
6502 			    loop_id == ha->loop_id) {
6503 				loop_id = ha->pha->free_loop_id++;
6504 				continue;
6505 			}
6506 
6507 			ADAPTER_STATE_UNLOCK(ha);
6508 			rval = ql_login_fabric_port(ha, tq, loop_id);
6509 
6510 			/*
6511 			 * If PORT_ID_USED is returned
6512 			 * the login_fabric_port() updates
6513 			 * with the correct loop ID
6514 			 */
6515 			switch (rval) {
6516 			case QL_PORT_ID_USED:
6517 				/*
6518 				 * use f/w handle and try to
6519 				 * login again.
6520 				 */
6521 				ADAPTER_STATE_LOCK(ha);
6522 				ha->pha->free_loop_id--;
6523 				ADAPTER_STATE_UNLOCK(ha);
6524 				loop_id = tq->loop_id;
6525 				break;
6526 
6527 			case QL_SUCCESS:
6528 				tq->flags |= TQF_FABRIC_DEVICE;
6529 				(void) ql_get_port_database(ha,
6530 				    tq, PDF_NONE);
6531 				index = 1;
6532 				break;
6533 
6534 			case QL_LOOP_ID_USED:
6535 				tq->loop_id = PORT_NO_LOOP_ID;
6536 				loop_id = ha->pha->free_loop_id++;
6537 				break;
6538 
6539 			case QL_ALL_IDS_IN_USE:
6540 				tq->loop_id = PORT_NO_LOOP_ID;
6541 				index = 1;
6542 				break;
6543 
6544 			default:
6545 				tq->loop_id = PORT_NO_LOOP_ID;
6546 				index = 1;
6547 				break;
6548 			}
6549 
6550 			ADAPTER_STATE_LOCK(ha);
6551 		}
6552 
6553 		ADAPTER_STATE_UNLOCK(ha);
6554 	} else {
6555 		rval = QL_FUNCTION_FAILED;
6556 	}
6557 
6558 	if (rval != QL_SUCCESS) {
6559 		EL(ha, "failed=%xh, d_id=%xh\n", rval, d_id.b24);
6560 	} else {
6561 		EL(ha, "d_id=%xh, loop_id=%xh, "
6562 		    "wwpn=%02x%02x%02x%02x%02x%02x%02x%02xh\n", tq->d_id.b24,
6563 		    tq->loop_id, tq->port_name[0], tq->port_name[1],
6564 		    tq->port_name[2], tq->port_name[3], tq->port_name[4],
6565 		    tq->port_name[5], tq->port_name[6], tq->port_name[7]);
6566 	}
6567 	return (rval);
6568 }
6569 
6570 /*
6571  * ql_login_fabric_port
6572  *	Issue login fabric port mailbox command.
6573  *
6574  * Input:
6575  *	ha:		adapter state pointer.
6576  *	tq:		target queue pointer.
6577  *	loop_id:	FC Loop ID.
6578  *
6579  * Returns:
6580  *	ql local function return status code.
6581  *
6582  * Context:
6583  *	Kernel context.
6584  */
6585 static int
ql_login_fabric_port(ql_adapter_state_t * ha,ql_tgt_t * tq,uint16_t loop_id)6586 ql_login_fabric_port(ql_adapter_state_t *ha, ql_tgt_t *tq, uint16_t loop_id)
6587 {
6588 	int		rval;
6589 	int		index;
6590 	int		retry = 0;
6591 	port_id_t	d_id;
6592 	ql_tgt_t	*newq;
6593 	ql_mbx_data_t	mr;
6594 
6595 	QL_PRINT_3(CE_CONT, "(%d): started, d_id=%xh\n", ha->instance,
6596 	    tq->d_id.b24);
6597 
6598 	/*
6599 	 * QL_PARAMETER_ERROR also means the firmware is
6600 	 * not able to allocate PCB entry due to resource
6601 	 * issues, or collision.
6602 	 */
6603 	do {
6604 		rval = ql_login_fport(ha, tq, loop_id, LFF_NONE, &mr);
6605 		if ((rval == QL_PARAMETER_ERROR) ||
6606 		    ((rval == QL_COMMAND_ERROR) && (mr.mb[1] == 2 ||
6607 		    mr.mb[1] == 3 || mr.mb[1] == 7 || mr.mb[1] == 0xd))) {
6608 			retry++;
6609 			drv_usecwait(10 * MILLISEC);
6610 		} else {
6611 			break;
6612 		}
6613 	} while (retry < 5);
6614 
6615 	switch (rval) {
6616 	case QL_SUCCESS:
6617 		tq->loop_id = loop_id;
6618 		break;
6619 
6620 	case QL_PORT_ID_USED:
6621 		/*
6622 		 * This Loop ID should NOT be in use in drivers
6623 		 */
6624 		newq = ql_loop_id_to_queue(ha, mr.mb[1]);
6625 
6626 		if (newq != NULL && newq != tq && tq->logout_sent == 0) {
6627 			cmn_err(CE_WARN, "ql_login_fabric_port(%d): logout of "
6628 			    "dup loop_id=%xh, d_id=%xh", ha->instance,
6629 			    newq->loop_id, newq->d_id.b24);
6630 			ql_send_logo(ha, newq, NULL);
6631 		}
6632 
6633 		tq->loop_id = mr.mb[1];
6634 		break;
6635 
6636 	case QL_LOOP_ID_USED:
6637 		d_id.b.al_pa = LSB(mr.mb[2]);
6638 		d_id.b.area = MSB(mr.mb[2]);
6639 		d_id.b.domain = LSB(mr.mb[1]);
6640 
6641 		newq = ql_d_id_to_queue(ha, d_id);
6642 		if (newq && (newq->loop_id != loop_id)) {
6643 			/*
6644 			 * This should NEVER ever happen; but this
6645 			 * code is needed to bail out when the worst
6646 			 * case happens - or as used to happen before
6647 			 */
6648 			QL_PRINT_2(CE_CONT, "(%d,%d): Loop ID is now "
6649 			    "reassigned; old pairs: [%xh, %xh] and [%xh, %xh];"
6650 			    "new pairs: [%xh, unknown] and [%xh, %xh]\n",
6651 			    ha->instance, ha->vp_index, tq->d_id.b24, loop_id,
6652 			    newq->d_id.b24, newq->loop_id, tq->d_id.b24,
6653 			    newq->d_id.b24, loop_id);
6654 
6655 			if ((newq->d_id.b24 & 0xff) != (d_id.b24 & 0xff)) {
6656 				ADAPTER_STATE_LOCK(ha);
6657 
6658 				index = ql_alpa_to_index[newq->d_id.b.al_pa];
6659 				ql_add_link_b(&ha->dev[index], &newq->device);
6660 
6661 				newq->d_id.b24 = d_id.b24;
6662 
6663 				index = ql_alpa_to_index[d_id.b.al_pa];
6664 				ql_add_link_b(&ha->dev[index], &newq->device);
6665 
6666 				ADAPTER_STATE_UNLOCK(ha);
6667 			}
6668 
6669 			(void) ql_get_port_database(ha, newq, PDF_NONE);
6670 
6671 		}
6672 
6673 		/*
6674 		 * Invalidate the loop ID for the
6675 		 * us to obtain a new one.
6676 		 */
6677 		tq->loop_id = PORT_NO_LOOP_ID;
6678 		break;
6679 
6680 	case QL_ALL_IDS_IN_USE:
6681 		rval = QL_FUNCTION_FAILED;
6682 		EL(ha, "no loop id's available\n");
6683 		break;
6684 
6685 	default:
6686 		if (rval == QL_COMMAND_ERROR) {
6687 			switch (mr.mb[1]) {
6688 			case 2:
6689 			case 3:
6690 				rval = QL_MEMORY_ALLOC_FAILED;
6691 				break;
6692 
6693 			case 4:
6694 				rval = QL_FUNCTION_TIMEOUT;
6695 				break;
6696 			case 7:
6697 				rval = QL_FABRIC_NOT_INITIALIZED;
6698 				break;
6699 			default:
6700 				EL(ha, "cmd rtn; mb1=%xh\n", mr.mb[1]);
6701 				break;
6702 			}
6703 		} else {
6704 			cmn_err(CE_WARN, "%s(%d): login fabric port failed"
6705 			    " D_ID=%xh, rval=%xh, mb1=%xh", QL_NAME,
6706 			    ha->instance, tq->d_id.b24, rval, mr.mb[1]);
6707 		}
6708 		break;
6709 	}
6710 
6711 	if (rval != QL_SUCCESS && rval != QL_PORT_ID_USED &&
6712 	    rval != QL_LOOP_ID_USED) {
6713 		EL(ha, "failed=%xh\n", rval);
6714 	} else {
6715 		/*EMPTY*/
6716 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
6717 	}
6718 	return (rval);
6719 }
6720 
6721 /*
6722  * ql_logout_port
6723  *	Logs out a device if possible.
6724  *
6725  * Input:
6726  *	ha:	adapter state pointer.
6727  *	d_id:	24 bit port ID.
6728  *
6729  * Returns:
6730  *	QL local function return status code.
6731  *
6732  * Context:
6733  *	Kernel context.
6734  */
6735 static int
ql_logout_port(ql_adapter_state_t * ha,port_id_t d_id)6736 ql_logout_port(ql_adapter_state_t *ha, port_id_t d_id)
6737 {
6738 	ql_link_t	*link;
6739 	ql_tgt_t	*tq;
6740 	uint16_t	index;
6741 
6742 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
6743 
6744 	/* Get head queue index. */
6745 	index = ql_alpa_to_index[d_id.b.al_pa];
6746 
6747 	/* Get device queue. */
6748 	tq = NULL;
6749 	for (link = ha->dev[index].first; link != NULL; link = link->next) {
6750 		tq = link->base_address;
6751 		if (tq->d_id.b24 == d_id.b24) {
6752 			break;
6753 		} else {
6754 			tq = NULL;
6755 		}
6756 	}
6757 
6758 	if (tq != NULL && tq->flags & TQF_FABRIC_DEVICE) {
6759 		(void) ql_logout_fabric_port(ha, tq);
6760 		tq->loop_id = PORT_NO_LOOP_ID;
6761 	}
6762 
6763 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
6764 
6765 	return (QL_SUCCESS);
6766 }
6767 
6768 /*
6769  * ql_dev_init
6770  *	Initialize/allocate device queue.
6771  *
6772  * Input:
6773  *	ha:		adapter state pointer.
6774  *	d_id:		device destination ID
6775  *	loop_id:	device loop ID
6776  *	ADAPTER_STATE_LOCK must be already obtained.
6777  *
6778  * Returns:
6779  *	NULL = failure
6780  *
6781  * Context:
6782  *	Kernel context.
6783  */
6784 ql_tgt_t *
ql_dev_init(ql_adapter_state_t * ha,port_id_t d_id,uint16_t loop_id)6785 ql_dev_init(ql_adapter_state_t *ha, port_id_t d_id, uint16_t loop_id)
6786 {
6787 	ql_link_t	*link;
6788 	uint16_t	index;
6789 	ql_tgt_t	*tq;
6790 
6791 	QL_PRINT_3(CE_CONT, "(%d): started, d_id=%xh, loop_id=%xh\n",
6792 	    ha->instance, d_id.b24, loop_id);
6793 
6794 	index = ql_alpa_to_index[d_id.b.al_pa];
6795 
6796 	/* If device queue exists, set proper loop ID. */
6797 	tq = NULL;
6798 	for (link = ha->dev[index].first; link != NULL; link = link->next) {
6799 		tq = link->base_address;
6800 		if (tq->d_id.b24 == d_id.b24) {
6801 			tq->loop_id = loop_id;
6802 
6803 			/* Reset port down retry count. */
6804 			tq->port_down_retry_count = ha->port_down_retry_count;
6805 			tq->qfull_retry_count = ha->qfull_retry_count;
6806 
6807 			break;
6808 		} else {
6809 			tq = NULL;
6810 		}
6811 	}
6812 
6813 	/* If device does not have queue. */
6814 	if (tq == NULL) {
6815 		tq = (ql_tgt_t *)kmem_zalloc(sizeof (ql_tgt_t), KM_SLEEP);
6816 		if (tq != NULL) {
6817 			/*
6818 			 * mutex to protect the device queue,
6819 			 * does not block interrupts.
6820 			 */
6821 			mutex_init(&tq->mutex, NULL, MUTEX_DRIVER,
6822 			    (ha->iflags & IFLG_INTR_AIF) ?
6823 			    (void *)(uintptr_t)ha->intr_pri :
6824 			    (void *)(uintptr_t)ha->iblock_cookie);
6825 
6826 			tq->d_id.b24 = d_id.b24;
6827 			tq->loop_id = loop_id;
6828 			tq->device.base_address = tq;
6829 			tq->iidma_rate = IIDMA_RATE_INIT;
6830 
6831 			/* Reset port down retry count. */
6832 			tq->port_down_retry_count = ha->port_down_retry_count;
6833 			tq->qfull_retry_count = ha->qfull_retry_count;
6834 
6835 			/* Add device to device queue. */
6836 			ql_add_link_b(&ha->dev[index], &tq->device);
6837 		}
6838 	}
6839 
6840 	if (tq == NULL) {
6841 		EL(ha, "failed, d_id=%xh, loop_id=%xh\n", d_id.b24, loop_id);
6842 	} else {
6843 		/*EMPTY*/
6844 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
6845 	}
6846 	return (tq);
6847 }
6848 
6849 /*
6850  * ql_dev_free
6851  *	Remove queue from device list and frees resources used by queue.
6852  *
6853  * Input:
6854  *	ha:	adapter state pointer.
6855  *	tq:	target queue pointer.
6856  *	ADAPTER_STATE_LOCK must be already obtained.
6857  *
6858  * Context:
6859  *	Kernel context.
6860  */
6861 void
ql_dev_free(ql_adapter_state_t * ha,ql_tgt_t * tq)6862 ql_dev_free(ql_adapter_state_t *ha, ql_tgt_t *tq)
6863 {
6864 	ql_link_t	*link;
6865 	uint16_t	index;
6866 	ql_lun_t	*lq;
6867 
6868 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
6869 
6870 	for (link = tq->lun_queues.first; link != NULL; link = link->next) {
6871 		lq = link->base_address;
6872 		if (lq->cmd.first != NULL) {
6873 			return;
6874 		}
6875 	}
6876 
6877 	if (tq->outcnt == 0) {
6878 		/* Get head queue index. */
6879 		index = ql_alpa_to_index[tq->d_id.b.al_pa];
6880 		for (link = ha->dev[index].first; link != NULL;
6881 		    link = link->next) {
6882 			if (link->base_address == tq) {
6883 				ql_remove_link(&ha->dev[index], link);
6884 
6885 				link = tq->lun_queues.first;
6886 				while (link != NULL) {
6887 					lq = link->base_address;
6888 					link = link->next;
6889 
6890 					ql_remove_link(&tq->lun_queues,
6891 					    &lq->link);
6892 					kmem_free(lq, sizeof (ql_lun_t));
6893 				}
6894 
6895 				mutex_destroy(&tq->mutex);
6896 				kmem_free(tq, sizeof (ql_tgt_t));
6897 				break;
6898 			}
6899 		}
6900 	}
6901 
6902 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
6903 }
6904 
6905 /*
6906  * ql_lun_queue
6907  *	Allocate LUN queue if does not exists.
6908  *
6909  * Input:
6910  *	ha:	adapter state pointer.
6911  *	tq:	target queue.
6912  *	lun:	LUN number.
6913  *
6914  * Returns:
6915  *	NULL = failure
6916  *
6917  * Context:
6918  *	Kernel context.
6919  */
6920 static ql_lun_t *
ql_lun_queue(ql_adapter_state_t * ha,ql_tgt_t * tq,uint16_t lun)6921 ql_lun_queue(ql_adapter_state_t *ha, ql_tgt_t *tq, uint16_t lun)
6922 {
6923 	ql_lun_t	*lq;
6924 	ql_link_t	*link;
6925 
6926 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
6927 
6928 	/* Fast path. */
6929 	if (tq->last_lun_queue != NULL && tq->last_lun_queue->lun_no == lun) {
6930 		QL_PRINT_3(CE_CONT, "(%d): fast done\n", ha->instance);
6931 		return (tq->last_lun_queue);
6932 	}
6933 
6934 	if (lun >= MAX_LUNS) {
6935 		EL(ha, "Exceeded MAX_LUN=%d, lun=%d\n", MAX_LUNS, lun);
6936 		return (NULL);
6937 	}
6938 	/* If device queue exists, set proper loop ID. */
6939 	lq = NULL;
6940 	for (link = tq->lun_queues.first; link != NULL; link = link->next) {
6941 		lq = link->base_address;
6942 		if (lq->lun_no == lun) {
6943 			QL_PRINT_3(CE_CONT, "(%d): found done\n", ha->instance);
6944 			tq->last_lun_queue = lq;
6945 			return (lq);
6946 		}
6947 	}
6948 
6949 	/* If queue does exist. */
6950 	lq = (ql_lun_t *)kmem_zalloc(sizeof (ql_lun_t), KM_SLEEP);
6951 
6952 	/* Initialize LUN queue. */
6953 	if (lq != NULL) {
6954 		lq->link.base_address = lq;
6955 
6956 		lq->lun_no = lun;
6957 		lq->target_queue = tq;
6958 
6959 		DEVICE_QUEUE_LOCK(tq);
6960 		ql_add_link_b(&tq->lun_queues, &lq->link);
6961 		DEVICE_QUEUE_UNLOCK(tq);
6962 		tq->last_lun_queue = lq;
6963 	}
6964 
6965 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
6966 
6967 	return (lq);
6968 }
6969 
6970 /*
6971  * ql_fcp_scsi_cmd
6972  *	Process fibre channel (FCP) SCSI protocol commands.
6973  *
6974  * Input:
6975  *	ha = adapter state pointer.
6976  *	pkt = pointer to fc_packet.
6977  *	sp = srb pointer.
6978  *
6979  * Returns:
6980  *	FC_SUCCESS - the packet was accepted for transport.
6981  *	FC_TRANSPORT_ERROR - a transport error occurred.
6982  *
6983  * Context:
6984  *	Kernel context.
6985  */
6986 static int
ql_fcp_scsi_cmd(ql_adapter_state_t * ha,fc_packet_t * pkt,ql_srb_t * sp)6987 ql_fcp_scsi_cmd(ql_adapter_state_t *ha, fc_packet_t *pkt, ql_srb_t *sp)
6988 {
6989 	port_id_t	d_id;
6990 	ql_tgt_t	*tq;
6991 	uint64_t	*ptr;
6992 	uint16_t	lun;
6993 
6994 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
6995 
6996 	tq = (ql_tgt_t *)pkt->pkt_fca_device;
6997 	if (tq == NULL) {
6998 		d_id.r.rsvd_1 = 0;
6999 		d_id.b24 = pkt->pkt_cmd_fhdr.d_id;
7000 		tq = ql_d_id_to_queue(ha, d_id);
7001 	}
7002 
7003 	sp->fcp = (struct fcp_cmd *)pkt->pkt_cmd;
7004 	lun = CHAR_TO_SHORT(lobyte(sp->fcp->fcp_ent_addr.ent_addr_0),
7005 	    hibyte(sp->fcp->fcp_ent_addr.ent_addr_0));
7006 
7007 	if (tq != NULL &&
7008 	    (sp->lun_queue = ql_lun_queue(ha, tq, lun)) != NULL) {
7009 
7010 		/*
7011 		 * zero out FCP response; 24 Bytes
7012 		 */
7013 		ptr = (uint64_t *)pkt->pkt_resp;
7014 		*ptr++ = 0; *ptr++ = 0; *ptr++ = 0;
7015 
7016 		/* Handle task management function. */
7017 		if ((sp->fcp->fcp_cntl.cntl_kill_tsk |
7018 		    sp->fcp->fcp_cntl.cntl_clr_aca |
7019 		    sp->fcp->fcp_cntl.cntl_reset_tgt |
7020 		    sp->fcp->fcp_cntl.cntl_reset_lun |
7021 		    sp->fcp->fcp_cntl.cntl_clr_tsk |
7022 		    sp->fcp->fcp_cntl.cntl_abort_tsk) != 0) {
7023 			ql_task_mgmt(ha, tq, pkt, sp);
7024 		} else {
7025 			ha->pha->xioctl->IosRequested++;
7026 			ha->pha->xioctl->BytesRequested += (uint32_t)
7027 			    sp->fcp->fcp_data_len;
7028 
7029 			/*
7030 			 * Setup for commands with data transfer
7031 			 */
7032 			sp->iocb = ha->fcp_cmd;
7033 			sp->req_cnt = 1;
7034 			if (sp->fcp->fcp_data_len != 0) {
7035 				/*
7036 				 * FCP data is bound to pkt_data_dma
7037 				 */
7038 				if (sp->fcp->fcp_cntl.cntl_write_data) {
7039 					(void) ddi_dma_sync(pkt->pkt_data_dma,
7040 					    0, 0, DDI_DMA_SYNC_FORDEV);
7041 				}
7042 
7043 				/* Setup IOCB count. */
7044 				if (pkt->pkt_data_cookie_cnt > ha->cmd_segs &&
7045 				    (!CFG_IST(ha, CFG_CTRL_8021) ||
7046 				    sp->sg_dma.dma_handle == NULL)) {
7047 					uint32_t	cnt;
7048 
7049 					cnt = pkt->pkt_data_cookie_cnt -
7050 					    ha->cmd_segs;
7051 					sp->req_cnt = (uint16_t)
7052 					    (cnt / ha->cmd_cont_segs);
7053 					if (cnt % ha->cmd_cont_segs) {
7054 						sp->req_cnt = (uint16_t)
7055 						    (sp->req_cnt + 2);
7056 					} else {
7057 						sp->req_cnt++;
7058 					}
7059 				}
7060 			}
7061 			QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
7062 
7063 			return (ql_start_cmd(ha, tq, pkt, sp));
7064 		}
7065 	} else {
7066 		pkt->pkt_state = FC_PKT_LOCAL_RJT;
7067 		pkt->pkt_reason = FC_REASON_NO_CONNECTION;
7068 
7069 		if (!(pkt->pkt_tran_flags & FC_TRAN_NO_INTR) && pkt->pkt_comp)
7070 			ql_awaken_task_daemon(ha, sp, 0, 0);
7071 	}
7072 
7073 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
7074 
7075 	return (FC_SUCCESS);
7076 }
7077 
7078 /*
7079  * ql_task_mgmt
7080  *	Task management function processor.
7081  *
7082  * Input:
7083  *	ha:	adapter state pointer.
7084  *	tq:	target queue pointer.
7085  *	pkt:	pointer to fc_packet.
7086  *	sp:	SRB pointer.
7087  *
7088  * Context:
7089  *	Kernel context.
7090  */
7091 static void
ql_task_mgmt(ql_adapter_state_t * ha,ql_tgt_t * tq,fc_packet_t * pkt,ql_srb_t * sp)7092 ql_task_mgmt(ql_adapter_state_t *ha, ql_tgt_t *tq, fc_packet_t *pkt,
7093     ql_srb_t *sp)
7094 {
7095 	fcp_rsp_t		*fcpr;
7096 	struct fcp_rsp_info	*rsp;
7097 	uint16_t		lun;
7098 
7099 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
7100 
7101 	fcpr = (fcp_rsp_t *)pkt->pkt_resp;
7102 	rsp = (struct fcp_rsp_info *)pkt->pkt_resp + sizeof (fcp_rsp_t);
7103 
7104 	bzero(fcpr, pkt->pkt_rsplen);
7105 
7106 	fcpr->fcp_u.fcp_status.rsp_len_set = 1;
7107 	fcpr->fcp_response_len = 8;
7108 	lun = CHAR_TO_SHORT(lobyte(sp->fcp->fcp_ent_addr.ent_addr_0),
7109 	    hibyte(sp->fcp->fcp_ent_addr.ent_addr_0));
7110 
7111 	if (sp->fcp->fcp_cntl.cntl_clr_aca) {
7112 		if (ql_clear_aca(ha, tq, lun) != QL_SUCCESS) {
7113 			rsp->rsp_code = FCP_TASK_MGMT_FAILED;
7114 		}
7115 	} else if (sp->fcp->fcp_cntl.cntl_reset_lun) {
7116 		if (ql_lun_reset(ha, tq, lun) != QL_SUCCESS) {
7117 			rsp->rsp_code = FCP_TASK_MGMT_FAILED;
7118 		}
7119 	} else if (sp->fcp->fcp_cntl.cntl_reset_tgt) {
7120 		if (ql_target_reset(ha, tq, ha->loop_reset_delay) !=
7121 		    QL_SUCCESS) {
7122 			rsp->rsp_code = FCP_TASK_MGMT_FAILED;
7123 		}
7124 	} else if (sp->fcp->fcp_cntl.cntl_clr_tsk) {
7125 		if (ql_clear_task_set(ha, tq, lun) != QL_SUCCESS) {
7126 			rsp->rsp_code = FCP_TASK_MGMT_FAILED;
7127 		}
7128 	} else if (sp->fcp->fcp_cntl.cntl_abort_tsk) {
7129 		if (ql_abort_task_set(ha, tq, lun) != QL_SUCCESS) {
7130 			rsp->rsp_code = FCP_TASK_MGMT_FAILED;
7131 		}
7132 	} else {
7133 		rsp->rsp_code = FCP_TASK_MGMT_NOT_SUPPTD;
7134 	}
7135 
7136 	pkt->pkt_state = FC_PKT_SUCCESS;
7137 
7138 	/* Do command callback. */
7139 	if (!(pkt->pkt_tran_flags & FC_TRAN_NO_INTR) && pkt->pkt_comp) {
7140 		ql_awaken_task_daemon(ha, sp, 0, 0);
7141 	}
7142 
7143 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
7144 }
7145 
7146 /*
7147  * ql_fcp_ip_cmd
7148  *	Process fibre channel (FCP) Internet (IP) protocols commands.
7149  *
7150  * Input:
7151  *	ha:	adapter state pointer.
7152  *	pkt:	pointer to fc_packet.
7153  *	sp:	SRB pointer.
7154  *
7155  * Returns:
7156  *	FC_SUCCESS - the packet was accepted for transport.
7157  *	FC_TRANSPORT_ERROR - a transport error occurred.
7158  *
7159  * Context:
7160  *	Kernel context.
7161  */
7162 static int
ql_fcp_ip_cmd(ql_adapter_state_t * ha,fc_packet_t * pkt,ql_srb_t * sp)7163 ql_fcp_ip_cmd(ql_adapter_state_t *ha, fc_packet_t *pkt, ql_srb_t *sp)
7164 {
7165 	port_id_t	d_id;
7166 	ql_tgt_t	*tq;
7167 
7168 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
7169 
7170 	tq = (ql_tgt_t *)pkt->pkt_fca_device;
7171 	if (tq == NULL) {
7172 		d_id.r.rsvd_1 = 0;
7173 		d_id.b24 = pkt->pkt_cmd_fhdr.d_id;
7174 		tq = ql_d_id_to_queue(ha, d_id);
7175 	}
7176 
7177 	if (tq != NULL && (sp->lun_queue = ql_lun_queue(ha, tq, 0)) != NULL) {
7178 		/*
7179 		 * IP data is bound to pkt_cmd_dma
7180 		 */
7181 		(void) ddi_dma_sync(pkt->pkt_cmd_dma,
7182 		    0, 0, DDI_DMA_SYNC_FORDEV);
7183 
7184 		/* Setup IOCB count. */
7185 		sp->iocb = ha->ip_cmd;
7186 		if (pkt->pkt_cmd_cookie_cnt > ha->cmd_segs) {
7187 			uint32_t	cnt;
7188 
7189 			cnt = pkt->pkt_cmd_cookie_cnt - ha->cmd_segs;
7190 			sp->req_cnt = (uint16_t)(cnt / ha->cmd_cont_segs);
7191 			if (cnt % ha->cmd_cont_segs) {
7192 				sp->req_cnt = (uint16_t)(sp->req_cnt + 2);
7193 			} else {
7194 				sp->req_cnt++;
7195 			}
7196 		} else {
7197 			sp->req_cnt = 1;
7198 		}
7199 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
7200 
7201 		return (ql_start_cmd(ha, tq, pkt, sp));
7202 	} else {
7203 		pkt->pkt_state = FC_PKT_LOCAL_RJT;
7204 		pkt->pkt_reason = FC_REASON_NO_CONNECTION;
7205 
7206 		if (!(pkt->pkt_tran_flags & FC_TRAN_NO_INTR) && pkt->pkt_comp)
7207 			ql_awaken_task_daemon(ha, sp, 0, 0);
7208 	}
7209 
7210 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
7211 
7212 	return (FC_SUCCESS);
7213 }
7214 
7215 /*
7216  * ql_fc_services
7217  *	Process fibre channel services (name server).
7218  *
7219  * Input:
7220  *	ha:	adapter state pointer.
7221  *	pkt:	pointer to fc_packet.
7222  *
7223  * Returns:
7224  *	FC_SUCCESS - the packet was accepted for transport.
7225  *	FC_TRANSPORT_ERROR - a transport error occurred.
7226  *
7227  * Context:
7228  *	Kernel context.
7229  */
7230 static int
ql_fc_services(ql_adapter_state_t * ha,fc_packet_t * pkt)7231 ql_fc_services(ql_adapter_state_t *ha, fc_packet_t *pkt)
7232 {
7233 	uint32_t	cnt;
7234 	fc_ct_header_t	hdr;
7235 	la_els_rjt_t	rjt;
7236 	port_id_t	d_id;
7237 	ql_tgt_t	*tq;
7238 	ql_srb_t	*sp;
7239 	int		rval;
7240 
7241 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
7242 
7243 	ddi_rep_get8(pkt->pkt_cmd_acc, (uint8_t *)&hdr,
7244 	    (uint8_t *)pkt->pkt_cmd, sizeof (hdr), DDI_DEV_AUTOINCR);
7245 
7246 	bzero(&rjt, sizeof (rjt));
7247 
7248 	/* Do some sanity checks */
7249 	cnt = (uint32_t)((uint32_t)(hdr.ct_aiusize * 4) +
7250 	    sizeof (fc_ct_header_t));
7251 	if (cnt > (uint32_t)pkt->pkt_rsplen) {
7252 		EL(ha, "FC_ELS_MALFORMED, cnt=%xh, size=%xh\n", cnt,
7253 		    pkt->pkt_rsplen);
7254 		return (FC_ELS_MALFORMED);
7255 	}
7256 
7257 	switch (hdr.ct_fcstype) {
7258 	case FCSTYPE_DIRECTORY:
7259 	case FCSTYPE_MGMTSERVICE:
7260 		/* An FCA must make sure that the header is in big endian */
7261 		ql_cthdr_endian(pkt->pkt_cmd_acc, pkt->pkt_cmd, B_FALSE);
7262 
7263 		d_id.b24 = pkt->pkt_cmd_fhdr.d_id;
7264 		tq = ql_d_id_to_queue(ha, d_id);
7265 		sp = (ql_srb_t *)pkt->pkt_fca_private;
7266 		if (tq == NULL ||
7267 		    (sp->lun_queue = ql_lun_queue(ha, tq, 0)) == NULL) {
7268 			pkt->pkt_state = FC_PKT_LOCAL_RJT;
7269 			pkt->pkt_reason = FC_REASON_NO_CONNECTION;
7270 			rval = QL_SUCCESS;
7271 			break;
7272 		}
7273 
7274 		/*
7275 		 * Services data is bound to pkt_cmd_dma
7276 		 */
7277 		(void) ddi_dma_sync(pkt->pkt_cmd_dma, 0, 0,
7278 		    DDI_DMA_SYNC_FORDEV);
7279 
7280 		sp->flags |= SRB_MS_PKT;
7281 		sp->retry_count = 32;
7282 
7283 		/* Setup IOCB count. */
7284 		sp->iocb = ha->ms_cmd;
7285 		if (pkt->pkt_resp_cookie_cnt > MS_DATA_SEGMENTS) {
7286 			cnt = pkt->pkt_resp_cookie_cnt - MS_DATA_SEGMENTS;
7287 			sp->req_cnt =
7288 			    (uint16_t)(cnt / CONT_TYPE_1_DATA_SEGMENTS);
7289 			if (cnt % CONT_TYPE_1_DATA_SEGMENTS) {
7290 				sp->req_cnt = (uint16_t)(sp->req_cnt + 2);
7291 			} else {
7292 				sp->req_cnt++;
7293 			}
7294 		} else {
7295 			sp->req_cnt = 1;
7296 		}
7297 		rval = ql_start_cmd(ha, tq, pkt, sp);
7298 
7299 		QL_PRINT_3(CE_CONT, "(%d): done, ql_start_cmd=%xh\n",
7300 		    ha->instance, rval);
7301 
7302 		return (rval);
7303 
7304 	default:
7305 		EL(ha, "unknown fcstype=%xh\n", hdr.ct_fcstype);
7306 		rval = QL_FUNCTION_PARAMETER_ERROR;
7307 		break;
7308 	}
7309 
7310 	if (rval != QL_SUCCESS) {
7311 		/* Build RJT. */
7312 		rjt.ls_code.ls_code = LA_ELS_RJT;
7313 		rjt.reason = FC_REASON_CMD_UNSUPPORTED;
7314 
7315 		ddi_rep_put8(pkt->pkt_resp_acc, (uint8_t *)&rjt,
7316 		    (uint8_t *)pkt->pkt_resp, sizeof (rjt), DDI_DEV_AUTOINCR);
7317 
7318 		pkt->pkt_state = FC_PKT_LOCAL_RJT;
7319 		pkt->pkt_reason = FC_REASON_UNSUPPORTED;
7320 		EL(ha, "LA_ELS_RJT, FC_REASON_UNSUPPORTED\n");
7321 	}
7322 
7323 	/* Do command callback. */
7324 	if (!(pkt->pkt_tran_flags & FC_TRAN_NO_INTR) && pkt->pkt_comp) {
7325 		ql_awaken_task_daemon(ha, (ql_srb_t *)pkt->pkt_fca_private,
7326 		    0, 0);
7327 	}
7328 
7329 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
7330 
7331 	return (FC_SUCCESS);
7332 }
7333 
7334 /*
7335  * ql_cthdr_endian
7336  *	Change endianess of ct passthrough header and payload.
7337  *
7338  * Input:
7339  *	acc_handle:	DMA buffer access handle.
7340  *	ct_hdr:		Pointer to header.
7341  *	restore:	Restore first flag.
7342  *
7343  * Context:
7344  *	Interrupt or Kernel context, no mailbox commands allowed.
7345  */
7346 void
ql_cthdr_endian(ddi_acc_handle_t acc_handle,caddr_t ct_hdr,boolean_t restore)7347 ql_cthdr_endian(ddi_acc_handle_t acc_handle, caddr_t ct_hdr,
7348     boolean_t restore)
7349 {
7350 	uint8_t		i, *bp;
7351 	fc_ct_header_t	hdr;
7352 	uint32_t	*hdrp = (uint32_t *)&hdr;
7353 
7354 	ddi_rep_get8(acc_handle, (uint8_t *)&hdr,
7355 	    (uint8_t *)ct_hdr, sizeof (hdr), DDI_DEV_AUTOINCR);
7356 
7357 	if (restore) {
7358 		for (i = 0; i < ((sizeof (hdr)) / (sizeof (uint32_t))); i++) {
7359 			*hdrp = BE_32(*hdrp);
7360 			hdrp++;
7361 		}
7362 	}
7363 
7364 	if (hdr.ct_fcstype == FCSTYPE_DIRECTORY) {
7365 		bp = (uint8_t *)ct_hdr + sizeof (fc_ct_header_t);
7366 
7367 		switch (hdr.ct_cmdrsp) {
7368 		case NS_GA_NXT:
7369 		case NS_GPN_ID:
7370 		case NS_GNN_ID:
7371 		case NS_GCS_ID:
7372 		case NS_GFT_ID:
7373 		case NS_GSPN_ID:
7374 		case NS_GPT_ID:
7375 		case NS_GID_FT:
7376 		case NS_GID_PT:
7377 		case NS_RPN_ID:
7378 		case NS_RNN_ID:
7379 		case NS_RSPN_ID:
7380 		case NS_DA_ID:
7381 			BIG_ENDIAN_32(bp);
7382 			break;
7383 		case NS_RFT_ID:
7384 		case NS_RCS_ID:
7385 		case NS_RPT_ID:
7386 			BIG_ENDIAN_32(bp);
7387 			bp += 4;
7388 			BIG_ENDIAN_32(bp);
7389 			break;
7390 		case NS_GNN_IP:
7391 		case NS_GIPA_IP:
7392 			BIG_ENDIAN(bp, 16);
7393 			break;
7394 		case NS_RIP_NN:
7395 			bp += 8;
7396 			BIG_ENDIAN(bp, 16);
7397 			break;
7398 		case NS_RIPA_NN:
7399 			bp += 8;
7400 			BIG_ENDIAN_64(bp);
7401 			break;
7402 		default:
7403 			break;
7404 		}
7405 	}
7406 
7407 	if (restore == B_FALSE) {
7408 		for (i = 0; i < ((sizeof (hdr)) / (sizeof (uint32_t))); i++) {
7409 			*hdrp = BE_32(*hdrp);
7410 			hdrp++;
7411 		}
7412 	}
7413 
7414 	ddi_rep_put8(acc_handle, (uint8_t *)&hdr,
7415 	    (uint8_t *)ct_hdr, sizeof (hdr), DDI_DEV_AUTOINCR);
7416 }
7417 
7418 /*
7419  * ql_start_cmd
7420  *	Finishes starting fibre channel protocol (FCP) command.
7421  *
7422  * Input:
7423  *	ha:	adapter state pointer.
7424  *	tq:	target queue pointer.
7425  *	pkt:	pointer to fc_packet.
7426  *	sp:	SRB pointer.
7427  *
7428  * Context:
7429  *	Kernel context.
7430  */
7431 static int
ql_start_cmd(ql_adapter_state_t * ha,ql_tgt_t * tq,fc_packet_t * pkt,ql_srb_t * sp)7432 ql_start_cmd(ql_adapter_state_t *ha, ql_tgt_t *tq, fc_packet_t *pkt,
7433     ql_srb_t *sp)
7434 {
7435 	int		rval = FC_SUCCESS;
7436 	time_t		poll_wait = 0;
7437 	ql_lun_t	*lq = sp->lun_queue;
7438 
7439 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
7440 
7441 	sp->handle = 0;
7442 
7443 	/* Set poll for finish. */
7444 	if (pkt->pkt_tran_flags & FC_TRAN_NO_INTR) {
7445 		sp->flags |= SRB_POLL;
7446 		if (pkt->pkt_timeout == 0) {
7447 			pkt->pkt_timeout = SCSI_POLL_TIMEOUT;
7448 		}
7449 	}
7450 
7451 	/* Acquire device queue lock. */
7452 	DEVICE_QUEUE_LOCK(tq);
7453 
7454 	/*
7455 	 * If we need authentication, report device busy to
7456 	 * upper layers to retry later
7457 	 */
7458 	if (tq->flags & (TQF_RSCN_RCVD | TQF_NEED_AUTHENTICATION)) {
7459 		DEVICE_QUEUE_UNLOCK(tq);
7460 		EL(ha, "failed, FC_DEVICE_BUSY=%xh, d_id=%xh\n", tq->flags,
7461 		    tq->d_id.b24);
7462 		return (FC_DEVICE_BUSY);
7463 	}
7464 
7465 	/* Insert command onto watchdog queue. */
7466 	if (!(pkt->pkt_tran_flags & FC_TRAN_DUMPING)) {
7467 		ql_timeout_insert(ha, tq, sp);
7468 	} else {
7469 		/*
7470 		 * Run dump requests in polled mode as kernel threads
7471 		 * and interrupts may have been disabled.
7472 		 */
7473 		sp->flags |= SRB_POLL;
7474 		sp->init_wdg_q_time = 0;
7475 		sp->isp_timeout = 0;
7476 	}
7477 
7478 	/* If a polling command setup wait time. */
7479 	if (sp->flags & SRB_POLL) {
7480 		if (sp->flags & SRB_WATCHDOG_ENABLED) {
7481 			poll_wait = (sp->wdg_q_time + 2) * WATCHDOG_TIME;
7482 		} else {
7483 			poll_wait = pkt->pkt_timeout;
7484 		}
7485 	}
7486 
7487 	if (ha->pha->flags & ABORT_CMDS_LOOP_DOWN_TMO &&
7488 	    (CFG_IST(ha, CFG_ENABLE_LINK_DOWN_REPORTING))) {
7489 		/* Set ending status. */
7490 		sp->pkt->pkt_reason = CS_PORT_UNAVAILABLE;
7491 
7492 		/* Call done routine to handle completions. */
7493 		sp->cmd.next = NULL;
7494 		DEVICE_QUEUE_UNLOCK(tq);
7495 		ql_done(&sp->cmd);
7496 	} else {
7497 		if (ddi_in_panic() && (sp->flags & SRB_POLL)) {
7498 			int do_lip = 0;
7499 
7500 			DEVICE_QUEUE_UNLOCK(tq);
7501 
7502 			ADAPTER_STATE_LOCK(ha);
7503 			if ((do_lip = ha->pha->lip_on_panic) == 0) {
7504 				ha->pha->lip_on_panic++;
7505 			}
7506 			ADAPTER_STATE_UNLOCK(ha);
7507 
7508 			if (!do_lip) {
7509 
7510 				/*
7511 				 * That Qlogic F/W performs PLOGI, PRLI, etc
7512 				 * is helpful here. If a PLOGI fails for some
7513 				 * reason, you would get CS_PORT_LOGGED_OUT
7514 				 * or some such error; and we should get a
7515 				 * careful polled mode login kicked off inside
7516 				 * of this driver itself. You don't have FC
7517 				 * transport's services as all threads are
7518 				 * suspended, interrupts disabled, and so
7519 				 * on. Right now we do re-login if the packet
7520 				 * state isn't FC_PKT_SUCCESS.
7521 				 */
7522 				(void) ql_abort_isp(ha);
7523 			}
7524 
7525 			ql_start_iocb(ha, sp);
7526 		} else {
7527 			/* Add the command to the device queue */
7528 			if (pkt->pkt_tran_flags & FC_TRAN_HI_PRIORITY) {
7529 				ql_add_link_t(&lq->cmd, &sp->cmd);
7530 			} else {
7531 				ql_add_link_b(&lq->cmd, &sp->cmd);
7532 			}
7533 
7534 			sp->flags |= SRB_IN_DEVICE_QUEUE;
7535 
7536 			/* Check whether next message can be processed */
7537 			ql_next(ha, lq);
7538 		}
7539 	}
7540 
7541 	/* If polling, wait for finish. */
7542 	if (poll_wait) {
7543 		if (ql_poll_cmd(ha, sp, poll_wait) != QL_SUCCESS) {
7544 			int	res;
7545 
7546 			res = ql_abort((opaque_t)ha, pkt, 0);
7547 			if (res != FC_SUCCESS && res != FC_ABORTED) {
7548 				DEVICE_QUEUE_LOCK(tq);
7549 				ql_remove_link(&lq->cmd, &sp->cmd);
7550 				sp->flags &= ~SRB_IN_DEVICE_QUEUE;
7551 				DEVICE_QUEUE_UNLOCK(tq);
7552 			}
7553 		}
7554 
7555 		if (pkt->pkt_state != FC_PKT_SUCCESS) {
7556 			EL(ha, "failed, FC_TRANSPORT_ERROR\n");
7557 			rval = FC_TRANSPORT_ERROR;
7558 		}
7559 
7560 		if (ddi_in_panic()) {
7561 			if (pkt->pkt_state != FC_PKT_SUCCESS) {
7562 				port_id_t d_id;
7563 
7564 				/*
7565 				 * successful LOGIN implies by design
7566 				 * that PRLI also succeeded for disks
7567 				 * Note also that there is no special
7568 				 * mailbox command to send PRLI.
7569 				 */
7570 				d_id.b24 = pkt->pkt_cmd_fhdr.d_id;
7571 				(void) ql_login_port(ha, d_id);
7572 			}
7573 		}
7574 
7575 		/*
7576 		 * This should only happen during CPR dumping
7577 		 */
7578 		if (!(pkt->pkt_tran_flags & FC_TRAN_NO_INTR) &&
7579 		    pkt->pkt_comp) {
7580 			sp->flags &= ~SRB_POLL;
7581 			(*pkt->pkt_comp)(pkt);
7582 		}
7583 	}
7584 
7585 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
7586 
7587 	return (rval);
7588 }
7589 
7590 /*
7591  * ql_poll_cmd
7592  *	Polls commands for completion.
7593  *
7594  * Input:
7595  *	ha = adapter state pointer.
7596  *	sp = SRB command pointer.
7597  *	poll_wait = poll wait time in seconds.
7598  *
7599  * Returns:
7600  *	QL local function return status code.
7601  *
7602  * Context:
7603  *	Kernel context.
7604  */
7605 static int
ql_poll_cmd(ql_adapter_state_t * vha,ql_srb_t * sp,time_t poll_wait)7606 ql_poll_cmd(ql_adapter_state_t *vha, ql_srb_t *sp, time_t poll_wait)
7607 {
7608 	int			rval = QL_SUCCESS;
7609 	time_t			msecs_left = poll_wait * 100;	/* 10ms inc */
7610 	ql_adapter_state_t	*ha = vha->pha;
7611 
7612 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
7613 
7614 	while (sp->flags & SRB_POLL) {
7615 
7616 		if ((ha->flags & INTERRUPTS_ENABLED) == 0 ||
7617 		    ha->idle_timer >= 15 || ddi_in_panic()) {
7618 
7619 			/* If waiting for restart, do it now. */
7620 			if (ha->port_retry_timer != 0) {
7621 				ADAPTER_STATE_LOCK(ha);
7622 				ha->port_retry_timer = 0;
7623 				ADAPTER_STATE_UNLOCK(ha);
7624 
7625 				TASK_DAEMON_LOCK(ha);
7626 				ha->task_daemon_flags |= PORT_RETRY_NEEDED;
7627 				TASK_DAEMON_UNLOCK(ha);
7628 			}
7629 
7630 			if (INTERRUPT_PENDING(ha)) {
7631 				(void) ql_isr((caddr_t)ha);
7632 				INTR_LOCK(ha);
7633 				ha->intr_claimed = TRUE;
7634 				INTR_UNLOCK(ha);
7635 			}
7636 
7637 			/*
7638 			 * Call task thread function in case the
7639 			 * daemon is not running.
7640 			 */
7641 			TASK_DAEMON_LOCK(ha);
7642 
7643 			if (!ddi_in_panic() && QL_DAEMON_NOT_ACTIVE(ha) &&
7644 			    QL_TASK_PENDING(ha)) {
7645 				ha->task_daemon_flags |= TASK_THREAD_CALLED;
7646 				ql_task_thread(ha);
7647 				ha->task_daemon_flags &= ~TASK_THREAD_CALLED;
7648 			}
7649 
7650 			TASK_DAEMON_UNLOCK(ha);
7651 		}
7652 
7653 		if (msecs_left < 10) {
7654 			rval = QL_FUNCTION_TIMEOUT;
7655 			break;
7656 		}
7657 
7658 		/*
7659 		 * Polling interval is 10 milli seconds; Increasing
7660 		 * the polling interval to seconds since disk IO
7661 		 * timeout values are ~60 seconds is tempting enough,
7662 		 * but CPR dump time increases, and so will the crash
7663 		 * dump time; Don't toy with the settings without due
7664 		 * consideration for all the scenarios that will be
7665 		 * impacted.
7666 		 */
7667 		ql_delay(ha, 10000);
7668 		msecs_left -= 10;
7669 	}
7670 
7671 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
7672 
7673 	return (rval);
7674 }
7675 
7676 /*
7677  * ql_next
7678  *	Retrieve and process next job in the device queue.
7679  *
7680  * Input:
7681  *	ha:	adapter state pointer.
7682  *	lq:	LUN queue pointer.
7683  *	DEVICE_QUEUE_LOCK must be already obtained.
7684  *
7685  * Output:
7686  *	Releases DEVICE_QUEUE_LOCK upon exit.
7687  *
7688  * Context:
7689  *	Interrupt or Kernel context, no mailbox commands allowed.
7690  */
7691 void
ql_next(ql_adapter_state_t * vha,ql_lun_t * lq)7692 ql_next(ql_adapter_state_t *vha, ql_lun_t *lq)
7693 {
7694 	ql_srb_t		*sp;
7695 	ql_link_t		*link;
7696 	ql_tgt_t		*tq = lq->target_queue;
7697 	ql_adapter_state_t	*ha = vha->pha;
7698 
7699 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
7700 
7701 	if (ddi_in_panic()) {
7702 		DEVICE_QUEUE_UNLOCK(tq);
7703 		QL_PRINT_3(CE_CONT, "(%d): panic/active exit\n",
7704 		    ha->instance);
7705 		return;
7706 	}
7707 
7708 	while ((link = lq->cmd.first) != NULL) {
7709 		sp = link->base_address;
7710 
7711 		/* Exit if can not start commands. */
7712 		if (DRIVER_SUSPENDED(ha) ||
7713 		    (ha->flags & ONLINE) == 0 ||
7714 		    !VALID_DEVICE_ID(ha, tq->loop_id) ||
7715 		    sp->flags & SRB_ABORT ||
7716 		    tq->flags & (TQF_RSCN_RCVD | TQF_NEED_AUTHENTICATION |
7717 		    TQF_QUEUE_SUSPENDED)) {
7718 			EL(vha, "break, d_id=%xh, tdf=%xh, tqf=%xh, spf=%xh, "
7719 			    "haf=%xh, loop_id=%xh\n", tq->d_id.b24,
7720 			    ha->task_daemon_flags, tq->flags, sp->flags,
7721 			    ha->flags, tq->loop_id);
7722 			break;
7723 		}
7724 
7725 		/*
7726 		 * Find out the LUN number for untagged command use.
7727 		 * If there is an untagged command pending for the LUN,
7728 		 * we would not submit another untagged command
7729 		 * or if reached LUN execution throttle.
7730 		 */
7731 		if (sp->flags & SRB_FCP_CMD_PKT) {
7732 			if (lq->flags & LQF_UNTAGGED_PENDING ||
7733 			    lq->lun_outcnt >= ha->execution_throttle) {
7734 				QL_PRINT_8(CE_CONT, "(%d): break, d_id=%xh, "
7735 				    "lf=%xh, lun_outcnt=%xh\n", ha->instance,
7736 				    tq->d_id.b24, lq->flags, lq->lun_outcnt);
7737 				break;
7738 			}
7739 			if (sp->fcp->fcp_cntl.cntl_qtype ==
7740 			    FCP_QTYPE_UNTAGGED) {
7741 				/*
7742 				 * Set the untagged-flag for the LUN
7743 				 * so that no more untagged commands
7744 				 * can be submitted for this LUN.
7745 				 */
7746 				lq->flags |= LQF_UNTAGGED_PENDING;
7747 			}
7748 
7749 			/* Count command as sent. */
7750 			lq->lun_outcnt++;
7751 		}
7752 
7753 		/* Remove srb from device queue. */
7754 		ql_remove_link(&lq->cmd, &sp->cmd);
7755 		sp->flags &= ~SRB_IN_DEVICE_QUEUE;
7756 
7757 		tq->outcnt++;
7758 
7759 		ql_start_iocb(vha, sp);
7760 	}
7761 
7762 	/* Release device queue lock. */
7763 	DEVICE_QUEUE_UNLOCK(tq);
7764 
7765 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
7766 }
7767 
7768 /*
7769  * ql_done
7770  *	Process completed commands.
7771  *
7772  * Input:
7773  *	link:	first command link in chain.
7774  *
7775  * Context:
7776  *	Interrupt or Kernel context, no mailbox commands allowed.
7777  */
7778 void
ql_done(ql_link_t * link)7779 ql_done(ql_link_t *link)
7780 {
7781 	ql_adapter_state_t	*ha;
7782 	ql_link_t		*next_link;
7783 	ql_srb_t		*sp;
7784 	ql_tgt_t		*tq;
7785 	ql_lun_t		*lq;
7786 
7787 	QL_PRINT_3(CE_CONT, "started\n");
7788 
7789 	for (; link != NULL; link = next_link) {
7790 		next_link = link->next;
7791 		sp = link->base_address;
7792 		ha = sp->ha;
7793 
7794 		if (sp->flags & SRB_UB_CALLBACK) {
7795 			QL_UB_LOCK(ha);
7796 			if (sp->flags & SRB_UB_IN_ISP) {
7797 				if (ha->ub_outcnt != 0) {
7798 					ha->ub_outcnt--;
7799 				}
7800 				QL_UB_UNLOCK(ha);
7801 				ql_isp_rcvbuf(ha);
7802 				QL_UB_LOCK(ha);
7803 			}
7804 			QL_UB_UNLOCK(ha);
7805 			ql_awaken_task_daemon(ha, sp, 0, 0);
7806 		} else {
7807 			/* Free outstanding command slot. */
7808 			if (sp->handle != 0) {
7809 				ha->outstanding_cmds[
7810 				    sp->handle & OSC_INDEX_MASK] = NULL;
7811 				sp->handle = 0;
7812 				sp->flags &= ~SRB_IN_TOKEN_ARRAY;
7813 			}
7814 
7815 			/* Acquire device queue lock. */
7816 			lq = sp->lun_queue;
7817 			tq = lq->target_queue;
7818 			DEVICE_QUEUE_LOCK(tq);
7819 
7820 			/* Decrement outstanding commands on device. */
7821 			if (tq->outcnt != 0) {
7822 				tq->outcnt--;
7823 			}
7824 
7825 			if (sp->flags & SRB_FCP_CMD_PKT) {
7826 				if (sp->fcp->fcp_cntl.cntl_qtype ==
7827 				    FCP_QTYPE_UNTAGGED) {
7828 					/*
7829 					 * Clear the flag for this LUN so that
7830 					 * untagged commands can be submitted
7831 					 * for it.
7832 					 */
7833 					lq->flags &= ~LQF_UNTAGGED_PENDING;
7834 				}
7835 
7836 				if (lq->lun_outcnt != 0) {
7837 					lq->lun_outcnt--;
7838 				}
7839 			}
7840 
7841 			/* Reset port down retry count on good completion. */
7842 			if (sp->pkt->pkt_reason == CS_COMPLETE) {
7843 				tq->port_down_retry_count =
7844 				    ha->port_down_retry_count;
7845 				tq->qfull_retry_count = ha->qfull_retry_count;
7846 			}
7847 
7848 
7849 			/* Alter aborted status for fast timeout feature */
7850 			if (CFG_IST(ha, CFG_FAST_TIMEOUT) &&
7851 			    (sp->flags & (SRB_MS_PKT | SRB_ELS_PKT) ||
7852 			    !(tq->flags & TQF_NEED_AUTHENTICATION)) &&
7853 			    sp->flags & SRB_RETRY &&
7854 			    (sp->flags & SRB_WATCHDOG_ENABLED &&
7855 			    sp->wdg_q_time > 1)) {
7856 				EL(ha, "fast abort modify change\n");
7857 				sp->flags &= ~(SRB_RETRY);
7858 				sp->pkt->pkt_reason = CS_TIMEOUT;
7859 			}
7860 
7861 			/* Place request back on top of target command queue */
7862 			if ((sp->flags & (SRB_MS_PKT | SRB_ELS_PKT) ||
7863 			    !(tq->flags & TQF_NEED_AUTHENTICATION)) &&
7864 			    sp->flags & SRB_RETRY &&
7865 			    (sp->flags & SRB_WATCHDOG_ENABLED &&
7866 			    sp->wdg_q_time > 1)) {
7867 				sp->flags &= ~(SRB_ISP_STARTED |
7868 				    SRB_ISP_COMPLETED | SRB_RETRY);
7869 
7870 				/* Reset watchdog timer */
7871 				sp->wdg_q_time = sp->init_wdg_q_time;
7872 
7873 				/* Issue marker command on reset status. */
7874 				if (!(ha->task_daemon_flags & LOOP_DOWN) &&
7875 				    (sp->pkt->pkt_reason == CS_RESET ||
7876 				    (CFG_IST(ha, CFG_CTRL_24258081) &&
7877 				    sp->pkt->pkt_reason == CS_ABORTED))) {
7878 					(void) ql_marker(ha, tq->loop_id, 0,
7879 					    MK_SYNC_ID);
7880 				}
7881 
7882 				ql_add_link_t(&lq->cmd, &sp->cmd);
7883 				sp->flags |= SRB_IN_DEVICE_QUEUE;
7884 				ql_next(ha, lq);
7885 			} else {
7886 				/* Remove command from watchdog queue. */
7887 				if (sp->flags & SRB_WATCHDOG_ENABLED) {
7888 					ql_remove_link(&tq->wdg, &sp->wdg);
7889 					sp->flags &= ~SRB_WATCHDOG_ENABLED;
7890 				}
7891 
7892 				if (lq->cmd.first != NULL) {
7893 					ql_next(ha, lq);
7894 				} else {
7895 					/* Release LU queue specific lock. */
7896 					DEVICE_QUEUE_UNLOCK(tq);
7897 					if (ha->pha->pending_cmds.first !=
7898 					    NULL) {
7899 						ql_start_iocb(ha, NULL);
7900 					}
7901 				}
7902 
7903 				/* Sync buffers if required.  */
7904 				if (sp->flags & (SRB_MS_PKT | SRB_ELS_PKT)) {
7905 					(void) ddi_dma_sync(
7906 					    sp->pkt->pkt_resp_dma,
7907 					    0, 0, DDI_DMA_SYNC_FORCPU);
7908 				}
7909 
7910 				/* Map ISP completion codes. */
7911 				sp->pkt->pkt_expln = FC_EXPLN_NONE;
7912 				sp->pkt->pkt_action = FC_ACTION_RETRYABLE;
7913 				switch (sp->pkt->pkt_reason) {
7914 				case CS_COMPLETE:
7915 					sp->pkt->pkt_state = FC_PKT_SUCCESS;
7916 					break;
7917 				case CS_RESET:
7918 					/* Issue marker command. */
7919 					if (!(ha->task_daemon_flags &
7920 					    LOOP_DOWN)) {
7921 						(void) ql_marker(ha,
7922 						    tq->loop_id, 0,
7923 						    MK_SYNC_ID);
7924 					}
7925 					sp->pkt->pkt_state =
7926 					    FC_PKT_PORT_OFFLINE;
7927 					sp->pkt->pkt_reason =
7928 					    FC_REASON_ABORTED;
7929 					break;
7930 				case CS_RESOUCE_UNAVAILABLE:
7931 					sp->pkt->pkt_state = FC_PKT_LOCAL_BSY;
7932 					sp->pkt->pkt_reason =
7933 					    FC_REASON_PKT_BUSY;
7934 					break;
7935 
7936 				case CS_TIMEOUT:
7937 					sp->pkt->pkt_state = FC_PKT_TIMEOUT;
7938 					sp->pkt->pkt_reason =
7939 					    FC_REASON_HW_ERROR;
7940 					break;
7941 				case CS_DATA_OVERRUN:
7942 					sp->pkt->pkt_state = FC_PKT_LOCAL_RJT;
7943 					sp->pkt->pkt_reason =
7944 					    FC_REASON_OVERRUN;
7945 					break;
7946 				case CS_PORT_UNAVAILABLE:
7947 				case CS_PORT_LOGGED_OUT:
7948 					sp->pkt->pkt_state =
7949 					    FC_PKT_PORT_OFFLINE;
7950 					sp->pkt->pkt_reason =
7951 					    FC_REASON_LOGIN_REQUIRED;
7952 					ql_send_logo(ha, tq, NULL);
7953 					break;
7954 				case CS_PORT_CONFIG_CHG:
7955 					sp->pkt->pkt_state =
7956 					    FC_PKT_PORT_OFFLINE;
7957 					sp->pkt->pkt_reason =
7958 					    FC_REASON_OFFLINE;
7959 					break;
7960 				case CS_QUEUE_FULL:
7961 					sp->pkt->pkt_state = FC_PKT_LOCAL_RJT;
7962 					sp->pkt->pkt_reason = FC_REASON_QFULL;
7963 					break;
7964 
7965 				case CS_ABORTED:
7966 					DEVICE_QUEUE_LOCK(tq);
7967 					if (tq->flags & (TQF_RSCN_RCVD |
7968 					    TQF_NEED_AUTHENTICATION)) {
7969 						sp->pkt->pkt_state =
7970 						    FC_PKT_PORT_OFFLINE;
7971 						sp->pkt->pkt_reason =
7972 						    FC_REASON_LOGIN_REQUIRED;
7973 					} else {
7974 						sp->pkt->pkt_state =
7975 						    FC_PKT_LOCAL_RJT;
7976 						sp->pkt->pkt_reason =
7977 						    FC_REASON_ABORTED;
7978 					}
7979 					DEVICE_QUEUE_UNLOCK(tq);
7980 					break;
7981 
7982 				case CS_TRANSPORT:
7983 					sp->pkt->pkt_state = FC_PKT_LOCAL_RJT;
7984 					sp->pkt->pkt_reason =
7985 					    FC_PKT_TRAN_ERROR;
7986 					break;
7987 
7988 				case CS_DATA_UNDERRUN:
7989 					sp->pkt->pkt_state = FC_PKT_LOCAL_RJT;
7990 					sp->pkt->pkt_reason =
7991 					    FC_REASON_UNDERRUN;
7992 					break;
7993 				case CS_DMA_ERROR:
7994 				case CS_BAD_PAYLOAD:
7995 				case CS_UNKNOWN:
7996 				case CS_CMD_FAILED:
7997 				default:
7998 					sp->pkt->pkt_state = FC_PKT_LOCAL_RJT;
7999 					sp->pkt->pkt_reason =
8000 					    FC_REASON_HW_ERROR;
8001 					break;
8002 				}
8003 
8004 				/* Now call the pkt completion callback */
8005 				if (sp->flags & SRB_POLL) {
8006 					sp->flags &= ~SRB_POLL;
8007 				} else if (sp->pkt->pkt_comp) {
8008 					if (sp->pkt->pkt_tran_flags &
8009 					    FC_TRAN_IMMEDIATE_CB) {
8010 						(*sp->pkt->pkt_comp)(sp->pkt);
8011 					} else {
8012 						ql_awaken_task_daemon(ha, sp,
8013 						    0, 0);
8014 					}
8015 				}
8016 			}
8017 		}
8018 	}
8019 
8020 	QL_PRINT_3(CE_CONT, "done\n");
8021 }
8022 
8023 /*
8024  * ql_awaken_task_daemon
8025  *	Adds command completion callback to callback queue and/or
8026  *	awakens task daemon thread.
8027  *
8028  * Input:
8029  *	ha:		adapter state pointer.
8030  *	sp:		srb pointer.
8031  *	set_flags:	task daemon flags to set.
8032  *	reset_flags:	task daemon flags to reset.
8033  *
8034  * Context:
8035  *	Interrupt or Kernel context, no mailbox commands allowed.
8036  */
8037 void
ql_awaken_task_daemon(ql_adapter_state_t * vha,ql_srb_t * sp,uint32_t set_flags,uint32_t reset_flags)8038 ql_awaken_task_daemon(ql_adapter_state_t *vha, ql_srb_t *sp,
8039     uint32_t set_flags, uint32_t reset_flags)
8040 {
8041 	ql_adapter_state_t	*ha = vha->pha;
8042 
8043 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
8044 
8045 	/* Acquire task daemon lock. */
8046 	TASK_DAEMON_LOCK(ha);
8047 
8048 	if (set_flags & ISP_ABORT_NEEDED) {
8049 		if (ha->task_daemon_flags & ABORT_ISP_ACTIVE) {
8050 			set_flags &= ~ISP_ABORT_NEEDED;
8051 		}
8052 	}
8053 
8054 	ha->task_daemon_flags |= set_flags;
8055 	ha->task_daemon_flags &= ~reset_flags;
8056 
8057 	if (QL_DAEMON_SUSPENDED(ha)) {
8058 		if (sp != NULL) {
8059 			TASK_DAEMON_UNLOCK(ha);
8060 
8061 			/* Do callback. */
8062 			if (sp->flags & SRB_UB_CALLBACK) {
8063 				ql_unsol_callback(sp);
8064 			} else {
8065 				(*sp->pkt->pkt_comp)(sp->pkt);
8066 			}
8067 		} else {
8068 			if (!(curthread->t_flag & T_INTR_THREAD) &&
8069 			    !(ha->task_daemon_flags & TASK_THREAD_CALLED)) {
8070 				ha->task_daemon_flags |= TASK_THREAD_CALLED;
8071 				ql_task_thread(ha);
8072 				ha->task_daemon_flags &= ~TASK_THREAD_CALLED;
8073 			}
8074 
8075 			TASK_DAEMON_UNLOCK(ha);
8076 		}
8077 	} else {
8078 		if (sp != NULL) {
8079 			ql_add_link_b(&ha->callback_queue, &sp->cmd);
8080 		}
8081 
8082 		if (ha->task_daemon_flags & TASK_DAEMON_SLEEPING_FLG) {
8083 			cv_broadcast(&ha->cv_task_daemon);
8084 		}
8085 		TASK_DAEMON_UNLOCK(ha);
8086 	}
8087 
8088 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
8089 }
8090 
8091 /*
8092  * ql_task_daemon
8093  *	Thread that is awaken by the driver when a
8094  *	background needs to be done.
8095  *
8096  * Input:
8097  *	arg = adapter state pointer.
8098  *
8099  * Context:
8100  *	Kernel context.
8101  */
8102 static void
ql_task_daemon(void * arg)8103 ql_task_daemon(void *arg)
8104 {
8105 	ql_adapter_state_t	*ha = (void *)arg;
8106 
8107 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
8108 
8109 	CALLB_CPR_INIT(&ha->cprinfo, &ha->task_daemon_mutex, callb_generic_cpr,
8110 	    "ql_task_daemon");
8111 
8112 	/* Acquire task daemon lock. */
8113 	TASK_DAEMON_LOCK(ha);
8114 
8115 	ha->task_daemon_flags |= TASK_DAEMON_ALIVE_FLG;
8116 
8117 	while ((ha->task_daemon_flags & TASK_DAEMON_STOP_FLG) == 0) {
8118 		ql_task_thread(ha);
8119 
8120 		QL_PRINT_3(CE_CONT, "(%d): Going to sleep\n", ha->instance);
8121 
8122 		/*
8123 		 * Before we wait on the conditional variable, we
8124 		 * need to check if STOP_FLG is set for us to terminate
8125 		 */
8126 		if (ha->task_daemon_flags & TASK_DAEMON_STOP_FLG) {
8127 			break;
8128 		}
8129 
8130 		/*LINTED [Solaris CALLB_CPR_SAFE_BEGIN Lint error]*/
8131 		CALLB_CPR_SAFE_BEGIN(&ha->cprinfo);
8132 
8133 		ha->task_daemon_flags |= TASK_DAEMON_SLEEPING_FLG;
8134 
8135 		/* If killed, stop task daemon */
8136 		if (cv_wait_sig(&ha->cv_task_daemon,
8137 		    &ha->task_daemon_mutex) == 0) {
8138 			ha->task_daemon_flags |= TASK_DAEMON_STOP_FLG;
8139 		}
8140 
8141 		ha->task_daemon_flags &= ~TASK_DAEMON_SLEEPING_FLG;
8142 
8143 		/*LINTED [Solaris CALLB_CPR_SAFE_END Lint error]*/
8144 		CALLB_CPR_SAFE_END(&ha->cprinfo, &ha->task_daemon_mutex);
8145 
8146 		QL_PRINT_3(CE_CONT, "(%d): Awakened\n", ha->instance);
8147 	}
8148 
8149 	ha->task_daemon_flags &= ~(TASK_DAEMON_STOP_FLG |
8150 	    TASK_DAEMON_ALIVE_FLG);
8151 
8152 	/*LINTED [Solaris CALLB_CPR_EXIT Lint error]*/
8153 	CALLB_CPR_EXIT(&ha->cprinfo);
8154 
8155 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
8156 
8157 	thread_exit();
8158 }
8159 
8160 /*
8161  * ql_task_thread
8162  *	Thread run by daemon.
8163  *
8164  * Input:
8165  *	ha = adapter state pointer.
8166  *	TASK_DAEMON_LOCK must be acquired prior to call.
8167  *
8168  * Context:
8169  *	Kernel context.
8170  */
8171 static void
ql_task_thread(ql_adapter_state_t * ha)8172 ql_task_thread(ql_adapter_state_t *ha)
8173 {
8174 	int			loop_again;
8175 	ql_srb_t		*sp;
8176 	ql_head_t		*head;
8177 	ql_link_t		*link;
8178 	caddr_t			msg;
8179 	ql_adapter_state_t	*vha;
8180 
8181 	do {
8182 		QL_PRINT_3(CE_CONT, "(%d): task_daemon_flags=%xh\n",
8183 		    ha->instance, ha->task_daemon_flags);
8184 
8185 		loop_again = FALSE;
8186 
8187 		QL_PM_LOCK(ha);
8188 		if (ha->power_level != PM_LEVEL_D0) {
8189 			QL_PM_UNLOCK(ha);
8190 			ha->task_daemon_flags |= TASK_DAEMON_STALLED_FLG;
8191 			break;
8192 		}
8193 		QL_PM_UNLOCK(ha);
8194 
8195 		/* IDC event. */
8196 		if (ha->task_daemon_flags & IDC_EVENT) {
8197 			ha->task_daemon_flags &= ~IDC_EVENT;
8198 			TASK_DAEMON_UNLOCK(ha);
8199 			ql_process_idc_event(ha);
8200 			TASK_DAEMON_LOCK(ha);
8201 			loop_again = TRUE;
8202 		}
8203 
8204 		if (ha->flags & ADAPTER_SUSPENDED || ha->task_daemon_flags &
8205 		    (TASK_DAEMON_STOP_FLG | DRIVER_STALL) ||
8206 		    (ha->flags & ONLINE) == 0) {
8207 			ha->task_daemon_flags |= TASK_DAEMON_STALLED_FLG;
8208 			break;
8209 		}
8210 		ha->task_daemon_flags &= ~TASK_DAEMON_STALLED_FLG;
8211 
8212 		if (ha->task_daemon_flags & ISP_ABORT_NEEDED) {
8213 			TASK_DAEMON_UNLOCK(ha);
8214 			if (ha->log_parity_pause == B_TRUE) {
8215 				(void) ql_flash_errlog(ha,
8216 				    FLASH_ERRLOG_PARITY_ERR, 0,
8217 				    MSW(ha->parity_stat_err),
8218 				    LSW(ha->parity_stat_err));
8219 				ha->log_parity_pause = B_FALSE;
8220 			}
8221 			ql_port_state(ha, FC_STATE_OFFLINE, FC_STATE_CHANGE);
8222 			TASK_DAEMON_LOCK(ha);
8223 			loop_again = TRUE;
8224 		}
8225 
8226 		/* Idle Check. */
8227 		if (ha->task_daemon_flags & TASK_DAEMON_IDLE_CHK_FLG) {
8228 			ha->task_daemon_flags &= ~TASK_DAEMON_IDLE_CHK_FLG;
8229 			if (!(ha->task_daemon_flags & QL_SUSPENDED)) {
8230 				TASK_DAEMON_UNLOCK(ha);
8231 				ql_idle_check(ha);
8232 				TASK_DAEMON_LOCK(ha);
8233 				loop_again = TRUE;
8234 			}
8235 		}
8236 
8237 		/* Crystal+ port#0 bypass transition */
8238 		if (ha->task_daemon_flags & HANDLE_PORT_BYPASS_CHANGE) {
8239 			ha->task_daemon_flags &= ~HANDLE_PORT_BYPASS_CHANGE;
8240 			TASK_DAEMON_UNLOCK(ha);
8241 			(void) ql_initiate_lip(ha);
8242 			TASK_DAEMON_LOCK(ha);
8243 			loop_again = TRUE;
8244 		}
8245 
8246 		/* Abort queues needed. */
8247 		if (ha->task_daemon_flags & ABORT_QUEUES_NEEDED) {
8248 			ha->task_daemon_flags &= ~ABORT_QUEUES_NEEDED;
8249 			TASK_DAEMON_UNLOCK(ha);
8250 			ql_abort_queues(ha);
8251 			TASK_DAEMON_LOCK(ha);
8252 		}
8253 
8254 		/* Not suspended, awaken waiting routines. */
8255 		if (!(ha->task_daemon_flags & QL_SUSPENDED) &&
8256 		    ha->task_daemon_flags & SUSPENDED_WAKEUP_FLG) {
8257 			ha->task_daemon_flags &= ~SUSPENDED_WAKEUP_FLG;
8258 			cv_broadcast(&ha->cv_dr_suspended);
8259 			loop_again = TRUE;
8260 		}
8261 
8262 		/* Handle RSCN changes. */
8263 		for (vha = ha; vha != NULL; vha = vha->vp_next) {
8264 			if (vha->task_daemon_flags & RSCN_UPDATE_NEEDED) {
8265 				vha->task_daemon_flags &= ~RSCN_UPDATE_NEEDED;
8266 				TASK_DAEMON_UNLOCK(ha);
8267 				(void) ql_handle_rscn_update(vha);
8268 				TASK_DAEMON_LOCK(ha);
8269 				loop_again = TRUE;
8270 			}
8271 		}
8272 
8273 		/* Handle state changes. */
8274 		for (vha = ha; vha != NULL; vha = vha->vp_next) {
8275 			if (vha->task_daemon_flags & FC_STATE_CHANGE &&
8276 			    !(ha->task_daemon_flags &
8277 			    TASK_DAEMON_POWERING_DOWN)) {
8278 				/* Report state change. */
8279 				EL(vha, "state change = %xh\n", vha->state);
8280 				vha->task_daemon_flags &= ~FC_STATE_CHANGE;
8281 
8282 				if (vha->task_daemon_flags &
8283 				    COMMAND_WAIT_NEEDED) {
8284 					vha->task_daemon_flags &=
8285 					    ~COMMAND_WAIT_NEEDED;
8286 					if (!(ha->task_daemon_flags &
8287 					    COMMAND_WAIT_ACTIVE)) {
8288 						ha->task_daemon_flags |=
8289 						    COMMAND_WAIT_ACTIVE;
8290 						TASK_DAEMON_UNLOCK(ha);
8291 						ql_cmd_wait(ha);
8292 						TASK_DAEMON_LOCK(ha);
8293 						ha->task_daemon_flags &=
8294 						    ~COMMAND_WAIT_ACTIVE;
8295 					}
8296 				}
8297 
8298 				msg = NULL;
8299 				if (FC_PORT_STATE_MASK(vha->state) ==
8300 				    FC_STATE_OFFLINE) {
8301 					if (vha->task_daemon_flags &
8302 					    STATE_ONLINE) {
8303 						if (ha->topology &
8304 						    QL_LOOP_CONNECTION) {
8305 							msg = "Loop OFFLINE";
8306 						} else {
8307 							msg = "Link OFFLINE";
8308 						}
8309 					}
8310 					vha->task_daemon_flags &=
8311 					    ~STATE_ONLINE;
8312 				} else if (FC_PORT_STATE_MASK(vha->state) ==
8313 				    FC_STATE_LOOP) {
8314 					if (!(vha->task_daemon_flags &
8315 					    STATE_ONLINE)) {
8316 						msg = "Loop ONLINE";
8317 					}
8318 					vha->task_daemon_flags |= STATE_ONLINE;
8319 				} else if (FC_PORT_STATE_MASK(vha->state) ==
8320 				    FC_STATE_ONLINE) {
8321 					if (!(vha->task_daemon_flags &
8322 					    STATE_ONLINE)) {
8323 						msg = "Link ONLINE";
8324 					}
8325 					vha->task_daemon_flags |= STATE_ONLINE;
8326 				} else {
8327 					msg = "Unknown Link state";
8328 				}
8329 
8330 				if (msg != NULL) {
8331 					cmn_err(CE_NOTE, "!Qlogic %s(%d,%d): "
8332 					    "%s", QL_NAME, ha->instance,
8333 					    vha->vp_index, msg);
8334 				}
8335 
8336 				if (vha->flags & FCA_BOUND) {
8337 					QL_PRINT_10(CE_CONT, "(%d,%d): statec_"
8338 					    "cb state=%xh\n", ha->instance,
8339 					    vha->vp_index, vha->state);
8340 					TASK_DAEMON_UNLOCK(ha);
8341 					(vha->bind_info.port_statec_cb)
8342 					    (vha->bind_info.port_handle,
8343 					    vha->state);
8344 					TASK_DAEMON_LOCK(ha);
8345 				}
8346 				loop_again = TRUE;
8347 			}
8348 		}
8349 
8350 		if (ha->task_daemon_flags & LIP_RESET_PENDING &&
8351 		    !(ha->task_daemon_flags & TASK_DAEMON_POWERING_DOWN)) {
8352 			EL(ha, "processing LIP reset\n");
8353 			ha->task_daemon_flags &= ~LIP_RESET_PENDING;
8354 			TASK_DAEMON_UNLOCK(ha);
8355 			for (vha = ha; vha != NULL; vha = vha->vp_next) {
8356 				if (vha->flags & FCA_BOUND) {
8357 					QL_PRINT_10(CE_CONT, "(%d,%d): statec_"
8358 					    "cb reset\n", ha->instance,
8359 					    vha->vp_index);
8360 					(vha->bind_info.port_statec_cb)
8361 					    (vha->bind_info.port_handle,
8362 					    FC_STATE_TARGET_PORT_RESET);
8363 				}
8364 			}
8365 			TASK_DAEMON_LOCK(ha);
8366 			loop_again = TRUE;
8367 		}
8368 
8369 		if (QL_IS_SET(ha->task_daemon_flags, NEED_UNSOLICITED_BUFFERS |
8370 		    FIRMWARE_UP)) {
8371 			/*
8372 			 * The firmware needs more unsolicited
8373 			 * buffers. We cannot allocate any new
8374 			 * buffers unless the ULP module requests
8375 			 * for new buffers. All we can do here is
8376 			 * to give received buffers from the pool
8377 			 * that is already allocated
8378 			 */
8379 			ha->task_daemon_flags &= ~NEED_UNSOLICITED_BUFFERS;
8380 			TASK_DAEMON_UNLOCK(ha);
8381 			ql_isp_rcvbuf(ha);
8382 			TASK_DAEMON_LOCK(ha);
8383 			loop_again = TRUE;
8384 		}
8385 
8386 		if (ha->task_daemon_flags & ISP_ABORT_NEEDED) {
8387 			TASK_DAEMON_UNLOCK(ha);
8388 			(void) ql_abort_isp(ha);
8389 			TASK_DAEMON_LOCK(ha);
8390 			loop_again = TRUE;
8391 		}
8392 
8393 		if (!(ha->task_daemon_flags & (LOOP_DOWN | DRIVER_STALL |
8394 		    COMMAND_WAIT_NEEDED))) {
8395 			if (QL_IS_SET(ha->task_daemon_flags,
8396 			    RESET_MARKER_NEEDED | FIRMWARE_UP)) {
8397 				ha->task_daemon_flags &= ~RESET_MARKER_NEEDED;
8398 				if (!(ha->task_daemon_flags & RESET_ACTIVE)) {
8399 					ha->task_daemon_flags |= RESET_ACTIVE;
8400 					TASK_DAEMON_UNLOCK(ha);
8401 					for (vha = ha; vha != NULL;
8402 					    vha = vha->vp_next) {
8403 						ql_rst_aen(vha);
8404 					}
8405 					TASK_DAEMON_LOCK(ha);
8406 					ha->task_daemon_flags &= ~RESET_ACTIVE;
8407 					loop_again = TRUE;
8408 				}
8409 			}
8410 
8411 			if (QL_IS_SET(ha->task_daemon_flags,
8412 			    LOOP_RESYNC_NEEDED | FIRMWARE_UP)) {
8413 				if (!(ha->task_daemon_flags &
8414 				    LOOP_RESYNC_ACTIVE)) {
8415 					ha->task_daemon_flags |=
8416 					    LOOP_RESYNC_ACTIVE;
8417 					TASK_DAEMON_UNLOCK(ha);
8418 					(void) ql_loop_resync(ha);
8419 					TASK_DAEMON_LOCK(ha);
8420 					loop_again = TRUE;
8421 				}
8422 			}
8423 		}
8424 
8425 		/* Port retry needed. */
8426 		if (ha->task_daemon_flags & PORT_RETRY_NEEDED) {
8427 			ha->task_daemon_flags &= ~PORT_RETRY_NEEDED;
8428 			ADAPTER_STATE_LOCK(ha);
8429 			ha->port_retry_timer = 0;
8430 			ADAPTER_STATE_UNLOCK(ha);
8431 
8432 			TASK_DAEMON_UNLOCK(ha);
8433 			ql_restart_queues(ha);
8434 			TASK_DAEMON_LOCK(ha);
8435 			loop_again = B_TRUE;
8436 		}
8437 
8438 		/* iiDMA setting needed? */
8439 		if (ha->task_daemon_flags & TD_IIDMA_NEEDED) {
8440 			ha->task_daemon_flags &= ~TD_IIDMA_NEEDED;
8441 
8442 			TASK_DAEMON_UNLOCK(ha);
8443 			ql_iidma(ha);
8444 			TASK_DAEMON_LOCK(ha);
8445 			loop_again = B_TRUE;
8446 		}
8447 
8448 		if (ha->task_daemon_flags & SEND_PLOGI) {
8449 			ha->task_daemon_flags &= ~SEND_PLOGI;
8450 			TASK_DAEMON_UNLOCK(ha);
8451 			(void) ql_n_port_plogi(ha);
8452 			TASK_DAEMON_LOCK(ha);
8453 		}
8454 
8455 		head = &ha->callback_queue;
8456 		if (head->first != NULL) {
8457 			sp = head->first->base_address;
8458 			link = &sp->cmd;
8459 
8460 			/* Dequeue command. */
8461 			ql_remove_link(head, link);
8462 
8463 			/* Release task daemon lock. */
8464 			TASK_DAEMON_UNLOCK(ha);
8465 
8466 			/* Do callback. */
8467 			if (sp->flags & SRB_UB_CALLBACK) {
8468 				ql_unsol_callback(sp);
8469 			} else {
8470 				(*sp->pkt->pkt_comp)(sp->pkt);
8471 			}
8472 
8473 			/* Acquire task daemon lock. */
8474 			TASK_DAEMON_LOCK(ha);
8475 
8476 			loop_again = TRUE;
8477 		}
8478 
8479 	} while (loop_again);
8480 }
8481 
8482 /*
8483  * ql_idle_check
8484  *	Test for adapter is alive and well.
8485  *
8486  * Input:
8487  *	ha:	adapter state pointer.
8488  *
8489  * Context:
8490  *	Kernel context.
8491  */
8492 static void
ql_idle_check(ql_adapter_state_t * ha)8493 ql_idle_check(ql_adapter_state_t *ha)
8494 {
8495 	ddi_devstate_t	state;
8496 	int		rval;
8497 	ql_mbx_data_t	mr;
8498 
8499 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
8500 
8501 	/* Firmware Ready Test. */
8502 	rval = ql_get_firmware_state(ha, &mr);
8503 	if (!(ha->task_daemon_flags & QL_SUSPENDED) &&
8504 	    (rval != QL_SUCCESS || mr.mb[1] != FSTATE_READY)) {
8505 		EL(ha, "failed, Firmware Ready Test = %xh\n", rval);
8506 		state = ddi_get_devstate(ha->dip);
8507 		if (state == DDI_DEVSTATE_UP) {
8508 			/*EMPTY*/
8509 			ddi_dev_report_fault(ha->dip, DDI_SERVICE_DEGRADED,
8510 			    DDI_DEVICE_FAULT, "Firmware Ready Test failed");
8511 		}
8512 		TASK_DAEMON_LOCK(ha);
8513 		if (!(ha->task_daemon_flags & ABORT_ISP_ACTIVE)) {
8514 			EL(ha, "fstate_ready, isp_abort_needed\n");
8515 			ha->task_daemon_flags |= ISP_ABORT_NEEDED;
8516 		}
8517 		TASK_DAEMON_UNLOCK(ha);
8518 	}
8519 
8520 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
8521 }
8522 
8523 /*
8524  * ql_unsol_callback
8525  *	Handle unsolicited buffer callbacks.
8526  *
8527  * Input:
8528  *	ha = adapter state pointer.
8529  *	sp = srb pointer.
8530  *
8531  * Context:
8532  *	Kernel context.
8533  */
8534 static void
ql_unsol_callback(ql_srb_t * sp)8535 ql_unsol_callback(ql_srb_t *sp)
8536 {
8537 	fc_affected_id_t	*af;
8538 	fc_unsol_buf_t		*ubp;
8539 	uchar_t			r_ctl;
8540 	uchar_t			ls_code;
8541 	ql_tgt_t		*tq;
8542 	ql_adapter_state_t	*ha = sp->ha, *pha = sp->ha->pha;
8543 
8544 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
8545 
8546 	ubp = ha->ub_array[sp->handle];
8547 	r_ctl = ubp->ub_frame.r_ctl;
8548 	ls_code = ubp->ub_buffer[0];
8549 
8550 	if (sp->lun_queue == NULL) {
8551 		tq = NULL;
8552 	} else {
8553 		tq = sp->lun_queue->target_queue;
8554 	}
8555 
8556 	QL_UB_LOCK(ha);
8557 	if (sp->flags & SRB_UB_FREE_REQUESTED ||
8558 	    pha->task_daemon_flags & TASK_DAEMON_POWERING_DOWN) {
8559 		sp->flags &= ~(SRB_UB_IN_ISP | SRB_UB_CALLBACK |
8560 		    SRB_UB_RSCN | SRB_UB_FCP | SRB_UB_ACQUIRED);
8561 		sp->flags |= SRB_UB_IN_FCA;
8562 		QL_UB_UNLOCK(ha);
8563 		return;
8564 	}
8565 
8566 	/* Process RSCN */
8567 	if (sp->flags & SRB_UB_RSCN) {
8568 		int sendup = 1;
8569 
8570 		/*
8571 		 * Defer RSCN posting until commands return
8572 		 */
8573 		QL_UB_UNLOCK(ha);
8574 
8575 		af = (fc_affected_id_t *)((caddr_t)ubp->ub_buffer + 4);
8576 
8577 		/* Abort outstanding commands */
8578 		sendup = ql_process_rscn(ha, af);
8579 		if (sendup == 0) {
8580 
8581 			TASK_DAEMON_LOCK(ha);
8582 			ql_add_link_b(&pha->callback_queue, &sp->cmd);
8583 			TASK_DAEMON_UNLOCK(ha);
8584 
8585 			/*
8586 			 * Wait for commands to drain in F/W (doesn't take
8587 			 * more than a few milliseconds)
8588 			 */
8589 			ql_delay(ha, 10000);
8590 
8591 			QL_PRINT_2(CE_CONT, "(%d,%d): done rscn_sendup=0, "
8592 			    "fmt=%xh, d_id=%xh\n", ha->instance, ha->vp_index,
8593 			    af->aff_format, af->aff_d_id);
8594 			return;
8595 		}
8596 
8597 		QL_UB_LOCK(ha);
8598 
8599 		EL(ha, "sending unsol rscn, fmt=%xh, d_id=%xh to transport\n",
8600 		    af->aff_format, af->aff_d_id);
8601 	}
8602 
8603 	/* Process UNSOL LOGO */
8604 	if ((r_ctl == R_CTL_ELS_REQ) && (ls_code == LA_ELS_LOGO)) {
8605 		QL_UB_UNLOCK(ha);
8606 
8607 		if (tq && (ql_process_logo_for_device(ha, tq) == 0)) {
8608 			TASK_DAEMON_LOCK(ha);
8609 			ql_add_link_b(&pha->callback_queue, &sp->cmd);
8610 			TASK_DAEMON_UNLOCK(ha);
8611 			QL_PRINT_2(CE_CONT, "(%d,%d): logo_sendup=0, d_id=%xh"
8612 			    "\n", ha->instance, ha->vp_index, tq->d_id.b24);
8613 			return;
8614 		}
8615 
8616 		QL_UB_LOCK(ha);
8617 		EL(ha, "sending unsol logout for %xh to transport\n",
8618 		    ubp->ub_frame.s_id);
8619 	}
8620 
8621 	sp->flags &= ~(SRB_UB_IN_FCA | SRB_UB_IN_ISP | SRB_UB_RSCN |
8622 	    SRB_UB_FCP);
8623 
8624 	if (sp->ub_type == FC_TYPE_IS8802_SNAP) {
8625 		(void) ddi_dma_sync(sp->ub_buffer.dma_handle, 0,
8626 		    ubp->ub_bufsize, DDI_DMA_SYNC_FORCPU);
8627 	}
8628 	QL_UB_UNLOCK(ha);
8629 
8630 	(ha->bind_info.port_unsol_cb)(ha->bind_info.port_handle,
8631 	    ubp, sp->ub_type);
8632 
8633 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
8634 }
8635 
8636 /*
8637  * ql_send_logo
8638  *
8639  * Input:
8640  *	ha:	adapter state pointer.
8641  *	tq:	target queue pointer.
8642  *	done_q:	done queue pointer.
8643  *
8644  * Context:
8645  *	Interrupt or Kernel context, no mailbox commands allowed.
8646  */
8647 void
ql_send_logo(ql_adapter_state_t * vha,ql_tgt_t * tq,ql_head_t * done_q)8648 ql_send_logo(ql_adapter_state_t *vha, ql_tgt_t *tq, ql_head_t *done_q)
8649 {
8650 	fc_unsol_buf_t		*ubp;
8651 	ql_srb_t		*sp;
8652 	la_els_logo_t		*payload;
8653 	ql_adapter_state_t	*ha = vha->pha;
8654 
8655 	QL_PRINT_3(CE_CONT, "(%d): started, d_id=%xh\n", ha->instance,
8656 	    tq->d_id.b24);
8657 
8658 	if ((tq->d_id.b24 == 0) || (tq->d_id.b24 == 0xffffff)) {
8659 		EL(ha, "no device, d_id=%xh\n", tq->d_id.b24);
8660 		return;
8661 	}
8662 
8663 	if ((tq->flags & (TQF_RSCN_RCVD | TQF_PLOGI_PROGRS)) == 0 &&
8664 	    tq->logout_sent == 0 && (ha->task_daemon_flags & LOOP_DOWN) == 0) {
8665 
8666 		/* Locate a buffer to use. */
8667 		ubp = ql_get_unsolicited_buffer(vha, FC_TYPE_EXTENDED_LS);
8668 		if (ubp == NULL) {
8669 			EL(vha, "Failed, get_unsolicited_buffer\n");
8670 			return;
8671 		}
8672 
8673 		DEVICE_QUEUE_LOCK(tq);
8674 		tq->flags |= TQF_NEED_AUTHENTICATION;
8675 		tq->logout_sent++;
8676 		DEVICE_QUEUE_UNLOCK(tq);
8677 
8678 		EL(vha, "Received LOGO from = %xh\n", tq->d_id.b24);
8679 
8680 		sp = ubp->ub_fca_private;
8681 
8682 		/* Set header. */
8683 		ubp->ub_frame.d_id = vha->d_id.b24;
8684 		ubp->ub_frame.r_ctl = R_CTL_ELS_REQ;
8685 		ubp->ub_frame.s_id = tq->d_id.b24;
8686 		ubp->ub_frame.rsvd = 0;
8687 		ubp->ub_frame.f_ctl = F_CTL_FIRST_SEQ | F_CTL_END_SEQ |
8688 		    F_CTL_SEQ_INITIATIVE;
8689 		ubp->ub_frame.type = FC_TYPE_EXTENDED_LS;
8690 		ubp->ub_frame.seq_cnt = 0;
8691 		ubp->ub_frame.df_ctl = 0;
8692 		ubp->ub_frame.seq_id = 0;
8693 		ubp->ub_frame.rx_id = 0xffff;
8694 		ubp->ub_frame.ox_id = 0xffff;
8695 
8696 		/* set payload. */
8697 		payload = (la_els_logo_t *)ubp->ub_buffer;
8698 		bzero(payload, sizeof (la_els_logo_t));
8699 		/* Make sure ls_code in payload is always big endian */
8700 		ubp->ub_buffer[0] = LA_ELS_LOGO;
8701 		ubp->ub_buffer[1] = 0;
8702 		ubp->ub_buffer[2] = 0;
8703 		ubp->ub_buffer[3] = 0;
8704 		bcopy(&vha->loginparams.node_ww_name.raw_wwn[0],
8705 		    &payload->nport_ww_name.raw_wwn[0], 8);
8706 		payload->nport_id.port_id = tq->d_id.b24;
8707 
8708 		QL_UB_LOCK(ha);
8709 		sp->flags |= SRB_UB_CALLBACK;
8710 		QL_UB_UNLOCK(ha);
8711 		if (tq->lun_queues.first != NULL) {
8712 			sp->lun_queue = (tq->lun_queues.first)->base_address;
8713 		} else {
8714 			sp->lun_queue = ql_lun_queue(vha, tq, 0);
8715 		}
8716 		if (done_q) {
8717 			ql_add_link_b(done_q, &sp->cmd);
8718 		} else {
8719 			ql_awaken_task_daemon(ha, sp, 0, 0);
8720 		}
8721 	}
8722 
8723 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
8724 }
8725 
8726 static int
ql_process_logo_for_device(ql_adapter_state_t * ha,ql_tgt_t * tq)8727 ql_process_logo_for_device(ql_adapter_state_t *ha, ql_tgt_t *tq)
8728 {
8729 	port_id_t	d_id;
8730 	ql_srb_t	*sp;
8731 	ql_link_t	*link;
8732 	int		sendup = 1;
8733 
8734 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
8735 
8736 	DEVICE_QUEUE_LOCK(tq);
8737 	if (tq->outcnt) {
8738 		DEVICE_QUEUE_UNLOCK(tq);
8739 		sendup = 0;
8740 		(void) ql_abort_device(ha, tq, 1);
8741 		ql_delay(ha, 10000);
8742 	} else {
8743 		DEVICE_QUEUE_UNLOCK(tq);
8744 		TASK_DAEMON_LOCK(ha);
8745 
8746 		for (link = ha->pha->callback_queue.first; link != NULL;
8747 		    link = link->next) {
8748 			sp = link->base_address;
8749 			if (sp->flags & SRB_UB_CALLBACK) {
8750 				continue;
8751 			}
8752 			d_id.b24 = sp->pkt->pkt_cmd_fhdr.d_id;
8753 
8754 			if (tq->d_id.b24 == d_id.b24) {
8755 				sendup = 0;
8756 				break;
8757 			}
8758 		}
8759 
8760 		TASK_DAEMON_UNLOCK(ha);
8761 	}
8762 
8763 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
8764 
8765 	return (sendup);
8766 }
8767 
8768 static int
ql_send_plogi(ql_adapter_state_t * ha,ql_tgt_t * tq,ql_head_t * done_q)8769 ql_send_plogi(ql_adapter_state_t *ha, ql_tgt_t *tq, ql_head_t *done_q)
8770 {
8771 	fc_unsol_buf_t		*ubp;
8772 	ql_srb_t		*sp;
8773 	la_els_logi_t		*payload;
8774 	class_svc_param_t	*class3_param;
8775 
8776 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
8777 
8778 	if ((tq->flags & TQF_RSCN_RCVD) || (ha->task_daemon_flags &
8779 	    LOOP_DOWN)) {
8780 		EL(ha, "Failed, tqf=%xh\n", tq->flags);
8781 		return (QL_FUNCTION_FAILED);
8782 	}
8783 
8784 	/* Locate a buffer to use. */
8785 	ubp = ql_get_unsolicited_buffer(ha, FC_TYPE_EXTENDED_LS);
8786 	if (ubp == NULL) {
8787 		EL(ha, "Failed\n");
8788 		return (QL_FUNCTION_FAILED);
8789 	}
8790 
8791 	QL_PRINT_3(CE_CONT, "(%d): Received LOGO from = %xh\n",
8792 	    ha->instance, tq->d_id.b24);
8793 
8794 	EL(ha, "Emulate PLOGI from = %xh tq = %x\n", tq->d_id.b24, tq);
8795 
8796 	sp = ubp->ub_fca_private;
8797 
8798 	/* Set header. */
8799 	ubp->ub_frame.d_id = ha->d_id.b24;
8800 	ubp->ub_frame.r_ctl = R_CTL_ELS_REQ;
8801 	ubp->ub_frame.s_id = tq->d_id.b24;
8802 	ubp->ub_frame.rsvd = 0;
8803 	ubp->ub_frame.f_ctl = F_CTL_FIRST_SEQ | F_CTL_END_SEQ |
8804 	    F_CTL_SEQ_INITIATIVE;
8805 	ubp->ub_frame.type = FC_TYPE_EXTENDED_LS;
8806 	ubp->ub_frame.seq_cnt = 0;
8807 	ubp->ub_frame.df_ctl = 0;
8808 	ubp->ub_frame.seq_id = 0;
8809 	ubp->ub_frame.rx_id = 0xffff;
8810 	ubp->ub_frame.ox_id = 0xffff;
8811 
8812 	/* set payload. */
8813 	payload = (la_els_logi_t *)ubp->ub_buffer;
8814 	bzero(payload, sizeof (payload));
8815 
8816 	payload->ls_code.ls_code = LA_ELS_PLOGI;
8817 	payload->common_service.fcph_version = 0x2006;
8818 	payload->common_service.cmn_features = 0x8800;
8819 
8820 	CFG_IST(ha, CFG_CTRL_24258081) ?
8821 	    (payload->common_service.rx_bufsize = CHAR_TO_SHORT(
8822 	    ha->init_ctrl_blk.cb24.max_frame_length[0],
8823 	    ha->init_ctrl_blk.cb24.max_frame_length[1])) :
8824 	    (payload->common_service.rx_bufsize = CHAR_TO_SHORT(
8825 	    ha->init_ctrl_blk.cb.max_frame_length[0],
8826 	    ha->init_ctrl_blk.cb.max_frame_length[1]));
8827 
8828 	payload->common_service.conc_sequences = 0xff;
8829 	payload->common_service.relative_offset = 0x03;
8830 	payload->common_service.e_d_tov = 0x7d0;
8831 
8832 	bcopy((void *)&tq->port_name[0],
8833 	    (void *)&payload->nport_ww_name.raw_wwn[0], 8);
8834 
8835 	bcopy((void *)&tq->node_name[0],
8836 	    (void *)&payload->node_ww_name.raw_wwn[0], 8);
8837 
8838 	class3_param = (class_svc_param_t *)&payload->class_3;
8839 	class3_param->class_valid_svc_opt = 0x8000;
8840 	class3_param->recipient_ctl = tq->class3_recipient_ctl;
8841 	class3_param->rcv_data_size = tq->class3_rcv_data_size;
8842 	class3_param->conc_sequences = tq->class3_conc_sequences;
8843 	class3_param->open_sequences_per_exch =
8844 	    tq->class3_open_sequences_per_exch;
8845 
8846 	QL_UB_LOCK(ha);
8847 	sp->flags |= SRB_UB_CALLBACK;
8848 	QL_UB_UNLOCK(ha);
8849 
8850 	ql_isp_els_handle_endian(ha, (uint8_t *)payload, LA_ELS_PLOGI);
8851 
8852 	if (done_q) {
8853 		ql_add_link_b(done_q, &sp->cmd);
8854 	} else {
8855 		ql_awaken_task_daemon(ha, sp, 0, 0);
8856 	}
8857 
8858 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
8859 
8860 	return (QL_SUCCESS);
8861 }
8862 
8863 /*
8864  * Abort outstanding commands in the Firmware, clear internally
8865  * queued commands in the driver, Synchronize the target with
8866  * the Firmware
8867  */
8868 int
ql_abort_device(ql_adapter_state_t * ha,ql_tgt_t * tq,int drain)8869 ql_abort_device(ql_adapter_state_t *ha, ql_tgt_t *tq, int drain)
8870 {
8871 	ql_link_t	*link, *link2;
8872 	ql_lun_t	*lq;
8873 	int		rval = QL_SUCCESS;
8874 	ql_srb_t	*sp;
8875 	ql_head_t	done_q = { NULL, NULL };
8876 
8877 	QL_PRINT_10(CE_CONT, "(%d,%d): started\n", ha->instance, ha->vp_index);
8878 
8879 	/*
8880 	 * First clear, internally queued commands
8881 	 */
8882 	DEVICE_QUEUE_LOCK(tq);
8883 	for (link = tq->lun_queues.first; link != NULL; link = link->next) {
8884 		lq = link->base_address;
8885 
8886 		link2 = lq->cmd.first;
8887 		while (link2 != NULL) {
8888 			sp = link2->base_address;
8889 			link2 = link2->next;
8890 
8891 			if (sp->flags & SRB_ABORT) {
8892 				continue;
8893 			}
8894 
8895 			/* Remove srb from device command queue. */
8896 			ql_remove_link(&lq->cmd, &sp->cmd);
8897 			sp->flags &= ~SRB_IN_DEVICE_QUEUE;
8898 
8899 			/* Set ending status. */
8900 			sp->pkt->pkt_reason = CS_ABORTED;
8901 
8902 			/* Call done routine to handle completions. */
8903 			ql_add_link_b(&done_q, &sp->cmd);
8904 		}
8905 	}
8906 	DEVICE_QUEUE_UNLOCK(tq);
8907 
8908 	if (done_q.first != NULL) {
8909 		ql_done(done_q.first);
8910 	}
8911 
8912 	if (drain && VALID_TARGET_ID(ha, tq->loop_id) && PD_PORT_LOGIN(tq)) {
8913 		rval = ql_abort_target(ha, tq, 0);
8914 	}
8915 
8916 	if (rval != QL_SUCCESS) {
8917 		EL(ha, "failed=%xh, d_id=%xh\n", rval, tq->d_id.b24);
8918 	} else {
8919 		/*EMPTY*/
8920 		QL_PRINT_10(CE_CONT, "(%d,%d): done\n", ha->instance,
8921 		    ha->vp_index);
8922 	}
8923 
8924 	return (rval);
8925 }
8926 
8927 /*
8928  * ql_rcv_rscn_els
8929  *	Processes received RSCN extended link service.
8930  *
8931  * Input:
8932  *	ha:	adapter state pointer.
8933  *	mb:	array containing input mailbox registers.
8934  *	done_q:	done queue pointer.
8935  *
8936  * Context:
8937  *	Interrupt or Kernel context, no mailbox commands allowed.
8938  */
8939 void
ql_rcv_rscn_els(ql_adapter_state_t * ha,uint16_t * mb,ql_head_t * done_q)8940 ql_rcv_rscn_els(ql_adapter_state_t *ha, uint16_t *mb, ql_head_t *done_q)
8941 {
8942 	fc_unsol_buf_t		*ubp;
8943 	ql_srb_t		*sp;
8944 	fc_rscn_t		*rn;
8945 	fc_affected_id_t	*af;
8946 	port_id_t		d_id;
8947 
8948 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
8949 
8950 	/* Locate a buffer to use. */
8951 	ubp = ql_get_unsolicited_buffer(ha, FC_TYPE_EXTENDED_LS);
8952 	if (ubp != NULL) {
8953 		sp = ubp->ub_fca_private;
8954 
8955 		/* Set header. */
8956 		ubp->ub_frame.d_id = ha->d_id.b24;
8957 		ubp->ub_frame.r_ctl = R_CTL_ELS_REQ;
8958 		ubp->ub_frame.s_id = FS_FABRIC_CONTROLLER;
8959 		ubp->ub_frame.rsvd = 0;
8960 		ubp->ub_frame.f_ctl = F_CTL_FIRST_SEQ | F_CTL_END_SEQ |
8961 		    F_CTL_SEQ_INITIATIVE;
8962 		ubp->ub_frame.type = FC_TYPE_EXTENDED_LS;
8963 		ubp->ub_frame.seq_cnt = 0;
8964 		ubp->ub_frame.df_ctl = 0;
8965 		ubp->ub_frame.seq_id = 0;
8966 		ubp->ub_frame.rx_id = 0xffff;
8967 		ubp->ub_frame.ox_id = 0xffff;
8968 
8969 		/* set payload. */
8970 		rn = (fc_rscn_t *)ubp->ub_buffer;
8971 		af = (fc_affected_id_t *)((caddr_t)ubp->ub_buffer + 4);
8972 
8973 		rn->rscn_code = LA_ELS_RSCN;
8974 		rn->rscn_len = 4;
8975 		rn->rscn_payload_len = 8;
8976 		d_id.b.al_pa = LSB(mb[2]);
8977 		d_id.b.area = MSB(mb[2]);
8978 		d_id.b.domain =	LSB(mb[1]);
8979 		af->aff_d_id = d_id.b24;
8980 		af->aff_format = MSB(mb[1]);
8981 
8982 		EL(ha, "LA_ELS_RSCN fmt=%xh, d_id=%xh\n", af->aff_format,
8983 		    af->aff_d_id);
8984 
8985 		ql_update_rscn(ha, af);
8986 
8987 		QL_UB_LOCK(ha);
8988 		sp->flags |= SRB_UB_CALLBACK | SRB_UB_RSCN;
8989 		QL_UB_UNLOCK(ha);
8990 		ql_add_link_b(done_q, &sp->cmd);
8991 	}
8992 
8993 	if (ubp == NULL) {
8994 		EL(ha, "Failed, get_unsolicited_buffer\n");
8995 	} else {
8996 		/*EMPTY*/
8997 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
8998 	}
8999 }
9000 
9001 /*
9002  * ql_update_rscn
9003  *	Update devices from received RSCN.
9004  *
9005  * Input:
9006  *	ha:	adapter state pointer.
9007  *	af:	pointer to RSCN data.
9008  *
9009  * Context:
9010  *	Interrupt or Kernel context, no mailbox commands allowed.
9011  */
9012 static void
ql_update_rscn(ql_adapter_state_t * ha,fc_affected_id_t * af)9013 ql_update_rscn(ql_adapter_state_t *ha, fc_affected_id_t *af)
9014 {
9015 	ql_link_t	*link;
9016 	uint16_t	index;
9017 	ql_tgt_t	*tq;
9018 
9019 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
9020 
9021 	if (af->aff_format == FC_RSCN_PORT_ADDRESS) {
9022 		port_id_t d_id;
9023 
9024 		d_id.r.rsvd_1 = 0;
9025 		d_id.b24 = af->aff_d_id;
9026 
9027 		tq = ql_d_id_to_queue(ha, d_id);
9028 		if (tq) {
9029 			EL(ha, "SD_RSCN_RCVD %xh RPA\n", d_id.b24);
9030 			DEVICE_QUEUE_LOCK(tq);
9031 			tq->flags |= TQF_RSCN_RCVD;
9032 			DEVICE_QUEUE_UNLOCK(tq);
9033 		}
9034 		QL_PRINT_3(CE_CONT, "(%d): FC_RSCN_PORT_ADDRESS done\n",
9035 		    ha->instance);
9036 
9037 		return;
9038 	}
9039 
9040 	for (index = 0; index < DEVICE_HEAD_LIST_SIZE; index++) {
9041 		for (link = ha->dev[index].first; link != NULL;
9042 		    link = link->next) {
9043 			tq = link->base_address;
9044 
9045 			switch (af->aff_format) {
9046 			case FC_RSCN_FABRIC_ADDRESS:
9047 				if (!RESERVED_LOOP_ID(ha, tq->loop_id)) {
9048 					EL(ha, "SD_RSCN_RCVD %xh RFA\n",
9049 					    tq->d_id.b24);
9050 					DEVICE_QUEUE_LOCK(tq);
9051 					tq->flags |= TQF_RSCN_RCVD;
9052 					DEVICE_QUEUE_UNLOCK(tq);
9053 				}
9054 				break;
9055 
9056 			case FC_RSCN_AREA_ADDRESS:
9057 				if ((tq->d_id.b24 & 0xffff00) == af->aff_d_id) {
9058 					EL(ha, "SD_RSCN_RCVD %xh RAA\n",
9059 					    tq->d_id.b24);
9060 					DEVICE_QUEUE_LOCK(tq);
9061 					tq->flags |= TQF_RSCN_RCVD;
9062 					DEVICE_QUEUE_UNLOCK(tq);
9063 				}
9064 				break;
9065 
9066 			case FC_RSCN_DOMAIN_ADDRESS:
9067 				if ((tq->d_id.b24 & 0xff0000) == af->aff_d_id) {
9068 					EL(ha, "SD_RSCN_RCVD %xh RDA\n",
9069 					    tq->d_id.b24);
9070 					DEVICE_QUEUE_LOCK(tq);
9071 					tq->flags |= TQF_RSCN_RCVD;
9072 					DEVICE_QUEUE_UNLOCK(tq);
9073 				}
9074 				break;
9075 
9076 			default:
9077 				break;
9078 			}
9079 		}
9080 	}
9081 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
9082 }
9083 
9084 /*
9085  * ql_process_rscn
9086  *
9087  * Input:
9088  *	ha:	adapter state pointer.
9089  *	af:	RSCN payload pointer.
9090  *
9091  * Context:
9092  *	Kernel context.
9093  */
9094 static int
ql_process_rscn(ql_adapter_state_t * ha,fc_affected_id_t * af)9095 ql_process_rscn(ql_adapter_state_t *ha, fc_affected_id_t *af)
9096 {
9097 	int		sendit;
9098 	int		sendup = 1;
9099 	ql_link_t	*link;
9100 	uint16_t	index;
9101 	ql_tgt_t	*tq;
9102 
9103 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
9104 
9105 	if (af->aff_format == FC_RSCN_PORT_ADDRESS) {
9106 		port_id_t d_id;
9107 
9108 		d_id.r.rsvd_1 = 0;
9109 		d_id.b24 = af->aff_d_id;
9110 
9111 		tq = ql_d_id_to_queue(ha, d_id);
9112 		if (tq) {
9113 			sendup = ql_process_rscn_for_device(ha, tq);
9114 		}
9115 
9116 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
9117 
9118 		return (sendup);
9119 	}
9120 
9121 	for (index = 0; index < DEVICE_HEAD_LIST_SIZE; index++) {
9122 		for (link = ha->dev[index].first; link != NULL;
9123 		    link = link->next) {
9124 
9125 			tq = link->base_address;
9126 			if (tq == NULL) {
9127 				continue;
9128 			}
9129 
9130 			switch (af->aff_format) {
9131 			case FC_RSCN_FABRIC_ADDRESS:
9132 				if (!RESERVED_LOOP_ID(ha, tq->loop_id)) {
9133 					sendit = ql_process_rscn_for_device(
9134 					    ha, tq);
9135 					if (sendup) {
9136 						sendup = sendit;
9137 					}
9138 				}
9139 				break;
9140 
9141 			case FC_RSCN_AREA_ADDRESS:
9142 				if ((tq->d_id.b24 & 0xffff00) ==
9143 				    af->aff_d_id) {
9144 					sendit = ql_process_rscn_for_device(
9145 					    ha, tq);
9146 
9147 					if (sendup) {
9148 						sendup = sendit;
9149 					}
9150 				}
9151 				break;
9152 
9153 			case FC_RSCN_DOMAIN_ADDRESS:
9154 				if ((tq->d_id.b24 & 0xff0000) ==
9155 				    af->aff_d_id) {
9156 					sendit = ql_process_rscn_for_device(
9157 					    ha, tq);
9158 
9159 					if (sendup) {
9160 						sendup = sendit;
9161 					}
9162 				}
9163 				break;
9164 
9165 			default:
9166 				break;
9167 			}
9168 		}
9169 	}
9170 
9171 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
9172 
9173 	return (sendup);
9174 }
9175 
9176 /*
9177  * ql_process_rscn_for_device
9178  *
9179  * Input:
9180  *	ha:	adapter state pointer.
9181  *	tq:	target queue pointer.
9182  *
9183  * Context:
9184  *	Kernel context.
9185  */
9186 static int
ql_process_rscn_for_device(ql_adapter_state_t * ha,ql_tgt_t * tq)9187 ql_process_rscn_for_device(ql_adapter_state_t *ha, ql_tgt_t *tq)
9188 {
9189 	int sendup = 1;
9190 
9191 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
9192 
9193 	DEVICE_QUEUE_LOCK(tq);
9194 
9195 	/*
9196 	 * Let FCP-2 compliant devices continue I/Os
9197 	 * with their low level recoveries.
9198 	 */
9199 	if (((tq->flags & TQF_INITIATOR_DEVICE) == 0) &&
9200 	    (tq->prli_svc_param_word_3 & PRLI_W3_RETRY)) {
9201 		/*
9202 		 * Cause ADISC to go out
9203 		 */
9204 		DEVICE_QUEUE_UNLOCK(tq);
9205 
9206 		(void) ql_get_port_database(ha, tq, PDF_NONE);
9207 
9208 		DEVICE_QUEUE_LOCK(tq);
9209 		tq->flags &= ~TQF_RSCN_RCVD;
9210 
9211 	} else if (tq->loop_id != PORT_NO_LOOP_ID) {
9212 		if (tq->d_id.b24 != BROADCAST_ADDR) {
9213 			tq->flags |= TQF_NEED_AUTHENTICATION;
9214 		}
9215 
9216 		DEVICE_QUEUE_UNLOCK(tq);
9217 
9218 		(void) ql_abort_device(ha, tq, 1);
9219 
9220 		DEVICE_QUEUE_LOCK(tq);
9221 
9222 		if (tq->outcnt) {
9223 			sendup = 0;
9224 		} else {
9225 			tq->flags &= ~TQF_RSCN_RCVD;
9226 		}
9227 	} else {
9228 		tq->flags &= ~TQF_RSCN_RCVD;
9229 	}
9230 
9231 	if (sendup) {
9232 		if (tq->d_id.b24 != BROADCAST_ADDR) {
9233 			tq->flags |= TQF_NEED_AUTHENTICATION;
9234 		}
9235 	}
9236 
9237 	DEVICE_QUEUE_UNLOCK(tq);
9238 
9239 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
9240 
9241 	return (sendup);
9242 }
9243 
9244 static int
ql_handle_rscn_update(ql_adapter_state_t * ha)9245 ql_handle_rscn_update(ql_adapter_state_t *ha)
9246 {
9247 	int			rval;
9248 	ql_tgt_t		*tq;
9249 	uint16_t		index, loop_id;
9250 	ql_dev_id_list_t	*list;
9251 	uint32_t		list_size;
9252 	port_id_t		d_id;
9253 	ql_mbx_data_t		mr;
9254 	ql_head_t		done_q = { NULL, NULL };
9255 
9256 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
9257 
9258 	list_size = sizeof (ql_dev_id_list_t) * DEVICE_LIST_ENTRIES;
9259 	list = kmem_zalloc(list_size, KM_SLEEP);
9260 	if (list == NULL) {
9261 		rval = QL_MEMORY_ALLOC_FAILED;
9262 		EL(ha, "kmem_zalloc failed=%xh\n", rval);
9263 		return (rval);
9264 	}
9265 
9266 	/*
9267 	 * Get data from RISC code d_id list to init each device queue.
9268 	 */
9269 	rval = ql_get_id_list(ha, (caddr_t)list, list_size, &mr);
9270 	if (rval != QL_SUCCESS) {
9271 		kmem_free(list, list_size);
9272 		EL(ha, "get_id_list failed=%xh\n", rval);
9273 		return (rval);
9274 	}
9275 
9276 	/* Acquire adapter state lock. */
9277 	ADAPTER_STATE_LOCK(ha);
9278 
9279 	/* Check for new devices */
9280 	for (index = 0; index < mr.mb[1]; index++) {
9281 		ql_dev_list(ha, list, index, &d_id, &loop_id);
9282 
9283 		if (VALID_DEVICE_ID(ha, loop_id)) {
9284 			d_id.r.rsvd_1 = 0;
9285 
9286 			tq = ql_d_id_to_queue(ha, d_id);
9287 			if (tq != NULL) {
9288 				continue;
9289 			}
9290 
9291 			tq = ql_dev_init(ha, d_id, loop_id);
9292 
9293 			/* Test for fabric device. */
9294 			if (d_id.b.domain != ha->d_id.b.domain ||
9295 			    d_id.b.area != ha->d_id.b.area) {
9296 				tq->flags |= TQF_FABRIC_DEVICE;
9297 			}
9298 
9299 			ADAPTER_STATE_UNLOCK(ha);
9300 			if (ql_get_port_database(ha, tq, PDF_NONE) !=
9301 			    QL_SUCCESS) {
9302 				tq->loop_id = PORT_NO_LOOP_ID;
9303 			}
9304 			ADAPTER_STATE_LOCK(ha);
9305 
9306 			/*
9307 			 * Send up a PLOGI about the new device
9308 			 */
9309 			if (VALID_DEVICE_ID(ha, tq->loop_id)) {
9310 				(void) ql_send_plogi(ha, tq, &done_q);
9311 			}
9312 		}
9313 	}
9314 
9315 	/* Release adapter state lock. */
9316 	ADAPTER_STATE_UNLOCK(ha);
9317 
9318 	if (done_q.first != NULL) {
9319 		ql_done(done_q.first);
9320 	}
9321 
9322 	kmem_free(list, list_size);
9323 
9324 	if (rval != QL_SUCCESS) {
9325 		EL(ha, "failed=%xh\n", rval);
9326 	} else {
9327 		/*EMPTY*/
9328 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
9329 	}
9330 
9331 	return (rval);
9332 }
9333 
9334 /*
9335  * ql_free_unsolicited_buffer
9336  *	Frees allocated buffer.
9337  *
9338  * Input:
9339  *	ha = adapter state pointer.
9340  *	index = buffer array index.
9341  *	ADAPTER_STATE_LOCK must be already obtained.
9342  *
9343  * Context:
9344  *	Kernel context.
9345  */
9346 static void
ql_free_unsolicited_buffer(ql_adapter_state_t * ha,fc_unsol_buf_t * ubp)9347 ql_free_unsolicited_buffer(ql_adapter_state_t *ha, fc_unsol_buf_t *ubp)
9348 {
9349 	ql_srb_t	*sp;
9350 	int		status;
9351 
9352 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
9353 
9354 	sp = ubp->ub_fca_private;
9355 	if (sp->ub_type == FC_TYPE_IS8802_SNAP) {
9356 		/* Disconnect IP from system buffers. */
9357 		if (ha->flags & IP_INITIALIZED) {
9358 			ADAPTER_STATE_UNLOCK(ha);
9359 			status = ql_shutdown_ip(ha);
9360 			ADAPTER_STATE_LOCK(ha);
9361 			if (status != QL_SUCCESS) {
9362 				cmn_err(CE_WARN,
9363 				    "!Qlogic %s(%d): Failed to shutdown IP",
9364 				    QL_NAME, ha->instance);
9365 				return;
9366 			}
9367 
9368 			ha->flags &= ~IP_ENABLED;
9369 		}
9370 
9371 		ql_free_phys(ha, &sp->ub_buffer);
9372 	} else {
9373 		kmem_free(ubp->ub_buffer, ubp->ub_bufsize);
9374 	}
9375 
9376 	kmem_free(sp, sizeof (ql_srb_t));
9377 	kmem_free(ubp, sizeof (fc_unsol_buf_t));
9378 
9379 	if (ha->ub_allocated != 0) {
9380 		ha->ub_allocated--;
9381 	}
9382 
9383 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
9384 }
9385 
9386 /*
9387  * ql_get_unsolicited_buffer
9388  *	Locates a free unsolicited buffer.
9389  *
9390  * Input:
9391  *	ha = adapter state pointer.
9392  *	type = buffer type.
9393  *
9394  * Returns:
9395  *	Unsolicited buffer pointer.
9396  *
9397  * Context:
9398  *	Interrupt or Kernel context, no mailbox commands allowed.
9399  */
9400 fc_unsol_buf_t *
ql_get_unsolicited_buffer(ql_adapter_state_t * ha,uint32_t type)9401 ql_get_unsolicited_buffer(ql_adapter_state_t *ha, uint32_t type)
9402 {
9403 	fc_unsol_buf_t	*ubp;
9404 	ql_srb_t	*sp;
9405 	uint16_t	index;
9406 
9407 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
9408 
9409 	/* Locate a buffer to use. */
9410 	ubp = NULL;
9411 
9412 	QL_UB_LOCK(ha);
9413 	for (index = 0; index < QL_UB_LIMIT; index++) {
9414 		ubp = ha->ub_array[index];
9415 		if (ubp != NULL) {
9416 			sp = ubp->ub_fca_private;
9417 			if ((sp->ub_type == type) &&
9418 			    (sp->flags & SRB_UB_IN_FCA) &&
9419 			    (!(sp->flags & (SRB_UB_CALLBACK |
9420 			    SRB_UB_FREE_REQUESTED | SRB_UB_ACQUIRED)))) {
9421 				sp->flags |= SRB_UB_ACQUIRED;
9422 				ubp->ub_resp_flags = 0;
9423 				break;
9424 			}
9425 			ubp = NULL;
9426 		}
9427 	}
9428 	QL_UB_UNLOCK(ha);
9429 
9430 	if (ubp) {
9431 		ubp->ub_resp_token = NULL;
9432 		ubp->ub_class = FC_TRAN_CLASS3;
9433 	}
9434 
9435 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
9436 
9437 	return (ubp);
9438 }
9439 
9440 /*
9441  * ql_ub_frame_hdr
9442  *	Processes received unsolicited buffers from ISP.
9443  *
9444  * Input:
9445  *	ha:	adapter state pointer.
9446  *	tq:	target queue pointer.
9447  *	index:	unsolicited buffer array index.
9448  *	done_q:	done queue pointer.
9449  *
9450  * Returns:
9451  *	ql local function return status code.
9452  *
9453  * Context:
9454  *	Interrupt or Kernel context, no mailbox commands allowed.
9455  */
9456 int
ql_ub_frame_hdr(ql_adapter_state_t * ha,ql_tgt_t * tq,uint16_t index,ql_head_t * done_q)9457 ql_ub_frame_hdr(ql_adapter_state_t *ha, ql_tgt_t *tq, uint16_t index,
9458     ql_head_t *done_q)
9459 {
9460 	fc_unsol_buf_t	*ubp;
9461 	ql_srb_t	*sp;
9462 	uint16_t	loop_id;
9463 	int		rval = QL_FUNCTION_FAILED;
9464 
9465 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
9466 
9467 	QL_UB_LOCK(ha);
9468 	if (index >= QL_UB_LIMIT || (ubp = ha->ub_array[index]) == NULL) {
9469 		EL(ha, "Invalid buffer index=%xh\n", index);
9470 		QL_UB_UNLOCK(ha);
9471 		return (rval);
9472 	}
9473 
9474 	sp = ubp->ub_fca_private;
9475 	if (sp->flags & SRB_UB_FREE_REQUESTED) {
9476 		EL(ha, "buffer freed index=%xh\n", index);
9477 		sp->flags &= ~(SRB_UB_IN_ISP | SRB_UB_CALLBACK |
9478 		    SRB_UB_RSCN | SRB_UB_FCP | SRB_UB_ACQUIRED);
9479 
9480 		sp->flags |= SRB_UB_IN_FCA;
9481 
9482 		QL_UB_UNLOCK(ha);
9483 		return (rval);
9484 	}
9485 
9486 	if ((sp->handle == index) &&
9487 	    (sp->flags & SRB_UB_IN_ISP) &&
9488 	    (sp->ub_type == FC_TYPE_IS8802_SNAP) &&
9489 	    (!(sp->flags & SRB_UB_ACQUIRED))) {
9490 		/* set broadcast D_ID */
9491 		loop_id = (uint16_t)(CFG_IST(ha, CFG_CTRL_24258081) ?
9492 		    BROADCAST_24XX_HDL : IP_BROADCAST_LOOP_ID);
9493 		if (tq->ub_loop_id == loop_id) {
9494 			if (ha->topology & QL_FL_PORT) {
9495 				ubp->ub_frame.d_id = 0x000000;
9496 			} else {
9497 				ubp->ub_frame.d_id = 0xffffff;
9498 			}
9499 		} else {
9500 			ubp->ub_frame.d_id = ha->d_id.b24;
9501 		}
9502 		ubp->ub_frame.r_ctl = R_CTL_UNSOL_DATA;
9503 		ubp->ub_frame.rsvd = 0;
9504 		ubp->ub_frame.s_id = tq->d_id.b24;
9505 		ubp->ub_frame.type = FC_TYPE_IS8802_SNAP;
9506 		ubp->ub_frame.seq_cnt = tq->ub_seq_cnt;
9507 		ubp->ub_frame.df_ctl = 0;
9508 		ubp->ub_frame.seq_id = tq->ub_seq_id;
9509 		ubp->ub_frame.rx_id = 0xffff;
9510 		ubp->ub_frame.ox_id = 0xffff;
9511 		ubp->ub_bufsize = sp->ub_size < tq->ub_sequence_length ?
9512 		    sp->ub_size : tq->ub_sequence_length;
9513 		ubp->ub_frame.ro = tq->ub_frame_ro;
9514 
9515 		tq->ub_sequence_length = (uint16_t)
9516 		    (tq->ub_sequence_length - ubp->ub_bufsize);
9517 		tq->ub_frame_ro += ubp->ub_bufsize;
9518 		tq->ub_seq_cnt++;
9519 
9520 		if (tq->ub_seq_cnt == tq->ub_total_seg_cnt) {
9521 			if (tq->ub_seq_cnt == 1) {
9522 				ubp->ub_frame.f_ctl = F_CTL_RO_PRESENT |
9523 				    F_CTL_FIRST_SEQ | F_CTL_END_SEQ;
9524 			} else {
9525 				ubp->ub_frame.f_ctl = F_CTL_RO_PRESENT |
9526 				    F_CTL_END_SEQ;
9527 			}
9528 			tq->ub_total_seg_cnt = 0;
9529 		} else if (tq->ub_seq_cnt == 1) {
9530 			ubp->ub_frame.f_ctl = F_CTL_RO_PRESENT |
9531 			    F_CTL_FIRST_SEQ;
9532 			ubp->ub_frame.df_ctl = 0x20;
9533 		}
9534 
9535 		QL_PRINT_3(CE_CONT, "(%d): ub_frame.d_id=%xh\n",
9536 		    ha->instance, ubp->ub_frame.d_id);
9537 		QL_PRINT_3(CE_CONT, "(%d): ub_frame.s_id=%xh\n",
9538 		    ha->instance, ubp->ub_frame.s_id);
9539 		QL_PRINT_3(CE_CONT, "(%d): ub_frame.seq_cnt=%xh\n",
9540 		    ha->instance, ubp->ub_frame.seq_cnt);
9541 		QL_PRINT_3(CE_CONT, "(%d): ub_frame.seq_id=%xh\n",
9542 		    ha->instance, ubp->ub_frame.seq_id);
9543 		QL_PRINT_3(CE_CONT, "(%d): ub_frame.ro=%xh\n",
9544 		    ha->instance, ubp->ub_frame.ro);
9545 		QL_PRINT_3(CE_CONT, "(%d): ub_frame.f_ctl=%xh\n",
9546 		    ha->instance, ubp->ub_frame.f_ctl);
9547 		QL_PRINT_3(CE_CONT, "(%d): ub_bufsize=%xh\n",
9548 		    ha->instance, ubp->ub_bufsize);
9549 		QL_DUMP_3(ubp->ub_buffer, 8,
9550 		    ubp->ub_bufsize < 64 ? ubp->ub_bufsize : 64);
9551 
9552 		sp->flags |= SRB_UB_CALLBACK | SRB_UB_ACQUIRED;
9553 		ql_add_link_b(done_q, &sp->cmd);
9554 		rval = QL_SUCCESS;
9555 	} else {
9556 		if (sp->handle != index) {
9557 			EL(ha, "Bad index=%xh, expect=%xh\n", index,
9558 			    sp->handle);
9559 		}
9560 		if ((sp->flags & SRB_UB_IN_ISP) == 0) {
9561 			EL(ha, "buffer was already in driver, index=%xh\n",
9562 			    index);
9563 		}
9564 		if ((sp->ub_type == FC_TYPE_IS8802_SNAP) == 0) {
9565 			EL(ha, "buffer was not an IP buffer, index=%xh\n",
9566 			    index);
9567 		}
9568 		if (sp->flags & SRB_UB_ACQUIRED) {
9569 			EL(ha, "buffer was being used by driver, index=%xh\n",
9570 			    index);
9571 		}
9572 	}
9573 	QL_UB_UNLOCK(ha);
9574 
9575 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
9576 
9577 	return (rval);
9578 }
9579 
9580 /*
9581  * ql_timer
9582  *	One second timer function.
9583  *
9584  * Input:
9585  *	ql_hba.first = first link in adapter list.
9586  *
9587  * Context:
9588  *	Interrupt context, no mailbox commands allowed.
9589  */
9590 static void
ql_timer(void * arg)9591 ql_timer(void *arg)
9592 {
9593 	ql_link_t		*link;
9594 	uint32_t		set_flags;
9595 	uint32_t		reset_flags;
9596 	ql_adapter_state_t	*ha = NULL, *vha;
9597 
9598 	QL_PRINT_6(CE_CONT, "started\n");
9599 
9600 	/* Acquire global state lock. */
9601 	GLOBAL_STATE_LOCK();
9602 	if (ql_timer_timeout_id == NULL) {
9603 		/* Release global state lock. */
9604 		GLOBAL_STATE_UNLOCK();
9605 		return;
9606 	}
9607 
9608 	for (link = ql_hba.first; link != NULL; link = link->next) {
9609 		ha = link->base_address;
9610 
9611 		/* Skip adapter if suspended of stalled. */
9612 		ADAPTER_STATE_LOCK(ha);
9613 		if (ha->flags & ADAPTER_SUSPENDED ||
9614 		    ha->task_daemon_flags & DRIVER_STALL) {
9615 			ADAPTER_STATE_UNLOCK(ha);
9616 			continue;
9617 		}
9618 		ha->flags |= ADAPTER_TIMER_BUSY;
9619 		ADAPTER_STATE_UNLOCK(ha);
9620 
9621 		QL_PM_LOCK(ha);
9622 		if (ha->power_level != PM_LEVEL_D0) {
9623 			QL_PM_UNLOCK(ha);
9624 
9625 			ADAPTER_STATE_LOCK(ha);
9626 			ha->flags &= ~ADAPTER_TIMER_BUSY;
9627 			ADAPTER_STATE_UNLOCK(ha);
9628 			continue;
9629 		}
9630 		ha->busy++;
9631 		QL_PM_UNLOCK(ha);
9632 
9633 		set_flags = 0;
9634 		reset_flags = 0;
9635 
9636 		/* Port retry timer handler. */
9637 		if (LOOP_READY(ha)) {
9638 			ADAPTER_STATE_LOCK(ha);
9639 			if (ha->port_retry_timer != 0) {
9640 				ha->port_retry_timer--;
9641 				if (ha->port_retry_timer == 0) {
9642 					set_flags |= PORT_RETRY_NEEDED;
9643 				}
9644 			}
9645 			ADAPTER_STATE_UNLOCK(ha);
9646 		}
9647 
9648 		/* Loop down timer handler. */
9649 		if (LOOP_RECONFIGURE(ha) == 0) {
9650 			if (ha->loop_down_timer > LOOP_DOWN_TIMER_END) {
9651 				ha->loop_down_timer--;
9652 				/*
9653 				 * give the firmware loop down dump flag
9654 				 * a chance to work.
9655 				 */
9656 				if (ha->loop_down_timer == LOOP_DOWN_RESET) {
9657 					if (CFG_IST(ha,
9658 					    CFG_DUMP_LOOP_OFFLINE_TIMEOUT)) {
9659 						(void) ql_binary_fw_dump(ha,
9660 						    TRUE);
9661 					}
9662 					EL(ha, "loop_down_reset, "
9663 					    "isp_abort_needed\n");
9664 					set_flags |= ISP_ABORT_NEEDED;
9665 				}
9666 			}
9667 			if (CFG_IST(ha, CFG_ENABLE_LINK_DOWN_REPORTING)) {
9668 				/* Command abort time handler. */
9669 				if (ha->loop_down_timer ==
9670 				    ha->loop_down_abort_time) {
9671 					ADAPTER_STATE_LOCK(ha);
9672 					ha->flags |= ABORT_CMDS_LOOP_DOWN_TMO;
9673 					ADAPTER_STATE_UNLOCK(ha);
9674 					set_flags |= ABORT_QUEUES_NEEDED;
9675 					EL(ha, "loop_down_abort_time, "
9676 					    "abort_queues_needed\n");
9677 				}
9678 
9679 				/* Watchdog timer handler. */
9680 				if (ha->watchdog_timer == 0) {
9681 					ha->watchdog_timer = WATCHDOG_TIME;
9682 				} else if (LOOP_READY(ha)) {
9683 					ha->watchdog_timer--;
9684 					if (ha->watchdog_timer == 0) {
9685 						for (vha = ha; vha != NULL;
9686 						    vha = vha->vp_next) {
9687 							ql_watchdog(vha,
9688 							    &set_flags,
9689 							    &reset_flags);
9690 						}
9691 						ha->watchdog_timer =
9692 						    WATCHDOG_TIME;
9693 					}
9694 				}
9695 			}
9696 		}
9697 
9698 		/* Idle timer handler. */
9699 		if (!DRIVER_SUSPENDED(ha)) {
9700 			if (++ha->idle_timer >= IDLE_CHECK_TIMER) {
9701 #if defined(QL_DEBUG_LEVEL_6) || !defined(QL_DEBUG_LEVEL_3)
9702 				set_flags |= TASK_DAEMON_IDLE_CHK_FLG;
9703 #endif
9704 				ha->idle_timer = 0;
9705 			}
9706 			if (ha->send_plogi_timer != NULL) {
9707 				ha->send_plogi_timer--;
9708 				if (ha->send_plogi_timer == NULL) {
9709 					set_flags |= SEND_PLOGI;
9710 				}
9711 			}
9712 		}
9713 		ADAPTER_STATE_LOCK(ha);
9714 		if (ha->idc_restart_timer != 0) {
9715 			ha->idc_restart_timer--;
9716 			if (ha->idc_restart_timer == 0) {
9717 				ha->idc_restart_cnt = 0;
9718 				reset_flags |= DRIVER_STALL;
9719 			}
9720 		}
9721 		if (ha->idc_flash_acc_timer != 0) {
9722 			ha->idc_flash_acc_timer--;
9723 			if (ha->idc_flash_acc_timer == 0 &&
9724 			    ha->idc_flash_acc != 0) {
9725 				ha->idc_flash_acc = 1;
9726 				ha->idc_mb[0] = MBA_IDC_NOTIFICATION;
9727 				ha->idc_mb[1] = 0;
9728 				ha->idc_mb[2] = IDC_OPC_DRV_START;
9729 				set_flags |= IDC_EVENT;
9730 			}
9731 		}
9732 		ADAPTER_STATE_UNLOCK(ha);
9733 
9734 		if (set_flags != 0 || reset_flags != 0) {
9735 			ql_awaken_task_daemon(ha, NULL, set_flags,
9736 			    reset_flags);
9737 		}
9738 
9739 		if (ha->xioctl->ledstate.BeaconState == BEACON_ON) {
9740 			ql_blink_led(ha);
9741 		}
9742 
9743 		/* Update the IO stats */
9744 		if (ha->xioctl->IOInputByteCnt >= 0x100000) {
9745 			ha->xioctl->IOInputMByteCnt +=
9746 			    (ha->xioctl->IOInputByteCnt / 0x100000);
9747 			ha->xioctl->IOInputByteCnt %= 0x100000;
9748 		}
9749 
9750 		if (ha->xioctl->IOOutputByteCnt >= 0x100000) {
9751 			ha->xioctl->IOOutputMByteCnt +=
9752 			    (ha->xioctl->IOOutputByteCnt / 0x100000);
9753 			ha->xioctl->IOOutputByteCnt %= 0x100000;
9754 		}
9755 
9756 		if (CFG_IST(ha, CFG_CTRL_8021)) {
9757 			(void) ql_8021_idc_handler(ha);
9758 		}
9759 
9760 		ADAPTER_STATE_LOCK(ha);
9761 		ha->flags &= ~ADAPTER_TIMER_BUSY;
9762 		ADAPTER_STATE_UNLOCK(ha);
9763 
9764 		QL_PM_LOCK(ha);
9765 		ha->busy--;
9766 		QL_PM_UNLOCK(ha);
9767 	}
9768 
9769 	/* Restart timer, if not being stopped. */
9770 	if (ql_timer_timeout_id != NULL) {
9771 		ql_timer_timeout_id = timeout(ql_timer, arg, ql_timer_ticks);
9772 	}
9773 
9774 	/* Release global state lock. */
9775 	GLOBAL_STATE_UNLOCK();
9776 
9777 	QL_PRINT_6(CE_CONT, "done\n");
9778 }
9779 
9780 /*
9781  * ql_timeout_insert
9782  *	Function used to insert a command block onto the
9783  *	watchdog timer queue.
9784  *
9785  *	Note: Must insure that pkt_time is not zero
9786  *			before calling ql_timeout_insert.
9787  *
9788  * Input:
9789  *	ha:	adapter state pointer.
9790  *	tq:	target queue pointer.
9791  *	sp:	SRB pointer.
9792  *	DEVICE_QUEUE_LOCK must be already obtained.
9793  *
9794  * Context:
9795  *	Kernel context.
9796  */
9797 /* ARGSUSED */
9798 static void
ql_timeout_insert(ql_adapter_state_t * ha,ql_tgt_t * tq,ql_srb_t * sp)9799 ql_timeout_insert(ql_adapter_state_t *ha, ql_tgt_t *tq, ql_srb_t *sp)
9800 {
9801 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
9802 
9803 	if (sp->pkt->pkt_timeout != 0 && sp->pkt->pkt_timeout < 0x10000) {
9804 		sp->isp_timeout = (uint16_t)(sp->pkt->pkt_timeout);
9805 		/*
9806 		 * The WATCHDOG_TIME must be rounded up + 1.  As an example,
9807 		 * consider a 1 second timeout. If the WATCHDOG_TIME is 1, it
9808 		 * will expire in the next watchdog call, which could be in
9809 		 * 1 microsecond.
9810 		 *
9811 		 */
9812 		sp->wdg_q_time = (sp->isp_timeout + WATCHDOG_TIME - 1) /
9813 		    WATCHDOG_TIME;
9814 		/*
9815 		 * Added an additional 10 to account for the
9816 		 * firmware timer drift which can occur with
9817 		 * very long timeout values.
9818 		 */
9819 		sp->wdg_q_time += 10;
9820 
9821 		/*
9822 		 * Add 6 more to insure watchdog does not timeout at the same
9823 		 * time as ISP RISC code timeout.
9824 		 */
9825 		sp->wdg_q_time += 6;
9826 
9827 		/* Save initial time for resetting watchdog time. */
9828 		sp->init_wdg_q_time = sp->wdg_q_time;
9829 
9830 		/* Insert command onto watchdog queue. */
9831 		ql_add_link_b(&tq->wdg, &sp->wdg);
9832 
9833 		sp->flags |= SRB_WATCHDOG_ENABLED;
9834 	} else {
9835 		sp->isp_timeout = 0;
9836 		sp->wdg_q_time = 0;
9837 		sp->init_wdg_q_time = 0;
9838 	}
9839 
9840 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
9841 }
9842 
9843 /*
9844  * ql_watchdog
9845  *	Timeout handler that runs in interrupt context. The
9846  *	ql_adapter_state_t * argument is the parameter set up when the
9847  *	timeout was initialized (state structure pointer).
9848  *	Function used to update timeout values and if timeout
9849  *	has occurred command will be aborted.
9850  *
9851  * Input:
9852  *	ha:		adapter state pointer.
9853  *	set_flags:	task daemon flags to set.
9854  *	reset_flags:	task daemon flags to reset.
9855  *
9856  * Context:
9857  *	Interrupt context, no mailbox commands allowed.
9858  */
9859 static void
ql_watchdog(ql_adapter_state_t * ha,uint32_t * set_flags,uint32_t * reset_flags)9860 ql_watchdog(ql_adapter_state_t *ha, uint32_t *set_flags, uint32_t *reset_flags)
9861 {
9862 	ql_srb_t	*sp;
9863 	ql_link_t	*link;
9864 	ql_link_t	*next_cmd;
9865 	ql_link_t	*next_device;
9866 	ql_tgt_t	*tq;
9867 	ql_lun_t	*lq;
9868 	uint16_t	index;
9869 	int		q_sane;
9870 
9871 	QL_PRINT_6(CE_CONT, "(%d): started\n", ha->instance);
9872 
9873 	/* Loop through all targets. */
9874 	for (index = 0; index < DEVICE_HEAD_LIST_SIZE; index++) {
9875 		for (link = ha->dev[index].first; link != NULL;
9876 		    link = next_device) {
9877 			tq = link->base_address;
9878 
9879 			/* Try to acquire device queue lock. */
9880 			if (TRY_DEVICE_QUEUE_LOCK(tq) == 0) {
9881 				next_device = NULL;
9882 				continue;
9883 			}
9884 
9885 			next_device = link->next;
9886 
9887 			if (!(CFG_IST(ha, CFG_ENABLE_LINK_DOWN_REPORTING)) &&
9888 			    (tq->port_down_retry_count == 0)) {
9889 				/* Release device queue lock. */
9890 				DEVICE_QUEUE_UNLOCK(tq);
9891 				continue;
9892 			}
9893 
9894 			/* Find out if this device is in a sane state. */
9895 			if (tq->flags & (TQF_RSCN_RCVD |
9896 			    TQF_NEED_AUTHENTICATION | TQF_QUEUE_SUSPENDED)) {
9897 				q_sane = 0;
9898 			} else {
9899 				q_sane = 1;
9900 			}
9901 			/* Loop through commands on watchdog queue. */
9902 			for (link = tq->wdg.first; link != NULL;
9903 			    link = next_cmd) {
9904 				next_cmd = link->next;
9905 				sp = link->base_address;
9906 				lq = sp->lun_queue;
9907 
9908 				/*
9909 				 * For SCSI commands, if everything seems to
9910 				 * be going fine and this packet is stuck
9911 				 * because of throttling at LUN or target
9912 				 * level then do not decrement the
9913 				 * sp->wdg_q_time
9914 				 */
9915 				if (ha->task_daemon_flags & STATE_ONLINE &&
9916 				    (sp->flags & SRB_ISP_STARTED) == 0 &&
9917 				    q_sane && sp->flags & SRB_FCP_CMD_PKT &&
9918 				    lq->lun_outcnt >= ha->execution_throttle) {
9919 					continue;
9920 				}
9921 
9922 				if (sp->wdg_q_time != 0) {
9923 					sp->wdg_q_time--;
9924 
9925 					/* Timeout? */
9926 					if (sp->wdg_q_time != 0) {
9927 						continue;
9928 					}
9929 
9930 					ql_remove_link(&tq->wdg, &sp->wdg);
9931 					sp->flags &= ~SRB_WATCHDOG_ENABLED;
9932 
9933 					if (sp->flags & SRB_ISP_STARTED) {
9934 						ql_cmd_timeout(ha, tq, sp,
9935 						    set_flags, reset_flags);
9936 
9937 						DEVICE_QUEUE_UNLOCK(tq);
9938 						tq = NULL;
9939 						next_cmd = NULL;
9940 						next_device = NULL;
9941 						index = DEVICE_HEAD_LIST_SIZE;
9942 					} else {
9943 						ql_cmd_timeout(ha, tq, sp,
9944 						    set_flags, reset_flags);
9945 					}
9946 				}
9947 			}
9948 
9949 			/* Release device queue lock. */
9950 			if (tq != NULL) {
9951 				DEVICE_QUEUE_UNLOCK(tq);
9952 			}
9953 		}
9954 	}
9955 
9956 	QL_PRINT_6(CE_CONT, "(%d): done\n", ha->instance);
9957 }
9958 
9959 /*
9960  * ql_cmd_timeout
9961  *	Command timeout handler.
9962  *
9963  * Input:
9964  *	ha:		adapter state pointer.
9965  *	tq:		target queue pointer.
9966  *	sp:		SRB pointer.
9967  *	set_flags:	task daemon flags to set.
9968  *	reset_flags:	task daemon flags to reset.
9969  *
9970  * Context:
9971  *	Interrupt context, no mailbox commands allowed.
9972  */
9973 /* ARGSUSED */
9974 static void
ql_cmd_timeout(ql_adapter_state_t * ha,ql_tgt_t * tq,ql_srb_t * sp,uint32_t * set_flags,uint32_t * reset_flags)9975 ql_cmd_timeout(ql_adapter_state_t *ha, ql_tgt_t *tq, ql_srb_t *sp,
9976     uint32_t *set_flags, uint32_t *reset_flags)
9977 {
9978 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
9979 
9980 	if (!(sp->flags & SRB_ISP_STARTED)) {
9981 
9982 		EL(ha, "command timed out in driver = %ph\n", (void *)sp);
9983 
9984 		REQUEST_RING_LOCK(ha);
9985 
9986 		/* if it's on a queue */
9987 		if (sp->cmd.head) {
9988 			/*
9989 			 * The pending_cmds que needs to be
9990 			 * protected by the ring lock
9991 			 */
9992 			ql_remove_link(sp->cmd.head, &sp->cmd);
9993 		}
9994 		sp->flags &= ~SRB_IN_DEVICE_QUEUE;
9995 
9996 		/* Release device queue lock. */
9997 		REQUEST_RING_UNLOCK(ha);
9998 		DEVICE_QUEUE_UNLOCK(tq);
9999 
10000 		/* Set timeout status */
10001 		sp->pkt->pkt_reason = CS_TIMEOUT;
10002 
10003 		/* Ensure no retry */
10004 		sp->flags &= ~SRB_RETRY;
10005 
10006 		/* Call done routine to handle completion. */
10007 		ql_done(&sp->cmd);
10008 
10009 		DEVICE_QUEUE_LOCK(tq);
10010 	} else if (CFG_IST(ha, CFG_CTRL_8021)) {
10011 		int		rval;
10012 		uint32_t	index;
10013 
10014 		EL(ha, "command timed out in isp=%ph, osc=%ph, index=%xh, "
10015 		    "spf=%xh\n", (void *)sp,
10016 		    (void *)ha->outstanding_cmds[sp->handle & OSC_INDEX_MASK],
10017 		    sp->handle & OSC_INDEX_MASK, sp->flags);
10018 
10019 		DEVICE_QUEUE_UNLOCK(tq);
10020 
10021 		INTR_LOCK(ha);
10022 		ha->pha->xioctl->ControllerErrorCount++;
10023 		if (sp->handle) {
10024 			ha->pha->timeout_cnt++;
10025 			index = sp->handle & OSC_INDEX_MASK;
10026 			if (ha->pha->outstanding_cmds[index] == sp) {
10027 				sp->request_ring_ptr->entry_type =
10028 				    INVALID_ENTRY_TYPE;
10029 				sp->request_ring_ptr->entry_count = 0;
10030 				ha->pha->outstanding_cmds[index] = 0;
10031 			}
10032 			INTR_UNLOCK(ha);
10033 
10034 			rval = ql_abort_command(ha, sp);
10035 			if (rval == QL_FUNCTION_TIMEOUT ||
10036 			    rval == QL_LOCK_TIMEOUT ||
10037 			    rval == QL_FUNCTION_PARAMETER_ERROR ||
10038 			    ha->pha->timeout_cnt > TIMEOUT_THRESHOLD) {
10039 				*set_flags |= ISP_ABORT_NEEDED;
10040 				EL(ha, "abort status=%xh, tc=%xh, isp_abort_"
10041 				    "needed\n", rval, ha->pha->timeout_cnt);
10042 			}
10043 
10044 			sp->handle = 0;
10045 			sp->flags &= ~SRB_IN_TOKEN_ARRAY;
10046 		} else {
10047 			INTR_UNLOCK(ha);
10048 		}
10049 
10050 		/* Set timeout status */
10051 		sp->pkt->pkt_reason = CS_TIMEOUT;
10052 
10053 		/* Ensure no retry */
10054 		sp->flags &= ~SRB_RETRY;
10055 
10056 		/* Call done routine to handle completion. */
10057 		ql_done(&sp->cmd);
10058 
10059 		DEVICE_QUEUE_LOCK(tq);
10060 
10061 	} else {
10062 		EL(ha, "command timed out in isp=%ph, osc=%ph, index=%xh, "
10063 		    "spf=%xh, isp_abort_needed\n", (void *)sp,
10064 		    (void *)ha->outstanding_cmds[sp->handle & OSC_INDEX_MASK],
10065 		    sp->handle & OSC_INDEX_MASK, sp->flags);
10066 
10067 		/* Release device queue lock. */
10068 		DEVICE_QUEUE_UNLOCK(tq);
10069 
10070 		INTR_LOCK(ha);
10071 		ha->pha->xioctl->ControllerErrorCount++;
10072 		INTR_UNLOCK(ha);
10073 
10074 		/* Set ISP needs to be reset */
10075 		sp->flags |= SRB_COMMAND_TIMEOUT;
10076 
10077 		if (CFG_IST(ha, CFG_DUMP_DRIVER_COMMAND_TIMEOUT)) {
10078 			(void) ql_binary_fw_dump(ha, TRUE);
10079 		}
10080 
10081 		*set_flags |= ISP_ABORT_NEEDED;
10082 
10083 		DEVICE_QUEUE_LOCK(tq);
10084 	}
10085 
10086 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
10087 }
10088 
10089 /*
10090  * ql_rst_aen
10091  *	Processes asynchronous reset.
10092  *
10093  * Input:
10094  *	ha = adapter state pointer.
10095  *
10096  * Context:
10097  *	Kernel context.
10098  */
10099 static void
ql_rst_aen(ql_adapter_state_t * ha)10100 ql_rst_aen(ql_adapter_state_t *ha)
10101 {
10102 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
10103 
10104 	/* Issue marker command. */
10105 	(void) ql_marker(ha, 0, 0, MK_SYNC_ALL);
10106 
10107 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
10108 }
10109 
10110 /*
10111  * ql_cmd_wait
10112  *	Stall driver until all outstanding commands are returned.
10113  *
10114  * Input:
10115  *	ha = adapter state pointer.
10116  *
10117  * Context:
10118  *	Kernel context.
10119  */
10120 void
ql_cmd_wait(ql_adapter_state_t * ha)10121 ql_cmd_wait(ql_adapter_state_t *ha)
10122 {
10123 	uint16_t		index;
10124 	ql_link_t		*link;
10125 	ql_tgt_t		*tq;
10126 	ql_adapter_state_t	*vha;
10127 
10128 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
10129 
10130 	/* Wait for all outstanding commands to be returned. */
10131 	(void) ql_wait_outstanding(ha);
10132 
10133 	/*
10134 	 * clear out internally queued commands
10135 	 */
10136 	for (vha = ha; vha != NULL; vha = vha->vp_next) {
10137 		for (index = 0; index < DEVICE_HEAD_LIST_SIZE; index++) {
10138 			for (link = vha->dev[index].first; link != NULL;
10139 			    link = link->next) {
10140 				tq = link->base_address;
10141 				if (tq &&
10142 				    (!(tq->prli_svc_param_word_3 &
10143 				    PRLI_W3_RETRY))) {
10144 					(void) ql_abort_device(vha, tq, 0);
10145 				}
10146 			}
10147 		}
10148 	}
10149 
10150 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
10151 }
10152 
10153 /*
10154  * ql_wait_outstanding
10155  *	Wait for all outstanding commands to complete.
10156  *
10157  * Input:
10158  *	ha = adapter state pointer.
10159  *
10160  * Returns:
10161  *	index - the index for ql_srb into outstanding_cmds.
10162  *
10163  * Context:
10164  *	Kernel context.
10165  */
10166 static uint16_t
ql_wait_outstanding(ql_adapter_state_t * ha)10167 ql_wait_outstanding(ql_adapter_state_t *ha)
10168 {
10169 	ql_srb_t	*sp;
10170 	uint16_t	index, count;
10171 
10172 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
10173 
10174 	count = ql_osc_wait_count;
10175 	for (index = 1; index < MAX_OUTSTANDING_COMMANDS; index++) {
10176 		if (ha->pha->pending_cmds.first != NULL) {
10177 			ql_start_iocb(ha, NULL);
10178 			index = 1;
10179 		}
10180 		if ((sp = ha->pha->outstanding_cmds[index]) != NULL &&
10181 		    (sp->flags & SRB_COMMAND_TIMEOUT) == 0) {
10182 			if (count-- != 0) {
10183 				ql_delay(ha, 10000);
10184 				index = 0;
10185 			} else {
10186 				EL(ha, "failed, sp=%ph, oci=%d, hdl=%xh\n",
10187 				    (void *)sp, index, sp->handle);
10188 				break;
10189 			}
10190 		}
10191 	}
10192 
10193 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
10194 
10195 	return (index);
10196 }
10197 
10198 /*
10199  * ql_restart_queues
10200  *	Restart device queues.
10201  *
10202  * Input:
10203  *	ha = adapter state pointer.
10204  *	DEVICE_QUEUE_LOCK must be released.
10205  *
10206  * Context:
10207  *	Interrupt or Kernel context, no mailbox commands allowed.
10208  */
10209 static void
ql_restart_queues(ql_adapter_state_t * ha)10210 ql_restart_queues(ql_adapter_state_t *ha)
10211 {
10212 	ql_link_t		*link, *link2;
10213 	ql_tgt_t		*tq;
10214 	ql_lun_t		*lq;
10215 	uint16_t		index;
10216 	ql_adapter_state_t	*vha;
10217 
10218 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
10219 
10220 	for (vha = ha->pha; vha != NULL; vha = vha->vp_next) {
10221 		for (index = 0; index < DEVICE_HEAD_LIST_SIZE; index++) {
10222 			for (link = vha->dev[index].first; link != NULL;
10223 			    link = link->next) {
10224 				tq = link->base_address;
10225 
10226 				/* Acquire device queue lock. */
10227 				DEVICE_QUEUE_LOCK(tq);
10228 
10229 				tq->flags &= ~TQF_QUEUE_SUSPENDED;
10230 
10231 				for (link2 = tq->lun_queues.first;
10232 				    link2 != NULL; link2 = link2->next) {
10233 					lq = link2->base_address;
10234 
10235 					if (lq->cmd.first != NULL) {
10236 						ql_next(vha, lq);
10237 						DEVICE_QUEUE_LOCK(tq);
10238 					}
10239 				}
10240 
10241 				/* Release device queue lock. */
10242 				DEVICE_QUEUE_UNLOCK(tq);
10243 			}
10244 		}
10245 	}
10246 
10247 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
10248 }
10249 
10250 /*
10251  * ql_iidma
10252  *	Setup iiDMA parameters to firmware
10253  *
10254  * Input:
10255  *	ha = adapter state pointer.
10256  *	DEVICE_QUEUE_LOCK must be released.
10257  *
10258  * Context:
10259  *	Interrupt or Kernel context, no mailbox commands allowed.
10260  */
10261 static void
ql_iidma(ql_adapter_state_t * ha)10262 ql_iidma(ql_adapter_state_t *ha)
10263 {
10264 	ql_link_t	*link;
10265 	ql_tgt_t	*tq;
10266 	uint16_t	index;
10267 	char		buf[256];
10268 	uint32_t	data;
10269 
10270 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
10271 
10272 	if ((CFG_IST(ha, CFG_CTRL_242581)) == 0) {
10273 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
10274 		return;
10275 	}
10276 
10277 	for (index = 0; index < DEVICE_HEAD_LIST_SIZE; index++) {
10278 		for (link = ha->dev[index].first; link != NULL;
10279 		    link = link->next) {
10280 			tq = link->base_address;
10281 
10282 			/* Acquire device queue lock. */
10283 			DEVICE_QUEUE_LOCK(tq);
10284 
10285 			if ((tq->flags & TQF_IIDMA_NEEDED) == 0) {
10286 				DEVICE_QUEUE_UNLOCK(tq);
10287 				continue;
10288 			}
10289 
10290 			tq->flags &= ~TQF_IIDMA_NEEDED;
10291 
10292 			if ((tq->loop_id > LAST_N_PORT_HDL) ||
10293 			    (tq->iidma_rate == IIDMA_RATE_NDEF)) {
10294 				DEVICE_QUEUE_UNLOCK(tq);
10295 				continue;
10296 			}
10297 
10298 			/* Get the iiDMA persistent data */
10299 			if (tq->iidma_rate == IIDMA_RATE_INIT) {
10300 				(void) sprintf(buf,
10301 				    "iidma-rate-%02x%02x%02x%02x%02x"
10302 				    "%02x%02x%02x", tq->port_name[0],
10303 				    tq->port_name[1], tq->port_name[2],
10304 				    tq->port_name[3], tq->port_name[4],
10305 				    tq->port_name[5], tq->port_name[6],
10306 				    tq->port_name[7]);
10307 
10308 				if ((data = ql_get_prop(ha, buf)) ==
10309 				    0xffffffff) {
10310 					tq->iidma_rate = IIDMA_RATE_NDEF;
10311 				} else {
10312 					switch (data) {
10313 					case IIDMA_RATE_1GB:
10314 					case IIDMA_RATE_2GB:
10315 					case IIDMA_RATE_4GB:
10316 					case IIDMA_RATE_10GB:
10317 						tq->iidma_rate = data;
10318 						break;
10319 					case IIDMA_RATE_8GB:
10320 						if (CFG_IST(ha,
10321 						    CFG_CTRL_25XX)) {
10322 							tq->iidma_rate = data;
10323 						} else {
10324 							tq->iidma_rate =
10325 							    IIDMA_RATE_4GB;
10326 						}
10327 						break;
10328 					default:
10329 						EL(ha, "invalid data for "
10330 						    "parameter: %s: %xh\n",
10331 						    buf, data);
10332 						tq->iidma_rate =
10333 						    IIDMA_RATE_NDEF;
10334 						break;
10335 					}
10336 				}
10337 			}
10338 
10339 			/* Set the firmware's iiDMA rate */
10340 			if (tq->iidma_rate <= IIDMA_RATE_MAX &&
10341 			    !(CFG_IST(ha, CFG_CTRL_8081))) {
10342 				data = ql_iidma_rate(ha, tq->loop_id,
10343 				    &tq->iidma_rate, EXT_IIDMA_MODE_SET);
10344 				if (data != QL_SUCCESS) {
10345 					EL(ha, "mbx failed: %xh\n", data);
10346 				}
10347 			}
10348 
10349 			/* Release device queue lock. */
10350 			DEVICE_QUEUE_UNLOCK(tq);
10351 		}
10352 	}
10353 
10354 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
10355 }
10356 
10357 /*
10358  * ql_abort_queues
10359  *	Abort all commands on device queues.
10360  *
10361  * Input:
10362  *	ha = adapter state pointer.
10363  *
10364  * Context:
10365  *	Interrupt or Kernel context, no mailbox commands allowed.
10366  */
10367 static void
ql_abort_queues(ql_adapter_state_t * ha)10368 ql_abort_queues(ql_adapter_state_t *ha)
10369 {
10370 	ql_link_t		*link;
10371 	ql_tgt_t		*tq;
10372 	ql_srb_t		*sp;
10373 	uint16_t		index;
10374 	ql_adapter_state_t	*vha;
10375 
10376 	QL_PRINT_10(CE_CONT, "(%d): started\n", ha->instance);
10377 
10378 	/* Return all commands in outstanding command list. */
10379 	INTR_LOCK(ha);
10380 
10381 	/* Place all commands in outstanding cmd list on device queue. */
10382 	for (index = 1; index < MAX_OUTSTANDING_COMMANDS; index++) {
10383 		if (ha->pending_cmds.first != NULL) {
10384 			INTR_UNLOCK(ha);
10385 			ql_start_iocb(ha, NULL);
10386 			/* Delay for system */
10387 			ql_delay(ha, 10000);
10388 			INTR_LOCK(ha);
10389 			index = 1;
10390 		}
10391 		sp = ha->outstanding_cmds[index];
10392 
10393 		/* skip devices capable of FCP2 retrys */
10394 		if ((sp != NULL) &&
10395 		    ((tq = sp->lun_queue->target_queue) != NULL) &&
10396 		    (!(tq->prli_svc_param_word_3 & PRLI_W3_RETRY))) {
10397 			ha->outstanding_cmds[index] = NULL;
10398 			sp->handle = 0;
10399 			sp->flags &= ~SRB_IN_TOKEN_ARRAY;
10400 
10401 			INTR_UNLOCK(ha);
10402 
10403 			/* Set ending status. */
10404 			sp->pkt->pkt_reason = CS_PORT_UNAVAILABLE;
10405 			sp->flags |= SRB_ISP_COMPLETED;
10406 
10407 			/* Call done routine to handle completions. */
10408 			sp->cmd.next = NULL;
10409 			ql_done(&sp->cmd);
10410 
10411 			INTR_LOCK(ha);
10412 		}
10413 	}
10414 	INTR_UNLOCK(ha);
10415 
10416 	for (vha = ha; vha != NULL; vha = vha->vp_next) {
10417 		QL_PRINT_10(CE_CONT, "(%d,%d): abort instance\n",
10418 		    vha->instance, vha->vp_index);
10419 		for (index = 0; index < DEVICE_HEAD_LIST_SIZE; index++) {
10420 			for (link = vha->dev[index].first; link != NULL;
10421 			    link = link->next) {
10422 				tq = link->base_address;
10423 				/* skip devices capable of FCP2 retrys */
10424 				if (!(tq->prli_svc_param_word_3 &
10425 				    PRLI_W3_RETRY)) {
10426 					/*
10427 					 * Set port unavailable status and
10428 					 * return all commands on a devices
10429 					 * queues.
10430 					 */
10431 					ql_abort_device_queues(ha, tq);
10432 				}
10433 			}
10434 		}
10435 	}
10436 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
10437 }
10438 
10439 /*
10440  * ql_abort_device_queues
10441  *	Abort all commands on device queues.
10442  *
10443  * Input:
10444  *	ha = adapter state pointer.
10445  *
10446  * Context:
10447  *	Interrupt or Kernel context, no mailbox commands allowed.
10448  */
10449 static void
ql_abort_device_queues(ql_adapter_state_t * ha,ql_tgt_t * tq)10450 ql_abort_device_queues(ql_adapter_state_t *ha, ql_tgt_t *tq)
10451 {
10452 	ql_link_t	*lun_link, *cmd_link;
10453 	ql_srb_t	*sp;
10454 	ql_lun_t	*lq;
10455 
10456 	QL_PRINT_10(CE_CONT, "(%d): started\n", ha->instance);
10457 
10458 	DEVICE_QUEUE_LOCK(tq);
10459 
10460 	for (lun_link = tq->lun_queues.first; lun_link != NULL;
10461 	    lun_link = lun_link->next) {
10462 		lq = lun_link->base_address;
10463 
10464 		cmd_link = lq->cmd.first;
10465 		while (cmd_link != NULL) {
10466 			sp = cmd_link->base_address;
10467 
10468 			if (sp->flags & SRB_ABORT) {
10469 				cmd_link = cmd_link->next;
10470 				continue;
10471 			}
10472 
10473 			/* Remove srb from device cmd queue. */
10474 			ql_remove_link(&lq->cmd, &sp->cmd);
10475 
10476 			sp->flags &= ~SRB_IN_DEVICE_QUEUE;
10477 
10478 			DEVICE_QUEUE_UNLOCK(tq);
10479 
10480 			/* Set ending status. */
10481 			sp->pkt->pkt_reason = CS_PORT_UNAVAILABLE;
10482 
10483 			/* Call done routine to handle completion. */
10484 			ql_done(&sp->cmd);
10485 
10486 			/* Delay for system */
10487 			ql_delay(ha, 10000);
10488 
10489 			DEVICE_QUEUE_LOCK(tq);
10490 			cmd_link = lq->cmd.first;
10491 		}
10492 	}
10493 	DEVICE_QUEUE_UNLOCK(tq);
10494 
10495 	QL_PRINT_10(CE_CONT, "(%d): done\n", ha->instance);
10496 }
10497 
10498 /*
10499  * ql_loop_resync
10500  *	Resync with fibre channel devices.
10501  *
10502  * Input:
10503  *	ha = adapter state pointer.
10504  *	DEVICE_QUEUE_LOCK must be released.
10505  *
10506  * Returns:
10507  *	ql local function return status code.
10508  *
10509  * Context:
10510  *	Kernel context.
10511  */
10512 static int
ql_loop_resync(ql_adapter_state_t * ha)10513 ql_loop_resync(ql_adapter_state_t *ha)
10514 {
10515 	int rval;
10516 
10517 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
10518 
10519 	if (ha->flags & IP_INITIALIZED) {
10520 		(void) ql_shutdown_ip(ha);
10521 	}
10522 
10523 	rval = ql_fw_ready(ha, 10);
10524 
10525 	TASK_DAEMON_LOCK(ha);
10526 	ha->task_daemon_flags &= ~LOOP_RESYNC_ACTIVE;
10527 	TASK_DAEMON_UNLOCK(ha);
10528 
10529 	/* Set loop online, if it really is. */
10530 	if (rval == QL_SUCCESS) {
10531 		ql_loop_online(ha);
10532 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
10533 	} else {
10534 		EL(ha, "failed, rval = %xh\n", rval);
10535 	}
10536 
10537 	return (rval);
10538 }
10539 
10540 /*
10541  * ql_loop_online
10542  *	Set loop online status if it really is online.
10543  *
10544  * Input:
10545  *	ha = adapter state pointer.
10546  *	DEVICE_QUEUE_LOCK must be released.
10547  *
10548  * Context:
10549  *	Kernel context.
10550  */
10551 void
ql_loop_online(ql_adapter_state_t * ha)10552 ql_loop_online(ql_adapter_state_t *ha)
10553 {
10554 	ql_adapter_state_t	*vha;
10555 
10556 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
10557 
10558 	/* Inform the FC Transport that the hardware is online. */
10559 	for (vha = ha->pha; vha != NULL; vha = vha->vp_next) {
10560 		if (!(vha->task_daemon_flags &
10561 		    (LOOP_RESYNC_NEEDED | LOOP_DOWN))) {
10562 			/* Restart IP if it was shutdown. */
10563 			if (vha->vp_index == 0 && vha->flags & IP_ENABLED &&
10564 			    !(vha->flags & IP_INITIALIZED)) {
10565 				(void) ql_initialize_ip(vha);
10566 				ql_isp_rcvbuf(vha);
10567 			}
10568 
10569 			if (FC_PORT_STATE_MASK(vha->state) != FC_STATE_LOOP &&
10570 			    FC_PORT_STATE_MASK(vha->state) !=
10571 			    FC_STATE_ONLINE) {
10572 				vha->state = FC_PORT_SPEED_MASK(vha->state);
10573 				if (vha->topology & QL_LOOP_CONNECTION) {
10574 					vha->state |= FC_STATE_LOOP;
10575 				} else {
10576 					vha->state |= FC_STATE_ONLINE;
10577 				}
10578 				TASK_DAEMON_LOCK(ha);
10579 				vha->task_daemon_flags |= FC_STATE_CHANGE;
10580 				TASK_DAEMON_UNLOCK(ha);
10581 			}
10582 		}
10583 	}
10584 
10585 	ql_awaken_task_daemon(ha, NULL, 0, 0);
10586 
10587 	/* Restart device queues that may have been stopped. */
10588 	ql_restart_queues(ha);
10589 
10590 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
10591 }
10592 
10593 /*
10594  * ql_fca_handle_to_state
10595  *	Verifies handle to be correct.
10596  *
10597  * Input:
10598  *	fca_handle = pointer to state structure.
10599  *
10600  * Returns:
10601  *	NULL = failure
10602  *
10603  * Context:
10604  *	Kernel context.
10605  */
10606 static ql_adapter_state_t *
ql_fca_handle_to_state(opaque_t fca_handle)10607 ql_fca_handle_to_state(opaque_t fca_handle)
10608 {
10609 #ifdef	QL_DEBUG_ROUTINES
10610 	ql_link_t		*link;
10611 	ql_adapter_state_t	*ha = NULL;
10612 	ql_adapter_state_t	*vha = NULL;
10613 
10614 	for (link = ql_hba.first; link != NULL; link = link->next) {
10615 		ha = link->base_address;
10616 		for (vha = ha->vp_next; vha != NULL; vha = vha->vp_next) {
10617 			if ((opaque_t)vha == fca_handle) {
10618 				ha = vha;
10619 				break;
10620 			}
10621 		}
10622 		if ((opaque_t)ha == fca_handle) {
10623 			break;
10624 		} else {
10625 			ha = NULL;
10626 		}
10627 	}
10628 
10629 	if (ha == NULL) {
10630 		/*EMPTY*/
10631 		QL_PRINT_2(CE_CONT, "failed\n");
10632 	}
10633 
10634 #endif /* QL_DEBUG_ROUTINES */
10635 
10636 	return ((ql_adapter_state_t *)fca_handle);
10637 }
10638 
10639 /*
10640  * ql_d_id_to_queue
10641  *	Locate device queue that matches destination ID.
10642  *
10643  * Input:
10644  *	ha = adapter state pointer.
10645  *	d_id = destination ID
10646  *
10647  * Returns:
10648  *	NULL = failure
10649  *
10650  * Context:
10651  *	Interrupt or Kernel context, no mailbox commands allowed.
10652  */
10653 ql_tgt_t *
ql_d_id_to_queue(ql_adapter_state_t * ha,port_id_t d_id)10654 ql_d_id_to_queue(ql_adapter_state_t *ha, port_id_t d_id)
10655 {
10656 	uint16_t	index;
10657 	ql_tgt_t	*tq;
10658 	ql_link_t	*link;
10659 
10660 	/* Get head queue index. */
10661 	index = ql_alpa_to_index[d_id.b.al_pa];
10662 
10663 	for (link = ha->dev[index].first; link != NULL; link = link->next) {
10664 		tq = link->base_address;
10665 		if (tq->d_id.b24 == d_id.b24 &&
10666 		    VALID_DEVICE_ID(ha, tq->loop_id)) {
10667 			return (tq);
10668 		}
10669 	}
10670 
10671 	return (NULL);
10672 }
10673 
10674 /*
10675  * ql_loop_id_to_queue
10676  *	Locate device queue that matches loop ID.
10677  *
10678  * Input:
10679  *	ha:		adapter state pointer.
10680  *	loop_id:	destination ID
10681  *
10682  * Returns:
10683  *	NULL = failure
10684  *
10685  * Context:
10686  *	Interrupt or Kernel context, no mailbox commands allowed.
10687  */
10688 ql_tgt_t *
ql_loop_id_to_queue(ql_adapter_state_t * ha,uint16_t loop_id)10689 ql_loop_id_to_queue(ql_adapter_state_t *ha, uint16_t loop_id)
10690 {
10691 	uint16_t	index;
10692 	ql_tgt_t	*tq;
10693 	ql_link_t	*link;
10694 
10695 	for (index = 0; index < DEVICE_HEAD_LIST_SIZE; index++) {
10696 		for (link = ha->dev[index].first; link != NULL;
10697 		    link = link->next) {
10698 			tq = link->base_address;
10699 			if (tq->loop_id == loop_id) {
10700 				return (tq);
10701 			}
10702 		}
10703 	}
10704 
10705 	return (NULL);
10706 }
10707 
10708 /*
10709  * ql_kstat_update
10710  *	Updates kernel statistics.
10711  *
10712  * Input:
10713  *	ksp - driver kernel statistics structure pointer.
10714  *	rw - function to perform
10715  *
10716  * Returns:
10717  *	0 or EACCES
10718  *
10719  * Context:
10720  *	Kernel context.
10721  */
10722 /* ARGSUSED */
10723 static int
ql_kstat_update(kstat_t * ksp,int rw)10724 ql_kstat_update(kstat_t *ksp, int rw)
10725 {
10726 	int			rval;
10727 
10728 	QL_PRINT_3(CE_CONT, "started\n");
10729 
10730 	if (rw == KSTAT_WRITE) {
10731 		rval = EACCES;
10732 	} else {
10733 		rval = 0;
10734 	}
10735 
10736 	if (rval != 0) {
10737 		/*EMPTY*/
10738 		QL_PRINT_2(CE_CONT, "failed, rval = %xh\n", rval);
10739 	} else {
10740 		/*EMPTY*/
10741 		QL_PRINT_3(CE_CONT, "done\n");
10742 	}
10743 	return (rval);
10744 }
10745 
10746 /*
10747  * ql_load_flash
10748  *	Loads flash.
10749  *
10750  * Input:
10751  *	ha:	adapter state pointer.
10752  *	dp:	data pointer.
10753  *	size:	data length.
10754  *
10755  * Returns:
10756  *	ql local function return status code.
10757  *
10758  * Context:
10759  *	Kernel context.
10760  */
10761 int
ql_load_flash(ql_adapter_state_t * ha,uint8_t * dp,uint32_t size)10762 ql_load_flash(ql_adapter_state_t *ha, uint8_t *dp, uint32_t size)
10763 {
10764 	uint32_t	cnt;
10765 	int		rval;
10766 	uint32_t	size_to_offset;
10767 	uint32_t	size_to_compare;
10768 	int		erase_all;
10769 
10770 	if (CFG_IST(ha, CFG_CTRL_24258081)) {
10771 		return (ql_24xx_load_flash(ha, dp, size, 0));
10772 	}
10773 
10774 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
10775 
10776 	size_to_compare = 0x20000;
10777 	size_to_offset = 0;
10778 	erase_all = 0;
10779 	if (CFG_IST(ha, CFG_SBUS_CARD)) {
10780 		if (size == 0x80000) {
10781 			/* Request to flash the entire chip. */
10782 			size_to_compare = 0x80000;
10783 			erase_all = 1;
10784 		} else {
10785 			size_to_compare = 0x40000;
10786 			if (ql_flash_sbus_fpga) {
10787 				size_to_offset = 0x40000;
10788 			}
10789 		}
10790 	}
10791 	if (size > size_to_compare) {
10792 		rval = QL_FUNCTION_PARAMETER_ERROR;
10793 		EL(ha, "failed=%xh\n", rval);
10794 		return (rval);
10795 	}
10796 
10797 	GLOBAL_HW_LOCK();
10798 
10799 	/* Enable Flash Read/Write. */
10800 	ql_flash_enable(ha);
10801 
10802 	/* Erase flash prior to write. */
10803 	rval = ql_erase_flash(ha, erase_all);
10804 
10805 	if (rval == QL_SUCCESS) {
10806 		/* Write data to flash. */
10807 		for (cnt = 0; cnt < size; cnt++) {
10808 			/* Allow other system activity. */
10809 			if (cnt % 0x1000 == 0) {
10810 				ql_delay(ha, 10000);
10811 			}
10812 			rval = ql_program_flash_address(ha,
10813 			    cnt + size_to_offset, *dp++);
10814 			if (rval != QL_SUCCESS) {
10815 				break;
10816 			}
10817 		}
10818 	}
10819 
10820 	ql_flash_disable(ha);
10821 
10822 	GLOBAL_HW_UNLOCK();
10823 
10824 	if (rval != QL_SUCCESS) {
10825 		EL(ha, "failed=%xh\n", rval);
10826 	} else {
10827 		/*EMPTY*/
10828 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
10829 	}
10830 	return (rval);
10831 }
10832 
10833 /*
10834  * ql_program_flash_address
10835  *	Program flash address.
10836  *
10837  * Input:
10838  *	ha = adapter state pointer.
10839  *	addr = flash byte address.
10840  *	data = data to be written to flash.
10841  *
10842  * Returns:
10843  *	ql local function return status code.
10844  *
10845  * Context:
10846  *	Kernel context.
10847  */
10848 static int
ql_program_flash_address(ql_adapter_state_t * ha,uint32_t addr,uint8_t data)10849 ql_program_flash_address(ql_adapter_state_t *ha, uint32_t addr, uint8_t data)
10850 {
10851 	int rval;
10852 
10853 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
10854 
10855 	if (CFG_IST(ha, CFG_SBUS_CARD)) {
10856 		ql_write_flash_byte(ha, 0x5555, 0xa0);
10857 		ql_write_flash_byte(ha, addr, data);
10858 	} else {
10859 		/* Write Program Command Sequence */
10860 		ql_write_flash_byte(ha, 0x5555, 0xaa);
10861 		ql_write_flash_byte(ha, 0x2aaa, 0x55);
10862 		ql_write_flash_byte(ha, 0x5555, 0xa0);
10863 		ql_write_flash_byte(ha, addr, data);
10864 	}
10865 
10866 	/* Wait for write to complete. */
10867 	rval = ql_poll_flash(ha, addr, data);
10868 
10869 	if (rval != QL_SUCCESS) {
10870 		EL(ha, "failed=%xh\n", rval);
10871 	} else {
10872 		/*EMPTY*/
10873 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
10874 	}
10875 	return (rval);
10876 }
10877 
10878 /*
10879  * ql_erase_flash
10880  *	Erases entire flash.
10881  *
10882  * Input:
10883  *	ha = adapter state pointer.
10884  *
10885  * Returns:
10886  *	ql local function return status code.
10887  *
10888  * Context:
10889  *	Kernel context.
10890  */
10891 int
ql_erase_flash(ql_adapter_state_t * ha,int erase_all)10892 ql_erase_flash(ql_adapter_state_t *ha, int erase_all)
10893 {
10894 	int		rval;
10895 	uint32_t	erase_delay = 2000000;
10896 	uint32_t	sStartAddr;
10897 	uint32_t	ssize;
10898 	uint32_t	cnt;
10899 	uint8_t		*bfp;
10900 	uint8_t		*tmp;
10901 
10902 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
10903 
10904 	if ((CFG_IST(ha, CFG_SBUS_CARD)) && !erase_all) {
10905 
10906 		if (ql_flash_sbus_fpga == 1) {
10907 			ssize = QL_SBUS_FCODE_SIZE;
10908 			sStartAddr = QL_FCODE_OFFSET;
10909 		} else {
10910 			ssize = QL_FPGA_SIZE;
10911 			sStartAddr = QL_FPGA_OFFSET;
10912 		}
10913 
10914 		erase_delay = 20000000;
10915 
10916 		bfp = (uint8_t *)kmem_zalloc(ssize, KM_SLEEP);
10917 
10918 		/* Save the section of flash we're not updating to buffer */
10919 		tmp = bfp;
10920 		for (cnt = sStartAddr; cnt < ssize+sStartAddr; cnt++) {
10921 			/* Allow other system activity. */
10922 			if (cnt % 0x1000 == 0) {
10923 				ql_delay(ha, 10000);
10924 			}
10925 			*tmp++ = (uint8_t)ql_read_flash_byte(ha, cnt);
10926 		}
10927 	}
10928 
10929 	/* Chip Erase Command Sequence */
10930 	ql_write_flash_byte(ha, 0x5555, 0xaa);
10931 	ql_write_flash_byte(ha, 0x2aaa, 0x55);
10932 	ql_write_flash_byte(ha, 0x5555, 0x80);
10933 	ql_write_flash_byte(ha, 0x5555, 0xaa);
10934 	ql_write_flash_byte(ha, 0x2aaa, 0x55);
10935 	ql_write_flash_byte(ha, 0x5555, 0x10);
10936 
10937 	ql_delay(ha, erase_delay);
10938 
10939 	/* Wait for erase to complete. */
10940 	rval = ql_poll_flash(ha, 0, 0x80);
10941 
10942 	if (rval != QL_SUCCESS) {
10943 		EL(ha, "failed=%xh\n", rval);
10944 		if (CFG_IST(ha, CFG_SBUS_CARD)) {
10945 			kmem_free(bfp, ssize);
10946 		}
10947 		return (rval);
10948 	}
10949 
10950 	/* restore the section we saved in the buffer */
10951 	if ((CFG_IST(ha, CFG_SBUS_CARD)) && !erase_all) {
10952 		/* Restore the section we saved off */
10953 		tmp = bfp;
10954 		for (cnt = sStartAddr; cnt < ssize+sStartAddr; cnt++) {
10955 			/* Allow other system activity. */
10956 			if (cnt % 0x1000 == 0) {
10957 				ql_delay(ha, 10000);
10958 			}
10959 			rval = ql_program_flash_address(ha, cnt, *tmp++);
10960 			if (rval != QL_SUCCESS) {
10961 				break;
10962 			}
10963 		}
10964 
10965 		kmem_free(bfp, ssize);
10966 	}
10967 
10968 	if (rval != QL_SUCCESS) {
10969 		EL(ha, "failed=%xh\n", rval);
10970 	} else {
10971 		/*EMPTY*/
10972 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
10973 	}
10974 	return (rval);
10975 }
10976 
10977 /*
10978  * ql_poll_flash
10979  *	Polls flash for completion.
10980  *
10981  * Input:
10982  *	ha = adapter state pointer.
10983  *	addr = flash byte address.
10984  *	data = data to be polled.
10985  *
10986  * Returns:
10987  *	ql local function return status code.
10988  *
10989  * Context:
10990  *	Kernel context.
10991  */
10992 int
ql_poll_flash(ql_adapter_state_t * ha,uint32_t addr,uint8_t poll_data)10993 ql_poll_flash(ql_adapter_state_t *ha, uint32_t addr, uint8_t poll_data)
10994 {
10995 	uint8_t		flash_data;
10996 	uint32_t	cnt;
10997 	int		rval = QL_FUNCTION_FAILED;
10998 
10999 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
11000 
11001 	poll_data = (uint8_t)(poll_data & BIT_7);
11002 
11003 	/* Wait for 30 seconds for command to finish. */
11004 	for (cnt = 30000000; cnt; cnt--) {
11005 		flash_data = (uint8_t)ql_read_flash_byte(ha, addr);
11006 
11007 		if ((flash_data & BIT_7) == poll_data) {
11008 			rval = QL_SUCCESS;
11009 			break;
11010 		}
11011 		if (flash_data & BIT_5 && cnt > 2) {
11012 			cnt = 2;
11013 		}
11014 		drv_usecwait(1);
11015 	}
11016 
11017 	if (rval != QL_SUCCESS) {
11018 		EL(ha, "failed=%xh\n", rval);
11019 	} else {
11020 		/*EMPTY*/
11021 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
11022 	}
11023 	return (rval);
11024 }
11025 
11026 /*
11027  * ql_flash_enable
11028  *	Setup flash for reading/writing.
11029  *
11030  * Input:
11031  *	ha = adapter state pointer.
11032  *
11033  * Context:
11034  *	Kernel context.
11035  */
11036 void
ql_flash_enable(ql_adapter_state_t * ha)11037 ql_flash_enable(ql_adapter_state_t *ha)
11038 {
11039 	uint16_t	data;
11040 
11041 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
11042 
11043 	/* Enable Flash Read/Write. */
11044 	if (CFG_IST(ha, CFG_SBUS_CARD)) {
11045 		data = (uint16_t)ddi_get16(ha->sbus_fpga_dev_handle,
11046 		    (uint16_t *)(ha->sbus_fpga_iobase + FPGA_CONF));
11047 		data = (uint16_t)(data | SBUS_FLASH_WRITE_ENABLE);
11048 		ddi_put16(ha->sbus_fpga_dev_handle,
11049 		    (uint16_t *)(ha->sbus_fpga_iobase + FPGA_CONF), data);
11050 		/* Read reset command sequence */
11051 		ql_write_flash_byte(ha, 0xaaa, 0xaa);
11052 		ql_write_flash_byte(ha, 0x555, 0x55);
11053 		ql_write_flash_byte(ha, 0xaaa, 0x20);
11054 		ql_write_flash_byte(ha, 0x555, 0xf0);
11055 	} else {
11056 		data = (uint16_t)(RD16_IO_REG(ha, ctrl_status) |
11057 		    ISP_FLASH_ENABLE);
11058 		WRT16_IO_REG(ha, ctrl_status, data);
11059 
11060 		/* Read/Reset Command Sequence */
11061 		ql_write_flash_byte(ha, 0x5555, 0xaa);
11062 		ql_write_flash_byte(ha, 0x2aaa, 0x55);
11063 		ql_write_flash_byte(ha, 0x5555, 0xf0);
11064 	}
11065 	(void) ql_read_flash_byte(ha, 0);
11066 
11067 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
11068 }
11069 
11070 /*
11071  * ql_flash_disable
11072  *	Disable flash and allow RISC to run.
11073  *
11074  * Input:
11075  *	ha = adapter state pointer.
11076  *
11077  * Context:
11078  *	Kernel context.
11079  */
11080 void
ql_flash_disable(ql_adapter_state_t * ha)11081 ql_flash_disable(ql_adapter_state_t *ha)
11082 {
11083 	uint16_t	data;
11084 
11085 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
11086 
11087 	if (CFG_IST(ha, CFG_SBUS_CARD)) {
11088 		/*
11089 		 * Lock the flash back up.
11090 		 */
11091 		ql_write_flash_byte(ha, 0x555, 0x90);
11092 		ql_write_flash_byte(ha, 0x555, 0x0);
11093 
11094 		data = (uint16_t)ddi_get16(ha->sbus_fpga_dev_handle,
11095 		    (uint16_t *)(ha->sbus_fpga_iobase + FPGA_CONF));
11096 		data = (uint16_t)(data & ~SBUS_FLASH_WRITE_ENABLE);
11097 		ddi_put16(ha->sbus_fpga_dev_handle,
11098 		    (uint16_t *)(ha->sbus_fpga_iobase + FPGA_CONF), data);
11099 	} else {
11100 		data = (uint16_t)(RD16_IO_REG(ha, ctrl_status) &
11101 		    ~ISP_FLASH_ENABLE);
11102 		WRT16_IO_REG(ha, ctrl_status, data);
11103 	}
11104 
11105 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
11106 }
11107 
11108 /*
11109  * ql_write_flash_byte
11110  *	Write byte to flash.
11111  *
11112  * Input:
11113  *	ha = adapter state pointer.
11114  *	addr = flash byte address.
11115  *	data = data to be written.
11116  *
11117  * Context:
11118  *	Kernel context.
11119  */
11120 void
ql_write_flash_byte(ql_adapter_state_t * ha,uint32_t addr,uint8_t data)11121 ql_write_flash_byte(ql_adapter_state_t *ha, uint32_t addr, uint8_t data)
11122 {
11123 	if (CFG_IST(ha, CFG_SBUS_CARD)) {
11124 		ddi_put16(ha->sbus_fpga_dev_handle,
11125 		    (uint16_t *)(ha->sbus_fpga_iobase + FPGA_EEPROM_LOADDR),
11126 		    LSW(addr));
11127 		ddi_put16(ha->sbus_fpga_dev_handle,
11128 		    (uint16_t *)(ha->sbus_fpga_iobase + FPGA_EEPROM_HIADDR),
11129 		    MSW(addr));
11130 		ddi_put16(ha->sbus_fpga_dev_handle,
11131 		    (uint16_t *)(ha->sbus_fpga_iobase + FPGA_EEPROM_DATA),
11132 		    (uint16_t)data);
11133 	} else {
11134 		uint16_t bank_select;
11135 
11136 		/* Setup bit 16 of flash address. */
11137 		bank_select = (uint16_t)RD16_IO_REG(ha, ctrl_status);
11138 
11139 		if (CFG_IST(ha, CFG_CTRL_6322)) {
11140 			bank_select = (uint16_t)(bank_select & ~0xf0);
11141 			bank_select = (uint16_t)(bank_select |
11142 			    ((addr >> 12 & 0xf0) | ISP_FLASH_64K_BANK));
11143 			WRT16_IO_REG(ha, ctrl_status, bank_select);
11144 		} else {
11145 			if (addr & BIT_16 && !(bank_select &
11146 			    ISP_FLASH_64K_BANK)) {
11147 				bank_select = (uint16_t)(bank_select |
11148 				    ISP_FLASH_64K_BANK);
11149 				WRT16_IO_REG(ha, ctrl_status, bank_select);
11150 			} else if (!(addr & BIT_16) && bank_select &
11151 			    ISP_FLASH_64K_BANK) {
11152 				bank_select = (uint16_t)(bank_select &
11153 				    ~ISP_FLASH_64K_BANK);
11154 				WRT16_IO_REG(ha, ctrl_status, bank_select);
11155 			}
11156 		}
11157 
11158 		if (CFG_IST(ha, CFG_SBUS_CARD)) {
11159 			WRT16_IO_REG(ha, flash_address, (uint16_t)addr);
11160 			WRT16_IO_REG(ha, flash_data, (uint16_t)data);
11161 		} else {
11162 			WRT16_IOMAP_REG(ha, flash_address, addr);
11163 			WRT16_IOMAP_REG(ha, flash_data, data);
11164 		}
11165 	}
11166 }
11167 
11168 /*
11169  * ql_read_flash_byte
11170  *	Reads byte from flash, but must read a word from chip.
11171  *
11172  * Input:
11173  *	ha = adapter state pointer.
11174  *	addr = flash byte address.
11175  *
11176  * Returns:
11177  *	byte from flash.
11178  *
11179  * Context:
11180  *	Kernel context.
11181  */
11182 uint8_t
ql_read_flash_byte(ql_adapter_state_t * ha,uint32_t addr)11183 ql_read_flash_byte(ql_adapter_state_t *ha, uint32_t addr)
11184 {
11185 	uint8_t	data;
11186 
11187 	if (CFG_IST(ha, CFG_SBUS_CARD)) {
11188 		ddi_put16(ha->sbus_fpga_dev_handle,
11189 		    (uint16_t *)(ha->sbus_fpga_iobase + FPGA_EEPROM_LOADDR),
11190 		    LSW(addr));
11191 		ddi_put16(ha->sbus_fpga_dev_handle,
11192 		    (uint16_t *)(ha->sbus_fpga_iobase + FPGA_EEPROM_HIADDR),
11193 		    MSW(addr));
11194 		data = (uint8_t)ddi_get16(ha->sbus_fpga_dev_handle,
11195 		    (uint16_t *)(ha->sbus_fpga_iobase + FPGA_EEPROM_DATA));
11196 	} else {
11197 		uint16_t	bank_select;
11198 
11199 		/* Setup bit 16 of flash address. */
11200 		bank_select = RD16_IO_REG(ha, ctrl_status);
11201 		if (CFG_IST(ha, CFG_CTRL_6322)) {
11202 			bank_select = (uint16_t)(bank_select & ~0xf0);
11203 			bank_select = (uint16_t)(bank_select |
11204 			    ((addr >> 12 & 0xf0) | ISP_FLASH_64K_BANK));
11205 			WRT16_IO_REG(ha, ctrl_status, bank_select);
11206 		} else {
11207 			if (addr & BIT_16 &&
11208 			    !(bank_select & ISP_FLASH_64K_BANK)) {
11209 				bank_select = (uint16_t)(bank_select |
11210 				    ISP_FLASH_64K_BANK);
11211 				WRT16_IO_REG(ha, ctrl_status, bank_select);
11212 			} else if (!(addr & BIT_16) &&
11213 			    bank_select & ISP_FLASH_64K_BANK) {
11214 				bank_select = (uint16_t)(bank_select &
11215 				    ~ISP_FLASH_64K_BANK);
11216 				WRT16_IO_REG(ha, ctrl_status, bank_select);
11217 			}
11218 		}
11219 
11220 		if (CFG_IST(ha, CFG_SBUS_CARD)) {
11221 			WRT16_IO_REG(ha, flash_address, addr);
11222 			data = (uint8_t)RD16_IO_REG(ha, flash_data);
11223 		} else {
11224 			WRT16_IOMAP_REG(ha, flash_address, addr);
11225 			data = (uint8_t)RD16_IOMAP_REG(ha, flash_data);
11226 		}
11227 	}
11228 
11229 	return (data);
11230 }
11231 
11232 /*
11233  * ql_24xx_flash_id
11234  *	Get flash IDs.
11235  *
11236  * Input:
11237  *	ha:		adapter state pointer.
11238  *
11239  * Returns:
11240  *	ql local function return status code.
11241  *
11242  * Context:
11243  *	Kernel context.
11244  */
11245 int
ql_24xx_flash_id(ql_adapter_state_t * vha)11246 ql_24xx_flash_id(ql_adapter_state_t *vha)
11247 {
11248 	int			rval;
11249 	uint32_t		fdata = 0;
11250 	ql_adapter_state_t	*ha = vha->pha;
11251 	ql_xioctl_t		*xp = ha->xioctl;
11252 
11253 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
11254 
11255 	rval = ql_24xx_read_flash(ha, FLASH_CONF_ADDR | 0x3AB, &fdata);
11256 
11257 	if (rval != QL_SUCCESS || fdata == 0 || CFG_IST(ha, CFG_CTRL_2581)) {
11258 		fdata = 0;
11259 		rval = ql_24xx_read_flash(ha, FLASH_CONF_ADDR |
11260 		    (CFG_IST(ha, CFG_CTRL_2422) ? 0x39F : 0x49F), &fdata);
11261 	}
11262 
11263 	if (rval != QL_SUCCESS) {
11264 		EL(ha, "24xx read_flash failed=%xh\n", rval);
11265 	} else if (fdata != 0) {
11266 		xp->fdesc.flash_manuf = LSB(LSW(fdata));
11267 		xp->fdesc.flash_id = MSB(LSW(fdata));
11268 		xp->fdesc.flash_len = LSB(MSW(fdata));
11269 	} else {
11270 		xp->fdesc.flash_manuf = ATMEL_FLASH;
11271 		xp->fdesc.flash_id = ATMEL_FLASHID_1024K;
11272 		xp->fdesc.flash_len = 0;
11273 	}
11274 
11275 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
11276 
11277 	return (rval);
11278 }
11279 
11280 /*
11281  * ql_24xx_load_flash
11282  *	Loads flash.
11283  *
11284  * Input:
11285  *	ha = adapter state pointer.
11286  *	dp = data pointer.
11287  *	size = data length in bytes.
11288  *	faddr = 32bit word flash byte address.
11289  *
11290  * Returns:
11291  *	ql local function return status code.
11292  *
11293  * Context:
11294  *	Kernel context.
11295  */
11296 int
ql_24xx_load_flash(ql_adapter_state_t * vha,uint8_t * dp,uint32_t size,uint32_t faddr)11297 ql_24xx_load_flash(ql_adapter_state_t *vha, uint8_t *dp, uint32_t size,
11298     uint32_t faddr)
11299 {
11300 	int			rval;
11301 	uint32_t		cnt, rest_addr, fdata, wc;
11302 	dma_mem_t		dmabuf = {0};
11303 	ql_adapter_state_t	*ha = vha->pha;
11304 	ql_xioctl_t		*xp = ha->xioctl;
11305 
11306 	QL_PRINT_3(CE_CONT, "(%d): started, faddr=%xh, size=%xh\n",
11307 	    ha->instance, faddr, size);
11308 
11309 	/* start address must be 32 bit word aligned */
11310 	if ((faddr & 0x3) != 0) {
11311 		EL(ha, "incorrect buffer size alignment\n");
11312 		return (QL_FUNCTION_PARAMETER_ERROR);
11313 	}
11314 
11315 	/* Allocate DMA buffer */
11316 	if (CFG_IST(ha, CFG_CTRL_2581)) {
11317 		if ((rval = ql_get_dma_mem(ha, &dmabuf, 0xffff,
11318 		    LITTLE_ENDIAN_DMA, QL_DMA_DATA_ALIGN)) !=
11319 		    QL_SUCCESS) {
11320 			EL(ha, "dma alloc failed, rval=%xh\n", rval);
11321 			return (rval);
11322 		}
11323 	}
11324 
11325 	GLOBAL_HW_LOCK();
11326 
11327 	/* Enable flash write */
11328 	if ((rval = ql_24xx_unprotect_flash(ha)) != QL_SUCCESS) {
11329 		GLOBAL_HW_UNLOCK();
11330 		EL(ha, "unprotect_flash failed, rval=%xh\n", rval);
11331 		ql_free_phys(ha, &dmabuf);
11332 		return (rval);
11333 	}
11334 
11335 	/* setup mask of address range within a sector */
11336 	rest_addr = (xp->fdesc.block_size - 1) >> 2;
11337 
11338 	faddr = faddr >> 2;	/* flash gets 32 bit words */
11339 
11340 	/*
11341 	 * Write data to flash.
11342 	 */
11343 	cnt = 0;
11344 	size = (size + 3) >> 2;	/* Round up & convert to dwords */
11345 
11346 	while (cnt < size) {
11347 		/* Beginning of a sector? */
11348 		if ((faddr & rest_addr) == 0) {
11349 			if (CFG_IST(ha, CFG_CTRL_8021)) {
11350 				fdata = ha->flash_data_addr | faddr;
11351 				rval = ql_8021_rom_erase(ha, fdata);
11352 				if (rval != QL_SUCCESS) {
11353 					EL(ha, "8021 erase sector status="
11354 					    "%xh, start=%xh, end=%xh"
11355 					    "\n", rval, fdata,
11356 					    fdata + rest_addr);
11357 					break;
11358 				}
11359 			} else if (CFG_IST(ha, CFG_CTRL_81XX)) {
11360 				fdata = ha->flash_data_addr | faddr;
11361 				rval = ql_flash_access(ha,
11362 				    FAC_ERASE_SECTOR, fdata, fdata +
11363 				    rest_addr, 0);
11364 				if (rval != QL_SUCCESS) {
11365 					EL(ha, "erase sector status="
11366 					    "%xh, start=%xh, end=%xh"
11367 					    "\n", rval, fdata,
11368 					    fdata + rest_addr);
11369 					break;
11370 				}
11371 			} else {
11372 				fdata = (faddr & ~rest_addr) << 2;
11373 				fdata = (fdata & 0xff00) |
11374 				    (fdata << 16 & 0xff0000) |
11375 				    (fdata >> 16 & 0xff);
11376 
11377 				if (rest_addr == 0x1fff) {
11378 					/* 32kb sector block erase */
11379 					rval = ql_24xx_write_flash(ha,
11380 					    FLASH_CONF_ADDR | 0x0352,
11381 					    fdata);
11382 				} else {
11383 					/* 64kb sector block erase */
11384 					rval = ql_24xx_write_flash(ha,
11385 					    FLASH_CONF_ADDR | 0x03d8,
11386 					    fdata);
11387 				}
11388 				if (rval != QL_SUCCESS) {
11389 					EL(ha, "Unable to flash sector"
11390 					    ": address=%xh\n", faddr);
11391 					break;
11392 				}
11393 			}
11394 		}
11395 
11396 		/* Write data */
11397 		if (CFG_IST(ha, CFG_CTRL_2581) &&
11398 		    ((faddr & 0x3f) == 0)) {
11399 			/*
11400 			 * Limit write up to sector boundary.
11401 			 */
11402 			wc = ((~faddr & (rest_addr>>1)) + 1);
11403 
11404 			if (size - cnt < wc) {
11405 				wc = size - cnt;
11406 			}
11407 
11408 			ddi_rep_put8(dmabuf.acc_handle, (uint8_t *)dp,
11409 			    (uint8_t *)dmabuf.bp, wc<<2,
11410 			    DDI_DEV_AUTOINCR);
11411 
11412 			rval = ql_wrt_risc_ram(ha, ha->flash_data_addr |
11413 			    faddr, dmabuf.cookie.dmac_laddress, wc);
11414 			if (rval != QL_SUCCESS) {
11415 				EL(ha, "unable to dma to flash "
11416 				    "address=%xh\n", faddr << 2);
11417 				break;
11418 			}
11419 
11420 			cnt += wc;
11421 			faddr += wc;
11422 			dp += wc << 2;
11423 		} else {
11424 			fdata = *dp++;
11425 			fdata |= *dp++ << 8;
11426 			fdata |= *dp++ << 16;
11427 			fdata |= *dp++ << 24;
11428 			rval = ql_24xx_write_flash(ha,
11429 			    ha->flash_data_addr | faddr, fdata);
11430 			if (rval != QL_SUCCESS) {
11431 				EL(ha, "Unable to program flash "
11432 				    "address=%xh data=%xh\n", faddr,
11433 				    *dp);
11434 				break;
11435 			}
11436 			cnt++;
11437 			faddr++;
11438 
11439 			/* Allow other system activity. */
11440 			if (cnt % 0x1000 == 0) {
11441 				ql_delay(ha, 10000);
11442 			}
11443 		}
11444 	}
11445 
11446 	ql_24xx_protect_flash(ha);
11447 
11448 	ql_free_phys(ha, &dmabuf);
11449 
11450 	GLOBAL_HW_UNLOCK();
11451 
11452 	if (rval != QL_SUCCESS) {
11453 		EL(ha, "failed=%xh\n", rval);
11454 	} else {
11455 		/*EMPTY*/
11456 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
11457 	}
11458 	return (rval);
11459 }
11460 
11461 /*
11462  * ql_24xx_read_flash
11463  *	Reads a 32bit word from ISP24xx NVRAM/FLASH.
11464  *
11465  * Input:
11466  *	ha:	adapter state pointer.
11467  *	faddr:	NVRAM/FLASH address.
11468  *	bp:	data pointer.
11469  *
11470  * Returns:
11471  *	ql local function return status code.
11472  *
11473  * Context:
11474  *	Kernel context.
11475  */
11476 int
ql_24xx_read_flash(ql_adapter_state_t * vha,uint32_t faddr,uint32_t * bp)11477 ql_24xx_read_flash(ql_adapter_state_t *vha, uint32_t faddr, uint32_t *bp)
11478 {
11479 	uint32_t		timer;
11480 	int			rval = QL_SUCCESS;
11481 	ql_adapter_state_t	*ha = vha->pha;
11482 
11483 	if (CFG_IST(ha, CFG_CTRL_8021)) {
11484 		if ((rval = ql_8021_rom_read(ha, faddr, bp)) != QL_SUCCESS) {
11485 			EL(ha, "8021 access error\n");
11486 		}
11487 		return (rval);
11488 	}
11489 
11490 	/* Clear access error flag */
11491 	WRT32_IO_REG(ha, ctrl_status,
11492 	    RD32_IO_REG(ha, ctrl_status) | FLASH_NVRAM_ACCESS_ERROR);
11493 
11494 	WRT32_IO_REG(ha, flash_address, faddr & ~FLASH_DATA_FLAG);
11495 
11496 	/* Wait for READ cycle to complete. */
11497 	for (timer = 300000; timer; timer--) {
11498 		if (RD32_IO_REG(ha, flash_address) & FLASH_DATA_FLAG) {
11499 			break;
11500 		}
11501 		drv_usecwait(10);
11502 	}
11503 
11504 	if (timer == 0) {
11505 		EL(ha, "failed, timeout\n");
11506 		rval = QL_FUNCTION_TIMEOUT;
11507 	} else if (RD32_IO_REG(ha, ctrl_status) & FLASH_NVRAM_ACCESS_ERROR) {
11508 		EL(ha, "failed, access error\n");
11509 		rval = QL_FUNCTION_FAILED;
11510 	}
11511 
11512 	*bp = RD32_IO_REG(ha, flash_data);
11513 
11514 	return (rval);
11515 }
11516 
11517 /*
11518  * ql_24xx_write_flash
11519  *	Writes a 32bit word to ISP24xx NVRAM/FLASH.
11520  *
11521  * Input:
11522  *	ha:	adapter state pointer.
11523  *	addr:	NVRAM/FLASH address.
11524  *	value:	data.
11525  *
11526  * Returns:
11527  *	ql local function return status code.
11528  *
11529  * Context:
11530  *	Kernel context.
11531  */
11532 int
ql_24xx_write_flash(ql_adapter_state_t * vha,uint32_t addr,uint32_t data)11533 ql_24xx_write_flash(ql_adapter_state_t *vha, uint32_t addr, uint32_t data)
11534 {
11535 	uint32_t		timer, fdata;
11536 	int			rval = QL_SUCCESS;
11537 	ql_adapter_state_t	*ha = vha->pha;
11538 
11539 	if (CFG_IST(ha, CFG_CTRL_8021)) {
11540 		if ((rval = ql_8021_rom_write(ha, addr, data)) != QL_SUCCESS) {
11541 			EL(ha, "8021 access error\n");
11542 		}
11543 		return (rval);
11544 	}
11545 	/* Clear access error flag */
11546 	WRT32_IO_REG(ha, ctrl_status,
11547 	    RD32_IO_REG(ha, ctrl_status) | FLASH_NVRAM_ACCESS_ERROR);
11548 
11549 	WRT32_IO_REG(ha, flash_data, data);
11550 	RD32_IO_REG(ha, flash_data);		/* PCI Posting. */
11551 	WRT32_IO_REG(ha, flash_address, addr | FLASH_DATA_FLAG);
11552 
11553 	/* Wait for Write cycle to complete. */
11554 	for (timer = 3000000; timer; timer--) {
11555 		if ((RD32_IO_REG(ha, flash_address) & FLASH_DATA_FLAG) == 0) {
11556 			/* Check flash write in progress. */
11557 			if ((addr & FLASH_ADDR_MASK) == FLASH_CONF_ADDR) {
11558 				(void) ql_24xx_read_flash(ha,
11559 				    FLASH_CONF_ADDR | 0x005, &fdata);
11560 				if (!(fdata & BIT_0)) {
11561 					break;
11562 				}
11563 			} else {
11564 				break;
11565 			}
11566 		}
11567 		drv_usecwait(10);
11568 	}
11569 	if (timer == 0) {
11570 		EL(ha, "failed, timeout\n");
11571 		rval = QL_FUNCTION_TIMEOUT;
11572 	} else if (RD32_IO_REG(ha, ctrl_status) & FLASH_NVRAM_ACCESS_ERROR) {
11573 		EL(ha, "access error\n");
11574 		rval = QL_FUNCTION_FAILED;
11575 	}
11576 
11577 	return (rval);
11578 }
11579 /*
11580  * ql_24xx_unprotect_flash
11581  *	Enable writes
11582  *
11583  * Input:
11584  *	ha:	adapter state pointer.
11585  *
11586  * Returns:
11587  *	ql local function return status code.
11588  *
11589  * Context:
11590  *	Kernel context.
11591  */
11592 int
ql_24xx_unprotect_flash(ql_adapter_state_t * vha)11593 ql_24xx_unprotect_flash(ql_adapter_state_t *vha)
11594 {
11595 	int			rval;
11596 	uint32_t		fdata;
11597 	ql_adapter_state_t	*ha = vha->pha;
11598 	ql_xioctl_t		*xp = ha->xioctl;
11599 
11600 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
11601 
11602 	if (CFG_IST(ha, CFG_CTRL_8021)) {
11603 		(void) ql_8021_rom_wrsr(ha, xp->fdesc.write_enable_bits);
11604 		rval = ql_8021_rom_wrsr(ha, xp->fdesc.write_enable_bits);
11605 		if (rval != QL_SUCCESS) {
11606 			EL(ha, "8021 access error\n");
11607 		}
11608 		return (rval);
11609 	}
11610 	if (CFG_IST(ha, CFG_CTRL_81XX)) {
11611 		if (ha->task_daemon_flags & FIRMWARE_UP) {
11612 			if ((rval = ql_flash_access(ha, FAC_WRT_ENABLE, 0, 0,
11613 			    0)) != QL_SUCCESS) {
11614 				EL(ha, "status=%xh\n", rval);
11615 			}
11616 			QL_PRINT_3(CE_CONT, "(%d): 8100 done\n",
11617 			    ha->instance);
11618 			return (rval);
11619 		}
11620 	} else {
11621 		/* Enable flash write. */
11622 		WRT32_IO_REG(ha, ctrl_status,
11623 		    RD32_IO_REG(ha, ctrl_status) | ISP_FLASH_ENABLE);
11624 		RD32_IO_REG(ha, ctrl_status);	/* PCI Posting. */
11625 	}
11626 
11627 	/*
11628 	 * Remove block write protection (SST and ST) and
11629 	 * Sector/Block Protection Register Lock (SST, ST, ATMEL).
11630 	 * Unprotect sectors.
11631 	 */
11632 	(void) ql_24xx_write_flash(ha, FLASH_CONF_ADDR | 0x100 |
11633 	    xp->fdesc.write_statusreg_cmd, xp->fdesc.write_enable_bits);
11634 
11635 	if (xp->fdesc.unprotect_sector_cmd != 0) {
11636 		for (fdata = 0; fdata < 0x10; fdata++) {
11637 			(void) ql_24xx_write_flash(ha, FLASH_CONF_ADDR |
11638 			    0x300 | xp->fdesc.unprotect_sector_cmd, fdata);
11639 		}
11640 
11641 		(void) ql_24xx_write_flash(ha, FLASH_CONF_ADDR | 0x300 |
11642 		    xp->fdesc.unprotect_sector_cmd, 0x00400f);
11643 		(void) ql_24xx_write_flash(ha, FLASH_CONF_ADDR | 0x300 |
11644 		    xp->fdesc.unprotect_sector_cmd, 0x00600f);
11645 		(void) ql_24xx_write_flash(ha, FLASH_CONF_ADDR | 0x300 |
11646 		    xp->fdesc.unprotect_sector_cmd, 0x00800f);
11647 	}
11648 
11649 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
11650 
11651 	return (QL_SUCCESS);
11652 }
11653 
11654 /*
11655  * ql_24xx_protect_flash
11656  *	Disable writes
11657  *
11658  * Input:
11659  *	ha:	adapter state pointer.
11660  *
11661  * Context:
11662  *	Kernel context.
11663  */
11664 void
ql_24xx_protect_flash(ql_adapter_state_t * vha)11665 ql_24xx_protect_flash(ql_adapter_state_t *vha)
11666 {
11667 	int			rval;
11668 	uint32_t		fdata;
11669 	ql_adapter_state_t	*ha = vha->pha;
11670 	ql_xioctl_t		*xp = ha->xioctl;
11671 
11672 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
11673 
11674 	if (CFG_IST(ha, CFG_CTRL_8021)) {
11675 		(void) ql_8021_rom_wrsr(ha, xp->fdesc.write_enable_bits);
11676 		rval = ql_8021_rom_wrsr(ha, xp->fdesc.write_disable_bits);
11677 		if (rval != QL_SUCCESS) {
11678 			EL(ha, "8021 access error\n");
11679 		}
11680 		return;
11681 	}
11682 	if (CFG_IST(ha, CFG_CTRL_81XX)) {
11683 		if (ha->task_daemon_flags & FIRMWARE_UP) {
11684 			if ((rval = ql_flash_access(ha, FAC_WRT_PROTECT, 0, 0,
11685 			    0)) != QL_SUCCESS) {
11686 				EL(ha, "status=%xh\n", rval);
11687 			}
11688 			QL_PRINT_3(CE_CONT, "(%d): 8100 done\n",
11689 			    ha->instance);
11690 			return;
11691 		}
11692 	} else {
11693 		/* Enable flash write. */
11694 		WRT32_IO_REG(ha, ctrl_status,
11695 		    RD32_IO_REG(ha, ctrl_status) | ISP_FLASH_ENABLE);
11696 		RD32_IO_REG(ha, ctrl_status);	/* PCI Posting. */
11697 	}
11698 
11699 	/*
11700 	 * Protect sectors.
11701 	 * Set block write protection (SST and ST) and
11702 	 * Sector/Block Protection Register Lock (SST, ST, ATMEL).
11703 	 */
11704 	if (xp->fdesc.protect_sector_cmd != 0) {
11705 		for (fdata = 0; fdata < 0x10; fdata++) {
11706 			(void) ql_24xx_write_flash(ha, FLASH_CONF_ADDR |
11707 			    0x330 | xp->fdesc.protect_sector_cmd, fdata);
11708 		}
11709 		(void) ql_24xx_write_flash(ha, FLASH_CONF_ADDR | 0x330 |
11710 		    xp->fdesc.protect_sector_cmd, 0x00400f);
11711 		(void) ql_24xx_write_flash(ha, FLASH_CONF_ADDR | 0x330 |
11712 		    xp->fdesc.protect_sector_cmd, 0x00600f);
11713 		(void) ql_24xx_write_flash(ha, FLASH_CONF_ADDR | 0x330 |
11714 		    xp->fdesc.protect_sector_cmd, 0x00800f);
11715 
11716 		/* TODO: ??? */
11717 		(void) ql_24xx_write_flash(ha,
11718 		    FLASH_CONF_ADDR | 0x101, 0x80);
11719 	} else {
11720 		(void) ql_24xx_write_flash(ha,
11721 		    FLASH_CONF_ADDR | 0x101, 0x9c);
11722 	}
11723 
11724 	/* Disable flash write. */
11725 	if (!(CFG_IST(ha, CFG_CTRL_81XX))) {
11726 		WRT32_IO_REG(ha, ctrl_status,
11727 		    RD32_IO_REG(ha, ctrl_status) & ~ISP_FLASH_ENABLE);
11728 		RD32_IO_REG(ha, ctrl_status);	/* PCI Posting. */
11729 	}
11730 
11731 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
11732 }
11733 
11734 /*
11735  * ql_dump_firmware
11736  *	Save RISC code state information.
11737  *
11738  * Input:
11739  *	ha = adapter state pointer.
11740  *
11741  * Returns:
11742  *	QL local function return status code.
11743  *
11744  * Context:
11745  *	Kernel context.
11746  */
11747 static int
ql_dump_firmware(ql_adapter_state_t * vha)11748 ql_dump_firmware(ql_adapter_state_t *vha)
11749 {
11750 	int			rval;
11751 	clock_t			timer = drv_usectohz(30000000);
11752 	ql_adapter_state_t	*ha = vha->pha;
11753 
11754 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
11755 
11756 	QL_DUMP_LOCK(ha);
11757 
11758 	if (ha->ql_dump_state & QL_DUMPING ||
11759 	    (ha->ql_dump_state & QL_DUMP_VALID &&
11760 	    !(ha->ql_dump_state & QL_DUMP_UPLOADED))) {
11761 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
11762 		QL_DUMP_UNLOCK(ha);
11763 		return (QL_SUCCESS);
11764 	}
11765 
11766 	QL_DUMP_UNLOCK(ha);
11767 
11768 	ql_awaken_task_daemon(ha, NULL, DRIVER_STALL, 0);
11769 
11770 	/*
11771 	 * Wait for all outstanding commands to complete
11772 	 */
11773 	(void) ql_wait_outstanding(ha);
11774 
11775 	/* Dump firmware. */
11776 	rval = ql_binary_fw_dump(ha, TRUE);
11777 
11778 	/* Do abort to force restart. */
11779 	ql_awaken_task_daemon(ha, NULL, ISP_ABORT_NEEDED, DRIVER_STALL);
11780 	EL(ha, "restarting, isp_abort_needed\n");
11781 
11782 	/* Acquire task daemon lock. */
11783 	TASK_DAEMON_LOCK(ha);
11784 
11785 	/* Wait for suspension to end. */
11786 	while (ha->task_daemon_flags & QL_SUSPENDED) {
11787 		ha->task_daemon_flags |= SUSPENDED_WAKEUP_FLG;
11788 
11789 		/* 30 seconds from now */
11790 		if (cv_reltimedwait(&ha->cv_dr_suspended,
11791 		    &ha->task_daemon_mutex, timer, TR_CLOCK_TICK) == -1) {
11792 			/*
11793 			 * The timeout time 'timer' was
11794 			 * reached without the condition
11795 			 * being signaled.
11796 			 */
11797 			break;
11798 		}
11799 	}
11800 
11801 	/* Release task daemon lock. */
11802 	TASK_DAEMON_UNLOCK(ha);
11803 
11804 	if (rval == QL_SUCCESS || rval == QL_DATA_EXISTS) {
11805 		/*EMPTY*/
11806 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
11807 	} else {
11808 		EL(ha, "failed, rval = %xh\n", rval);
11809 	}
11810 	return (rval);
11811 }
11812 
11813 /*
11814  * ql_binary_fw_dump
11815  *	Dumps binary data from firmware.
11816  *
11817  * Input:
11818  *	ha = adapter state pointer.
11819  *	lock_needed = mailbox lock needed.
11820  *
11821  * Returns:
11822  *	ql local function return status code.
11823  *
11824  * Context:
11825  *	Interrupt or Kernel context, no mailbox commands allowed.
11826  */
11827 int
ql_binary_fw_dump(ql_adapter_state_t * vha,int lock_needed)11828 ql_binary_fw_dump(ql_adapter_state_t *vha, int lock_needed)
11829 {
11830 	clock_t			timer;
11831 	mbx_cmd_t		mc;
11832 	mbx_cmd_t		*mcp = &mc;
11833 	int			rval = QL_SUCCESS;
11834 	ql_adapter_state_t	*ha = vha->pha;
11835 
11836 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
11837 
11838 	if (CFG_IST(ha, CFG_CTRL_8021)) {
11839 		EL(ha, "8021 not supported\n");
11840 		return (QL_NOT_SUPPORTED);
11841 	}
11842 
11843 	QL_DUMP_LOCK(ha);
11844 
11845 	if (ha->ql_dump_state & QL_DUMPING ||
11846 	    (ha->ql_dump_state & QL_DUMP_VALID &&
11847 	    !(ha->ql_dump_state & QL_DUMP_UPLOADED))) {
11848 		EL(ha, "dump already done, qds=%x\n", ha->ql_dump_state);
11849 		QL_DUMP_UNLOCK(ha);
11850 		return (QL_DATA_EXISTS);
11851 	}
11852 
11853 	ha->ql_dump_state &= ~(QL_DUMP_VALID | QL_DUMP_UPLOADED);
11854 	ha->ql_dump_state |= QL_DUMPING;
11855 
11856 	QL_DUMP_UNLOCK(ha);
11857 
11858 	if (CFG_IST(ha, CFG_ENABLE_FWEXTTRACE)) {
11859 
11860 		/* Insert Time Stamp */
11861 		rval = ql_fw_etrace(ha, &ha->fwexttracebuf,
11862 		    FTO_INSERT_TIME_STAMP);
11863 		if (rval != QL_SUCCESS) {
11864 			EL(ha, "f/w extended trace insert"
11865 			    "time stamp failed: %xh\n", rval);
11866 		}
11867 	}
11868 
11869 	if (lock_needed == TRUE) {
11870 		/* Acquire mailbox register lock. */
11871 		MBX_REGISTER_LOCK(ha);
11872 		timer = (ha->mcp->timeout + 2) * drv_usectohz(1000000);
11873 
11874 		/* Check for mailbox available, if not wait for signal. */
11875 		while (ha->mailbox_flags & MBX_BUSY_FLG) {
11876 			ha->mailbox_flags = (uint8_t)
11877 			    (ha->mailbox_flags | MBX_WANT_FLG);
11878 
11879 			/* 30 seconds from now */
11880 			if (cv_reltimedwait(&ha->cv_mbx_wait, &ha->mbx_mutex,
11881 			    timer, TR_CLOCK_TICK) == -1) {
11882 				/*
11883 				 * The timeout time 'timer' was
11884 				 * reached without the condition
11885 				 * being signaled.
11886 				 */
11887 
11888 				/* Release mailbox register lock. */
11889 				MBX_REGISTER_UNLOCK(ha);
11890 
11891 				EL(ha, "failed, rval = %xh\n",
11892 				    QL_FUNCTION_TIMEOUT);
11893 				return (QL_FUNCTION_TIMEOUT);
11894 			}
11895 		}
11896 
11897 		/* Set busy flag. */
11898 		ha->mailbox_flags = (uint8_t)
11899 		    (ha->mailbox_flags | MBX_BUSY_FLG);
11900 		mcp->timeout = 120;
11901 		ha->mcp = mcp;
11902 
11903 		/* Release mailbox register lock. */
11904 		MBX_REGISTER_UNLOCK(ha);
11905 	}
11906 
11907 	/* Free previous dump buffer. */
11908 	if (ha->ql_dump_ptr != NULL) {
11909 		kmem_free(ha->ql_dump_ptr, ha->ql_dump_size);
11910 		ha->ql_dump_ptr = NULL;
11911 	}
11912 
11913 	if (CFG_IST(ha, CFG_CTRL_2422)) {
11914 		ha->ql_dump_size = (uint32_t)(sizeof (ql_24xx_fw_dump_t) +
11915 		    ha->fw_ext_memory_size);
11916 	} else if (CFG_IST(ha, CFG_CTRL_25XX)) {
11917 		ha->ql_dump_size = (uint32_t)(sizeof (ql_25xx_fw_dump_t) +
11918 		    ha->fw_ext_memory_size);
11919 	} else if (CFG_IST(ha, CFG_CTRL_81XX)) {
11920 		ha->ql_dump_size = (uint32_t)(sizeof (ql_81xx_fw_dump_t) +
11921 		    ha->fw_ext_memory_size);
11922 	} else {
11923 		ha->ql_dump_size = sizeof (ql_fw_dump_t);
11924 	}
11925 
11926 	if ((ha->ql_dump_ptr = kmem_zalloc(ha->ql_dump_size, KM_NOSLEEP)) ==
11927 	    NULL) {
11928 		rval = QL_MEMORY_ALLOC_FAILED;
11929 	} else {
11930 		if (CFG_IST(ha, (CFG_CTRL_2300 | CFG_CTRL_6322))) {
11931 			rval = ql_2300_binary_fw_dump(ha, ha->ql_dump_ptr);
11932 		} else if (CFG_IST(ha, CFG_CTRL_81XX)) {
11933 			rval = ql_81xx_binary_fw_dump(ha, ha->ql_dump_ptr);
11934 		} else if (CFG_IST(ha, CFG_CTRL_25XX)) {
11935 			rval = ql_25xx_binary_fw_dump(ha, ha->ql_dump_ptr);
11936 		} else if (CFG_IST(ha, CFG_CTRL_2422)) {
11937 			rval = ql_24xx_binary_fw_dump(ha, ha->ql_dump_ptr);
11938 		} else {
11939 			rval = ql_2200_binary_fw_dump(ha, ha->ql_dump_ptr);
11940 		}
11941 	}
11942 
11943 	/* Reset ISP chip. */
11944 	ql_reset_chip(ha);
11945 
11946 	QL_DUMP_LOCK(ha);
11947 
11948 	if (rval != QL_SUCCESS) {
11949 		if (ha->ql_dump_ptr != NULL) {
11950 			kmem_free(ha->ql_dump_ptr, ha->ql_dump_size);
11951 			ha->ql_dump_ptr = NULL;
11952 		}
11953 		ha->ql_dump_state &= ~(QL_DUMPING | QL_DUMP_VALID |
11954 		    QL_DUMP_UPLOADED);
11955 		EL(ha, "failed, rval = %xh\n", rval);
11956 	} else {
11957 		ha->ql_dump_state &= ~(QL_DUMPING | QL_DUMP_UPLOADED);
11958 		ha->ql_dump_state |= QL_DUMP_VALID;
11959 		EL(ha, "done\n");
11960 	}
11961 
11962 	QL_DUMP_UNLOCK(ha);
11963 
11964 	return (rval);
11965 }
11966 
11967 /*
11968  * ql_ascii_fw_dump
11969  *	Converts firmware binary dump to ascii.
11970  *
11971  * Input:
11972  *	ha = adapter state pointer.
11973  *	bptr = buffer pointer.
11974  *
11975  * Returns:
11976  *	Amount of data buffer used.
11977  *
11978  * Context:
11979  *	Kernel context.
11980  */
11981 size_t
ql_ascii_fw_dump(ql_adapter_state_t * vha,caddr_t bufp)11982 ql_ascii_fw_dump(ql_adapter_state_t *vha, caddr_t bufp)
11983 {
11984 	uint32_t		cnt;
11985 	caddr_t			bp;
11986 	int			mbox_cnt;
11987 	ql_adapter_state_t	*ha = vha->pha;
11988 	ql_fw_dump_t		*fw = ha->ql_dump_ptr;
11989 
11990 	if (CFG_IST(ha, CFG_CTRL_2422)) {
11991 		return (ql_24xx_ascii_fw_dump(ha, bufp));
11992 	} else if (CFG_IST(ha, CFG_CTRL_2581)) {
11993 		return (ql_2581_ascii_fw_dump(ha, bufp));
11994 	}
11995 
11996 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
11997 
11998 	if (CFG_IST(ha, CFG_CTRL_2300)) {
11999 		(void) sprintf(bufp, "\nISP 2300IP ");
12000 	} else if (CFG_IST(ha, CFG_CTRL_6322)) {
12001 		(void) sprintf(bufp, "\nISP 6322FLX ");
12002 	} else {
12003 		(void) sprintf(bufp, "\nISP 2200IP ");
12004 	}
12005 
12006 	bp = bufp + strlen(bufp);
12007 	(void) sprintf(bp, "Firmware Version %d.%d.%d\n",
12008 	    ha->fw_major_version, ha->fw_minor_version,
12009 	    ha->fw_subminor_version);
12010 
12011 	(void) strcat(bufp, "\nPBIU Registers:");
12012 	bp = bufp + strlen(bufp);
12013 	for (cnt = 0; cnt < sizeof (fw->pbiu_reg) / 2; cnt++) {
12014 		if (cnt % 8 == 0) {
12015 			*bp++ = '\n';
12016 		}
12017 		(void) sprintf(bp, "%04x  ", fw->pbiu_reg[cnt]);
12018 		bp = bp + 6;
12019 	}
12020 
12021 	if (CFG_IST(ha, (CFG_CTRL_2300 | CFG_CTRL_6322))) {
12022 		(void) strcat(bufp, "\n\nReqQ-RspQ-Risc2Host Status "
12023 		    "registers:");
12024 		bp = bufp + strlen(bufp);
12025 		for (cnt = 0; cnt < sizeof (fw->risc_host_reg) / 2; cnt++) {
12026 			if (cnt % 8 == 0) {
12027 				*bp++ = '\n';
12028 			}
12029 			(void) sprintf(bp, "%04x  ", fw->risc_host_reg[cnt]);
12030 			bp = bp + 6;
12031 		}
12032 	}
12033 
12034 	(void) strcat(bp, "\n\nMailbox Registers:");
12035 	bp = bufp + strlen(bufp);
12036 	mbox_cnt = (CFG_IST(ha, (CFG_CTRL_2300 | CFG_CTRL_6322))) ? 16 : 8;
12037 	for (cnt = 0; cnt < mbox_cnt; cnt++) {
12038 		if (cnt % 8 == 0) {
12039 			*bp++ = '\n';
12040 		}
12041 		(void) sprintf(bp, "%04x  ", fw->mailbox_reg[cnt]);
12042 		bp = bp + 6;
12043 	}
12044 
12045 	if (CFG_IST(ha, (CFG_CTRL_2300 | CFG_CTRL_6322))) {
12046 		(void) strcat(bp, "\n\nAuto Request Response DMA Registers:");
12047 		bp = bufp + strlen(bufp);
12048 		for (cnt = 0; cnt < sizeof (fw->resp_dma_reg) / 2; cnt++) {
12049 			if (cnt % 8 == 0) {
12050 				*bp++ = '\n';
12051 			}
12052 			(void) sprintf(bp, "%04x  ", fw->resp_dma_reg[cnt]);
12053 			bp = bp + 6;
12054 		}
12055 	}
12056 
12057 	(void) strcat(bp, "\n\nDMA Registers:");
12058 	bp = bufp + strlen(bufp);
12059 	for (cnt = 0; cnt < sizeof (fw->dma_reg) / 2; cnt++) {
12060 		if (cnt % 8 == 0) {
12061 			*bp++ = '\n';
12062 		}
12063 		(void) sprintf(bp, "%04x  ", fw->dma_reg[cnt]);
12064 		bp = bp + 6;
12065 	}
12066 
12067 	(void) strcat(bp, "\n\nRISC Hardware Registers:");
12068 	bp = bufp + strlen(bufp);
12069 	for (cnt = 0; cnt < sizeof (fw->risc_hdw_reg) / 2; cnt++) {
12070 		if (cnt % 8 == 0) {
12071 			*bp++ = '\n';
12072 		}
12073 		(void) sprintf(bp, "%04x  ", fw->risc_hdw_reg[cnt]);
12074 		bp = bp + 6;
12075 	}
12076 
12077 	(void) strcat(bp, "\n\nRISC GP0 Registers:");
12078 	bp = bufp + strlen(bufp);
12079 	for (cnt = 0; cnt < sizeof (fw->risc_gp0_reg) / 2; cnt++) {
12080 		if (cnt % 8 == 0) {
12081 			*bp++ = '\n';
12082 		}
12083 		(void) sprintf(bp, "%04x  ", fw->risc_gp0_reg[cnt]);
12084 		bp = bp + 6;
12085 	}
12086 
12087 	(void) strcat(bp, "\n\nRISC GP1 Registers:");
12088 	bp = bufp + strlen(bufp);
12089 	for (cnt = 0; cnt < sizeof (fw->risc_gp1_reg) / 2; cnt++) {
12090 		if (cnt % 8 == 0) {
12091 			*bp++ = '\n';
12092 		}
12093 		(void) sprintf(bp, "%04x  ", fw->risc_gp1_reg[cnt]);
12094 		bp = bp + 6;
12095 	}
12096 
12097 	(void) strcat(bp, "\n\nRISC GP2 Registers:");
12098 	bp = bufp + strlen(bufp);
12099 	for (cnt = 0; cnt < sizeof (fw->risc_gp2_reg) / 2; cnt++) {
12100 		if (cnt % 8 == 0) {
12101 			*bp++ = '\n';
12102 		}
12103 		(void) sprintf(bp, "%04x  ", fw->risc_gp2_reg[cnt]);
12104 		bp = bp + 6;
12105 	}
12106 
12107 	(void) strcat(bp, "\n\nRISC GP3 Registers:");
12108 	bp = bufp + strlen(bufp);
12109 	for (cnt = 0; cnt < sizeof (fw->risc_gp3_reg) / 2; cnt++) {
12110 		if (cnt % 8 == 0) {
12111 			*bp++ = '\n';
12112 		}
12113 		(void) sprintf(bp, "%04x  ", fw->risc_gp3_reg[cnt]);
12114 		bp = bp + 6;
12115 	}
12116 
12117 	(void) strcat(bp, "\n\nRISC GP4 Registers:");
12118 	bp = bufp + strlen(bufp);
12119 	for (cnt = 0; cnt < sizeof (fw->risc_gp4_reg) / 2; cnt++) {
12120 		if (cnt % 8 == 0) {
12121 			*bp++ = '\n';
12122 		}
12123 		(void) sprintf(bp, "%04x  ", fw->risc_gp4_reg[cnt]);
12124 		bp = bp + 6;
12125 	}
12126 
12127 	(void) strcat(bp, "\n\nRISC GP5 Registers:");
12128 	bp = bufp + strlen(bufp);
12129 	for (cnt = 0; cnt < sizeof (fw->risc_gp5_reg) / 2; cnt++) {
12130 		if (cnt % 8 == 0) {
12131 			*bp++ = '\n';
12132 		}
12133 		(void) sprintf(bp, "%04x  ", fw->risc_gp5_reg[cnt]);
12134 		bp = bp + 6;
12135 	}
12136 
12137 	(void) strcat(bp, "\n\nRISC GP6 Registers:");
12138 	bp = bufp + strlen(bufp);
12139 	for (cnt = 0; cnt < sizeof (fw->risc_gp6_reg) / 2; cnt++) {
12140 		if (cnt % 8 == 0) {
12141 			*bp++ = '\n';
12142 		}
12143 		(void) sprintf(bp, "%04x  ", fw->risc_gp6_reg[cnt]);
12144 		bp = bp + 6;
12145 	}
12146 
12147 	(void) strcat(bp, "\n\nRISC GP7 Registers:");
12148 	bp = bufp + strlen(bufp);
12149 	for (cnt = 0; cnt < sizeof (fw->risc_gp7_reg) / 2; cnt++) {
12150 		if (cnt % 8 == 0) {
12151 			*bp++ = '\n';
12152 		}
12153 		(void) sprintf(bp, "%04x  ", fw->risc_gp7_reg[cnt]);
12154 		bp = bp + 6;
12155 	}
12156 
12157 	(void) strcat(bp, "\n\nFrame Buffer Hardware Registers:");
12158 	bp = bufp + strlen(bufp);
12159 	for (cnt = 0; cnt < sizeof (fw->frame_buf_hdw_reg) / 2; cnt++) {
12160 		if ((cnt == 16) && ((CFG_IST(ha, (CFG_CTRL_2300 |
12161 		    CFG_CTRL_6322)) == 0))) {
12162 			break;
12163 		}
12164 		if (cnt % 8 == 0) {
12165 			*bp++ = '\n';
12166 		}
12167 		(void) sprintf(bp, "%04x  ", fw->frame_buf_hdw_reg[cnt]);
12168 		bp = bp + 6;
12169 	}
12170 
12171 	(void) strcat(bp, "\n\nFPM B0 Registers:");
12172 	bp = bufp + strlen(bufp);
12173 	for (cnt = 0; cnt < sizeof (fw->fpm_b0_reg) / 2; cnt++) {
12174 		if (cnt % 8 == 0) {
12175 			*bp++ = '\n';
12176 		}
12177 		(void) sprintf(bp, "%04x  ", fw->fpm_b0_reg[cnt]);
12178 		bp = bp + 6;
12179 	}
12180 
12181 	(void) strcat(bp, "\n\nFPM B1 Registers:");
12182 	bp = bufp + strlen(bufp);
12183 	for (cnt = 0; cnt < sizeof (fw->fpm_b1_reg) / 2; cnt++) {
12184 		if (cnt % 8 == 0) {
12185 			*bp++ = '\n';
12186 		}
12187 		(void) sprintf(bp, "%04x  ", fw->fpm_b1_reg[cnt]);
12188 		bp = bp + 6;
12189 	}
12190 
12191 	if (CFG_IST(ha, (CFG_CTRL_2300 | CFG_CTRL_6322))) {
12192 		(void) strcat(bp, "\n\nCode RAM Dump:");
12193 		bp = bufp + strlen(bufp);
12194 		for (cnt = 0; cnt < sizeof (fw->risc_ram) / 2; cnt++) {
12195 			if (cnt % 8 == 0) {
12196 				(void) sprintf(bp, "\n%05x: ", cnt + 0x0800);
12197 				bp = bp + 8;
12198 			}
12199 			(void) sprintf(bp, "%04x  ", fw->risc_ram[cnt]);
12200 			bp = bp + 6;
12201 		}
12202 
12203 		(void) strcat(bp, "\n\nStack RAM Dump:");
12204 		bp = bufp + strlen(bufp);
12205 		for (cnt = 0; cnt < sizeof (fw->stack_ram) / 2; cnt++) {
12206 			if (cnt % 8 == 0) {
12207 				(void) sprintf(bp, "\n%05x: ", cnt + 0x010000);
12208 				bp = bp + 8;
12209 			}
12210 			(void) sprintf(bp, "%04x  ", fw->stack_ram[cnt]);
12211 			bp = bp + 6;
12212 		}
12213 
12214 		(void) strcat(bp, "\n\nData RAM Dump:");
12215 		bp = bufp + strlen(bufp);
12216 		for (cnt = 0; cnt < sizeof (fw->data_ram) / 2; cnt++) {
12217 			if (cnt % 8 == 0) {
12218 				(void) sprintf(bp, "\n%05x: ", cnt + 0x010800);
12219 				bp = bp + 8;
12220 			}
12221 			(void) sprintf(bp, "%04x  ", fw->data_ram[cnt]);
12222 			bp = bp + 6;
12223 		}
12224 	} else {
12225 		(void) strcat(bp, "\n\nRISC SRAM:");
12226 		bp = bufp + strlen(bufp);
12227 		for (cnt = 0; cnt < 0xf000; cnt++) {
12228 			if (cnt % 8 == 0) {
12229 				(void) sprintf(bp, "\n%04x: ", cnt + 0x1000);
12230 				bp = bp + 7;
12231 			}
12232 			(void) sprintf(bp, "%04x  ", fw->risc_ram[cnt]);
12233 			bp = bp + 6;
12234 		}
12235 	}
12236 
12237 	(void) strcat(bp, "\n\n[<==END] ISP Debug Dump.");
12238 	bp += strlen(bp);
12239 
12240 	(void) sprintf(bp, "\n\nRequest Queue");
12241 	bp += strlen(bp);
12242 	for (cnt = 0; cnt < REQUEST_QUEUE_SIZE / 4; cnt++) {
12243 		if (cnt % 8 == 0) {
12244 			(void) sprintf(bp, "\n%08x: ", cnt);
12245 			bp += strlen(bp);
12246 		}
12247 		(void) sprintf(bp, "%08x ", fw->req_q[cnt]);
12248 		bp += strlen(bp);
12249 	}
12250 
12251 	(void) sprintf(bp, "\n\nResponse Queue");
12252 	bp += strlen(bp);
12253 	for (cnt = 0; cnt < RESPONSE_QUEUE_SIZE / 4; cnt++) {
12254 		if (cnt % 8 == 0) {
12255 			(void) sprintf(bp, "\n%08x: ", cnt);
12256 			bp += strlen(bp);
12257 		}
12258 		(void) sprintf(bp, "%08x ", fw->rsp_q[cnt]);
12259 		bp += strlen(bp);
12260 	}
12261 
12262 	(void) sprintf(bp, "\n");
12263 
12264 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
12265 
12266 	return (strlen(bufp));
12267 }
12268 
12269 /*
12270  * ql_24xx_ascii_fw_dump
12271  *	Converts ISP24xx firmware binary dump to ascii.
12272  *
12273  * Input:
12274  *	ha = adapter state pointer.
12275  *	bptr = buffer pointer.
12276  *
12277  * Returns:
12278  *	Amount of data buffer used.
12279  *
12280  * Context:
12281  *	Kernel context.
12282  */
12283 static size_t
ql_24xx_ascii_fw_dump(ql_adapter_state_t * ha,caddr_t bufp)12284 ql_24xx_ascii_fw_dump(ql_adapter_state_t *ha, caddr_t bufp)
12285 {
12286 	uint32_t		cnt;
12287 	caddr_t			bp = bufp;
12288 	ql_24xx_fw_dump_t	*fw = ha->ql_dump_ptr;
12289 
12290 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
12291 
12292 	(void) sprintf(bp, "ISP FW Version %d.%02d.%02d Attributes %X\n",
12293 	    ha->fw_major_version, ha->fw_minor_version,
12294 	    ha->fw_subminor_version, ha->fw_attributes);
12295 	bp += strlen(bp);
12296 
12297 	(void) sprintf(bp, "\nHCCR Register\n%08x\n", fw->hccr);
12298 
12299 	(void) strcat(bp, "\nHost Interface Registers");
12300 	bp += strlen(bp);
12301 	for (cnt = 0; cnt < sizeof (fw->host_reg) / 4; cnt++) {
12302 		if (cnt % 8 == 0) {
12303 			(void) sprintf(bp++, "\n");
12304 		}
12305 
12306 		(void) sprintf(bp, "%08x ", fw->host_reg[cnt]);
12307 		bp += 9;
12308 	}
12309 
12310 	(void) sprintf(bp, "\n\nMailbox Registers");
12311 	bp += strlen(bp);
12312 	for (cnt = 0; cnt < sizeof (fw->mailbox_reg) / 2; cnt++) {
12313 		if (cnt % 16 == 0) {
12314 			(void) sprintf(bp++, "\n");
12315 		}
12316 
12317 		(void) sprintf(bp, "%04x ", fw->mailbox_reg[cnt]);
12318 		bp += 5;
12319 	}
12320 
12321 	(void) sprintf(bp, "\n\nXSEQ GP Registers");
12322 	bp += strlen(bp);
12323 	for (cnt = 0; cnt < sizeof (fw->xseq_gp_reg) / 4; cnt++) {
12324 		if (cnt % 8 == 0) {
12325 			(void) sprintf(bp++, "\n");
12326 		}
12327 
12328 		(void) sprintf(bp, "%08x ", fw->xseq_gp_reg[cnt]);
12329 		bp += 9;
12330 	}
12331 
12332 	(void) sprintf(bp, "\n\nXSEQ-0 Registers");
12333 	bp += strlen(bp);
12334 	for (cnt = 0; cnt < sizeof (fw->xseq_0_reg) / 4; cnt++) {
12335 		if (cnt % 8 == 0) {
12336 			(void) sprintf(bp++, "\n");
12337 		}
12338 
12339 		(void) sprintf(bp, "%08x ", fw->xseq_0_reg[cnt]);
12340 		bp += 9;
12341 	}
12342 
12343 	(void) sprintf(bp, "\n\nXSEQ-1 Registers");
12344 	bp += strlen(bp);
12345 	for (cnt = 0; cnt < sizeof (fw->xseq_1_reg) / 4; cnt++) {
12346 		if (cnt % 8 == 0) {
12347 			(void) sprintf(bp++, "\n");
12348 		}
12349 
12350 		(void) sprintf(bp, "%08x ", fw->xseq_1_reg[cnt]);
12351 		bp += 9;
12352 	}
12353 
12354 	(void) sprintf(bp, "\n\nRSEQ GP Registers");
12355 	bp += strlen(bp);
12356 	for (cnt = 0; cnt < sizeof (fw->rseq_gp_reg) / 4; cnt++) {
12357 		if (cnt % 8 == 0) {
12358 			(void) sprintf(bp++, "\n");
12359 		}
12360 
12361 		(void) sprintf(bp, "%08x ", fw->rseq_gp_reg[cnt]);
12362 		bp += 9;
12363 	}
12364 
12365 	(void) sprintf(bp, "\n\nRSEQ-0 Registers");
12366 	bp += strlen(bp);
12367 	for (cnt = 0; cnt < sizeof (fw->rseq_0_reg) / 4; cnt++) {
12368 		if (cnt % 8 == 0) {
12369 			(void) sprintf(bp++, "\n");
12370 		}
12371 
12372 		(void) sprintf(bp, "%08x ", fw->rseq_0_reg[cnt]);
12373 		bp += 9;
12374 	}
12375 
12376 	(void) sprintf(bp, "\n\nRSEQ-1 Registers");
12377 	bp += strlen(bp);
12378 	for (cnt = 0; cnt < sizeof (fw->rseq_1_reg) / 4; cnt++) {
12379 		if (cnt % 8 == 0) {
12380 			(void) sprintf(bp++, "\n");
12381 		}
12382 
12383 		(void) sprintf(bp, "%08x ", fw->rseq_1_reg[cnt]);
12384 		bp += 9;
12385 	}
12386 
12387 	(void) sprintf(bp, "\n\nRSEQ-2 Registers");
12388 	bp += strlen(bp);
12389 	for (cnt = 0; cnt < sizeof (fw->rseq_2_reg) / 4; cnt++) {
12390 		if (cnt % 8 == 0) {
12391 			(void) sprintf(bp++, "\n");
12392 		}
12393 
12394 		(void) sprintf(bp, "%08x ", fw->rseq_2_reg[cnt]);
12395 		bp += 9;
12396 	}
12397 
12398 	(void) sprintf(bp, "\n\nCommand DMA Registers");
12399 	bp += strlen(bp);
12400 	for (cnt = 0; cnt < sizeof (fw->cmd_dma_reg) / 4; cnt++) {
12401 		if (cnt % 8 == 0) {
12402 			(void) sprintf(bp++, "\n");
12403 		}
12404 
12405 		(void) sprintf(bp, "%08x ", fw->cmd_dma_reg[cnt]);
12406 		bp += 9;
12407 	}
12408 
12409 	(void) sprintf(bp, "\n\nRequest0 Queue DMA Channel Registers");
12410 	bp += strlen(bp);
12411 	for (cnt = 0; cnt < sizeof (fw->req0_dma_reg) / 4; cnt++) {
12412 		if (cnt % 8 == 0) {
12413 			(void) sprintf(bp++, "\n");
12414 		}
12415 
12416 		(void) sprintf(bp, "%08x ", fw->req0_dma_reg[cnt]);
12417 		bp += 9;
12418 	}
12419 
12420 	(void) sprintf(bp, "\n\nResponse0 Queue DMA Channel Registers");
12421 	bp += strlen(bp);
12422 	for (cnt = 0; cnt < sizeof (fw->resp0_dma_reg) / 4; cnt++) {
12423 		if (cnt % 8 == 0) {
12424 			(void) sprintf(bp++, "\n");
12425 		}
12426 
12427 		(void) sprintf(bp, "%08x ", fw->resp0_dma_reg[cnt]);
12428 		bp += 9;
12429 	}
12430 
12431 	(void) sprintf(bp, "\n\nRequest1 Queue DMA Channel Registers");
12432 	bp += strlen(bp);
12433 	for (cnt = 0; cnt < sizeof (fw->req1_dma_reg) / 4; cnt++) {
12434 		if (cnt % 8 == 0) {
12435 			(void) sprintf(bp++, "\n");
12436 		}
12437 
12438 		(void) sprintf(bp, "%08x ", fw->req1_dma_reg[cnt]);
12439 		bp += 9;
12440 	}
12441 
12442 	(void) sprintf(bp, "\n\nXMT0 Data DMA Registers");
12443 	bp += strlen(bp);
12444 	for (cnt = 0; cnt < sizeof (fw->xmt0_dma_reg) / 4; cnt++) {
12445 		if (cnt % 8 == 0) {
12446 			(void) sprintf(bp++, "\n");
12447 		}
12448 
12449 		(void) sprintf(bp, "%08x ", fw->xmt0_dma_reg[cnt]);
12450 		bp += 9;
12451 	}
12452 
12453 	(void) sprintf(bp, "\n\nXMT1 Data DMA Registers");
12454 	bp += strlen(bp);
12455 	for (cnt = 0; cnt < sizeof (fw->xmt1_dma_reg) / 4; cnt++) {
12456 		if (cnt % 8 == 0) {
12457 			(void) sprintf(bp++, "\n");
12458 		}
12459 
12460 		(void) sprintf(bp, "%08x ", fw->xmt1_dma_reg[cnt]);
12461 		bp += 9;
12462 	}
12463 
12464 	(void) sprintf(bp, "\n\nXMT2 Data DMA Registers");
12465 	bp += strlen(bp);
12466 	for (cnt = 0; cnt < sizeof (fw->xmt2_dma_reg) / 4; cnt++) {
12467 		if (cnt % 8 == 0) {
12468 			(void) sprintf(bp++, "\n");
12469 		}
12470 
12471 		(void) sprintf(bp, "%08x ", fw->xmt2_dma_reg[cnt]);
12472 		bp += 9;
12473 	}
12474 
12475 	(void) sprintf(bp, "\n\nXMT3 Data DMA Registers");
12476 	bp += strlen(bp);
12477 	for (cnt = 0; cnt < sizeof (fw->xmt3_dma_reg) / 4; cnt++) {
12478 		if (cnt % 8 == 0) {
12479 			(void) sprintf(bp++, "\n");
12480 		}
12481 
12482 		(void) sprintf(bp, "%08x ", fw->xmt3_dma_reg[cnt]);
12483 		bp += 9;
12484 	}
12485 
12486 	(void) sprintf(bp, "\n\nXMT4 Data DMA Registers");
12487 	bp += strlen(bp);
12488 	for (cnt = 0; cnt < sizeof (fw->xmt4_dma_reg) / 4; cnt++) {
12489 		if (cnt % 8 == 0) {
12490 			(void) sprintf(bp++, "\n");
12491 		}
12492 
12493 		(void) sprintf(bp, "%08x ", fw->xmt4_dma_reg[cnt]);
12494 		bp += 9;
12495 	}
12496 
12497 	(void) sprintf(bp, "\n\nXMT Data DMA Common Registers");
12498 	bp += strlen(bp);
12499 	for (cnt = 0; cnt < sizeof (fw->xmt_data_dma_reg) / 4; cnt++) {
12500 		if (cnt % 8 == 0) {
12501 			(void) sprintf(bp++, "\n");
12502 		}
12503 
12504 		(void) sprintf(bp, "%08x ", fw->xmt_data_dma_reg[cnt]);
12505 		bp += 9;
12506 	}
12507 
12508 	(void) sprintf(bp, "\n\nRCV Thread 0 Data DMA Registers");
12509 	bp += strlen(bp);
12510 	for (cnt = 0; cnt < sizeof (fw->rcvt0_data_dma_reg) / 4; cnt++) {
12511 		if (cnt % 8 == 0) {
12512 			(void) sprintf(bp++, "\n");
12513 		}
12514 
12515 		(void) sprintf(bp, "%08x ", fw->rcvt0_data_dma_reg[cnt]);
12516 		bp += 9;
12517 	}
12518 
12519 	(void) sprintf(bp, "\n\nRCV Thread 1 Data DMA Registers");
12520 	bp += strlen(bp);
12521 	for (cnt = 0; cnt < sizeof (fw->rcvt1_data_dma_reg) / 4; cnt++) {
12522 		if (cnt % 8 == 0) {
12523 			(void) sprintf(bp++, "\n");
12524 		}
12525 
12526 		(void) sprintf(bp, "%08x ", fw->rcvt1_data_dma_reg[cnt]);
12527 		bp += 9;
12528 	}
12529 
12530 	(void) sprintf(bp, "\n\nRISC GP Registers");
12531 	bp += strlen(bp);
12532 	for (cnt = 0; cnt < sizeof (fw->risc_gp_reg) / 4; cnt++) {
12533 		if (cnt % 8 == 0) {
12534 			(void) sprintf(bp++, "\n");
12535 		}
12536 
12537 		(void) sprintf(bp, "%08x ", fw->risc_gp_reg[cnt]);
12538 		bp += 9;
12539 	}
12540 
12541 	(void) sprintf(bufp + strlen(bufp), "\n\nShadow Registers");
12542 	bp += strlen(bp);
12543 	for (cnt = 0; cnt < sizeof (fw->shadow_reg) / 4; cnt++) {
12544 		if (cnt % 8 == 0) {
12545 			(void) sprintf(bp++, "\n");
12546 		}
12547 
12548 		(void) sprintf(bp, "%08x ", fw->shadow_reg[cnt]);
12549 		bp += 9;
12550 	}
12551 
12552 	(void) sprintf(bp, "\n\nLMC Registers");
12553 	bp += strlen(bp);
12554 	for (cnt = 0; cnt < sizeof (fw->lmc_reg) / 4; cnt++) {
12555 		if (cnt % 8 == 0) {
12556 			(void) sprintf(bp++, "\n");
12557 		}
12558 
12559 		(void) sprintf(bp, "%08x ", fw->lmc_reg[cnt]);
12560 		bp += 9;
12561 	}
12562 
12563 	(void) sprintf(bp, "\n\nFPM Hardware Registers");
12564 	bp += strlen(bp);
12565 	for (cnt = 0; cnt < sizeof (fw->fpm_hdw_reg) / 4; cnt++) {
12566 		if (cnt % 8 == 0) {
12567 			(void) sprintf(bp++, "\n");
12568 		}
12569 
12570 		(void) sprintf(bp, "%08x ", fw->fpm_hdw_reg[cnt]);
12571 		bp += 9;
12572 	}
12573 
12574 	(void) sprintf(bp, "\n\nFB Hardware Registers");
12575 	bp += strlen(bp);
12576 	for (cnt = 0; cnt < sizeof (fw->fb_hdw_reg) / 4; cnt++) {
12577 		if (cnt % 8 == 0) {
12578 			(void) sprintf(bp++, "\n");
12579 		}
12580 
12581 		(void) sprintf(bp, "%08x ", fw->fb_hdw_reg[cnt]);
12582 		bp += 9;
12583 	}
12584 
12585 	(void) sprintf(bp, "\n\nCode RAM");
12586 	bp += strlen(bp);
12587 	for (cnt = 0; cnt < sizeof (fw->code_ram) / 4; cnt++) {
12588 		if (cnt % 8 == 0) {
12589 			(void) sprintf(bp, "\n%08x: ", cnt + 0x20000);
12590 			bp += 11;
12591 		}
12592 
12593 		(void) sprintf(bp, "%08x ", fw->code_ram[cnt]);
12594 		bp += 9;
12595 	}
12596 
12597 	(void) sprintf(bp, "\n\nExternal Memory");
12598 	bp += strlen(bp);
12599 	for (cnt = 0; cnt < ha->fw_ext_memory_size / 4; cnt++) {
12600 		if (cnt % 8 == 0) {
12601 			(void) sprintf(bp, "\n%08x: ", cnt + 0x100000);
12602 			bp += 11;
12603 		}
12604 		(void) sprintf(bp, "%08x ", fw->ext_mem[cnt]);
12605 		bp += 9;
12606 	}
12607 
12608 	(void) sprintf(bp, "\n[<==END] ISP Debug Dump");
12609 	bp += strlen(bp);
12610 
12611 	(void) sprintf(bp, "\n\nRequest Queue");
12612 	bp += strlen(bp);
12613 	for (cnt = 0; cnt < REQUEST_QUEUE_SIZE / 4; cnt++) {
12614 		if (cnt % 8 == 0) {
12615 			(void) sprintf(bp, "\n%08x: ", cnt);
12616 			bp += strlen(bp);
12617 		}
12618 		(void) sprintf(bp, "%08x ", fw->req_q[cnt]);
12619 		bp += strlen(bp);
12620 	}
12621 
12622 	(void) sprintf(bp, "\n\nResponse Queue");
12623 	bp += strlen(bp);
12624 	for (cnt = 0; cnt < RESPONSE_QUEUE_SIZE / 4; cnt++) {
12625 		if (cnt % 8 == 0) {
12626 			(void) sprintf(bp, "\n%08x: ", cnt);
12627 			bp += strlen(bp);
12628 		}
12629 		(void) sprintf(bp, "%08x ", fw->rsp_q[cnt]);
12630 		bp += strlen(bp);
12631 	}
12632 
12633 	if (CFG_IST(ha, CFG_ENABLE_FWEXTTRACE) &&
12634 	    (ha->fwexttracebuf.bp != NULL)) {
12635 		uint32_t cnt_b = 0;
12636 		uint64_t w64 = (uintptr_t)ha->fwexttracebuf.bp;
12637 
12638 		(void) sprintf(bp, "\n\nExtended Trace Buffer Memory");
12639 		bp += strlen(bp);
12640 		/* show data address as a byte address, data as long words */
12641 		for (cnt = 0; cnt < FWEXTSIZE / 4; cnt++) {
12642 			cnt_b = cnt * 4;
12643 			if (cnt_b % 32 == 0) {
12644 				(void) sprintf(bp, "\n%08x: ",
12645 				    (int)(w64 + cnt_b));
12646 				bp += 11;
12647 			}
12648 			(void) sprintf(bp, "%08x ", fw->ext_trace_buf[cnt]);
12649 			bp += 9;
12650 		}
12651 	}
12652 
12653 	if (CFG_IST(ha, CFG_ENABLE_FWFCETRACE) &&
12654 	    (ha->fwfcetracebuf.bp != NULL)) {
12655 		uint32_t cnt_b = 0;
12656 		uint64_t w64 = (uintptr_t)ha->fwfcetracebuf.bp;
12657 
12658 		(void) sprintf(bp, "\n\nFC Event Trace Buffer Memory");
12659 		bp += strlen(bp);
12660 		/* show data address as a byte address, data as long words */
12661 		for (cnt = 0; cnt < FWFCESIZE / 4; cnt++) {
12662 			cnt_b = cnt * 4;
12663 			if (cnt_b % 32 == 0) {
12664 				(void) sprintf(bp, "\n%08x: ",
12665 				    (int)(w64 + cnt_b));
12666 				bp += 11;
12667 			}
12668 			(void) sprintf(bp, "%08x ", fw->fce_trace_buf[cnt]);
12669 			bp += 9;
12670 		}
12671 	}
12672 
12673 	(void) sprintf(bp, "\n\n");
12674 	bp += strlen(bp);
12675 
12676 	cnt = (uint32_t)((uintptr_t)bp - (uintptr_t)bufp);
12677 
12678 	QL_PRINT_3(CE_CONT, "(%d): done=%xh\n", ha->instance, cnt);
12679 
12680 	return (cnt);
12681 }
12682 
12683 /*
12684  * ql_2581_ascii_fw_dump
12685  *	Converts ISP25xx or ISP81xx firmware binary dump to ascii.
12686  *
12687  * Input:
12688  *	ha = adapter state pointer.
12689  *	bptr = buffer pointer.
12690  *
12691  * Returns:
12692  *	Amount of data buffer used.
12693  *
12694  * Context:
12695  *	Kernel context.
12696  */
12697 static size_t
ql_2581_ascii_fw_dump(ql_adapter_state_t * ha,caddr_t bufp)12698 ql_2581_ascii_fw_dump(ql_adapter_state_t *ha, caddr_t bufp)
12699 {
12700 	uint32_t		cnt;
12701 	uint32_t		cnt1;
12702 	caddr_t			bp = bufp;
12703 	ql_25xx_fw_dump_t	*fw = ha->ql_dump_ptr;
12704 
12705 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
12706 
12707 	(void) sprintf(bp, "\nISP FW Version %d.%02d.%02d Attributes %X\n",
12708 	    ha->fw_major_version, ha->fw_minor_version,
12709 	    ha->fw_subminor_version, ha->fw_attributes);
12710 	bp += strlen(bp);
12711 
12712 	(void) sprintf(bp, "\nR2H Status Register\n%08x\n", fw->r2h_status);
12713 	bp += strlen(bp);
12714 
12715 	(void) sprintf(bp, "\nHostRisc Registers");
12716 	bp += strlen(bp);
12717 	for (cnt = 0; cnt < sizeof (fw->hostrisc_reg) / 4; cnt++) {
12718 		if (cnt % 8 == 0) {
12719 			(void) sprintf(bp++, "\n");
12720 		}
12721 		(void) sprintf(bp, "%08x ", fw->hostrisc_reg[cnt]);
12722 		bp += 9;
12723 	}
12724 
12725 	(void) sprintf(bp, "\n\nPCIe Registers");
12726 	bp += strlen(bp);
12727 	for (cnt = 0; cnt < sizeof (fw->pcie_reg) / 4; cnt++) {
12728 		if (cnt % 8 == 0) {
12729 			(void) sprintf(bp++, "\n");
12730 		}
12731 		(void) sprintf(bp, "%08x ", fw->pcie_reg[cnt]);
12732 		bp += 9;
12733 	}
12734 
12735 	(void) strcat(bp, "\n\nHost Interface Registers");
12736 	bp += strlen(bp);
12737 	for (cnt = 0; cnt < sizeof (fw->host_reg) / 4; cnt++) {
12738 		if (cnt % 8 == 0) {
12739 			(void) sprintf(bp++, "\n");
12740 		}
12741 		(void) sprintf(bp, "%08x ", fw->host_reg[cnt]);
12742 		bp += 9;
12743 	}
12744 
12745 	(void) sprintf(bufp + strlen(bufp), "\n\nShadow Registers");
12746 	bp += strlen(bp);
12747 	for (cnt = 0; cnt < sizeof (fw->shadow_reg) / 4; cnt++) {
12748 		if (cnt % 8 == 0) {
12749 			(void) sprintf(bp++, "\n");
12750 		}
12751 		(void) sprintf(bp, "%08x ", fw->shadow_reg[cnt]);
12752 		bp += 9;
12753 	}
12754 
12755 	(void) sprintf(bufp + strlen(bufp), "\n\nRISC IO Register\n%08x",
12756 	    fw->risc_io);
12757 	bp += strlen(bp);
12758 
12759 	(void) sprintf(bp, "\n\nMailbox Registers");
12760 	bp += strlen(bp);
12761 	for (cnt = 0; cnt < sizeof (fw->mailbox_reg) / 2; cnt++) {
12762 		if (cnt % 16 == 0) {
12763 			(void) sprintf(bp++, "\n");
12764 		}
12765 		(void) sprintf(bp, "%04x ", fw->mailbox_reg[cnt]);
12766 		bp += 5;
12767 	}
12768 
12769 	(void) sprintf(bp, "\n\nXSEQ GP Registers");
12770 	bp += strlen(bp);
12771 	for (cnt = 0; cnt < sizeof (fw->xseq_gp_reg) / 4; cnt++) {
12772 		if (cnt % 8 == 0) {
12773 			(void) sprintf(bp++, "\n");
12774 		}
12775 		(void) sprintf(bp, "%08x ", fw->xseq_gp_reg[cnt]);
12776 		bp += 9;
12777 	}
12778 
12779 	(void) sprintf(bp, "\n\nXSEQ-0 Registers");
12780 	bp += strlen(bp);
12781 	for (cnt = 0; cnt < sizeof (fw->xseq_0_reg) / 4; cnt++) {
12782 		if (cnt % 8 == 0) {
12783 			(void) sprintf(bp++, "\n");
12784 		}
12785 		(void) sprintf(bp, "%08x ", fw->xseq_0_reg[cnt]);
12786 		bp += 9;
12787 	}
12788 
12789 	(void) sprintf(bp, "\n\nXSEQ-1 Registers");
12790 	bp += strlen(bp);
12791 	for (cnt = 0; cnt < sizeof (fw->xseq_1_reg) / 4; cnt++) {
12792 		if (cnt % 8 == 0) {
12793 			(void) sprintf(bp++, "\n");
12794 		}
12795 		(void) sprintf(bp, "%08x ", fw->xseq_1_reg[cnt]);
12796 		bp += 9;
12797 	}
12798 
12799 	(void) sprintf(bp, "\n\nRSEQ GP Registers");
12800 	bp += strlen(bp);
12801 	for (cnt = 0; cnt < sizeof (fw->rseq_gp_reg) / 4; cnt++) {
12802 		if (cnt % 8 == 0) {
12803 			(void) sprintf(bp++, "\n");
12804 		}
12805 		(void) sprintf(bp, "%08x ", fw->rseq_gp_reg[cnt]);
12806 		bp += 9;
12807 	}
12808 
12809 	(void) sprintf(bp, "\n\nRSEQ-0 Registers");
12810 	bp += strlen(bp);
12811 	for (cnt = 0; cnt < sizeof (fw->rseq_0_reg) / 4; cnt++) {
12812 		if (cnt % 8 == 0) {
12813 			(void) sprintf(bp++, "\n");
12814 		}
12815 		(void) sprintf(bp, "%08x ", fw->rseq_0_reg[cnt]);
12816 		bp += 9;
12817 	}
12818 
12819 	(void) sprintf(bp, "\n\nRSEQ-1 Registers");
12820 	bp += strlen(bp);
12821 	for (cnt = 0; cnt < sizeof (fw->rseq_1_reg) / 4; cnt++) {
12822 		if (cnt % 8 == 0) {
12823 			(void) sprintf(bp++, "\n");
12824 		}
12825 		(void) sprintf(bp, "%08x ", fw->rseq_1_reg[cnt]);
12826 		bp += 9;
12827 	}
12828 
12829 	(void) sprintf(bp, "\n\nRSEQ-2 Registers");
12830 	bp += strlen(bp);
12831 	for (cnt = 0; cnt < sizeof (fw->rseq_2_reg) / 4; cnt++) {
12832 		if (cnt % 8 == 0) {
12833 			(void) sprintf(bp++, "\n");
12834 		}
12835 		(void) sprintf(bp, "%08x ", fw->rseq_2_reg[cnt]);
12836 		bp += 9;
12837 	}
12838 
12839 	(void) sprintf(bp, "\n\nASEQ GP Registers");
12840 	bp += strlen(bp);
12841 	for (cnt = 0; cnt < sizeof (fw->aseq_gp_reg) / 4; cnt++) {
12842 		if (cnt % 8 == 0) {
12843 			(void) sprintf(bp++, "\n");
12844 		}
12845 		(void) sprintf(bp, "%08x ", fw->aseq_gp_reg[cnt]);
12846 		bp += 9;
12847 	}
12848 
12849 	(void) sprintf(bp, "\n\nASEQ-0 Registers");
12850 	bp += strlen(bp);
12851 	for (cnt = 0; cnt < sizeof (fw->aseq_0_reg) / 4; cnt++) {
12852 		if (cnt % 8 == 0) {
12853 			(void) sprintf(bp++, "\n");
12854 		}
12855 		(void) sprintf(bp, "%08x ", fw->aseq_0_reg[cnt]);
12856 		bp += 9;
12857 	}
12858 
12859 	(void) sprintf(bp, "\n\nASEQ-1 Registers");
12860 	bp += strlen(bp);
12861 	for (cnt = 0; cnt < sizeof (fw->aseq_1_reg) / 4; cnt++) {
12862 		if (cnt % 8 == 0) {
12863 			(void) sprintf(bp++, "\n");
12864 		}
12865 		(void) sprintf(bp, "%08x ", fw->aseq_1_reg[cnt]);
12866 		bp += 9;
12867 	}
12868 
12869 	(void) sprintf(bp, "\n\nASEQ-2 Registers");
12870 	bp += strlen(bp);
12871 	for (cnt = 0; cnt < sizeof (fw->aseq_2_reg) / 4; cnt++) {
12872 		if (cnt % 8 == 0) {
12873 			(void) sprintf(bp++, "\n");
12874 		}
12875 		(void) sprintf(bp, "%08x ", fw->aseq_2_reg[cnt]);
12876 		bp += 9;
12877 	}
12878 
12879 	(void) sprintf(bp, "\n\nCommand DMA Registers");
12880 	bp += strlen(bp);
12881 	for (cnt = 0; cnt < sizeof (fw->cmd_dma_reg) / 4; cnt++) {
12882 		if (cnt % 8 == 0) {
12883 			(void) sprintf(bp++, "\n");
12884 		}
12885 		(void)  sprintf(bp, "%08x ", fw->cmd_dma_reg[cnt]);
12886 		bp += 9;
12887 	}
12888 
12889 	(void) sprintf(bp, "\n\nRequest0 Queue DMA Channel Registers");
12890 	bp += strlen(bp);
12891 	for (cnt = 0; cnt < sizeof (fw->req0_dma_reg) / 4; cnt++) {
12892 		if (cnt % 8 == 0) {
12893 			(void) sprintf(bp++, "\n");
12894 		}
12895 		(void) sprintf(bp, "%08x ", fw->req0_dma_reg[cnt]);
12896 		bp += 9;
12897 	}
12898 
12899 	(void) sprintf(bp, "\n\nResponse0 Queue DMA Channel Registers");
12900 	bp += strlen(bp);
12901 	for (cnt = 0; cnt < sizeof (fw->resp0_dma_reg) / 4; cnt++) {
12902 		if (cnt % 8 == 0) {
12903 			(void) sprintf(bp++, "\n");
12904 		}
12905 		(void) sprintf(bp, "%08x ", fw->resp0_dma_reg[cnt]);
12906 		bp += 9;
12907 	}
12908 
12909 	(void) sprintf(bp, "\n\nRequest1 Queue DMA Channel Registers");
12910 	bp += strlen(bp);
12911 	for (cnt = 0; cnt < sizeof (fw->req1_dma_reg) / 4; cnt++) {
12912 		if (cnt % 8 == 0) {
12913 			(void) sprintf(bp++, "\n");
12914 		}
12915 		(void) sprintf(bp, "%08x ", fw->req1_dma_reg[cnt]);
12916 		bp += 9;
12917 	}
12918 
12919 	(void) sprintf(bp, "\n\nXMT0 Data DMA Registers");
12920 	bp += strlen(bp);
12921 	for (cnt = 0; cnt < sizeof (fw->xmt0_dma_reg) / 4; cnt++) {
12922 		if (cnt % 8 == 0) {
12923 			(void) sprintf(bp++, "\n");
12924 		}
12925 		(void) sprintf(bp, "%08x ", fw->xmt0_dma_reg[cnt]);
12926 		bp += 9;
12927 	}
12928 
12929 	(void) sprintf(bp, "\n\nXMT1 Data DMA Registers");
12930 	bp += strlen(bp);
12931 	for (cnt = 0; cnt < sizeof (fw->xmt1_dma_reg) / 4; cnt++) {
12932 		if (cnt % 8 == 0) {
12933 			(void) sprintf(bp++, "\n");
12934 		}
12935 		(void) sprintf(bp, "%08x ", fw->xmt1_dma_reg[cnt]);
12936 		bp += 9;
12937 	}
12938 
12939 	(void) sprintf(bp, "\n\nXMT2 Data DMA Registers");
12940 	bp += strlen(bp);
12941 	for (cnt = 0; cnt < sizeof (fw->xmt2_dma_reg) / 4; cnt++) {
12942 		if (cnt % 8 == 0) {
12943 			(void) sprintf(bp++, "\n");
12944 		}
12945 		(void) sprintf(bp, "%08x ", fw->xmt2_dma_reg[cnt]);
12946 		bp += 9;
12947 	}
12948 
12949 	(void) sprintf(bp, "\n\nXMT3 Data DMA Registers");
12950 	bp += strlen(bp);
12951 	for (cnt = 0; cnt < sizeof (fw->xmt3_dma_reg) / 4; cnt++) {
12952 		if (cnt % 8 == 0) {
12953 			(void) sprintf(bp++, "\n");
12954 		}
12955 		(void) sprintf(bp, "%08x ", fw->xmt3_dma_reg[cnt]);
12956 		bp += 9;
12957 	}
12958 
12959 	(void) sprintf(bp, "\n\nXMT4 Data DMA Registers");
12960 	bp += strlen(bp);
12961 	for (cnt = 0; cnt < sizeof (fw->xmt4_dma_reg) / 4; cnt++) {
12962 		if (cnt % 8 == 0) {
12963 			(void) sprintf(bp++, "\n");
12964 		}
12965 		(void) sprintf(bp, "%08x ", fw->xmt4_dma_reg[cnt]);
12966 		bp += 9;
12967 	}
12968 
12969 	(void) sprintf(bp, "\n\nXMT Data DMA Common Registers");
12970 	bp += strlen(bp);
12971 	for (cnt = 0; cnt < sizeof (fw->xmt_data_dma_reg) / 4; cnt++) {
12972 		if (cnt % 8 == 0) {
12973 			(void) sprintf(bp++, "\n");
12974 		}
12975 		(void) sprintf(bp, "%08x ", fw->xmt_data_dma_reg[cnt]);
12976 		bp += 9;
12977 	}
12978 
12979 	(void) sprintf(bp, "\n\nRCV Thread 0 Data DMA Registers");
12980 	bp += strlen(bp);
12981 	for (cnt = 0; cnt < sizeof (fw->rcvt0_data_dma_reg) / 4; cnt++) {
12982 		if (cnt % 8 == 0) {
12983 			(void) sprintf(bp++, "\n");
12984 		}
12985 		(void) sprintf(bp, "%08x ", fw->rcvt0_data_dma_reg[cnt]);
12986 		bp += 9;
12987 	}
12988 
12989 	(void) sprintf(bp, "\n\nRCV Thread 1 Data DMA Registers");
12990 	bp += strlen(bp);
12991 	for (cnt = 0; cnt < sizeof (fw->rcvt1_data_dma_reg) / 4; cnt++) {
12992 		if (cnt % 8 == 0) {
12993 			(void) sprintf(bp++, "\n");
12994 		}
12995 		(void) sprintf(bp, "%08x ", fw->rcvt1_data_dma_reg[cnt]);
12996 		bp += 9;
12997 	}
12998 
12999 	(void) sprintf(bp, "\n\nRISC GP Registers");
13000 	bp += strlen(bp);
13001 	for (cnt = 0; cnt < sizeof (fw->risc_gp_reg) / 4; cnt++) {
13002 		if (cnt % 8 == 0) {
13003 			(void) sprintf(bp++, "\n");
13004 		}
13005 		(void) sprintf(bp, "%08x ", fw->risc_gp_reg[cnt]);
13006 		bp += 9;
13007 	}
13008 
13009 	(void) sprintf(bp, "\n\nLMC Registers");
13010 	bp += strlen(bp);
13011 	for (cnt = 0; cnt < sizeof (fw->lmc_reg) / 4; cnt++) {
13012 		if (cnt % 8 == 0) {
13013 			(void) sprintf(bp++, "\n");
13014 		}
13015 		(void) sprintf(bp, "%08x ", fw->lmc_reg[cnt]);
13016 		bp += 9;
13017 	}
13018 
13019 	(void) sprintf(bp, "\n\nFPM Hardware Registers");
13020 	bp += strlen(bp);
13021 	cnt1 = CFG_IST(ha, CFG_CTRL_81XX) ?
13022 	    (uint32_t)(sizeof (((ql_81xx_fw_dump_t *)(fw))->fpm_hdw_reg)) :
13023 	    (uint32_t)(sizeof (fw->fpm_hdw_reg));
13024 	for (cnt = 0; cnt < cnt1 / 4; cnt++) {
13025 		if (cnt % 8 == 0) {
13026 			(void) sprintf(bp++, "\n");
13027 		}
13028 		(void) sprintf(bp, "%08x ", fw->fpm_hdw_reg[cnt]);
13029 		bp += 9;
13030 	}
13031 
13032 	(void) sprintf(bp, "\n\nFB Hardware Registers");
13033 	bp += strlen(bp);
13034 	cnt1 = CFG_IST(ha, CFG_CTRL_81XX) ?
13035 	    (uint32_t)(sizeof (((ql_81xx_fw_dump_t *)(fw))->fb_hdw_reg)) :
13036 	    (uint32_t)(sizeof (fw->fb_hdw_reg));
13037 	for (cnt = 0; cnt < cnt1 / 4; cnt++) {
13038 		if (cnt % 8 == 0) {
13039 			(void) sprintf(bp++, "\n");
13040 		}
13041 		(void) sprintf(bp, "%08x ", fw->fb_hdw_reg[cnt]);
13042 		bp += 9;
13043 	}
13044 
13045 	(void) sprintf(bp, "\n\nCode RAM");
13046 	bp += strlen(bp);
13047 	for (cnt = 0; cnt < sizeof (fw->code_ram) / 4; cnt++) {
13048 		if (cnt % 8 == 0) {
13049 			(void) sprintf(bp, "\n%08x: ", cnt + 0x20000);
13050 			bp += 11;
13051 		}
13052 		(void) sprintf(bp, "%08x ", fw->code_ram[cnt]);
13053 		bp += 9;
13054 	}
13055 
13056 	(void) sprintf(bp, "\n\nExternal Memory");
13057 	bp += strlen(bp);
13058 	for (cnt = 0; cnt < ha->fw_ext_memory_size / 4; cnt++) {
13059 		if (cnt % 8 == 0) {
13060 			(void) sprintf(bp, "\n%08x: ", cnt + 0x100000);
13061 			bp += 11;
13062 		}
13063 		(void) sprintf(bp, "%08x ", fw->ext_mem[cnt]);
13064 		bp += 9;
13065 	}
13066 
13067 	(void) sprintf(bp, "\n[<==END] ISP Debug Dump");
13068 	bp += strlen(bp);
13069 
13070 	(void) sprintf(bp, "\n\nRequest Queue");
13071 	bp += strlen(bp);
13072 	for (cnt = 0; cnt < REQUEST_QUEUE_SIZE / 4; cnt++) {
13073 		if (cnt % 8 == 0) {
13074 			(void) sprintf(bp, "\n%08x: ", cnt);
13075 			bp += strlen(bp);
13076 		}
13077 		(void) sprintf(bp, "%08x ", fw->req_q[cnt]);
13078 		bp += strlen(bp);
13079 	}
13080 
13081 	(void) sprintf(bp, "\n\nResponse Queue");
13082 	bp += strlen(bp);
13083 	for (cnt = 0; cnt < RESPONSE_QUEUE_SIZE / 4; cnt++) {
13084 		if (cnt % 8 == 0) {
13085 			(void) sprintf(bp, "\n%08x: ", cnt);
13086 			bp += strlen(bp);
13087 		}
13088 		(void) sprintf(bp, "%08x ", fw->rsp_q[cnt]);
13089 		bp += strlen(bp);
13090 	}
13091 
13092 	if (CFG_IST(ha, CFG_ENABLE_FWEXTTRACE) &&
13093 	    (ha->fwexttracebuf.bp != NULL)) {
13094 		uint32_t cnt_b = 0;
13095 		uint64_t w64 = (uintptr_t)ha->fwexttracebuf.bp;
13096 
13097 		(void) sprintf(bp, "\n\nExtended Trace Buffer Memory");
13098 		bp += strlen(bp);
13099 		/* show data address as a byte address, data as long words */
13100 		for (cnt = 0; cnt < FWEXTSIZE / 4; cnt++) {
13101 			cnt_b = cnt * 4;
13102 			if (cnt_b % 32 == 0) {
13103 				(void) sprintf(bp, "\n%08x: ",
13104 				    (int)(w64 + cnt_b));
13105 				bp += 11;
13106 			}
13107 			(void) sprintf(bp, "%08x ", fw->ext_trace_buf[cnt]);
13108 			bp += 9;
13109 		}
13110 	}
13111 
13112 	if (CFG_IST(ha, CFG_ENABLE_FWFCETRACE) &&
13113 	    (ha->fwfcetracebuf.bp != NULL)) {
13114 		uint32_t cnt_b = 0;
13115 		uint64_t w64 = (uintptr_t)ha->fwfcetracebuf.bp;
13116 
13117 		(void) sprintf(bp, "\n\nFC Event Trace Buffer Memory");
13118 		bp += strlen(bp);
13119 		/* show data address as a byte address, data as long words */
13120 		for (cnt = 0; cnt < FWFCESIZE / 4; cnt++) {
13121 			cnt_b = cnt * 4;
13122 			if (cnt_b % 32 == 0) {
13123 				(void) sprintf(bp, "\n%08x: ",
13124 				    (int)(w64 + cnt_b));
13125 				bp += 11;
13126 			}
13127 			(void) sprintf(bp, "%08x ", fw->fce_trace_buf[cnt]);
13128 			bp += 9;
13129 		}
13130 	}
13131 
13132 	(void) sprintf(bp, "\n\n");
13133 	bp += strlen(bp);
13134 
13135 	cnt = (uint32_t)((uintptr_t)bp - (uintptr_t)bufp);
13136 
13137 	QL_PRINT_3(CE_CONT, "(%d): done=%xh\n", ha->instance, cnt);
13138 
13139 	return (cnt);
13140 }
13141 
13142 /*
13143  * ql_2200_binary_fw_dump
13144  *
13145  * Input:
13146  *	ha:	adapter state pointer.
13147  *	fw:	firmware dump context pointer.
13148  *
13149  * Returns:
13150  *	ql local function return status code.
13151  *
13152  * Context:
13153  *	Interrupt or Kernel context, no mailbox commands allowed.
13154  */
13155 static int
ql_2200_binary_fw_dump(ql_adapter_state_t * ha,ql_fw_dump_t * fw)13156 ql_2200_binary_fw_dump(ql_adapter_state_t *ha, ql_fw_dump_t *fw)
13157 {
13158 	uint32_t	cnt;
13159 	uint16_t	risc_address;
13160 	clock_t		timer;
13161 	mbx_cmd_t	mc;
13162 	mbx_cmd_t	*mcp = &mc;
13163 	int		rval = QL_SUCCESS;
13164 
13165 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
13166 
13167 	/* Disable ISP interrupts. */
13168 	WRT16_IO_REG(ha, ictrl, 0);
13169 	ADAPTER_STATE_LOCK(ha);
13170 	ha->flags &= ~INTERRUPTS_ENABLED;
13171 	ADAPTER_STATE_UNLOCK(ha);
13172 
13173 	/* Release mailbox registers. */
13174 	WRT16_IO_REG(ha, semaphore, 0);
13175 
13176 	/* Pause RISC. */
13177 	WRT16_IO_REG(ha, hccr, HC_PAUSE_RISC);
13178 	timer = 30000;
13179 	while ((RD16_IO_REG(ha, hccr) & HC_RISC_PAUSE) == 0) {
13180 		if (timer-- != 0) {
13181 			drv_usecwait(MILLISEC);
13182 		} else {
13183 			rval = QL_FUNCTION_TIMEOUT;
13184 			break;
13185 		}
13186 	}
13187 
13188 	if (rval == QL_SUCCESS) {
13189 		(void) ql_read_regs(ha, fw->pbiu_reg, ha->iobase,
13190 		    sizeof (fw->pbiu_reg) / 2, 16);
13191 
13192 		/* In 2200 we only read 8 mailboxes */
13193 		(void) ql_read_regs(ha, fw->mailbox_reg, ha->iobase + 0x10,
13194 		    8, 16);
13195 
13196 		(void) ql_read_regs(ha, fw->dma_reg, ha->iobase + 0x20,
13197 		    sizeof (fw->dma_reg) / 2, 16);
13198 
13199 		WRT16_IO_REG(ha, ctrl_status, 0);
13200 		(void) ql_read_regs(ha, fw->risc_hdw_reg, ha->iobase + 0xA0,
13201 		    sizeof (fw->risc_hdw_reg) / 2, 16);
13202 
13203 		WRT16_IO_REG(ha, pcr, 0x2000);
13204 		(void) ql_read_regs(ha, fw->risc_gp0_reg, ha->iobase + 0x80,
13205 		    sizeof (fw->risc_gp0_reg) / 2, 16);
13206 
13207 		WRT16_IO_REG(ha, pcr, 0x2100);
13208 		(void) ql_read_regs(ha, fw->risc_gp1_reg, ha->iobase + 0x80,
13209 		    sizeof (fw->risc_gp1_reg) / 2, 16);
13210 
13211 		WRT16_IO_REG(ha, pcr, 0x2200);
13212 		(void) ql_read_regs(ha, fw->risc_gp2_reg, ha->iobase + 0x80,
13213 		    sizeof (fw->risc_gp2_reg) / 2, 16);
13214 
13215 		WRT16_IO_REG(ha, pcr, 0x2300);
13216 		(void) ql_read_regs(ha, fw->risc_gp3_reg, ha->iobase + 0x80,
13217 		    sizeof (fw->risc_gp3_reg) / 2, 16);
13218 
13219 		WRT16_IO_REG(ha, pcr, 0x2400);
13220 		(void) ql_read_regs(ha, fw->risc_gp4_reg, ha->iobase + 0x80,
13221 		    sizeof (fw->risc_gp4_reg) / 2, 16);
13222 
13223 		WRT16_IO_REG(ha, pcr, 0x2500);
13224 		(void) ql_read_regs(ha, fw->risc_gp5_reg, ha->iobase + 0x80,
13225 		    sizeof (fw->risc_gp5_reg) / 2, 16);
13226 
13227 		WRT16_IO_REG(ha, pcr, 0x2600);
13228 		(void) ql_read_regs(ha, fw->risc_gp6_reg, ha->iobase + 0x80,
13229 		    sizeof (fw->risc_gp6_reg) / 2, 16);
13230 
13231 		WRT16_IO_REG(ha, pcr, 0x2700);
13232 		(void) ql_read_regs(ha, fw->risc_gp7_reg, ha->iobase + 0x80,
13233 		    sizeof (fw->risc_gp7_reg) / 2, 16);
13234 
13235 		WRT16_IO_REG(ha, ctrl_status, 0x10);
13236 		/* 2200 has only 16 registers */
13237 		(void) ql_read_regs(ha, fw->frame_buf_hdw_reg,
13238 		    ha->iobase + 0x80, 16, 16);
13239 
13240 		WRT16_IO_REG(ha, ctrl_status, 0x20);
13241 		(void) ql_read_regs(ha, fw->fpm_b0_reg, ha->iobase + 0x80,
13242 		    sizeof (fw->fpm_b0_reg) / 2, 16);
13243 
13244 		WRT16_IO_REG(ha, ctrl_status, 0x30);
13245 		(void) ql_read_regs(ha, fw->fpm_b1_reg, ha->iobase + 0x80,
13246 		    sizeof (fw->fpm_b1_reg) / 2, 16);
13247 
13248 		/* Select FPM registers. */
13249 		WRT16_IO_REG(ha, ctrl_status, 0x20);
13250 
13251 		/* FPM Soft Reset. */
13252 		WRT16_IO_REG(ha, fpm_diag_config, 0x100);
13253 
13254 		/* Select frame buffer registers. */
13255 		WRT16_IO_REG(ha, ctrl_status, 0x10);
13256 
13257 		/* Reset frame buffer FIFOs. */
13258 		WRT16_IO_REG(ha, fb_cmd, 0xa000);
13259 
13260 		/* Select RISC module registers. */
13261 		WRT16_IO_REG(ha, ctrl_status, 0);
13262 
13263 		/* Reset RISC module. */
13264 		WRT16_IO_REG(ha, hccr, HC_RESET_RISC);
13265 
13266 		/* Reset ISP semaphore. */
13267 		WRT16_IO_REG(ha, semaphore, 0);
13268 
13269 		/* Release RISC module. */
13270 		WRT16_IO_REG(ha, hccr, HC_RELEASE_RISC);
13271 
13272 		/* Wait for RISC to recover from reset. */
13273 		timer = 30000;
13274 		while (RD16_IO_REG(ha, mailbox_out[0]) == MBS_BUSY) {
13275 			if (timer-- != 0) {
13276 				drv_usecwait(MILLISEC);
13277 			} else {
13278 				rval = QL_FUNCTION_TIMEOUT;
13279 				break;
13280 			}
13281 		}
13282 
13283 		/* Disable RISC pause on FPM parity error. */
13284 		WRT16_IO_REG(ha, hccr, HC_DISABLE_PARITY_PAUSE);
13285 	}
13286 
13287 	if (rval == QL_SUCCESS) {
13288 		/* Pause RISC. */
13289 		WRT16_IO_REG(ha, hccr, HC_PAUSE_RISC);
13290 		timer = 30000;
13291 		while ((RD16_IO_REG(ha, hccr) & HC_RISC_PAUSE) == 0) {
13292 			if (timer-- != 0) {
13293 				drv_usecwait(MILLISEC);
13294 			} else {
13295 				rval = QL_FUNCTION_TIMEOUT;
13296 				break;
13297 			}
13298 		}
13299 	}
13300 
13301 	if (rval == QL_SUCCESS) {
13302 		/* Set memory configuration and timing. */
13303 		WRT16_IO_REG(ha, mctr, 0xf2);
13304 
13305 		/* Release RISC. */
13306 		WRT16_IO_REG(ha, hccr, HC_RELEASE_RISC);
13307 
13308 		/* Get RISC SRAM. */
13309 		risc_address = 0x1000;
13310 		WRT16_IO_REG(ha, mailbox_in[0], MBC_READ_RAM_WORD);
13311 		for (cnt = 0; cnt < 0xf000; cnt++) {
13312 			WRT16_IO_REG(ha, mailbox_in[1], risc_address++);
13313 			WRT16_IO_REG(ha, hccr, HC_SET_HOST_INT);
13314 			for (timer = 6000000; timer != 0; timer--) {
13315 				/* Check for pending interrupts. */
13316 				if (INTERRUPT_PENDING(ha)) {
13317 					if (RD16_IO_REG(ha, semaphore) &
13318 					    BIT_0) {
13319 						WRT16_IO_REG(ha, hccr,
13320 						    HC_CLR_RISC_INT);
13321 						mcp->mb[0] = RD16_IO_REG(ha,
13322 						    mailbox_out[0]);
13323 						fw->risc_ram[cnt] =
13324 						    RD16_IO_REG(ha,
13325 						    mailbox_out[2]);
13326 						WRT16_IO_REG(ha,
13327 						    semaphore, 0);
13328 						break;
13329 					}
13330 					WRT16_IO_REG(ha, hccr,
13331 					    HC_CLR_RISC_INT);
13332 				}
13333 				drv_usecwait(5);
13334 			}
13335 
13336 			if (timer == 0) {
13337 				rval = QL_FUNCTION_TIMEOUT;
13338 			} else {
13339 				rval = mcp->mb[0];
13340 			}
13341 
13342 			if (rval != QL_SUCCESS) {
13343 				break;
13344 			}
13345 		}
13346 	}
13347 
13348 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
13349 
13350 	return (rval);
13351 }
13352 
13353 /*
13354  * ql_2300_binary_fw_dump
13355  *
13356  * Input:
13357  *	ha:	adapter state pointer.
13358  *	fw:	firmware dump context pointer.
13359  *
13360  * Returns:
13361  *	ql local function return status code.
13362  *
13363  * Context:
13364  *	Interrupt or Kernel context, no mailbox commands allowed.
13365  */
13366 static int
ql_2300_binary_fw_dump(ql_adapter_state_t * ha,ql_fw_dump_t * fw)13367 ql_2300_binary_fw_dump(ql_adapter_state_t *ha, ql_fw_dump_t *fw)
13368 {
13369 	clock_t	timer;
13370 	int	rval = QL_SUCCESS;
13371 
13372 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
13373 
13374 	/* Disable ISP interrupts. */
13375 	WRT16_IO_REG(ha, ictrl, 0);
13376 	ADAPTER_STATE_LOCK(ha);
13377 	ha->flags &= ~INTERRUPTS_ENABLED;
13378 	ADAPTER_STATE_UNLOCK(ha);
13379 
13380 	/* Release mailbox registers. */
13381 	WRT16_IO_REG(ha, semaphore, 0);
13382 
13383 	/* Pause RISC. */
13384 	WRT16_IO_REG(ha, hccr, HC_PAUSE_RISC);
13385 	timer = 30000;
13386 	while ((RD16_IO_REG(ha, hccr) & HC_RISC_PAUSE) == 0) {
13387 		if (timer-- != 0) {
13388 			drv_usecwait(MILLISEC);
13389 		} else {
13390 			rval = QL_FUNCTION_TIMEOUT;
13391 			break;
13392 		}
13393 	}
13394 
13395 	if (rval == QL_SUCCESS) {
13396 		(void) ql_read_regs(ha, fw->pbiu_reg, ha->iobase,
13397 		    sizeof (fw->pbiu_reg) / 2, 16);
13398 
13399 		(void) ql_read_regs(ha, fw->risc_host_reg, ha->iobase + 0x10,
13400 		    sizeof (fw->risc_host_reg) / 2, 16);
13401 
13402 		(void) ql_read_regs(ha, fw->mailbox_reg, ha->iobase + 0x40,
13403 		    sizeof (fw->mailbox_reg) / 2, 16);
13404 
13405 		WRT16_IO_REG(ha, ctrl_status, 0x40);
13406 		(void) ql_read_regs(ha, fw->resp_dma_reg, ha->iobase + 0x80,
13407 		    sizeof (fw->resp_dma_reg) / 2, 16);
13408 
13409 		WRT16_IO_REG(ha, ctrl_status, 0x50);
13410 		(void) ql_read_regs(ha, fw->dma_reg, ha->iobase + 0x80,
13411 		    sizeof (fw->dma_reg) / 2, 16);
13412 
13413 		WRT16_IO_REG(ha, ctrl_status, 0);
13414 		(void) ql_read_regs(ha, fw->risc_hdw_reg, ha->iobase + 0xA0,
13415 		    sizeof (fw->risc_hdw_reg) / 2, 16);
13416 
13417 		WRT16_IO_REG(ha, pcr, 0x2000);
13418 		(void) ql_read_regs(ha, fw->risc_gp0_reg, ha->iobase + 0x80,
13419 		    sizeof (fw->risc_gp0_reg) / 2, 16);
13420 
13421 		WRT16_IO_REG(ha, pcr, 0x2200);
13422 		(void) ql_read_regs(ha, fw->risc_gp1_reg, ha->iobase + 0x80,
13423 		    sizeof (fw->risc_gp1_reg) / 2, 16);
13424 
13425 		WRT16_IO_REG(ha, pcr, 0x2400);
13426 		(void) ql_read_regs(ha, fw->risc_gp2_reg, ha->iobase + 0x80,
13427 		    sizeof (fw->risc_gp2_reg) / 2, 16);
13428 
13429 		WRT16_IO_REG(ha, pcr, 0x2600);
13430 		(void) ql_read_regs(ha, fw->risc_gp3_reg, ha->iobase + 0x80,
13431 		    sizeof (fw->risc_gp3_reg) / 2, 16);
13432 
13433 		WRT16_IO_REG(ha, pcr, 0x2800);
13434 		(void) ql_read_regs(ha, fw->risc_gp4_reg, ha->iobase + 0x80,
13435 		    sizeof (fw->risc_gp4_reg) / 2, 16);
13436 
13437 		WRT16_IO_REG(ha, pcr, 0x2A00);
13438 		(void) ql_read_regs(ha, fw->risc_gp5_reg, ha->iobase + 0x80,
13439 		    sizeof (fw->risc_gp5_reg) / 2, 16);
13440 
13441 		WRT16_IO_REG(ha, pcr, 0x2C00);
13442 		(void) ql_read_regs(ha, fw->risc_gp6_reg, ha->iobase + 0x80,
13443 		    sizeof (fw->risc_gp6_reg) / 2, 16);
13444 
13445 		WRT16_IO_REG(ha, pcr, 0x2E00);
13446 		(void) ql_read_regs(ha, fw->risc_gp7_reg, ha->iobase + 0x80,
13447 		    sizeof (fw->risc_gp7_reg) / 2, 16);
13448 
13449 		WRT16_IO_REG(ha, ctrl_status, 0x10);
13450 		(void) ql_read_regs(ha, fw->frame_buf_hdw_reg,
13451 		    ha->iobase + 0x80, sizeof (fw->frame_buf_hdw_reg) / 2, 16);
13452 
13453 		WRT16_IO_REG(ha, ctrl_status, 0x20);
13454 		(void) ql_read_regs(ha, fw->fpm_b0_reg, ha->iobase + 0x80,
13455 		    sizeof (fw->fpm_b0_reg) / 2, 16);
13456 
13457 		WRT16_IO_REG(ha, ctrl_status, 0x30);
13458 		(void) ql_read_regs(ha, fw->fpm_b1_reg, ha->iobase + 0x80,
13459 		    sizeof (fw->fpm_b1_reg) / 2, 16);
13460 
13461 		/* Select FPM registers. */
13462 		WRT16_IO_REG(ha, ctrl_status, 0x20);
13463 
13464 		/* FPM Soft Reset. */
13465 		WRT16_IO_REG(ha, fpm_diag_config, 0x100);
13466 
13467 		/* Select frame buffer registers. */
13468 		WRT16_IO_REG(ha, ctrl_status, 0x10);
13469 
13470 		/* Reset frame buffer FIFOs. */
13471 		WRT16_IO_REG(ha, fb_cmd, 0xa000);
13472 
13473 		/* Select RISC module registers. */
13474 		WRT16_IO_REG(ha, ctrl_status, 0);
13475 
13476 		/* Reset RISC module. */
13477 		WRT16_IO_REG(ha, hccr, HC_RESET_RISC);
13478 
13479 		/* Reset ISP semaphore. */
13480 		WRT16_IO_REG(ha, semaphore, 0);
13481 
13482 		/* Release RISC module. */
13483 		WRT16_IO_REG(ha, hccr, HC_RELEASE_RISC);
13484 
13485 		/* Wait for RISC to recover from reset. */
13486 		timer = 30000;
13487 		while (RD16_IO_REG(ha, mailbox_out[0]) == MBS_BUSY) {
13488 			if (timer-- != 0) {
13489 				drv_usecwait(MILLISEC);
13490 			} else {
13491 				rval = QL_FUNCTION_TIMEOUT;
13492 				break;
13493 			}
13494 		}
13495 
13496 		/* Disable RISC pause on FPM parity error. */
13497 		WRT16_IO_REG(ha, hccr, HC_DISABLE_PARITY_PAUSE);
13498 	}
13499 
13500 	/* Get RISC SRAM. */
13501 	if (rval == QL_SUCCESS) {
13502 		rval = ql_read_risc_ram(ha, 0x800, 0xf800, fw->risc_ram);
13503 	}
13504 	/* Get STACK SRAM. */
13505 	if (rval == QL_SUCCESS) {
13506 		rval = ql_read_risc_ram(ha, 0x10000, 0x800, fw->stack_ram);
13507 	}
13508 	/* Get DATA SRAM. */
13509 	if (rval == QL_SUCCESS) {
13510 		rval = ql_read_risc_ram(ha, 0x10800, 0xf800, fw->data_ram);
13511 	}
13512 
13513 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
13514 
13515 	return (rval);
13516 }
13517 
13518 /*
13519  * ql_24xx_binary_fw_dump
13520  *
13521  * Input:
13522  *	ha:	adapter state pointer.
13523  *	fw:	firmware dump context pointer.
13524  *
13525  * Returns:
13526  *	ql local function return status code.
13527  *
13528  * Context:
13529  *	Interrupt or Kernel context, no mailbox commands allowed.
13530  */
13531 static int
ql_24xx_binary_fw_dump(ql_adapter_state_t * ha,ql_24xx_fw_dump_t * fw)13532 ql_24xx_binary_fw_dump(ql_adapter_state_t *ha, ql_24xx_fw_dump_t *fw)
13533 {
13534 	uint32_t	*reg32;
13535 	void		*bp;
13536 	clock_t		timer;
13537 	int		rval = QL_SUCCESS;
13538 
13539 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
13540 
13541 	fw->hccr = RD32_IO_REG(ha, hccr);
13542 
13543 	/* Pause RISC. */
13544 	if ((RD32_IO_REG(ha, risc2host) & RH_RISC_PAUSED) == 0) {
13545 		/* Disable ISP interrupts. */
13546 		WRT16_IO_REG(ha, ictrl, 0);
13547 
13548 		WRT32_IO_REG(ha, hccr, HC24_PAUSE_RISC);
13549 		for (timer = 30000;
13550 		    (RD32_IO_REG(ha, risc2host) & RH_RISC_PAUSED) == 0 &&
13551 		    rval == QL_SUCCESS; timer--) {
13552 			if (timer) {
13553 				drv_usecwait(100);
13554 			} else {
13555 				rval = QL_FUNCTION_TIMEOUT;
13556 			}
13557 		}
13558 	}
13559 
13560 	if (rval == QL_SUCCESS) {
13561 		/* Host interface registers. */
13562 		(void) ql_read_regs(ha, fw->host_reg, ha->iobase,
13563 		    sizeof (fw->host_reg) / 4, 32);
13564 
13565 		/* Disable ISP interrupts. */
13566 		WRT32_IO_REG(ha, ictrl, 0);
13567 		RD32_IO_REG(ha, ictrl);
13568 		ADAPTER_STATE_LOCK(ha);
13569 		ha->flags &= ~INTERRUPTS_ENABLED;
13570 		ADAPTER_STATE_UNLOCK(ha);
13571 
13572 		/* Shadow registers. */
13573 
13574 		WRT32_IO_REG(ha, io_base_addr, 0x0F70);
13575 		RD32_IO_REG(ha, io_base_addr);
13576 
13577 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
13578 		WRT_REG_DWORD(ha, reg32, 0xB0000000);
13579 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
13580 		fw->shadow_reg[0] = RD_REG_DWORD(ha, reg32);
13581 
13582 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
13583 		WRT_REG_DWORD(ha, reg32, 0xB0100000);
13584 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
13585 		fw->shadow_reg[1] = RD_REG_DWORD(ha, reg32);
13586 
13587 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
13588 		WRT_REG_DWORD(ha, reg32, 0xB0200000);
13589 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
13590 		fw->shadow_reg[2] = RD_REG_DWORD(ha, reg32);
13591 
13592 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
13593 		WRT_REG_DWORD(ha, reg32, 0xB0300000);
13594 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
13595 		fw->shadow_reg[3] = RD_REG_DWORD(ha, reg32);
13596 
13597 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
13598 		WRT_REG_DWORD(ha, reg32, 0xB0400000);
13599 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
13600 		fw->shadow_reg[4] = RD_REG_DWORD(ha, reg32);
13601 
13602 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
13603 		WRT_REG_DWORD(ha, reg32, 0xB0500000);
13604 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
13605 		fw->shadow_reg[5] = RD_REG_DWORD(ha, reg32);
13606 
13607 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
13608 		WRT_REG_DWORD(ha, reg32, 0xB0600000);
13609 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
13610 		fw->shadow_reg[6] = RD_REG_DWORD(ha, reg32);
13611 
13612 		/* Mailbox registers. */
13613 		(void) ql_read_regs(ha, fw->mailbox_reg, ha->iobase + 0x80,
13614 		    sizeof (fw->mailbox_reg) / 2, 16);
13615 
13616 		/* Transfer sequence registers. */
13617 
13618 		/* XSEQ GP */
13619 		WRT32_IO_REG(ha, io_base_addr, 0xBF00);
13620 		bp = ql_read_regs(ha, fw->xseq_gp_reg, ha->iobase + 0xC0,
13621 		    16, 32);
13622 		WRT32_IO_REG(ha, io_base_addr, 0xBF10);
13623 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13624 		WRT32_IO_REG(ha, io_base_addr, 0xBF20);
13625 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13626 		WRT32_IO_REG(ha, io_base_addr, 0xBF30);
13627 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13628 		WRT32_IO_REG(ha, io_base_addr, 0xBF40);
13629 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13630 		WRT32_IO_REG(ha, io_base_addr, 0xBF50);
13631 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13632 		WRT32_IO_REG(ha, io_base_addr, 0xBF60);
13633 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13634 		WRT32_IO_REG(ha, io_base_addr, 0xBF70);
13635 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13636 
13637 		/* XSEQ-0 */
13638 		WRT32_IO_REG(ha, io_base_addr, 0xBFE0);
13639 		(void) ql_read_regs(ha, fw->xseq_0_reg, ha->iobase + 0xC0,
13640 		    sizeof (fw->xseq_0_reg) / 4, 32);
13641 
13642 		/* XSEQ-1 */
13643 		WRT32_IO_REG(ha, io_base_addr, 0xBFF0);
13644 		(void) ql_read_regs(ha, fw->xseq_1_reg, ha->iobase + 0xC0,
13645 		    sizeof (fw->xseq_1_reg) / 4, 32);
13646 
13647 		/* Receive sequence registers. */
13648 
13649 		/* RSEQ GP */
13650 		WRT32_IO_REG(ha, io_base_addr, 0xFF00);
13651 		bp = ql_read_regs(ha, fw->rseq_gp_reg, ha->iobase + 0xC0,
13652 		    16, 32);
13653 		WRT32_IO_REG(ha, io_base_addr, 0xFF10);
13654 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13655 		WRT32_IO_REG(ha, io_base_addr, 0xFF20);
13656 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13657 		WRT32_IO_REG(ha, io_base_addr, 0xFF30);
13658 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13659 		WRT32_IO_REG(ha, io_base_addr, 0xFF40);
13660 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13661 		WRT32_IO_REG(ha, io_base_addr, 0xFF50);
13662 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13663 		WRT32_IO_REG(ha, io_base_addr, 0xFF60);
13664 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13665 		WRT32_IO_REG(ha, io_base_addr, 0xFF70);
13666 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13667 
13668 		/* RSEQ-0 */
13669 		WRT32_IO_REG(ha, io_base_addr, 0xFFD0);
13670 		(void) ql_read_regs(ha, fw->rseq_0_reg, ha->iobase + 0xC0,
13671 		    sizeof (fw->rseq_0_reg) / 4, 32);
13672 
13673 		/* RSEQ-1 */
13674 		WRT32_IO_REG(ha, io_base_addr, 0xFFE0);
13675 		(void) ql_read_regs(ha, fw->rseq_1_reg, ha->iobase + 0xC0,
13676 		    sizeof (fw->rseq_1_reg) / 4, 32);
13677 
13678 		/* RSEQ-2 */
13679 		WRT32_IO_REG(ha, io_base_addr, 0xFFF0);
13680 		(void) ql_read_regs(ha, fw->rseq_2_reg, ha->iobase + 0xC0,
13681 		    sizeof (fw->rseq_2_reg) / 4, 32);
13682 
13683 		/* Command DMA registers. */
13684 
13685 		WRT32_IO_REG(ha, io_base_addr, 0x7100);
13686 		(void) ql_read_regs(ha, fw->cmd_dma_reg, ha->iobase + 0xC0,
13687 		    sizeof (fw->cmd_dma_reg) / 4, 32);
13688 
13689 		/* Queues. */
13690 
13691 		/* RequestQ0 */
13692 		WRT32_IO_REG(ha, io_base_addr, 0x7200);
13693 		bp = ql_read_regs(ha, fw->req0_dma_reg, ha->iobase + 0xC0,
13694 		    8, 32);
13695 		(void) ql_read_regs(ha, bp, ha->iobase + 0xE4, 7, 32);
13696 
13697 		/* ResponseQ0 */
13698 		WRT32_IO_REG(ha, io_base_addr, 0x7300);
13699 		bp = ql_read_regs(ha, fw->resp0_dma_reg, ha->iobase + 0xC0,
13700 		    8, 32);
13701 		(void) ql_read_regs(ha, bp, ha->iobase + 0xE4, 7, 32);
13702 
13703 		/* RequestQ1 */
13704 		WRT32_IO_REG(ha, io_base_addr, 0x7400);
13705 		bp = ql_read_regs(ha, fw->req1_dma_reg, ha->iobase + 0xC0,
13706 		    8, 32);
13707 		(void) ql_read_regs(ha, bp, ha->iobase + 0xE4, 7, 32);
13708 
13709 		/* Transmit DMA registers. */
13710 
13711 		/* XMT0 */
13712 		WRT32_IO_REG(ha, io_base_addr, 0x7600);
13713 		bp = ql_read_regs(ha, fw->xmt0_dma_reg, ha->iobase + 0xC0,
13714 		    16, 32);
13715 		WRT32_IO_REG(ha, io_base_addr, 0x7610);
13716 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13717 
13718 		/* XMT1 */
13719 		WRT32_IO_REG(ha, io_base_addr, 0x7620);
13720 		bp = ql_read_regs(ha, fw->xmt1_dma_reg, ha->iobase + 0xC0,
13721 		    16, 32);
13722 		WRT32_IO_REG(ha, io_base_addr, 0x7630);
13723 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13724 
13725 		/* XMT2 */
13726 		WRT32_IO_REG(ha, io_base_addr, 0x7640);
13727 		bp = ql_read_regs(ha, fw->xmt2_dma_reg, ha->iobase + 0xC0,
13728 		    16, 32);
13729 		WRT32_IO_REG(ha, io_base_addr, 0x7650);
13730 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13731 
13732 		/* XMT3 */
13733 		WRT32_IO_REG(ha, io_base_addr, 0x7660);
13734 		bp = ql_read_regs(ha, fw->xmt3_dma_reg, ha->iobase + 0xC0,
13735 		    16, 32);
13736 		WRT32_IO_REG(ha, io_base_addr, 0x7670);
13737 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13738 
13739 		/* XMT4 */
13740 		WRT32_IO_REG(ha, io_base_addr, 0x7680);
13741 		bp = ql_read_regs(ha, fw->xmt4_dma_reg, ha->iobase + 0xC0,
13742 		    16, 32);
13743 		WRT32_IO_REG(ha, io_base_addr, 0x7690);
13744 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13745 
13746 		/* XMT Common */
13747 		WRT32_IO_REG(ha, io_base_addr, 0x76A0);
13748 		(void) ql_read_regs(ha, fw->xmt_data_dma_reg,
13749 		    ha->iobase + 0xC0, sizeof (fw->xmt_data_dma_reg) / 4, 32);
13750 
13751 		/* Receive DMA registers. */
13752 
13753 		/* RCVThread0 */
13754 		WRT32_IO_REG(ha, io_base_addr, 0x7700);
13755 		bp = ql_read_regs(ha, fw->rcvt0_data_dma_reg,
13756 		    ha->iobase + 0xC0, 16, 32);
13757 		WRT32_IO_REG(ha, io_base_addr, 0x7710);
13758 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13759 
13760 		/* RCVThread1 */
13761 		WRT32_IO_REG(ha, io_base_addr, 0x7720);
13762 		bp = ql_read_regs(ha, fw->rcvt1_data_dma_reg,
13763 		    ha->iobase + 0xC0, 16, 32);
13764 		WRT32_IO_REG(ha, io_base_addr, 0x7730);
13765 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13766 
13767 		/* RISC registers. */
13768 
13769 		/* RISC GP */
13770 		WRT32_IO_REG(ha, io_base_addr, 0x0F00);
13771 		bp = ql_read_regs(ha, fw->risc_gp_reg, ha->iobase + 0xC0,
13772 		    16, 32);
13773 		WRT32_IO_REG(ha, io_base_addr, 0x0F10);
13774 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13775 		WRT32_IO_REG(ha, io_base_addr, 0x0F20);
13776 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13777 		WRT32_IO_REG(ha, io_base_addr, 0x0F30);
13778 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13779 		WRT32_IO_REG(ha, io_base_addr, 0x0F40);
13780 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13781 		WRT32_IO_REG(ha, io_base_addr, 0x0F50);
13782 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13783 		WRT32_IO_REG(ha, io_base_addr, 0x0F60);
13784 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13785 		WRT32_IO_REG(ha, io_base_addr, 0x0F70);
13786 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13787 
13788 		/* Local memory controller registers. */
13789 
13790 		/* LMC */
13791 		WRT32_IO_REG(ha, io_base_addr, 0x3000);
13792 		bp = ql_read_regs(ha, fw->lmc_reg, ha->iobase + 0xC0,
13793 		    16, 32);
13794 		WRT32_IO_REG(ha, io_base_addr, 0x3010);
13795 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13796 		WRT32_IO_REG(ha, io_base_addr, 0x3020);
13797 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13798 		WRT32_IO_REG(ha, io_base_addr, 0x3030);
13799 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13800 		WRT32_IO_REG(ha, io_base_addr, 0x3040);
13801 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13802 		WRT32_IO_REG(ha, io_base_addr, 0x3050);
13803 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13804 		WRT32_IO_REG(ha, io_base_addr, 0x3060);
13805 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13806 
13807 		/* Fibre Protocol Module registers. */
13808 
13809 		/* FPM hardware */
13810 		WRT32_IO_REG(ha, io_base_addr, 0x4000);
13811 		bp = ql_read_regs(ha, fw->fpm_hdw_reg, ha->iobase + 0xC0,
13812 		    16, 32);
13813 		WRT32_IO_REG(ha, io_base_addr, 0x4010);
13814 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13815 		WRT32_IO_REG(ha, io_base_addr, 0x4020);
13816 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13817 		WRT32_IO_REG(ha, io_base_addr, 0x4030);
13818 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13819 		WRT32_IO_REG(ha, io_base_addr, 0x4040);
13820 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13821 		WRT32_IO_REG(ha, io_base_addr, 0x4050);
13822 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13823 		WRT32_IO_REG(ha, io_base_addr, 0x4060);
13824 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13825 		WRT32_IO_REG(ha, io_base_addr, 0x4070);
13826 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13827 		WRT32_IO_REG(ha, io_base_addr, 0x4080);
13828 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13829 		WRT32_IO_REG(ha, io_base_addr, 0x4090);
13830 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13831 		WRT32_IO_REG(ha, io_base_addr, 0x40A0);
13832 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13833 		WRT32_IO_REG(ha, io_base_addr, 0x40B0);
13834 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13835 
13836 		/* Frame Buffer registers. */
13837 
13838 		/* FB hardware */
13839 		WRT32_IO_REG(ha, io_base_addr, 0x6000);
13840 		bp = ql_read_regs(ha, fw->fb_hdw_reg, ha->iobase + 0xC0,
13841 		    16, 32);
13842 		WRT32_IO_REG(ha, io_base_addr, 0x6010);
13843 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13844 		WRT32_IO_REG(ha, io_base_addr, 0x6020);
13845 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13846 		WRT32_IO_REG(ha, io_base_addr, 0x6030);
13847 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13848 		WRT32_IO_REG(ha, io_base_addr, 0x6040);
13849 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13850 		WRT32_IO_REG(ha, io_base_addr, 0x6100);
13851 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13852 		WRT32_IO_REG(ha, io_base_addr, 0x6130);
13853 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13854 		WRT32_IO_REG(ha, io_base_addr, 0x6150);
13855 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13856 		WRT32_IO_REG(ha, io_base_addr, 0x6170);
13857 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13858 		WRT32_IO_REG(ha, io_base_addr, 0x6190);
13859 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13860 		WRT32_IO_REG(ha, io_base_addr, 0x61B0);
13861 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
13862 	}
13863 
13864 	/* Get the request queue */
13865 	if (rval == QL_SUCCESS) {
13866 		uint32_t	cnt;
13867 		uint32_t	*w32 = (uint32_t *)ha->request_ring_bp;
13868 
13869 		/* Sync DMA buffer. */
13870 		(void) ddi_dma_sync(ha->hba_buf.dma_handle,
13871 		    REQUEST_Q_BUFFER_OFFSET, sizeof (fw->req_q),
13872 		    DDI_DMA_SYNC_FORKERNEL);
13873 
13874 		for (cnt = 0; cnt < sizeof (fw->req_q) / 4; cnt++) {
13875 			fw->req_q[cnt] = *w32++;
13876 			LITTLE_ENDIAN_32(&fw->req_q[cnt]);
13877 		}
13878 	}
13879 
13880 	/* Get the response queue */
13881 	if (rval == QL_SUCCESS) {
13882 		uint32_t	cnt;
13883 		uint32_t	*w32 = (uint32_t *)ha->response_ring_bp;
13884 
13885 		/* Sync DMA buffer. */
13886 		(void) ddi_dma_sync(ha->hba_buf.dma_handle,
13887 		    RESPONSE_Q_BUFFER_OFFSET, sizeof (fw->rsp_q),
13888 		    DDI_DMA_SYNC_FORKERNEL);
13889 
13890 		for (cnt = 0; cnt < sizeof (fw->rsp_q) / 4; cnt++) {
13891 			fw->rsp_q[cnt] = *w32++;
13892 			LITTLE_ENDIAN_32(&fw->rsp_q[cnt]);
13893 		}
13894 	}
13895 
13896 	/* Reset RISC. */
13897 	ql_reset_chip(ha);
13898 
13899 	/* Memory. */
13900 	if (rval == QL_SUCCESS) {
13901 		/* Code RAM. */
13902 		rval = ql_read_risc_ram(ha, 0x20000,
13903 		    sizeof (fw->code_ram) / 4, fw->code_ram);
13904 	}
13905 	if (rval == QL_SUCCESS) {
13906 		/* External Memory. */
13907 		rval = ql_read_risc_ram(ha, 0x100000,
13908 		    ha->fw_ext_memory_size / 4, fw->ext_mem);
13909 	}
13910 
13911 	/* Get the extended trace buffer */
13912 	if (rval == QL_SUCCESS) {
13913 		if (CFG_IST(ha, CFG_ENABLE_FWEXTTRACE) &&
13914 		    (ha->fwexttracebuf.bp != NULL)) {
13915 			uint32_t	cnt;
13916 			uint32_t	*w32 = ha->fwexttracebuf.bp;
13917 
13918 			/* Sync DMA buffer. */
13919 			(void) ddi_dma_sync(ha->fwexttracebuf.dma_handle, 0,
13920 			    FWEXTSIZE, DDI_DMA_SYNC_FORKERNEL);
13921 
13922 			for (cnt = 0; cnt < FWEXTSIZE / 4; cnt++) {
13923 				fw->ext_trace_buf[cnt] = *w32++;
13924 			}
13925 		}
13926 	}
13927 
13928 	/* Get the FC event trace buffer */
13929 	if (rval == QL_SUCCESS) {
13930 		if (CFG_IST(ha, CFG_ENABLE_FWFCETRACE) &&
13931 		    (ha->fwfcetracebuf.bp != NULL)) {
13932 			uint32_t	cnt;
13933 			uint32_t	*w32 = ha->fwfcetracebuf.bp;
13934 
13935 			/* Sync DMA buffer. */
13936 			(void) ddi_dma_sync(ha->fwfcetracebuf.dma_handle, 0,
13937 			    FWFCESIZE, DDI_DMA_SYNC_FORKERNEL);
13938 
13939 			for (cnt = 0; cnt < FWFCESIZE / 4; cnt++) {
13940 				fw->fce_trace_buf[cnt] = *w32++;
13941 			}
13942 		}
13943 	}
13944 
13945 	if (rval != QL_SUCCESS) {
13946 		EL(ha, "failed=%xh\n", rval);
13947 	} else {
13948 		/*EMPTY*/
13949 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
13950 	}
13951 
13952 	return (rval);
13953 }
13954 
13955 /*
13956  * ql_25xx_binary_fw_dump
13957  *
13958  * Input:
13959  *	ha:	adapter state pointer.
13960  *	fw:	firmware dump context pointer.
13961  *
13962  * Returns:
13963  *	ql local function return status code.
13964  *
13965  * Context:
13966  *	Interrupt or Kernel context, no mailbox commands allowed.
13967  */
13968 static int
ql_25xx_binary_fw_dump(ql_adapter_state_t * ha,ql_25xx_fw_dump_t * fw)13969 ql_25xx_binary_fw_dump(ql_adapter_state_t *ha, ql_25xx_fw_dump_t *fw)
13970 {
13971 	uint32_t	*reg32;
13972 	void		*bp;
13973 	clock_t		timer;
13974 	int		rval = QL_SUCCESS;
13975 
13976 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
13977 
13978 	fw->r2h_status = RD32_IO_REG(ha, risc2host);
13979 
13980 	/* Pause RISC. */
13981 	if ((RD32_IO_REG(ha, risc2host) & RH_RISC_PAUSED) == 0) {
13982 		/* Disable ISP interrupts. */
13983 		WRT16_IO_REG(ha, ictrl, 0);
13984 
13985 		WRT32_IO_REG(ha, hccr, HC24_PAUSE_RISC);
13986 		for (timer = 30000;
13987 		    (RD32_IO_REG(ha, risc2host) & RH_RISC_PAUSED) == 0 &&
13988 		    rval == QL_SUCCESS; timer--) {
13989 			if (timer) {
13990 				drv_usecwait(100);
13991 				if (timer % 10000 == 0) {
13992 					EL(ha, "risc pause %d\n", timer);
13993 				}
13994 			} else {
13995 				EL(ha, "risc pause timeout\n");
13996 				rval = QL_FUNCTION_TIMEOUT;
13997 			}
13998 		}
13999 	}
14000 
14001 	if (rval == QL_SUCCESS) {
14002 
14003 		/* Host Interface registers */
14004 
14005 		/* HostRisc registers. */
14006 		WRT32_IO_REG(ha, io_base_addr, 0x7000);
14007 		bp = ql_read_regs(ha, fw->hostrisc_reg, ha->iobase + 0xC0,
14008 		    16, 32);
14009 		WRT32_IO_REG(ha, io_base_addr, 0x7010);
14010 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14011 
14012 		/* PCIe registers. */
14013 		WRT32_IO_REG(ha, io_base_addr, 0x7c00);
14014 		WRT_REG_DWORD(ha, ha->iobase + 0xc0, 0x1);
14015 		bp = ql_read_regs(ha, fw->pcie_reg, ha->iobase + 0xC4,
14016 		    3, 32);
14017 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 1, 32);
14018 		WRT_REG_DWORD(ha, ha->iobase + 0xc0, 0x0);
14019 
14020 		/* Host interface registers. */
14021 		(void) ql_read_regs(ha, fw->host_reg, ha->iobase,
14022 		    sizeof (fw->host_reg) / 4, 32);
14023 
14024 		/* Disable ISP interrupts. */
14025 
14026 		WRT32_IO_REG(ha, ictrl, 0);
14027 		RD32_IO_REG(ha, ictrl);
14028 		ADAPTER_STATE_LOCK(ha);
14029 		ha->flags &= ~INTERRUPTS_ENABLED;
14030 		ADAPTER_STATE_UNLOCK(ha);
14031 
14032 		/* Shadow registers. */
14033 
14034 		WRT32_IO_REG(ha, io_base_addr, 0x0F70);
14035 		RD32_IO_REG(ha, io_base_addr);
14036 
14037 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
14038 		WRT_REG_DWORD(ha, reg32, 0xB0000000);
14039 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
14040 		fw->shadow_reg[0] = RD_REG_DWORD(ha, reg32);
14041 
14042 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
14043 		WRT_REG_DWORD(ha, reg32, 0xB0100000);
14044 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
14045 		fw->shadow_reg[1] = RD_REG_DWORD(ha, reg32);
14046 
14047 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
14048 		WRT_REG_DWORD(ha, reg32, 0xB0200000);
14049 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
14050 		fw->shadow_reg[2] = RD_REG_DWORD(ha, reg32);
14051 
14052 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
14053 		WRT_REG_DWORD(ha, reg32, 0xB0300000);
14054 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
14055 		fw->shadow_reg[3] = RD_REG_DWORD(ha, reg32);
14056 
14057 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
14058 		WRT_REG_DWORD(ha, reg32, 0xB0400000);
14059 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
14060 		fw->shadow_reg[4] = RD_REG_DWORD(ha, reg32);
14061 
14062 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
14063 		WRT_REG_DWORD(ha, reg32, 0xB0500000);
14064 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
14065 		fw->shadow_reg[5] = RD_REG_DWORD(ha, reg32);
14066 
14067 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
14068 		WRT_REG_DWORD(ha, reg32, 0xB0600000);
14069 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
14070 		fw->shadow_reg[6] = RD_REG_DWORD(ha, reg32);
14071 
14072 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
14073 		WRT_REG_DWORD(ha, reg32, 0xB0700000);
14074 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
14075 		fw->shadow_reg[7] = RD_REG_DWORD(ha, reg32);
14076 
14077 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
14078 		WRT_REG_DWORD(ha, reg32, 0xB0800000);
14079 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
14080 		fw->shadow_reg[8] = RD_REG_DWORD(ha, reg32);
14081 
14082 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
14083 		WRT_REG_DWORD(ha, reg32, 0xB0900000);
14084 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
14085 		fw->shadow_reg[9] = RD_REG_DWORD(ha, reg32);
14086 
14087 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
14088 		WRT_REG_DWORD(ha, reg32, 0xB0A00000);
14089 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
14090 		fw->shadow_reg[0xa] = RD_REG_DWORD(ha, reg32);
14091 
14092 		/* RISC I/O register. */
14093 
14094 		WRT32_IO_REG(ha, io_base_addr, 0x0010);
14095 		(void) ql_read_regs(ha, &fw->risc_io, ha->iobase + 0xC0,
14096 		    1, 32);
14097 
14098 		/* Mailbox registers. */
14099 
14100 		(void) ql_read_regs(ha, fw->mailbox_reg, ha->iobase + 0x80,
14101 		    sizeof (fw->mailbox_reg) / 2, 16);
14102 
14103 		/* Transfer sequence registers. */
14104 
14105 		/* XSEQ GP */
14106 		WRT32_IO_REG(ha, io_base_addr, 0xBF00);
14107 		bp = ql_read_regs(ha, fw->xseq_gp_reg, ha->iobase + 0xC0,
14108 		    16, 32);
14109 		WRT32_IO_REG(ha, io_base_addr, 0xBF10);
14110 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14111 		WRT32_IO_REG(ha, io_base_addr, 0xBF20);
14112 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14113 		WRT32_IO_REG(ha, io_base_addr, 0xBF30);
14114 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14115 		WRT32_IO_REG(ha, io_base_addr, 0xBF40);
14116 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14117 		WRT32_IO_REG(ha, io_base_addr, 0xBF50);
14118 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14119 		WRT32_IO_REG(ha, io_base_addr, 0xBF60);
14120 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14121 		WRT32_IO_REG(ha, io_base_addr, 0xBF70);
14122 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14123 
14124 		/* XSEQ-0 */
14125 		WRT32_IO_REG(ha, io_base_addr, 0xBFC0);
14126 		bp = ql_read_regs(ha, fw->xseq_0_reg, ha->iobase + 0xC0,
14127 		    16, 32);
14128 		WRT32_IO_REG(ha, io_base_addr, 0xBFD0);
14129 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14130 		WRT32_IO_REG(ha, io_base_addr, 0xBFE0);
14131 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14132 
14133 		/* XSEQ-1 */
14134 		WRT32_IO_REG(ha, io_base_addr, 0xBFF0);
14135 		(void) ql_read_regs(ha, fw->xseq_1_reg, ha->iobase + 0xC0,
14136 		    16, 32);
14137 
14138 		/* Receive sequence registers. */
14139 
14140 		/* RSEQ GP */
14141 		WRT32_IO_REG(ha, io_base_addr, 0xFF00);
14142 		bp = ql_read_regs(ha, fw->rseq_gp_reg, ha->iobase + 0xC0,
14143 		    16, 32);
14144 		WRT32_IO_REG(ha, io_base_addr, 0xFF10);
14145 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14146 		WRT32_IO_REG(ha, io_base_addr, 0xFF20);
14147 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14148 		WRT32_IO_REG(ha, io_base_addr, 0xFF30);
14149 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14150 		WRT32_IO_REG(ha, io_base_addr, 0xFF40);
14151 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14152 		WRT32_IO_REG(ha, io_base_addr, 0xFF50);
14153 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14154 		WRT32_IO_REG(ha, io_base_addr, 0xFF60);
14155 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14156 		WRT32_IO_REG(ha, io_base_addr, 0xFF70);
14157 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14158 
14159 		/* RSEQ-0 */
14160 		WRT32_IO_REG(ha, io_base_addr, 0xFFC0);
14161 		bp = ql_read_regs(ha, fw->rseq_0_reg, ha->iobase + 0xC0,
14162 		    16, 32);
14163 		WRT32_IO_REG(ha, io_base_addr, 0xFFD0);
14164 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14165 
14166 		/* RSEQ-1 */
14167 		WRT32_IO_REG(ha, io_base_addr, 0xFFE0);
14168 		(void) ql_read_regs(ha, fw->rseq_1_reg, ha->iobase + 0xC0,
14169 		    sizeof (fw->rseq_1_reg) / 4, 32);
14170 
14171 		/* RSEQ-2 */
14172 		WRT32_IO_REG(ha, io_base_addr, 0xFFF0);
14173 		(void) ql_read_regs(ha, fw->rseq_2_reg, ha->iobase + 0xC0,
14174 		    sizeof (fw->rseq_2_reg) / 4, 32);
14175 
14176 		/* Auxiliary sequencer registers. */
14177 
14178 		/* ASEQ GP */
14179 		WRT32_IO_REG(ha, io_base_addr, 0xB000);
14180 		bp = ql_read_regs(ha, fw->aseq_gp_reg, ha->iobase + 0xC0,
14181 		    16, 32);
14182 		WRT32_IO_REG(ha, io_base_addr, 0xB010);
14183 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14184 		WRT32_IO_REG(ha, io_base_addr, 0xB020);
14185 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14186 		WRT32_IO_REG(ha, io_base_addr, 0xB030);
14187 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14188 		WRT32_IO_REG(ha, io_base_addr, 0xB040);
14189 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14190 		WRT32_IO_REG(ha, io_base_addr, 0xB050);
14191 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14192 		WRT32_IO_REG(ha, io_base_addr, 0xB060);
14193 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14194 		WRT32_IO_REG(ha, io_base_addr, 0xB070);
14195 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14196 
14197 		/* ASEQ-0 */
14198 		WRT32_IO_REG(ha, io_base_addr, 0xB0C0);
14199 		bp = ql_read_regs(ha, fw->aseq_0_reg, ha->iobase + 0xC0,
14200 		    16, 32);
14201 		WRT32_IO_REG(ha, io_base_addr, 0xB0D0);
14202 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14203 
14204 		/* ASEQ-1 */
14205 		WRT32_IO_REG(ha, io_base_addr, 0xB0E0);
14206 		(void) ql_read_regs(ha, fw->aseq_1_reg, ha->iobase + 0xC0,
14207 		    16, 32);
14208 
14209 		/* ASEQ-2 */
14210 		WRT32_IO_REG(ha, io_base_addr, 0xB0F0);
14211 		(void) ql_read_regs(ha, fw->aseq_2_reg, ha->iobase + 0xC0,
14212 		    16, 32);
14213 
14214 		/* Command DMA registers. */
14215 
14216 		WRT32_IO_REG(ha, io_base_addr, 0x7100);
14217 		(void) ql_read_regs(ha, fw->cmd_dma_reg, ha->iobase + 0xC0,
14218 		    sizeof (fw->cmd_dma_reg) / 4, 32);
14219 
14220 		/* Queues. */
14221 
14222 		/* RequestQ0 */
14223 		WRT32_IO_REG(ha, io_base_addr, 0x7200);
14224 		bp = ql_read_regs(ha, fw->req0_dma_reg, ha->iobase + 0xC0,
14225 		    8, 32);
14226 		(void) ql_read_regs(ha, bp, ha->iobase + 0xE4, 7, 32);
14227 
14228 		/* ResponseQ0 */
14229 		WRT32_IO_REG(ha, io_base_addr, 0x7300);
14230 		bp = ql_read_regs(ha, fw->resp0_dma_reg, ha->iobase + 0xC0,
14231 		    8, 32);
14232 		(void) ql_read_regs(ha, bp, ha->iobase + 0xE4, 7, 32);
14233 
14234 		/* RequestQ1 */
14235 		WRT32_IO_REG(ha, io_base_addr, 0x7400);
14236 		bp = ql_read_regs(ha, fw->req1_dma_reg, ha->iobase + 0xC0,
14237 		    8, 32);
14238 		(void) ql_read_regs(ha, bp, ha->iobase + 0xE4, 7, 32);
14239 
14240 		/* Transmit DMA registers. */
14241 
14242 		/* XMT0 */
14243 		WRT32_IO_REG(ha, io_base_addr, 0x7600);
14244 		bp = ql_read_regs(ha, fw->xmt0_dma_reg, ha->iobase + 0xC0,
14245 		    16, 32);
14246 		WRT32_IO_REG(ha, io_base_addr, 0x7610);
14247 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14248 
14249 		/* XMT1 */
14250 		WRT32_IO_REG(ha, io_base_addr, 0x7620);
14251 		bp = ql_read_regs(ha, fw->xmt1_dma_reg, ha->iobase + 0xC0,
14252 		    16, 32);
14253 		WRT32_IO_REG(ha, io_base_addr, 0x7630);
14254 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14255 
14256 		/* XMT2 */
14257 		WRT32_IO_REG(ha, io_base_addr, 0x7640);
14258 		bp = ql_read_regs(ha, fw->xmt2_dma_reg, ha->iobase + 0xC0,
14259 		    16, 32);
14260 		WRT32_IO_REG(ha, io_base_addr, 0x7650);
14261 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14262 
14263 		/* XMT3 */
14264 		WRT32_IO_REG(ha, io_base_addr, 0x7660);
14265 		bp = ql_read_regs(ha, fw->xmt3_dma_reg, ha->iobase + 0xC0,
14266 		    16, 32);
14267 		WRT32_IO_REG(ha, io_base_addr, 0x7670);
14268 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14269 
14270 		/* XMT4 */
14271 		WRT32_IO_REG(ha, io_base_addr, 0x7680);
14272 		bp = ql_read_regs(ha, fw->xmt4_dma_reg, ha->iobase + 0xC0,
14273 		    16, 32);
14274 		WRT32_IO_REG(ha, io_base_addr, 0x7690);
14275 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14276 
14277 		/* XMT Common */
14278 		WRT32_IO_REG(ha, io_base_addr, 0x76A0);
14279 		(void) ql_read_regs(ha, fw->xmt_data_dma_reg,
14280 		    ha->iobase + 0xC0, sizeof (fw->xmt_data_dma_reg) / 4, 32);
14281 
14282 		/* Receive DMA registers. */
14283 
14284 		/* RCVThread0 */
14285 		WRT32_IO_REG(ha, io_base_addr, 0x7700);
14286 		bp = ql_read_regs(ha, fw->rcvt0_data_dma_reg,
14287 		    ha->iobase + 0xC0, 16, 32);
14288 		WRT32_IO_REG(ha, io_base_addr, 0x7710);
14289 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14290 
14291 		/* RCVThread1 */
14292 		WRT32_IO_REG(ha, io_base_addr, 0x7720);
14293 		bp = ql_read_regs(ha, fw->rcvt1_data_dma_reg,
14294 		    ha->iobase + 0xC0, 16, 32);
14295 		WRT32_IO_REG(ha, io_base_addr, 0x7730);
14296 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14297 
14298 		/* RISC registers. */
14299 
14300 		/* RISC GP */
14301 		WRT32_IO_REG(ha, io_base_addr, 0x0F00);
14302 		bp = ql_read_regs(ha, fw->risc_gp_reg, ha->iobase + 0xC0,
14303 		    16, 32);
14304 		WRT32_IO_REG(ha, io_base_addr, 0x0F10);
14305 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14306 		WRT32_IO_REG(ha, io_base_addr, 0x0F20);
14307 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14308 		WRT32_IO_REG(ha, io_base_addr, 0x0F30);
14309 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14310 		WRT32_IO_REG(ha, io_base_addr, 0x0F40);
14311 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14312 		WRT32_IO_REG(ha, io_base_addr, 0x0F50);
14313 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14314 		WRT32_IO_REG(ha, io_base_addr, 0x0F60);
14315 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14316 		WRT32_IO_REG(ha, io_base_addr, 0x0F70);
14317 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14318 
14319 		/* Local memory controller (LMC) registers. */
14320 
14321 		/* LMC */
14322 		WRT32_IO_REG(ha, io_base_addr, 0x3000);
14323 		bp = ql_read_regs(ha, fw->lmc_reg, ha->iobase + 0xC0,
14324 		    16, 32);
14325 		WRT32_IO_REG(ha, io_base_addr, 0x3010);
14326 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14327 		WRT32_IO_REG(ha, io_base_addr, 0x3020);
14328 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14329 		WRT32_IO_REG(ha, io_base_addr, 0x3030);
14330 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14331 		WRT32_IO_REG(ha, io_base_addr, 0x3040);
14332 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14333 		WRT32_IO_REG(ha, io_base_addr, 0x3050);
14334 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14335 		WRT32_IO_REG(ha, io_base_addr, 0x3060);
14336 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14337 		WRT32_IO_REG(ha, io_base_addr, 0x3070);
14338 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14339 
14340 		/* Fibre Protocol Module registers. */
14341 
14342 		/* FPM hardware */
14343 		WRT32_IO_REG(ha, io_base_addr, 0x4000);
14344 		bp = ql_read_regs(ha, fw->fpm_hdw_reg, ha->iobase + 0xC0,
14345 		    16, 32);
14346 		WRT32_IO_REG(ha, io_base_addr, 0x4010);
14347 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14348 		WRT32_IO_REG(ha, io_base_addr, 0x4020);
14349 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14350 		WRT32_IO_REG(ha, io_base_addr, 0x4030);
14351 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14352 		WRT32_IO_REG(ha, io_base_addr, 0x4040);
14353 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14354 		WRT32_IO_REG(ha, io_base_addr, 0x4050);
14355 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14356 		WRT32_IO_REG(ha, io_base_addr, 0x4060);
14357 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14358 		WRT32_IO_REG(ha, io_base_addr, 0x4070);
14359 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14360 		WRT32_IO_REG(ha, io_base_addr, 0x4080);
14361 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14362 		WRT32_IO_REG(ha, io_base_addr, 0x4090);
14363 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14364 		WRT32_IO_REG(ha, io_base_addr, 0x40A0);
14365 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14366 		WRT32_IO_REG(ha, io_base_addr, 0x40B0);
14367 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14368 
14369 		/* Frame Buffer registers. */
14370 
14371 		/* FB hardware */
14372 		WRT32_IO_REG(ha, io_base_addr, 0x6000);
14373 		bp = ql_read_regs(ha, fw->fb_hdw_reg, ha->iobase + 0xC0,
14374 		    16, 32);
14375 		WRT32_IO_REG(ha, io_base_addr, 0x6010);
14376 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14377 		WRT32_IO_REG(ha, io_base_addr, 0x6020);
14378 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14379 		WRT32_IO_REG(ha, io_base_addr, 0x6030);
14380 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14381 		WRT32_IO_REG(ha, io_base_addr, 0x6040);
14382 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14383 		WRT32_IO_REG(ha, io_base_addr, 0x6100);
14384 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14385 		WRT32_IO_REG(ha, io_base_addr, 0x6130);
14386 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14387 		WRT32_IO_REG(ha, io_base_addr, 0x6150);
14388 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14389 		WRT32_IO_REG(ha, io_base_addr, 0x6170);
14390 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14391 		WRT32_IO_REG(ha, io_base_addr, 0x6190);
14392 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14393 		WRT32_IO_REG(ha, io_base_addr, 0x61B0);
14394 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14395 		WRT32_IO_REG(ha, io_base_addr, 0x6F00);
14396 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14397 	}
14398 
14399 	/* Get the request queue */
14400 	if (rval == QL_SUCCESS) {
14401 		uint32_t	cnt;
14402 		uint32_t	*w32 = (uint32_t *)ha->request_ring_bp;
14403 
14404 		/* Sync DMA buffer. */
14405 		(void) ddi_dma_sync(ha->hba_buf.dma_handle,
14406 		    REQUEST_Q_BUFFER_OFFSET, sizeof (fw->req_q),
14407 		    DDI_DMA_SYNC_FORKERNEL);
14408 
14409 		for (cnt = 0; cnt < sizeof (fw->req_q) / 4; cnt++) {
14410 			fw->req_q[cnt] = *w32++;
14411 			LITTLE_ENDIAN_32(&fw->req_q[cnt]);
14412 		}
14413 	}
14414 
14415 	/* Get the respons queue */
14416 	if (rval == QL_SUCCESS) {
14417 		uint32_t	cnt;
14418 		uint32_t	*w32 = (uint32_t *)ha->response_ring_bp;
14419 
14420 		/* Sync DMA buffer. */
14421 		(void) ddi_dma_sync(ha->hba_buf.dma_handle,
14422 		    RESPONSE_Q_BUFFER_OFFSET, sizeof (fw->rsp_q),
14423 		    DDI_DMA_SYNC_FORKERNEL);
14424 
14425 		for (cnt = 0; cnt < sizeof (fw->rsp_q) / 4; cnt++) {
14426 			fw->rsp_q[cnt] = *w32++;
14427 			LITTLE_ENDIAN_32(&fw->rsp_q[cnt]);
14428 		}
14429 	}
14430 
14431 	/* Reset RISC. */
14432 
14433 	ql_reset_chip(ha);
14434 
14435 	/* Memory. */
14436 
14437 	if (rval == QL_SUCCESS) {
14438 		/* Code RAM. */
14439 		rval = ql_read_risc_ram(ha, 0x20000,
14440 		    sizeof (fw->code_ram) / 4, fw->code_ram);
14441 	}
14442 	if (rval == QL_SUCCESS) {
14443 		/* External Memory. */
14444 		rval = ql_read_risc_ram(ha, 0x100000,
14445 		    ha->fw_ext_memory_size / 4, fw->ext_mem);
14446 	}
14447 
14448 	/* Get the FC event trace buffer */
14449 	if (rval == QL_SUCCESS) {
14450 		if (CFG_IST(ha, CFG_ENABLE_FWFCETRACE) &&
14451 		    (ha->fwfcetracebuf.bp != NULL)) {
14452 			uint32_t	cnt;
14453 			uint32_t	*w32 = ha->fwfcetracebuf.bp;
14454 
14455 			/* Sync DMA buffer. */
14456 			(void) ddi_dma_sync(ha->fwfcetracebuf.dma_handle, 0,
14457 			    FWFCESIZE, DDI_DMA_SYNC_FORKERNEL);
14458 
14459 			for (cnt = 0; cnt < FWFCESIZE / 4; cnt++) {
14460 				fw->fce_trace_buf[cnt] = *w32++;
14461 			}
14462 		}
14463 	}
14464 
14465 	/* Get the extended trace buffer */
14466 	if (rval == QL_SUCCESS) {
14467 		if (CFG_IST(ha, CFG_ENABLE_FWEXTTRACE) &&
14468 		    (ha->fwexttracebuf.bp != NULL)) {
14469 			uint32_t	cnt;
14470 			uint32_t	*w32 = ha->fwexttracebuf.bp;
14471 
14472 			/* Sync DMA buffer. */
14473 			(void) ddi_dma_sync(ha->fwexttracebuf.dma_handle, 0,
14474 			    FWEXTSIZE, DDI_DMA_SYNC_FORKERNEL);
14475 
14476 			for (cnt = 0; cnt < FWEXTSIZE / 4; cnt++) {
14477 				fw->ext_trace_buf[cnt] = *w32++;
14478 			}
14479 		}
14480 	}
14481 
14482 	if (rval != QL_SUCCESS) {
14483 		EL(ha, "failed=%xh\n", rval);
14484 	} else {
14485 		/*EMPTY*/
14486 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
14487 	}
14488 
14489 	return (rval);
14490 }
14491 
14492 /*
14493  * ql_81xx_binary_fw_dump
14494  *
14495  * Input:
14496  *	ha:	adapter state pointer.
14497  *	fw:	firmware dump context pointer.
14498  *
14499  * Returns:
14500  *	ql local function return status code.
14501  *
14502  * Context:
14503  *	Interrupt or Kernel context, no mailbox commands allowed.
14504  */
14505 static int
ql_81xx_binary_fw_dump(ql_adapter_state_t * ha,ql_81xx_fw_dump_t * fw)14506 ql_81xx_binary_fw_dump(ql_adapter_state_t *ha, ql_81xx_fw_dump_t *fw)
14507 {
14508 	uint32_t	*reg32;
14509 	void		*bp;
14510 	clock_t		timer;
14511 	int		rval = QL_SUCCESS;
14512 
14513 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
14514 
14515 	fw->r2h_status = RD32_IO_REG(ha, risc2host);
14516 
14517 	/* Pause RISC. */
14518 	if ((RD32_IO_REG(ha, risc2host) & RH_RISC_PAUSED) == 0) {
14519 		/* Disable ISP interrupts. */
14520 		WRT16_IO_REG(ha, ictrl, 0);
14521 
14522 		WRT32_IO_REG(ha, hccr, HC24_PAUSE_RISC);
14523 		for (timer = 30000;
14524 		    (RD32_IO_REG(ha, risc2host) & RH_RISC_PAUSED) == 0 &&
14525 		    rval == QL_SUCCESS; timer--) {
14526 			if (timer) {
14527 				drv_usecwait(100);
14528 				if (timer % 10000 == 0) {
14529 					EL(ha, "risc pause %d\n", timer);
14530 				}
14531 			} else {
14532 				EL(ha, "risc pause timeout\n");
14533 				rval = QL_FUNCTION_TIMEOUT;
14534 			}
14535 		}
14536 	}
14537 
14538 	if (rval == QL_SUCCESS) {
14539 
14540 		/* Host Interface registers */
14541 
14542 		/* HostRisc registers. */
14543 		WRT32_IO_REG(ha, io_base_addr, 0x7000);
14544 		bp = ql_read_regs(ha, fw->hostrisc_reg, ha->iobase + 0xC0,
14545 		    16, 32);
14546 		WRT32_IO_REG(ha, io_base_addr, 0x7010);
14547 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14548 
14549 		/* PCIe registers. */
14550 		WRT32_IO_REG(ha, io_base_addr, 0x7c00);
14551 		WRT_REG_DWORD(ha, ha->iobase + 0xc0, 0x1);
14552 		bp = ql_read_regs(ha, fw->pcie_reg, ha->iobase + 0xC4,
14553 		    3, 32);
14554 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 1, 32);
14555 		WRT_REG_DWORD(ha, ha->iobase + 0xc0, 0x0);
14556 
14557 		/* Host interface registers. */
14558 		(void) ql_read_regs(ha, fw->host_reg, ha->iobase,
14559 		    sizeof (fw->host_reg) / 4, 32);
14560 
14561 		/* Disable ISP interrupts. */
14562 
14563 		WRT32_IO_REG(ha, ictrl, 0);
14564 		RD32_IO_REG(ha, ictrl);
14565 		ADAPTER_STATE_LOCK(ha);
14566 		ha->flags &= ~INTERRUPTS_ENABLED;
14567 		ADAPTER_STATE_UNLOCK(ha);
14568 
14569 		/* Shadow registers. */
14570 
14571 		WRT32_IO_REG(ha, io_base_addr, 0x0F70);
14572 		RD32_IO_REG(ha, io_base_addr);
14573 
14574 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
14575 		WRT_REG_DWORD(ha, reg32, 0xB0000000);
14576 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
14577 		fw->shadow_reg[0] = RD_REG_DWORD(ha, reg32);
14578 
14579 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
14580 		WRT_REG_DWORD(ha, reg32, 0xB0100000);
14581 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
14582 		fw->shadow_reg[1] = RD_REG_DWORD(ha, reg32);
14583 
14584 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
14585 		WRT_REG_DWORD(ha, reg32, 0xB0200000);
14586 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
14587 		fw->shadow_reg[2] = RD_REG_DWORD(ha, reg32);
14588 
14589 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
14590 		WRT_REG_DWORD(ha, reg32, 0xB0300000);
14591 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
14592 		fw->shadow_reg[3] = RD_REG_DWORD(ha, reg32);
14593 
14594 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
14595 		WRT_REG_DWORD(ha, reg32, 0xB0400000);
14596 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
14597 		fw->shadow_reg[4] = RD_REG_DWORD(ha, reg32);
14598 
14599 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
14600 		WRT_REG_DWORD(ha, reg32, 0xB0500000);
14601 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
14602 		fw->shadow_reg[5] = RD_REG_DWORD(ha, reg32);
14603 
14604 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
14605 		WRT_REG_DWORD(ha, reg32, 0xB0600000);
14606 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
14607 		fw->shadow_reg[6] = RD_REG_DWORD(ha, reg32);
14608 
14609 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
14610 		WRT_REG_DWORD(ha, reg32, 0xB0700000);
14611 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
14612 		fw->shadow_reg[7] = RD_REG_DWORD(ha, reg32);
14613 
14614 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
14615 		WRT_REG_DWORD(ha, reg32, 0xB0800000);
14616 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
14617 		fw->shadow_reg[8] = RD_REG_DWORD(ha, reg32);
14618 
14619 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
14620 		WRT_REG_DWORD(ha, reg32, 0xB0900000);
14621 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
14622 		fw->shadow_reg[9] = RD_REG_DWORD(ha, reg32);
14623 
14624 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xF0);
14625 		WRT_REG_DWORD(ha, reg32, 0xB0A00000);
14626 		reg32 = (uint32_t *)((caddr_t)ha->iobase + 0xFC);
14627 		fw->shadow_reg[0xa] = RD_REG_DWORD(ha, reg32);
14628 
14629 		/* RISC I/O register. */
14630 
14631 		WRT32_IO_REG(ha, io_base_addr, 0x0010);
14632 		(void) ql_read_regs(ha, &fw->risc_io, ha->iobase + 0xC0,
14633 		    1, 32);
14634 
14635 		/* Mailbox registers. */
14636 
14637 		(void) ql_read_regs(ha, fw->mailbox_reg, ha->iobase + 0x80,
14638 		    sizeof (fw->mailbox_reg) / 2, 16);
14639 
14640 		/* Transfer sequence registers. */
14641 
14642 		/* XSEQ GP */
14643 		WRT32_IO_REG(ha, io_base_addr, 0xBF00);
14644 		bp = ql_read_regs(ha, fw->xseq_gp_reg, ha->iobase + 0xC0,
14645 		    16, 32);
14646 		WRT32_IO_REG(ha, io_base_addr, 0xBF10);
14647 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14648 		WRT32_IO_REG(ha, io_base_addr, 0xBF20);
14649 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14650 		WRT32_IO_REG(ha, io_base_addr, 0xBF30);
14651 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14652 		WRT32_IO_REG(ha, io_base_addr, 0xBF40);
14653 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14654 		WRT32_IO_REG(ha, io_base_addr, 0xBF50);
14655 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14656 		WRT32_IO_REG(ha, io_base_addr, 0xBF60);
14657 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14658 		WRT32_IO_REG(ha, io_base_addr, 0xBF70);
14659 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14660 
14661 		/* XSEQ-0 */
14662 		WRT32_IO_REG(ha, io_base_addr, 0xBFC0);
14663 		bp = ql_read_regs(ha, fw->xseq_0_reg, ha->iobase + 0xC0,
14664 		    16, 32);
14665 		WRT32_IO_REG(ha, io_base_addr, 0xBFD0);
14666 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14667 		WRT32_IO_REG(ha, io_base_addr, 0xBFE0);
14668 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14669 
14670 		/* XSEQ-1 */
14671 		WRT32_IO_REG(ha, io_base_addr, 0xBFF0);
14672 		(void) ql_read_regs(ha, fw->xseq_1_reg, ha->iobase + 0xC0,
14673 		    16, 32);
14674 
14675 		/* Receive sequence registers. */
14676 
14677 		/* RSEQ GP */
14678 		WRT32_IO_REG(ha, io_base_addr, 0xFF00);
14679 		bp = ql_read_regs(ha, fw->rseq_gp_reg, ha->iobase + 0xC0,
14680 		    16, 32);
14681 		WRT32_IO_REG(ha, io_base_addr, 0xFF10);
14682 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14683 		WRT32_IO_REG(ha, io_base_addr, 0xFF20);
14684 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14685 		WRT32_IO_REG(ha, io_base_addr, 0xFF30);
14686 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14687 		WRT32_IO_REG(ha, io_base_addr, 0xFF40);
14688 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14689 		WRT32_IO_REG(ha, io_base_addr, 0xFF50);
14690 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14691 		WRT32_IO_REG(ha, io_base_addr, 0xFF60);
14692 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14693 		WRT32_IO_REG(ha, io_base_addr, 0xFF70);
14694 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14695 
14696 		/* RSEQ-0 */
14697 		WRT32_IO_REG(ha, io_base_addr, 0xFFC0);
14698 		bp = ql_read_regs(ha, fw->rseq_0_reg, ha->iobase + 0xC0,
14699 		    16, 32);
14700 		WRT32_IO_REG(ha, io_base_addr, 0xFFD0);
14701 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14702 
14703 		/* RSEQ-1 */
14704 		WRT32_IO_REG(ha, io_base_addr, 0xFFE0);
14705 		(void) ql_read_regs(ha, fw->rseq_1_reg, ha->iobase + 0xC0,
14706 		    sizeof (fw->rseq_1_reg) / 4, 32);
14707 
14708 		/* RSEQ-2 */
14709 		WRT32_IO_REG(ha, io_base_addr, 0xFFF0);
14710 		(void) ql_read_regs(ha, fw->rseq_2_reg, ha->iobase + 0xC0,
14711 		    sizeof (fw->rseq_2_reg) / 4, 32);
14712 
14713 		/* Auxiliary sequencer registers. */
14714 
14715 		/* ASEQ GP */
14716 		WRT32_IO_REG(ha, io_base_addr, 0xB000);
14717 		bp = ql_read_regs(ha, fw->aseq_gp_reg, ha->iobase + 0xC0,
14718 		    16, 32);
14719 		WRT32_IO_REG(ha, io_base_addr, 0xB010);
14720 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14721 		WRT32_IO_REG(ha, io_base_addr, 0xB020);
14722 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14723 		WRT32_IO_REG(ha, io_base_addr, 0xB030);
14724 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14725 		WRT32_IO_REG(ha, io_base_addr, 0xB040);
14726 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14727 		WRT32_IO_REG(ha, io_base_addr, 0xB050);
14728 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14729 		WRT32_IO_REG(ha, io_base_addr, 0xB060);
14730 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14731 		WRT32_IO_REG(ha, io_base_addr, 0xB070);
14732 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14733 
14734 		/* ASEQ-0 */
14735 		WRT32_IO_REG(ha, io_base_addr, 0xB0C0);
14736 		bp = ql_read_regs(ha, fw->aseq_0_reg, ha->iobase + 0xC0,
14737 		    16, 32);
14738 		WRT32_IO_REG(ha, io_base_addr, 0xB0D0);
14739 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14740 
14741 		/* ASEQ-1 */
14742 		WRT32_IO_REG(ha, io_base_addr, 0xB0E0);
14743 		(void) ql_read_regs(ha, fw->aseq_1_reg, ha->iobase + 0xC0,
14744 		    16, 32);
14745 
14746 		/* ASEQ-2 */
14747 		WRT32_IO_REG(ha, io_base_addr, 0xB0F0);
14748 		(void) ql_read_regs(ha, fw->aseq_2_reg, ha->iobase + 0xC0,
14749 		    16, 32);
14750 
14751 		/* Command DMA registers. */
14752 
14753 		WRT32_IO_REG(ha, io_base_addr, 0x7100);
14754 		(void) ql_read_regs(ha, fw->cmd_dma_reg, ha->iobase + 0xC0,
14755 		    sizeof (fw->cmd_dma_reg) / 4, 32);
14756 
14757 		/* Queues. */
14758 
14759 		/* RequestQ0 */
14760 		WRT32_IO_REG(ha, io_base_addr, 0x7200);
14761 		bp = ql_read_regs(ha, fw->req0_dma_reg, ha->iobase + 0xC0,
14762 		    8, 32);
14763 		(void) ql_read_regs(ha, bp, ha->iobase + 0xE4, 7, 32);
14764 
14765 		/* ResponseQ0 */
14766 		WRT32_IO_REG(ha, io_base_addr, 0x7300);
14767 		bp = ql_read_regs(ha, fw->resp0_dma_reg, ha->iobase + 0xC0,
14768 		    8, 32);
14769 		(void) ql_read_regs(ha, bp, ha->iobase + 0xE4, 7, 32);
14770 
14771 		/* RequestQ1 */
14772 		WRT32_IO_REG(ha, io_base_addr, 0x7400);
14773 		bp = ql_read_regs(ha, fw->req1_dma_reg, ha->iobase + 0xC0,
14774 		    8, 32);
14775 		(void) ql_read_regs(ha, bp, ha->iobase + 0xE4, 7, 32);
14776 
14777 		/* Transmit DMA registers. */
14778 
14779 		/* XMT0 */
14780 		WRT32_IO_REG(ha, io_base_addr, 0x7600);
14781 		bp = ql_read_regs(ha, fw->xmt0_dma_reg, ha->iobase + 0xC0,
14782 		    16, 32);
14783 		WRT32_IO_REG(ha, io_base_addr, 0x7610);
14784 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14785 
14786 		/* XMT1 */
14787 		WRT32_IO_REG(ha, io_base_addr, 0x7620);
14788 		bp = ql_read_regs(ha, fw->xmt1_dma_reg, ha->iobase + 0xC0,
14789 		    16, 32);
14790 		WRT32_IO_REG(ha, io_base_addr, 0x7630);
14791 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14792 
14793 		/* XMT2 */
14794 		WRT32_IO_REG(ha, io_base_addr, 0x7640);
14795 		bp = ql_read_regs(ha, fw->xmt2_dma_reg, ha->iobase + 0xC0,
14796 		    16, 32);
14797 		WRT32_IO_REG(ha, io_base_addr, 0x7650);
14798 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14799 
14800 		/* XMT3 */
14801 		WRT32_IO_REG(ha, io_base_addr, 0x7660);
14802 		bp = ql_read_regs(ha, fw->xmt3_dma_reg, ha->iobase + 0xC0,
14803 		    16, 32);
14804 		WRT32_IO_REG(ha, io_base_addr, 0x7670);
14805 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14806 
14807 		/* XMT4 */
14808 		WRT32_IO_REG(ha, io_base_addr, 0x7680);
14809 		bp = ql_read_regs(ha, fw->xmt4_dma_reg, ha->iobase + 0xC0,
14810 		    16, 32);
14811 		WRT32_IO_REG(ha, io_base_addr, 0x7690);
14812 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14813 
14814 		/* XMT Common */
14815 		WRT32_IO_REG(ha, io_base_addr, 0x76A0);
14816 		(void) ql_read_regs(ha, fw->xmt_data_dma_reg,
14817 		    ha->iobase + 0xC0, sizeof (fw->xmt_data_dma_reg) / 4, 32);
14818 
14819 		/* Receive DMA registers. */
14820 
14821 		/* RCVThread0 */
14822 		WRT32_IO_REG(ha, io_base_addr, 0x7700);
14823 		bp = ql_read_regs(ha, fw->rcvt0_data_dma_reg,
14824 		    ha->iobase + 0xC0, 16, 32);
14825 		WRT32_IO_REG(ha, io_base_addr, 0x7710);
14826 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14827 
14828 		/* RCVThread1 */
14829 		WRT32_IO_REG(ha, io_base_addr, 0x7720);
14830 		bp = ql_read_regs(ha, fw->rcvt1_data_dma_reg,
14831 		    ha->iobase + 0xC0, 16, 32);
14832 		WRT32_IO_REG(ha, io_base_addr, 0x7730);
14833 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14834 
14835 		/* RISC registers. */
14836 
14837 		/* RISC GP */
14838 		WRT32_IO_REG(ha, io_base_addr, 0x0F00);
14839 		bp = ql_read_regs(ha, fw->risc_gp_reg, ha->iobase + 0xC0,
14840 		    16, 32);
14841 		WRT32_IO_REG(ha, io_base_addr, 0x0F10);
14842 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14843 		WRT32_IO_REG(ha, io_base_addr, 0x0F20);
14844 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14845 		WRT32_IO_REG(ha, io_base_addr, 0x0F30);
14846 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14847 		WRT32_IO_REG(ha, io_base_addr, 0x0F40);
14848 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14849 		WRT32_IO_REG(ha, io_base_addr, 0x0F50);
14850 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14851 		WRT32_IO_REG(ha, io_base_addr, 0x0F60);
14852 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14853 		WRT32_IO_REG(ha, io_base_addr, 0x0F70);
14854 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14855 
14856 		/* Local memory controller (LMC) registers. */
14857 
14858 		/* LMC */
14859 		WRT32_IO_REG(ha, io_base_addr, 0x3000);
14860 		bp = ql_read_regs(ha, fw->lmc_reg, ha->iobase + 0xC0,
14861 		    16, 32);
14862 		WRT32_IO_REG(ha, io_base_addr, 0x3010);
14863 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14864 		WRT32_IO_REG(ha, io_base_addr, 0x3020);
14865 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14866 		WRT32_IO_REG(ha, io_base_addr, 0x3030);
14867 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14868 		WRT32_IO_REG(ha, io_base_addr, 0x3040);
14869 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14870 		WRT32_IO_REG(ha, io_base_addr, 0x3050);
14871 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14872 		WRT32_IO_REG(ha, io_base_addr, 0x3060);
14873 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14874 		WRT32_IO_REG(ha, io_base_addr, 0x3070);
14875 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14876 
14877 		/* Fibre Protocol Module registers. */
14878 
14879 		/* FPM hardware */
14880 		WRT32_IO_REG(ha, io_base_addr, 0x4000);
14881 		bp = ql_read_regs(ha, fw->fpm_hdw_reg, ha->iobase + 0xC0,
14882 		    16, 32);
14883 		WRT32_IO_REG(ha, io_base_addr, 0x4010);
14884 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14885 		WRT32_IO_REG(ha, io_base_addr, 0x4020);
14886 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14887 		WRT32_IO_REG(ha, io_base_addr, 0x4030);
14888 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14889 		WRT32_IO_REG(ha, io_base_addr, 0x4040);
14890 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14891 		WRT32_IO_REG(ha, io_base_addr, 0x4050);
14892 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14893 		WRT32_IO_REG(ha, io_base_addr, 0x4060);
14894 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14895 		WRT32_IO_REG(ha, io_base_addr, 0x4070);
14896 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14897 		WRT32_IO_REG(ha, io_base_addr, 0x4080);
14898 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14899 		WRT32_IO_REG(ha, io_base_addr, 0x4090);
14900 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14901 		WRT32_IO_REG(ha, io_base_addr, 0x40A0);
14902 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14903 		WRT32_IO_REG(ha, io_base_addr, 0x40B0);
14904 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14905 		WRT32_IO_REG(ha, io_base_addr, 0x40C0);
14906 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14907 		WRT32_IO_REG(ha, io_base_addr, 0x40D0);
14908 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14909 
14910 		/* Frame Buffer registers. */
14911 
14912 		/* FB hardware */
14913 		WRT32_IO_REG(ha, io_base_addr, 0x6000);
14914 		bp = ql_read_regs(ha, fw->fb_hdw_reg, ha->iobase + 0xC0,
14915 		    16, 32);
14916 		WRT32_IO_REG(ha, io_base_addr, 0x6010);
14917 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14918 		WRT32_IO_REG(ha, io_base_addr, 0x6020);
14919 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14920 		WRT32_IO_REG(ha, io_base_addr, 0x6030);
14921 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14922 		WRT32_IO_REG(ha, io_base_addr, 0x6040);
14923 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14924 		WRT32_IO_REG(ha, io_base_addr, 0x6100);
14925 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14926 		WRT32_IO_REG(ha, io_base_addr, 0x6130);
14927 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14928 		WRT32_IO_REG(ha, io_base_addr, 0x6150);
14929 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14930 		WRT32_IO_REG(ha, io_base_addr, 0x6170);
14931 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14932 		WRT32_IO_REG(ha, io_base_addr, 0x6190);
14933 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14934 		WRT32_IO_REG(ha, io_base_addr, 0x61B0);
14935 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14936 		WRT32_IO_REG(ha, io_base_addr, 0x61C0);
14937 		bp = ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14938 		WRT32_IO_REG(ha, io_base_addr, 0x6F00);
14939 		(void) ql_read_regs(ha, bp, ha->iobase + 0xC0, 16, 32);
14940 	}
14941 
14942 	/* Get the request queue */
14943 	if (rval == QL_SUCCESS) {
14944 		uint32_t	cnt;
14945 		uint32_t	*w32 = (uint32_t *)ha->request_ring_bp;
14946 
14947 		/* Sync DMA buffer. */
14948 		(void) ddi_dma_sync(ha->hba_buf.dma_handle,
14949 		    REQUEST_Q_BUFFER_OFFSET, sizeof (fw->req_q),
14950 		    DDI_DMA_SYNC_FORKERNEL);
14951 
14952 		for (cnt = 0; cnt < sizeof (fw->req_q) / 4; cnt++) {
14953 			fw->req_q[cnt] = *w32++;
14954 			LITTLE_ENDIAN_32(&fw->req_q[cnt]);
14955 		}
14956 	}
14957 
14958 	/* Get the response queue */
14959 	if (rval == QL_SUCCESS) {
14960 		uint32_t	cnt;
14961 		uint32_t	*w32 = (uint32_t *)ha->response_ring_bp;
14962 
14963 		/* Sync DMA buffer. */
14964 		(void) ddi_dma_sync(ha->hba_buf.dma_handle,
14965 		    RESPONSE_Q_BUFFER_OFFSET, sizeof (fw->rsp_q),
14966 		    DDI_DMA_SYNC_FORKERNEL);
14967 
14968 		for (cnt = 0; cnt < sizeof (fw->rsp_q) / 4; cnt++) {
14969 			fw->rsp_q[cnt] = *w32++;
14970 			LITTLE_ENDIAN_32(&fw->rsp_q[cnt]);
14971 		}
14972 	}
14973 
14974 	/* Reset RISC. */
14975 
14976 	ql_reset_chip(ha);
14977 
14978 	/* Memory. */
14979 
14980 	if (rval == QL_SUCCESS) {
14981 		/* Code RAM. */
14982 		rval = ql_read_risc_ram(ha, 0x20000,
14983 		    sizeof (fw->code_ram) / 4, fw->code_ram);
14984 	}
14985 	if (rval == QL_SUCCESS) {
14986 		/* External Memory. */
14987 		rval = ql_read_risc_ram(ha, 0x100000,
14988 		    ha->fw_ext_memory_size / 4, fw->ext_mem);
14989 	}
14990 
14991 	/* Get the FC event trace buffer */
14992 	if (rval == QL_SUCCESS) {
14993 		if (CFG_IST(ha, CFG_ENABLE_FWFCETRACE) &&
14994 		    (ha->fwfcetracebuf.bp != NULL)) {
14995 			uint32_t	cnt;
14996 			uint32_t	*w32 = ha->fwfcetracebuf.bp;
14997 
14998 			/* Sync DMA buffer. */
14999 			(void) ddi_dma_sync(ha->fwfcetracebuf.dma_handle, 0,
15000 			    FWFCESIZE, DDI_DMA_SYNC_FORKERNEL);
15001 
15002 			for (cnt = 0; cnt < FWFCESIZE / 4; cnt++) {
15003 				fw->fce_trace_buf[cnt] = *w32++;
15004 			}
15005 		}
15006 	}
15007 
15008 	/* Get the extended trace buffer */
15009 	if (rval == QL_SUCCESS) {
15010 		if (CFG_IST(ha, CFG_ENABLE_FWEXTTRACE) &&
15011 		    (ha->fwexttracebuf.bp != NULL)) {
15012 			uint32_t	cnt;
15013 			uint32_t	*w32 = ha->fwexttracebuf.bp;
15014 
15015 			/* Sync DMA buffer. */
15016 			(void) ddi_dma_sync(ha->fwexttracebuf.dma_handle, 0,
15017 			    FWEXTSIZE, DDI_DMA_SYNC_FORKERNEL);
15018 
15019 			for (cnt = 0; cnt < FWEXTSIZE / 4; cnt++) {
15020 				fw->ext_trace_buf[cnt] = *w32++;
15021 			}
15022 		}
15023 	}
15024 
15025 	if (rval != QL_SUCCESS) {
15026 		EL(ha, "failed=%xh\n", rval);
15027 	} else {
15028 		/*EMPTY*/
15029 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
15030 	}
15031 
15032 	return (rval);
15033 }
15034 
15035 /*
15036  * ql_read_risc_ram
15037  *	Reads RISC RAM one word at a time.
15038  *	Risc interrupts must be disabled when this routine is called.
15039  *
15040  * Input:
15041  *	ha:	adapter state pointer.
15042  *	risc_address:	RISC code start address.
15043  *	len:		Number of words.
15044  *	buf:		buffer pointer.
15045  *
15046  * Returns:
15047  *	ql local function return status code.
15048  *
15049  * Context:
15050  *	Interrupt or Kernel context, no mailbox commands allowed.
15051  */
15052 static int
ql_read_risc_ram(ql_adapter_state_t * ha,uint32_t risc_address,uint32_t len,void * buf)15053 ql_read_risc_ram(ql_adapter_state_t *ha, uint32_t risc_address, uint32_t len,
15054     void *buf)
15055 {
15056 	uint32_t	cnt;
15057 	uint16_t	stat;
15058 	clock_t		timer;
15059 	uint16_t	*buf16 = (uint16_t *)buf;
15060 	uint32_t	*buf32 = (uint32_t *)buf;
15061 	int		rval = QL_SUCCESS;
15062 
15063 	for (cnt = 0; cnt < len; cnt++, risc_address++) {
15064 		WRT16_IO_REG(ha, mailbox_in[0], MBC_READ_RAM_EXTENDED);
15065 		WRT16_IO_REG(ha, mailbox_in[1], LSW(risc_address));
15066 		WRT16_IO_REG(ha, mailbox_in[8], MSW(risc_address));
15067 		if (CFG_IST(ha, CFG_CTRL_8021)) {
15068 			WRT32_IO_REG(ha, nx_host_int, NX_MBX_CMD);
15069 		} else if (CFG_IST(ha, CFG_CTRL_242581)) {
15070 			WRT32_IO_REG(ha, hccr, HC24_SET_HOST_INT);
15071 		} else {
15072 			WRT16_IO_REG(ha, hccr, HC_SET_HOST_INT);
15073 		}
15074 		for (timer = 6000000; timer && rval == QL_SUCCESS; timer--) {
15075 			if (INTERRUPT_PENDING(ha)) {
15076 				stat = (uint16_t)
15077 				    (RD16_IO_REG(ha, risc2host) & 0xff);
15078 				if ((stat == 1) || (stat == 0x10)) {
15079 					if (CFG_IST(ha, CFG_CTRL_24258081)) {
15080 						buf32[cnt] = SHORT_TO_LONG(
15081 						    RD16_IO_REG(ha,
15082 						    mailbox_out[2]),
15083 						    RD16_IO_REG(ha,
15084 						    mailbox_out[3]));
15085 					} else {
15086 						buf16[cnt] =
15087 						    RD16_IO_REG(ha,
15088 						    mailbox_out[2]);
15089 					}
15090 
15091 					break;
15092 				} else if ((stat == 2) || (stat == 0x11)) {
15093 					rval = RD16_IO_REG(ha, mailbox_out[0]);
15094 					break;
15095 				}
15096 				if (CFG_IST(ha, CFG_CTRL_8021)) {
15097 					ql_8021_clr_hw_intr(ha);
15098 					ql_8021_clr_fw_intr(ha);
15099 				} else if (CFG_IST(ha, CFG_CTRL_242581)) {
15100 					WRT32_IO_REG(ha, hccr,
15101 					    HC24_CLR_RISC_INT);
15102 					RD32_IO_REG(ha, hccr);
15103 				} else {
15104 					WRT16_IO_REG(ha, hccr,
15105 					    HC_CLR_RISC_INT);
15106 				}
15107 			}
15108 			drv_usecwait(5);
15109 		}
15110 		if (CFG_IST(ha, CFG_CTRL_8021)) {
15111 			ql_8021_clr_hw_intr(ha);
15112 			ql_8021_clr_fw_intr(ha);
15113 		} else if (CFG_IST(ha, CFG_CTRL_242581)) {
15114 			WRT32_IO_REG(ha, hccr, HC24_CLR_RISC_INT);
15115 			RD32_IO_REG(ha, hccr);
15116 		} else {
15117 			WRT16_IO_REG(ha, hccr, HC_CLR_RISC_INT);
15118 			WRT16_IO_REG(ha, semaphore, 0);
15119 		}
15120 
15121 		if (timer == 0) {
15122 			rval = QL_FUNCTION_TIMEOUT;
15123 		}
15124 	}
15125 
15126 	return (rval);
15127 }
15128 
15129 /*
15130  * ql_read_regs
15131  *	Reads adapter registers to buffer.
15132  *
15133  * Input:
15134  *	ha:	adapter state pointer.
15135  *	buf:	buffer pointer.
15136  *	reg:	start address.
15137  *	count:	number of registers.
15138  *	wds:	register size.
15139  *
15140  * Context:
15141  *	Interrupt or Kernel context, no mailbox commands allowed.
15142  */
15143 static void *
ql_read_regs(ql_adapter_state_t * ha,void * buf,void * reg,uint32_t count,uint8_t wds)15144 ql_read_regs(ql_adapter_state_t *ha, void *buf, void *reg, uint32_t count,
15145     uint8_t wds)
15146 {
15147 	uint32_t	*bp32, *reg32;
15148 	uint16_t	*bp16, *reg16;
15149 	uint8_t		*bp8, *reg8;
15150 
15151 	switch (wds) {
15152 	case 32:
15153 		bp32 = buf;
15154 		reg32 = reg;
15155 		while (count--) {
15156 			*bp32++ = RD_REG_DWORD(ha, reg32++);
15157 		}
15158 		return (bp32);
15159 	case 16:
15160 		bp16 = buf;
15161 		reg16 = reg;
15162 		while (count--) {
15163 			*bp16++ = RD_REG_WORD(ha, reg16++);
15164 		}
15165 		return (bp16);
15166 	case 8:
15167 		bp8 = buf;
15168 		reg8 = reg;
15169 		while (count--) {
15170 			*bp8++ = RD_REG_BYTE(ha, reg8++);
15171 		}
15172 		return (bp8);
15173 	default:
15174 		EL(ha, "Unknown word size=%d\n", wds);
15175 		return (buf);
15176 	}
15177 }
15178 
15179 static int
ql_save_config_regs(dev_info_t * dip)15180 ql_save_config_regs(dev_info_t *dip)
15181 {
15182 	ql_adapter_state_t	*ha;
15183 	int			ret;
15184 	ql_config_space_t	chs;
15185 	caddr_t			prop = "ql-config-space";
15186 
15187 	ha = ddi_get_soft_state(ql_state, ddi_get_instance(dip));
15188 	if (ha == NULL) {
15189 		QL_PRINT_2(CE_CONT, "(%d): no adapter\n",
15190 		    ddi_get_instance(dip));
15191 		return (DDI_FAILURE);
15192 	}
15193 
15194 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
15195 
15196 	/*LINTED [Solaris DDI_DEV_T_ANY Lint warning]*/
15197 	if (ddi_prop_exists(DDI_DEV_T_ANY, dip, DDI_PROP_DONTPASS, prop) ==
15198 	    1) {
15199 		QL_PRINT_2(CE_CONT, "(%d): no prop exit\n", ha->instance);
15200 		return (DDI_SUCCESS);
15201 	}
15202 
15203 	chs.chs_command = (uint16_t)ql_pci_config_get16(ha, PCI_CONF_COMM);
15204 	chs.chs_header_type = (uint8_t)ql_pci_config_get8(ha,
15205 	    PCI_CONF_HEADER);
15206 	if ((chs.chs_header_type & PCI_HEADER_TYPE_M) == PCI_HEADER_ONE) {
15207 		chs.chs_bridge_control = (uint8_t)ql_pci_config_get8(ha,
15208 		    PCI_BCNF_BCNTRL);
15209 	}
15210 
15211 	chs.chs_cache_line_size = (uint8_t)ql_pci_config_get8(ha,
15212 	    PCI_CONF_CACHE_LINESZ);
15213 
15214 	chs.chs_latency_timer = (uint8_t)ql_pci_config_get8(ha,
15215 	    PCI_CONF_LATENCY_TIMER);
15216 
15217 	if ((chs.chs_header_type & PCI_HEADER_TYPE_M) == PCI_HEADER_ONE) {
15218 		chs.chs_sec_latency_timer = (uint8_t)ql_pci_config_get8(ha,
15219 		    PCI_BCNF_LATENCY_TIMER);
15220 	}
15221 
15222 	chs.chs_base0 = ql_pci_config_get32(ha, PCI_CONF_BASE0);
15223 	chs.chs_base1 = ql_pci_config_get32(ha, PCI_CONF_BASE1);
15224 	chs.chs_base2 = ql_pci_config_get32(ha, PCI_CONF_BASE2);
15225 	chs.chs_base3 = ql_pci_config_get32(ha, PCI_CONF_BASE3);
15226 	chs.chs_base4 = ql_pci_config_get32(ha, PCI_CONF_BASE4);
15227 	chs.chs_base5 = ql_pci_config_get32(ha, PCI_CONF_BASE5);
15228 
15229 	/*LINTED [Solaris DDI_DEV_T_NONE Lint warning]*/
15230 	ret = ndi_prop_update_byte_array(DDI_DEV_T_NONE, dip, prop,
15231 	    (uchar_t *)&chs, sizeof (ql_config_space_t));
15232 
15233 	if (ret != DDI_PROP_SUCCESS) {
15234 		cmn_err(CE_WARN, "!Qlogic %s(%d) can't update prop %s",
15235 		    QL_NAME, ddi_get_instance(dip), prop);
15236 		return (DDI_FAILURE);
15237 	}
15238 
15239 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
15240 
15241 	return (DDI_SUCCESS);
15242 }
15243 
15244 static int
ql_restore_config_regs(dev_info_t * dip)15245 ql_restore_config_regs(dev_info_t *dip)
15246 {
15247 	ql_adapter_state_t	*ha;
15248 	uint_t			elements;
15249 	ql_config_space_t	*chs_p;
15250 	caddr_t			prop = "ql-config-space";
15251 
15252 	ha = ddi_get_soft_state(ql_state, ddi_get_instance(dip));
15253 	if (ha == NULL) {
15254 		QL_PRINT_2(CE_CONT, "(%d): no adapter\n",
15255 		    ddi_get_instance(dip));
15256 		return (DDI_FAILURE);
15257 	}
15258 
15259 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
15260 
15261 	/*LINTED [Solaris DDI_DEV_T_ANY Lint warning]*/
15262 	if (ddi_prop_lookup_byte_array(DDI_DEV_T_ANY, dip,
15263 	    DDI_PROP_DONTPASS | DDI_PROP_NOTPROM, prop,
15264 	    (uchar_t **)&chs_p, &elements) != DDI_PROP_SUCCESS) {
15265 		QL_PRINT_2(CE_CONT, "(%d): no prop exit\n", ha->instance);
15266 		return (DDI_FAILURE);
15267 	}
15268 
15269 	ql_pci_config_put16(ha, PCI_CONF_COMM, chs_p->chs_command);
15270 
15271 	if ((chs_p->chs_header_type & PCI_HEADER_TYPE_M) == PCI_HEADER_ONE) {
15272 		ql_pci_config_put16(ha, PCI_BCNF_BCNTRL,
15273 		    chs_p->chs_bridge_control);
15274 	}
15275 
15276 	ql_pci_config_put8(ha, PCI_CONF_CACHE_LINESZ,
15277 	    chs_p->chs_cache_line_size);
15278 
15279 	ql_pci_config_put8(ha, PCI_CONF_LATENCY_TIMER,
15280 	    chs_p->chs_latency_timer);
15281 
15282 	if ((chs_p->chs_header_type & PCI_HEADER_TYPE_M) == PCI_HEADER_ONE) {
15283 		ql_pci_config_put8(ha, PCI_BCNF_LATENCY_TIMER,
15284 		    chs_p->chs_sec_latency_timer);
15285 	}
15286 
15287 	ql_pci_config_put32(ha, PCI_CONF_BASE0, chs_p->chs_base0);
15288 	ql_pci_config_put32(ha, PCI_CONF_BASE1, chs_p->chs_base1);
15289 	ql_pci_config_put32(ha, PCI_CONF_BASE2, chs_p->chs_base2);
15290 	ql_pci_config_put32(ha, PCI_CONF_BASE3, chs_p->chs_base3);
15291 	ql_pci_config_put32(ha, PCI_CONF_BASE4, chs_p->chs_base4);
15292 	ql_pci_config_put32(ha, PCI_CONF_BASE5, chs_p->chs_base5);
15293 
15294 	ddi_prop_free(chs_p);
15295 
15296 	/*LINTED [Solaris DDI_DEV_T_NONE Lint warning]*/
15297 	if (ndi_prop_remove(DDI_DEV_T_NONE, dip, prop) != DDI_PROP_SUCCESS) {
15298 		cmn_err(CE_WARN, "!Qlogic %s(%d): can't remove prop %s",
15299 		    QL_NAME, ddi_get_instance(dip), prop);
15300 	}
15301 
15302 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
15303 
15304 	return (DDI_SUCCESS);
15305 }
15306 
15307 uint8_t
ql_pci_config_get8(ql_adapter_state_t * ha,off_t off)15308 ql_pci_config_get8(ql_adapter_state_t *ha, off_t off)
15309 {
15310 	if (CFG_IST(ha, CFG_SBUS_CARD)) {
15311 		return (ddi_get8(ha->sbus_config_handle,
15312 		    (uint8_t *)(ha->sbus_config_base + off)));
15313 	}
15314 
15315 #ifdef KERNEL_32
15316 	return (pci_config_getb(ha->pci_handle, off));
15317 #else
15318 	return (pci_config_get8(ha->pci_handle, off));
15319 #endif
15320 }
15321 
15322 uint16_t
ql_pci_config_get16(ql_adapter_state_t * ha,off_t off)15323 ql_pci_config_get16(ql_adapter_state_t *ha, off_t off)
15324 {
15325 	if (CFG_IST(ha, CFG_SBUS_CARD)) {
15326 		return (ddi_get16(ha->sbus_config_handle,
15327 		    (uint16_t *)(ha->sbus_config_base + off)));
15328 	}
15329 
15330 #ifdef KERNEL_32
15331 	return (pci_config_getw(ha->pci_handle, off));
15332 #else
15333 	return (pci_config_get16(ha->pci_handle, off));
15334 #endif
15335 }
15336 
15337 uint32_t
ql_pci_config_get32(ql_adapter_state_t * ha,off_t off)15338 ql_pci_config_get32(ql_adapter_state_t *ha, off_t off)
15339 {
15340 	if (CFG_IST(ha, CFG_SBUS_CARD)) {
15341 		return (ddi_get32(ha->sbus_config_handle,
15342 		    (uint32_t *)(ha->sbus_config_base + off)));
15343 	}
15344 
15345 #ifdef KERNEL_32
15346 	return (pci_config_getl(ha->pci_handle, off));
15347 #else
15348 	return (pci_config_get32(ha->pci_handle, off));
15349 #endif
15350 }
15351 
15352 void
ql_pci_config_put8(ql_adapter_state_t * ha,off_t off,uint8_t val)15353 ql_pci_config_put8(ql_adapter_state_t *ha, off_t off, uint8_t val)
15354 {
15355 	if (CFG_IST(ha, CFG_SBUS_CARD)) {
15356 		ddi_put8(ha->sbus_config_handle,
15357 		    (uint8_t *)(ha->sbus_config_base + off), val);
15358 	} else {
15359 #ifdef KERNEL_32
15360 		pci_config_putb(ha->pci_handle, off, val);
15361 #else
15362 		pci_config_put8(ha->pci_handle, off, val);
15363 #endif
15364 	}
15365 }
15366 
15367 void
ql_pci_config_put16(ql_adapter_state_t * ha,off_t off,uint16_t val)15368 ql_pci_config_put16(ql_adapter_state_t *ha, off_t off, uint16_t val)
15369 {
15370 	if (CFG_IST(ha, CFG_SBUS_CARD)) {
15371 		ddi_put16(ha->sbus_config_handle,
15372 		    (uint16_t *)(ha->sbus_config_base + off), val);
15373 	} else {
15374 #ifdef KERNEL_32
15375 		pci_config_putw(ha->pci_handle, off, val);
15376 #else
15377 		pci_config_put16(ha->pci_handle, off, val);
15378 #endif
15379 	}
15380 }
15381 
15382 void
ql_pci_config_put32(ql_adapter_state_t * ha,off_t off,uint32_t val)15383 ql_pci_config_put32(ql_adapter_state_t *ha, off_t off, uint32_t val)
15384 {
15385 	if (CFG_IST(ha, CFG_SBUS_CARD)) {
15386 		ddi_put32(ha->sbus_config_handle,
15387 		    (uint32_t *)(ha->sbus_config_base + off), val);
15388 	} else {
15389 #ifdef KERNEL_32
15390 		pci_config_putl(ha->pci_handle, off, val);
15391 #else
15392 		pci_config_put32(ha->pci_handle, off, val);
15393 #endif
15394 	}
15395 }
15396 
15397 /*
15398  * ql_halt
15399  *	Waits for commands that are running to finish and
15400  *	if they do not, commands are aborted.
15401  *	Finally the adapter is reset.
15402  *
15403  * Input:
15404  *	ha:	adapter state pointer.
15405  *	pwr:	power state.
15406  *
15407  * Context:
15408  *	Kernel context.
15409  */
15410 static void
ql_halt(ql_adapter_state_t * ha,int pwr)15411 ql_halt(ql_adapter_state_t *ha, int pwr)
15412 {
15413 	uint32_t	cnt;
15414 	ql_tgt_t	*tq;
15415 	ql_srb_t	*sp;
15416 	uint16_t	index;
15417 	ql_link_t	*link;
15418 
15419 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
15420 
15421 	/* Wait for all commands running to finish. */
15422 	for (index = 0; index < DEVICE_HEAD_LIST_SIZE; index++) {
15423 		for (link = ha->dev[index].first; link != NULL;
15424 		    link = link->next) {
15425 			tq = link->base_address;
15426 			(void) ql_abort_device(ha, tq, 0);
15427 
15428 			/* Wait for 30 seconds for commands to finish. */
15429 			for (cnt = 3000; cnt != 0; cnt--) {
15430 				/* Acquire device queue lock. */
15431 				DEVICE_QUEUE_LOCK(tq);
15432 				if (tq->outcnt == 0) {
15433 					/* Release device queue lock. */
15434 					DEVICE_QUEUE_UNLOCK(tq);
15435 					break;
15436 				} else {
15437 					/* Release device queue lock. */
15438 					DEVICE_QUEUE_UNLOCK(tq);
15439 					ql_delay(ha, 10000);
15440 				}
15441 			}
15442 
15443 			/* Finish any commands waiting for more status. */
15444 			if (ha->status_srb != NULL) {
15445 				sp = ha->status_srb;
15446 				ha->status_srb = NULL;
15447 				sp->cmd.next = NULL;
15448 				ql_done(&sp->cmd);
15449 			}
15450 
15451 			/* Abort commands that did not finish. */
15452 			if (cnt == 0) {
15453 				for (cnt = 1; cnt < MAX_OUTSTANDING_COMMANDS;
15454 				    cnt++) {
15455 					if (ha->pending_cmds.first != NULL) {
15456 						ql_start_iocb(ha, NULL);
15457 						cnt = 1;
15458 					}
15459 					sp = ha->outstanding_cmds[cnt];
15460 					if (sp != NULL &&
15461 					    sp->lun_queue->target_queue ==
15462 					    tq) {
15463 						(void) ql_abort((opaque_t)ha,
15464 						    sp->pkt, 0);
15465 					}
15466 				}
15467 			}
15468 		}
15469 	}
15470 
15471 	/* Shutdown IP. */
15472 	if (ha->flags & IP_INITIALIZED) {
15473 		(void) ql_shutdown_ip(ha);
15474 	}
15475 
15476 	/* Stop all timers. */
15477 	ADAPTER_STATE_LOCK(ha);
15478 	ha->port_retry_timer = 0;
15479 	ha->loop_down_timer = LOOP_DOWN_TIMER_OFF;
15480 	ha->watchdog_timer = 0;
15481 	ADAPTER_STATE_UNLOCK(ha);
15482 
15483 	if (pwr == PM_LEVEL_D3) {
15484 		ADAPTER_STATE_LOCK(ha);
15485 		ha->flags &= ~ONLINE;
15486 		ADAPTER_STATE_UNLOCK(ha);
15487 
15488 		/* Reset ISP chip. */
15489 		ql_reset_chip(ha);
15490 	}
15491 
15492 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
15493 }
15494 
15495 /*
15496  * ql_get_dma_mem
15497  *	Function used to allocate dma memory.
15498  *
15499  * Input:
15500  *	ha:			adapter state pointer.
15501  *	mem:			pointer to dma memory object.
15502  *	size:			size of the request in bytes
15503  *
15504  * Returns:
15505  *	qn local function return status code.
15506  *
15507  * Context:
15508  *	Kernel context.
15509  */
15510 int
ql_get_dma_mem(ql_adapter_state_t * ha,dma_mem_t * mem,uint32_t size,mem_alloc_type_t allocation_type,mem_alignment_t alignment)15511 ql_get_dma_mem(ql_adapter_state_t *ha, dma_mem_t *mem, uint32_t size,
15512     mem_alloc_type_t allocation_type, mem_alignment_t alignment)
15513 {
15514 	int	rval;
15515 
15516 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
15517 
15518 	mem->size = size;
15519 	mem->type = allocation_type;
15520 	mem->cookie_count = 1;
15521 
15522 	switch (alignment) {
15523 	case QL_DMA_DATA_ALIGN:
15524 		mem->alignment = QL_DMA_ALIGN_8_BYTE_BOUNDARY;
15525 		break;
15526 	case QL_DMA_RING_ALIGN:
15527 		mem->alignment = QL_DMA_ALIGN_64_BYTE_BOUNDARY;
15528 		break;
15529 	default:
15530 		EL(ha, "failed, unknown alignment type %x\n", alignment);
15531 		break;
15532 	}
15533 
15534 	if ((rval = ql_alloc_phys(ha, mem, KM_SLEEP)) != QL_SUCCESS) {
15535 		ql_free_phys(ha, mem);
15536 		EL(ha, "failed, alloc_phys=%xh\n", rval);
15537 	}
15538 
15539 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
15540 
15541 	return (rval);
15542 }
15543 
15544 /*
15545  * ql_alloc_phys
15546  *	Function used to allocate memory and zero it.
15547  *	Memory is below 4 GB.
15548  *
15549  * Input:
15550  *	ha:			adapter state pointer.
15551  *	mem:			pointer to dma memory object.
15552  *	sleep:			KM_SLEEP/KM_NOSLEEP flag.
15553  *	mem->cookie_count	number of segments allowed.
15554  *	mem->type		memory allocation type.
15555  *	mem->size		memory size.
15556  *	mem->alignment		memory alignment.
15557  *
15558  * Returns:
15559  *	qn local function return status code.
15560  *
15561  * Context:
15562  *	Kernel context.
15563  */
15564 int
ql_alloc_phys(ql_adapter_state_t * ha,dma_mem_t * mem,int sleep)15565 ql_alloc_phys(ql_adapter_state_t *ha, dma_mem_t *mem, int sleep)
15566 {
15567 	size_t			rlen;
15568 	ddi_dma_attr_t		dma_attr;
15569 	ddi_device_acc_attr_t	acc_attr = ql_dev_acc_attr;
15570 
15571 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
15572 
15573 	dma_attr = CFG_IST(ha, CFG_ENABLE_64BIT_ADDRESSING) ?
15574 	    ql_64bit_io_dma_attr : ql_32bit_io_dma_attr;
15575 
15576 	dma_attr.dma_attr_align = mem->alignment; /* DMA address alignment */
15577 	dma_attr.dma_attr_sgllen = (int)mem->cookie_count;
15578 
15579 	/*
15580 	 * Workaround for SUN XMITS buffer must end and start on 8 byte
15581 	 * boundary. Else, hardware will overrun the buffer. Simple fix is
15582 	 * to make sure buffer has enough room for overrun.
15583 	 */
15584 	if (mem->size & 7) {
15585 		mem->size += 8 - (mem->size & 7);
15586 	}
15587 
15588 	mem->flags = DDI_DMA_CONSISTENT;
15589 
15590 	/*
15591 	 * Allocate DMA memory for command.
15592 	 */
15593 	if (ddi_dma_alloc_handle(ha->dip, &dma_attr, (sleep == KM_SLEEP) ?
15594 	    DDI_DMA_SLEEP : DDI_DMA_DONTWAIT, NULL, &mem->dma_handle) !=
15595 	    DDI_SUCCESS) {
15596 		EL(ha, "failed, ddi_dma_alloc_handle\n");
15597 		mem->dma_handle = NULL;
15598 		return (QL_MEMORY_ALLOC_FAILED);
15599 	}
15600 
15601 	switch (mem->type) {
15602 	case KERNEL_MEM:
15603 		mem->bp = kmem_zalloc(mem->size, sleep);
15604 		break;
15605 	case BIG_ENDIAN_DMA:
15606 	case LITTLE_ENDIAN_DMA:
15607 	case NO_SWAP_DMA:
15608 		if (mem->type == BIG_ENDIAN_DMA) {
15609 			acc_attr.devacc_attr_endian_flags =
15610 			    DDI_STRUCTURE_BE_ACC;
15611 		} else if (mem->type == NO_SWAP_DMA) {
15612 			acc_attr.devacc_attr_endian_flags = DDI_NEVERSWAP_ACC;
15613 		}
15614 		if (ddi_dma_mem_alloc(mem->dma_handle, mem->size, &acc_attr,
15615 		    mem->flags, (sleep == KM_SLEEP) ? DDI_DMA_SLEEP :
15616 		    DDI_DMA_DONTWAIT, NULL, (caddr_t *)&mem->bp, &rlen,
15617 		    &mem->acc_handle) == DDI_SUCCESS) {
15618 			bzero(mem->bp, mem->size);
15619 			/* ensure we got what we asked for (32bit) */
15620 			if (dma_attr.dma_attr_addr_hi == NULL) {
15621 				if (mem->cookie.dmac_notused != NULL) {
15622 					EL(ha, "failed, ddi_dma_mem_alloc "
15623 					    "returned 64 bit DMA address\n");
15624 					ql_free_phys(ha, mem);
15625 					return (QL_MEMORY_ALLOC_FAILED);
15626 				}
15627 			}
15628 		} else {
15629 			mem->acc_handle = NULL;
15630 			mem->bp = NULL;
15631 		}
15632 		break;
15633 	default:
15634 		EL(ha, "failed, unknown type=%xh\n", mem->type);
15635 		mem->acc_handle = NULL;
15636 		mem->bp = NULL;
15637 		break;
15638 	}
15639 
15640 	if (mem->bp == NULL) {
15641 		EL(ha, "failed, ddi_dma_mem_alloc\n");
15642 		ddi_dma_free_handle(&mem->dma_handle);
15643 		mem->dma_handle = NULL;
15644 		return (QL_MEMORY_ALLOC_FAILED);
15645 	}
15646 
15647 	mem->flags |= DDI_DMA_RDWR;
15648 
15649 	if (ql_bind_dma_buffer(ha, mem, sleep) != DDI_DMA_MAPPED) {
15650 		EL(ha, "failed, ddi_dma_addr_bind_handle\n");
15651 		ql_free_phys(ha, mem);
15652 		return (QL_MEMORY_ALLOC_FAILED);
15653 	}
15654 
15655 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
15656 
15657 	return (QL_SUCCESS);
15658 }
15659 
15660 /*
15661  * ql_free_phys
15662  *	Function used to free physical memory.
15663  *
15664  * Input:
15665  *	ha:	adapter state pointer.
15666  *	mem:	pointer to dma memory object.
15667  *
15668  * Context:
15669  *	Kernel context.
15670  */
15671 void
ql_free_phys(ql_adapter_state_t * ha,dma_mem_t * mem)15672 ql_free_phys(ql_adapter_state_t *ha, dma_mem_t *mem)
15673 {
15674 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
15675 
15676 	if (mem != NULL && mem->dma_handle != NULL) {
15677 		ql_unbind_dma_buffer(ha, mem);
15678 		switch (mem->type) {
15679 		case KERNEL_MEM:
15680 			if (mem->bp != NULL) {
15681 				kmem_free(mem->bp, mem->size);
15682 			}
15683 			break;
15684 		case LITTLE_ENDIAN_DMA:
15685 		case BIG_ENDIAN_DMA:
15686 		case NO_SWAP_DMA:
15687 			if (mem->acc_handle != NULL) {
15688 				ddi_dma_mem_free(&mem->acc_handle);
15689 				mem->acc_handle = NULL;
15690 			}
15691 			break;
15692 		default:
15693 			break;
15694 		}
15695 		mem->bp = NULL;
15696 		ddi_dma_free_handle(&mem->dma_handle);
15697 		mem->dma_handle = NULL;
15698 	}
15699 
15700 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
15701 }
15702 
15703 /*
15704  * ql_alloc_dma_resouce.
15705  *	Allocates DMA resource for buffer.
15706  *
15707  * Input:
15708  *	ha:			adapter state pointer.
15709  *	mem:			pointer to dma memory object.
15710  *	sleep:			KM_SLEEP/KM_NOSLEEP flag.
15711  *	mem->cookie_count	number of segments allowed.
15712  *	mem->type		memory allocation type.
15713  *	mem->size		memory size.
15714  *	mem->bp			pointer to memory or struct buf
15715  *
15716  * Returns:
15717  *	qn local function return status code.
15718  *
15719  * Context:
15720  *	Kernel context.
15721  */
15722 int
ql_alloc_dma_resouce(ql_adapter_state_t * ha,dma_mem_t * mem,int sleep)15723 ql_alloc_dma_resouce(ql_adapter_state_t *ha, dma_mem_t *mem, int sleep)
15724 {
15725 	ddi_dma_attr_t	dma_attr;
15726 
15727 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
15728 
15729 	dma_attr = CFG_IST(ha, CFG_ENABLE_64BIT_ADDRESSING) ?
15730 	    ql_64bit_io_dma_attr : ql_32bit_io_dma_attr;
15731 	dma_attr.dma_attr_sgllen = (int)mem->cookie_count;
15732 
15733 	/*
15734 	 * Allocate DMA handle for command.
15735 	 */
15736 	if (ddi_dma_alloc_handle(ha->dip, &dma_attr, (sleep == KM_SLEEP) ?
15737 	    DDI_DMA_SLEEP : DDI_DMA_DONTWAIT, NULL, &mem->dma_handle) !=
15738 	    DDI_SUCCESS) {
15739 		EL(ha, "failed, ddi_dma_alloc_handle\n");
15740 		mem->dma_handle = NULL;
15741 		return (QL_MEMORY_ALLOC_FAILED);
15742 	}
15743 
15744 	mem->flags = DDI_DMA_RDWR | DDI_DMA_CONSISTENT;
15745 
15746 	if (ql_bind_dma_buffer(ha, mem, sleep) != DDI_DMA_MAPPED) {
15747 		EL(ha, "failed, bind_dma_buffer\n");
15748 		ddi_dma_free_handle(&mem->dma_handle);
15749 		mem->dma_handle = NULL;
15750 		return (QL_MEMORY_ALLOC_FAILED);
15751 	}
15752 
15753 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
15754 
15755 	return (QL_SUCCESS);
15756 }
15757 
15758 /*
15759  * ql_free_dma_resource
15760  *	Frees DMA resources.
15761  *
15762  * Input:
15763  *	ha:		adapter state pointer.
15764  *	mem:		pointer to dma memory object.
15765  *	mem->dma_handle	DMA memory handle.
15766  *
15767  * Context:
15768  *	Kernel context.
15769  */
15770 void
ql_free_dma_resource(ql_adapter_state_t * ha,dma_mem_t * mem)15771 ql_free_dma_resource(ql_adapter_state_t *ha, dma_mem_t *mem)
15772 {
15773 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
15774 
15775 	ql_free_phys(ha, mem);
15776 
15777 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
15778 }
15779 
15780 /*
15781  * ql_bind_dma_buffer
15782  *	Binds DMA buffer.
15783  *
15784  * Input:
15785  *	ha:			adapter state pointer.
15786  *	mem:			pointer to dma memory object.
15787  *	sleep:			KM_SLEEP or KM_NOSLEEP.
15788  *	mem->dma_handle		DMA memory handle.
15789  *	mem->cookie_count	number of segments allowed.
15790  *	mem->type		memory allocation type.
15791  *	mem->size		memory size.
15792  *	mem->bp			pointer to memory or struct buf
15793  *
15794  * Returns:
15795  *	mem->cookies		pointer to list of cookies.
15796  *	mem->cookie_count	number of cookies.
15797  *	status			success = DDI_DMA_MAPPED
15798  *				DDI_DMA_PARTIAL_MAP, DDI_DMA_INUSE,
15799  *				DDI_DMA_NORESOURCES, DDI_DMA_NOMAPPING or
15800  *				DDI_DMA_TOOBIG
15801  *
15802  * Context:
15803  *	Kernel context.
15804  */
15805 static int
ql_bind_dma_buffer(ql_adapter_state_t * ha,dma_mem_t * mem,int sleep)15806 ql_bind_dma_buffer(ql_adapter_state_t *ha, dma_mem_t *mem, int sleep)
15807 {
15808 	int			rval;
15809 	ddi_dma_cookie_t	*cookiep;
15810 	uint32_t		cnt = mem->cookie_count;
15811 
15812 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
15813 
15814 	if (mem->type == STRUCT_BUF_MEMORY) {
15815 		rval = ddi_dma_buf_bind_handle(mem->dma_handle, mem->bp,
15816 		    mem->flags, (sleep == KM_SLEEP) ? DDI_DMA_SLEEP :
15817 		    DDI_DMA_DONTWAIT, NULL, &mem->cookie, &mem->cookie_count);
15818 	} else {
15819 		rval = ddi_dma_addr_bind_handle(mem->dma_handle, NULL, mem->bp,
15820 		    mem->size, mem->flags, (sleep == KM_SLEEP) ?
15821 		    DDI_DMA_SLEEP : DDI_DMA_DONTWAIT, NULL, &mem->cookie,
15822 		    &mem->cookie_count);
15823 	}
15824 
15825 	if (rval == DDI_DMA_MAPPED) {
15826 		if (mem->cookie_count > cnt) {
15827 			(void) ddi_dma_unbind_handle(mem->dma_handle);
15828 			EL(ha, "failed, cookie_count %d > %d\n",
15829 			    mem->cookie_count, cnt);
15830 			rval = DDI_DMA_TOOBIG;
15831 		} else {
15832 			if (mem->cookie_count > 1) {
15833 				if (mem->cookies = kmem_zalloc(
15834 				    sizeof (ddi_dma_cookie_t) *
15835 				    mem->cookie_count, sleep)) {
15836 					*mem->cookies = mem->cookie;
15837 					cookiep = mem->cookies;
15838 					for (cnt = 1; cnt < mem->cookie_count;
15839 					    cnt++) {
15840 						ddi_dma_nextcookie(
15841 						    mem->dma_handle,
15842 						    ++cookiep);
15843 					}
15844 				} else {
15845 					(void) ddi_dma_unbind_handle(
15846 					    mem->dma_handle);
15847 					EL(ha, "failed, kmem_zalloc\n");
15848 					rval = DDI_DMA_NORESOURCES;
15849 				}
15850 			} else {
15851 				/*
15852 				 * It has been reported that dmac_size at times
15853 				 * may be incorrect on sparc machines so for
15854 				 * sparc machines that only have one segment
15855 				 * use the buffer size instead.
15856 				 */
15857 				mem->cookies = &mem->cookie;
15858 				mem->cookies->dmac_size = mem->size;
15859 			}
15860 		}
15861 	}
15862 
15863 	if (rval != DDI_DMA_MAPPED) {
15864 		EL(ha, "failed=%xh\n", rval);
15865 	} else {
15866 		/*EMPTY*/
15867 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
15868 	}
15869 
15870 	return (rval);
15871 }
15872 
15873 /*
15874  * ql_unbind_dma_buffer
15875  *	Unbinds DMA buffer.
15876  *
15877  * Input:
15878  *	ha:			adapter state pointer.
15879  *	mem:			pointer to dma memory object.
15880  *	mem->dma_handle		DMA memory handle.
15881  *	mem->cookies		pointer to cookie list.
15882  *	mem->cookie_count	number of cookies.
15883  *
15884  * Context:
15885  *	Kernel context.
15886  */
15887 /* ARGSUSED */
15888 static void
ql_unbind_dma_buffer(ql_adapter_state_t * ha,dma_mem_t * mem)15889 ql_unbind_dma_buffer(ql_adapter_state_t *ha, dma_mem_t *mem)
15890 {
15891 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
15892 
15893 	(void) ddi_dma_unbind_handle(mem->dma_handle);
15894 	if (mem->cookie_count > 1) {
15895 		kmem_free(mem->cookies, sizeof (ddi_dma_cookie_t) *
15896 		    mem->cookie_count);
15897 		mem->cookies = NULL;
15898 	}
15899 	mem->cookie_count = 0;
15900 
15901 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
15902 }
15903 
15904 static int
ql_suspend_adapter(ql_adapter_state_t * ha)15905 ql_suspend_adapter(ql_adapter_state_t *ha)
15906 {
15907 	clock_t timer = 32 * drv_usectohz(1000000);
15908 
15909 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
15910 
15911 	/*
15912 	 * First we will claim mbox ownership so that no
15913 	 * thread using mbox hangs when we disable the
15914 	 * interrupt in the middle of it.
15915 	 */
15916 	MBX_REGISTER_LOCK(ha);
15917 
15918 	/* Check for mailbox available, if not wait for signal. */
15919 	while (ha->mailbox_flags & MBX_BUSY_FLG) {
15920 		ha->mailbox_flags = (uint8_t)
15921 		    (ha->mailbox_flags | MBX_WANT_FLG);
15922 
15923 		/* 30 seconds from now */
15924 		if (cv_reltimedwait(&ha->cv_mbx_wait, &ha->mbx_mutex,
15925 		    timer, TR_CLOCK_TICK) == -1) {
15926 
15927 			/* Release mailbox register lock. */
15928 			MBX_REGISTER_UNLOCK(ha);
15929 			EL(ha, "failed, Suspend mbox");
15930 			return (QL_FUNCTION_TIMEOUT);
15931 		}
15932 	}
15933 
15934 	/* Set busy flag. */
15935 	ha->mailbox_flags = (uint8_t)(ha->mailbox_flags | MBX_BUSY_FLG);
15936 	MBX_REGISTER_UNLOCK(ha);
15937 
15938 	(void) ql_wait_outstanding(ha);
15939 
15940 	/*
15941 	 * here we are sure that there will not be any mbox interrupt.
15942 	 * So, let's make sure that we return back all the outstanding
15943 	 * cmds as well as internally queued commands.
15944 	 */
15945 	ql_halt(ha, PM_LEVEL_D0);
15946 
15947 	if (ha->power_level != PM_LEVEL_D3) {
15948 		/* Disable ISP interrupts. */
15949 		WRT16_IO_REG(ha, ictrl, 0);
15950 	}
15951 
15952 	ADAPTER_STATE_LOCK(ha);
15953 	ha->flags &= ~INTERRUPTS_ENABLED;
15954 	ADAPTER_STATE_UNLOCK(ha);
15955 
15956 	MBX_REGISTER_LOCK(ha);
15957 	/* Reset busy status. */
15958 	ha->mailbox_flags = (uint8_t)(ha->mailbox_flags & ~MBX_BUSY_FLG);
15959 
15960 	/* If thread is waiting for mailbox go signal it to start. */
15961 	if (ha->mailbox_flags & MBX_WANT_FLG) {
15962 		ha->mailbox_flags = (uint8_t)
15963 		    (ha->mailbox_flags & ~MBX_WANT_FLG);
15964 		cv_broadcast(&ha->cv_mbx_wait);
15965 	}
15966 	/* Release mailbox register lock. */
15967 	MBX_REGISTER_UNLOCK(ha);
15968 
15969 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
15970 
15971 	return (QL_SUCCESS);
15972 }
15973 
15974 /*
15975  * ql_add_link_b
15976  *	Add link to the end of the chain.
15977  *
15978  * Input:
15979  *	head = Head of link list.
15980  *	link = link to be added.
15981  *	LOCK must be already obtained.
15982  *
15983  * Context:
15984  *	Interrupt or Kernel context, no mailbox commands allowed.
15985  */
15986 void
ql_add_link_b(ql_head_t * head,ql_link_t * link)15987 ql_add_link_b(ql_head_t *head, ql_link_t *link)
15988 {
15989 	/* at the end there isn't a next */
15990 	link->next = NULL;
15991 
15992 	if ((link->prev = head->last) == NULL) {
15993 		head->first = link;
15994 	} else {
15995 		head->last->next = link;
15996 	}
15997 
15998 	head->last = link;
15999 	link->head = head;	/* the queue we're on */
16000 }
16001 
16002 /*
16003  * ql_add_link_t
16004  *	Add link to the beginning of the chain.
16005  *
16006  * Input:
16007  *	head = Head of link list.
16008  *	link = link to be added.
16009  *	LOCK must be already obtained.
16010  *
16011  * Context:
16012  *	Interrupt or Kernel context, no mailbox commands allowed.
16013  */
16014 void
ql_add_link_t(ql_head_t * head,ql_link_t * link)16015 ql_add_link_t(ql_head_t *head, ql_link_t *link)
16016 {
16017 	link->prev = NULL;
16018 
16019 	if ((link->next = head->first) == NULL)	{
16020 		head->last = link;
16021 	} else {
16022 		head->first->prev = link;
16023 	}
16024 
16025 	head->first = link;
16026 	link->head = head;	/* the queue we're on */
16027 }
16028 
16029 /*
16030  * ql_remove_link
16031  *	Remove a link from the chain.
16032  *
16033  * Input:
16034  *	head = Head of link list.
16035  *	link = link to be removed.
16036  *	LOCK must be already obtained.
16037  *
16038  * Context:
16039  *	Interrupt or Kernel context, no mailbox commands allowed.
16040  */
16041 void
ql_remove_link(ql_head_t * head,ql_link_t * link)16042 ql_remove_link(ql_head_t *head, ql_link_t *link)
16043 {
16044 	if (link->prev != NULL) {
16045 		if ((link->prev->next = link->next) == NULL) {
16046 			head->last = link->prev;
16047 		} else {
16048 			link->next->prev = link->prev;
16049 		}
16050 	} else if ((head->first = link->next) == NULL) {
16051 		head->last = NULL;
16052 	} else {
16053 		head->first->prev = NULL;
16054 	}
16055 
16056 	/* not on a queue any more */
16057 	link->prev = link->next = NULL;
16058 	link->head = NULL;
16059 }
16060 
16061 /*
16062  * ql_chg_endian
16063  *	Change endianess of byte array.
16064  *
16065  * Input:
16066  *	buf = array pointer.
16067  *	size = size of array in bytes.
16068  *
16069  * Context:
16070  *	Interrupt or Kernel context, no mailbox commands allowed.
16071  */
16072 void
ql_chg_endian(uint8_t buf[],size_t size)16073 ql_chg_endian(uint8_t buf[], size_t size)
16074 {
16075 	uint8_t byte;
16076 	size_t  cnt1;
16077 	size_t  cnt;
16078 
16079 	cnt1 = size - 1;
16080 	for (cnt = 0; cnt < size / 2; cnt++) {
16081 		byte = buf[cnt1];
16082 		buf[cnt1] = buf[cnt];
16083 		buf[cnt] = byte;
16084 		cnt1--;
16085 	}
16086 }
16087 
16088 /*
16089  * ql_bstr_to_dec
16090  *	Convert decimal byte string to number.
16091  *
16092  * Input:
16093  *	s:	byte string pointer.
16094  *	ans:	interger pointer for number.
16095  *	size:	number of ascii bytes.
16096  *
16097  * Returns:
16098  *	success = number of ascii bytes processed.
16099  *
16100  * Context:
16101  *	Kernel/Interrupt context.
16102  */
16103 static int
ql_bstr_to_dec(char * s,uint32_t * ans,uint32_t size)16104 ql_bstr_to_dec(char *s, uint32_t *ans, uint32_t size)
16105 {
16106 	int			mul, num, cnt, pos;
16107 	char			*str;
16108 
16109 	/* Calculate size of number. */
16110 	if (size == 0) {
16111 		for (str = s; *str >= '0' && *str <= '9'; str++) {
16112 			size++;
16113 		}
16114 	}
16115 
16116 	*ans = 0;
16117 	for (cnt = 0; *s != '\0' && size; size--, cnt++) {
16118 		if (*s >= '0' && *s <= '9') {
16119 			num = *s++ - '0';
16120 		} else {
16121 			break;
16122 		}
16123 
16124 		for (mul = 1, pos = 1; pos < size; pos++) {
16125 			mul *= 10;
16126 		}
16127 		*ans += num * mul;
16128 	}
16129 
16130 	return (cnt);
16131 }
16132 
16133 /*
16134  * ql_delay
16135  *	Calls delay routine if threads are not suspended, otherwise, busy waits
16136  *	Minimum = 1 tick = 10ms
16137  *
16138  * Input:
16139  *	dly = delay time in microseconds.
16140  *
16141  * Context:
16142  *	Kernel or Interrupt context, no mailbox commands allowed.
16143  */
16144 void
ql_delay(ql_adapter_state_t * ha,clock_t usecs)16145 ql_delay(ql_adapter_state_t *ha, clock_t usecs)
16146 {
16147 	if (QL_DAEMON_SUSPENDED(ha) || ddi_in_panic()) {
16148 		drv_usecwait(usecs);
16149 	} else {
16150 		delay(drv_usectohz(usecs));
16151 	}
16152 }
16153 
16154 /*
16155  * ql_stall_drv
16156  *	Stalls one or all driver instances, waits for 30 seconds.
16157  *
16158  * Input:
16159  *	ha:		adapter state pointer or NULL for all.
16160  *	options:	BIT_0 --> leave driver stalled on exit if
16161  *				  failed.
16162  *
16163  * Returns:
16164  *	ql local function return status code.
16165  *
16166  * Context:
16167  *	Kernel context.
16168  */
16169 int
ql_stall_driver(ql_adapter_state_t * ha,uint32_t options)16170 ql_stall_driver(ql_adapter_state_t *ha, uint32_t options)
16171 {
16172 	ql_link_t		*link;
16173 	ql_adapter_state_t	*ha2;
16174 	uint32_t		timer;
16175 
16176 	QL_PRINT_3(CE_CONT, "started\n");
16177 
16178 	/* Wait for 30 seconds for daemons unstall. */
16179 	timer = 3000;
16180 	link = ha == NULL ? ql_hba.first : &ha->hba;
16181 	while (link != NULL && timer) {
16182 		ha2 = link->base_address;
16183 
16184 		ql_awaken_task_daemon(ha2, NULL, DRIVER_STALL, 0);
16185 
16186 		if ((ha2->task_daemon_flags & TASK_DAEMON_ALIVE_FLG) == 0 ||
16187 		    (ha2->task_daemon_flags & TASK_DAEMON_STOP_FLG) != 0 ||
16188 		    (ha2->task_daemon_flags & TASK_DAEMON_STALLED_FLG &&
16189 		    ql_wait_outstanding(ha2) == MAX_OUTSTANDING_COMMANDS)) {
16190 			link = ha == NULL ? link->next : NULL;
16191 			continue;
16192 		}
16193 
16194 		ql_delay(ha2, 10000);
16195 		timer--;
16196 		link = ha == NULL ? ql_hba.first : &ha->hba;
16197 	}
16198 
16199 	if (ha2 != NULL && timer == 0) {
16200 		EL(ha2, "failed, tdf=%xh, exiting state is: %s\n",
16201 		    ha2->task_daemon_flags, (options & BIT_0 ? "stalled" :
16202 		    "unstalled"));
16203 		if (options & BIT_0) {
16204 			ql_awaken_task_daemon(ha2, NULL, 0, DRIVER_STALL);
16205 		}
16206 		return (QL_FUNCTION_TIMEOUT);
16207 	}
16208 
16209 	QL_PRINT_3(CE_CONT, "done\n");
16210 
16211 	return (QL_SUCCESS);
16212 }
16213 
16214 /*
16215  * ql_restart_driver
16216  *	Restarts one or all driver instances.
16217  *
16218  * Input:
16219  *	ha:	adapter state pointer or NULL for all.
16220  *
16221  * Context:
16222  *	Kernel context.
16223  */
16224 void
ql_restart_driver(ql_adapter_state_t * ha)16225 ql_restart_driver(ql_adapter_state_t *ha)
16226 {
16227 	ql_link_t		*link;
16228 	ql_adapter_state_t	*ha2;
16229 	uint32_t		timer;
16230 
16231 	QL_PRINT_3(CE_CONT, "started\n");
16232 
16233 	/* Tell all daemons to unstall. */
16234 	link = ha == NULL ? ql_hba.first : &ha->hba;
16235 	while (link != NULL) {
16236 		ha2 = link->base_address;
16237 
16238 		ql_awaken_task_daemon(ha2, NULL, 0, DRIVER_STALL);
16239 
16240 		link = ha == NULL ? link->next : NULL;
16241 	}
16242 
16243 	/* Wait for 30 seconds for all daemons unstall. */
16244 	timer = 3000;
16245 	link = ha == NULL ? ql_hba.first : &ha->hba;
16246 	while (link != NULL && timer) {
16247 		ha2 = link->base_address;
16248 
16249 		if ((ha2->task_daemon_flags & TASK_DAEMON_ALIVE_FLG) == 0 ||
16250 		    (ha2->task_daemon_flags & TASK_DAEMON_STOP_FLG) != 0 ||
16251 		    (ha2->task_daemon_flags & TASK_DAEMON_STALLED_FLG) == 0) {
16252 			QL_PRINT_2(CE_CONT, "(%d,%d): restarted\n",
16253 			    ha2->instance, ha2->vp_index);
16254 			ql_restart_queues(ha2);
16255 			link = ha == NULL ? link->next : NULL;
16256 			continue;
16257 		}
16258 
16259 		QL_PRINT_2(CE_CONT, "(%d,%d): failed, tdf=%xh\n",
16260 		    ha2->instance, ha2->vp_index, ha2->task_daemon_flags);
16261 
16262 		ql_delay(ha2, 10000);
16263 		timer--;
16264 		link = ha == NULL ? ql_hba.first : &ha->hba;
16265 	}
16266 
16267 	QL_PRINT_3(CE_CONT, "done\n");
16268 }
16269 
16270 /*
16271  * ql_setup_interrupts
16272  *	Sets up interrupts based on the HBA's and platform's
16273  *	capabilities (e.g., legacy / MSI / FIXED).
16274  *
16275  * Input:
16276  *	ha = adapter state pointer.
16277  *
16278  * Returns:
16279  *	DDI_SUCCESS or DDI_FAILURE.
16280  *
16281  * Context:
16282  *	Kernel context.
16283  */
16284 static int
ql_setup_interrupts(ql_adapter_state_t * ha)16285 ql_setup_interrupts(ql_adapter_state_t *ha)
16286 {
16287 	int32_t		rval = DDI_FAILURE;
16288 	int32_t		i;
16289 	int32_t		itypes = 0;
16290 
16291 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
16292 
16293 	/*
16294 	 * The Solaris Advanced Interrupt Functions (aif) are only
16295 	 * supported on s10U1 or greater.
16296 	 */
16297 	if (ql_os_release_level < 10 || ql_disable_aif != 0) {
16298 		EL(ha, "interrupt framework is not supported or is "
16299 		    "disabled, using legacy\n");
16300 		return (ql_legacy_intr(ha));
16301 	} else if (ql_os_release_level == 10) {
16302 		/*
16303 		 * See if the advanced interrupt functions (aif) are
16304 		 * in the kernel
16305 		 */
16306 		void	*fptr = (void *)&ddi_intr_get_supported_types;
16307 
16308 		if (fptr == NULL) {
16309 			EL(ha, "aif is not supported, using legacy "
16310 			    "interrupts (rev)\n");
16311 			return (ql_legacy_intr(ha));
16312 		}
16313 	}
16314 
16315 	/* See what types of interrupts this HBA and platform support */
16316 	if ((i = ddi_intr_get_supported_types(ha->dip, &itypes)) !=
16317 	    DDI_SUCCESS) {
16318 		EL(ha, "get supported types failed, rval=%xh, "
16319 		    "assuming FIXED\n", i);
16320 		itypes = DDI_INTR_TYPE_FIXED;
16321 	}
16322 
16323 	EL(ha, "supported types are: %xh\n", itypes);
16324 
16325 	if ((itypes & DDI_INTR_TYPE_MSIX) &&
16326 	    (rval = ql_setup_msix(ha)) == DDI_SUCCESS) {
16327 		EL(ha, "successful MSI-X setup\n");
16328 	} else if ((itypes & DDI_INTR_TYPE_MSI) &&
16329 	    (rval = ql_setup_msi(ha)) == DDI_SUCCESS) {
16330 		EL(ha, "successful MSI setup\n");
16331 	} else {
16332 		rval = ql_setup_fixed(ha);
16333 	}
16334 
16335 	if (rval != DDI_SUCCESS) {
16336 		EL(ha, "failed, aif, rval=%xh\n", rval);
16337 	} else {
16338 		/*EMPTY*/
16339 		QL_PRINT_3(CE_CONT, "(%d): done\n");
16340 	}
16341 
16342 	return (rval);
16343 }
16344 
16345 /*
16346  * ql_setup_msi
16347  *	Set up aif MSI interrupts
16348  *
16349  * Input:
16350  *	ha = adapter state pointer.
16351  *
16352  * Returns:
16353  *	DDI_SUCCESS or DDI_FAILURE.
16354  *
16355  * Context:
16356  *	Kernel context.
16357  */
16358 static int
ql_setup_msi(ql_adapter_state_t * ha)16359 ql_setup_msi(ql_adapter_state_t *ha)
16360 {
16361 	int32_t		count = 0;
16362 	int32_t		avail = 0;
16363 	int32_t		actual = 0;
16364 	int32_t		msitype = DDI_INTR_TYPE_MSI;
16365 	int32_t		ret;
16366 	ql_ifunc_t	itrfun[10] = {0};
16367 
16368 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
16369 
16370 	if (ql_disable_msi != 0) {
16371 		EL(ha, "MSI is disabled by user\n");
16372 		return (DDI_FAILURE);
16373 	}
16374 
16375 	/* MSI support is only suported on 24xx HBA's. */
16376 	if (!(CFG_IST(ha, CFG_CTRL_24258081))) {
16377 		EL(ha, "HBA does not support MSI\n");
16378 		return (DDI_FAILURE);
16379 	}
16380 
16381 	/* Get number of MSI interrupts the system supports */
16382 	if (((ret = ddi_intr_get_nintrs(ha->dip, msitype, &count)) !=
16383 	    DDI_SUCCESS) || count == 0) {
16384 		EL(ha, "failed, nintrs ret=%xh, cnt=%xh\n", ret, count);
16385 		return (DDI_FAILURE);
16386 	}
16387 
16388 	/* Get number of available MSI interrupts */
16389 	if (((ret = ddi_intr_get_navail(ha->dip, msitype, &avail)) !=
16390 	    DDI_SUCCESS) || avail == 0) {
16391 		EL(ha, "failed, navail ret=%xh, avail=%xh\n", ret, avail);
16392 		return (DDI_FAILURE);
16393 	}
16394 
16395 	/* MSI requires only 1.  */
16396 	count = 1;
16397 	itrfun[0].ifunc = &ql_isr_aif;
16398 
16399 	/* Allocate space for interrupt handles */
16400 	ha->hsize = ((uint32_t)(sizeof (ddi_intr_handle_t)) * count);
16401 	ha->htable = kmem_zalloc(ha->hsize, KM_SLEEP);
16402 
16403 	ha->iflags |= IFLG_INTR_MSI;
16404 
16405 	/* Allocate the interrupts */
16406 	if ((ret = ddi_intr_alloc(ha->dip, ha->htable, msitype, 0, count,
16407 	    &actual, 0)) != DDI_SUCCESS || actual < count) {
16408 		EL(ha, "failed, intr_alloc ret=%xh, count = %xh, "
16409 		    "actual=%xh\n", ret, count, actual);
16410 		ql_release_intr(ha);
16411 		return (DDI_FAILURE);
16412 	}
16413 
16414 	ha->intr_cnt = actual;
16415 
16416 	/* Get interrupt priority */
16417 	if ((ret = ddi_intr_get_pri(ha->htable[0], &ha->intr_pri)) !=
16418 	    DDI_SUCCESS) {
16419 		EL(ha, "failed, get_pri ret=%xh\n", ret);
16420 		ql_release_intr(ha);
16421 		return (ret);
16422 	}
16423 
16424 	/* Add the interrupt handler */
16425 	if ((ret = ddi_intr_add_handler(ha->htable[0], itrfun[0].ifunc,
16426 	    (caddr_t)ha, (caddr_t)0)) != DDI_SUCCESS) {
16427 		EL(ha, "failed, intr_add ret=%xh\n", ret);
16428 		ql_release_intr(ha);
16429 		return (ret);
16430 	}
16431 
16432 	/* Setup mutexes */
16433 	if ((ret = ql_init_mutex(ha)) != DDI_SUCCESS) {
16434 		EL(ha, "failed, mutex init ret=%xh\n", ret);
16435 		ql_release_intr(ha);
16436 		return (ret);
16437 	}
16438 
16439 	/* Get the capabilities */
16440 	(void) ddi_intr_get_cap(ha->htable[0], &ha->intr_cap);
16441 
16442 	/* Enable interrupts */
16443 	if (ha->intr_cap & DDI_INTR_FLAG_BLOCK) {
16444 		if ((ret = ddi_intr_block_enable(ha->htable, ha->intr_cnt)) !=
16445 		    DDI_SUCCESS) {
16446 			EL(ha, "failed, block enable, ret=%xh\n", ret);
16447 			ql_destroy_mutex(ha);
16448 			ql_release_intr(ha);
16449 			return (ret);
16450 		}
16451 	} else {
16452 		if ((ret = ddi_intr_enable(ha->htable[0])) != DDI_SUCCESS) {
16453 			EL(ha, "failed, intr enable, ret=%xh\n", ret);
16454 			ql_destroy_mutex(ha);
16455 			ql_release_intr(ha);
16456 			return (ret);
16457 		}
16458 	}
16459 
16460 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
16461 
16462 	return (DDI_SUCCESS);
16463 }
16464 
16465 /*
16466  * ql_setup_msix
16467  *	Set up aif MSI-X interrupts
16468  *
16469  * Input:
16470  *	ha = adapter state pointer.
16471  *
16472  * Returns:
16473  *	DDI_SUCCESS or DDI_FAILURE.
16474  *
16475  * Context:
16476  *	Kernel context.
16477  */
16478 static int
ql_setup_msix(ql_adapter_state_t * ha)16479 ql_setup_msix(ql_adapter_state_t *ha)
16480 {
16481 	uint16_t	hwvect;
16482 	int32_t		count = 0;
16483 	int32_t		avail = 0;
16484 	int32_t		actual = 0;
16485 	int32_t		msitype = DDI_INTR_TYPE_MSIX;
16486 	int32_t		ret;
16487 	uint32_t	i;
16488 	ql_ifunc_t	itrfun[QL_MSIX_MAXAIF] = {0};
16489 
16490 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
16491 
16492 	if (ql_disable_msix != 0) {
16493 		EL(ha, "MSI-X is disabled by user\n");
16494 		return (DDI_FAILURE);
16495 	}
16496 
16497 	/*
16498 	 * MSI-X support is only available on 24xx HBA's that have
16499 	 * rev A2 parts (revid = 3) or greater.
16500 	 */
16501 	if (!((ha->device_id == 0x2532) || (ha->device_id == 0x2432) ||
16502 	    (ha->device_id == 0x8432) || (ha->device_id == 0x8001) ||
16503 	    (ha->device_id == 0x8021))) {
16504 		EL(ha, "HBA does not support MSI-X\n");
16505 		return (DDI_FAILURE);
16506 	}
16507 
16508 	if (CFG_IST(ha, CFG_CTRL_2422) && (ha->rev_id < 3)) {
16509 		EL(ha, "HBA does not support MSI-X (revid)\n");
16510 		return (DDI_FAILURE);
16511 	}
16512 
16513 	/* Per HP, these HP branded HBA's are not supported with MSI-X */
16514 	if (ha->ven_id == 0x103C && (ha->subsys_id == 0x7041 ||
16515 	    ha->subsys_id == 0x7040 || ha->subsys_id == 0x1705)) {
16516 		EL(ha, "HBA does not support MSI-X (subdevid)\n");
16517 		return (DDI_FAILURE);
16518 	}
16519 
16520 	/* Get the number of 24xx/25xx MSI-X h/w vectors */
16521 	hwvect = (uint16_t)(((CFG_IST(ha, CFG_CTRL_2422) ?
16522 	    ql_pci_config_get16(ha, 0x7e) :
16523 	    ql_pci_config_get16(ha, 0xa2)) & 0x3ff) + 1);
16524 
16525 	EL(ha, "pcie config space hwvect = %d\n", hwvect);
16526 
16527 	if (hwvect < QL_MSIX_MAXAIF) {
16528 		EL(ha, "failed, min h/w vectors req'd: %d, avail: %d\n",
16529 		    QL_MSIX_MAXAIF, hwvect);
16530 		return (DDI_FAILURE);
16531 	}
16532 
16533 	/* Get number of MSI-X interrupts the platform h/w supports */
16534 	if (((ret = ddi_intr_get_nintrs(ha->dip, msitype, &count)) !=
16535 	    DDI_SUCCESS) || count == 0) {
16536 		EL(ha, "failed, nintrs ret=%xh, cnt=%xh\n", ret, count);
16537 		return (DDI_FAILURE);
16538 	}
16539 
16540 	/* Get number of available system interrupts */
16541 	if (((ret = ddi_intr_get_navail(ha->dip, msitype, &avail)) !=
16542 	    DDI_SUCCESS) || avail == 0) {
16543 		EL(ha, "failed, navail ret=%xh, avail=%xh\n", ret, avail);
16544 		return (DDI_FAILURE);
16545 	}
16546 
16547 	/* Fill out the intr table */
16548 	count = QL_MSIX_MAXAIF;
16549 	itrfun[QL_MSIX_AIF].ifunc = &ql_isr_aif;
16550 	itrfun[QL_MSIX_RSPQ].ifunc = &ql_isr_aif;
16551 
16552 	/* Allocate space for interrupt handles */
16553 	ha->hsize = ((uint32_t)(sizeof (ddi_intr_handle_t)) * hwvect);
16554 	if ((ha->htable = kmem_zalloc(ha->hsize, KM_SLEEP)) == NULL) {
16555 		ha->hsize = 0;
16556 		EL(ha, "failed, unable to allocate htable space\n");
16557 		return (DDI_FAILURE);
16558 	}
16559 
16560 	ha->iflags |= IFLG_INTR_MSIX;
16561 
16562 	/* Allocate the interrupts */
16563 	if (((ret = ddi_intr_alloc(ha->dip, ha->htable, msitype,
16564 	    DDI_INTR_ALLOC_NORMAL, count, &actual, 0)) != DDI_SUCCESS) ||
16565 	    actual < QL_MSIX_MAXAIF) {
16566 		EL(ha, "failed, intr_alloc ret=%xh, count = %xh, "
16567 		    "actual=%xh\n", ret, count, actual);
16568 		ql_release_intr(ha);
16569 		return (DDI_FAILURE);
16570 	}
16571 
16572 	ha->intr_cnt = actual;
16573 
16574 	/* Get interrupt priority */
16575 	if ((ret = ddi_intr_get_pri(ha->htable[0], &ha->intr_pri)) !=
16576 	    DDI_SUCCESS) {
16577 		EL(ha, "failed, get_pri ret=%xh\n", ret);
16578 		ql_release_intr(ha);
16579 		return (ret);
16580 	}
16581 
16582 	/* Add the interrupt handlers */
16583 	for (i = 0; i < actual; i++) {
16584 		if ((ret = ddi_intr_add_handler(ha->htable[i], itrfun[i].ifunc,
16585 		    (void *)ha, (void *)((ulong_t)i))) != DDI_SUCCESS) {
16586 			EL(ha, "failed, addh#=%xh, act=%xh, ret=%xh\n", i,
16587 			    actual, ret);
16588 			ql_release_intr(ha);
16589 			return (ret);
16590 		}
16591 	}
16592 
16593 	/*
16594 	 * duplicate the rest of the intr's
16595 	 * ddi_intr_dup_handler() isn't working on x86 just yet...
16596 	 */
16597 #ifdef __sparc
16598 	for (i = actual; i < hwvect; i++) {
16599 		if ((ret = ddi_intr_dup_handler(ha->htable[0], (int)i,
16600 		    &ha->htable[i])) != DDI_SUCCESS) {
16601 			EL(ha, "failed, intr_dup#=%xh, act=%xh, ret=%xh\n",
16602 			    i, actual, ret);
16603 			ql_release_intr(ha);
16604 			return (ret);
16605 		}
16606 	}
16607 #endif
16608 
16609 	/* Setup mutexes */
16610 	if ((ret = ql_init_mutex(ha)) != DDI_SUCCESS) {
16611 		EL(ha, "failed, mutex init ret=%xh\n", ret);
16612 		ql_release_intr(ha);
16613 		return (ret);
16614 	}
16615 
16616 	/* Get the capabilities */
16617 	(void) ddi_intr_get_cap(ha->htable[0], &ha->intr_cap);
16618 
16619 	/* Enable interrupts */
16620 	if (ha->intr_cap & DDI_INTR_FLAG_BLOCK) {
16621 		if ((ret = ddi_intr_block_enable(ha->htable, ha->intr_cnt)) !=
16622 		    DDI_SUCCESS) {
16623 			EL(ha, "failed, block enable, ret=%xh\n", ret);
16624 			ql_destroy_mutex(ha);
16625 			ql_release_intr(ha);
16626 			return (ret);
16627 		}
16628 	} else {
16629 		for (i = 0; i < ha->intr_cnt; i++) {
16630 			if ((ret = ddi_intr_enable(ha->htable[i])) !=
16631 			    DDI_SUCCESS) {
16632 				EL(ha, "failed, intr enable, ret=%xh\n", ret);
16633 				ql_destroy_mutex(ha);
16634 				ql_release_intr(ha);
16635 				return (ret);
16636 			}
16637 		}
16638 	}
16639 
16640 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
16641 
16642 	return (DDI_SUCCESS);
16643 }
16644 
16645 /*
16646  * ql_setup_fixed
16647  *	Sets up aif FIXED interrupts
16648  *
16649  * Input:
16650  *	ha = adapter state pointer.
16651  *
16652  * Returns:
16653  *	DDI_SUCCESS or DDI_FAILURE.
16654  *
16655  * Context:
16656  *	Kernel context.
16657  */
16658 static int
ql_setup_fixed(ql_adapter_state_t * ha)16659 ql_setup_fixed(ql_adapter_state_t *ha)
16660 {
16661 	int32_t		count = 0;
16662 	int32_t		actual = 0;
16663 	int32_t		ret;
16664 	uint32_t	i;
16665 
16666 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
16667 
16668 	/* Get number of fixed interrupts the system supports */
16669 	if (((ret = ddi_intr_get_nintrs(ha->dip, DDI_INTR_TYPE_FIXED,
16670 	    &count)) != DDI_SUCCESS) || count == 0) {
16671 		EL(ha, "failed, nintrs ret=%xh, cnt=%xh\n", ret, count);
16672 		return (DDI_FAILURE);
16673 	}
16674 
16675 	ha->iflags |= IFLG_INTR_FIXED;
16676 
16677 	/* Allocate space for interrupt handles */
16678 	ha->hsize = ((uint32_t)(sizeof (ddi_intr_handle_t)) * count);
16679 	ha->htable = kmem_zalloc(ha->hsize, KM_SLEEP);
16680 
16681 	/* Allocate the interrupts */
16682 	if (((ret = ddi_intr_alloc(ha->dip, ha->htable, DDI_INTR_TYPE_FIXED,
16683 	    0, count, &actual, DDI_INTR_ALLOC_STRICT)) != DDI_SUCCESS) ||
16684 	    actual < count) {
16685 		EL(ha, "failed, intr_alloc ret=%xh, count=%xh, "
16686 		    "actual=%xh\n", ret, count, actual);
16687 		ql_release_intr(ha);
16688 		return (DDI_FAILURE);
16689 	}
16690 
16691 	ha->intr_cnt = actual;
16692 
16693 	/* Get interrupt priority */
16694 	if ((ret = ddi_intr_get_pri(ha->htable[0], &ha->intr_pri)) !=
16695 	    DDI_SUCCESS) {
16696 		EL(ha, "failed, get_pri ret=%xh\n", ret);
16697 		ql_release_intr(ha);
16698 		return (ret);
16699 	}
16700 
16701 	/* Add the interrupt handlers */
16702 	for (i = 0; i < ha->intr_cnt; i++) {
16703 		if ((ret = ddi_intr_add_handler(ha->htable[i], &ql_isr_aif,
16704 		    (void *)ha, (void *)((ulong_t)(i)))) != DDI_SUCCESS) {
16705 			EL(ha, "failed, intr_add ret=%xh\n", ret);
16706 			ql_release_intr(ha);
16707 			return (ret);
16708 		}
16709 	}
16710 
16711 	/* Setup mutexes */
16712 	if ((ret = ql_init_mutex(ha)) != DDI_SUCCESS) {
16713 		EL(ha, "failed, mutex init ret=%xh\n", ret);
16714 		ql_release_intr(ha);
16715 		return (ret);
16716 	}
16717 
16718 	/* Enable interrupts */
16719 	for (i = 0; i < ha->intr_cnt; i++) {
16720 		if ((ret = ddi_intr_enable(ha->htable[i])) != DDI_SUCCESS) {
16721 			EL(ha, "failed, intr enable, ret=%xh\n", ret);
16722 			ql_destroy_mutex(ha);
16723 			ql_release_intr(ha);
16724 			return (ret);
16725 		}
16726 	}
16727 
16728 	EL(ha, "using FIXED interupts\n");
16729 
16730 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
16731 
16732 	return (DDI_SUCCESS);
16733 }
16734 
16735 /*
16736  * ql_disable_intr
16737  *	Disables interrupts
16738  *
16739  * Input:
16740  *	ha = adapter state pointer.
16741  *
16742  * Returns:
16743  *
16744  * Context:
16745  *	Kernel context.
16746  */
16747 static void
ql_disable_intr(ql_adapter_state_t * ha)16748 ql_disable_intr(ql_adapter_state_t *ha)
16749 {
16750 	uint32_t	i, rval;
16751 
16752 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
16753 
16754 	if (!(ha->iflags & IFLG_INTR_AIF)) {
16755 
16756 		/* Disable legacy interrupts */
16757 		(void) ddi_remove_intr(ha->dip, 0, ha->iblock_cookie);
16758 
16759 	} else if ((ha->intr_cap & DDI_INTR_FLAG_BLOCK) &&
16760 	    (ha->iflags & (IFLG_INTR_MSI | IFLG_INTR_MSIX))) {
16761 
16762 		/* Remove AIF block interrupts (MSI) */
16763 		if ((rval = ddi_intr_block_disable(ha->htable, ha->intr_cnt))
16764 		    != DDI_SUCCESS) {
16765 			EL(ha, "failed intr block disable, rval=%x\n", rval);
16766 		}
16767 
16768 	} else {
16769 
16770 		/* Remove AIF non-block interrupts (fixed).  */
16771 		for (i = 0; i < ha->intr_cnt; i++) {
16772 			if ((rval = ddi_intr_disable(ha->htable[i])) !=
16773 			    DDI_SUCCESS) {
16774 				EL(ha, "failed intr disable, intr#=%xh, "
16775 				    "rval=%xh\n", i, rval);
16776 			}
16777 		}
16778 	}
16779 
16780 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
16781 }
16782 
16783 /*
16784  * ql_release_intr
16785  *	Releases aif legacy interrupt resources
16786  *
16787  * Input:
16788  *	ha = adapter state pointer.
16789  *
16790  * Returns:
16791  *
16792  * Context:
16793  *	Kernel context.
16794  */
16795 static void
ql_release_intr(ql_adapter_state_t * ha)16796 ql_release_intr(ql_adapter_state_t *ha)
16797 {
16798 	int32_t 	i;
16799 
16800 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
16801 
16802 	if (!(ha->iflags & IFLG_INTR_AIF)) {
16803 		QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
16804 		return;
16805 	}
16806 
16807 	ha->iflags &= ~(IFLG_INTR_AIF);
16808 	if (ha->htable != NULL && ha->hsize > 0) {
16809 		i = (int32_t)ha->hsize / (int32_t)sizeof (ddi_intr_handle_t);
16810 		while (i-- > 0) {
16811 			if (ha->htable[i] == 0) {
16812 				EL(ha, "htable[%x]=0h\n", i);
16813 				continue;
16814 			}
16815 
16816 			(void) ddi_intr_disable(ha->htable[i]);
16817 
16818 			if (i < ha->intr_cnt) {
16819 				(void) ddi_intr_remove_handler(ha->htable[i]);
16820 			}
16821 
16822 			(void) ddi_intr_free(ha->htable[i]);
16823 		}
16824 
16825 		kmem_free(ha->htable, ha->hsize);
16826 		ha->htable = NULL;
16827 	}
16828 
16829 	ha->hsize = 0;
16830 	ha->intr_cnt = 0;
16831 	ha->intr_pri = 0;
16832 	ha->intr_cap = 0;
16833 
16834 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
16835 }
16836 
16837 /*
16838  * ql_legacy_intr
16839  *	Sets up legacy interrupts.
16840  *
16841  *	NB: Only to be used if AIF (Advanced Interupt Framework)
16842  *	    if NOT in the kernel.
16843  *
16844  * Input:
16845  *	ha = adapter state pointer.
16846  *
16847  * Returns:
16848  *	DDI_SUCCESS or DDI_FAILURE.
16849  *
16850  * Context:
16851  *	Kernel context.
16852  */
16853 static int
ql_legacy_intr(ql_adapter_state_t * ha)16854 ql_legacy_intr(ql_adapter_state_t *ha)
16855 {
16856 	int	rval = DDI_SUCCESS;
16857 
16858 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
16859 
16860 	/* Setup mutexes */
16861 	if (ql_init_mutex(ha) != DDI_SUCCESS) {
16862 		EL(ha, "failed, mutex init\n");
16863 		return (DDI_FAILURE);
16864 	}
16865 
16866 	/* Setup standard/legacy interrupt handler */
16867 	if (ddi_add_intr(ha->dip, (uint_t)0, &ha->iblock_cookie,
16868 	    (ddi_idevice_cookie_t *)0, ql_isr, (caddr_t)ha) != DDI_SUCCESS) {
16869 		cmn_err(CE_WARN, "%s(%d): Failed to add legacy interrupt",
16870 		    QL_NAME, ha->instance);
16871 		ql_destroy_mutex(ha);
16872 		rval = DDI_FAILURE;
16873 	}
16874 
16875 	if (rval == DDI_SUCCESS) {
16876 		ha->iflags |= IFLG_INTR_LEGACY;
16877 		EL(ha, "using legacy interrupts\n");
16878 	}
16879 
16880 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
16881 
16882 	return (rval);
16883 }
16884 
16885 /*
16886  * ql_init_mutex
16887  *	Initializes mutex's
16888  *
16889  * Input:
16890  *	ha = adapter state pointer.
16891  *
16892  * Returns:
16893  *	DDI_SUCCESS or DDI_FAILURE.
16894  *
16895  * Context:
16896  *	Kernel context.
16897  */
16898 static int
ql_init_mutex(ql_adapter_state_t * ha)16899 ql_init_mutex(ql_adapter_state_t *ha)
16900 {
16901 	int	ret;
16902 	void	*intr;
16903 
16904 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
16905 
16906 	if (ha->iflags & IFLG_INTR_AIF) {
16907 		intr = (void *)(uintptr_t)ha->intr_pri;
16908 	} else {
16909 		/* Get iblock cookies to initialize mutexes */
16910 		if ((ret = ddi_get_iblock_cookie(ha->dip, 0,
16911 		    &ha->iblock_cookie)) != DDI_SUCCESS) {
16912 			EL(ha, "failed, get_iblock: %xh\n", ret);
16913 			return (DDI_FAILURE);
16914 		}
16915 		intr = (void *)ha->iblock_cookie;
16916 	}
16917 
16918 	/* mutexes to protect the adapter state structure. */
16919 	mutex_init(&ha->mutex, NULL, MUTEX_DRIVER, intr);
16920 
16921 	/* mutex to protect the ISP response ring. */
16922 	mutex_init(&ha->intr_mutex, NULL, MUTEX_DRIVER, intr);
16923 
16924 	/* mutex to protect the mailbox registers. */
16925 	mutex_init(&ha->mbx_mutex, NULL, MUTEX_DRIVER, intr);
16926 
16927 	/* power management protection */
16928 	mutex_init(&ha->pm_mutex, NULL, MUTEX_DRIVER, intr);
16929 
16930 	/* Mailbox wait and interrupt conditional variable. */
16931 	cv_init(&ha->cv_mbx_wait, NULL, CV_DRIVER, NULL);
16932 	cv_init(&ha->cv_mbx_intr, NULL, CV_DRIVER, NULL);
16933 
16934 	/* mutex to protect the ISP request ring. */
16935 	mutex_init(&ha->req_ring_mutex, NULL, MUTEX_DRIVER, intr);
16936 
16937 	/* Unsolicited buffer conditional variable. */
16938 	cv_init(&ha->cv_ub, NULL, CV_DRIVER, NULL);
16939 
16940 	mutex_init(&ha->ub_mutex, NULL, MUTEX_DRIVER, intr);
16941 	mutex_init(&ha->cache_mutex, NULL, MUTEX_DRIVER, intr);
16942 
16943 	/* Suspended conditional variable. */
16944 	cv_init(&ha->cv_dr_suspended, NULL, CV_DRIVER, NULL);
16945 
16946 	/* mutex to protect task daemon context. */
16947 	mutex_init(&ha->task_daemon_mutex, NULL, MUTEX_DRIVER, intr);
16948 
16949 	/* Task_daemon thread conditional variable. */
16950 	cv_init(&ha->cv_task_daemon, NULL, CV_DRIVER, NULL);
16951 
16952 	/* mutex to protect diag port manage interface */
16953 	mutex_init(&ha->portmutex, NULL, MUTEX_DRIVER, intr);
16954 
16955 	/* mutex to protect per instance f/w dump flags and buffer */
16956 	mutex_init(&ha->dump_mutex, NULL, MUTEX_DRIVER, intr);
16957 
16958 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
16959 
16960 	return (DDI_SUCCESS);
16961 }
16962 
16963 /*
16964  * ql_destroy_mutex
16965  *	Destroys mutex's
16966  *
16967  * Input:
16968  *	ha = adapter state pointer.
16969  *
16970  * Returns:
16971  *
16972  * Context:
16973  *	Kernel context.
16974  */
16975 static void
ql_destroy_mutex(ql_adapter_state_t * ha)16976 ql_destroy_mutex(ql_adapter_state_t *ha)
16977 {
16978 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
16979 
16980 	mutex_destroy(&ha->dump_mutex);
16981 	mutex_destroy(&ha->portmutex);
16982 	cv_destroy(&ha->cv_task_daemon);
16983 	mutex_destroy(&ha->task_daemon_mutex);
16984 	cv_destroy(&ha->cv_dr_suspended);
16985 	mutex_destroy(&ha->cache_mutex);
16986 	mutex_destroy(&ha->ub_mutex);
16987 	cv_destroy(&ha->cv_ub);
16988 	mutex_destroy(&ha->req_ring_mutex);
16989 	cv_destroy(&ha->cv_mbx_intr);
16990 	cv_destroy(&ha->cv_mbx_wait);
16991 	mutex_destroy(&ha->pm_mutex);
16992 	mutex_destroy(&ha->mbx_mutex);
16993 	mutex_destroy(&ha->intr_mutex);
16994 	mutex_destroy(&ha->mutex);
16995 
16996 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
16997 }
16998 
16999 /*
17000  * ql_fwmodule_resolve
17001  *	Loads and resolves external firmware module and symbols
17002  *
17003  * Input:
17004  *	ha:		adapter state pointer.
17005  *
17006  * Returns:
17007  *	ql local function return status code:
17008  *		QL_SUCCESS - external f/w module module and symbols resolved
17009  *		QL_FW_NOT_SUPPORTED - Driver does not support ISP type
17010  *		QL_FWMODLOAD_FAILED - Could not load f/w module (ddi failed)
17011  *		QL_FWSYM_NOT_FOUND - Unable to resolve internal f/w symbol
17012  * Context:
17013  *	Kernel context.
17014  *
17015  * NOTE: We currently ddi_modopen/ddi_modclose at attach/detach time.  We
17016  * could switch to a tighter scope around acutal download (and add an extra
17017  * ddi_modopen for module opens that occur before root is mounted).
17018  *
17019  */
17020 uint32_t
ql_fwmodule_resolve(ql_adapter_state_t * ha)17021 ql_fwmodule_resolve(ql_adapter_state_t *ha)
17022 {
17023 	int8_t			module[128];
17024 	int8_t			fw_version[128];
17025 	uint32_t		rval = QL_SUCCESS;
17026 	caddr_t			code, code02;
17027 	uint8_t			*p_ucfw;
17028 	uint16_t		*p_usaddr, *p_uslen;
17029 	uint32_t		*p_uiaddr, *p_uilen, *p_uifw;
17030 	uint32_t		*p_uiaddr02, *p_uilen02;
17031 	struct fw_table		*fwt;
17032 	extern struct fw_table	fw_table[];
17033 
17034 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
17035 
17036 	if (ha->fw_module != NULL) {
17037 		EL(ha, "%x f/w module %d.%02d.%02d is already loaded\n",
17038 		    ha->fw_class, ha->fw_major_version, ha->fw_minor_version,
17039 		    ha->fw_subminor_version);
17040 		return (rval);
17041 	}
17042 
17043 	/* make sure the fw_class is in the fw_table of supported classes */
17044 	for (fwt = &fw_table[0]; fwt->fw_version; fwt++) {
17045 		if (fwt->fw_class == ha->fw_class)
17046 			break;			/* match */
17047 	}
17048 	if (fwt->fw_version == NULL) {
17049 		cmn_err(CE_WARN, "%s(%d): can't find f/w class %x "
17050 		    "in driver's fw_table", QL_NAME, ha->instance,
17051 		    ha->fw_class);
17052 		return (QL_FW_NOT_SUPPORTED);
17053 	}
17054 
17055 	/*
17056 	 * open the module related to the fw_class
17057 	 */
17058 	(void) snprintf(module, sizeof (module), "misc/qlc/qlc_fw_%x",
17059 	    ha->fw_class);
17060 
17061 	ha->fw_module = ddi_modopen(module, KRTLD_MODE_FIRST, NULL);
17062 	if (ha->fw_module == NULL) {
17063 		cmn_err(CE_WARN, "%s(%d): can't load firmware file %s",
17064 		    QL_NAME, ha->instance, module);
17065 		return (QL_FWMODLOAD_FAILED);
17066 	}
17067 
17068 	/*
17069 	 * resolve the fw module symbols, data types depend on fw_class
17070 	 */
17071 
17072 	switch (ha->fw_class) {
17073 	case 0x2200:
17074 	case 0x2300:
17075 	case 0x6322:
17076 
17077 		if ((code = ddi_modsym(ha->fw_module, "risc_code01",
17078 		    NULL)) == NULL) {
17079 			rval = QL_FWSYM_NOT_FOUND;
17080 			EL(ha, "failed, f/w module %d rc01 symbol\n", module);
17081 		} else if ((p_usaddr = ddi_modsym(ha->fw_module,
17082 		    "risc_code_addr01", NULL)) == NULL) {
17083 			rval = QL_FWSYM_NOT_FOUND;
17084 			EL(ha, "failed, f/w module %d rca01 symbol\n", module);
17085 		} else if ((p_uslen = ddi_modsym(ha->fw_module,
17086 		    "risc_code_length01", NULL)) == NULL) {
17087 			rval = QL_FWSYM_NOT_FOUND;
17088 			EL(ha, "failed, f/w module %d rcl01 symbol\n", module);
17089 		} else if ((p_ucfw = ddi_modsym(ha->fw_module,
17090 		    "firmware_version", NULL)) == NULL) {
17091 			rval = QL_FWSYM_NOT_FOUND;
17092 			EL(ha, "failed, f/w module %d fwver symbol\n", module);
17093 		}
17094 
17095 		if (rval == QL_SUCCESS) {
17096 			ha->risc_fw[0].code = code;
17097 			ha->risc_fw[0].addr = *p_usaddr;
17098 			ha->risc_fw[0].length = *p_uslen;
17099 
17100 			(void) snprintf(fw_version, sizeof (fw_version),
17101 			    "%d.%02d.%02d", p_ucfw[0], p_ucfw[1], p_ucfw[2]);
17102 		}
17103 		break;
17104 
17105 	case 0x2400:
17106 	case 0x2500:
17107 	case 0x8100:
17108 
17109 		if ((code = ddi_modsym(ha->fw_module, "risc_code01",
17110 		    NULL)) == NULL) {
17111 			rval = QL_FWSYM_NOT_FOUND;
17112 			EL(ha, "failed, f/w module %d rc01 symbol\n", module);
17113 		} else if ((p_uiaddr = ddi_modsym(ha->fw_module,
17114 		    "risc_code_addr01", NULL)) == NULL) {
17115 			rval = QL_FWSYM_NOT_FOUND;
17116 			EL(ha, "failed, f/w module %d rca01 symbol\n", module);
17117 		} else if ((p_uilen = ddi_modsym(ha->fw_module,
17118 		    "risc_code_length01", NULL)) == NULL) {
17119 			rval = QL_FWSYM_NOT_FOUND;
17120 			EL(ha, "failed, f/w module %d rcl01 symbol\n", module);
17121 		} else if ((p_uifw = ddi_modsym(ha->fw_module,
17122 		    "firmware_version", NULL)) == NULL) {
17123 			rval = QL_FWSYM_NOT_FOUND;
17124 			EL(ha, "failed, f/w module %d fwver symbol\n", module);
17125 		}
17126 
17127 		if ((code02 = ddi_modsym(ha->fw_module, "risc_code02",
17128 		    NULL)) == NULL) {
17129 			rval = QL_FWSYM_NOT_FOUND;
17130 			EL(ha, "failed, f/w module %d rc02 symbol\n", module);
17131 		} else if ((p_uiaddr02 = ddi_modsym(ha->fw_module,
17132 		    "risc_code_addr02", NULL)) == NULL) {
17133 			rval = QL_FWSYM_NOT_FOUND;
17134 			EL(ha, "failed, f/w module %d rca02 symbol\n", module);
17135 		} else if ((p_uilen02 = ddi_modsym(ha->fw_module,
17136 		    "risc_code_length02", NULL)) == NULL) {
17137 			rval = QL_FWSYM_NOT_FOUND;
17138 			EL(ha, "failed, f/w module %d rcl02 symbol\n", module);
17139 		}
17140 
17141 		if (rval == QL_SUCCESS) {
17142 			ha->risc_fw[0].code = code;
17143 			ha->risc_fw[0].addr = *p_uiaddr;
17144 			ha->risc_fw[0].length = *p_uilen;
17145 			ha->risc_fw[1].code = code02;
17146 			ha->risc_fw[1].addr = *p_uiaddr02;
17147 			ha->risc_fw[1].length = *p_uilen02;
17148 
17149 			(void) snprintf(fw_version, sizeof (fw_version),
17150 			    "%d.%02d.%02d", p_uifw[0], p_uifw[1], p_uifw[2]);
17151 		}
17152 		break;
17153 
17154 	default:
17155 		EL(ha, "fw_class: '%x' is not supported\n", ha->fw_class);
17156 		rval = QL_FW_NOT_SUPPORTED;
17157 	}
17158 
17159 	if (rval != QL_SUCCESS) {
17160 		cmn_err(CE_WARN, "%s(%d): can't resolve firmware "
17161 		    "module %s (%x)", QL_NAME, ha->instance, module, rval);
17162 		if (ha->fw_module != NULL) {
17163 			(void) ddi_modclose(ha->fw_module);
17164 			ha->fw_module = NULL;
17165 		}
17166 	} else {
17167 		/*
17168 		 * check for firmware version mismatch between module and
17169 		 * compiled in fw_table version.
17170 		 */
17171 
17172 		if (strcmp(fwt->fw_version, fw_version) != 0) {
17173 
17174 			/*
17175 			 * If f/w / driver version mismatches then
17176 			 * return a successful status -- however warn
17177 			 * the user that this is NOT recommended.
17178 			 */
17179 
17180 			cmn_err(CE_WARN, "%s(%d): driver / f/w version "
17181 			    "mismatch for %x: driver-%s module-%s", QL_NAME,
17182 			    ha->instance, ha->fw_class, fwt->fw_version,
17183 			    fw_version);
17184 
17185 			ha->cfg_flags |= CFG_FW_MISMATCH;
17186 		} else {
17187 			ha->cfg_flags &= ~CFG_FW_MISMATCH;
17188 		}
17189 	}
17190 
17191 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
17192 
17193 	return (rval);
17194 }
17195 
17196 /*
17197  * ql_port_state
17198  *	Set the state on all adapter ports.
17199  *
17200  * Input:
17201  *	ha:	parent adapter state pointer.
17202  *	state:	port state.
17203  *	flags:	task daemon flags to set.
17204  *
17205  * Context:
17206  *	Interrupt or Kernel context, no mailbox commands allowed.
17207  */
17208 void
ql_port_state(ql_adapter_state_t * ha,uint32_t state,uint32_t flags)17209 ql_port_state(ql_adapter_state_t *ha, uint32_t state, uint32_t flags)
17210 {
17211 	ql_adapter_state_t	*vha;
17212 
17213 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
17214 
17215 	TASK_DAEMON_LOCK(ha);
17216 	for (vha = ha->pha; vha != NULL; vha = vha->vp_next) {
17217 		if (FC_PORT_STATE_MASK(vha->state) != state) {
17218 			vha->state = state != FC_STATE_OFFLINE ?
17219 			    (FC_PORT_SPEED_MASK(vha->state) | state) : state;
17220 			vha->task_daemon_flags |= flags;
17221 		}
17222 	}
17223 	ha->pha->task_daemon_flags |= flags & LOOP_DOWN;
17224 	TASK_DAEMON_UNLOCK(ha);
17225 
17226 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
17227 }
17228 
17229 /*
17230  * ql_el_trace_desc_ctor - Construct an extended logging trace descriptor.
17231  *
17232  * Input:	Pointer to the adapter state structure.
17233  * Returns:	Success or Failure.
17234  * Context:	Kernel context.
17235  */
17236 int
ql_el_trace_desc_ctor(ql_adapter_state_t * ha)17237 ql_el_trace_desc_ctor(ql_adapter_state_t *ha)
17238 {
17239 	int	rval = DDI_SUCCESS;
17240 
17241 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
17242 
17243 	ha->el_trace_desc =
17244 	    (el_trace_desc_t *)kmem_zalloc(sizeof (el_trace_desc_t), KM_SLEEP);
17245 
17246 	if (ha->el_trace_desc == NULL) {
17247 		cmn_err(CE_WARN, "%s(%d): can't construct trace descriptor",
17248 		    QL_NAME, ha->instance);
17249 		rval = DDI_FAILURE;
17250 	} else {
17251 		ha->el_trace_desc->next		= 0;
17252 		ha->el_trace_desc->trace_buffer =
17253 		    (char *)kmem_zalloc(EL_TRACE_BUF_SIZE, KM_SLEEP);
17254 
17255 		if (ha->el_trace_desc->trace_buffer == NULL) {
17256 			cmn_err(CE_WARN, "%s(%d): can't get trace buffer",
17257 			    QL_NAME, ha->instance);
17258 			kmem_free(ha->el_trace_desc, sizeof (el_trace_desc_t));
17259 			rval = DDI_FAILURE;
17260 		} else {
17261 			ha->el_trace_desc->trace_buffer_size =
17262 			    EL_TRACE_BUF_SIZE;
17263 			mutex_init(&ha->el_trace_desc->mutex, NULL,
17264 			    MUTEX_DRIVER, NULL);
17265 		}
17266 	}
17267 
17268 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
17269 
17270 	return (rval);
17271 }
17272 
17273 /*
17274  * ql_el_trace_desc_dtor - Destroy an extended logging trace descriptor.
17275  *
17276  * Input:	Pointer to the adapter state structure.
17277  * Returns:	Success or Failure.
17278  * Context:	Kernel context.
17279  */
17280 int
ql_el_trace_desc_dtor(ql_adapter_state_t * ha)17281 ql_el_trace_desc_dtor(ql_adapter_state_t *ha)
17282 {
17283 	int	rval = DDI_SUCCESS;
17284 
17285 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
17286 
17287 	if (ha->el_trace_desc == NULL) {
17288 		cmn_err(CE_WARN, "%s(%d): can't destroy el trace descriptor",
17289 		    QL_NAME, ha->instance);
17290 		rval = DDI_FAILURE;
17291 	} else {
17292 		if (ha->el_trace_desc->trace_buffer != NULL) {
17293 			kmem_free(ha->el_trace_desc->trace_buffer,
17294 			    ha->el_trace_desc->trace_buffer_size);
17295 		}
17296 		mutex_destroy(&ha->el_trace_desc->mutex);
17297 		kmem_free(ha->el_trace_desc, sizeof (el_trace_desc_t));
17298 	}
17299 
17300 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
17301 
17302 	return (rval);
17303 }
17304 
17305 /*
17306  * els_cmd_text	- Return a pointer to a string describing the command
17307  *
17308  * Input:	els_cmd = the els command opcode.
17309  * Returns:	pointer to a string.
17310  * Context:	Kernel context.
17311  */
17312 char *
els_cmd_text(int els_cmd)17313 els_cmd_text(int els_cmd)
17314 {
17315 	cmd_table_t *entry = &els_cmd_tbl[0];
17316 
17317 	return (cmd_text(entry, els_cmd));
17318 }
17319 
17320 /*
17321  * mbx_cmd_text - Return a pointer to a string describing the command
17322  *
17323  * Input:	mbx_cmd = the mailbox command opcode.
17324  * Returns:	pointer to a string.
17325  * Context:	Kernel context.
17326  */
17327 char *
mbx_cmd_text(int mbx_cmd)17328 mbx_cmd_text(int mbx_cmd)
17329 {
17330 	cmd_table_t *entry = &mbox_cmd_tbl[0];
17331 
17332 	return (cmd_text(entry, mbx_cmd));
17333 }
17334 
17335 /*
17336  * cmd_text	Return a pointer to a string describing the command
17337  *
17338  * Input:	entry = the command table
17339  *		cmd = the command.
17340  * Returns:	pointer to a string.
17341  * Context:	Kernel context.
17342  */
17343 char *
cmd_text(cmd_table_t * entry,int cmd)17344 cmd_text(cmd_table_t *entry, int cmd)
17345 {
17346 	for (; entry->cmd != 0; entry++) {
17347 		if (entry->cmd == cmd) {
17348 			break;
17349 		}
17350 	}
17351 	return (entry->string);
17352 }
17353 
17354 /*
17355  * ql_els_24xx_mbox_cmd_iocb - els request indication.
17356  *
17357  * Input:	ha = adapter state pointer.
17358  *		srb = scsi request block pointer.
17359  *		arg = els passthru entry iocb pointer.
17360  * Returns:
17361  * Context:	Kernel context.
17362  */
17363 void
ql_els_24xx_iocb(ql_adapter_state_t * ha,ql_srb_t * srb,void * arg)17364 ql_els_24xx_iocb(ql_adapter_state_t *ha, ql_srb_t *srb, void *arg)
17365 {
17366 	els_descriptor_t	els_desc;
17367 
17368 	/* Extract the ELS information */
17369 	ql_fca_isp_els_request(ha, (fc_packet_t *)srb->pkt, &els_desc);
17370 
17371 	/* Construct the passthru entry */
17372 	ql_isp_els_request_ctor(&els_desc, (els_passthru_entry_t *)arg);
17373 
17374 	/* Ensure correct endianness */
17375 	ql_isp_els_handle_cmd_endian(ha, srb);
17376 }
17377 
17378 /*
17379  * ql_fca_isp_els_request - Extract into an els descriptor the info required
17380  *			    to build an els_passthru iocb from an fc packet.
17381  *
17382  * Input:	ha = adapter state pointer.
17383  *		pkt = fc packet pointer
17384  *		els_desc = els descriptor pointer
17385  * Returns:
17386  * Context:	Kernel context.
17387  */
17388 static void
ql_fca_isp_els_request(ql_adapter_state_t * ha,fc_packet_t * pkt,els_descriptor_t * els_desc)17389 ql_fca_isp_els_request(ql_adapter_state_t *ha, fc_packet_t *pkt,
17390     els_descriptor_t *els_desc)
17391 {
17392 	ls_code_t	els;
17393 
17394 	ddi_rep_get8(pkt->pkt_cmd_acc, (uint8_t *)&els,
17395 	    (uint8_t *)pkt->pkt_cmd, sizeof (els), DDI_DEV_AUTOINCR);
17396 
17397 	els_desc->els = els.ls_code;
17398 
17399 	els_desc->els_handle = ha->hba_buf.acc_handle;
17400 	els_desc->d_id.b24 = pkt->pkt_cmd_fhdr.d_id;
17401 	els_desc->s_id.b24 = pkt->pkt_cmd_fhdr.s_id;
17402 	/* if n_port_handle is not < 0x7d use 0 */
17403 	if (LOCAL_LOOP_ID(ha->n_port->n_port_handle)) {
17404 		els_desc->n_port_handle = ha->n_port->n_port_handle;
17405 	} else {
17406 		els_desc->n_port_handle = 0;
17407 	}
17408 	els_desc->control_flags = 0;
17409 	els_desc->cmd_byte_count = pkt->pkt_cmdlen;
17410 	/*
17411 	 * Transmit DSD. This field defines the Fibre Channel Frame payload
17412 	 * (without the frame header) in system memory.
17413 	 */
17414 	els_desc->tx_dsd.addr[0] = LSD(pkt->pkt_cmd_cookie->dmac_laddress);
17415 	els_desc->tx_dsd.addr[1] = MSD(pkt->pkt_cmd_cookie->dmac_laddress);
17416 	els_desc->tx_dsd.length = (uint32_t)pkt->pkt_cmd_cookie->dmac_size;
17417 
17418 	els_desc->rsp_byte_count = pkt->pkt_rsplen;
17419 	/*
17420 	 * Receive DSD. This field defines the ELS response payload buffer
17421 	 * for the ISP24xx firmware transferring the received ELS
17422 	 * response frame to a location in host memory.
17423 	 */
17424 	els_desc->rx_dsd.addr[0] = LSD(pkt->pkt_resp_cookie->dmac_laddress);
17425 	els_desc->rx_dsd.addr[1] = MSD(pkt->pkt_resp_cookie->dmac_laddress);
17426 	els_desc->rx_dsd.length = (uint32_t)pkt->pkt_resp_cookie->dmac_size;
17427 }
17428 
17429 /*
17430  * ql_isp_els_request_ctor - Construct an els_passthru_entry iocb
17431  * using the els descriptor.
17432  *
17433  * Input:	ha = adapter state pointer.
17434  *		els_desc = els descriptor pointer.
17435  *		els_entry = els passthru entry iocb pointer.
17436  * Returns:
17437  * Context:	Kernel context.
17438  */
17439 static void
ql_isp_els_request_ctor(els_descriptor_t * els_desc,els_passthru_entry_t * els_entry)17440 ql_isp_els_request_ctor(els_descriptor_t *els_desc,
17441     els_passthru_entry_t *els_entry)
17442 {
17443 	uint32_t	*ptr32;
17444 
17445 	/*
17446 	 * Construct command packet.
17447 	 */
17448 	ddi_put8(els_desc->els_handle, &els_entry->entry_type,
17449 	    (uint8_t)ELS_PASSTHRU_TYPE);
17450 	ddi_put16(els_desc->els_handle, &els_entry->n_port_hdl,
17451 	    els_desc->n_port_handle);
17452 	ddi_put8(els_desc->els_handle, &els_entry->sof_type, (uint8_t)BIT_4);
17453 	ddi_put32(els_desc->els_handle, &els_entry->rcv_exch_address,
17454 	    (uint32_t)0);
17455 	ddi_put8(els_desc->els_handle, &els_entry->els_cmd_opcode,
17456 	    els_desc->els);
17457 	ddi_put8(els_desc->els_handle, &els_entry->d_id_7_0,
17458 	    els_desc->d_id.b.al_pa);
17459 	ddi_put8(els_desc->els_handle, &els_entry->d_id_15_8,
17460 	    els_desc->d_id.b.area);
17461 	ddi_put8(els_desc->els_handle, &els_entry->d_id_23_16,
17462 	    els_desc->d_id.b.domain);
17463 	ddi_put8(els_desc->els_handle, &els_entry->s_id_7_0,
17464 	    els_desc->s_id.b.al_pa);
17465 	ddi_put8(els_desc->els_handle, &els_entry->s_id_15_8,
17466 	    els_desc->s_id.b.area);
17467 	ddi_put8(els_desc->els_handle, &els_entry->s_id_23_16,
17468 	    els_desc->s_id.b.domain);
17469 	ddi_put16(els_desc->els_handle, &els_entry->control_flags,
17470 	    els_desc->control_flags);
17471 	ddi_put32(els_desc->els_handle, &els_entry->rcv_payld_data_bcnt,
17472 	    els_desc->rsp_byte_count);
17473 	ddi_put32(els_desc->els_handle, &els_entry->xmt_payld_data_bcnt,
17474 	    els_desc->cmd_byte_count);
17475 	/* Load transmit data segments and count. */
17476 	ptr32 = (uint32_t *)&els_entry->xmt_dseg_0_address;
17477 	ddi_put16(els_desc->els_handle, &els_entry->xmt_dseg_count, 1);
17478 	ddi_put32(els_desc->els_handle, ptr32++, els_desc->tx_dsd.addr[0]);
17479 	ddi_put32(els_desc->els_handle, ptr32++, els_desc->tx_dsd.addr[1]);
17480 	ddi_put32(els_desc->els_handle, ptr32++, els_desc->tx_dsd.length);
17481 	ddi_put16(els_desc->els_handle, &els_entry->rcv_dseg_count, 1);
17482 	ddi_put32(els_desc->els_handle, ptr32++, els_desc->rx_dsd.addr[0]);
17483 	ddi_put32(els_desc->els_handle, ptr32++, els_desc->rx_dsd.addr[1]);
17484 	ddi_put32(els_desc->els_handle, ptr32++, els_desc->rx_dsd.length);
17485 }
17486 
17487 /*
17488  * ql_isp_els_handle_cmd_endian - els requests must be in big endian
17489  *				  in host memory.
17490  *
17491  * Input:	ha = adapter state pointer.
17492  *		srb = scsi request block
17493  * Returns:
17494  * Context:	Kernel context.
17495  */
17496 void
ql_isp_els_handle_cmd_endian(ql_adapter_state_t * ha,ql_srb_t * srb)17497 ql_isp_els_handle_cmd_endian(ql_adapter_state_t *ha, ql_srb_t *srb)
17498 {
17499 	ls_code_t	els;
17500 	fc_packet_t	*pkt;
17501 	uint8_t		*ptr;
17502 
17503 	pkt = srb->pkt;
17504 
17505 	ddi_rep_get8(pkt->pkt_cmd_acc, (uint8_t *)&els,
17506 	    (uint8_t *)pkt->pkt_cmd, sizeof (els), DDI_DEV_AUTOINCR);
17507 
17508 	ptr = (uint8_t *)pkt->pkt_cmd;
17509 
17510 	ql_isp_els_handle_endian(ha, ptr, els.ls_code);
17511 }
17512 
17513 /*
17514  * ql_isp_els_handle_rsp_endian - els responses must be in big endian
17515  *				  in host memory.
17516  * Input:	ha = adapter state pointer.
17517  *		srb = scsi request block
17518  * Returns:
17519  * Context:	Kernel context.
17520  */
17521 void
ql_isp_els_handle_rsp_endian(ql_adapter_state_t * ha,ql_srb_t * srb)17522 ql_isp_els_handle_rsp_endian(ql_adapter_state_t *ha, ql_srb_t *srb)
17523 {
17524 	ls_code_t	els;
17525 	fc_packet_t	*pkt;
17526 	uint8_t		*ptr;
17527 
17528 	pkt = srb->pkt;
17529 
17530 	ddi_rep_get8(pkt->pkt_cmd_acc, (uint8_t *)&els,
17531 	    (uint8_t *)pkt->pkt_cmd, sizeof (els), DDI_DEV_AUTOINCR);
17532 
17533 	ptr = (uint8_t *)pkt->pkt_resp;
17534 	BIG_ENDIAN_32(&els);
17535 	ql_isp_els_handle_endian(ha, ptr, els.ls_code);
17536 }
17537 
17538 /*
17539  * ql_isp_els_handle_endian - els requests/responses must be in big endian
17540  *			      in host memory.
17541  * Input:	ha = adapter state pointer.
17542  *		ptr = els request/response buffer pointer.
17543  *		ls_code = els command code.
17544  * Returns:
17545  * Context:	Kernel context.
17546  */
17547 void
ql_isp_els_handle_endian(ql_adapter_state_t * ha,uint8_t * ptr,uint8_t ls_code)17548 ql_isp_els_handle_endian(ql_adapter_state_t *ha, uint8_t *ptr, uint8_t ls_code)
17549 {
17550 	switch (ls_code) {
17551 	case LA_ELS_PLOGI: {
17552 		BIG_ENDIAN_32(ptr);	/* Command Code */
17553 		ptr += 4;
17554 		BIG_ENDIAN_16(ptr);	/* FC-PH version */
17555 		ptr += 2;
17556 		BIG_ENDIAN_16(ptr);	/* b2b credit */
17557 		ptr += 2;
17558 		BIG_ENDIAN_16(ptr);	/* Cmn Feature flags */
17559 		ptr += 2;
17560 		BIG_ENDIAN_16(ptr);	/* Rcv data size */
17561 		ptr += 2;
17562 		BIG_ENDIAN_16(ptr);	/* Concurrent Seq */
17563 		ptr += 2;
17564 		BIG_ENDIAN_16(ptr);	/* Rel offset */
17565 		ptr += 2;
17566 		BIG_ENDIAN_32(ptr);	/* E_D_TOV */
17567 		ptr += 4;		/* Port Name */
17568 		ptr += 8;		/* Node Name */
17569 		ptr += 8;		/* Class 1 */
17570 		ptr += 16;		/* Class 2 */
17571 		ptr += 16;		/* Class 3 */
17572 		BIG_ENDIAN_16(ptr);	/* Service options */
17573 		ptr += 2;
17574 		BIG_ENDIAN_16(ptr);	/* Initiator control */
17575 		ptr += 2;
17576 		BIG_ENDIAN_16(ptr);	/* Recipient Control */
17577 		ptr += 2;
17578 		BIG_ENDIAN_16(ptr);	/* Rcv size */
17579 		ptr += 2;
17580 		BIG_ENDIAN_16(ptr);	/* Concurrent Seq */
17581 		ptr += 2;
17582 		BIG_ENDIAN_16(ptr);	/* N_Port e2e credit */
17583 		ptr += 2;
17584 		BIG_ENDIAN_16(ptr);	/* Open Seq/Exch */
17585 		break;
17586 	}
17587 	case LA_ELS_PRLI: {
17588 		BIG_ENDIAN_32(ptr);	/* Command Code/Page length */
17589 		ptr += 4;		/* Type */
17590 		ptr += 2;
17591 		BIG_ENDIAN_16(ptr);	/* Flags */
17592 		ptr += 2;
17593 		BIG_ENDIAN_32(ptr);	/* Originator Process associator  */
17594 		ptr += 4;
17595 		BIG_ENDIAN_32(ptr);	/* Responder Process associator */
17596 		ptr += 4;
17597 		BIG_ENDIAN_32(ptr);	/* Flags */
17598 		break;
17599 	}
17600 	default:
17601 		EL(ha, "can't handle els code %x\n", ls_code);
17602 		break;
17603 	}
17604 }
17605 
17606 /*
17607  * ql_n_port_plogi
17608  *	In N port 2 N port topology where an N Port has logged in with the
17609  *	firmware because it has the N_Port login initiative, we send up
17610  *	a plogi by proxy which stimulates the login procedure to continue.
17611  *
17612  * Input:
17613  *	ha = adapter state pointer.
17614  * Returns:
17615  *
17616  * Context:
17617  *	Kernel context.
17618  */
17619 static int
ql_n_port_plogi(ql_adapter_state_t * ha)17620 ql_n_port_plogi(ql_adapter_state_t *ha)
17621 {
17622 	int		rval;
17623 	ql_tgt_t	*tq;
17624 	ql_head_t done_q = { NULL, NULL };
17625 
17626 	rval = QL_SUCCESS;
17627 
17628 	if (ha->topology & QL_N_PORT) {
17629 		/* if we're doing this the n_port_handle must be good */
17630 		if (LOCAL_LOOP_ID(ha->n_port->n_port_handle)) {
17631 			tq = ql_loop_id_to_queue(ha,
17632 			    ha->n_port->n_port_handle);
17633 			if (tq != NULL) {
17634 				(void) ql_send_plogi(ha, tq, &done_q);
17635 			} else {
17636 				EL(ha, "n_port_handle = %x, tq = %x\n",
17637 				    ha->n_port->n_port_handle, tq);
17638 			}
17639 		} else {
17640 			EL(ha, "n_port_handle = %x, tq = %x\n",
17641 			    ha->n_port->n_port_handle, tq);
17642 		}
17643 		if (done_q.first != NULL) {
17644 			ql_done(done_q.first);
17645 		}
17646 	}
17647 	return (rval);
17648 }
17649 
17650 /*
17651  * Compare two WWNs. The NAA is omitted for comparison.
17652  *
17653  * Note particularly that the indentation used in this
17654  * function  isn't according to Sun recommendations. It
17655  * is indented to make reading a bit easy.
17656  *
17657  * Return Values:
17658  *   if first == second return  0
17659  *   if first > second  return  1
17660  *   if first < second  return -1
17661  */
17662 int
ql_wwn_cmp(ql_adapter_state_t * ha,la_wwn_t * first,la_wwn_t * second)17663 ql_wwn_cmp(ql_adapter_state_t *ha, la_wwn_t *first, la_wwn_t *second)
17664 {
17665 	la_wwn_t t1, t2;
17666 	int rval;
17667 
17668 	EL(ha, "WWPN=%08x%08x\n",
17669 	    BE_32(first->i_wwn[0]), BE_32(first->i_wwn[1]));
17670 	EL(ha, "WWPN=%08x%08x\n",
17671 	    BE_32(second->i_wwn[0]), BE_32(second->i_wwn[1]));
17672 	/*
17673 	 * Fibre Channel protocol is big endian, so compare
17674 	 * as big endian values
17675 	 */
17676 	t1.i_wwn[0] = BE_32(first->i_wwn[0]);
17677 	t1.i_wwn[1] = BE_32(first->i_wwn[1]);
17678 
17679 	t2.i_wwn[0] = BE_32(second->i_wwn[0]);
17680 	t2.i_wwn[1] = BE_32(second->i_wwn[1]);
17681 
17682 	if (t1.i_wwn[0] == t2.i_wwn[0]) {
17683 		if (t1.i_wwn[1] == t2.i_wwn[1]) {
17684 			rval = 0;
17685 		} else if (t1.i_wwn[1] > t2.i_wwn[1]) {
17686 			rval = 1;
17687 		} else {
17688 			rval = -1;
17689 		}
17690 	} else {
17691 		if (t1.i_wwn[0] > t2.i_wwn[0]) {
17692 			rval = 1;
17693 		} else {
17694 			rval = -1;
17695 		}
17696 	}
17697 	return (rval);
17698 }
17699 
17700 /*
17701  * ql_wait_for_td_stop
17702  *	Wait for task daemon to stop running.  Internal command timeout
17703  *	is approximately 30 seconds, so it may help in some corner
17704  *	cases to wait that long
17705  *
17706  * Input:
17707  *	ha = adapter state pointer.
17708  *
17709  * Returns:
17710  *	DDI_SUCCESS or DDI_FAILURE.
17711  *
17712  * Context:
17713  *	Kernel context.
17714  */
17715 
17716 static int
ql_wait_for_td_stop(ql_adapter_state_t * ha)17717 ql_wait_for_td_stop(ql_adapter_state_t *ha)
17718 {
17719 	int	rval = DDI_FAILURE;
17720 	UINT16	wait_cnt;
17721 
17722 	for (wait_cnt = 0; wait_cnt < 3000; wait_cnt++) {
17723 		/* The task daemon clears the stop flag on exit. */
17724 		if (ha->task_daemon_flags & TASK_DAEMON_STOP_FLG) {
17725 			if (ha->cprinfo.cc_events & CALLB_CPR_START ||
17726 			    ddi_in_panic()) {
17727 				drv_usecwait(10000);
17728 			} else {
17729 				delay(drv_usectohz(10000));
17730 			}
17731 		} else {
17732 			rval = DDI_SUCCESS;
17733 			break;
17734 		}
17735 	}
17736 	return (rval);
17737 }
17738 
17739 /*
17740  * ql_nvram_cache_desc_ctor - Construct an nvram cache descriptor.
17741  *
17742  * Input:	Pointer to the adapter state structure.
17743  * Returns:	Success or Failure.
17744  * Context:	Kernel context.
17745  */
17746 int
ql_nvram_cache_desc_ctor(ql_adapter_state_t * ha)17747 ql_nvram_cache_desc_ctor(ql_adapter_state_t *ha)
17748 {
17749 	int	rval = DDI_SUCCESS;
17750 
17751 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
17752 
17753 	ha->nvram_cache =
17754 	    (nvram_cache_desc_t *)kmem_zalloc(sizeof (nvram_cache_desc_t),
17755 	    KM_SLEEP);
17756 
17757 	if (ha->nvram_cache == NULL) {
17758 		cmn_err(CE_WARN, "%s(%d): can't construct nvram cache"
17759 		    " descriptor", QL_NAME, ha->instance);
17760 		rval = DDI_FAILURE;
17761 	} else {
17762 		if (CFG_IST(ha, CFG_CTRL_24258081)) {
17763 			ha->nvram_cache->size = sizeof (nvram_24xx_t);
17764 		} else {
17765 			ha->nvram_cache->size = sizeof (nvram_t);
17766 		}
17767 		ha->nvram_cache->cache =
17768 		    (void *)kmem_zalloc(ha->nvram_cache->size, KM_SLEEP);
17769 		if (ha->nvram_cache->cache == NULL) {
17770 			cmn_err(CE_WARN, "%s(%d): can't get nvram cache buffer",
17771 			    QL_NAME, ha->instance);
17772 			kmem_free(ha->nvram_cache,
17773 			    sizeof (nvram_cache_desc_t));
17774 			ha->nvram_cache = 0;
17775 			rval = DDI_FAILURE;
17776 		} else {
17777 			mutex_init(&ha->nvram_cache->mutex, NULL,
17778 			    MUTEX_DRIVER, NULL);
17779 			ha->nvram_cache->valid = 0;
17780 		}
17781 	}
17782 
17783 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
17784 
17785 	return (rval);
17786 }
17787 
17788 /*
17789  * ql_nvram_cache_desc_dtor - Destroy an nvram cache descriptor.
17790  *
17791  * Input:	Pointer to the adapter state structure.
17792  * Returns:	Success or Failure.
17793  * Context:	Kernel context.
17794  */
17795 int
ql_nvram_cache_desc_dtor(ql_adapter_state_t * ha)17796 ql_nvram_cache_desc_dtor(ql_adapter_state_t *ha)
17797 {
17798 	int	rval = DDI_SUCCESS;
17799 
17800 	QL_PRINT_3(CE_CONT, "(%d): started\n", ha->instance);
17801 
17802 	if (ha->nvram_cache == NULL) {
17803 		cmn_err(CE_WARN, "%s(%d): can't destroy nvram descriptor",
17804 		    QL_NAME, ha->instance);
17805 		rval = DDI_FAILURE;
17806 	} else {
17807 		if (ha->nvram_cache->cache != NULL) {
17808 			kmem_free(ha->nvram_cache->cache,
17809 			    ha->nvram_cache->size);
17810 		}
17811 		mutex_destroy(&ha->nvram_cache->mutex);
17812 		kmem_free(ha->nvram_cache, sizeof (nvram_cache_desc_t));
17813 	}
17814 
17815 	QL_PRINT_3(CE_CONT, "(%d): done\n", ha->instance);
17816 
17817 	return (rval);
17818 }
17819 
17820 /*
17821  * ql_process_idc_event - Handle an Inter-Driver Communication async event.
17822  *
17823  * Input:	Pointer to the adapter state structure.
17824  * Returns:	void
17825  * Context:	Kernel context.
17826  */
17827 static void
ql_process_idc_event(ql_adapter_state_t * ha)17828 ql_process_idc_event(ql_adapter_state_t *ha)
17829 {
17830 	int	rval;
17831 
17832 	switch (ha->idc_mb[0]) {
17833 	case MBA_IDC_NOTIFICATION:
17834 		/*
17835 		 * The informational opcode (idc_mb[2]) can be a
17836 		 * defined value or the mailbox command being executed
17837 		 * on another function which stimulated this IDC message.
17838 		 */
17839 		ADAPTER_STATE_LOCK(ha);
17840 		switch (ha->idc_mb[2]) {
17841 		case IDC_OPC_DRV_START:
17842 			if (ha->idc_flash_acc != 0) {
17843 				ha->idc_flash_acc--;
17844 				if (ha->idc_flash_acc == 0) {
17845 					ha->idc_flash_acc_timer = 0;
17846 					GLOBAL_HW_UNLOCK();
17847 				}
17848 			}
17849 			if (ha->idc_restart_cnt != 0) {
17850 				ha->idc_restart_cnt--;
17851 				if (ha->idc_restart_cnt == 0) {
17852 					ha->idc_restart_timer = 0;
17853 					ADAPTER_STATE_UNLOCK(ha);
17854 					TASK_DAEMON_LOCK(ha);
17855 					ha->task_daemon_flags &= ~DRIVER_STALL;
17856 					TASK_DAEMON_UNLOCK(ha);
17857 					ql_restart_queues(ha);
17858 				} else {
17859 					ADAPTER_STATE_UNLOCK(ha);
17860 				}
17861 			} else {
17862 				ADAPTER_STATE_UNLOCK(ha);
17863 			}
17864 			break;
17865 		case IDC_OPC_FLASH_ACC:
17866 			ha->idc_flash_acc_timer = 30;
17867 			if (ha->idc_flash_acc == 0) {
17868 				GLOBAL_HW_LOCK();
17869 			}
17870 			ha->idc_flash_acc++;
17871 			ADAPTER_STATE_UNLOCK(ha);
17872 			break;
17873 		case IDC_OPC_RESTART_MPI:
17874 			ha->idc_restart_timer = 30;
17875 			ha->idc_restart_cnt++;
17876 			ADAPTER_STATE_UNLOCK(ha);
17877 			TASK_DAEMON_LOCK(ha);
17878 			ha->task_daemon_flags |= DRIVER_STALL;
17879 			TASK_DAEMON_UNLOCK(ha);
17880 			break;
17881 		case IDC_OPC_PORT_RESET_MBC:
17882 		case IDC_OPC_SET_PORT_CONFIG_MBC:
17883 			ha->idc_restart_timer = 30;
17884 			ha->idc_restart_cnt++;
17885 			ADAPTER_STATE_UNLOCK(ha);
17886 			TASK_DAEMON_LOCK(ha);
17887 			ha->task_daemon_flags |= DRIVER_STALL;
17888 			TASK_DAEMON_UNLOCK(ha);
17889 			(void) ql_wait_outstanding(ha);
17890 			break;
17891 		default:
17892 			ADAPTER_STATE_UNLOCK(ha);
17893 			EL(ha, "Unknown IDC opcode=%xh %xh\n", ha->idc_mb[0],
17894 			    ha->idc_mb[2]);
17895 			break;
17896 		}
17897 		/*
17898 		 * If there is a timeout value associated with this IDC
17899 		 * notification then there is an implied requirement
17900 		 * that we return an ACK.
17901 		 */
17902 		if (ha->idc_mb[1] & IDC_TIMEOUT_MASK) {
17903 			rval = ql_idc_ack(ha);
17904 			if (rval != QL_SUCCESS) {
17905 				EL(ha, "idc_ack status=%xh %xh\n", rval,
17906 				    ha->idc_mb[2]);
17907 			}
17908 		}
17909 		break;
17910 	case MBA_IDC_COMPLETE:
17911 		/*
17912 		 * We don't ACK completions, only these require action.
17913 		 */
17914 		switch (ha->idc_mb[2]) {
17915 		case IDC_OPC_PORT_RESET_MBC:
17916 		case IDC_OPC_SET_PORT_CONFIG_MBC:
17917 			ADAPTER_STATE_LOCK(ha);
17918 			if (ha->idc_restart_cnt != 0) {
17919 				ha->idc_restart_cnt--;
17920 				if (ha->idc_restart_cnt == 0) {
17921 					ha->idc_restart_timer = 0;
17922 					ADAPTER_STATE_UNLOCK(ha);
17923 					TASK_DAEMON_LOCK(ha);
17924 					ha->task_daemon_flags &= ~DRIVER_STALL;
17925 					TASK_DAEMON_UNLOCK(ha);
17926 					ql_restart_queues(ha);
17927 				} else {
17928 					ADAPTER_STATE_UNLOCK(ha);
17929 				}
17930 			} else {
17931 				ADAPTER_STATE_UNLOCK(ha);
17932 			}
17933 			break;
17934 		default:
17935 			break; /* Don't care... */
17936 		}
17937 		break;
17938 	case MBA_IDC_TIME_EXTENDED:
17939 		QL_PRINT_10(CE_CONT, "(%d): MBA_IDC_TIME_EXTENDED="
17940 		    "%xh\n", ha->instance, ha->idc_mb[2]);
17941 		break;
17942 	default:
17943 		EL(ha, "Inconsistent IDC event =%xh %xh\n", ha->idc_mb[0],
17944 		    ha->idc_mb[2]);
17945 		ADAPTER_STATE_UNLOCK(ha);
17946 		break;
17947 	}
17948 }
17949