xref: /titanic_44/usr/src/uts/common/io/fibre-channel/ulp/fcp.c (revision 379c004d1f26b343f034bba8a350290691d00d38)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 /*
22  * Copyright 2009 Sun Microsystems, Inc.  All rights reserved.
23  * Use is subject to license terms.
24  *
25  * Fibre Channel SCSI ULP Mapping driver
26  */
27 
28 #include <sys/scsi/scsi.h>
29 #include <sys/types.h>
30 #include <sys/varargs.h>
31 #include <sys/devctl.h>
32 #include <sys/thread.h>
33 #include <sys/thread.h>
34 #include <sys/open.h>
35 #include <sys/file.h>
36 #include <sys/sunndi.h>
37 #include <sys/console.h>
38 #include <sys/proc.h>
39 #include <sys/time.h>
40 #include <sys/utsname.h>
41 #include <sys/scsi/impl/scsi_reset_notify.h>
42 #include <sys/ndi_impldefs.h>
43 #include <sys/byteorder.h>
44 #include <sys/fs/dv_node.h>
45 #include <sys/ctype.h>
46 #include <sys/sunmdi.h>
47 
48 #include <sys/fibre-channel/fc.h>
49 #include <sys/fibre-channel/impl/fc_ulpif.h>
50 #include <sys/fibre-channel/ulp/fcpvar.h>
51 
52 /*
53  * Discovery Process
54  * =================
55  *
56  *    The discovery process is a major function of FCP.	 In order to help
57  * understand that function a flow diagram is given here.  This diagram
58  * doesn't claim to cover all the cases and the events that can occur during
59  * the discovery process nor the subtleties of the code.  The code paths shown
60  * are simplified.  Its purpose is to help the reader (and potentially bug
61  * fixer) have an overall view of the logic of the code.  For that reason the
62  * diagram covers the simple case of the line coming up cleanly or of a new
63  * port attaching to FCP the link being up.  The reader must keep in mind
64  * that:
65  *
66  *	- There are special cases where bringing devices online and offline
67  *	  is driven by Ioctl.
68  *
69  *	- The behavior of the discovery process can be modified through the
70  *	  .conf file.
71  *
72  *	- The line can go down and come back up at any time during the
73  *	  discovery process which explains some of the complexity of the code.
74  *
75  * ............................................................................
76  *
77  * STEP 1: The line comes up or a new Fibre Channel port attaches to FCP.
78  *
79  *
80  *			+-------------------------+
81  *   fp/fctl module --->|    fcp_port_attach	  |
82  *			+-------------------------+
83  *	   |			     |
84  *	   |			     |
85  *	   |			     v
86  *	   |		+-------------------------+
87  *	   |		| fcp_handle_port_attach  |
88  *	   |		+-------------------------+
89  *	   |				|
90  *	   |				|
91  *	   +--------------------+	|
92  *				|	|
93  *				v	v
94  *			+-------------------------+
95  *			|   fcp_statec_callback	  |
96  *			+-------------------------+
97  *				    |
98  *				    |
99  *				    v
100  *			+-------------------------+
101  *			|    fcp_handle_devices	  |
102  *			+-------------------------+
103  *				    |
104  *				    |
105  *				    v
106  *			+-------------------------+
107  *			|   fcp_handle_mapflags	  |
108  *			+-------------------------+
109  *				    |
110  *				    |
111  *				    v
112  *			+-------------------------+
113  *			|     fcp_send_els	  |
114  *			|			  |
115  *			| PLOGI or PRLI To all the|
116  *			| reachable devices.	  |
117  *			+-------------------------+
118  *
119  *
120  * ............................................................................
121  *
122  * STEP 2: The callback functions of the PLOGI and/or PRLI requests sent during
123  *	   STEP 1 are called (it is actually the same function).
124  *
125  *
126  *			+-------------------------+
127  *			|    fcp_icmd_callback	  |
128  *   fp/fctl module --->|			  |
129  *			| callback for PLOGI and  |
130  *			| PRLI.			  |
131  *			+-------------------------+
132  *				     |
133  *				     |
134  *	    Received PLOGI Accept   /-\	  Received PRLI Accept
135  *		       _ _ _ _ _ _ /   \_ _ _ _ _ _
136  *		      |		   \   /	   |
137  *		      |		    \-/		   |
138  *		      |				   |
139  *		      v				   v
140  *	+-------------------------+	+-------------------------+
141  *	|     fcp_send_els	  |	|     fcp_send_scsi	  |
142  *	|			  |	|			  |
143  *	|	  PRLI		  |	|	REPORT_LUN	  |
144  *	+-------------------------+	+-------------------------+
145  *
146  * ............................................................................
147  *
148  * STEP 3: The callback functions of the SCSI commands issued by FCP are called
149  *	   (It is actually the same function).
150  *
151  *
152  *			    +-------------------------+
153  *   fp/fctl module ------->|	 fcp_scsi_callback    |
154  *			    +-------------------------+
155  *					|
156  *					|
157  *					|
158  *	Receive REPORT_LUN reply       /-\	Receive INQUIRY PAGE83 reply
159  *		  _ _ _ _ _ _ _ _ _ _ /	  \_ _ _ _ _ _ _ _ _ _ _ _
160  *		 |		      \	  /			  |
161  *		 |		       \-/			  |
162  *		 |			|			  |
163  *		 | Receive INQUIRY reply|			  |
164  *		 |			|			  |
165  *		 v			v			  v
166  * +------------------------+ +----------------------+ +----------------------+
167  * |  fcp_handle_reportlun  | |	 fcp_handle_inquiry  | |  fcp_handle_page83   |
168  * |(Called for each Target)| | (Called for each LUN)| |(Called for each LUN) |
169  * +------------------------+ +----------------------+ +----------------------+
170  *		 |			|			  |
171  *		 |			|			  |
172  *		 |			|			  |
173  *		 v			v			  |
174  *     +-----------------+	+-----------------+		  |
175  *     |  fcp_send_scsi	 |	|  fcp_send_scsi  |		  |
176  *     |		 |	|		  |		  |
177  *     |     INQUIRY	 |	| INQUIRY PAGE83  |		  |
178  *     |  (To each LUN)	 |	+-----------------+		  |
179  *     +-----------------+					  |
180  *								  |
181  *								  v
182  *						      +------------------------+
183  *						      |	 fcp_call_finish_init  |
184  *						      +------------------------+
185  *								  |
186  *								  v
187  *						 +-----------------------------+
188  *						 |  fcp_call_finish_init_held  |
189  *						 +-----------------------------+
190  *								  |
191  *								  |
192  *			   All LUNs scanned			 /-\
193  *			       _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ /   \
194  *			      |					\   /
195  *			      |					 \-/
196  *			      v					  |
197  *		     +------------------+			  |
198  *		     |	fcp_finish_tgt	|			  |
199  *		     +------------------+			  |
200  *			      |	  Target Not Offline and	  |
201  *  Target Not Offline and    |	  not marked and tgt_node_state	  |
202  *  marked		     /-\  not FCP_TGT_NODE_ON_DEMAND	  |
203  *		_ _ _ _ _ _ /	\_ _ _ _ _ _ _ _		  |
204  *	       |	    \	/		|		  |
205  *	       |	     \-/		|		  |
206  *	       v				v		  |
207  * +----------------------------+     +-------------------+	  |
208  * |	 fcp_offline_target	|     |	 fcp_create_luns  |	  |
209  * |				|     +-------------------+	  |
210  * | A structure fcp_tgt_elem	|		|		  |
211  * | is created and queued in	|		v		  |
212  * | the FCP port list		|     +-------------------+	  |
213  * | port_offline_tgts.	 It	|     |	 fcp_pass_to_hp	  |	  |
214  * | will be unqueued by the	|     |			  |	  |
215  * | watchdog timer.		|     | Called for each	  |	  |
216  * +----------------------------+     | LUN. Dispatches	  |	  |
217  *		  |		      | fcp_hp_task	  |	  |
218  *		  |		      +-------------------+	  |
219  *		  |				|		  |
220  *		  |				|		  |
221  *		  |				|		  |
222  *		  |				+---------------->|
223  *		  |						  |
224  *		  +---------------------------------------------->|
225  *								  |
226  *								  |
227  *		All the targets (devices) have been scanned	 /-\
228  *				_ _ _ _	_ _ _ _	_ _ _ _ _ _ _ _ /   \
229  *			       |				\   /
230  *			       |				 \-/
231  *	    +-------------------------------------+		  |
232  *	    |		fcp_finish_init		  |		  |
233  *	    |					  |		  |
234  *	    | Signal broadcasts the condition	  |		  |
235  *	    | variable port_config_cv of the FCP  |		  |
236  *	    | port.  One potential code sequence  |		  |
237  *	    | waiting on the condition variable	  |		  |
238  *	    | the code sequence handling	  |		  |
239  *	    | BUS_CONFIG_ALL and BUS_CONFIG_DRIVER|		  |
240  *	    | The other is in the function	  |		  |
241  *	    | fcp_reconfig_wait which is called	  |		  |
242  *	    | in the transmit path preventing IOs |		  |
243  *	    | from going through till the disco-  |		  |
244  *	    | very process is over.		  |		  |
245  *	    +-------------------------------------+		  |
246  *			       |				  |
247  *			       |				  |
248  *			       +--------------------------------->|
249  *								  |
250  *								  v
251  *								Return
252  *
253  * ............................................................................
254  *
255  * STEP 4: The hot plug task is called (for each fcp_hp_elem).
256  *
257  *
258  *			+-------------------------+
259  *			|      fcp_hp_task	  |
260  *			+-------------------------+
261  *				     |
262  *				     |
263  *				     v
264  *			+-------------------------+
265  *			|     fcp_trigger_lun	  |
266  *			+-------------------------+
267  *				     |
268  *				     |
269  *				     v
270  *		   Bring offline    /-\	 Bring online
271  *		  _ _ _ _ _ _ _ _ _/   \_ _ _ _ _ _ _ _ _ _
272  *		 |		   \   /		   |
273  *		 |		    \-/			   |
274  *		 v					   v
275  *    +---------------------+			+-----------------------+
276  *    |	 fcp_offline_child  |			|      fcp_get_cip	|
277  *    +---------------------+			|			|
278  *						| Creates a dev_info_t	|
279  *						| or a mdi_pathinfo_t	|
280  *						| depending on whether	|
281  *						| mpxio is on or off.	|
282  *						+-----------------------+
283  *							   |
284  *							   |
285  *							   v
286  *						+-----------------------+
287  *						|  fcp_online_child	|
288  *						|			|
289  *						| Set device online	|
290  *						| using NDI or MDI.	|
291  *						+-----------------------+
292  *
293  * ............................................................................
294  *
295  * STEP 5: The watchdog timer expires.	The watch dog timer does much more that
296  *	   what is described here.  We only show the target offline path.
297  *
298  *
299  *			 +--------------------------+
300  *			 |	  fcp_watch	    |
301  *			 +--------------------------+
302  *				       |
303  *				       |
304  *				       v
305  *			 +--------------------------+
306  *			 |  fcp_scan_offline_tgts   |
307  *			 +--------------------------+
308  *				       |
309  *				       |
310  *				       v
311  *			 +--------------------------+
312  *			 |  fcp_offline_target_now  |
313  *			 +--------------------------+
314  *				       |
315  *				       |
316  *				       v
317  *			 +--------------------------+
318  *			 |   fcp_offline_tgt_luns   |
319  *			 +--------------------------+
320  *				       |
321  *				       |
322  *				       v
323  *			 +--------------------------+
324  *			 |     fcp_offline_lun	    |
325  *			 +--------------------------+
326  *				       |
327  *				       |
328  *				       v
329  *		     +----------------------------------+
330  *		     |	     fcp_offline_lun_now	|
331  *		     |					|
332  *		     | A request (or two if mpxio) is	|
333  *		     | sent to the hot plug task using	|
334  *		     | a fcp_hp_elem structure.		|
335  *		     +----------------------------------+
336  */
337 
338 /*
339  * Functions registered with DDI framework
340  */
341 static int fcp_attach(dev_info_t *devi, ddi_attach_cmd_t cmd);
342 static int fcp_detach(dev_info_t *devi, ddi_detach_cmd_t cmd);
343 static int fcp_open(dev_t *devp, int flag, int otype, cred_t *credp);
344 static int fcp_close(dev_t dev, int flag, int otype, cred_t *credp);
345 static int fcp_ioctl(dev_t dev, int cmd, intptr_t data, int mode,
346     cred_t *credp, int *rval);
347 
348 /*
349  * Functions registered with FC Transport framework
350  */
351 static int fcp_port_attach(opaque_t ulph, fc_ulp_port_info_t *pinfo,
352     fc_attach_cmd_t cmd,  uint32_t s_id);
353 static int fcp_port_detach(opaque_t ulph, fc_ulp_port_info_t *info,
354     fc_detach_cmd_t cmd);
355 static int fcp_port_ioctl(opaque_t ulph, opaque_t port_handle, dev_t dev,
356     int cmd, intptr_t data, int mode, cred_t *credp, int *rval,
357     uint32_t claimed);
358 static int fcp_els_callback(opaque_t ulph, opaque_t port_handle,
359     fc_unsol_buf_t *buf, uint32_t claimed);
360 static int fcp_data_callback(opaque_t ulph, opaque_t port_handle,
361     fc_unsol_buf_t *buf, uint32_t claimed);
362 static void fcp_statec_callback(opaque_t ulph, opaque_t port_handle,
363     uint32_t port_state, uint32_t port_top, fc_portmap_t *devlist,
364     uint32_t  dev_cnt, uint32_t port_sid);
365 
366 /*
367  * Functions registered with SCSA framework
368  */
369 static int fcp_phys_tgt_init(dev_info_t *hba_dip, dev_info_t *tgt_dip,
370     scsi_hba_tran_t *hba_tran, struct scsi_device *sd);
371 static int fcp_scsi_tgt_init(dev_info_t *hba_dip, dev_info_t *tgt_dip,
372     scsi_hba_tran_t *hba_tran, struct scsi_device *sd);
373 static void fcp_scsi_tgt_free(dev_info_t *hba_dip, dev_info_t *tgt_dip,
374     scsi_hba_tran_t *hba_tran, struct scsi_device *sd);
375 static int fcp_scsi_start(struct scsi_address *ap, struct scsi_pkt *pkt);
376 static int fcp_scsi_abort(struct scsi_address *ap, struct scsi_pkt *pkt);
377 static int fcp_scsi_reset(struct scsi_address *ap, int level);
378 static int fcp_scsi_getcap(struct scsi_address *ap, char *cap, int whom);
379 static int fcp_scsi_setcap(struct scsi_address *ap, char *cap, int value,
380     int whom);
381 static void fcp_pkt_teardown(struct scsi_pkt *pkt);
382 static int fcp_scsi_reset_notify(struct scsi_address *ap, int flag,
383     void (*callback)(caddr_t), caddr_t arg);
384 static int fcp_scsi_bus_get_eventcookie(dev_info_t *dip, dev_info_t *rdip,
385     char *name, ddi_eventcookie_t *event_cookiep);
386 static int fcp_scsi_bus_add_eventcall(dev_info_t *dip, dev_info_t *rdip,
387     ddi_eventcookie_t eventid, void (*callback)(), void *arg,
388     ddi_callback_id_t *cb_id);
389 static int fcp_scsi_bus_remove_eventcall(dev_info_t *devi,
390     ddi_callback_id_t cb_id);
391 static int fcp_scsi_bus_post_event(dev_info_t *dip, dev_info_t *rdip,
392     ddi_eventcookie_t eventid, void *impldata);
393 static int fcp_scsi_bus_config(dev_info_t *parent, uint_t flag,
394     ddi_bus_config_op_t op, void *arg, dev_info_t **childp);
395 static int fcp_scsi_bus_unconfig(dev_info_t *parent, uint_t flag,
396     ddi_bus_config_op_t op, void *arg);
397 
398 /*
399  * Internal functions
400  */
401 static int fcp_setup_device_data_ioctl(int cmd, struct fcp_ioctl *data,
402     int mode, int *rval);
403 
404 static int fcp_setup_scsi_ioctl(struct fcp_scsi_cmd *u_fscsi,
405     int mode, int *rval);
406 static int fcp_copyin_scsi_cmd(caddr_t base_addr,
407     struct fcp_scsi_cmd *fscsi, int mode);
408 static int fcp_copyout_scsi_cmd(struct fcp_scsi_cmd *fscsi,
409     caddr_t base_addr, int mode);
410 static int fcp_send_scsi_ioctl(struct fcp_scsi_cmd *fscsi);
411 
412 static struct fcp_tgt *fcp_port_create_tgt(struct fcp_port *pptr,
413     la_wwn_t *pwwn, int	*ret_val, int *fc_status, int *fc_pkt_state,
414     int *fc_pkt_reason, int *fc_pkt_action);
415 static int fcp_tgt_send_plogi(struct fcp_tgt *ptgt, int *fc_status,
416     int *fc_pkt_state, int *fc_pkt_reason, int *fc_pkt_action);
417 static int fcp_tgt_send_prli(struct fcp_tgt	*ptgt, int *fc_status,
418     int *fc_pkt_state, int *fc_pkt_reason, int *fc_pkt_action);
419 static void fcp_ipkt_sema_init(struct fcp_ipkt *icmd);
420 static int fcp_ipkt_sema_wait(struct fcp_ipkt *icmd);
421 static void fcp_ipkt_sema_callback(struct fc_packet *fpkt);
422 static void fcp_ipkt_sema_cleanup(struct fcp_ipkt *icmd);
423 
424 static void fcp_handle_devices(struct fcp_port *pptr,
425     fc_portmap_t devlist[], uint32_t dev_cnt, int link_cnt,
426     fcp_map_tag_t *map_tag, int cause);
427 static int fcp_handle_mapflags(struct fcp_port *pptr,
428     struct fcp_tgt *ptgt, fc_portmap_t *map_entry, int link_cnt,
429     int tgt_cnt, int cause);
430 static int fcp_send_els(struct fcp_port *pptr, struct fcp_tgt *ptgt,
431     struct fcp_ipkt *icmd, uchar_t opcode, int lcount, int tcount, int cause);
432 static void fcp_update_state(struct fcp_port *pptr, uint32_t state,
433     int cause);
434 static void fcp_update_tgt_state(struct fcp_tgt *ptgt, int flag,
435     uint32_t state);
436 static struct fcp_port *fcp_get_port(opaque_t port_handle);
437 static void fcp_unsol_callback(fc_packet_t *fpkt);
438 static void fcp_unsol_resp_init(fc_packet_t *pkt, fc_unsol_buf_t *buf,
439     uchar_t r_ctl, uchar_t type);
440 static int fcp_unsol_prli(struct fcp_port *pptr, fc_unsol_buf_t *buf);
441 static struct fcp_ipkt *fcp_icmd_alloc(struct fcp_port *pptr,
442     struct fcp_tgt *ptgt, int cmd_len, int resp_len, int data_len,
443     int nodma, int lcount, int tcount, int cause, uint32_t rscn_count);
444 static void fcp_icmd_free(struct fcp_port *pptr, struct fcp_ipkt *icmd);
445 static int fcp_alloc_dma(struct fcp_port *pptr, struct fcp_ipkt *icmd,
446     int nodma, int flags);
447 static void fcp_free_dma(struct fcp_port *pptr, struct fcp_ipkt *icmd);
448 static struct fcp_tgt *fcp_lookup_target(struct fcp_port *pptr,
449     uchar_t *wwn);
450 static struct fcp_tgt *fcp_get_target_by_did(struct fcp_port *pptr,
451     uint32_t d_id);
452 static void fcp_icmd_callback(fc_packet_t *fpkt);
453 static int fcp_send_scsi(struct fcp_lun *plun, uchar_t opcode,
454     int len, int lcount, int tcount, int cause, uint32_t rscn_count);
455 static int fcp_check_reportlun(struct fcp_rsp *rsp, fc_packet_t *fpkt);
456 static void fcp_scsi_callback(fc_packet_t *fpkt);
457 static void fcp_retry_scsi_cmd(fc_packet_t *fpkt);
458 static void fcp_handle_inquiry(fc_packet_t *fpkt, struct fcp_ipkt *icmd);
459 static void fcp_handle_reportlun(fc_packet_t *fpkt, struct fcp_ipkt *icmd);
460 static struct fcp_lun *fcp_get_lun(struct fcp_tgt *ptgt,
461     uint16_t lun_num);
462 static int fcp_finish_tgt(struct fcp_port *pptr, struct fcp_tgt *ptgt,
463     int link_cnt, int tgt_cnt, int cause);
464 static void fcp_finish_init(struct fcp_port *pptr);
465 static void fcp_create_luns(struct fcp_tgt *ptgt, int link_cnt,
466     int tgt_cnt, int cause);
467 static int fcp_trigger_lun(struct fcp_lun *plun, child_info_t *cip,
468     int old_mpxio, int online, int link_cnt, int tgt_cnt, int flags);
469 static int fcp_offline_target(struct fcp_port *pptr, struct fcp_tgt *ptgt,
470     int link_cnt, int tgt_cnt, int nowait, int flags);
471 static void fcp_offline_target_now(struct fcp_port *pptr,
472     struct fcp_tgt *ptgt, int link_cnt, int tgt_cnt, int flags);
473 static void fcp_offline_tgt_luns(struct fcp_tgt *ptgt, int link_cnt,
474     int tgt_cnt, int flags);
475 static void fcp_offline_lun(struct fcp_lun *plun, int link_cnt, int tgt_cnt,
476     int nowait, int flags);
477 static void fcp_prepare_offline_lun(struct fcp_lun *plun, int link_cnt,
478     int tgt_cnt);
479 static void fcp_offline_lun_now(struct fcp_lun *plun, int link_cnt,
480     int tgt_cnt, int flags);
481 static void fcp_scan_offline_luns(struct fcp_port *pptr);
482 static void fcp_scan_offline_tgts(struct fcp_port *pptr);
483 static void fcp_update_offline_flags(struct fcp_lun *plun);
484 static struct fcp_pkt *fcp_scan_commands(struct fcp_lun *plun);
485 static void fcp_abort_commands(struct fcp_pkt *head, struct
486     fcp_port *pptr);
487 static void fcp_cmd_callback(fc_packet_t *fpkt);
488 static void fcp_complete_pkt(fc_packet_t *fpkt);
489 static int fcp_validate_fcp_response(struct fcp_rsp *rsp,
490     struct fcp_port *pptr);
491 static int fcp_device_changed(struct fcp_port *pptr, struct fcp_tgt *ptgt,
492     fc_portmap_t *map_entry, int link_cnt, int tgt_cnt, int cause);
493 static struct fcp_lun *fcp_alloc_lun(struct fcp_tgt *ptgt);
494 static void fcp_dealloc_lun(struct fcp_lun *plun);
495 static struct fcp_tgt *fcp_alloc_tgt(struct fcp_port *pptr,
496     fc_portmap_t *map_entry, int link_cnt);
497 static void fcp_dealloc_tgt(struct fcp_tgt *ptgt);
498 static void fcp_queue_ipkt(struct fcp_port *pptr, fc_packet_t *fpkt);
499 static int fcp_transport(opaque_t port_handle, fc_packet_t *fpkt,
500     int internal);
501 static void fcp_log(int level, dev_info_t *dip, const char *fmt, ...);
502 static int fcp_handle_port_attach(opaque_t ulph, fc_ulp_port_info_t *pinfo,
503     uint32_t s_id, int instance);
504 static int fcp_handle_port_detach(struct fcp_port *pptr, int flag,
505     int instance);
506 static void fcp_cleanup_port(struct fcp_port *pptr, int instance);
507 static int fcp_kmem_cache_constructor(struct scsi_pkt *, scsi_hba_tran_t *,
508     int);
509 static void fcp_kmem_cache_destructor(struct  scsi_pkt *, scsi_hba_tran_t *);
510 static int fcp_pkt_setup(struct scsi_pkt *, int (*)(), caddr_t);
511 static int fcp_alloc_cmd_resp(struct fcp_port *pptr, fc_packet_t *fpkt,
512     int flags);
513 static void fcp_free_cmd_resp(struct fcp_port *pptr, fc_packet_t *fpkt);
514 static int fcp_reset_target(struct scsi_address *ap, int level);
515 static int fcp_commoncap(struct scsi_address *ap, char *cap,
516     int val, int tgtonly, int doset);
517 static int fcp_scsi_get_name(struct scsi_device *sd, char *name, int len);
518 static int fcp_scsi_get_bus_addr(struct scsi_device *sd, char *name, int len);
519 static int fcp_linkreset(struct fcp_port *pptr, struct scsi_address *ap,
520     int sleep);
521 static int fcp_handle_port_resume(opaque_t ulph, fc_ulp_port_info_t *pinfo,
522     uint32_t s_id, fc_attach_cmd_t cmd, int instance);
523 static void fcp_cp_pinfo(struct fcp_port *pptr, fc_ulp_port_info_t *pinfo);
524 static void fcp_process_elem(struct fcp_hp_elem *elem, int result);
525 static child_info_t *fcp_get_cip(struct fcp_lun *plun, child_info_t *cip,
526     int lcount, int tcount);
527 static int fcp_is_dip_present(struct fcp_lun *plun, dev_info_t *cdip);
528 static int fcp_is_child_present(struct fcp_lun *plun, child_info_t *cip);
529 static dev_info_t *fcp_create_dip(struct fcp_lun *plun, int link_cnt,
530     int tgt_cnt);
531 static dev_info_t *fcp_find_existing_dip(struct fcp_lun *plun,
532     dev_info_t *pdip, caddr_t name);
533 static int fcp_online_child(struct fcp_lun *plun, child_info_t *cip,
534     int lcount, int tcount, int flags, int *circ);
535 static int fcp_offline_child(struct fcp_lun *plun, child_info_t *cip,
536     int lcount, int tcount, int flags, int *circ);
537 static void fcp_remove_child(struct fcp_lun *plun);
538 static void fcp_watch(void *arg);
539 static void fcp_check_reset_delay(struct fcp_port *pptr);
540 static void fcp_abort_all(struct fcp_port *pptr, struct fcp_tgt *ttgt,
541     struct fcp_lun *rlun, int tgt_cnt);
542 struct fcp_port *fcp_soft_state_unlink(struct fcp_port *pptr);
543 static struct fcp_lun *fcp_lookup_lun(struct fcp_port *pptr,
544     uchar_t *wwn, uint16_t lun);
545 static void fcp_prepare_pkt(struct fcp_port *pptr, struct fcp_pkt *cmd,
546     struct fcp_lun *plun);
547 static void fcp_post_callback(struct fcp_pkt *cmd);
548 static int fcp_dopoll(struct fcp_port *pptr, struct fcp_pkt *cmd);
549 static struct fcp_port *fcp_dip2port(dev_info_t *dip);
550 struct fcp_lun *fcp_get_lun_from_cip(struct fcp_port *pptr,
551     child_info_t *cip);
552 static int fcp_pass_to_hp_and_wait(struct fcp_port *pptr,
553     struct fcp_lun *plun, child_info_t *cip, int what, int link_cnt,
554     int tgt_cnt, int flags);
555 static struct fcp_hp_elem *fcp_pass_to_hp(struct fcp_port *pptr,
556     struct fcp_lun *plun, child_info_t *cip, int what, int link_cnt,
557     int tgt_cnt, int flags, int wait);
558 static void fcp_retransport_cmd(struct fcp_port *pptr,
559     struct fcp_pkt *cmd);
560 static void fcp_fail_cmd(struct fcp_pkt *cmd, uchar_t reason,
561     uint_t statistics);
562 static void fcp_queue_pkt(struct fcp_port *pptr, struct fcp_pkt *cmd);
563 static void fcp_update_targets(struct fcp_port *pptr,
564     fc_portmap_t *dev_list, uint32_t count, uint32_t state, int cause);
565 static int fcp_call_finish_init(struct fcp_port *pptr,
566     struct fcp_tgt *ptgt, int lcount, int tcount, int cause);
567 static int fcp_call_finish_init_held(struct fcp_port *pptr,
568     struct fcp_tgt *ptgt, int lcount, int tcount, int cause);
569 static void fcp_reconfigure_luns(void * tgt_handle);
570 static void fcp_free_targets(struct fcp_port *pptr);
571 static void fcp_free_target(struct fcp_tgt *ptgt);
572 static int fcp_is_retryable(struct fcp_ipkt *icmd);
573 static int fcp_create_on_demand(struct fcp_port *pptr, uchar_t *pwwn);
574 static void fcp_ascii_to_wwn(caddr_t string, uchar_t bytes[], unsigned int);
575 static void fcp_wwn_to_ascii(uchar_t bytes[], char *string);
576 static void fcp_print_error(fc_packet_t *fpkt);
577 static int fcp_handle_ipkt_errors(struct fcp_port *pptr,
578     struct fcp_tgt *ptgt, struct fcp_ipkt *icmd, int rval, caddr_t op);
579 static int fcp_outstanding_lun_cmds(struct fcp_tgt *ptgt);
580 static fc_portmap_t *fcp_construct_map(struct fcp_port *pptr,
581     uint32_t *dev_cnt);
582 static void fcp_offline_all(struct fcp_port *pptr, int lcount, int cause);
583 static int fcp_get_statec_count(struct fcp_ioctl *data, int mode, int *rval);
584 static int fcp_copyin_fcp_ioctl_data(struct fcp_ioctl *, int, int *,
585     struct fcp_ioctl *, struct fcp_port **);
586 static char *fcp_get_lun_path(struct fcp_lun *plun);
587 static int fcp_get_target_mappings(struct fcp_ioctl *data, int mode,
588     int *rval);
589 static int fcp_do_ns_registry(struct fcp_port *pptr, uint32_t s_id);
590 static void fcp_retry_ns_registry(struct fcp_port *pptr, uint32_t s_id);
591 static char *fcp_get_lun_path(struct fcp_lun *plun);
592 static int fcp_get_target_mappings(struct fcp_ioctl *data, int mode,
593     int *rval);
594 static void fcp_reconfig_wait(struct fcp_port *pptr);
595 
596 /*
597  * New functions added for mpxio support
598  */
599 static int fcp_virt_tgt_init(dev_info_t *hba_dip, dev_info_t *tgt_dip,
600     scsi_hba_tran_t *hba_tran, struct scsi_device *sd);
601 static mdi_pathinfo_t *fcp_create_pip(struct fcp_lun *plun, int lcount,
602     int tcount);
603 static mdi_pathinfo_t *fcp_find_existing_pip(struct fcp_lun *plun,
604     dev_info_t *pdip);
605 static int fcp_is_pip_present(struct fcp_lun *plun, mdi_pathinfo_t *pip);
606 static void fcp_handle_page83(fc_packet_t *, struct fcp_ipkt *, int);
607 static void fcp_update_mpxio_path_verifybusy(struct fcp_port *pptr);
608 static int fcp_copy_guid_2_lun_block(struct fcp_lun *plun, char *guidp);
609 static int fcp_update_mpxio_path(struct fcp_lun *plun, child_info_t *cip,
610     int what);
611 static int fcp_is_reconfig_needed(struct fcp_tgt *ptgt,
612     fc_packet_t *fpkt);
613 static int fcp_symmetric_device_probe(struct fcp_lun *plun);
614 
615 /*
616  * New functions added for lun masking support
617  */
618 static void fcp_read_blacklist(dev_info_t *dip,
619     struct fcp_black_list_entry **pplun_blacklist);
620 static void fcp_mask_pwwn_lun(char *curr_pwwn, char *curr_lun,
621     struct fcp_black_list_entry **pplun_blacklist);
622 static void fcp_add_one_mask(char *curr_pwwn, uint32_t lun_id,
623     struct fcp_black_list_entry **pplun_blacklist);
624 static int fcp_should_mask(la_wwn_t *wwn, uint32_t lun_id);
625 static void fcp_cleanup_blacklist(struct fcp_black_list_entry **lun_blacklist);
626 
627 extern struct mod_ops	mod_driverops;
628 /*
629  * This variable is defined in modctl.c and set to '1' after the root driver
630  * and fs are loaded.  It serves as an indication that the root filesystem can
631  * be used.
632  */
633 extern int		modrootloaded;
634 /*
635  * This table contains strings associated with the SCSI sense key codes.  It
636  * is used by FCP to print a clear explanation of the code returned in the
637  * sense information by a device.
638  */
639 extern char		*sense_keys[];
640 /*
641  * This device is created by the SCSI pseudo nexus driver (SCSI vHCI).	It is
642  * under this device that the paths to a physical device are created when
643  * MPxIO is used.
644  */
645 extern dev_info_t	*scsi_vhci_dip;
646 
647 /*
648  * Report lun processing
649  */
650 #define	FCP_LUN_ADDRESSING		0x80
651 #define	FCP_PD_ADDRESSING		0x00
652 #define	FCP_VOLUME_ADDRESSING		0x40
653 
654 #define	FCP_SVE_THROTTLE		0x28 /* Vicom */
655 #define	MAX_INT_DMA			0x7fffffff
656 #define	FCP_MAX_SENSE_LEN		252
657 #define	FCP_MAX_RESPONSE_LEN		0xffffff
658 /*
659  * Property definitions
660  */
661 #define	NODE_WWN_PROP	(char *)fcp_node_wwn_prop
662 #define	PORT_WWN_PROP	(char *)fcp_port_wwn_prop
663 #define	TARGET_PROP	(char *)fcp_target_prop
664 #define	LUN_PROP	(char *)fcp_lun_prop
665 #define	SAM_LUN_PROP	(char *)fcp_sam_lun_prop
666 #define	CONF_WWN_PROP	(char *)fcp_conf_wwn_prop
667 #define	OBP_BOOT_WWN	(char *)fcp_obp_boot_wwn
668 #define	MANUAL_CFG_ONLY	(char *)fcp_manual_config_only
669 #define	INIT_PORT_PROP	(char *)fcp_init_port_prop
670 #define	TGT_PORT_PROP	(char *)fcp_tgt_port_prop
671 #define	LUN_BLACKLIST_PROP	(char *)fcp_lun_blacklist_prop
672 /*
673  * Short hand macros.
674  */
675 #define	LUN_PORT	(plun->lun_tgt->tgt_port)
676 #define	LUN_TGT		(plun->lun_tgt)
677 
678 /*
679  * Driver private macros
680  */
681 #define	FCP_ATOB(x)	(((x) >= '0' && (x) <= '9') ? ((x) - '0') :	\
682 			((x) >= 'a' && (x) <= 'f') ?			\
683 			((x) - 'a' + 10) : ((x) - 'A' + 10))
684 
685 #define	FCP_MAX(a, b)	((a) > (b) ? (a) : (b))
686 
687 #define	FCP_N_NDI_EVENTS						\
688 	(sizeof (fcp_ndi_event_defs) / sizeof (ndi_event_definition_t))
689 
690 #define	FCP_LINK_STATE_CHANGED(p, c)			\
691 	((p)->port_link_cnt != (c)->ipkt_link_cnt)
692 
693 #define	FCP_TGT_STATE_CHANGED(t, c)			\
694 	((t)->tgt_change_cnt != (c)->ipkt_change_cnt)
695 
696 #define	FCP_STATE_CHANGED(p, t, c)		\
697 	(FCP_TGT_STATE_CHANGED(t, c))
698 
699 #define	FCP_MUST_RETRY(fpkt)				\
700 	((fpkt)->pkt_state == FC_PKT_LOCAL_BSY ||	\
701 	(fpkt)->pkt_state == FC_PKT_LOCAL_RJT ||	\
702 	(fpkt)->pkt_state == FC_PKT_TRAN_BSY ||	\
703 	(fpkt)->pkt_state == FC_PKT_ELS_IN_PROGRESS ||	\
704 	(fpkt)->pkt_state == FC_PKT_NPORT_BSY ||	\
705 	(fpkt)->pkt_state == FC_PKT_FABRIC_BSY ||	\
706 	(fpkt)->pkt_state == FC_PKT_PORT_OFFLINE ||	\
707 	(fpkt)->pkt_reason == FC_REASON_OFFLINE)
708 
709 #define	FCP_SENSE_REPORTLUN_CHANGED(es)		\
710 	((es)->es_key == KEY_UNIT_ATTENTION &&	\
711 	(es)->es_add_code == 0x3f &&		\
712 	(es)->es_qual_code == 0x0e)
713 
714 #define	FCP_SENSE_NO_LUN(es)			\
715 	((es)->es_key == KEY_ILLEGAL_REQUEST &&	\
716 	(es)->es_add_code == 0x25 &&		\
717 	(es)->es_qual_code == 0x0)
718 
719 #define	FCP_VERSION		"1.187"
720 #define	FCP_NAME_VERSION	"SunFC FCP v" FCP_VERSION
721 
722 #define	FCP_NUM_ELEMENTS(array)			\
723 	(sizeof (array) / sizeof ((array)[0]))
724 
725 /*
726  * Debugging, Error reporting, and tracing
727  */
728 #define	FCP_LOG_SIZE		1024 * 1024
729 
730 #define	FCP_LEVEL_1		0x00001		/* attach/detach PM CPR */
731 #define	FCP_LEVEL_2		0x00002		/* failures/Invalid data */
732 #define	FCP_LEVEL_3		0x00004		/* state change, discovery */
733 #define	FCP_LEVEL_4		0x00008		/* ULP messages */
734 #define	FCP_LEVEL_5		0x00010		/* ELS/SCSI cmds */
735 #define	FCP_LEVEL_6		0x00020		/* Transport failures */
736 #define	FCP_LEVEL_7		0x00040
737 #define	FCP_LEVEL_8		0x00080		/* I/O tracing */
738 #define	FCP_LEVEL_9		0x00100		/* I/O tracing */
739 
740 
741 
742 /*
743  * Log contents to system messages file
744  */
745 #define	FCP_MSG_LEVEL_1	(FCP_LEVEL_1 | FC_TRACE_LOG_MSG)
746 #define	FCP_MSG_LEVEL_2	(FCP_LEVEL_2 | FC_TRACE_LOG_MSG)
747 #define	FCP_MSG_LEVEL_3	(FCP_LEVEL_3 | FC_TRACE_LOG_MSG)
748 #define	FCP_MSG_LEVEL_4	(FCP_LEVEL_4 | FC_TRACE_LOG_MSG)
749 #define	FCP_MSG_LEVEL_5	(FCP_LEVEL_5 | FC_TRACE_LOG_MSG)
750 #define	FCP_MSG_LEVEL_6	(FCP_LEVEL_6 | FC_TRACE_LOG_MSG)
751 #define	FCP_MSG_LEVEL_7	(FCP_LEVEL_7 | FC_TRACE_LOG_MSG)
752 #define	FCP_MSG_LEVEL_8	(FCP_LEVEL_8 | FC_TRACE_LOG_MSG)
753 #define	FCP_MSG_LEVEL_9	(FCP_LEVEL_9 | FC_TRACE_LOG_MSG)
754 
755 
756 /*
757  * Log contents to trace buffer
758  */
759 #define	FCP_BUF_LEVEL_1	(FCP_LEVEL_1 | FC_TRACE_LOG_BUF)
760 #define	FCP_BUF_LEVEL_2	(FCP_LEVEL_2 | FC_TRACE_LOG_BUF)
761 #define	FCP_BUF_LEVEL_3	(FCP_LEVEL_3 | FC_TRACE_LOG_BUF)
762 #define	FCP_BUF_LEVEL_4	(FCP_LEVEL_4 | FC_TRACE_LOG_BUF)
763 #define	FCP_BUF_LEVEL_5	(FCP_LEVEL_5 | FC_TRACE_LOG_BUF)
764 #define	FCP_BUF_LEVEL_6	(FCP_LEVEL_6 | FC_TRACE_LOG_BUF)
765 #define	FCP_BUF_LEVEL_7	(FCP_LEVEL_7 | FC_TRACE_LOG_BUF)
766 #define	FCP_BUF_LEVEL_8	(FCP_LEVEL_8 | FC_TRACE_LOG_BUF)
767 #define	FCP_BUF_LEVEL_9	(FCP_LEVEL_9 | FC_TRACE_LOG_BUF)
768 
769 
770 /*
771  * Log contents to both system messages file and trace buffer
772  */
773 #define	FCP_MSG_BUF_LEVEL_1	(FCP_LEVEL_1 | FC_TRACE_LOG_BUF |	\
774 				FC_TRACE_LOG_MSG)
775 #define	FCP_MSG_BUF_LEVEL_2	(FCP_LEVEL_2 | FC_TRACE_LOG_BUF |	\
776 				FC_TRACE_LOG_MSG)
777 #define	FCP_MSG_BUF_LEVEL_3	(FCP_LEVEL_3 | FC_TRACE_LOG_BUF |	\
778 				FC_TRACE_LOG_MSG)
779 #define	FCP_MSG_BUF_LEVEL_4	(FCP_LEVEL_4 | FC_TRACE_LOG_BUF |	\
780 				FC_TRACE_LOG_MSG)
781 #define	FCP_MSG_BUF_LEVEL_5	(FCP_LEVEL_5 | FC_TRACE_LOG_BUF |	\
782 				FC_TRACE_LOG_MSG)
783 #define	FCP_MSG_BUF_LEVEL_6	(FCP_LEVEL_6 | FC_TRACE_LOG_BUF |	\
784 				FC_TRACE_LOG_MSG)
785 #define	FCP_MSG_BUF_LEVEL_7	(FCP_LEVEL_7 | FC_TRACE_LOG_BUF |	\
786 				FC_TRACE_LOG_MSG)
787 #define	FCP_MSG_BUF_LEVEL_8	(FCP_LEVEL_8 | FC_TRACE_LOG_BUF |	\
788 				FC_TRACE_LOG_MSG)
789 #define	FCP_MSG_BUF_LEVEL_9	(FCP_LEVEL_9 | FC_TRACE_LOG_BUF |	\
790 				FC_TRACE_LOG_MSG)
791 #ifdef DEBUG
792 #define	FCP_DTRACE	fc_trace_debug
793 #else
794 #define	FCP_DTRACE
795 #endif
796 
797 #define	FCP_TRACE	fc_trace_debug
798 
799 static struct cb_ops fcp_cb_ops = {
800 	fcp_open,			/* open */
801 	fcp_close,			/* close */
802 	nodev,				/* strategy */
803 	nodev,				/* print */
804 	nodev,				/* dump */
805 	nodev,				/* read */
806 	nodev,				/* write */
807 	fcp_ioctl,			/* ioctl */
808 	nodev,				/* devmap */
809 	nodev,				/* mmap */
810 	nodev,				/* segmap */
811 	nochpoll,			/* chpoll */
812 	ddi_prop_op,			/* cb_prop_op */
813 	0,				/* streamtab */
814 	D_NEW | D_MP | D_HOTPLUG,	/* cb_flag */
815 	CB_REV,				/* rev */
816 	nodev,				/* aread */
817 	nodev				/* awrite */
818 };
819 
820 
821 static struct dev_ops fcp_ops = {
822 	DEVO_REV,
823 	0,
824 	ddi_getinfo_1to1,
825 	nulldev,		/* identify */
826 	nulldev,		/* probe */
827 	fcp_attach,		/* attach and detach are mandatory */
828 	fcp_detach,
829 	nodev,			/* reset */
830 	&fcp_cb_ops,		/* cb_ops */
831 	NULL,			/* bus_ops */
832 	NULL,			/* power */
833 };
834 
835 
836 char *fcp_version = FCP_NAME_VERSION;
837 
838 static struct modldrv modldrv = {
839 	&mod_driverops,
840 	FCP_NAME_VERSION,
841 	&fcp_ops
842 };
843 
844 
845 static struct modlinkage modlinkage = {
846 	MODREV_1,
847 	&modldrv,
848 	NULL
849 };
850 
851 
852 static fc_ulp_modinfo_t fcp_modinfo = {
853 	&fcp_modinfo,			/* ulp_handle */
854 	FCTL_ULP_MODREV_4,		/* ulp_rev */
855 	FC4_SCSI_FCP,			/* ulp_type */
856 	"fcp",				/* ulp_name */
857 	FCP_STATEC_MASK,		/* ulp_statec_mask */
858 	fcp_port_attach,		/* ulp_port_attach */
859 	fcp_port_detach,		/* ulp_port_detach */
860 	fcp_port_ioctl,			/* ulp_port_ioctl */
861 	fcp_els_callback,		/* ulp_els_callback */
862 	fcp_data_callback,		/* ulp_data_callback */
863 	fcp_statec_callback		/* ulp_statec_callback */
864 };
865 
866 #ifdef	DEBUG
867 #define	FCP_TRACE_DEFAULT	(FC_TRACE_LOG_MASK | FCP_LEVEL_1 |	\
868 				FCP_LEVEL_2 | FCP_LEVEL_3 |		\
869 				FCP_LEVEL_4 | FCP_LEVEL_5 |		\
870 				FCP_LEVEL_6 | FCP_LEVEL_7)
871 #else
872 #define	FCP_TRACE_DEFAULT	(FC_TRACE_LOG_MASK | FCP_LEVEL_1 |	\
873 				FCP_LEVEL_2 | FCP_LEVEL_3 |		\
874 				FCP_LEVEL_4 | FCP_LEVEL_5 |		\
875 				FCP_LEVEL_6 | FCP_LEVEL_7)
876 #endif
877 
878 /* FCP global variables */
879 int			fcp_bus_config_debug = 0;
880 static int		fcp_log_size = FCP_LOG_SIZE;
881 static int		fcp_trace = FCP_TRACE_DEFAULT;
882 static fc_trace_logq_t	*fcp_logq = NULL;
883 static struct fcp_black_list_entry	*fcp_lun_blacklist = NULL;
884 /*
885  * The auto-configuration is set by default.  The only way of disabling it is
886  * through the property MANUAL_CFG_ONLY in the fcp.conf file.
887  */
888 static int		fcp_enable_auto_configuration = 1;
889 static int		fcp_max_bus_config_retries	= 4;
890 static int		fcp_lun_ready_retry = 300;
891 /*
892  * The value assigned to the following variable has changed several times due
893  * to a problem with the data underruns reporting of some firmware(s).	The
894  * current value of 50 gives a timeout value of 25 seconds for a max number
895  * of 256 LUNs.
896  */
897 static int		fcp_max_target_retries = 50;
898 /*
899  * Watchdog variables
900  * ------------------
901  *
902  * fcp_watchdog_init
903  *
904  *	Indicates if the watchdog timer is running or not.  This is actually
905  *	a counter of the number of Fibre Channel ports that attached.  When
906  *	the first port attaches the watchdog is started.  When the last port
907  *	detaches the watchdog timer is stopped.
908  *
909  * fcp_watchdog_time
910  *
911  *	This is the watchdog clock counter.  It is incremented by
912  *	fcp_watchdog_time each time the watchdog timer expires.
913  *
914  * fcp_watchdog_timeout
915  *
916  *	Increment value of the variable fcp_watchdog_time as well as the
917  *	the timeout value of the watchdog timer.  The unit is 1 second.	 It
918  *	is strange that this is not a #define	but a variable since the code
919  *	never changes this value.  The reason why it can be said that the
920  *	unit is 1 second is because the number of ticks for the watchdog
921  *	timer is determined like this:
922  *
923  *	    fcp_watchdog_tick = fcp_watchdog_timeout *
924  *				  drv_usectohz(1000000);
925  *
926  *	The value 1000000 is hard coded in the code.
927  *
928  * fcp_watchdog_tick
929  *
930  *	Watchdog timer value in ticks.
931  */
932 static int		fcp_watchdog_init = 0;
933 static int		fcp_watchdog_time = 0;
934 static int		fcp_watchdog_timeout = 1;
935 static int		fcp_watchdog_tick;
936 
937 /*
938  * fcp_offline_delay is a global variable to enable customisation of
939  * the timeout on link offlines or RSCNs. The default value is set
940  * to match FCP_OFFLINE_DELAY (20sec), which is 2*RA_TOV_els as
941  * specified in FCP4 Chapter 11 (see www.t10.org).
942  *
943  * The variable fcp_offline_delay is specified in SECONDS.
944  *
945  * If we made this a static var then the user would not be able to
946  * change it. This variable is set in fcp_attach().
947  */
948 unsigned int		fcp_offline_delay = FCP_OFFLINE_DELAY;
949 
950 static void		*fcp_softstate = NULL; /* for soft state */
951 static uchar_t		fcp_oflag = FCP_IDLE; /* open flag */
952 static kmutex_t		fcp_global_mutex;
953 static kmutex_t		fcp_ioctl_mutex;
954 static dev_info_t	*fcp_global_dip = NULL;
955 static timeout_id_t	fcp_watchdog_id;
956 const char		*fcp_lun_prop = "lun";
957 const char		*fcp_sam_lun_prop = "sam-lun";
958 const char		*fcp_target_prop = "target";
959 /*
960  * NOTE: consumers of "node-wwn" property include stmsboot in ON
961  * consolidation.
962  */
963 const char		*fcp_node_wwn_prop = "node-wwn";
964 const char		*fcp_port_wwn_prop = "port-wwn";
965 const char		*fcp_conf_wwn_prop = "fc-port-wwn";
966 const char		*fcp_obp_boot_wwn = "fc-boot-dev-portwwn";
967 const char		*fcp_manual_config_only = "manual_configuration_only";
968 const char		*fcp_init_port_prop = "initiator-port";
969 const char		*fcp_tgt_port_prop = "target-port";
970 const char		*fcp_lun_blacklist_prop = "pwwn-lun-blacklist";
971 
972 static struct fcp_port	*fcp_port_head = NULL;
973 static ddi_eventcookie_t	fcp_insert_eid;
974 static ddi_eventcookie_t	fcp_remove_eid;
975 
976 static ndi_event_definition_t	fcp_ndi_event_defs[] = {
977 	{ FCP_EVENT_TAG_INSERT, FCAL_INSERT_EVENT, EPL_KERNEL },
978 	{ FCP_EVENT_TAG_REMOVE, FCAL_REMOVE_EVENT, EPL_INTERRUPT }
979 };
980 
981 /*
982  * List of valid commands for the scsi_ioctl call
983  */
984 static uint8_t scsi_ioctl_list[] = {
985 	SCMD_INQUIRY,
986 	SCMD_REPORT_LUN,
987 	SCMD_READ_CAPACITY
988 };
989 
990 /*
991  * this is used to dummy up a report lun response for cases
992  * where the target doesn't support it
993  */
994 static uchar_t fcp_dummy_lun[] = {
995 	0x00,		/* MSB length (length = no of luns * 8) */
996 	0x00,
997 	0x00,
998 	0x08,		/* LSB length */
999 	0x00,		/* MSB reserved */
1000 	0x00,
1001 	0x00,
1002 	0x00,		/* LSB reserved */
1003 	FCP_PD_ADDRESSING,
1004 	0x00,		/* LUN is ZERO at the first level */
1005 	0x00,
1006 	0x00,		/* second level is zero */
1007 	0x00,
1008 	0x00,		/* third level is zero */
1009 	0x00,
1010 	0x00		/* fourth level is zero */
1011 };
1012 
1013 static uchar_t fcp_alpa_to_switch[] = {
1014 	0x00, 0x7d, 0x7c, 0x00, 0x7b, 0x00, 0x00, 0x00, 0x7a, 0x00,
1015 	0x00, 0x00, 0x00, 0x00, 0x00, 0x79, 0x78, 0x00, 0x00, 0x00,
1016 	0x00, 0x00, 0x00, 0x77, 0x76, 0x00, 0x00, 0x75, 0x00, 0x74,
1017 	0x73, 0x72, 0x00, 0x00, 0x00, 0x71, 0x00, 0x70, 0x6f, 0x6e,
1018 	0x00, 0x6d, 0x6c, 0x6b, 0x6a, 0x69, 0x68, 0x00, 0x00, 0x67,
1019 	0x66, 0x65, 0x64, 0x63, 0x62, 0x00, 0x00, 0x61, 0x60, 0x00,
1020 	0x5f, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x5e, 0x00, 0x5d,
1021 	0x5c, 0x5b, 0x00, 0x5a, 0x59, 0x58, 0x57, 0x56, 0x55, 0x00,
1022 	0x00, 0x54, 0x53, 0x52, 0x51, 0x50, 0x4f, 0x00, 0x00, 0x4e,
1023 	0x4d, 0x00, 0x4c, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x4b,
1024 	0x00, 0x4a, 0x49, 0x48, 0x00, 0x47, 0x46, 0x45, 0x44, 0x43,
1025 	0x42, 0x00, 0x00, 0x41, 0x40, 0x3f, 0x3e, 0x3d, 0x3c, 0x00,
1026 	0x00, 0x3b, 0x3a, 0x00, 0x39, 0x00, 0x00, 0x00, 0x38, 0x37,
1027 	0x36, 0x00, 0x35, 0x00, 0x00, 0x00, 0x34, 0x00, 0x00, 0x00,
1028 	0x00, 0x00, 0x00, 0x33, 0x32, 0x00, 0x00, 0x00, 0x00, 0x00,
1029 	0x00, 0x31, 0x30, 0x00, 0x00, 0x2f, 0x00, 0x2e, 0x2d, 0x2c,
1030 	0x00, 0x00, 0x00, 0x2b, 0x00, 0x2a, 0x29, 0x28, 0x00, 0x27,
1031 	0x26, 0x25, 0x24, 0x23, 0x22, 0x00, 0x00, 0x21, 0x20, 0x1f,
1032 	0x1e, 0x1d, 0x1c, 0x00, 0x00, 0x1b, 0x1a, 0x00, 0x19, 0x00,
1033 	0x00, 0x00, 0x00, 0x00, 0x00, 0x18, 0x00, 0x17, 0x16, 0x15,
1034 	0x00, 0x14, 0x13, 0x12, 0x11, 0x10, 0x0f, 0x00, 0x00, 0x0e,
1035 	0x0d, 0x0c, 0x0b, 0x0a, 0x09, 0x00, 0x00, 0x08, 0x07, 0x00,
1036 	0x06, 0x00, 0x00, 0x00, 0x05, 0x04, 0x03, 0x00, 0x02, 0x00,
1037 	0x00, 0x00, 0x01, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00,
1038 };
1039 
1040 static caddr_t pid = "SESS01	      ";
1041 
1042 #if	!defined(lint)
1043 
1044 _NOTE(MUTEX_PROTECTS_DATA(fcp_global_mutex,
1045     fcp_port::fcp_next fcp_watchdog_id))
1046 
1047 _NOTE(DATA_READABLE_WITHOUT_LOCK(fcp_watchdog_time))
1048 
1049 _NOTE(SCHEME_PROTECTS_DATA("Unshared",
1050     fcp_insert_eid
1051     fcp_remove_eid
1052     fcp_watchdog_time))
1053 
1054 _NOTE(SCHEME_PROTECTS_DATA("Unshared",
1055     fcp_cb_ops
1056     fcp_ops
1057     callb_cpr))
1058 
1059 #endif /* lint */
1060 
1061 /*
1062  * This table is used to determine whether or not it's safe to copy in
1063  * the target node name for a lun.  Since all luns behind the same target
1064  * have the same wwnn, only tagets that do not support multiple luns are
1065  * eligible to be enumerated under mpxio if they aren't page83 compliant.
1066  */
1067 
1068 char *fcp_symmetric_disk_table[] = {
1069 	"SEAGATE ST",
1070 	"IBM	 DDYFT",
1071 	"SUNW	 SUNWGS",	/* Daktari enclosure */
1072 	"SUN	 SENA",		/* SES device */
1073 	"SUN	 SESS01"	/* VICOM SVE box */
1074 };
1075 
1076 int fcp_symmetric_disk_table_size =
1077 	sizeof (fcp_symmetric_disk_table)/sizeof (char *);
1078 
1079 /*
1080  * The _init(9e) return value should be that of mod_install(9f). Under
1081  * some circumstances, a failure may not be related mod_install(9f) and
1082  * one would then require a return value to indicate the failure. Looking
1083  * at mod_install(9f), it is expected to return 0 for success and non-zero
1084  * for failure. mod_install(9f) for device drivers, further goes down the
1085  * calling chain and ends up in ddi_installdrv(), whose return values are
1086  * DDI_SUCCESS and DDI_FAILURE - There are also other functions in the
1087  * calling chain of mod_install(9f) which return values like EINVAL and
1088  * in some even return -1.
1089  *
1090  * To work around the vagaries of the mod_install() calling chain, return
1091  * either 0 or ENODEV depending on the success or failure of mod_install()
1092  */
1093 int
1094 _init(void)
1095 {
1096 	int rval;
1097 
1098 	/*
1099 	 * Allocate soft state and prepare to do ddi_soft_state_zalloc()
1100 	 * before registering with the transport first.
1101 	 */
1102 	if (ddi_soft_state_init(&fcp_softstate,
1103 	    sizeof (struct fcp_port), FCP_INIT_ITEMS) != 0) {
1104 		return (EINVAL);
1105 	}
1106 
1107 	mutex_init(&fcp_global_mutex, NULL, MUTEX_DRIVER, NULL);
1108 	mutex_init(&fcp_ioctl_mutex, NULL, MUTEX_DRIVER, NULL);
1109 
1110 	if ((rval = fc_ulp_add(&fcp_modinfo)) != FC_SUCCESS) {
1111 		cmn_err(CE_WARN, "fcp: fc_ulp_add failed");
1112 		mutex_destroy(&fcp_global_mutex);
1113 		mutex_destroy(&fcp_ioctl_mutex);
1114 		ddi_soft_state_fini(&fcp_softstate);
1115 		return (ENODEV);
1116 	}
1117 
1118 	fcp_logq = fc_trace_alloc_logq(fcp_log_size);
1119 
1120 	if ((rval = mod_install(&modlinkage)) != 0) {
1121 		fc_trace_free_logq(fcp_logq);
1122 		(void) fc_ulp_remove(&fcp_modinfo);
1123 		mutex_destroy(&fcp_global_mutex);
1124 		mutex_destroy(&fcp_ioctl_mutex);
1125 		ddi_soft_state_fini(&fcp_softstate);
1126 		rval = ENODEV;
1127 	}
1128 
1129 	return (rval);
1130 }
1131 
1132 
1133 /*
1134  * the system is done with us as a driver, so clean up
1135  */
1136 int
1137 _fini(void)
1138 {
1139 	int rval;
1140 
1141 	/*
1142 	 * don't start cleaning up until we know that the module remove
1143 	 * has worked  -- if this works, then we know that each instance
1144 	 * has successfully been DDI_DETACHed
1145 	 */
1146 	if ((rval = mod_remove(&modlinkage)) != 0) {
1147 		return (rval);
1148 	}
1149 
1150 	(void) fc_ulp_remove(&fcp_modinfo);
1151 
1152 	ddi_soft_state_fini(&fcp_softstate);
1153 	mutex_destroy(&fcp_global_mutex);
1154 	mutex_destroy(&fcp_ioctl_mutex);
1155 	fc_trace_free_logq(fcp_logq);
1156 
1157 	return (rval);
1158 }
1159 
1160 
1161 int
1162 _info(struct modinfo *modinfop)
1163 {
1164 	return (mod_info(&modlinkage, modinfop));
1165 }
1166 
1167 
1168 /*
1169  * attach the module
1170  */
1171 static int
1172 fcp_attach(dev_info_t *devi, ddi_attach_cmd_t cmd)
1173 {
1174 	int rval = DDI_SUCCESS;
1175 
1176 	FCP_DTRACE(fcp_logq, "fcp", fcp_trace,
1177 	    FCP_BUF_LEVEL_8, 0, "fcp module attach: cmd=0x%x", cmd);
1178 
1179 	if (cmd == DDI_ATTACH) {
1180 		/* The FCP pseudo device is created here. */
1181 		mutex_enter(&fcp_global_mutex);
1182 		fcp_global_dip = devi;
1183 		mutex_exit(&fcp_global_mutex);
1184 
1185 		if (ddi_create_minor_node(fcp_global_dip, "fcp", S_IFCHR,
1186 		    0, DDI_PSEUDO, 0) == DDI_SUCCESS) {
1187 			ddi_report_dev(fcp_global_dip);
1188 		} else {
1189 			cmn_err(CE_WARN, "FCP: Cannot create minor node");
1190 			mutex_enter(&fcp_global_mutex);
1191 			fcp_global_dip = NULL;
1192 			mutex_exit(&fcp_global_mutex);
1193 
1194 			rval = DDI_FAILURE;
1195 		}
1196 		/*
1197 		 * We check the fcp_offline_delay property at this
1198 		 * point. This variable is global for the driver,
1199 		 * not specific to an instance.
1200 		 *
1201 		 * We do not recommend setting the value to less
1202 		 * than 10 seconds (RA_TOV_els), or greater than
1203 		 * 60 seconds.
1204 		 */
1205 		fcp_offline_delay = ddi_prop_get_int(DDI_DEV_T_ANY,
1206 		    devi, DDI_PROP_DONTPASS | DDI_PROP_NOTPROM,
1207 		    "fcp_offline_delay", FCP_OFFLINE_DELAY);
1208 		if ((fcp_offline_delay < 10) ||
1209 		    (fcp_offline_delay > 60)) {
1210 			cmn_err(CE_WARN, "Setting fcp_offline_delay "
1211 			    "to %d second(s). This is outside the "
1212 			    "recommended range of 10..60 seconds.",
1213 			    fcp_offline_delay);
1214 		}
1215 	}
1216 
1217 	return (rval);
1218 }
1219 
1220 
1221 /*ARGSUSED*/
1222 static int
1223 fcp_detach(dev_info_t *devi, ddi_detach_cmd_t cmd)
1224 {
1225 	int	res = DDI_SUCCESS;
1226 
1227 	FCP_DTRACE(fcp_logq, "fcp", fcp_trace,
1228 	    FCP_BUF_LEVEL_8, 0,	 "module detach: cmd=0x%x", cmd);
1229 
1230 	if (cmd == DDI_DETACH) {
1231 		/*
1232 		 * Check if there are active ports/threads. If there
1233 		 * are any, we will fail, else we will succeed (there
1234 		 * should not be much to clean up)
1235 		 */
1236 		mutex_enter(&fcp_global_mutex);
1237 		FCP_DTRACE(fcp_logq, "fcp",
1238 		    fcp_trace, FCP_BUF_LEVEL_8, 0,  "port_head=%p",
1239 		    (void *) fcp_port_head);
1240 
1241 		if (fcp_port_head == NULL) {
1242 			ddi_remove_minor_node(fcp_global_dip, NULL);
1243 			fcp_global_dip = NULL;
1244 			mutex_exit(&fcp_global_mutex);
1245 		} else {
1246 			mutex_exit(&fcp_global_mutex);
1247 			res = DDI_FAILURE;
1248 		}
1249 	}
1250 	FCP_DTRACE(fcp_logq, "fcp", fcp_trace,
1251 	    FCP_BUF_LEVEL_8, 0,	 "module detach returning %d", res);
1252 
1253 	return (res);
1254 }
1255 
1256 
1257 /* ARGSUSED */
1258 static int
1259 fcp_open(dev_t *devp, int flag, int otype, cred_t *credp)
1260 {
1261 	if (otype != OTYP_CHR) {
1262 		return (EINVAL);
1263 	}
1264 
1265 	/*
1266 	 * Allow only root to talk;
1267 	 */
1268 	if (drv_priv(credp)) {
1269 		return (EPERM);
1270 	}
1271 
1272 	mutex_enter(&fcp_global_mutex);
1273 	if (fcp_oflag & FCP_EXCL) {
1274 		mutex_exit(&fcp_global_mutex);
1275 		return (EBUSY);
1276 	}
1277 
1278 	if (flag & FEXCL) {
1279 		if (fcp_oflag & FCP_OPEN) {
1280 			mutex_exit(&fcp_global_mutex);
1281 			return (EBUSY);
1282 		}
1283 		fcp_oflag |= FCP_EXCL;
1284 	}
1285 	fcp_oflag |= FCP_OPEN;
1286 	mutex_exit(&fcp_global_mutex);
1287 
1288 	return (0);
1289 }
1290 
1291 
1292 /* ARGSUSED */
1293 static int
1294 fcp_close(dev_t dev, int flag, int otype, cred_t *credp)
1295 {
1296 	if (otype != OTYP_CHR) {
1297 		return (EINVAL);
1298 	}
1299 
1300 	mutex_enter(&fcp_global_mutex);
1301 	if (!(fcp_oflag & FCP_OPEN)) {
1302 		mutex_exit(&fcp_global_mutex);
1303 		return (ENODEV);
1304 	}
1305 	fcp_oflag = FCP_IDLE;
1306 	mutex_exit(&fcp_global_mutex);
1307 
1308 	return (0);
1309 }
1310 
1311 
1312 /*
1313  * fcp_ioctl
1314  *	Entry point for the FCP ioctls
1315  *
1316  * Input:
1317  *	See ioctl(9E)
1318  *
1319  * Output:
1320  *	See ioctl(9E)
1321  *
1322  * Returns:
1323  *	See ioctl(9E)
1324  *
1325  * Context:
1326  *	Kernel context.
1327  */
1328 /* ARGSUSED */
1329 static int
1330 fcp_ioctl(dev_t dev, int cmd, intptr_t data, int mode, cred_t *credp,
1331     int *rval)
1332 {
1333 	int			ret = 0;
1334 
1335 	mutex_enter(&fcp_global_mutex);
1336 	if (!(fcp_oflag & FCP_OPEN)) {
1337 		mutex_exit(&fcp_global_mutex);
1338 		return (ENXIO);
1339 	}
1340 	mutex_exit(&fcp_global_mutex);
1341 
1342 	switch (cmd) {
1343 	case FCP_TGT_INQUIRY:
1344 	case FCP_TGT_CREATE:
1345 	case FCP_TGT_DELETE:
1346 		ret = fcp_setup_device_data_ioctl(cmd,
1347 		    (struct fcp_ioctl *)data, mode, rval);
1348 		break;
1349 
1350 	case FCP_TGT_SEND_SCSI:
1351 		mutex_enter(&fcp_ioctl_mutex);
1352 		ret = fcp_setup_scsi_ioctl(
1353 		    (struct fcp_scsi_cmd *)data, mode, rval);
1354 		mutex_exit(&fcp_ioctl_mutex);
1355 		break;
1356 
1357 	case FCP_STATE_COUNT:
1358 		ret = fcp_get_statec_count((struct fcp_ioctl *)data,
1359 		    mode, rval);
1360 		break;
1361 	case FCP_GET_TARGET_MAPPINGS:
1362 		ret = fcp_get_target_mappings((struct fcp_ioctl *)data,
1363 		    mode, rval);
1364 		break;
1365 	default:
1366 		fcp_log(CE_WARN, NULL,
1367 		    "!Invalid ioctl opcode = 0x%x", cmd);
1368 		ret	= EINVAL;
1369 	}
1370 
1371 	return (ret);
1372 }
1373 
1374 
1375 /*
1376  * fcp_setup_device_data_ioctl
1377  *	Setup handler for the "device data" style of
1378  *	ioctl for FCP.	See "fcp_util.h" for data structure
1379  *	definition.
1380  *
1381  * Input:
1382  *	cmd	= FCP ioctl command
1383  *	data	= ioctl data
1384  *	mode	= See ioctl(9E)
1385  *
1386  * Output:
1387  *	data	= ioctl data
1388  *	rval	= return value - see ioctl(9E)
1389  *
1390  * Returns:
1391  *	See ioctl(9E)
1392  *
1393  * Context:
1394  *	Kernel context.
1395  */
1396 /* ARGSUSED */
1397 static int
1398 fcp_setup_device_data_ioctl(int cmd, struct fcp_ioctl *data, int mode,
1399     int *rval)
1400 {
1401 	struct fcp_port	*pptr;
1402 	struct	device_data	*dev_data;
1403 	uint32_t		link_cnt;
1404 	la_wwn_t		*wwn_ptr = NULL;
1405 	struct fcp_tgt		*ptgt = NULL;
1406 	struct fcp_lun		*plun = NULL;
1407 	int			i, error;
1408 	struct fcp_ioctl	fioctl;
1409 
1410 #ifdef	_MULTI_DATAMODEL
1411 	switch (ddi_model_convert_from(mode & FMODELS)) {
1412 	case DDI_MODEL_ILP32: {
1413 		struct fcp32_ioctl f32_ioctl;
1414 
1415 		if (ddi_copyin((void *)data, (void *)&f32_ioctl,
1416 		    sizeof (struct fcp32_ioctl), mode)) {
1417 			return (EFAULT);
1418 		}
1419 		fioctl.fp_minor = f32_ioctl.fp_minor;
1420 		fioctl.listlen = f32_ioctl.listlen;
1421 		fioctl.list = (caddr_t)(long)f32_ioctl.list;
1422 		break;
1423 	}
1424 	case DDI_MODEL_NONE:
1425 		if (ddi_copyin((void *)data, (void *)&fioctl,
1426 		    sizeof (struct fcp_ioctl), mode)) {
1427 			return (EFAULT);
1428 		}
1429 		break;
1430 	}
1431 
1432 #else	/* _MULTI_DATAMODEL */
1433 	if (ddi_copyin((void *)data, (void *)&fioctl,
1434 	    sizeof (struct fcp_ioctl), mode)) {
1435 		return (EFAULT);
1436 	}
1437 #endif	/* _MULTI_DATAMODEL */
1438 
1439 	/*
1440 	 * Right now we can assume that the minor number matches with
1441 	 * this instance of fp. If this changes we will need to
1442 	 * revisit this logic.
1443 	 */
1444 	mutex_enter(&fcp_global_mutex);
1445 	pptr = fcp_port_head;
1446 	while (pptr) {
1447 		if (pptr->port_instance == (uint32_t)fioctl.fp_minor) {
1448 			break;
1449 		} else {
1450 			pptr = pptr->port_next;
1451 		}
1452 	}
1453 	mutex_exit(&fcp_global_mutex);
1454 	if (pptr == NULL) {
1455 		return (ENXIO);
1456 	}
1457 	mutex_enter(&pptr->port_mutex);
1458 
1459 
1460 	if ((dev_data = kmem_zalloc((sizeof (struct device_data)) *
1461 	    fioctl.listlen, KM_NOSLEEP)) == NULL) {
1462 		mutex_exit(&pptr->port_mutex);
1463 		return (ENOMEM);
1464 	}
1465 
1466 	if (ddi_copyin(fioctl.list, dev_data,
1467 	    (sizeof (struct device_data)) * fioctl.listlen, mode)) {
1468 		kmem_free(dev_data, sizeof (*dev_data) * fioctl.listlen);
1469 		mutex_exit(&pptr->port_mutex);
1470 		return (EFAULT);
1471 	}
1472 	link_cnt = pptr->port_link_cnt;
1473 
1474 	if (cmd == FCP_TGT_INQUIRY) {
1475 		wwn_ptr = (la_wwn_t *)&(dev_data[0].dev_pwwn);
1476 		if (bcmp(wwn_ptr->raw_wwn, pptr->port_pwwn.raw_wwn,
1477 		    sizeof (wwn_ptr->raw_wwn)) == 0) {
1478 			/* This ioctl is requesting INQ info of local HBA */
1479 			mutex_exit(&pptr->port_mutex);
1480 			dev_data[0].dev0_type = DTYPE_UNKNOWN;
1481 			dev_data[0].dev_status = 0;
1482 			if (ddi_copyout(dev_data, fioctl.list,
1483 			    (sizeof (struct device_data)) * fioctl.listlen,
1484 			    mode)) {
1485 				kmem_free(dev_data,
1486 				    sizeof (*dev_data) * fioctl.listlen);
1487 				return (EFAULT);
1488 			}
1489 			kmem_free(dev_data,
1490 			    sizeof (*dev_data) * fioctl.listlen);
1491 #ifdef	_MULTI_DATAMODEL
1492 			switch (ddi_model_convert_from(mode & FMODELS)) {
1493 			case DDI_MODEL_ILP32: {
1494 				struct fcp32_ioctl f32_ioctl;
1495 				f32_ioctl.fp_minor = fioctl.fp_minor;
1496 				f32_ioctl.listlen = fioctl.listlen;
1497 				f32_ioctl.list = (caddr32_t)(long)fioctl.list;
1498 				if (ddi_copyout((void *)&f32_ioctl,
1499 				    (void *)data,
1500 				    sizeof (struct fcp32_ioctl), mode)) {
1501 					return (EFAULT);
1502 				}
1503 				break;
1504 			}
1505 			case DDI_MODEL_NONE:
1506 				if (ddi_copyout((void *)&fioctl, (void *)data,
1507 				    sizeof (struct fcp_ioctl), mode)) {
1508 					return (EFAULT);
1509 				}
1510 				break;
1511 			}
1512 #else	/* _MULTI_DATAMODEL */
1513 			if (ddi_copyout((void *)&fioctl, (void *)data,
1514 			    sizeof (struct fcp_ioctl), mode)) {
1515 				return (EFAULT);
1516 			}
1517 #endif	/* _MULTI_DATAMODEL */
1518 			return (0);
1519 		}
1520 	}
1521 
1522 	if (pptr->port_state & (FCP_STATE_INIT | FCP_STATE_OFFLINE)) {
1523 		kmem_free(dev_data, sizeof (*dev_data) * fioctl.listlen);
1524 		mutex_exit(&pptr->port_mutex);
1525 		return (ENXIO);
1526 	}
1527 
1528 	for (i = 0; (i < fioctl.listlen) && (link_cnt == pptr->port_link_cnt);
1529 	    i++) {
1530 		wwn_ptr = (la_wwn_t *)&(dev_data[i].dev_pwwn);
1531 
1532 		dev_data[i].dev0_type = DTYPE_UNKNOWN;
1533 
1534 
1535 		dev_data[i].dev_status = ENXIO;
1536 
1537 		if ((ptgt = fcp_lookup_target(pptr,
1538 		    (uchar_t *)wwn_ptr)) == NULL) {
1539 			mutex_exit(&pptr->port_mutex);
1540 			if (fc_ulp_get_remote_port(pptr->port_fp_handle,
1541 			    wwn_ptr, &error, 0) == NULL) {
1542 				dev_data[i].dev_status = ENODEV;
1543 				mutex_enter(&pptr->port_mutex);
1544 				continue;
1545 			} else {
1546 
1547 				dev_data[i].dev_status = EAGAIN;
1548 
1549 				mutex_enter(&pptr->port_mutex);
1550 				continue;
1551 			}
1552 		} else {
1553 			mutex_enter(&ptgt->tgt_mutex);
1554 			if (ptgt->tgt_state & (FCP_TGT_MARK |
1555 			    FCP_TGT_BUSY)) {
1556 				dev_data[i].dev_status = EAGAIN;
1557 				mutex_exit(&ptgt->tgt_mutex);
1558 				continue;
1559 			}
1560 
1561 			if (ptgt->tgt_state & FCP_TGT_OFFLINE) {
1562 				if (ptgt->tgt_icap && !ptgt->tgt_tcap) {
1563 					dev_data[i].dev_status = ENOTSUP;
1564 				} else {
1565 					dev_data[i].dev_status = ENXIO;
1566 				}
1567 				mutex_exit(&ptgt->tgt_mutex);
1568 				continue;
1569 			}
1570 
1571 			switch (cmd) {
1572 			case FCP_TGT_INQUIRY:
1573 				/*
1574 				 * The reason we give device type of
1575 				 * lun 0 only even though in some
1576 				 * cases(like maxstrat) lun 0 device
1577 				 * type may be 0x3f(invalid) is that
1578 				 * for bridge boxes target will appear
1579 				 * as luns and the first lun could be
1580 				 * a device that utility may not care
1581 				 * about (like a tape device).
1582 				 */
1583 				dev_data[i].dev_lun_cnt = ptgt->tgt_lun_cnt;
1584 				dev_data[i].dev_status = 0;
1585 				mutex_exit(&ptgt->tgt_mutex);
1586 
1587 				if ((plun = fcp_get_lun(ptgt, 0)) == NULL) {
1588 					dev_data[i].dev0_type = DTYPE_UNKNOWN;
1589 				} else {
1590 					dev_data[i].dev0_type = plun->lun_type;
1591 				}
1592 				mutex_enter(&ptgt->tgt_mutex);
1593 				break;
1594 
1595 			case FCP_TGT_CREATE:
1596 				mutex_exit(&ptgt->tgt_mutex);
1597 				mutex_exit(&pptr->port_mutex);
1598 
1599 				/*
1600 				 * serialize state change call backs.
1601 				 * only one call back will be handled
1602 				 * at a time.
1603 				 */
1604 				mutex_enter(&fcp_global_mutex);
1605 				if (fcp_oflag & FCP_BUSY) {
1606 					mutex_exit(&fcp_global_mutex);
1607 					if (dev_data) {
1608 						kmem_free(dev_data,
1609 						    sizeof (*dev_data) *
1610 						    fioctl.listlen);
1611 					}
1612 					return (EBUSY);
1613 				}
1614 				fcp_oflag |= FCP_BUSY;
1615 				mutex_exit(&fcp_global_mutex);
1616 
1617 				dev_data[i].dev_status =
1618 				    fcp_create_on_demand(pptr,
1619 				    wwn_ptr->raw_wwn);
1620 
1621 				if (dev_data[i].dev_status != 0) {
1622 					char	buf[25];
1623 
1624 					for (i = 0; i < FC_WWN_SIZE; i++) {
1625 						(void) sprintf(&buf[i << 1],
1626 						    "%02x",
1627 						    wwn_ptr->raw_wwn[i]);
1628 					}
1629 
1630 					fcp_log(CE_WARN, pptr->port_dip,
1631 					    "!Failed to create nodes for"
1632 					    " pwwn=%s; error=%x", buf,
1633 					    dev_data[i].dev_status);
1634 				}
1635 
1636 				/* allow state change call backs again */
1637 				mutex_enter(&fcp_global_mutex);
1638 				fcp_oflag &= ~FCP_BUSY;
1639 				mutex_exit(&fcp_global_mutex);
1640 
1641 				mutex_enter(&pptr->port_mutex);
1642 				mutex_enter(&ptgt->tgt_mutex);
1643 
1644 				break;
1645 
1646 			case FCP_TGT_DELETE:
1647 				break;
1648 
1649 			default:
1650 				fcp_log(CE_WARN, pptr->port_dip,
1651 				    "!Invalid device data ioctl "
1652 				    "opcode = 0x%x", cmd);
1653 			}
1654 			mutex_exit(&ptgt->tgt_mutex);
1655 		}
1656 	}
1657 	mutex_exit(&pptr->port_mutex);
1658 
1659 	if (ddi_copyout(dev_data, fioctl.list,
1660 	    (sizeof (struct device_data)) * fioctl.listlen, mode)) {
1661 		kmem_free(dev_data, sizeof (*dev_data) * fioctl.listlen);
1662 		return (EFAULT);
1663 	}
1664 	kmem_free(dev_data, sizeof (*dev_data) * fioctl.listlen);
1665 
1666 #ifdef	_MULTI_DATAMODEL
1667 	switch (ddi_model_convert_from(mode & FMODELS)) {
1668 	case DDI_MODEL_ILP32: {
1669 		struct fcp32_ioctl f32_ioctl;
1670 
1671 		f32_ioctl.fp_minor = fioctl.fp_minor;
1672 		f32_ioctl.listlen = fioctl.listlen;
1673 		f32_ioctl.list = (caddr32_t)(long)fioctl.list;
1674 		if (ddi_copyout((void *)&f32_ioctl, (void *)data,
1675 		    sizeof (struct fcp32_ioctl), mode)) {
1676 			return (EFAULT);
1677 		}
1678 		break;
1679 	}
1680 	case DDI_MODEL_NONE:
1681 		if (ddi_copyout((void *)&fioctl, (void *)data,
1682 		    sizeof (struct fcp_ioctl), mode)) {
1683 			return (EFAULT);
1684 		}
1685 		break;
1686 	}
1687 #else	/* _MULTI_DATAMODEL */
1688 
1689 	if (ddi_copyout((void *)&fioctl, (void *)data,
1690 	    sizeof (struct fcp_ioctl), mode)) {
1691 		return (EFAULT);
1692 	}
1693 #endif	/* _MULTI_DATAMODEL */
1694 
1695 	return (0);
1696 }
1697 
1698 /*
1699  * Fetch the target mappings (path, etc.) for all LUNs
1700  * on this port.
1701  */
1702 /* ARGSUSED */
1703 static int
1704 fcp_get_target_mappings(struct fcp_ioctl *data,
1705     int mode, int *rval)
1706 {
1707 	struct fcp_port	    *pptr;
1708 	fc_hba_target_mappings_t    *mappings;
1709 	fc_hba_mapping_entry_t	    *map;
1710 	struct fcp_tgt	    *ptgt = NULL;
1711 	struct fcp_lun	    *plun = NULL;
1712 	int			    i, mapIndex, mappingSize;
1713 	int			    listlen;
1714 	struct fcp_ioctl	    fioctl;
1715 	char			    *path;
1716 	fcp_ent_addr_t		    sam_lun_addr;
1717 
1718 #ifdef	_MULTI_DATAMODEL
1719 	switch (ddi_model_convert_from(mode & FMODELS)) {
1720 	case DDI_MODEL_ILP32: {
1721 		struct fcp32_ioctl f32_ioctl;
1722 
1723 		if (ddi_copyin((void *)data, (void *)&f32_ioctl,
1724 		    sizeof (struct fcp32_ioctl), mode)) {
1725 			return (EFAULT);
1726 		}
1727 		fioctl.fp_minor = f32_ioctl.fp_minor;
1728 		fioctl.listlen = f32_ioctl.listlen;
1729 		fioctl.list = (caddr_t)(long)f32_ioctl.list;
1730 		break;
1731 	}
1732 	case DDI_MODEL_NONE:
1733 		if (ddi_copyin((void *)data, (void *)&fioctl,
1734 		    sizeof (struct fcp_ioctl), mode)) {
1735 			return (EFAULT);
1736 		}
1737 		break;
1738 	}
1739 
1740 #else	/* _MULTI_DATAMODEL */
1741 	if (ddi_copyin((void *)data, (void *)&fioctl,
1742 	    sizeof (struct fcp_ioctl), mode)) {
1743 		return (EFAULT);
1744 	}
1745 #endif	/* _MULTI_DATAMODEL */
1746 
1747 	/*
1748 	 * Right now we can assume that the minor number matches with
1749 	 * this instance of fp. If this changes we will need to
1750 	 * revisit this logic.
1751 	 */
1752 	mutex_enter(&fcp_global_mutex);
1753 	pptr = fcp_port_head;
1754 	while (pptr) {
1755 		if (pptr->port_instance == (uint32_t)fioctl.fp_minor) {
1756 			break;
1757 		} else {
1758 			pptr = pptr->port_next;
1759 		}
1760 	}
1761 	mutex_exit(&fcp_global_mutex);
1762 	if (pptr == NULL) {
1763 		cmn_err(CE_NOTE, "target mappings: unknown instance number: %d",
1764 		    fioctl.fp_minor);
1765 		return (ENXIO);
1766 	}
1767 
1768 
1769 	/* We use listlen to show the total buffer size */
1770 	mappingSize = fioctl.listlen;
1771 
1772 	/* Now calculate how many mapping entries will fit */
1773 	listlen = fioctl.listlen + sizeof (fc_hba_mapping_entry_t)
1774 	    - sizeof (fc_hba_target_mappings_t);
1775 	if (listlen <= 0) {
1776 		cmn_err(CE_NOTE, "target mappings: Insufficient buffer");
1777 		return (ENXIO);
1778 	}
1779 	listlen = listlen / sizeof (fc_hba_mapping_entry_t);
1780 
1781 	if ((mappings = kmem_zalloc(mappingSize, KM_SLEEP)) == NULL) {
1782 		return (ENOMEM);
1783 	}
1784 	mappings->version = FC_HBA_TARGET_MAPPINGS_VERSION;
1785 
1786 	/* Now get to work */
1787 	mapIndex = 0;
1788 
1789 	mutex_enter(&pptr->port_mutex);
1790 	/* Loop through all targets on this port */
1791 	for (i = 0; i < FCP_NUM_HASH; i++) {
1792 		for (ptgt = pptr->port_tgt_hash_table[i]; ptgt != NULL;
1793 		    ptgt = ptgt->tgt_next) {
1794 
1795 
1796 			/* Loop through all LUNs on this target */
1797 			for (plun = ptgt->tgt_lun; plun != NULL;
1798 			    plun = plun->lun_next) {
1799 				if (plun->lun_state & FCP_LUN_OFFLINE) {
1800 					continue;
1801 				}
1802 
1803 				path = fcp_get_lun_path(plun);
1804 				if (path == NULL) {
1805 					continue;
1806 				}
1807 
1808 				if (mapIndex >= listlen) {
1809 					mapIndex ++;
1810 					kmem_free(path, MAXPATHLEN);
1811 					continue;
1812 				}
1813 				map = &mappings->entries[mapIndex++];
1814 				bcopy(path, map->targetDriver,
1815 				    sizeof (map->targetDriver));
1816 				map->d_id = ptgt->tgt_d_id;
1817 				map->busNumber = 0;
1818 				map->targetNumber = ptgt->tgt_d_id;
1819 				map->osLUN = plun->lun_num;
1820 
1821 				/*
1822 				 * We had swapped lun when we stored it in
1823 				 * lun_addr. We need to swap it back before
1824 				 * returning it to user land
1825 				 */
1826 
1827 				sam_lun_addr.ent_addr_0 =
1828 				    BE_16(plun->lun_addr.ent_addr_0);
1829 				sam_lun_addr.ent_addr_1 =
1830 				    BE_16(plun->lun_addr.ent_addr_1);
1831 				sam_lun_addr.ent_addr_2 =
1832 				    BE_16(plun->lun_addr.ent_addr_2);
1833 				sam_lun_addr.ent_addr_3 =
1834 				    BE_16(plun->lun_addr.ent_addr_3);
1835 
1836 				bcopy(&sam_lun_addr, &map->samLUN,
1837 				    FCP_LUN_SIZE);
1838 				bcopy(ptgt->tgt_node_wwn.raw_wwn,
1839 				    map->NodeWWN.raw_wwn, sizeof (la_wwn_t));
1840 				bcopy(ptgt->tgt_port_wwn.raw_wwn,
1841 				    map->PortWWN.raw_wwn, sizeof (la_wwn_t));
1842 
1843 				if (plun->lun_guid) {
1844 
1845 					/* convert ascii wwn to bytes */
1846 					fcp_ascii_to_wwn(plun->lun_guid,
1847 					    map->guid, sizeof (map->guid));
1848 
1849 					if ((sizeof (map->guid)) <
1850 					    plun->lun_guid_size / 2) {
1851 						cmn_err(CE_WARN,
1852 						    "fcp_get_target_mappings:"
1853 						    "guid copy space "
1854 						    "insufficient."
1855 						    "Copy Truncation - "
1856 						    "available %d; need %d",
1857 						    (int)sizeof (map->guid),
1858 						    (int)
1859 						    plun->lun_guid_size / 2);
1860 					}
1861 				}
1862 				kmem_free(path, MAXPATHLEN);
1863 			}
1864 		}
1865 	}
1866 	mutex_exit(&pptr->port_mutex);
1867 	mappings->numLuns = mapIndex;
1868 
1869 	if (ddi_copyout(mappings, fioctl.list, mappingSize, mode)) {
1870 		kmem_free(mappings, mappingSize);
1871 		return (EFAULT);
1872 	}
1873 	kmem_free(mappings, mappingSize);
1874 
1875 #ifdef	_MULTI_DATAMODEL
1876 	switch (ddi_model_convert_from(mode & FMODELS)) {
1877 	case DDI_MODEL_ILP32: {
1878 		struct fcp32_ioctl f32_ioctl;
1879 
1880 		f32_ioctl.fp_minor = fioctl.fp_minor;
1881 		f32_ioctl.listlen = fioctl.listlen;
1882 		f32_ioctl.list = (caddr32_t)(long)fioctl.list;
1883 		if (ddi_copyout((void *)&f32_ioctl, (void *)data,
1884 		    sizeof (struct fcp32_ioctl), mode)) {
1885 			return (EFAULT);
1886 		}
1887 		break;
1888 	}
1889 	case DDI_MODEL_NONE:
1890 		if (ddi_copyout((void *)&fioctl, (void *)data,
1891 		    sizeof (struct fcp_ioctl), mode)) {
1892 			return (EFAULT);
1893 		}
1894 		break;
1895 	}
1896 #else	/* _MULTI_DATAMODEL */
1897 
1898 	if (ddi_copyout((void *)&fioctl, (void *)data,
1899 	    sizeof (struct fcp_ioctl), mode)) {
1900 		return (EFAULT);
1901 	}
1902 #endif	/* _MULTI_DATAMODEL */
1903 
1904 	return (0);
1905 }
1906 
1907 /*
1908  * fcp_setup_scsi_ioctl
1909  *	Setup handler for the "scsi passthru" style of
1910  *	ioctl for FCP.	See "fcp_util.h" for data structure
1911  *	definition.
1912  *
1913  * Input:
1914  *	u_fscsi	= ioctl data (user address space)
1915  *	mode	= See ioctl(9E)
1916  *
1917  * Output:
1918  *	u_fscsi	= ioctl data (user address space)
1919  *	rval	= return value - see ioctl(9E)
1920  *
1921  * Returns:
1922  *	0	= OK
1923  *	EAGAIN	= See errno.h
1924  *	EBUSY	= See errno.h
1925  *	EFAULT	= See errno.h
1926  *	EINTR	= See errno.h
1927  *	EINVAL	= See errno.h
1928  *	EIO	= See errno.h
1929  *	ENOMEM	= See errno.h
1930  *	ENXIO	= See errno.h
1931  *
1932  * Context:
1933  *	Kernel context.
1934  */
1935 /* ARGSUSED */
1936 static int
1937 fcp_setup_scsi_ioctl(struct fcp_scsi_cmd *u_fscsi,
1938     int mode, int *rval)
1939 {
1940 	int			ret		= 0;
1941 	int			temp_ret;
1942 	caddr_t			k_cdbbufaddr	= NULL;
1943 	caddr_t			k_bufaddr	= NULL;
1944 	caddr_t			k_rqbufaddr	= NULL;
1945 	caddr_t			u_cdbbufaddr;
1946 	caddr_t			u_bufaddr;
1947 	caddr_t			u_rqbufaddr;
1948 	struct fcp_scsi_cmd	k_fscsi;
1949 
1950 	/*
1951 	 * Get fcp_scsi_cmd array element from user address space
1952 	 */
1953 	if ((ret = fcp_copyin_scsi_cmd((caddr_t)u_fscsi, &k_fscsi, mode))
1954 	    != 0) {
1955 		return (ret);
1956 	}
1957 
1958 
1959 	/*
1960 	 * Even though kmem_alloc() checks the validity of the
1961 	 * buffer length, this check is needed when the
1962 	 * kmem_flags set and the zero buffer length is passed.
1963 	 */
1964 	if ((k_fscsi.scsi_cdblen <= 0) ||
1965 	    (k_fscsi.scsi_buflen <= 0) ||
1966 	    (k_fscsi.scsi_buflen > FCP_MAX_RESPONSE_LEN) ||
1967 	    (k_fscsi.scsi_rqlen <= 0) ||
1968 	    (k_fscsi.scsi_rqlen > FCP_MAX_SENSE_LEN)) {
1969 		return (EINVAL);
1970 	}
1971 
1972 	/*
1973 	 * Allocate data for fcp_scsi_cmd pointer fields
1974 	 */
1975 	if (ret == 0) {
1976 		k_cdbbufaddr = kmem_alloc(k_fscsi.scsi_cdblen, KM_NOSLEEP);
1977 		k_bufaddr    = kmem_alloc(k_fscsi.scsi_buflen, KM_NOSLEEP);
1978 		k_rqbufaddr  = kmem_alloc(k_fscsi.scsi_rqlen,  KM_NOSLEEP);
1979 
1980 		if (k_cdbbufaddr == NULL ||
1981 		    k_bufaddr	 == NULL ||
1982 		    k_rqbufaddr	 == NULL) {
1983 			ret = ENOMEM;
1984 		}
1985 	}
1986 
1987 	/*
1988 	 * Get fcp_scsi_cmd pointer fields from user
1989 	 * address space
1990 	 */
1991 	if (ret == 0) {
1992 		u_cdbbufaddr = k_fscsi.scsi_cdbbufaddr;
1993 		u_bufaddr    = k_fscsi.scsi_bufaddr;
1994 		u_rqbufaddr  = k_fscsi.scsi_rqbufaddr;
1995 
1996 		if (ddi_copyin(u_cdbbufaddr,
1997 		    k_cdbbufaddr,
1998 		    k_fscsi.scsi_cdblen,
1999 		    mode)) {
2000 			ret = EFAULT;
2001 		} else if (ddi_copyin(u_bufaddr,
2002 		    k_bufaddr,
2003 		    k_fscsi.scsi_buflen,
2004 		    mode)) {
2005 			ret = EFAULT;
2006 		} else if (ddi_copyin(u_rqbufaddr,
2007 		    k_rqbufaddr,
2008 		    k_fscsi.scsi_rqlen,
2009 		    mode)) {
2010 			ret = EFAULT;
2011 		}
2012 	}
2013 
2014 	/*
2015 	 * Send scsi command (blocking)
2016 	 */
2017 	if (ret == 0) {
2018 		/*
2019 		 * Prior to sending the scsi command, the
2020 		 * fcp_scsi_cmd data structure must contain kernel,
2021 		 * not user, addresses.
2022 		 */
2023 		k_fscsi.scsi_cdbbufaddr	= k_cdbbufaddr;
2024 		k_fscsi.scsi_bufaddr	= k_bufaddr;
2025 		k_fscsi.scsi_rqbufaddr	= k_rqbufaddr;
2026 
2027 		ret = fcp_send_scsi_ioctl(&k_fscsi);
2028 
2029 		/*
2030 		 * After sending the scsi command, the
2031 		 * fcp_scsi_cmd data structure must contain user,
2032 		 * not kernel, addresses.
2033 		 */
2034 		k_fscsi.scsi_cdbbufaddr	= u_cdbbufaddr;
2035 		k_fscsi.scsi_bufaddr	= u_bufaddr;
2036 		k_fscsi.scsi_rqbufaddr	= u_rqbufaddr;
2037 	}
2038 
2039 	/*
2040 	 * Put fcp_scsi_cmd pointer fields to user address space
2041 	 */
2042 	if (ret == 0) {
2043 		if (ddi_copyout(k_cdbbufaddr,
2044 		    u_cdbbufaddr,
2045 		    k_fscsi.scsi_cdblen,
2046 		    mode)) {
2047 			ret = EFAULT;
2048 		} else if (ddi_copyout(k_bufaddr,
2049 		    u_bufaddr,
2050 		    k_fscsi.scsi_buflen,
2051 		    mode)) {
2052 			ret = EFAULT;
2053 		} else if (ddi_copyout(k_rqbufaddr,
2054 		    u_rqbufaddr,
2055 		    k_fscsi.scsi_rqlen,
2056 		    mode)) {
2057 			ret = EFAULT;
2058 		}
2059 	}
2060 
2061 	/*
2062 	 * Free data for fcp_scsi_cmd pointer fields
2063 	 */
2064 	if (k_cdbbufaddr != NULL) {
2065 		kmem_free(k_cdbbufaddr, k_fscsi.scsi_cdblen);
2066 	}
2067 	if (k_bufaddr != NULL) {
2068 		kmem_free(k_bufaddr, k_fscsi.scsi_buflen);
2069 	}
2070 	if (k_rqbufaddr != NULL) {
2071 		kmem_free(k_rqbufaddr, k_fscsi.scsi_rqlen);
2072 	}
2073 
2074 	/*
2075 	 * Put fcp_scsi_cmd array element to user address space
2076 	 */
2077 	temp_ret = fcp_copyout_scsi_cmd(&k_fscsi, (caddr_t)u_fscsi, mode);
2078 	if (temp_ret != 0) {
2079 		ret = temp_ret;
2080 	}
2081 
2082 	/*
2083 	 * Return status
2084 	 */
2085 	return (ret);
2086 }
2087 
2088 
2089 /*
2090  * fcp_copyin_scsi_cmd
2091  *	Copy in fcp_scsi_cmd data structure from user address space.
2092  *	The data may be in 32 bit or 64 bit modes.
2093  *
2094  * Input:
2095  *	base_addr	= from address (user address space)
2096  *	mode		= See ioctl(9E) and ddi_copyin(9F)
2097  *
2098  * Output:
2099  *	fscsi		= to address (kernel address space)
2100  *
2101  * Returns:
2102  *	0	= OK
2103  *	EFAULT	= Error
2104  *
2105  * Context:
2106  *	Kernel context.
2107  */
2108 static int
2109 fcp_copyin_scsi_cmd(caddr_t base_addr, struct fcp_scsi_cmd *fscsi, int mode)
2110 {
2111 #ifdef	_MULTI_DATAMODEL
2112 	struct fcp32_scsi_cmd	f32scsi;
2113 
2114 	switch (ddi_model_convert_from(mode & FMODELS)) {
2115 	case DDI_MODEL_ILP32:
2116 		/*
2117 		 * Copy data from user address space
2118 		 */
2119 		if (ddi_copyin((void *)base_addr,
2120 		    &f32scsi,
2121 		    sizeof (struct fcp32_scsi_cmd),
2122 		    mode)) {
2123 			return (EFAULT);
2124 		}
2125 		/*
2126 		 * Convert from 32 bit to 64 bit
2127 		 */
2128 		FCP32_SCSI_CMD_TO_FCP_SCSI_CMD(&f32scsi, fscsi);
2129 		break;
2130 	case DDI_MODEL_NONE:
2131 		/*
2132 		 * Copy data from user address space
2133 		 */
2134 		if (ddi_copyin((void *)base_addr,
2135 		    fscsi,
2136 		    sizeof (struct fcp_scsi_cmd),
2137 		    mode)) {
2138 			return (EFAULT);
2139 		}
2140 		break;
2141 	}
2142 #else	/* _MULTI_DATAMODEL */
2143 	/*
2144 	 * Copy data from user address space
2145 	 */
2146 	if (ddi_copyin((void *)base_addr,
2147 	    fscsi,
2148 	    sizeof (struct fcp_scsi_cmd),
2149 	    mode)) {
2150 		return (EFAULT);
2151 	}
2152 #endif	/* _MULTI_DATAMODEL */
2153 
2154 	return (0);
2155 }
2156 
2157 
2158 /*
2159  * fcp_copyout_scsi_cmd
2160  *	Copy out fcp_scsi_cmd data structure to user address space.
2161  *	The data may be in 32 bit or 64 bit modes.
2162  *
2163  * Input:
2164  *	fscsi		= to address (kernel address space)
2165  *	mode		= See ioctl(9E) and ddi_copyin(9F)
2166  *
2167  * Output:
2168  *	base_addr	= from address (user address space)
2169  *
2170  * Returns:
2171  *	0	= OK
2172  *	EFAULT	= Error
2173  *
2174  * Context:
2175  *	Kernel context.
2176  */
2177 static int
2178 fcp_copyout_scsi_cmd(struct fcp_scsi_cmd *fscsi, caddr_t base_addr, int mode)
2179 {
2180 #ifdef	_MULTI_DATAMODEL
2181 	struct fcp32_scsi_cmd	f32scsi;
2182 
2183 	switch (ddi_model_convert_from(mode & FMODELS)) {
2184 	case DDI_MODEL_ILP32:
2185 		/*
2186 		 * Convert from 64 bit to 32 bit
2187 		 */
2188 		FCP_SCSI_CMD_TO_FCP32_SCSI_CMD(fscsi, &f32scsi);
2189 		/*
2190 		 * Copy data to user address space
2191 		 */
2192 		if (ddi_copyout(&f32scsi,
2193 		    (void *)base_addr,
2194 		    sizeof (struct fcp32_scsi_cmd),
2195 		    mode)) {
2196 			return (EFAULT);
2197 		}
2198 		break;
2199 	case DDI_MODEL_NONE:
2200 		/*
2201 		 * Copy data to user address space
2202 		 */
2203 		if (ddi_copyout(fscsi,
2204 		    (void *)base_addr,
2205 		    sizeof (struct fcp_scsi_cmd),
2206 		    mode)) {
2207 			return (EFAULT);
2208 		}
2209 		break;
2210 	}
2211 #else	/* _MULTI_DATAMODEL */
2212 	/*
2213 	 * Copy data to user address space
2214 	 */
2215 	if (ddi_copyout(fscsi,
2216 	    (void *)base_addr,
2217 	    sizeof (struct fcp_scsi_cmd),
2218 	    mode)) {
2219 		return (EFAULT);
2220 	}
2221 #endif	/* _MULTI_DATAMODEL */
2222 
2223 	return (0);
2224 }
2225 
2226 
2227 /*
2228  * fcp_send_scsi_ioctl
2229  *	Sends the SCSI command in blocking mode.
2230  *
2231  * Input:
2232  *	fscsi		= SCSI command data structure
2233  *
2234  * Output:
2235  *	fscsi		= SCSI command data structure
2236  *
2237  * Returns:
2238  *	0	= OK
2239  *	EAGAIN	= See errno.h
2240  *	EBUSY	= See errno.h
2241  *	EINTR	= See errno.h
2242  *	EINVAL	= See errno.h
2243  *	EIO	= See errno.h
2244  *	ENOMEM	= See errno.h
2245  *	ENXIO	= See errno.h
2246  *
2247  * Context:
2248  *	Kernel context.
2249  */
2250 static int
2251 fcp_send_scsi_ioctl(struct fcp_scsi_cmd *fscsi)
2252 {
2253 	struct fcp_lun	*plun		= NULL;
2254 	struct fcp_port	*pptr		= NULL;
2255 	struct fcp_tgt	*ptgt		= NULL;
2256 	fc_packet_t		*fpkt		= NULL;
2257 	struct fcp_ipkt	*icmd		= NULL;
2258 	int			target_created	= FALSE;
2259 	fc_frame_hdr_t		*hp;
2260 	struct fcp_cmd		fcp_cmd;
2261 	struct fcp_cmd		*fcmd;
2262 	union scsi_cdb		*scsi_cdb;
2263 	la_wwn_t		*wwn_ptr;
2264 	int			nodma;
2265 	struct fcp_rsp		*rsp;
2266 	struct fcp_rsp_info	*rsp_info;
2267 	caddr_t			rsp_sense;
2268 	int			buf_len;
2269 	int			info_len;
2270 	int			sense_len;
2271 	struct scsi_extended_sense	*sense_to = NULL;
2272 	timeout_id_t		tid;
2273 	uint8_t			reconfig_lun = FALSE;
2274 	uint8_t			reconfig_pending = FALSE;
2275 	uint8_t			scsi_cmd;
2276 	int			rsp_len;
2277 	int			cmd_index;
2278 	int			fc_status;
2279 	int			pkt_state;
2280 	int			pkt_action;
2281 	int			pkt_reason;
2282 	int			ret, xport_retval = ~FC_SUCCESS;
2283 	int			lcount;
2284 	int			tcount;
2285 	int			reconfig_status;
2286 	int			port_busy = FALSE;
2287 	uchar_t			*lun_string;
2288 
2289 	/*
2290 	 * Check valid SCSI command
2291 	 */
2292 	scsi_cmd = ((uint8_t *)fscsi->scsi_cdbbufaddr)[0];
2293 	ret = EINVAL;
2294 	for (cmd_index = 0;
2295 	    cmd_index < FCP_NUM_ELEMENTS(scsi_ioctl_list) &&
2296 	    ret != 0;
2297 	    cmd_index++) {
2298 		/*
2299 		 * First byte of CDB is the SCSI command
2300 		 */
2301 		if (scsi_ioctl_list[cmd_index] == scsi_cmd) {
2302 			ret = 0;
2303 		}
2304 	}
2305 
2306 	/*
2307 	 * Check inputs
2308 	 */
2309 	if (fscsi->scsi_flags != FCP_SCSI_READ) {
2310 		ret = EINVAL;
2311 	} else if (fscsi->scsi_cdblen > FCP_CDB_SIZE) {
2312 		/* no larger than */
2313 		ret = EINVAL;
2314 	}
2315 
2316 
2317 	/*
2318 	 * Find FC port
2319 	 */
2320 	if (ret == 0) {
2321 		/*
2322 		 * Acquire global mutex
2323 		 */
2324 		mutex_enter(&fcp_global_mutex);
2325 
2326 		pptr = fcp_port_head;
2327 		while (pptr) {
2328 			if (pptr->port_instance ==
2329 			    (uint32_t)fscsi->scsi_fc_port_num) {
2330 				break;
2331 			} else {
2332 				pptr = pptr->port_next;
2333 			}
2334 		}
2335 
2336 		if (pptr == NULL) {
2337 			ret = ENXIO;
2338 		} else {
2339 			/*
2340 			 * fc_ulp_busy_port can raise power
2341 			 *  so, we must not hold any mutexes involved in PM
2342 			 */
2343 			mutex_exit(&fcp_global_mutex);
2344 			ret = fc_ulp_busy_port(pptr->port_fp_handle);
2345 		}
2346 
2347 		if (ret == 0) {
2348 
2349 			/* remember port is busy, so we will release later */
2350 			port_busy = TRUE;
2351 
2352 			/*
2353 			 * If there is a reconfiguration in progress, wait
2354 			 * for it to complete.
2355 			 */
2356 
2357 			fcp_reconfig_wait(pptr);
2358 
2359 			/* reacquire mutexes in order */
2360 			mutex_enter(&fcp_global_mutex);
2361 			mutex_enter(&pptr->port_mutex);
2362 
2363 			/*
2364 			 * Will port accept DMA?
2365 			 */
2366 			nodma = (pptr->port_fcp_dma == FC_NO_DVMA_SPACE)
2367 			    ? 1 : 0;
2368 
2369 			/*
2370 			 * If init or offline, device not known
2371 			 *
2372 			 * If we are discovering (onlining), we can
2373 			 * NOT obviously provide reliable data about
2374 			 * devices until it is complete
2375 			 */
2376 			if (pptr->port_state &	  (FCP_STATE_INIT |
2377 			    FCP_STATE_OFFLINE)) {
2378 				ret = ENXIO;
2379 			} else if (pptr->port_state & FCP_STATE_ONLINING) {
2380 				ret = EBUSY;
2381 			} else {
2382 				/*
2383 				 * Find target from pwwn
2384 				 *
2385 				 * The wwn must be put into a local
2386 				 * variable to ensure alignment.
2387 				 */
2388 				wwn_ptr = (la_wwn_t *)&(fscsi->scsi_fc_pwwn);
2389 				ptgt = fcp_lookup_target(pptr,
2390 				    (uchar_t *)wwn_ptr);
2391 
2392 				/*
2393 				 * If target not found,
2394 				 */
2395 				if (ptgt == NULL) {
2396 					/*
2397 					 * Note: Still have global &
2398 					 * port mutexes
2399 					 */
2400 					mutex_exit(&pptr->port_mutex);
2401 					ptgt = fcp_port_create_tgt(pptr,
2402 					    wwn_ptr, &ret, &fc_status,
2403 					    &pkt_state, &pkt_action,
2404 					    &pkt_reason);
2405 					mutex_enter(&pptr->port_mutex);
2406 
2407 					fscsi->scsi_fc_status  = fc_status;
2408 					fscsi->scsi_pkt_state  =
2409 					    (uchar_t)pkt_state;
2410 					fscsi->scsi_pkt_reason = pkt_reason;
2411 					fscsi->scsi_pkt_action =
2412 					    (uchar_t)pkt_action;
2413 
2414 					if (ptgt != NULL) {
2415 						target_created = TRUE;
2416 					} else if (ret == 0) {
2417 						ret = ENOMEM;
2418 					}
2419 				}
2420 
2421 				if (ret == 0) {
2422 					/*
2423 					 * Acquire target
2424 					 */
2425 					mutex_enter(&ptgt->tgt_mutex);
2426 
2427 					/*
2428 					 * If target is mark or busy,
2429 					 * then target can not be used
2430 					 */
2431 					if (ptgt->tgt_state &
2432 					    (FCP_TGT_MARK |
2433 					    FCP_TGT_BUSY)) {
2434 						ret = EBUSY;
2435 					} else {
2436 						/*
2437 						 * Mark target as busy
2438 						 */
2439 						ptgt->tgt_state |=
2440 						    FCP_TGT_BUSY;
2441 					}
2442 
2443 					/*
2444 					 * Release target
2445 					 */
2446 					lcount = pptr->port_link_cnt;
2447 					tcount = ptgt->tgt_change_cnt;
2448 					mutex_exit(&ptgt->tgt_mutex);
2449 				}
2450 			}
2451 
2452 			/*
2453 			 * Release port
2454 			 */
2455 			mutex_exit(&pptr->port_mutex);
2456 		}
2457 
2458 		/*
2459 		 * Release global mutex
2460 		 */
2461 		mutex_exit(&fcp_global_mutex);
2462 	}
2463 
2464 	if (ret == 0) {
2465 		uint64_t belun = BE_64(fscsi->scsi_lun);
2466 
2467 		/*
2468 		 * If it's a target device, find lun from pwwn
2469 		 * The wwn must be put into a local
2470 		 * variable to ensure alignment.
2471 		 */
2472 		mutex_enter(&pptr->port_mutex);
2473 		wwn_ptr = (la_wwn_t *)&(fscsi->scsi_fc_pwwn);
2474 		if (!ptgt->tgt_tcap && ptgt->tgt_icap) {
2475 			/* this is not a target */
2476 			fscsi->scsi_fc_status = FC_DEVICE_NOT_TGT;
2477 			ret = ENXIO;
2478 		} else if ((belun << 16) != 0) {
2479 			/*
2480 			 * Since fcp only support PD and LU addressing method
2481 			 * so far, the last 6 bytes of a valid LUN are expected
2482 			 * to be filled with 00h.
2483 			 */
2484 			fscsi->scsi_fc_status = FC_INVALID_LUN;
2485 			cmn_err(CE_WARN, "fcp: Unsupported LUN addressing"
2486 			    " method 0x%02x with LUN number 0x%016" PRIx64,
2487 			    (uint8_t)(belun >> 62), belun);
2488 			ret = ENXIO;
2489 		} else if ((plun = fcp_lookup_lun(pptr, (uchar_t *)wwn_ptr,
2490 		    (uint16_t)((belun >> 48) & 0x3fff))) == NULL) {
2491 			/*
2492 			 * This is a SCSI target, but no LUN at this
2493 			 * address.
2494 			 *
2495 			 * In the future, we may want to send this to
2496 			 * the target, and let it respond
2497 			 * appropriately
2498 			 */
2499 			ret = ENXIO;
2500 		}
2501 		mutex_exit(&pptr->port_mutex);
2502 	}
2503 
2504 	/*
2505 	 * Finished grabbing external resources
2506 	 * Allocate internal packet (icmd)
2507 	 */
2508 	if (ret == 0) {
2509 		/*
2510 		 * Calc rsp len assuming rsp info included
2511 		 */
2512 		rsp_len = sizeof (struct fcp_rsp) +
2513 		    sizeof (struct fcp_rsp_info) + fscsi->scsi_rqlen;
2514 
2515 		icmd = fcp_icmd_alloc(pptr, ptgt,
2516 		    sizeof (struct fcp_cmd),
2517 		    rsp_len,
2518 		    fscsi->scsi_buflen,
2519 		    nodma,
2520 		    lcount,			/* ipkt_link_cnt */
2521 		    tcount,			/* ipkt_change_cnt */
2522 		    0,				/* cause */
2523 		    FC_INVALID_RSCN_COUNT);	/* invalidate the count */
2524 
2525 		if (icmd == NULL) {
2526 			ret = ENOMEM;
2527 		} else {
2528 			/*
2529 			 * Setup internal packet as sema sync
2530 			 */
2531 			fcp_ipkt_sema_init(icmd);
2532 		}
2533 	}
2534 
2535 	if (ret == 0) {
2536 		/*
2537 		 * Init fpkt pointer for use.
2538 		 */
2539 
2540 		fpkt = icmd->ipkt_fpkt;
2541 
2542 		fpkt->pkt_tran_flags	= FC_TRAN_CLASS3 | FC_TRAN_INTR;
2543 		fpkt->pkt_tran_type	= FC_PKT_FCP_READ; /* only rd for now */
2544 		fpkt->pkt_timeout	= fscsi->scsi_timeout;
2545 
2546 		/*
2547 		 * Init fcmd pointer for use by SCSI command
2548 		 */
2549 
2550 		if (nodma) {
2551 			fcmd = (struct fcp_cmd *)fpkt->pkt_cmd;
2552 		} else {
2553 			fcmd = &fcp_cmd;
2554 		}
2555 		bzero(fcmd, sizeof (struct fcp_cmd));
2556 		ptgt = plun->lun_tgt;
2557 
2558 		lun_string = (uchar_t *)&fscsi->scsi_lun;
2559 
2560 		fcmd->fcp_ent_addr.ent_addr_0 =
2561 		    BE_16(*(uint16_t *)&(lun_string[0]));
2562 		fcmd->fcp_ent_addr.ent_addr_1 =
2563 		    BE_16(*(uint16_t *)&(lun_string[2]));
2564 		fcmd->fcp_ent_addr.ent_addr_2 =
2565 		    BE_16(*(uint16_t *)&(lun_string[4]));
2566 		fcmd->fcp_ent_addr.ent_addr_3 =
2567 		    BE_16(*(uint16_t *)&(lun_string[6]));
2568 
2569 		/*
2570 		 * Setup internal packet(icmd)
2571 		 */
2572 		icmd->ipkt_lun		= plun;
2573 		icmd->ipkt_restart	= 0;
2574 		icmd->ipkt_retries	= 0;
2575 		icmd->ipkt_opcode	= 0;
2576 
2577 		/*
2578 		 * Init the frame HEADER Pointer for use
2579 		 */
2580 		hp = &fpkt->pkt_cmd_fhdr;
2581 
2582 		hp->s_id	= pptr->port_id;
2583 		hp->d_id	= ptgt->tgt_d_id;
2584 		hp->r_ctl	= R_CTL_COMMAND;
2585 		hp->type	= FC_TYPE_SCSI_FCP;
2586 		hp->f_ctl	= F_CTL_SEQ_INITIATIVE | F_CTL_FIRST_SEQ;
2587 		hp->rsvd	= 0;
2588 		hp->seq_id	= 0;
2589 		hp->seq_cnt	= 0;
2590 		hp->ox_id	= 0xffff;
2591 		hp->rx_id	= 0xffff;
2592 		hp->ro		= 0;
2593 
2594 		fcmd->fcp_cntl.cntl_qtype	= FCP_QTYPE_SIMPLE;
2595 		fcmd->fcp_cntl.cntl_read_data	= 1;	/* only rd for now */
2596 		fcmd->fcp_cntl.cntl_write_data	= 0;
2597 		fcmd->fcp_data_len	= fscsi->scsi_buflen;
2598 
2599 		scsi_cdb = (union scsi_cdb *)fcmd->fcp_cdb;
2600 		bcopy((char *)fscsi->scsi_cdbbufaddr, (char *)scsi_cdb,
2601 		    fscsi->scsi_cdblen);
2602 
2603 		if (!nodma) {
2604 			FCP_CP_OUT((uint8_t *)fcmd, fpkt->pkt_cmd,
2605 			    fpkt->pkt_cmd_acc, sizeof (struct fcp_cmd));
2606 		}
2607 
2608 		/*
2609 		 * Send SCSI command to FC transport
2610 		 */
2611 
2612 		if (ret == 0) {
2613 			mutex_enter(&ptgt->tgt_mutex);
2614 
2615 			if (!FCP_TGT_STATE_CHANGED(ptgt, icmd)) {
2616 				mutex_exit(&ptgt->tgt_mutex);
2617 				fscsi->scsi_fc_status = xport_retval =
2618 				    fc_ulp_transport(pptr->port_fp_handle,
2619 				    fpkt);
2620 				if (fscsi->scsi_fc_status != FC_SUCCESS) {
2621 					ret = EIO;
2622 				}
2623 			} else {
2624 				mutex_exit(&ptgt->tgt_mutex);
2625 				ret = EBUSY;
2626 			}
2627 		}
2628 	}
2629 
2630 	/*
2631 	 * Wait for completion only if fc_ulp_transport was called and it
2632 	 * returned a success. This is the only time callback will happen.
2633 	 * Otherwise, there is no point in waiting
2634 	 */
2635 	if ((ret == 0) && (xport_retval == FC_SUCCESS)) {
2636 		ret = fcp_ipkt_sema_wait(icmd);
2637 	}
2638 
2639 	/*
2640 	 * Copy data to IOCTL data structures
2641 	 */
2642 	rsp = NULL;
2643 	if ((ret == 0) && (xport_retval == FC_SUCCESS)) {
2644 		rsp = (struct fcp_rsp *)fpkt->pkt_resp;
2645 
2646 		if (fcp_validate_fcp_response(rsp, pptr) != FC_SUCCESS) {
2647 			fcp_log(CE_WARN, pptr->port_dip,
2648 			    "!SCSI command to d_id=0x%x lun=0x%x"
2649 			    " failed, Bad FCP response values:"
2650 			    " rsvd1=%x, rsvd2=%x, sts-rsvd1=%x,"
2651 			    " sts-rsvd2=%x, rsplen=%x, senselen=%x",
2652 			    ptgt->tgt_d_id, plun->lun_num,
2653 			    rsp->reserved_0, rsp->reserved_1,
2654 			    rsp->fcp_u.fcp_status.reserved_0,
2655 			    rsp->fcp_u.fcp_status.reserved_1,
2656 			    rsp->fcp_response_len, rsp->fcp_sense_len);
2657 
2658 			ret = EIO;
2659 		}
2660 	}
2661 
2662 	if ((ret == 0) && (rsp != NULL)) {
2663 		/*
2664 		 * Calc response lengths
2665 		 */
2666 		sense_len = 0;
2667 		info_len = 0;
2668 
2669 		if (rsp->fcp_u.fcp_status.rsp_len_set) {
2670 			info_len = rsp->fcp_response_len;
2671 		}
2672 
2673 		rsp_info   = (struct fcp_rsp_info *)
2674 		    ((uint8_t *)rsp + sizeof (struct fcp_rsp));
2675 
2676 		/*
2677 		 * Get SCSI status
2678 		 */
2679 		fscsi->scsi_bufstatus = rsp->fcp_u.fcp_status.scsi_status;
2680 		/*
2681 		 * If a lun was just added or removed and the next command
2682 		 * comes through this interface, we need to capture the check
2683 		 * condition so we can discover the new topology.
2684 		 */
2685 		if (fscsi->scsi_bufstatus != STATUS_GOOD &&
2686 		    rsp->fcp_u.fcp_status.sense_len_set) {
2687 			sense_len = rsp->fcp_sense_len;
2688 			rsp_sense  = (caddr_t)((uint8_t *)rsp_info + info_len);
2689 			sense_to = (struct scsi_extended_sense *)rsp_sense;
2690 			if ((FCP_SENSE_REPORTLUN_CHANGED(sense_to)) ||
2691 			    (FCP_SENSE_NO_LUN(sense_to))) {
2692 				reconfig_lun = TRUE;
2693 			}
2694 		}
2695 
2696 		if (fscsi->scsi_bufstatus == STATUS_GOOD && (ptgt != NULL) &&
2697 		    (reconfig_lun || (scsi_cdb->scc_cmd == SCMD_REPORT_LUN))) {
2698 			if (reconfig_lun == FALSE) {
2699 				reconfig_status =
2700 				    fcp_is_reconfig_needed(ptgt, fpkt);
2701 			}
2702 
2703 			if ((reconfig_lun == TRUE) ||
2704 			    (reconfig_status == TRUE)) {
2705 				mutex_enter(&ptgt->tgt_mutex);
2706 				if (ptgt->tgt_tid == NULL) {
2707 					/*
2708 					 * Either we've been notified the
2709 					 * REPORT_LUN data has changed, or
2710 					 * we've determined on our own that
2711 					 * we're out of date.  Kick off
2712 					 * rediscovery.
2713 					 */
2714 					tid = timeout(fcp_reconfigure_luns,
2715 					    (caddr_t)ptgt, drv_usectohz(1));
2716 
2717 					ptgt->tgt_tid = tid;
2718 					ptgt->tgt_state |= FCP_TGT_BUSY;
2719 					ret = EBUSY;
2720 					reconfig_pending = TRUE;
2721 				}
2722 				mutex_exit(&ptgt->tgt_mutex);
2723 			}
2724 		}
2725 
2726 		/*
2727 		 * Calc residuals and buffer lengths
2728 		 */
2729 
2730 		if (ret == 0) {
2731 			buf_len = fscsi->scsi_buflen;
2732 			fscsi->scsi_bufresid	= 0;
2733 			if (rsp->fcp_u.fcp_status.resid_under) {
2734 				if (rsp->fcp_resid <= fscsi->scsi_buflen) {
2735 					fscsi->scsi_bufresid = rsp->fcp_resid;
2736 				} else {
2737 					cmn_err(CE_WARN, "fcp: bad residue %x "
2738 					    "for txfer len %x", rsp->fcp_resid,
2739 					    fscsi->scsi_buflen);
2740 					fscsi->scsi_bufresid =
2741 					    fscsi->scsi_buflen;
2742 				}
2743 				buf_len -= fscsi->scsi_bufresid;
2744 			}
2745 			if (rsp->fcp_u.fcp_status.resid_over) {
2746 				fscsi->scsi_bufresid = -rsp->fcp_resid;
2747 			}
2748 
2749 			fscsi->scsi_rqresid	= fscsi->scsi_rqlen - sense_len;
2750 			if (fscsi->scsi_rqlen < sense_len) {
2751 				sense_len = fscsi->scsi_rqlen;
2752 			}
2753 
2754 			fscsi->scsi_fc_rspcode	= 0;
2755 			if (rsp->fcp_u.fcp_status.rsp_len_set) {
2756 				fscsi->scsi_fc_rspcode	= rsp_info->rsp_code;
2757 			}
2758 			fscsi->scsi_pkt_state	= fpkt->pkt_state;
2759 			fscsi->scsi_pkt_action	= fpkt->pkt_action;
2760 			fscsi->scsi_pkt_reason	= fpkt->pkt_reason;
2761 
2762 			/*
2763 			 * Copy data and request sense
2764 			 *
2765 			 * Data must be copied by using the FCP_CP_IN macro.
2766 			 * This will ensure the proper byte order since the data
2767 			 * is being copied directly from the memory mapped
2768 			 * device register.
2769 			 *
2770 			 * The response (and request sense) will be in the
2771 			 * correct byte order.	No special copy is necessary.
2772 			 */
2773 
2774 			if (buf_len) {
2775 				FCP_CP_IN(fpkt->pkt_data,
2776 				    fscsi->scsi_bufaddr,
2777 				    fpkt->pkt_data_acc,
2778 				    buf_len);
2779 			}
2780 			bcopy((void *)rsp_sense,
2781 			    (void *)fscsi->scsi_rqbufaddr,
2782 			    sense_len);
2783 		}
2784 	}
2785 
2786 	/*
2787 	 * Cleanup transport data structures if icmd was alloc-ed
2788 	 * So, cleanup happens in the same thread that icmd was alloc-ed
2789 	 */
2790 	if (icmd != NULL) {
2791 		fcp_ipkt_sema_cleanup(icmd);
2792 	}
2793 
2794 	/* restore pm busy/idle status */
2795 	if (port_busy) {
2796 		fc_ulp_idle_port(pptr->port_fp_handle);
2797 	}
2798 
2799 	/*
2800 	 * Cleanup target.  if a reconfig is pending, don't clear the BUSY
2801 	 * flag, it'll be cleared when the reconfig is complete.
2802 	 */
2803 	if ((ptgt != NULL) && !reconfig_pending) {
2804 		/*
2805 		 * If target was created,
2806 		 */
2807 		if (target_created) {
2808 			mutex_enter(&ptgt->tgt_mutex);
2809 			ptgt->tgt_state &= ~FCP_TGT_BUSY;
2810 			mutex_exit(&ptgt->tgt_mutex);
2811 		} else {
2812 			/*
2813 			 * De-mark target as busy
2814 			 */
2815 			mutex_enter(&ptgt->tgt_mutex);
2816 			ptgt->tgt_state &= ~FCP_TGT_BUSY;
2817 			mutex_exit(&ptgt->tgt_mutex);
2818 		}
2819 	}
2820 	return (ret);
2821 }
2822 
2823 
2824 static int
2825 fcp_is_reconfig_needed(struct fcp_tgt *ptgt,
2826     fc_packet_t	*fpkt)
2827 {
2828 	uchar_t			*lun_string;
2829 	uint16_t		lun_num, i;
2830 	int			num_luns;
2831 	int			actual_luns;
2832 	int			num_masked_luns;
2833 	int			lun_buflen;
2834 	struct fcp_lun	*plun	= NULL;
2835 	struct fcp_reportlun_resp	*report_lun;
2836 	uint8_t			reconfig_needed = FALSE;
2837 	uint8_t			lun_exists = FALSE;
2838 
2839 	report_lun = kmem_zalloc(fpkt->pkt_datalen, KM_SLEEP);
2840 
2841 	FCP_CP_IN(fpkt->pkt_data, report_lun, fpkt->pkt_data_acc,
2842 	    fpkt->pkt_datalen);
2843 
2844 	/* get number of luns (which is supplied as LUNS * 8) */
2845 	num_luns = BE_32(report_lun->num_lun) >> 3;
2846 
2847 	/*
2848 	 * Figure out exactly how many lun strings our response buffer
2849 	 * can hold.
2850 	 */
2851 	lun_buflen = (fpkt->pkt_datalen -
2852 	    2 * sizeof (uint32_t)) / sizeof (longlong_t);
2853 
2854 	/*
2855 	 * Is our response buffer full or not? We don't want to
2856 	 * potentially walk beyond the number of luns we have.
2857 	 */
2858 	if (num_luns <= lun_buflen) {
2859 		actual_luns = num_luns;
2860 	} else {
2861 		actual_luns = lun_buflen;
2862 	}
2863 
2864 	mutex_enter(&ptgt->tgt_mutex);
2865 
2866 	/* Scan each lun to see if we have masked it. */
2867 	num_masked_luns = 0;
2868 	if (fcp_lun_blacklist != NULL) {
2869 		for (i = 0; i < actual_luns; i++) {
2870 			lun_string = (uchar_t *)&(report_lun->lun_string[i]);
2871 			switch (lun_string[0] & 0xC0) {
2872 			case FCP_LUN_ADDRESSING:
2873 			case FCP_PD_ADDRESSING:
2874 				lun_num = ((lun_string[0] & 0x3F) << 8)
2875 				    | lun_string[1];
2876 				if (fcp_should_mask(&ptgt->tgt_port_wwn,
2877 				    lun_num) == TRUE) {
2878 					num_masked_luns++;
2879 				}
2880 				break;
2881 			default:
2882 				break;
2883 			}
2884 		}
2885 	}
2886 
2887 	/*
2888 	 * The quick and easy check.  If the number of LUNs reported
2889 	 * doesn't match the number we currently know about, we need
2890 	 * to reconfigure.
2891 	 */
2892 	if (num_luns && num_luns != (ptgt->tgt_lun_cnt + num_masked_luns)) {
2893 		mutex_exit(&ptgt->tgt_mutex);
2894 		kmem_free(report_lun, fpkt->pkt_datalen);
2895 		return (TRUE);
2896 	}
2897 
2898 	/*
2899 	 * If the quick and easy check doesn't turn up anything, we walk
2900 	 * the list of luns from the REPORT_LUN response and look for
2901 	 * any luns we don't know about.  If we find one, we know we need
2902 	 * to reconfigure. We will skip LUNs that are masked because of the
2903 	 * blacklist.
2904 	 */
2905 	for (i = 0; i < actual_luns; i++) {
2906 		lun_string = (uchar_t *)&(report_lun->lun_string[i]);
2907 		lun_exists = FALSE;
2908 		switch (lun_string[0] & 0xC0) {
2909 		case FCP_LUN_ADDRESSING:
2910 		case FCP_PD_ADDRESSING:
2911 			lun_num = ((lun_string[0] & 0x3F) << 8) | lun_string[1];
2912 
2913 			if ((fcp_lun_blacklist != NULL) && (fcp_should_mask(
2914 			    &ptgt->tgt_port_wwn, lun_num) == TRUE)) {
2915 				lun_exists = TRUE;
2916 				break;
2917 			}
2918 
2919 			for (plun = ptgt->tgt_lun; plun;
2920 			    plun = plun->lun_next) {
2921 				if (plun->lun_num == lun_num) {
2922 					lun_exists = TRUE;
2923 					break;
2924 				}
2925 			}
2926 			break;
2927 		default:
2928 			break;
2929 		}
2930 
2931 		if (lun_exists == FALSE) {
2932 			reconfig_needed = TRUE;
2933 			break;
2934 		}
2935 	}
2936 
2937 	mutex_exit(&ptgt->tgt_mutex);
2938 	kmem_free(report_lun, fpkt->pkt_datalen);
2939 
2940 	return (reconfig_needed);
2941 }
2942 
2943 /*
2944  * This function is called by fcp_handle_page83 and uses inquiry response data
2945  * stored in plun->lun_inq to determine whether or not a device is a member of
2946  * the table fcp_symmetric_disk_table_size. We return 0 if it is in the table,
2947  * otherwise 1.
2948  */
2949 static int
2950 fcp_symmetric_device_probe(struct fcp_lun *plun)
2951 {
2952 	struct scsi_inquiry	*stdinq = &plun->lun_inq;
2953 	char			*devidptr;
2954 	int			i, len;
2955 
2956 	for (i = 0; i < fcp_symmetric_disk_table_size; i++) {
2957 		devidptr = fcp_symmetric_disk_table[i];
2958 		len = (int)strlen(devidptr);
2959 
2960 		if (bcmp(stdinq->inq_vid, devidptr, len) == 0) {
2961 			return (0);
2962 		}
2963 	}
2964 	return (1);
2965 }
2966 
2967 
2968 /*
2969  * This function is called by fcp_ioctl for the FCP_STATE_COUNT ioctl
2970  * It basically returns the current count of # of state change callbacks
2971  * i.e the value of tgt_change_cnt.
2972  *
2973  * INPUT:
2974  *   fcp_ioctl.fp_minor -> The minor # of the fp port
2975  *   fcp_ioctl.listlen	-> 1
2976  *   fcp_ioctl.list	-> Pointer to a 32 bit integer
2977  */
2978 /*ARGSUSED2*/
2979 static int
2980 fcp_get_statec_count(struct fcp_ioctl *data, int mode, int *rval)
2981 {
2982 	int			ret;
2983 	uint32_t		link_cnt;
2984 	struct fcp_ioctl	fioctl;
2985 	struct fcp_port	*pptr = NULL;
2986 
2987 	if ((ret = fcp_copyin_fcp_ioctl_data(data, mode, rval, &fioctl,
2988 	    &pptr)) != 0) {
2989 		return (ret);
2990 	}
2991 
2992 	ASSERT(pptr != NULL);
2993 
2994 	if (fioctl.listlen != 1) {
2995 		return (EINVAL);
2996 	}
2997 
2998 	mutex_enter(&pptr->port_mutex);
2999 	if (pptr->port_state & FCP_STATE_OFFLINE) {
3000 		mutex_exit(&pptr->port_mutex);
3001 		return (ENXIO);
3002 	}
3003 
3004 	/*
3005 	 * FCP_STATE_INIT is set in 2 cases (not sure why it is overloaded):
3006 	 * When the fcp initially attaches to the port and there are nothing
3007 	 * hanging out of the port or if there was a repeat offline state change
3008 	 * callback (refer fcp_statec_callback() FC_STATE_OFFLINE case).
3009 	 * In the latter case, port_tmp_cnt will be non-zero and that is how we
3010 	 * will differentiate the 2 cases.
3011 	 */
3012 	if ((pptr->port_state & FCP_STATE_INIT) && pptr->port_tmp_cnt) {
3013 		mutex_exit(&pptr->port_mutex);
3014 		return (ENXIO);
3015 	}
3016 
3017 	link_cnt = pptr->port_link_cnt;
3018 	mutex_exit(&pptr->port_mutex);
3019 
3020 	if (ddi_copyout(&link_cnt, fioctl.list, (sizeof (uint32_t)), mode)) {
3021 		return (EFAULT);
3022 	}
3023 
3024 #ifdef	_MULTI_DATAMODEL
3025 	switch (ddi_model_convert_from(mode & FMODELS)) {
3026 	case DDI_MODEL_ILP32: {
3027 		struct fcp32_ioctl f32_ioctl;
3028 
3029 		f32_ioctl.fp_minor = fioctl.fp_minor;
3030 		f32_ioctl.listlen = fioctl.listlen;
3031 		f32_ioctl.list = (caddr32_t)(long)fioctl.list;
3032 		if (ddi_copyout((void *)&f32_ioctl, (void *)data,
3033 		    sizeof (struct fcp32_ioctl), mode)) {
3034 			return (EFAULT);
3035 		}
3036 		break;
3037 	}
3038 	case DDI_MODEL_NONE:
3039 		if (ddi_copyout((void *)&fioctl, (void *)data,
3040 		    sizeof (struct fcp_ioctl), mode)) {
3041 			return (EFAULT);
3042 		}
3043 		break;
3044 	}
3045 #else	/* _MULTI_DATAMODEL */
3046 
3047 	if (ddi_copyout((void *)&fioctl, (void *)data,
3048 	    sizeof (struct fcp_ioctl), mode)) {
3049 		return (EFAULT);
3050 	}
3051 #endif	/* _MULTI_DATAMODEL */
3052 
3053 	return (0);
3054 }
3055 
3056 /*
3057  * This function copies the fcp_ioctl structure passed in from user land
3058  * into kernel land. Handles 32 bit applications.
3059  */
3060 /*ARGSUSED*/
3061 static int
3062 fcp_copyin_fcp_ioctl_data(struct fcp_ioctl *data, int mode, int *rval,
3063     struct fcp_ioctl *fioctl, struct fcp_port **pptr)
3064 {
3065 	struct fcp_port	*t_pptr;
3066 
3067 #ifdef	_MULTI_DATAMODEL
3068 	switch (ddi_model_convert_from(mode & FMODELS)) {
3069 	case DDI_MODEL_ILP32: {
3070 		struct fcp32_ioctl f32_ioctl;
3071 
3072 		if (ddi_copyin((void *)data, (void *)&f32_ioctl,
3073 		    sizeof (struct fcp32_ioctl), mode)) {
3074 			return (EFAULT);
3075 		}
3076 		fioctl->fp_minor = f32_ioctl.fp_minor;
3077 		fioctl->listlen = f32_ioctl.listlen;
3078 		fioctl->list = (caddr_t)(long)f32_ioctl.list;
3079 		break;
3080 	}
3081 	case DDI_MODEL_NONE:
3082 		if (ddi_copyin((void *)data, (void *)fioctl,
3083 		    sizeof (struct fcp_ioctl), mode)) {
3084 			return (EFAULT);
3085 		}
3086 		break;
3087 	}
3088 
3089 #else	/* _MULTI_DATAMODEL */
3090 	if (ddi_copyin((void *)data, (void *)fioctl,
3091 	    sizeof (struct fcp_ioctl), mode)) {
3092 		return (EFAULT);
3093 	}
3094 #endif	/* _MULTI_DATAMODEL */
3095 
3096 	/*
3097 	 * Right now we can assume that the minor number matches with
3098 	 * this instance of fp. If this changes we will need to
3099 	 * revisit this logic.
3100 	 */
3101 	mutex_enter(&fcp_global_mutex);
3102 	t_pptr = fcp_port_head;
3103 	while (t_pptr) {
3104 		if (t_pptr->port_instance == (uint32_t)fioctl->fp_minor) {
3105 			break;
3106 		} else {
3107 			t_pptr = t_pptr->port_next;
3108 		}
3109 	}
3110 	*pptr = t_pptr;
3111 	mutex_exit(&fcp_global_mutex);
3112 	if (t_pptr == NULL) {
3113 		return (ENXIO);
3114 	}
3115 
3116 	return (0);
3117 }
3118 
3119 /*
3120  *     Function: fcp_port_create_tgt
3121  *
3122  *  Description: As the name suggest this function creates the target context
3123  *		 specified by the the WWN provided by the caller.  If the
3124  *		 creation goes well and the target is known by fp/fctl a PLOGI
3125  *		 followed by a PRLI are issued.
3126  *
3127  *     Argument: pptr		fcp port structure
3128  *		 pwwn		WWN of the target
3129  *		 ret_val	Address of the return code.  It could be:
3130  *				EIO, ENOMEM or 0.
3131  *		 fc_status	PLOGI or PRLI status completion
3132  *		 fc_pkt_state	PLOGI or PRLI state completion
3133  *		 fc_pkt_reason	PLOGI or PRLI reason completion
3134  *		 fc_pkt_action	PLOGI or PRLI action completion
3135  *
3136  * Return Value: NULL if it failed
3137  *		 Target structure address if it succeeds
3138  */
3139 static struct fcp_tgt *
3140 fcp_port_create_tgt(struct fcp_port *pptr, la_wwn_t *pwwn, int *ret_val,
3141     int *fc_status, int *fc_pkt_state, int *fc_pkt_reason, int *fc_pkt_action)
3142 {
3143 	struct fcp_tgt	*ptgt = NULL;
3144 	fc_portmap_t		devlist;
3145 	int			lcount;
3146 	int			error;
3147 
3148 	*ret_val = 0;
3149 
3150 	/*
3151 	 * Check FC port device & get port map
3152 	 */
3153 	if (fc_ulp_get_remote_port(pptr->port_fp_handle, pwwn,
3154 	    &error, 1) == NULL) {
3155 		*ret_val = EIO;
3156 	} else {
3157 		if (fc_ulp_pwwn_to_portmap(pptr->port_fp_handle, pwwn,
3158 		    &devlist) != FC_SUCCESS) {
3159 			*ret_val = EIO;
3160 		}
3161 	}
3162 
3163 	/* Set port map flags */
3164 	devlist.map_type = PORT_DEVICE_USER_CREATE;
3165 
3166 	/* Allocate target */
3167 	if (*ret_val == 0) {
3168 		lcount = pptr->port_link_cnt;
3169 		ptgt = fcp_alloc_tgt(pptr, &devlist, lcount);
3170 		if (ptgt == NULL) {
3171 			fcp_log(CE_WARN, pptr->port_dip,
3172 			    "!FC target allocation failed");
3173 			*ret_val = ENOMEM;
3174 		} else {
3175 			/* Setup target */
3176 			mutex_enter(&ptgt->tgt_mutex);
3177 
3178 			ptgt->tgt_statec_cause	= FCP_CAUSE_TGT_CHANGE;
3179 			ptgt->tgt_tmp_cnt	= 1;
3180 			ptgt->tgt_d_id		= devlist.map_did.port_id;
3181 			ptgt->tgt_hard_addr	=
3182 			    devlist.map_hard_addr.hard_addr;
3183 			ptgt->tgt_pd_handle	= devlist.map_pd;
3184 			ptgt->tgt_fca_dev	= NULL;
3185 
3186 			bcopy(&devlist.map_nwwn, &ptgt->tgt_node_wwn.raw_wwn[0],
3187 			    FC_WWN_SIZE);
3188 			bcopy(&devlist.map_pwwn, &ptgt->tgt_port_wwn.raw_wwn[0],
3189 			    FC_WWN_SIZE);
3190 
3191 			mutex_exit(&ptgt->tgt_mutex);
3192 		}
3193 	}
3194 
3195 	/* Release global mutex for PLOGI and PRLI */
3196 	mutex_exit(&fcp_global_mutex);
3197 
3198 	/* Send PLOGI (If necessary) */
3199 	if (*ret_val == 0) {
3200 		*ret_val = fcp_tgt_send_plogi(ptgt, fc_status,
3201 		    fc_pkt_state, fc_pkt_reason, fc_pkt_action);
3202 	}
3203 
3204 	/* Send PRLI (If necessary) */
3205 	if (*ret_val == 0) {
3206 		*ret_val = fcp_tgt_send_prli(ptgt, fc_status,
3207 		    fc_pkt_state, fc_pkt_reason, fc_pkt_action);
3208 	}
3209 
3210 	mutex_enter(&fcp_global_mutex);
3211 
3212 	return (ptgt);
3213 }
3214 
3215 /*
3216  *     Function: fcp_tgt_send_plogi
3217  *
3218  *  Description: This function sends a PLOGI to the target specified by the
3219  *		 caller and waits till it completes.
3220  *
3221  *     Argument: ptgt		Target to send the plogi to.
3222  *		 fc_status	Status returned by fp/fctl in the PLOGI request.
3223  *		 fc_pkt_state	State returned by fp/fctl in the PLOGI request.
3224  *		 fc_pkt_reason	Reason returned by fp/fctl in the PLOGI request.
3225  *		 fc_pkt_action	Action returned by fp/fctl in the PLOGI request.
3226  *
3227  * Return Value: 0
3228  *		 ENOMEM
3229  *		 EIO
3230  *
3231  *	Context: User context.
3232  */
3233 static int
3234 fcp_tgt_send_plogi(struct fcp_tgt *ptgt, int *fc_status, int *fc_pkt_state,
3235     int *fc_pkt_reason, int *fc_pkt_action)
3236 {
3237 	struct fcp_port	*pptr;
3238 	struct fcp_ipkt	*icmd;
3239 	struct fc_packet	*fpkt;
3240 	fc_frame_hdr_t		*hp;
3241 	struct la_els_logi	logi;
3242 	int			tcount;
3243 	int			lcount;
3244 	int			ret, login_retval = ~FC_SUCCESS;
3245 
3246 	ret = 0;
3247 
3248 	pptr = ptgt->tgt_port;
3249 
3250 	lcount = pptr->port_link_cnt;
3251 	tcount = ptgt->tgt_change_cnt;
3252 
3253 	/* Alloc internal packet */
3254 	icmd = fcp_icmd_alloc(pptr, ptgt, sizeof (la_els_logi_t),
3255 	    sizeof (la_els_logi_t), 0, 0, lcount, tcount, 0,
3256 	    FC_INVALID_RSCN_COUNT);
3257 
3258 	if (icmd == NULL) {
3259 		ret = ENOMEM;
3260 	} else {
3261 		/*
3262 		 * Setup internal packet as sema sync
3263 		 */
3264 		fcp_ipkt_sema_init(icmd);
3265 
3266 		/*
3267 		 * Setup internal packet (icmd)
3268 		 */
3269 		icmd->ipkt_lun		= NULL;
3270 		icmd->ipkt_restart	= 0;
3271 		icmd->ipkt_retries	= 0;
3272 		icmd->ipkt_opcode	= LA_ELS_PLOGI;
3273 
3274 		/*
3275 		 * Setup fc_packet
3276 		 */
3277 		fpkt = icmd->ipkt_fpkt;
3278 
3279 		fpkt->pkt_tran_flags	= FC_TRAN_CLASS3 | FC_TRAN_INTR;
3280 		fpkt->pkt_tran_type	= FC_PKT_EXCHANGE;
3281 		fpkt->pkt_timeout	= FCP_ELS_TIMEOUT;
3282 
3283 		/*
3284 		 * Setup FC frame header
3285 		 */
3286 		hp = &fpkt->pkt_cmd_fhdr;
3287 
3288 		hp->s_id	= pptr->port_id;	/* source ID */
3289 		hp->d_id	= ptgt->tgt_d_id;	/* dest ID */
3290 		hp->r_ctl	= R_CTL_ELS_REQ;
3291 		hp->type	= FC_TYPE_EXTENDED_LS;
3292 		hp->f_ctl	= F_CTL_SEQ_INITIATIVE | F_CTL_FIRST_SEQ;
3293 		hp->seq_id	= 0;
3294 		hp->rsvd	= 0;
3295 		hp->df_ctl	= 0;
3296 		hp->seq_cnt	= 0;
3297 		hp->ox_id	= 0xffff;		/* i.e. none */
3298 		hp->rx_id	= 0xffff;		/* i.e. none */
3299 		hp->ro		= 0;
3300 
3301 		/*
3302 		 * Setup PLOGI
3303 		 */
3304 		bzero(&logi, sizeof (struct la_els_logi));
3305 		logi.ls_code.ls_code = LA_ELS_PLOGI;
3306 
3307 		FCP_CP_OUT((uint8_t *)&logi, fpkt->pkt_cmd,
3308 		    fpkt->pkt_cmd_acc, sizeof (struct la_els_logi));
3309 
3310 		/*
3311 		 * Send PLOGI
3312 		 */
3313 		*fc_status = login_retval =
3314 		    fc_ulp_login(pptr->port_fp_handle, &fpkt, 1);
3315 		if (*fc_status != FC_SUCCESS) {
3316 			ret = EIO;
3317 		}
3318 	}
3319 
3320 	/*
3321 	 * Wait for completion
3322 	 */
3323 	if ((ret == 0) && (login_retval == FC_SUCCESS)) {
3324 		ret = fcp_ipkt_sema_wait(icmd);
3325 
3326 		*fc_pkt_state	= fpkt->pkt_state;
3327 		*fc_pkt_reason	= fpkt->pkt_reason;
3328 		*fc_pkt_action	= fpkt->pkt_action;
3329 	}
3330 
3331 	/*
3332 	 * Cleanup transport data structures if icmd was alloc-ed AND if there
3333 	 * is going to be no callback (i.e if fc_ulp_login() failed).
3334 	 * Otherwise, cleanup happens in callback routine.
3335 	 */
3336 	if (icmd != NULL) {
3337 		fcp_ipkt_sema_cleanup(icmd);
3338 	}
3339 
3340 	return (ret);
3341 }
3342 
3343 /*
3344  *     Function: fcp_tgt_send_prli
3345  *
3346  *  Description: Does nothing as of today.
3347  *
3348  *     Argument: ptgt		Target to send the prli to.
3349  *		 fc_status	Status returned by fp/fctl in the PRLI request.
3350  *		 fc_pkt_state	State returned by fp/fctl in the PRLI request.
3351  *		 fc_pkt_reason	Reason returned by fp/fctl in the PRLI request.
3352  *		 fc_pkt_action	Action returned by fp/fctl in the PRLI request.
3353  *
3354  * Return Value: 0
3355  */
3356 /*ARGSUSED*/
3357 static int
3358 fcp_tgt_send_prli(struct fcp_tgt *ptgt, int *fc_status, int *fc_pkt_state,
3359     int *fc_pkt_reason, int *fc_pkt_action)
3360 {
3361 	return (0);
3362 }
3363 
3364 /*
3365  *     Function: fcp_ipkt_sema_init
3366  *
3367  *  Description: Initializes the semaphore contained in the internal packet.
3368  *
3369  *     Argument: icmd	Internal packet the semaphore of which must be
3370  *			initialized.
3371  *
3372  * Return Value: None
3373  *
3374  *	Context: User context only.
3375  */
3376 static void
3377 fcp_ipkt_sema_init(struct fcp_ipkt *icmd)
3378 {
3379 	struct fc_packet	*fpkt;
3380 
3381 	fpkt = icmd->ipkt_fpkt;
3382 
3383 	/* Create semaphore for sync */
3384 	sema_init(&(icmd->ipkt_sema), 0, NULL, SEMA_DRIVER, NULL);
3385 
3386 	/* Setup the completion callback */
3387 	fpkt->pkt_comp = fcp_ipkt_sema_callback;
3388 }
3389 
3390 /*
3391  *     Function: fcp_ipkt_sema_wait
3392  *
3393  *  Description: Wait on the semaphore embedded in the internal packet.	 The
3394  *		 semaphore is released in the callback.
3395  *
3396  *     Argument: icmd	Internal packet to wait on for completion.
3397  *
3398  * Return Value: 0
3399  *		 EIO
3400  *		 EBUSY
3401  *		 EAGAIN
3402  *
3403  *	Context: User context only.
3404  *
3405  * This function does a conversion between the field pkt_state of the fc_packet
3406  * embedded in the internal packet (icmd) and the code it returns.
3407  */
3408 static int
3409 fcp_ipkt_sema_wait(struct fcp_ipkt *icmd)
3410 {
3411 	struct fc_packet	*fpkt;
3412 	int	ret;
3413 
3414 	ret = EIO;
3415 	fpkt = icmd->ipkt_fpkt;
3416 
3417 	/*
3418 	 * Wait on semaphore
3419 	 */
3420 	sema_p(&(icmd->ipkt_sema));
3421 
3422 	/*
3423 	 * Check the status of the FC packet
3424 	 */
3425 	switch (fpkt->pkt_state) {
3426 	case FC_PKT_SUCCESS:
3427 		ret = 0;
3428 		break;
3429 	case FC_PKT_LOCAL_RJT:
3430 		switch (fpkt->pkt_reason) {
3431 		case FC_REASON_SEQ_TIMEOUT:
3432 		case FC_REASON_RX_BUF_TIMEOUT:
3433 			ret = EAGAIN;
3434 			break;
3435 		case FC_REASON_PKT_BUSY:
3436 			ret = EBUSY;
3437 			break;
3438 		}
3439 		break;
3440 	case FC_PKT_TIMEOUT:
3441 		ret = EAGAIN;
3442 		break;
3443 	case FC_PKT_LOCAL_BSY:
3444 	case FC_PKT_TRAN_BSY:
3445 	case FC_PKT_NPORT_BSY:
3446 	case FC_PKT_FABRIC_BSY:
3447 		ret = EBUSY;
3448 		break;
3449 	case FC_PKT_LS_RJT:
3450 	case FC_PKT_BA_RJT:
3451 		switch (fpkt->pkt_reason) {
3452 		case FC_REASON_LOGICAL_BSY:
3453 			ret = EBUSY;
3454 			break;
3455 		}
3456 		break;
3457 	case FC_PKT_FS_RJT:
3458 		switch (fpkt->pkt_reason) {
3459 		case FC_REASON_FS_LOGICAL_BUSY:
3460 			ret = EBUSY;
3461 			break;
3462 		}
3463 		break;
3464 	}
3465 
3466 	return (ret);
3467 }
3468 
3469 /*
3470  *     Function: fcp_ipkt_sema_callback
3471  *
3472  *  Description: Registered as the completion callback function for the FC
3473  *		 transport when the ipkt semaphore is used for sync. This will
3474  *		 cleanup the used data structures, if necessary and wake up
3475  *		 the user thread to complete the transaction.
3476  *
3477  *     Argument: fpkt	FC packet (points to the icmd)
3478  *
3479  * Return Value: None
3480  *
3481  *	Context: User context only
3482  */
3483 static void
3484 fcp_ipkt_sema_callback(struct fc_packet *fpkt)
3485 {
3486 	struct fcp_ipkt	*icmd;
3487 
3488 	icmd = (struct fcp_ipkt *)fpkt->pkt_ulp_private;
3489 
3490 	/*
3491 	 * Wake up user thread
3492 	 */
3493 	sema_v(&(icmd->ipkt_sema));
3494 }
3495 
3496 /*
3497  *     Function: fcp_ipkt_sema_cleanup
3498  *
3499  *  Description: Called to cleanup (if necessary) the data structures used
3500  *		 when ipkt sema is used for sync.  This function will detect
3501  *		 whether the caller is the last thread (via counter) and
3502  *		 cleanup only if necessary.
3503  *
3504  *     Argument: icmd	Internal command packet
3505  *
3506  * Return Value: None
3507  *
3508  *	Context: User context only
3509  */
3510 static void
3511 fcp_ipkt_sema_cleanup(struct fcp_ipkt *icmd)
3512 {
3513 	struct fcp_tgt	*ptgt;
3514 	struct fcp_port	*pptr;
3515 
3516 	ptgt = icmd->ipkt_tgt;
3517 	pptr = icmd->ipkt_port;
3518 
3519 	/*
3520 	 * Acquire data structure
3521 	 */
3522 	mutex_enter(&ptgt->tgt_mutex);
3523 
3524 	/*
3525 	 * Destroy semaphore
3526 	 */
3527 	sema_destroy(&(icmd->ipkt_sema));
3528 
3529 	/*
3530 	 * Cleanup internal packet
3531 	 */
3532 	mutex_exit(&ptgt->tgt_mutex);
3533 	fcp_icmd_free(pptr, icmd);
3534 }
3535 
3536 /*
3537  *     Function: fcp_port_attach
3538  *
3539  *  Description: Called by the transport framework to resume, suspend or
3540  *		 attach a new port.
3541  *
3542  *     Argument: ulph		Port handle
3543  *		 *pinfo		Port information
3544  *		 cmd		Command
3545  *		 s_id		Port ID
3546  *
3547  * Return Value: FC_FAILURE or FC_SUCCESS
3548  */
3549 /*ARGSUSED*/
3550 static int
3551 fcp_port_attach(opaque_t ulph, fc_ulp_port_info_t *pinfo,
3552     fc_attach_cmd_t cmd, uint32_t s_id)
3553 {
3554 	int	instance;
3555 	int	res = FC_FAILURE; /* default result */
3556 
3557 	ASSERT(pinfo != NULL);
3558 
3559 	instance = ddi_get_instance(pinfo->port_dip);
3560 
3561 	switch (cmd) {
3562 	case FC_CMD_ATTACH:
3563 		/*
3564 		 * this port instance attaching for the first time (or after
3565 		 * being detached before)
3566 		 */
3567 		if (fcp_handle_port_attach(ulph, pinfo, s_id,
3568 		    instance) == DDI_SUCCESS) {
3569 			res = FC_SUCCESS;
3570 		} else {
3571 			ASSERT(ddi_get_soft_state(fcp_softstate,
3572 			    instance) == NULL);
3573 		}
3574 		break;
3575 
3576 	case FC_CMD_RESUME:
3577 	case FC_CMD_POWER_UP:
3578 		/*
3579 		 * this port instance was attached and the suspended and
3580 		 * will now be resumed
3581 		 */
3582 		if (fcp_handle_port_resume(ulph, pinfo, s_id, cmd,
3583 		    instance) == DDI_SUCCESS) {
3584 			res = FC_SUCCESS;
3585 		}
3586 		break;
3587 
3588 	default:
3589 		/* shouldn't happen */
3590 		FCP_TRACE(fcp_logq, "fcp",
3591 		    fcp_trace, FCP_BUF_LEVEL_2, 0,
3592 		    "port_attach: unknown cmdcommand: %d", cmd);
3593 		break;
3594 	}
3595 
3596 	/* return result */
3597 	FCP_DTRACE(fcp_logq, "fcp", fcp_trace,
3598 	    FCP_BUF_LEVEL_1, 0, "fcp_port_attach returning %d", res);
3599 
3600 	return (res);
3601 }
3602 
3603 
3604 /*
3605  * detach or suspend this port instance
3606  *
3607  * acquires and releases the global mutex
3608  *
3609  * acquires and releases the mutex for this port
3610  *
3611  * acquires and releases the hotplug mutex for this port
3612  */
3613 /*ARGSUSED*/
3614 static int
3615 fcp_port_detach(opaque_t ulph, fc_ulp_port_info_t *info,
3616     fc_detach_cmd_t cmd)
3617 {
3618 	int			flag;
3619 	int			instance;
3620 	struct fcp_port		*pptr;
3621 
3622 	instance = ddi_get_instance(info->port_dip);
3623 	pptr = ddi_get_soft_state(fcp_softstate, instance);
3624 
3625 	switch (cmd) {
3626 	case FC_CMD_SUSPEND:
3627 		FCP_DTRACE(fcp_logq, "fcp",
3628 		    fcp_trace, FCP_BUF_LEVEL_8, 0,
3629 		    "port suspend called for port %d", instance);
3630 		flag = FCP_STATE_SUSPENDED;
3631 		break;
3632 
3633 	case FC_CMD_POWER_DOWN:
3634 		FCP_DTRACE(fcp_logq, "fcp",
3635 		    fcp_trace, FCP_BUF_LEVEL_8, 0,
3636 		    "port power down called for port %d", instance);
3637 		flag = FCP_STATE_POWER_DOWN;
3638 		break;
3639 
3640 	case FC_CMD_DETACH:
3641 		FCP_DTRACE(fcp_logq, "fcp",
3642 		    fcp_trace, FCP_BUF_LEVEL_8, 0,
3643 		    "port detach called for port %d", instance);
3644 		flag = FCP_STATE_DETACHING;
3645 		break;
3646 
3647 	default:
3648 		/* shouldn't happen */
3649 		return (FC_FAILURE);
3650 	}
3651 	FCP_DTRACE(fcp_logq, "fcp", fcp_trace,
3652 	    FCP_BUF_LEVEL_1, 0, "fcp_port_detach returning");
3653 
3654 	return (fcp_handle_port_detach(pptr, flag, instance));
3655 }
3656 
3657 
3658 /*
3659  * called for ioctls on the transport's devctl interface, and the transport
3660  * has passed it to us
3661  *
3662  * this will only be called for device control ioctls (i.e. hotplugging stuff)
3663  *
3664  * return FC_SUCCESS if we decide to claim the ioctl,
3665  * else return FC_UNCLAIMED
3666  *
3667  * *rval is set iff we decide to claim the ioctl
3668  */
3669 /*ARGSUSED*/
3670 static int
3671 fcp_port_ioctl(opaque_t ulph, opaque_t port_handle, dev_t dev, int cmd,
3672     intptr_t data, int mode, cred_t *credp, int *rval, uint32_t claimed)
3673 {
3674 	int			retval = FC_UNCLAIMED;	/* return value */
3675 	struct fcp_port		*pptr = NULL;		/* our soft state */
3676 	struct devctl_iocdata	*dcp = NULL;		/* for devctl */
3677 	dev_info_t		*cdip;
3678 	mdi_pathinfo_t		*pip = NULL;
3679 	char			*ndi_nm;		/* NDI name */
3680 	char			*ndi_addr;		/* NDI addr */
3681 	int			is_mpxio, circ;
3682 	int			devi_entered = 0;
3683 	time_t			end_time;
3684 
3685 	ASSERT(rval != NULL);
3686 
3687 	FCP_DTRACE(fcp_logq, "fcp",
3688 	    fcp_trace, FCP_BUF_LEVEL_8, 0,
3689 	    "fcp_port_ioctl(cmd=0x%x, claimed=%d)", cmd, claimed);
3690 
3691 	/* if already claimed then forget it */
3692 	if (claimed) {
3693 		/*
3694 		 * for now, if this ioctl has already been claimed, then
3695 		 * we just ignore it
3696 		 */
3697 		return (retval);
3698 	}
3699 
3700 	/* get our port info */
3701 	if ((pptr = fcp_get_port(port_handle)) == NULL) {
3702 		fcp_log(CE_WARN, NULL,
3703 		    "!fcp:Invalid port handle handle in ioctl");
3704 		*rval = ENXIO;
3705 		return (retval);
3706 	}
3707 	is_mpxio = pptr->port_mpxio;
3708 
3709 	switch (cmd) {
3710 	case DEVCTL_BUS_GETSTATE:
3711 	case DEVCTL_BUS_QUIESCE:
3712 	case DEVCTL_BUS_UNQUIESCE:
3713 	case DEVCTL_BUS_RESET:
3714 	case DEVCTL_BUS_RESETALL:
3715 
3716 	case DEVCTL_BUS_DEV_CREATE:
3717 		if (ndi_dc_allochdl((void *)data, &dcp) != NDI_SUCCESS) {
3718 			return (retval);
3719 		}
3720 		break;
3721 
3722 	case DEVCTL_DEVICE_GETSTATE:
3723 	case DEVCTL_DEVICE_OFFLINE:
3724 	case DEVCTL_DEVICE_ONLINE:
3725 	case DEVCTL_DEVICE_REMOVE:
3726 	case DEVCTL_DEVICE_RESET:
3727 		if (ndi_dc_allochdl((void *)data, &dcp) != NDI_SUCCESS) {
3728 			return (retval);
3729 		}
3730 
3731 		ASSERT(dcp != NULL);
3732 
3733 		/* ensure we have a name and address */
3734 		if (((ndi_nm = ndi_dc_getname(dcp)) == NULL) ||
3735 		    ((ndi_addr = ndi_dc_getaddr(dcp)) == NULL)) {
3736 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
3737 			    fcp_trace, FCP_BUF_LEVEL_2, 0,
3738 			    "ioctl: can't get name (%s) or addr (%s)",
3739 			    ndi_nm ? ndi_nm : "<null ptr>",
3740 			    ndi_addr ? ndi_addr : "<null ptr>");
3741 			ndi_dc_freehdl(dcp);
3742 			return (retval);
3743 		}
3744 
3745 
3746 		/* get our child's DIP */
3747 		ASSERT(pptr != NULL);
3748 		if (is_mpxio) {
3749 			mdi_devi_enter(pptr->port_dip, &circ);
3750 		} else {
3751 			ndi_devi_enter(pptr->port_dip, &circ);
3752 		}
3753 		devi_entered = 1;
3754 
3755 		if ((cdip = ndi_devi_find(pptr->port_dip, ndi_nm,
3756 		    ndi_addr)) == NULL) {
3757 			/* Look for virtually enumerated devices. */
3758 			pip = mdi_pi_find(pptr->port_dip, NULL, ndi_addr);
3759 			if (pip == NULL ||
3760 			    ((cdip = mdi_pi_get_client(pip)) == NULL)) {
3761 				*rval = ENXIO;
3762 				goto out;
3763 			}
3764 		}
3765 		break;
3766 
3767 	default:
3768 		*rval = ENOTTY;
3769 		return (retval);
3770 	}
3771 
3772 	/* this ioctl is ours -- process it */
3773 
3774 	retval = FC_SUCCESS;		/* just means we claim the ioctl */
3775 
3776 	/* we assume it will be a success; else we'll set error value */
3777 	*rval = 0;
3778 
3779 
3780 	FCP_DTRACE(fcp_logq, pptr->port_instbuf,
3781 	    fcp_trace, FCP_BUF_LEVEL_8, 0,
3782 	    "ioctl: claiming this one");
3783 
3784 	/* handle ioctls now */
3785 	switch (cmd) {
3786 	case DEVCTL_DEVICE_GETSTATE:
3787 		ASSERT(cdip != NULL);
3788 		ASSERT(dcp != NULL);
3789 		if (ndi_dc_return_dev_state(cdip, dcp) != NDI_SUCCESS) {
3790 			*rval = EFAULT;
3791 		}
3792 		break;
3793 
3794 	case DEVCTL_DEVICE_REMOVE:
3795 	case DEVCTL_DEVICE_OFFLINE: {
3796 		int			flag = 0;
3797 		int			lcount;
3798 		int			tcount;
3799 		struct fcp_pkt	*head = NULL;
3800 		struct fcp_lun	*plun;
3801 		child_info_t		*cip = CIP(cdip);
3802 		int			all = 1;
3803 		struct fcp_lun	*tplun;
3804 		struct fcp_tgt	*ptgt;
3805 
3806 		ASSERT(pptr != NULL);
3807 		ASSERT(cdip != NULL);
3808 
3809 		mutex_enter(&pptr->port_mutex);
3810 		if (pip != NULL) {
3811 			cip = CIP(pip);
3812 		}
3813 		if ((plun = fcp_get_lun_from_cip(pptr, cip)) == NULL) {
3814 			mutex_exit(&pptr->port_mutex);
3815 			*rval = ENXIO;
3816 			break;
3817 		}
3818 
3819 		head = fcp_scan_commands(plun);
3820 		if (head != NULL) {
3821 			fcp_abort_commands(head, LUN_PORT);
3822 		}
3823 		lcount = pptr->port_link_cnt;
3824 		tcount = plun->lun_tgt->tgt_change_cnt;
3825 		mutex_exit(&pptr->port_mutex);
3826 
3827 		if (cmd == DEVCTL_DEVICE_REMOVE) {
3828 			flag = NDI_DEVI_REMOVE;
3829 		}
3830 
3831 		if (is_mpxio) {
3832 			mdi_devi_exit(pptr->port_dip, circ);
3833 		} else {
3834 			ndi_devi_exit(pptr->port_dip, circ);
3835 		}
3836 		devi_entered = 0;
3837 
3838 		*rval = fcp_pass_to_hp_and_wait(pptr, plun, cip,
3839 		    FCP_OFFLINE, lcount, tcount, flag);
3840 
3841 		if (*rval != NDI_SUCCESS) {
3842 			*rval = (*rval == NDI_BUSY) ? EBUSY : EIO;
3843 			break;
3844 		}
3845 
3846 		fcp_update_offline_flags(plun);
3847 
3848 		ptgt = plun->lun_tgt;
3849 		mutex_enter(&ptgt->tgt_mutex);
3850 		for (tplun = ptgt->tgt_lun; tplun != NULL; tplun =
3851 		    tplun->lun_next) {
3852 			mutex_enter(&tplun->lun_mutex);
3853 			if (!(tplun->lun_state & FCP_LUN_OFFLINE)) {
3854 				all = 0;
3855 			}
3856 			mutex_exit(&tplun->lun_mutex);
3857 		}
3858 
3859 		if (all) {
3860 			ptgt->tgt_node_state = FCP_TGT_NODE_NONE;
3861 			/*
3862 			 * The user is unconfiguring/offlining the device.
3863 			 * If fabric and the auto configuration is set
3864 			 * then make sure the user is the only one who
3865 			 * can reconfigure the device.
3866 			 */
3867 			if (FC_TOP_EXTERNAL(pptr->port_topology) &&
3868 			    fcp_enable_auto_configuration) {
3869 				ptgt->tgt_manual_config_only = 1;
3870 			}
3871 		}
3872 		mutex_exit(&ptgt->tgt_mutex);
3873 		break;
3874 	}
3875 
3876 	case DEVCTL_DEVICE_ONLINE: {
3877 		int			lcount;
3878 		int			tcount;
3879 		struct fcp_lun	*plun;
3880 		child_info_t		*cip = CIP(cdip);
3881 
3882 		ASSERT(cdip != NULL);
3883 		ASSERT(pptr != NULL);
3884 
3885 		mutex_enter(&pptr->port_mutex);
3886 		if (pip != NULL) {
3887 			cip = CIP(pip);
3888 		}
3889 		if ((plun = fcp_get_lun_from_cip(pptr, cip)) == NULL) {
3890 			mutex_exit(&pptr->port_mutex);
3891 			*rval = ENXIO;
3892 			break;
3893 		}
3894 		lcount = pptr->port_link_cnt;
3895 		tcount = plun->lun_tgt->tgt_change_cnt;
3896 		mutex_exit(&pptr->port_mutex);
3897 
3898 		/*
3899 		 * The FCP_LUN_ONLINING flag is used in fcp_scsi_start()
3900 		 * to allow the device attach to occur when the device is
3901 		 * FCP_LUN_OFFLINE (so we don't reject the INQUIRY command
3902 		 * from the scsi_probe()).
3903 		 */
3904 		mutex_enter(&LUN_TGT->tgt_mutex);
3905 		plun->lun_state |= FCP_LUN_ONLINING;
3906 		mutex_exit(&LUN_TGT->tgt_mutex);
3907 
3908 		if (is_mpxio) {
3909 			mdi_devi_exit(pptr->port_dip, circ);
3910 		} else {
3911 			ndi_devi_exit(pptr->port_dip, circ);
3912 		}
3913 		devi_entered = 0;
3914 
3915 		*rval = fcp_pass_to_hp_and_wait(pptr, plun, cip,
3916 		    FCP_ONLINE, lcount, tcount, 0);
3917 
3918 		if (*rval != NDI_SUCCESS) {
3919 			/* Reset the FCP_LUN_ONLINING bit */
3920 			mutex_enter(&LUN_TGT->tgt_mutex);
3921 			plun->lun_state &= ~FCP_LUN_ONLINING;
3922 			mutex_exit(&LUN_TGT->tgt_mutex);
3923 			*rval = EIO;
3924 			break;
3925 		}
3926 		mutex_enter(&LUN_TGT->tgt_mutex);
3927 		plun->lun_state &= ~(FCP_LUN_OFFLINE | FCP_LUN_BUSY |
3928 		    FCP_LUN_ONLINING);
3929 		mutex_exit(&LUN_TGT->tgt_mutex);
3930 		break;
3931 	}
3932 
3933 	case DEVCTL_BUS_DEV_CREATE: {
3934 		uchar_t			*bytes = NULL;
3935 		uint_t			nbytes;
3936 		struct fcp_tgt		*ptgt = NULL;
3937 		struct fcp_lun		*plun = NULL;
3938 		dev_info_t		*useless_dip = NULL;
3939 
3940 		*rval = ndi_dc_devi_create(dcp, pptr->port_dip,
3941 		    DEVCTL_CONSTRUCT, &useless_dip);
3942 		if (*rval != 0 || useless_dip == NULL) {
3943 			break;
3944 		}
3945 
3946 		if ((ddi_prop_lookup_byte_array(DDI_DEV_T_ANY, useless_dip,
3947 		    DDI_PROP_DONTPASS | DDI_PROP_NOTPROM, PORT_WWN_PROP, &bytes,
3948 		    &nbytes) != DDI_PROP_SUCCESS) || nbytes != FC_WWN_SIZE) {
3949 			*rval = EINVAL;
3950 			(void) ndi_devi_free(useless_dip);
3951 			if (bytes != NULL) {
3952 				ddi_prop_free(bytes);
3953 			}
3954 			break;
3955 		}
3956 
3957 		*rval = fcp_create_on_demand(pptr, bytes);
3958 		if (*rval == 0) {
3959 			mutex_enter(&pptr->port_mutex);
3960 			ptgt = fcp_lookup_target(pptr, (uchar_t *)bytes);
3961 			if (ptgt) {
3962 				/*
3963 				 * We now have a pointer to the target that
3964 				 * was created. Lets point to the first LUN on
3965 				 * this new target.
3966 				 */
3967 				mutex_enter(&ptgt->tgt_mutex);
3968 
3969 				plun = ptgt->tgt_lun;
3970 				/*
3971 				 * There may be stale/offline LUN entries on
3972 				 * this list (this is by design) and so we have
3973 				 * to make sure we point to the first online
3974 				 * LUN
3975 				 */
3976 				while (plun &&
3977 				    plun->lun_state & FCP_LUN_OFFLINE) {
3978 					plun = plun->lun_next;
3979 				}
3980 
3981 				mutex_exit(&ptgt->tgt_mutex);
3982 			}
3983 			mutex_exit(&pptr->port_mutex);
3984 		}
3985 
3986 		if (*rval == 0 && ptgt && plun) {
3987 			mutex_enter(&plun->lun_mutex);
3988 			/*
3989 			 * Allow up to fcp_lun_ready_retry seconds to
3990 			 * configure all the luns behind the target.
3991 			 *
3992 			 * The intent here is to allow targets with long
3993 			 * reboot/reset-recovery times to become available
3994 			 * while limiting the maximum wait time for an
3995 			 * unresponsive target.
3996 			 */
3997 			end_time = ddi_get_lbolt() +
3998 			    SEC_TO_TICK(fcp_lun_ready_retry);
3999 
4000 			while (ddi_get_lbolt() < end_time) {
4001 				retval = FC_SUCCESS;
4002 
4003 				/*
4004 				 * The new ndi interfaces for on-demand creation
4005 				 * are inflexible, Do some more work to pass on
4006 				 * a path name of some LUN (design is broken !)
4007 				 */
4008 				if (plun->lun_cip) {
4009 					if (plun->lun_mpxio == 0) {
4010 						cdip = DIP(plun->lun_cip);
4011 					} else {
4012 						cdip = mdi_pi_get_client(
4013 						    PIP(plun->lun_cip));
4014 					}
4015 					if (cdip == NULL) {
4016 						*rval = ENXIO;
4017 						break;
4018 					}
4019 
4020 					if (!i_ddi_devi_attached(cdip)) {
4021 						mutex_exit(&plun->lun_mutex);
4022 						delay(drv_usectohz(1000000));
4023 						mutex_enter(&plun->lun_mutex);
4024 					} else {
4025 						/*
4026 						 * This Lun is ready, lets
4027 						 * check the next one.
4028 						 */
4029 						mutex_exit(&plun->lun_mutex);
4030 						plun = plun->lun_next;
4031 						while (plun && (plun->lun_state
4032 						    & FCP_LUN_OFFLINE)) {
4033 							plun = plun->lun_next;
4034 						}
4035 						if (!plun) {
4036 							break;
4037 						}
4038 						mutex_enter(&plun->lun_mutex);
4039 					}
4040 				} else {
4041 					/*
4042 					 * lun_cip field for a valid lun
4043 					 * should never be NULL. Fail the
4044 					 * command.
4045 					 */
4046 					*rval = ENXIO;
4047 					break;
4048 				}
4049 			}
4050 			if (plun) {
4051 				mutex_exit(&plun->lun_mutex);
4052 			} else {
4053 				char devnm[MAXNAMELEN];
4054 				int nmlen;
4055 
4056 				nmlen = snprintf(devnm, MAXNAMELEN, "%s@%s",
4057 				    ddi_node_name(cdip),
4058 				    ddi_get_name_addr(cdip));
4059 
4060 				if (copyout(&devnm, dcp->cpyout_buf, nmlen) !=
4061 				    0) {
4062 					*rval = EFAULT;
4063 				}
4064 			}
4065 		} else {
4066 			int	i;
4067 			char	buf[25];
4068 
4069 			for (i = 0; i < FC_WWN_SIZE; i++) {
4070 				(void) sprintf(&buf[i << 1], "%02x", bytes[i]);
4071 			}
4072 
4073 			fcp_log(CE_WARN, pptr->port_dip,
4074 			    "!Failed to create nodes for pwwn=%s; error=%x",
4075 			    buf, *rval);
4076 		}
4077 
4078 		(void) ndi_devi_free(useless_dip);
4079 		ddi_prop_free(bytes);
4080 		break;
4081 	}
4082 
4083 	case DEVCTL_DEVICE_RESET: {
4084 		struct fcp_lun	*plun;
4085 		struct scsi_address	ap;
4086 		child_info_t		*cip = CIP(cdip);
4087 
4088 		ASSERT(cdip != NULL);
4089 		ASSERT(pptr != NULL);
4090 		mutex_enter(&pptr->port_mutex);
4091 		if (pip != NULL) {
4092 			cip = CIP(pip);
4093 		}
4094 		if ((plun = fcp_get_lun_from_cip(pptr, cip)) == NULL) {
4095 			mutex_exit(&pptr->port_mutex);
4096 			*rval = ENXIO;
4097 			break;
4098 		}
4099 		mutex_exit(&pptr->port_mutex);
4100 
4101 		mutex_enter(&plun->lun_tgt->tgt_mutex);
4102 		if (!(plun->lun_state & FCP_SCSI_LUN_TGT_INIT)) {
4103 			mutex_exit(&plun->lun_tgt->tgt_mutex);
4104 			*rval = ENXIO;
4105 			break;
4106 		}
4107 		ap.a_hba_tran = plun->lun_tran;
4108 		ASSERT(pptr->port_tran != NULL);
4109 		mutex_exit(&plun->lun_tgt->tgt_mutex);
4110 
4111 		/*
4112 		 * There is a chance lun_tran is NULL at this point. So check
4113 		 * for it. If it is NULL, it basically means that the tgt has
4114 		 * been freed. So, just return a "No such device or address"
4115 		 * error.
4116 		 */
4117 		if (ap.a_hba_tran == NULL) {
4118 			*rval = ENXIO;
4119 			break;
4120 		}
4121 
4122 		/*
4123 		 * set up ap so that fcp_reset can figure out
4124 		 * which target to reset
4125 		 */
4126 		if (fcp_scsi_reset(&ap, RESET_TARGET) == FALSE) {
4127 			*rval = EIO;
4128 		}
4129 		break;
4130 	}
4131 
4132 	case DEVCTL_BUS_GETSTATE:
4133 		ASSERT(dcp != NULL);
4134 		ASSERT(pptr != NULL);
4135 		ASSERT(pptr->port_dip != NULL);
4136 		if (ndi_dc_return_bus_state(pptr->port_dip, dcp) !=
4137 		    NDI_SUCCESS) {
4138 			*rval = EFAULT;
4139 		}
4140 		break;
4141 
4142 	case DEVCTL_BUS_QUIESCE:
4143 	case DEVCTL_BUS_UNQUIESCE:
4144 		*rval = ENOTSUP;
4145 		break;
4146 
4147 	case DEVCTL_BUS_RESET:
4148 	case DEVCTL_BUS_RESETALL:
4149 		ASSERT(pptr != NULL);
4150 		(void) fcp_linkreset(pptr, NULL,  KM_SLEEP);
4151 		break;
4152 
4153 	default:
4154 		ASSERT(dcp != NULL);
4155 		*rval = ENOTTY;
4156 		break;
4157 	}
4158 
4159 	/* all done -- clean up and return */
4160 out:	if (devi_entered) {
4161 		if (is_mpxio) {
4162 			mdi_devi_exit(pptr->port_dip, circ);
4163 		} else {
4164 			ndi_devi_exit(pptr->port_dip, circ);
4165 		}
4166 	}
4167 
4168 	if (dcp != NULL) {
4169 		ndi_dc_freehdl(dcp);
4170 	}
4171 
4172 	return (retval);
4173 }
4174 
4175 
4176 /*ARGSUSED*/
4177 static int
4178 fcp_els_callback(opaque_t ulph, opaque_t port_handle, fc_unsol_buf_t *buf,
4179     uint32_t claimed)
4180 {
4181 	uchar_t			r_ctl;
4182 	uchar_t			ls_code;
4183 	struct fcp_port	*pptr;
4184 
4185 	if ((pptr = fcp_get_port(port_handle)) == NULL || claimed) {
4186 		return (FC_UNCLAIMED);
4187 	}
4188 
4189 	mutex_enter(&pptr->port_mutex);
4190 	if (pptr->port_state & (FCP_STATE_DETACHING |
4191 	    FCP_STATE_SUSPENDED | FCP_STATE_POWER_DOWN)) {
4192 		mutex_exit(&pptr->port_mutex);
4193 		return (FC_UNCLAIMED);
4194 	}
4195 	mutex_exit(&pptr->port_mutex);
4196 
4197 	r_ctl = buf->ub_frame.r_ctl;
4198 
4199 	switch (r_ctl & R_CTL_ROUTING) {
4200 	case R_CTL_EXTENDED_SVC:
4201 		if (r_ctl == R_CTL_ELS_REQ) {
4202 			ls_code = buf->ub_buffer[0];
4203 
4204 			switch (ls_code) {
4205 			case LA_ELS_PRLI:
4206 				/*
4207 				 * We really don't care if something fails.
4208 				 * If the PRLI was not sent out, then the
4209 				 * other end will time it out.
4210 				 */
4211 				if (fcp_unsol_prli(pptr, buf) == FC_SUCCESS) {
4212 					return (FC_SUCCESS);
4213 				}
4214 				return (FC_UNCLAIMED);
4215 				/* NOTREACHED */
4216 
4217 			default:
4218 				break;
4219 			}
4220 		}
4221 		/* FALLTHROUGH */
4222 
4223 	default:
4224 		return (FC_UNCLAIMED);
4225 	}
4226 }
4227 
4228 
4229 /*ARGSUSED*/
4230 static int
4231 fcp_data_callback(opaque_t ulph, opaque_t port_handle, fc_unsol_buf_t *buf,
4232     uint32_t claimed)
4233 {
4234 	return (FC_UNCLAIMED);
4235 }
4236 
4237 /*
4238  *     Function: fcp_statec_callback
4239  *
4240  *  Description: The purpose of this function is to handle a port state change.
4241  *		 It is called from fp/fctl and, in a few instances, internally.
4242  *
4243  *     Argument: ulph		fp/fctl port handle
4244  *		 port_handle	fcp_port structure
4245  *		 port_state	Physical state of the port
4246  *		 port_top	Topology
4247  *		 *devlist	Pointer to the first entry of a table
4248  *				containing the remote ports that can be
4249  *				reached.
4250  *		 dev_cnt	Number of entries pointed by devlist.
4251  *		 port_sid	Port ID of the local port.
4252  *
4253  * Return Value: None
4254  */
4255 /*ARGSUSED*/
4256 static void
4257 fcp_statec_callback(opaque_t ulph, opaque_t port_handle,
4258     uint32_t port_state, uint32_t port_top, fc_portmap_t *devlist,
4259     uint32_t dev_cnt, uint32_t port_sid)
4260 {
4261 	uint32_t		link_count;
4262 	int			map_len = 0;
4263 	struct fcp_port	*pptr;
4264 	fcp_map_tag_t		*map_tag = NULL;
4265 
4266 	if ((pptr = fcp_get_port(port_handle)) == NULL) {
4267 		fcp_log(CE_WARN, NULL, "!Invalid port handle in callback");
4268 		return;			/* nothing to work with! */
4269 	}
4270 
4271 	FCP_TRACE(fcp_logq, pptr->port_instbuf,
4272 	    fcp_trace, FCP_BUF_LEVEL_2, 0,
4273 	    "fcp_statec_callback: port state/dev_cnt/top ="
4274 	    "%d/%d/%d", FC_PORT_STATE_MASK(port_state),
4275 	    dev_cnt, port_top);
4276 
4277 	mutex_enter(&pptr->port_mutex);
4278 
4279 	/*
4280 	 * If a thread is in detach, don't do anything.
4281 	 */
4282 	if (pptr->port_state & (FCP_STATE_DETACHING |
4283 	    FCP_STATE_SUSPENDED | FCP_STATE_POWER_DOWN)) {
4284 		mutex_exit(&pptr->port_mutex);
4285 		return;
4286 	}
4287 
4288 	/*
4289 	 * First thing we do is set the FCP_STATE_IN_CB_DEVC flag so that if
4290 	 * init_pkt is called, it knows whether or not the target's status
4291 	 * (or pd) might be changing.
4292 	 */
4293 
4294 	if (FC_PORT_STATE_MASK(port_state) == FC_STATE_DEVICE_CHANGE) {
4295 		pptr->port_state |= FCP_STATE_IN_CB_DEVC;
4296 	}
4297 
4298 	/*
4299 	 * the transport doesn't allocate or probe unless being
4300 	 * asked to by either the applications or ULPs
4301 	 *
4302 	 * in cases where the port is OFFLINE at the time of port
4303 	 * attach callback and the link comes ONLINE later, for
4304 	 * easier automatic node creation (i.e. without you having to
4305 	 * go out and run the utility to perform LOGINs) the
4306 	 * following conditional is helpful
4307 	 */
4308 	pptr->port_phys_state = port_state;
4309 
4310 	if (dev_cnt) {
4311 		mutex_exit(&pptr->port_mutex);
4312 
4313 		map_len = sizeof (*map_tag) * dev_cnt;
4314 		map_tag = kmem_alloc(map_len, KM_NOSLEEP);
4315 		if (map_tag == NULL) {
4316 			fcp_log(CE_WARN, pptr->port_dip,
4317 			    "!fcp%d: failed to allocate for map tags; "
4318 			    " state change will not be processed",
4319 			    pptr->port_instance);
4320 
4321 			mutex_enter(&pptr->port_mutex);
4322 			pptr->port_state &= ~FCP_STATE_IN_CB_DEVC;
4323 			mutex_exit(&pptr->port_mutex);
4324 
4325 			return;
4326 		}
4327 
4328 		mutex_enter(&pptr->port_mutex);
4329 	}
4330 
4331 	if (pptr->port_id != port_sid) {
4332 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
4333 		    fcp_trace, FCP_BUF_LEVEL_3, 0,
4334 		    "fcp: Port S_ID=0x%x => 0x%x", pptr->port_id,
4335 		    port_sid);
4336 		/*
4337 		 * The local port changed ID. It is the first time a port ID
4338 		 * is assigned or something drastic happened.  We might have
4339 		 * been unplugged and replugged on another loop or fabric port
4340 		 * or somebody grabbed the AL_PA we had or somebody rezoned
4341 		 * the fabric we were plugged into.
4342 		 */
4343 		pptr->port_id = port_sid;
4344 	}
4345 
4346 	switch (FC_PORT_STATE_MASK(port_state)) {
4347 	case FC_STATE_OFFLINE:
4348 	case FC_STATE_RESET_REQUESTED:
4349 		/*
4350 		 * link has gone from online to offline -- just update the
4351 		 * state of this port to BUSY and MARKed to go offline
4352 		 */
4353 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
4354 		    fcp_trace, FCP_BUF_LEVEL_3, 0,
4355 		    "link went offline");
4356 		if ((pptr->port_state & FCP_STATE_OFFLINE) && dev_cnt) {
4357 			/*
4358 			 * We were offline a while ago and this one
4359 			 * seems to indicate that the loop has gone
4360 			 * dead forever.
4361 			 */
4362 			pptr->port_tmp_cnt += dev_cnt;
4363 			pptr->port_state &= ~FCP_STATE_OFFLINE;
4364 			pptr->port_state |= FCP_STATE_INIT;
4365 			link_count = pptr->port_link_cnt;
4366 			fcp_handle_devices(pptr, devlist, dev_cnt,
4367 			    link_count, map_tag, FCP_CAUSE_LINK_DOWN);
4368 		} else {
4369 			pptr->port_link_cnt++;
4370 			ASSERT(!(pptr->port_state & FCP_STATE_SUSPENDED));
4371 			fcp_update_state(pptr, (FCP_LUN_BUSY |
4372 			    FCP_LUN_MARK), FCP_CAUSE_LINK_DOWN);
4373 			if (pptr->port_mpxio) {
4374 				fcp_update_mpxio_path_verifybusy(pptr);
4375 			}
4376 			pptr->port_state |= FCP_STATE_OFFLINE;
4377 			pptr->port_state &=
4378 			    ~(FCP_STATE_ONLINING | FCP_STATE_ONLINE);
4379 			pptr->port_tmp_cnt = 0;
4380 		}
4381 		mutex_exit(&pptr->port_mutex);
4382 		break;
4383 
4384 	case FC_STATE_ONLINE:
4385 	case FC_STATE_LIP:
4386 	case FC_STATE_LIP_LBIT_SET:
4387 		/*
4388 		 * link has gone from offline to online
4389 		 */
4390 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
4391 		    fcp_trace, FCP_BUF_LEVEL_3, 0,
4392 		    "link went online");
4393 
4394 		pptr->port_link_cnt++;
4395 
4396 		while (pptr->port_ipkt_cnt) {
4397 			mutex_exit(&pptr->port_mutex);
4398 			delay(drv_usectohz(1000000));
4399 			mutex_enter(&pptr->port_mutex);
4400 		}
4401 
4402 		pptr->port_topology = port_top;
4403 
4404 		/*
4405 		 * The state of the targets and luns accessible through this
4406 		 * port is updated.
4407 		 */
4408 		fcp_update_state(pptr, FCP_LUN_BUSY | FCP_LUN_MARK,
4409 		    FCP_CAUSE_LINK_CHANGE);
4410 
4411 		pptr->port_state &= ~(FCP_STATE_INIT | FCP_STATE_OFFLINE);
4412 		pptr->port_state |= FCP_STATE_ONLINING;
4413 		pptr->port_tmp_cnt = dev_cnt;
4414 		link_count = pptr->port_link_cnt;
4415 
4416 		pptr->port_deadline = fcp_watchdog_time +
4417 		    FCP_ICMD_DEADLINE;
4418 
4419 		if (!dev_cnt) {
4420 			/*
4421 			 * We go directly to the online state if no remote
4422 			 * ports were discovered.
4423 			 */
4424 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
4425 			    fcp_trace, FCP_BUF_LEVEL_3, 0,
4426 			    "No remote ports discovered");
4427 
4428 			pptr->port_state &= ~FCP_STATE_ONLINING;
4429 			pptr->port_state |= FCP_STATE_ONLINE;
4430 		}
4431 
4432 		switch (port_top) {
4433 		case FC_TOP_FABRIC:
4434 		case FC_TOP_PUBLIC_LOOP:
4435 		case FC_TOP_PRIVATE_LOOP:
4436 		case FC_TOP_PT_PT:
4437 
4438 			if (pptr->port_state & FCP_STATE_NS_REG_FAILED) {
4439 				fcp_retry_ns_registry(pptr, port_sid);
4440 			}
4441 
4442 			fcp_handle_devices(pptr, devlist, dev_cnt, link_count,
4443 			    map_tag, FCP_CAUSE_LINK_CHANGE);
4444 			break;
4445 
4446 		default:
4447 			/*
4448 			 * We got here because we were provided with an unknown
4449 			 * topology.
4450 			 */
4451 			if (pptr->port_state & FCP_STATE_NS_REG_FAILED) {
4452 				pptr->port_state &= ~FCP_STATE_NS_REG_FAILED;
4453 			}
4454 
4455 			pptr->port_tmp_cnt -= dev_cnt;
4456 			fcp_log(CE_WARN, pptr->port_dip,
4457 			    "!unknown/unsupported topology (0x%x)", port_top);
4458 			break;
4459 		}
4460 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
4461 		    fcp_trace, FCP_BUF_LEVEL_3, 0,
4462 		    "Notify ssd of the reset to reinstate the reservations");
4463 
4464 		scsi_hba_reset_notify_callback(&pptr->port_mutex,
4465 		    &pptr->port_reset_notify_listf);
4466 
4467 		mutex_exit(&pptr->port_mutex);
4468 
4469 		break;
4470 
4471 	case FC_STATE_RESET:
4472 		ASSERT(pptr->port_state & FCP_STATE_OFFLINE);
4473 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
4474 		    fcp_trace, FCP_BUF_LEVEL_3, 0,
4475 		    "RESET state, waiting for Offline/Online state_cb");
4476 		mutex_exit(&pptr->port_mutex);
4477 		break;
4478 
4479 	case FC_STATE_DEVICE_CHANGE:
4480 		/*
4481 		 * We come here when an application has requested
4482 		 * Dynamic node creation/deletion in Fabric connectivity.
4483 		 */
4484 		if (pptr->port_state & (FCP_STATE_OFFLINE |
4485 		    FCP_STATE_INIT)) {
4486 			/*
4487 			 * This case can happen when the FCTL is in the
4488 			 * process of giving us on online and the host on
4489 			 * the other side issues a PLOGI/PLOGO. Ideally
4490 			 * the state changes should be serialized unless
4491 			 * they are opposite (online-offline).
4492 			 * The transport will give us a final state change
4493 			 * so we can ignore this for the time being.
4494 			 */
4495 			pptr->port_state &= ~FCP_STATE_IN_CB_DEVC;
4496 			mutex_exit(&pptr->port_mutex);
4497 			break;
4498 		}
4499 
4500 		if (pptr->port_state & FCP_STATE_NS_REG_FAILED) {
4501 			fcp_retry_ns_registry(pptr, port_sid);
4502 		}
4503 
4504 		/*
4505 		 * Extend the deadline under steady state conditions
4506 		 * to provide more time for the device-change-commands
4507 		 */
4508 		if (!pptr->port_ipkt_cnt) {
4509 			pptr->port_deadline = fcp_watchdog_time +
4510 			    FCP_ICMD_DEADLINE;
4511 		}
4512 
4513 		/*
4514 		 * There is another race condition here, where if we were
4515 		 * in ONLINEING state and a devices in the map logs out,
4516 		 * fp will give another state change as DEVICE_CHANGE
4517 		 * and OLD. This will result in that target being offlined.
4518 		 * The pd_handle is freed. If from the first statec callback
4519 		 * we were going to fire a PLOGI/PRLI, the system will
4520 		 * panic in fc_ulp_transport with invalid pd_handle.
4521 		 * The fix is to check for the link_cnt before issuing
4522 		 * any command down.
4523 		 */
4524 		fcp_update_targets(pptr, devlist, dev_cnt,
4525 		    FCP_LUN_BUSY | FCP_LUN_MARK, FCP_CAUSE_TGT_CHANGE);
4526 
4527 		link_count = pptr->port_link_cnt;
4528 
4529 		fcp_handle_devices(pptr, devlist, dev_cnt,
4530 		    link_count, map_tag, FCP_CAUSE_TGT_CHANGE);
4531 
4532 		pptr->port_state &= ~FCP_STATE_IN_CB_DEVC;
4533 
4534 		mutex_exit(&pptr->port_mutex);
4535 		break;
4536 
4537 	case FC_STATE_TARGET_PORT_RESET:
4538 		if (pptr->port_state & FCP_STATE_NS_REG_FAILED) {
4539 			fcp_retry_ns_registry(pptr, port_sid);
4540 		}
4541 
4542 		/* Do nothing else */
4543 		mutex_exit(&pptr->port_mutex);
4544 		break;
4545 
4546 	default:
4547 		fcp_log(CE_WARN, pptr->port_dip,
4548 		    "!Invalid state change=0x%x", port_state);
4549 		mutex_exit(&pptr->port_mutex);
4550 		break;
4551 	}
4552 
4553 	if (map_tag) {
4554 		kmem_free(map_tag, map_len);
4555 	}
4556 }
4557 
4558 /*
4559  *     Function: fcp_handle_devices
4560  *
4561  *  Description: This function updates the devices currently known by
4562  *		 walking the list provided by the caller.  The list passed
4563  *		 by the caller is supposed to be the list of reachable
4564  *		 devices.
4565  *
4566  *     Argument: *pptr		Fcp port structure.
4567  *		 *devlist	Pointer to the first entry of a table
4568  *				containing the remote ports that can be
4569  *				reached.
4570  *		 dev_cnt	Number of entries pointed by devlist.
4571  *		 link_cnt	Link state count.
4572  *		 *map_tag	Array of fcp_map_tag_t structures.
4573  *		 cause		What caused this function to be called.
4574  *
4575  * Return Value: None
4576  *
4577  *	  Notes: The pptr->port_mutex must be held.
4578  */
4579 static void
4580 fcp_handle_devices(struct fcp_port *pptr, fc_portmap_t devlist[],
4581     uint32_t dev_cnt, int link_cnt, fcp_map_tag_t *map_tag, int cause)
4582 {
4583 	int			i;
4584 	int			check_finish_init = 0;
4585 	fc_portmap_t		*map_entry;
4586 	struct fcp_tgt	*ptgt = NULL;
4587 
4588 	FCP_TRACE(fcp_logq, pptr->port_instbuf,
4589 	    fcp_trace, FCP_BUF_LEVEL_3, 0,
4590 	    "fcp_handle_devices: called for %d dev(s)", dev_cnt);
4591 
4592 	if (dev_cnt) {
4593 		ASSERT(map_tag != NULL);
4594 	}
4595 
4596 	/*
4597 	 * The following code goes through the list of remote ports that are
4598 	 * accessible through this (pptr) local port (The list walked is the
4599 	 * one provided by the caller which is the list of the remote ports
4600 	 * currently reachable).  It checks if any of them was already
4601 	 * known by looking for the corresponding target structure based on
4602 	 * the world wide name.	 If a target is part of the list it is tagged
4603 	 * (ptgt->tgt_aux_state = FCP_TGT_TAGGED).
4604 	 *
4605 	 * Old comment
4606 	 * -----------
4607 	 * Before we drop port mutex; we MUST get the tags updated; This
4608 	 * two step process is somewhat slow, but more reliable.
4609 	 */
4610 	for (i = 0; (i < dev_cnt) && (pptr->port_link_cnt == link_cnt); i++) {
4611 		map_entry = &(devlist[i]);
4612 
4613 		/*
4614 		 * get ptr to this map entry in our port's
4615 		 * list (if any)
4616 		 */
4617 		ptgt = fcp_lookup_target(pptr,
4618 		    (uchar_t *)&(map_entry->map_pwwn));
4619 
4620 		if (ptgt) {
4621 			map_tag[i] = ptgt->tgt_change_cnt;
4622 			if (cause == FCP_CAUSE_LINK_CHANGE) {
4623 				ptgt->tgt_aux_state = FCP_TGT_TAGGED;
4624 			}
4625 		}
4626 	}
4627 
4628 	/*
4629 	 * At this point we know which devices of the new list were already
4630 	 * known (The field tgt_aux_state of the target structure has been
4631 	 * set to FCP_TGT_TAGGED).
4632 	 *
4633 	 * The following code goes through the list of targets currently known
4634 	 * by the local port (the list is actually a hashing table).  If a
4635 	 * target is found and is not tagged, it means the target cannot
4636 	 * be reached anymore through the local port (pptr).  It is offlined.
4637 	 * The offlining only occurs if the cause is FCP_CAUSE_LINK_CHANGE.
4638 	 */
4639 	for (i = 0; i < FCP_NUM_HASH; i++) {
4640 		for (ptgt = pptr->port_tgt_hash_table[i]; ptgt != NULL;
4641 		    ptgt = ptgt->tgt_next) {
4642 			mutex_enter(&ptgt->tgt_mutex);
4643 			if ((ptgt->tgt_aux_state != FCP_TGT_TAGGED) &&
4644 			    (cause == FCP_CAUSE_LINK_CHANGE) &&
4645 			    !(ptgt->tgt_state & FCP_TGT_OFFLINE)) {
4646 				fcp_offline_target_now(pptr, ptgt,
4647 				    link_cnt, ptgt->tgt_change_cnt, 0);
4648 			}
4649 			mutex_exit(&ptgt->tgt_mutex);
4650 		}
4651 	}
4652 
4653 	/*
4654 	 * At this point, the devices that were known but cannot be reached
4655 	 * anymore, have most likely been offlined.
4656 	 *
4657 	 * The following section of code seems to go through the list of
4658 	 * remote ports that can now be reached.  For every single one it
4659 	 * checks if it is already known or if it is a new port.
4660 	 */
4661 	for (i = 0; (i < dev_cnt) && (pptr->port_link_cnt == link_cnt); i++) {
4662 
4663 		if (check_finish_init) {
4664 			ASSERT(i > 0);
4665 			(void) fcp_call_finish_init_held(pptr, ptgt, link_cnt,
4666 			    map_tag[i - 1], cause);
4667 			check_finish_init = 0;
4668 		}
4669 
4670 		/* get a pointer to this map entry */
4671 		map_entry = &(devlist[i]);
4672 
4673 		/*
4674 		 * Check for the duplicate map entry flag. If we have marked
4675 		 * this entry as a duplicate we skip it since the correct
4676 		 * (perhaps even same) state change will be encountered
4677 		 * later in the list.
4678 		 */
4679 		if (map_entry->map_flags & PORT_DEVICE_DUPLICATE_MAP_ENTRY) {
4680 			continue;
4681 		}
4682 
4683 		/* get ptr to this map entry in our port's list (if any) */
4684 		ptgt = fcp_lookup_target(pptr,
4685 		    (uchar_t *)&(map_entry->map_pwwn));
4686 
4687 		if (ptgt) {
4688 			/*
4689 			 * This device was already known.  The field
4690 			 * tgt_aux_state is reset (was probably set to
4691 			 * FCP_TGT_TAGGED previously in this routine).
4692 			 */
4693 			ptgt->tgt_aux_state = 0;
4694 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
4695 			    fcp_trace, FCP_BUF_LEVEL_3, 0,
4696 			    "handle_devices: map did/state/type/flags = "
4697 			    "0x%x/0x%x/0x%x/0x%x, tgt_d_id=0x%x, "
4698 			    "tgt_state=%d",
4699 			    map_entry->map_did.port_id, map_entry->map_state,
4700 			    map_entry->map_type, map_entry->map_flags,
4701 			    ptgt->tgt_d_id, ptgt->tgt_state);
4702 		}
4703 
4704 		if (map_entry->map_type == PORT_DEVICE_OLD ||
4705 		    map_entry->map_type == PORT_DEVICE_NEW ||
4706 		    map_entry->map_type == PORT_DEVICE_CHANGED) {
4707 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
4708 			    fcp_trace, FCP_BUF_LEVEL_2, 0,
4709 			    "map_type=%x, did = %x",
4710 			    map_entry->map_type,
4711 			    map_entry->map_did.port_id);
4712 		}
4713 
4714 		switch (map_entry->map_type) {
4715 		case PORT_DEVICE_NOCHANGE:
4716 		case PORT_DEVICE_USER_CREATE:
4717 		case PORT_DEVICE_USER_LOGIN:
4718 		case PORT_DEVICE_NEW:
4719 			FCP_TGT_TRACE(ptgt, map_tag[i], FCP_TGT_TRACE_1);
4720 
4721 			if (fcp_handle_mapflags(pptr, ptgt, map_entry,
4722 			    link_cnt, (ptgt) ? map_tag[i] : 0,
4723 			    cause) == TRUE) {
4724 
4725 				FCP_TGT_TRACE(ptgt, map_tag[i],
4726 				    FCP_TGT_TRACE_2);
4727 				check_finish_init++;
4728 			}
4729 			break;
4730 
4731 		case PORT_DEVICE_OLD:
4732 			if (ptgt != NULL) {
4733 				FCP_TGT_TRACE(ptgt, map_tag[i],
4734 				    FCP_TGT_TRACE_3);
4735 
4736 				mutex_enter(&ptgt->tgt_mutex);
4737 				if (!(ptgt->tgt_state & FCP_TGT_OFFLINE)) {
4738 					/*
4739 					 * Must do an in-line wait for I/Os
4740 					 * to get drained
4741 					 */
4742 					mutex_exit(&ptgt->tgt_mutex);
4743 					mutex_exit(&pptr->port_mutex);
4744 
4745 					mutex_enter(&ptgt->tgt_mutex);
4746 					while (ptgt->tgt_ipkt_cnt ||
4747 					    fcp_outstanding_lun_cmds(ptgt)
4748 					    == FC_SUCCESS) {
4749 						mutex_exit(&ptgt->tgt_mutex);
4750 						delay(drv_usectohz(1000000));
4751 						mutex_enter(&ptgt->tgt_mutex);
4752 					}
4753 					mutex_exit(&ptgt->tgt_mutex);
4754 
4755 					mutex_enter(&pptr->port_mutex);
4756 					mutex_enter(&ptgt->tgt_mutex);
4757 
4758 					(void) fcp_offline_target(pptr, ptgt,
4759 					    link_cnt, map_tag[i], 0, 0);
4760 				}
4761 				mutex_exit(&ptgt->tgt_mutex);
4762 			}
4763 			check_finish_init++;
4764 			break;
4765 
4766 		case PORT_DEVICE_USER_DELETE:
4767 		case PORT_DEVICE_USER_LOGOUT:
4768 			if (ptgt != NULL) {
4769 				FCP_TGT_TRACE(ptgt, map_tag[i],
4770 				    FCP_TGT_TRACE_4);
4771 
4772 				mutex_enter(&ptgt->tgt_mutex);
4773 				if (!(ptgt->tgt_state & FCP_TGT_OFFLINE)) {
4774 					(void) fcp_offline_target(pptr, ptgt,
4775 					    link_cnt, map_tag[i], 1, 0);
4776 				}
4777 				mutex_exit(&ptgt->tgt_mutex);
4778 			}
4779 			check_finish_init++;
4780 			break;
4781 
4782 		case PORT_DEVICE_CHANGED:
4783 			if (ptgt != NULL) {
4784 				FCP_TGT_TRACE(ptgt, map_tag[i],
4785 				    FCP_TGT_TRACE_5);
4786 
4787 				if (fcp_device_changed(pptr, ptgt,
4788 				    map_entry, link_cnt, map_tag[i],
4789 				    cause) == TRUE) {
4790 					check_finish_init++;
4791 				}
4792 			} else {
4793 				if (fcp_handle_mapflags(pptr, ptgt,
4794 				    map_entry, link_cnt, 0, cause) == TRUE) {
4795 					check_finish_init++;
4796 				}
4797 			}
4798 			break;
4799 
4800 		default:
4801 			fcp_log(CE_WARN, pptr->port_dip,
4802 			    "!Invalid map_type=0x%x", map_entry->map_type);
4803 			check_finish_init++;
4804 			break;
4805 		}
4806 	}
4807 
4808 	if (check_finish_init && pptr->port_link_cnt == link_cnt) {
4809 		ASSERT(i > 0);
4810 		(void) fcp_call_finish_init_held(pptr, ptgt, link_cnt,
4811 		    map_tag[i-1], cause);
4812 	} else if (dev_cnt == 0 && pptr->port_link_cnt == link_cnt) {
4813 		fcp_offline_all(pptr, link_cnt, cause);
4814 	}
4815 }
4816 
4817 /*
4818  *     Function: fcp_handle_mapflags
4819  *
4820  *  Description: This function creates a target structure if the ptgt passed
4821  *		 is NULL.  It also kicks off the PLOGI if we are not logged
4822  *		 into the target yet or the PRLI if we are logged into the
4823  *		 target already.  The rest of the treatment is done in the
4824  *		 callbacks of the PLOGI or PRLI.
4825  *
4826  *     Argument: *pptr		FCP Port structure.
4827  *		 *ptgt		Target structure.
4828  *		 *map_entry	Array of fc_portmap_t structures.
4829  *		 link_cnt	Link state count.
4830  *		 tgt_cnt	Target state count.
4831  *		 cause		What caused this function to be called.
4832  *
4833  * Return Value: TRUE	Failed
4834  *		 FALSE	Succeeded
4835  *
4836  *	  Notes: pptr->port_mutex must be owned.
4837  */
4838 static int
4839 fcp_handle_mapflags(struct fcp_port	*pptr, struct fcp_tgt	*ptgt,
4840     fc_portmap_t *map_entry, int link_cnt, int tgt_cnt, int cause)
4841 {
4842 	int			lcount;
4843 	int			tcount;
4844 	int			ret = TRUE;
4845 	int			alloc;
4846 	struct fcp_ipkt	*icmd;
4847 	struct fcp_lun	*pseq_lun = NULL;
4848 	uchar_t			opcode;
4849 	int			valid_ptgt_was_passed = FALSE;
4850 
4851 	ASSERT(mutex_owned(&pptr->port_mutex));
4852 
4853 	/*
4854 	 * This case is possible where the FCTL has come up and done discovery
4855 	 * before FCP was loaded and attached. FCTL would have discovered the
4856 	 * devices and later the ULP came online. In this case ULP's would get
4857 	 * PORT_DEVICE_NOCHANGE but target would be NULL.
4858 	 */
4859 	if (ptgt == NULL) {
4860 		/* don't already have a target */
4861 		mutex_exit(&pptr->port_mutex);
4862 		ptgt = fcp_alloc_tgt(pptr, map_entry, link_cnt);
4863 		mutex_enter(&pptr->port_mutex);
4864 
4865 		if (ptgt == NULL) {
4866 			fcp_log(CE_WARN, pptr->port_dip,
4867 			    "!FC target allocation failed");
4868 			return (ret);
4869 		}
4870 		mutex_enter(&ptgt->tgt_mutex);
4871 		ptgt->tgt_statec_cause = cause;
4872 		ptgt->tgt_tmp_cnt = 1;
4873 		mutex_exit(&ptgt->tgt_mutex);
4874 	} else {
4875 		valid_ptgt_was_passed = TRUE;
4876 	}
4877 
4878 	/*
4879 	 * Copy in the target parameters
4880 	 */
4881 	mutex_enter(&ptgt->tgt_mutex);
4882 	ptgt->tgt_d_id = map_entry->map_did.port_id;
4883 	ptgt->tgt_hard_addr = map_entry->map_hard_addr.hard_addr;
4884 	ptgt->tgt_pd_handle = map_entry->map_pd;
4885 	ptgt->tgt_fca_dev = NULL;
4886 
4887 	/* Copy port and node WWNs */
4888 	bcopy(&map_entry->map_nwwn, &ptgt->tgt_node_wwn.raw_wwn[0],
4889 	    FC_WWN_SIZE);
4890 	bcopy(&map_entry->map_pwwn, &ptgt->tgt_port_wwn.raw_wwn[0],
4891 	    FC_WWN_SIZE);
4892 
4893 	if (!(map_entry->map_flags & PORT_DEVICE_NO_SKIP_DEVICE_DISCOVERY) &&
4894 	    (map_entry->map_type == PORT_DEVICE_NOCHANGE) &&
4895 	    (map_entry->map_state == PORT_DEVICE_LOGGED_IN) &&
4896 	    valid_ptgt_was_passed) {
4897 		/*
4898 		 * determine if there are any tape LUNs on this target
4899 		 */
4900 		for (pseq_lun = ptgt->tgt_lun;
4901 		    pseq_lun != NULL;
4902 		    pseq_lun = pseq_lun->lun_next) {
4903 			if ((pseq_lun->lun_type == DTYPE_SEQUENTIAL) &&
4904 			    !(pseq_lun->lun_state & FCP_LUN_OFFLINE)) {
4905 				fcp_update_tgt_state(ptgt, FCP_RESET,
4906 				    FCP_LUN_MARK);
4907 				mutex_exit(&ptgt->tgt_mutex);
4908 				return (ret);
4909 			}
4910 		}
4911 	}
4912 
4913 	/*
4914 	 * If ptgt was NULL when this function was entered, then tgt_node_state
4915 	 * was never specifically initialized but zeroed out which means
4916 	 * FCP_TGT_NODE_NONE.
4917 	 */
4918 	switch (ptgt->tgt_node_state) {
4919 	case FCP_TGT_NODE_NONE:
4920 	case FCP_TGT_NODE_ON_DEMAND:
4921 		if (FC_TOP_EXTERNAL(pptr->port_topology) &&
4922 		    !fcp_enable_auto_configuration &&
4923 		    map_entry->map_type != PORT_DEVICE_USER_CREATE) {
4924 			ptgt->tgt_node_state = FCP_TGT_NODE_ON_DEMAND;
4925 		} else if (FC_TOP_EXTERNAL(pptr->port_topology) &&
4926 		    fcp_enable_auto_configuration &&
4927 		    (ptgt->tgt_manual_config_only == 1) &&
4928 		    map_entry->map_type != PORT_DEVICE_USER_CREATE) {
4929 			/*
4930 			 * If auto configuration is set and
4931 			 * the tgt_manual_config_only flag is set then
4932 			 * we only want the user to be able to change
4933 			 * the state through create_on_demand.
4934 			 */
4935 			ptgt->tgt_node_state = FCP_TGT_NODE_ON_DEMAND;
4936 		} else {
4937 			ptgt->tgt_node_state = FCP_TGT_NODE_NONE;
4938 		}
4939 		break;
4940 
4941 	case FCP_TGT_NODE_PRESENT:
4942 		break;
4943 	}
4944 	/*
4945 	 * If we are booting from a fabric device, make sure we
4946 	 * mark the node state appropriately for this target to be
4947 	 * enumerated
4948 	 */
4949 	if (FC_TOP_EXTERNAL(pptr->port_topology) && pptr->port_boot_wwn[0]) {
4950 		if (bcmp((caddr_t)pptr->port_boot_wwn,
4951 		    (caddr_t)&ptgt->tgt_port_wwn.raw_wwn[0],
4952 		    sizeof (ptgt->tgt_port_wwn)) == 0) {
4953 			ptgt->tgt_node_state = FCP_TGT_NODE_NONE;
4954 		}
4955 	}
4956 	mutex_exit(&ptgt->tgt_mutex);
4957 
4958 	FCP_TRACE(fcp_logq, pptr->port_instbuf,
4959 	    fcp_trace, FCP_BUF_LEVEL_3, 0,
4960 	    "map_pd=%p, map_type=%x, did = %x, ulp_rscn_count=0x%x",
4961 	    map_entry->map_pd, map_entry->map_type, map_entry->map_did.port_id,
4962 	    map_entry->map_rscn_info.ulp_rscn_count);
4963 
4964 	mutex_enter(&ptgt->tgt_mutex);
4965 
4966 	/*
4967 	 * Reset target OFFLINE state and mark the target BUSY
4968 	 */
4969 	ptgt->tgt_state &= ~FCP_TGT_OFFLINE;
4970 	ptgt->tgt_state |= (FCP_TGT_BUSY | FCP_TGT_MARK);
4971 
4972 	tcount = tgt_cnt ? tgt_cnt : ptgt->tgt_change_cnt;
4973 	lcount = link_cnt;
4974 
4975 	mutex_exit(&ptgt->tgt_mutex);
4976 	mutex_exit(&pptr->port_mutex);
4977 
4978 	/*
4979 	 * if we are already logged in, then we do a PRLI, else
4980 	 * we do a PLOGI first (to get logged in)
4981 	 *
4982 	 * We will not check if we are the PLOGI initiator
4983 	 */
4984 	opcode = (map_entry->map_state == PORT_DEVICE_LOGGED_IN &&
4985 	    map_entry->map_pd != NULL) ? LA_ELS_PRLI : LA_ELS_PLOGI;
4986 
4987 	alloc = FCP_MAX(sizeof (la_els_logi_t), sizeof (la_els_prli_t));
4988 
4989 	icmd = fcp_icmd_alloc(pptr, ptgt, alloc, alloc, 0, 0, lcount, tcount,
4990 	    cause, map_entry->map_rscn_info.ulp_rscn_count);
4991 
4992 	if (icmd == NULL) {
4993 		FCP_TGT_TRACE(ptgt, tgt_cnt, FCP_TGT_TRACE_29);
4994 		/*
4995 		 * We've exited port_mutex before calling fcp_icmd_alloc,
4996 		 * we need to make sure we reacquire it before returning.
4997 		 */
4998 		mutex_enter(&pptr->port_mutex);
4999 		return (FALSE);
5000 	}
5001 
5002 	/* TRUE is only returned while target is intended skipped */
5003 	ret = FALSE;
5004 	/* discover info about this target */
5005 	if ((fcp_send_els(pptr, ptgt, icmd, opcode,
5006 	    lcount, tcount, cause)) == DDI_SUCCESS) {
5007 		FCP_TGT_TRACE(ptgt, tgt_cnt, FCP_TGT_TRACE_9);
5008 	} else {
5009 		fcp_icmd_free(pptr, icmd);
5010 		ret = TRUE;
5011 	}
5012 	mutex_enter(&pptr->port_mutex);
5013 
5014 	return (ret);
5015 }
5016 
5017 /*
5018  *     Function: fcp_send_els
5019  *
5020  *  Description: Sends an ELS to the target specified by the caller.  Supports
5021  *		 PLOGI and PRLI.
5022  *
5023  *     Argument: *pptr		Fcp port.
5024  *		 *ptgt		Target to send the ELS to.
5025  *		 *icmd		Internal packet
5026  *		 opcode		ELS opcode
5027  *		 lcount		Link state change counter
5028  *		 tcount		Target state change counter
5029  *		 cause		What caused the call
5030  *
5031  * Return Value: DDI_SUCCESS
5032  *		 Others
5033  */
5034 static int
5035 fcp_send_els(struct fcp_port *pptr, struct fcp_tgt *ptgt,
5036     struct fcp_ipkt *icmd, uchar_t opcode, int lcount, int tcount, int cause)
5037 {
5038 	fc_packet_t		*fpkt;
5039 	fc_frame_hdr_t		*hp;
5040 	int			internal = 0;
5041 	int			alloc;
5042 	int			cmd_len;
5043 	int			resp_len;
5044 	int			res = DDI_FAILURE; /* default result */
5045 	int			rval = DDI_FAILURE;
5046 
5047 	ASSERT(opcode == LA_ELS_PLOGI || opcode == LA_ELS_PRLI);
5048 	ASSERT(ptgt->tgt_port == pptr);
5049 
5050 	FCP_TRACE(fcp_logq, pptr->port_instbuf,
5051 	    fcp_trace, FCP_BUF_LEVEL_5, 0,
5052 	    "fcp_send_els: d_id=0x%x ELS 0x%x (%s)", ptgt->tgt_d_id, opcode,
5053 	    (opcode == LA_ELS_PLOGI) ? "PLOGI" : "PRLI");
5054 
5055 	if (opcode == LA_ELS_PLOGI) {
5056 		cmd_len = sizeof (la_els_logi_t);
5057 		resp_len = sizeof (la_els_logi_t);
5058 	} else {
5059 		ASSERT(opcode == LA_ELS_PRLI);
5060 		cmd_len = sizeof (la_els_prli_t);
5061 		resp_len = sizeof (la_els_prli_t);
5062 	}
5063 
5064 	if (icmd == NULL) {
5065 		alloc = FCP_MAX(sizeof (la_els_logi_t),
5066 		    sizeof (la_els_prli_t));
5067 		icmd = fcp_icmd_alloc(pptr, ptgt, alloc, alloc, 0, 0,
5068 		    lcount, tcount, cause, FC_INVALID_RSCN_COUNT);
5069 		if (icmd == NULL) {
5070 			FCP_TGT_TRACE(ptgt, tcount, FCP_TGT_TRACE_10);
5071 			return (res);
5072 		}
5073 		internal++;
5074 	}
5075 	fpkt = icmd->ipkt_fpkt;
5076 
5077 	fpkt->pkt_cmdlen = cmd_len;
5078 	fpkt->pkt_rsplen = resp_len;
5079 	fpkt->pkt_datalen = 0;
5080 	icmd->ipkt_retries = 0;
5081 
5082 	/* fill in fpkt info */
5083 	fpkt->pkt_tran_flags = FC_TRAN_CLASS3 | FC_TRAN_INTR;
5084 	fpkt->pkt_tran_type = FC_PKT_EXCHANGE;
5085 	fpkt->pkt_timeout = FCP_ELS_TIMEOUT;
5086 
5087 	/* get ptr to frame hdr in fpkt */
5088 	hp = &fpkt->pkt_cmd_fhdr;
5089 
5090 	/*
5091 	 * fill in frame hdr
5092 	 */
5093 	hp->r_ctl = R_CTL_ELS_REQ;
5094 	hp->s_id = pptr->port_id;	/* source ID */
5095 	hp->d_id = ptgt->tgt_d_id;	/* dest ID */
5096 	hp->type = FC_TYPE_EXTENDED_LS;
5097 	hp->f_ctl = F_CTL_SEQ_INITIATIVE | F_CTL_FIRST_SEQ;
5098 	hp->seq_id = 0;
5099 	hp->rsvd = 0;
5100 	hp->df_ctl  = 0;
5101 	hp->seq_cnt = 0;
5102 	hp->ox_id = 0xffff;		/* i.e. none */
5103 	hp->rx_id = 0xffff;		/* i.e. none */
5104 	hp->ro = 0;
5105 
5106 	/*
5107 	 * at this point we have a filled in cmd pkt
5108 	 *
5109 	 * fill in the respective info, then use the transport to send
5110 	 * the packet
5111 	 *
5112 	 * for a PLOGI call fc_ulp_login(), and
5113 	 * for a PRLI call fc_ulp_issue_els()
5114 	 */
5115 	switch (opcode) {
5116 	case LA_ELS_PLOGI: {
5117 		struct la_els_logi logi;
5118 
5119 		bzero(&logi, sizeof (struct la_els_logi));
5120 
5121 		hp = &fpkt->pkt_cmd_fhdr;
5122 		hp->r_ctl = R_CTL_ELS_REQ;
5123 		logi.ls_code.ls_code = LA_ELS_PLOGI;
5124 		logi.ls_code.mbz = 0;
5125 
5126 		FCP_CP_OUT((uint8_t *)&logi, fpkt->pkt_cmd,
5127 		    fpkt->pkt_cmd_acc, sizeof (struct la_els_logi));
5128 
5129 		icmd->ipkt_opcode = LA_ELS_PLOGI;
5130 
5131 		mutex_enter(&pptr->port_mutex);
5132 		if (!FCP_TGT_STATE_CHANGED(ptgt, icmd)) {
5133 
5134 			mutex_exit(&pptr->port_mutex);
5135 
5136 			rval = fc_ulp_login(pptr->port_fp_handle, &fpkt, 1);
5137 			if (rval == FC_SUCCESS) {
5138 				res = DDI_SUCCESS;
5139 				break;
5140 			}
5141 
5142 			FCP_TGT_TRACE(ptgt, tcount, FCP_TGT_TRACE_11);
5143 
5144 			res = fcp_handle_ipkt_errors(pptr, ptgt, icmd,
5145 			    rval, "PLOGI");
5146 		} else {
5147 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
5148 			    fcp_trace, FCP_BUF_LEVEL_5, 0,
5149 			    "fcp_send_els1: state change occured"
5150 			    " for D_ID=0x%x", ptgt->tgt_d_id);
5151 			mutex_exit(&pptr->port_mutex);
5152 			FCP_TGT_TRACE(ptgt, tcount, FCP_TGT_TRACE_12);
5153 		}
5154 		break;
5155 	}
5156 
5157 	case LA_ELS_PRLI: {
5158 		struct la_els_prli	prli;
5159 		struct fcp_prli		*fprli;
5160 
5161 		bzero(&prli, sizeof (struct la_els_prli));
5162 
5163 		hp = &fpkt->pkt_cmd_fhdr;
5164 		hp->r_ctl = R_CTL_ELS_REQ;
5165 
5166 		/* fill in PRLI cmd ELS fields */
5167 		prli.ls_code = LA_ELS_PRLI;
5168 		prli.page_length = 0x10;	/* huh? */
5169 		prli.payload_length = sizeof (struct la_els_prli);
5170 
5171 		icmd->ipkt_opcode = LA_ELS_PRLI;
5172 
5173 		/* get ptr to PRLI service params */
5174 		fprli = (struct fcp_prli *)prli.service_params;
5175 
5176 		/* fill in service params */
5177 		fprli->type = 0x08;
5178 		fprli->resvd1 = 0;
5179 		fprli->orig_process_assoc_valid = 0;
5180 		fprli->resp_process_assoc_valid = 0;
5181 		fprli->establish_image_pair = 1;
5182 		fprli->resvd2 = 0;
5183 		fprli->resvd3 = 0;
5184 		fprli->obsolete_1 = 0;
5185 		fprli->obsolete_2 = 0;
5186 		fprli->data_overlay_allowed = 0;
5187 		fprli->initiator_fn = 1;
5188 		fprli->confirmed_compl_allowed = 1;
5189 
5190 		if (fc_ulp_is_name_present("ltct") == FC_SUCCESS) {
5191 			fprli->target_fn = 1;
5192 		} else {
5193 			fprli->target_fn = 0;
5194 		}
5195 
5196 		fprli->retry = 1;
5197 		fprli->read_xfer_rdy_disabled = 1;
5198 		fprli->write_xfer_rdy_disabled = 0;
5199 
5200 		FCP_CP_OUT((uint8_t *)&prli, fpkt->pkt_cmd,
5201 		    fpkt->pkt_cmd_acc, sizeof (struct la_els_prli));
5202 
5203 		/* issue the PRLI request */
5204 
5205 		mutex_enter(&pptr->port_mutex);
5206 		if (!FCP_TGT_STATE_CHANGED(ptgt, icmd)) {
5207 
5208 			mutex_exit(&pptr->port_mutex);
5209 
5210 			rval = fc_ulp_issue_els(pptr->port_fp_handle, fpkt);
5211 			if (rval == FC_SUCCESS) {
5212 				res = DDI_SUCCESS;
5213 				break;
5214 			}
5215 
5216 			FCP_TGT_TRACE(ptgt, tcount, FCP_TGT_TRACE_13);
5217 
5218 			res = fcp_handle_ipkt_errors(pptr, ptgt, icmd,
5219 			    rval, "PRLI");
5220 		} else {
5221 			mutex_exit(&pptr->port_mutex);
5222 			FCP_TGT_TRACE(ptgt, tcount, FCP_TGT_TRACE_14);
5223 		}
5224 		break;
5225 	}
5226 
5227 	default:
5228 		fcp_log(CE_WARN, NULL, "!invalid ELS opcode=0x%x", opcode);
5229 		break;
5230 	}
5231 
5232 	FCP_TRACE(fcp_logq, pptr->port_instbuf,
5233 	    fcp_trace, FCP_BUF_LEVEL_5, 0,
5234 	    "fcp_send_els: returning %d", res);
5235 
5236 	if (res != DDI_SUCCESS) {
5237 		if (internal) {
5238 			fcp_icmd_free(pptr, icmd);
5239 		}
5240 	}
5241 
5242 	return (res);
5243 }
5244 
5245 
5246 /*
5247  * called internally update the state of all of the tgts and each LUN
5248  * for this port (i.e. each target  known to be attached to this port)
5249  * if they are not already offline
5250  *
5251  * must be called with the port mutex owned
5252  *
5253  * acquires and releases the target mutexes for each target attached
5254  * to this port
5255  */
5256 void
5257 fcp_update_state(struct fcp_port *pptr, uint32_t state, int cause)
5258 {
5259 	int i;
5260 	struct fcp_tgt *ptgt;
5261 
5262 	ASSERT(mutex_owned(&pptr->port_mutex));
5263 
5264 	for (i = 0; i < FCP_NUM_HASH; i++) {
5265 		for (ptgt = pptr->port_tgt_hash_table[i]; ptgt != NULL;
5266 		    ptgt = ptgt->tgt_next) {
5267 			mutex_enter(&ptgt->tgt_mutex);
5268 			fcp_update_tgt_state(ptgt, FCP_SET, state);
5269 			ptgt->tgt_change_cnt++;
5270 			ptgt->tgt_statec_cause = cause;
5271 			ptgt->tgt_tmp_cnt = 1;
5272 			ptgt->tgt_done = 0;
5273 			mutex_exit(&ptgt->tgt_mutex);
5274 		}
5275 	}
5276 }
5277 
5278 
5279 static void
5280 fcp_offline_all(struct fcp_port *pptr, int lcount, int cause)
5281 {
5282 	int i;
5283 	int ndevs;
5284 	struct fcp_tgt *ptgt;
5285 
5286 	ASSERT(mutex_owned(&pptr->port_mutex));
5287 
5288 	for (ndevs = 0, i = 0; i < FCP_NUM_HASH; i++) {
5289 		for (ptgt = pptr->port_tgt_hash_table[i]; ptgt != NULL;
5290 		    ptgt = ptgt->tgt_next) {
5291 			ndevs++;
5292 		}
5293 	}
5294 
5295 	if (ndevs == 0) {
5296 		return;
5297 	}
5298 	pptr->port_tmp_cnt = ndevs;
5299 
5300 	for (i = 0; i < FCP_NUM_HASH; i++) {
5301 		for (ptgt = pptr->port_tgt_hash_table[i]; ptgt != NULL;
5302 		    ptgt = ptgt->tgt_next) {
5303 			(void) fcp_call_finish_init_held(pptr, ptgt,
5304 			    lcount, ptgt->tgt_change_cnt, cause);
5305 		}
5306 	}
5307 }
5308 
5309 /*
5310  *     Function: fcp_update_tgt_state
5311  *
5312  *  Description: This function updates the field tgt_state of a target.	 That
5313  *		 field is a bitmap and which bit can be set or reset
5314  *		 individually.	The action applied to the target state is also
5315  *		 applied to all the LUNs belonging to the target (provided the
5316  *		 LUN is not offline).  A side effect of applying the state
5317  *		 modification to the target and the LUNs is the field tgt_trace
5318  *		 of the target and lun_trace of the LUNs is set to zero.
5319  *
5320  *
5321  *     Argument: *ptgt	Target structure.
5322  *		 flag	Flag indication what action to apply (set/reset).
5323  *		 state	State bits to update.
5324  *
5325  * Return Value: None
5326  *
5327  *	Context: Interrupt, Kernel or User context.
5328  *		 The mutex of the target (ptgt->tgt_mutex) must be owned when
5329  *		 calling this function.
5330  */
5331 void
5332 fcp_update_tgt_state(struct fcp_tgt *ptgt, int flag, uint32_t state)
5333 {
5334 	struct fcp_lun *plun;
5335 
5336 	ASSERT(mutex_owned(&ptgt->tgt_mutex));
5337 
5338 	if (!(ptgt->tgt_state & FCP_TGT_OFFLINE)) {
5339 		/* The target is not offline. */
5340 		if (flag == FCP_SET) {
5341 			ptgt->tgt_state |= state;
5342 			ptgt->tgt_trace = 0;
5343 		} else {
5344 			ptgt->tgt_state &= ~state;
5345 		}
5346 
5347 		for (plun = ptgt->tgt_lun; plun != NULL;
5348 		    plun = plun->lun_next) {
5349 			if (!(plun->lun_state & FCP_LUN_OFFLINE)) {
5350 				/* The LUN is not offline. */
5351 				if (flag == FCP_SET) {
5352 					plun->lun_state |= state;
5353 					plun->lun_trace = 0;
5354 				} else {
5355 					plun->lun_state &= ~state;
5356 				}
5357 			}
5358 		}
5359 	}
5360 }
5361 
5362 /*
5363  *     Function: fcp_update_tgt_state
5364  *
5365  *  Description: This function updates the field lun_state of a LUN.  That
5366  *		 field is a bitmap and which bit can be set or reset
5367  *		 individually.
5368  *
5369  *     Argument: *plun	LUN structure.
5370  *		 flag	Flag indication what action to apply (set/reset).
5371  *		 state	State bits to update.
5372  *
5373  * Return Value: None
5374  *
5375  *	Context: Interrupt, Kernel or User context.
5376  *		 The mutex of the target (ptgt->tgt_mutex) must be owned when
5377  *		 calling this function.
5378  */
5379 void
5380 fcp_update_lun_state(struct fcp_lun *plun, int flag, uint32_t state)
5381 {
5382 	struct fcp_tgt	*ptgt = plun->lun_tgt;
5383 
5384 	ASSERT(mutex_owned(&ptgt->tgt_mutex));
5385 
5386 	if (!(plun->lun_state & FCP_TGT_OFFLINE)) {
5387 		if (flag == FCP_SET) {
5388 			plun->lun_state |= state;
5389 		} else {
5390 			plun->lun_state &= ~state;
5391 		}
5392 	}
5393 }
5394 
5395 /*
5396  *     Function: fcp_get_port
5397  *
5398  *  Description: This function returns the fcp_port structure from the opaque
5399  *		 handle passed by the caller.  That opaque handle is the handle
5400  *		 used by fp/fctl to identify a particular local port.  That
5401  *		 handle has been stored in the corresponding fcp_port
5402  *		 structure.  This function is going to walk the global list of
5403  *		 fcp_port structures till one has a port_fp_handle that matches
5404  *		 the handle passed by the caller.  This function enters the
5405  *		 mutex fcp_global_mutex while walking the global list and then
5406  *		 releases it.
5407  *
5408  *     Argument: port_handle	Opaque handle that fp/fctl uses to identify a
5409  *				particular port.
5410  *
5411  * Return Value: NULL		Not found.
5412  *		 Not NULL	Pointer to the fcp_port structure.
5413  *
5414  *	Context: Interrupt, Kernel or User context.
5415  */
5416 static struct fcp_port *
5417 fcp_get_port(opaque_t port_handle)
5418 {
5419 	struct fcp_port *pptr;
5420 
5421 	ASSERT(port_handle != NULL);
5422 
5423 	mutex_enter(&fcp_global_mutex);
5424 	for (pptr = fcp_port_head; pptr != NULL; pptr = pptr->port_next) {
5425 		if (pptr->port_fp_handle == port_handle) {
5426 			break;
5427 		}
5428 	}
5429 	mutex_exit(&fcp_global_mutex);
5430 
5431 	return (pptr);
5432 }
5433 
5434 
5435 static void
5436 fcp_unsol_callback(fc_packet_t *fpkt)
5437 {
5438 	struct fcp_ipkt *icmd = (struct fcp_ipkt *)fpkt->pkt_ulp_private;
5439 	struct fcp_port *pptr = icmd->ipkt_port;
5440 
5441 	if (fpkt->pkt_state != FC_PKT_SUCCESS) {
5442 		caddr_t state, reason, action, expln;
5443 
5444 		(void) fc_ulp_pkt_error(fpkt, &state, &reason,
5445 		    &action, &expln);
5446 
5447 		fcp_log(CE_WARN, pptr->port_dip,
5448 		    "!couldn't post response to unsolicited request: "
5449 		    " state=%s reason=%s rx_id=%x ox_id=%x",
5450 		    state, reason, fpkt->pkt_cmd_fhdr.ox_id,
5451 		    fpkt->pkt_cmd_fhdr.rx_id);
5452 	}
5453 	fcp_icmd_free(pptr, icmd);
5454 }
5455 
5456 
5457 /*
5458  * Perform general purpose preparation of a response to an unsolicited request
5459  */
5460 static void
5461 fcp_unsol_resp_init(fc_packet_t *pkt, fc_unsol_buf_t *buf,
5462     uchar_t r_ctl, uchar_t type)
5463 {
5464 	pkt->pkt_cmd_fhdr.r_ctl = r_ctl;
5465 	pkt->pkt_cmd_fhdr.d_id = buf->ub_frame.s_id;
5466 	pkt->pkt_cmd_fhdr.s_id = buf->ub_frame.d_id;
5467 	pkt->pkt_cmd_fhdr.type = type;
5468 	pkt->pkt_cmd_fhdr.f_ctl = F_CTL_LAST_SEQ | F_CTL_XCHG_CONTEXT;
5469 	pkt->pkt_cmd_fhdr.seq_id = buf->ub_frame.seq_id;
5470 	pkt->pkt_cmd_fhdr.df_ctl  = buf->ub_frame.df_ctl;
5471 	pkt->pkt_cmd_fhdr.seq_cnt = buf->ub_frame.seq_cnt;
5472 	pkt->pkt_cmd_fhdr.ox_id = buf->ub_frame.ox_id;
5473 	pkt->pkt_cmd_fhdr.rx_id = buf->ub_frame.rx_id;
5474 	pkt->pkt_cmd_fhdr.ro = 0;
5475 	pkt->pkt_cmd_fhdr.rsvd = 0;
5476 	pkt->pkt_comp = fcp_unsol_callback;
5477 	pkt->pkt_pd = NULL;
5478 }
5479 
5480 
5481 /*ARGSUSED*/
5482 static int
5483 fcp_unsol_prli(struct fcp_port *pptr, fc_unsol_buf_t *buf)
5484 {
5485 	fc_packet_t		*fpkt;
5486 	struct la_els_prli	prli;
5487 	struct fcp_prli		*fprli;
5488 	struct fcp_ipkt	*icmd;
5489 	struct la_els_prli	*from;
5490 	struct fcp_prli		*orig;
5491 	struct fcp_tgt	*ptgt;
5492 	int			tcount = 0;
5493 	int			lcount;
5494 
5495 	from = (struct la_els_prli *)buf->ub_buffer;
5496 	orig = (struct fcp_prli *)from->service_params;
5497 
5498 	if ((ptgt = fcp_get_target_by_did(pptr, buf->ub_frame.s_id)) !=
5499 	    NULL) {
5500 		mutex_enter(&ptgt->tgt_mutex);
5501 		tcount = ptgt->tgt_change_cnt;
5502 		mutex_exit(&ptgt->tgt_mutex);
5503 	}
5504 	mutex_enter(&pptr->port_mutex);
5505 	lcount = pptr->port_link_cnt;
5506 	mutex_exit(&pptr->port_mutex);
5507 
5508 	if ((icmd = fcp_icmd_alloc(pptr, ptgt, sizeof (la_els_prli_t),
5509 	    sizeof (la_els_prli_t), 0, 0, lcount, tcount, 0,
5510 	    FC_INVALID_RSCN_COUNT)) == NULL) {
5511 		return (FC_FAILURE);
5512 	}
5513 	fpkt = icmd->ipkt_fpkt;
5514 	fpkt->pkt_tran_flags = FC_TRAN_CLASS3 | FC_TRAN_INTR;
5515 	fpkt->pkt_tran_type = FC_PKT_OUTBOUND;
5516 	fpkt->pkt_timeout = FCP_ELS_TIMEOUT;
5517 	fpkt->pkt_cmdlen = sizeof (la_els_prli_t);
5518 	fpkt->pkt_rsplen = 0;
5519 	fpkt->pkt_datalen = 0;
5520 
5521 	icmd->ipkt_opcode = LA_ELS_PRLI;
5522 
5523 	bzero(&prli, sizeof (struct la_els_prli));
5524 	fprli = (struct fcp_prli *)prli.service_params;
5525 	prli.ls_code = LA_ELS_ACC;
5526 	prli.page_length = 0x10;
5527 	prli.payload_length = sizeof (struct la_els_prli);
5528 
5529 	/* fill in service params */
5530 	fprli->type = 0x08;
5531 	fprli->resvd1 = 0;
5532 	fprli->orig_process_assoc_valid = orig->orig_process_assoc_valid;
5533 	fprli->orig_process_associator = orig->orig_process_associator;
5534 	fprli->resp_process_assoc_valid = 0;
5535 	fprli->establish_image_pair = 1;
5536 	fprli->resvd2 = 0;
5537 	fprli->resvd3 = 0;
5538 	fprli->obsolete_1 = 0;
5539 	fprli->obsolete_2 = 0;
5540 	fprli->data_overlay_allowed = 0;
5541 	fprli->initiator_fn = 1;
5542 	fprli->confirmed_compl_allowed = 1;
5543 
5544 	if (fc_ulp_is_name_present("ltct") == FC_SUCCESS) {
5545 		fprli->target_fn = 1;
5546 	} else {
5547 		fprli->target_fn = 0;
5548 	}
5549 
5550 	fprli->retry = 1;
5551 	fprli->read_xfer_rdy_disabled = 1;
5552 	fprli->write_xfer_rdy_disabled = 0;
5553 
5554 	/* save the unsol prli payload first */
5555 	FCP_CP_OUT((uint8_t *)from, fpkt->pkt_resp,
5556 	    fpkt->pkt_resp_acc, sizeof (struct la_els_prli));
5557 
5558 	FCP_CP_OUT((uint8_t *)&prli, fpkt->pkt_cmd,
5559 	    fpkt->pkt_cmd_acc, sizeof (struct la_els_prli));
5560 
5561 	fcp_unsol_resp_init(fpkt, buf, R_CTL_ELS_RSP, FC_TYPE_EXTENDED_LS);
5562 
5563 	mutex_enter(&pptr->port_mutex);
5564 	if (!FCP_LINK_STATE_CHANGED(pptr, icmd)) {
5565 		int rval;
5566 		mutex_exit(&pptr->port_mutex);
5567 
5568 		if ((rval = fc_ulp_issue_els(pptr->port_fp_handle, fpkt)) !=
5569 		    FC_SUCCESS) {
5570 			if (rval == FC_STATEC_BUSY || rval == FC_OFFLINE) {
5571 				fcp_queue_ipkt(pptr, fpkt);
5572 				return (FC_SUCCESS);
5573 			}
5574 			/* Let it timeout */
5575 			fcp_icmd_free(pptr, icmd);
5576 			return (FC_FAILURE);
5577 		}
5578 	} else {
5579 		mutex_exit(&pptr->port_mutex);
5580 		fcp_icmd_free(pptr, icmd);
5581 		return (FC_FAILURE);
5582 	}
5583 
5584 	(void) fc_ulp_ubrelease(pptr->port_fp_handle, 1, &buf->ub_token);
5585 
5586 	return (FC_SUCCESS);
5587 }
5588 
5589 /*
5590  *     Function: fcp_icmd_alloc
5591  *
5592  *  Description: This function allocated a fcp_ipkt structure.	The pkt_comp
5593  *		 field is initialized to fcp_icmd_callback.  Sometimes it is
5594  *		 modified by the caller (such as fcp_send_scsi).  The
5595  *		 structure is also tied to the state of the line and of the
5596  *		 target at a particular time.  That link is established by
5597  *		 setting the fields ipkt_link_cnt and ipkt_change_cnt to lcount
5598  *		 and tcount which came respectively from pptr->link_cnt and
5599  *		 ptgt->tgt_change_cnt.
5600  *
5601  *     Argument: *pptr		Fcp port.
5602  *		 *ptgt		Target (destination of the command).
5603  *		 cmd_len	Length of the command.
5604  *		 resp_len	Length of the expected response.
5605  *		 data_len	Length of the data.
5606  *		 nodma		Indicates weither the command and response.
5607  *				will be transfer through DMA or not.
5608  *		 lcount		Link state change counter.
5609  *		 tcount		Target state change counter.
5610  *		 cause		Reason that lead to this call.
5611  *
5612  * Return Value: NULL		Failed.
5613  *		 Not NULL	Internal packet address.
5614  */
5615 static struct fcp_ipkt *
5616 fcp_icmd_alloc(struct fcp_port *pptr, struct fcp_tgt *ptgt, int cmd_len,
5617     int resp_len, int data_len, int nodma, int lcount, int tcount, int cause,
5618     uint32_t rscn_count)
5619 {
5620 	int			dma_setup = 0;
5621 	fc_packet_t		*fpkt;
5622 	struct fcp_ipkt	*icmd = NULL;
5623 
5624 	icmd = kmem_zalloc(sizeof (struct fcp_ipkt) +
5625 	    pptr->port_dmacookie_sz + pptr->port_priv_pkt_len,
5626 	    KM_NOSLEEP);
5627 	if (icmd == NULL) {
5628 		fcp_log(CE_WARN, pptr->port_dip,
5629 		    "!internal packet allocation failed");
5630 		return (NULL);
5631 	}
5632 
5633 	/*
5634 	 * initialize the allocated packet
5635 	 */
5636 	icmd->ipkt_nodma = nodma;
5637 	icmd->ipkt_next = icmd->ipkt_prev = NULL;
5638 	icmd->ipkt_lun = NULL;
5639 
5640 	icmd->ipkt_link_cnt = lcount;
5641 	icmd->ipkt_change_cnt = tcount;
5642 	icmd->ipkt_cause = cause;
5643 
5644 	mutex_enter(&pptr->port_mutex);
5645 	icmd->ipkt_port = pptr;
5646 	mutex_exit(&pptr->port_mutex);
5647 
5648 	/* keep track of amt of data to be sent in pkt */
5649 	icmd->ipkt_cmdlen = cmd_len;
5650 	icmd->ipkt_resplen = resp_len;
5651 	icmd->ipkt_datalen = data_len;
5652 
5653 	/* set up pkt's ptr to the fc_packet_t struct, just after the ipkt */
5654 	icmd->ipkt_fpkt = (fc_packet_t *)(&icmd->ipkt_fc_packet);
5655 
5656 	/* set pkt's private ptr to point to cmd pkt */
5657 	icmd->ipkt_fpkt->pkt_ulp_private = (opaque_t)icmd;
5658 
5659 	/* set FCA private ptr to memory just beyond */
5660 	icmd->ipkt_fpkt->pkt_fca_private = (opaque_t)
5661 	    ((char *)icmd + sizeof (struct fcp_ipkt) +
5662 	    pptr->port_dmacookie_sz);
5663 
5664 	/* get ptr to fpkt substruct and fill it in */
5665 	fpkt = icmd->ipkt_fpkt;
5666 	fpkt->pkt_data_cookie = (ddi_dma_cookie_t *)((caddr_t)icmd +
5667 	    sizeof (struct fcp_ipkt));
5668 
5669 	if (ptgt != NULL) {
5670 		icmd->ipkt_tgt = ptgt;
5671 		fpkt->pkt_fca_device = ptgt->tgt_fca_dev;
5672 	}
5673 
5674 	fpkt->pkt_comp = fcp_icmd_callback;
5675 	fpkt->pkt_tran_flags = (FC_TRAN_CLASS3 | FC_TRAN_INTR);
5676 	fpkt->pkt_cmdlen = cmd_len;
5677 	fpkt->pkt_rsplen = resp_len;
5678 	fpkt->pkt_datalen = data_len;
5679 
5680 	/*
5681 	 * The pkt_ulp_rscn_infop (aka pkt_ulp_rsvd1) field is used to pass the
5682 	 * rscn_count as fcp knows down to the transport. If a valid count was
5683 	 * passed into this function, we allocate memory to actually pass down
5684 	 * this info.
5685 	 *
5686 	 * BTW, if the kmem_zalloc fails, we won't try too hard. This will
5687 	 * basically mean that fcp will not be able to help transport
5688 	 * distinguish if a new RSCN has come after fcp was last informed about
5689 	 * it. In such cases, it might lead to the problem mentioned in CR/bug #
5690 	 * 5068068 where the device might end up going offline in case of RSCN
5691 	 * storms.
5692 	 */
5693 	fpkt->pkt_ulp_rscn_infop = NULL;
5694 	if (rscn_count != FC_INVALID_RSCN_COUNT) {
5695 		fpkt->pkt_ulp_rscn_infop = kmem_zalloc(
5696 		    sizeof (fc_ulp_rscn_info_t), KM_NOSLEEP);
5697 		if (fpkt->pkt_ulp_rscn_infop == NULL) {
5698 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
5699 			    fcp_trace, FCP_BUF_LEVEL_6, 0,
5700 			    "Failed to alloc memory to pass rscn info");
5701 		}
5702 	}
5703 
5704 	if (fpkt->pkt_ulp_rscn_infop != NULL) {
5705 		fc_ulp_rscn_info_t	*rscnp;
5706 
5707 		rscnp = (fc_ulp_rscn_info_t *)fpkt->pkt_ulp_rscn_infop;
5708 		rscnp->ulp_rscn_count = rscn_count;
5709 	}
5710 
5711 	if (fcp_alloc_dma(pptr, icmd, nodma, KM_NOSLEEP) != FC_SUCCESS) {
5712 		goto fail;
5713 	}
5714 	dma_setup++;
5715 
5716 	/*
5717 	 * Must hold target mutex across setting of pkt_pd and call to
5718 	 * fc_ulp_init_packet to ensure the handle to the target doesn't go
5719 	 * away while we're not looking.
5720 	 */
5721 	if (ptgt != NULL) {
5722 		mutex_enter(&ptgt->tgt_mutex);
5723 		fpkt->pkt_pd = ptgt->tgt_pd_handle;
5724 
5725 		/* ask transport to do its initialization on this pkt */
5726 		if (fc_ulp_init_packet(pptr->port_fp_handle, fpkt, KM_NOSLEEP)
5727 		    != FC_SUCCESS) {
5728 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
5729 			    fcp_trace, FCP_BUF_LEVEL_6, 0,
5730 			    "fc_ulp_init_packet failed");
5731 			mutex_exit(&ptgt->tgt_mutex);
5732 			goto fail;
5733 		}
5734 		mutex_exit(&ptgt->tgt_mutex);
5735 	} else {
5736 		if (fc_ulp_init_packet(pptr->port_fp_handle, fpkt, KM_NOSLEEP)
5737 		    != FC_SUCCESS) {
5738 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
5739 			    fcp_trace, FCP_BUF_LEVEL_6, 0,
5740 			    "fc_ulp_init_packet failed");
5741 			goto fail;
5742 		}
5743 	}
5744 
5745 	mutex_enter(&pptr->port_mutex);
5746 	if (pptr->port_state & (FCP_STATE_DETACHING |
5747 	    FCP_STATE_SUSPENDED | FCP_STATE_POWER_DOWN)) {
5748 		int rval;
5749 
5750 		mutex_exit(&pptr->port_mutex);
5751 
5752 		rval = fc_ulp_uninit_packet(pptr->port_fp_handle, fpkt);
5753 		ASSERT(rval == FC_SUCCESS);
5754 
5755 		goto fail;
5756 	}
5757 
5758 	if (ptgt != NULL) {
5759 		mutex_enter(&ptgt->tgt_mutex);
5760 		ptgt->tgt_ipkt_cnt++;
5761 		mutex_exit(&ptgt->tgt_mutex);
5762 	}
5763 
5764 	pptr->port_ipkt_cnt++;
5765 
5766 	mutex_exit(&pptr->port_mutex);
5767 
5768 	return (icmd);
5769 
5770 fail:
5771 	if (fpkt->pkt_ulp_rscn_infop != NULL) {
5772 		kmem_free(fpkt->pkt_ulp_rscn_infop,
5773 		    sizeof (fc_ulp_rscn_info_t));
5774 		fpkt->pkt_ulp_rscn_infop = NULL;
5775 	}
5776 
5777 	if (dma_setup) {
5778 		fcp_free_dma(pptr, icmd);
5779 	}
5780 	kmem_free(icmd, sizeof (struct fcp_ipkt) + pptr->port_priv_pkt_len +
5781 	    (size_t)pptr->port_dmacookie_sz);
5782 
5783 	return (NULL);
5784 }
5785 
5786 /*
5787  *     Function: fcp_icmd_free
5788  *
5789  *  Description: Frees the internal command passed by the caller.
5790  *
5791  *     Argument: *pptr		Fcp port.
5792  *		 *icmd		Internal packet to free.
5793  *
5794  * Return Value: None
5795  */
5796 static void
5797 fcp_icmd_free(struct fcp_port *pptr, struct fcp_ipkt *icmd)
5798 {
5799 	struct fcp_tgt	*ptgt = icmd->ipkt_tgt;
5800 
5801 	/* Let the underlying layers do their cleanup. */
5802 	(void) fc_ulp_uninit_packet(pptr->port_fp_handle,
5803 	    icmd->ipkt_fpkt);
5804 
5805 	if (icmd->ipkt_fpkt->pkt_ulp_rscn_infop) {
5806 		kmem_free(icmd->ipkt_fpkt->pkt_ulp_rscn_infop,
5807 		    sizeof (fc_ulp_rscn_info_t));
5808 	}
5809 
5810 	fcp_free_dma(pptr, icmd);
5811 
5812 	kmem_free(icmd, sizeof (struct fcp_ipkt) + pptr->port_priv_pkt_len +
5813 	    (size_t)pptr->port_dmacookie_sz);
5814 
5815 	mutex_enter(&pptr->port_mutex);
5816 
5817 	if (ptgt) {
5818 		mutex_enter(&ptgt->tgt_mutex);
5819 		ptgt->tgt_ipkt_cnt--;
5820 		mutex_exit(&ptgt->tgt_mutex);
5821 	}
5822 
5823 	pptr->port_ipkt_cnt--;
5824 	mutex_exit(&pptr->port_mutex);
5825 }
5826 
5827 /*
5828  *     Function: fcp_alloc_dma
5829  *
5830  *  Description: Allocated the DMA resources required for the internal
5831  *		 packet.
5832  *
5833  *     Argument: *pptr	FCP port.
5834  *		 *icmd	Internal FCP packet.
5835  *		 nodma	Indicates if the Cmd and Resp will be DMAed.
5836  *		 flags	Allocation flags (Sleep or NoSleep).
5837  *
5838  * Return Value: FC_SUCCESS
5839  *		 FC_NOMEM
5840  */
5841 static int
5842 fcp_alloc_dma(struct fcp_port *pptr, struct fcp_ipkt *icmd,
5843     int nodma, int flags)
5844 {
5845 	int		rval;
5846 	size_t		real_size;
5847 	uint_t		ccount;
5848 	int		bound = 0;
5849 	int		cmd_resp = 0;
5850 	fc_packet_t	*fpkt;
5851 	ddi_dma_cookie_t	pkt_data_cookie;
5852 	ddi_dma_cookie_t	*cp;
5853 	uint32_t		cnt;
5854 
5855 	fpkt = &icmd->ipkt_fc_packet;
5856 
5857 	ASSERT(fpkt->pkt_cmd_dma == NULL && fpkt->pkt_data_dma == NULL &&
5858 	    fpkt->pkt_resp_dma == NULL);
5859 
5860 	icmd->ipkt_nodma = nodma;
5861 
5862 	if (nodma) {
5863 		fpkt->pkt_cmd = kmem_zalloc(fpkt->pkt_cmdlen, flags);
5864 		if (fpkt->pkt_cmd == NULL) {
5865 			goto fail;
5866 		}
5867 
5868 		fpkt->pkt_resp = kmem_zalloc(fpkt->pkt_rsplen, flags);
5869 		if (fpkt->pkt_resp == NULL) {
5870 			goto fail;
5871 		}
5872 	} else {
5873 		ASSERT(fpkt->pkt_cmdlen && fpkt->pkt_rsplen);
5874 
5875 		rval = fcp_alloc_cmd_resp(pptr, fpkt, flags);
5876 		if (rval == FC_FAILURE) {
5877 			ASSERT(fpkt->pkt_cmd_dma == NULL &&
5878 			    fpkt->pkt_resp_dma == NULL);
5879 			goto fail;
5880 		}
5881 		cmd_resp++;
5882 	}
5883 
5884 	if (fpkt->pkt_datalen != 0) {
5885 		/*
5886 		 * set up DMA handle and memory for the data in this packet
5887 		 */
5888 		if (ddi_dma_alloc_handle(pptr->port_dip,
5889 		    &pptr->port_data_dma_attr, DDI_DMA_DONTWAIT,
5890 		    NULL, &fpkt->pkt_data_dma) != DDI_SUCCESS) {
5891 			goto fail;
5892 		}
5893 
5894 		if (ddi_dma_mem_alloc(fpkt->pkt_data_dma, fpkt->pkt_datalen,
5895 		    &pptr->port_dma_acc_attr, DDI_DMA_CONSISTENT,
5896 		    DDI_DMA_DONTWAIT, NULL, &fpkt->pkt_data,
5897 		    &real_size, &fpkt->pkt_data_acc) != DDI_SUCCESS) {
5898 			goto fail;
5899 		}
5900 
5901 		/* was DMA mem size gotten < size asked for/needed ?? */
5902 		if (real_size < fpkt->pkt_datalen) {
5903 			goto fail;
5904 		}
5905 
5906 		/* bind DMA address and handle together */
5907 		if (ddi_dma_addr_bind_handle(fpkt->pkt_data_dma,
5908 		    NULL, fpkt->pkt_data, real_size, DDI_DMA_READ |
5909 		    DDI_DMA_CONSISTENT, DDI_DMA_DONTWAIT, NULL,
5910 		    &pkt_data_cookie, &ccount) != DDI_DMA_MAPPED) {
5911 			goto fail;
5912 		}
5913 		bound++;
5914 
5915 		if (ccount > pptr->port_data_dma_attr.dma_attr_sgllen) {
5916 			goto fail;
5917 		}
5918 
5919 		fpkt->pkt_data_cookie_cnt = ccount;
5920 
5921 		cp = fpkt->pkt_data_cookie;
5922 		*cp = pkt_data_cookie;
5923 		cp++;
5924 
5925 		for (cnt = 1; cnt < ccount; cnt++, cp++) {
5926 			ddi_dma_nextcookie(fpkt->pkt_data_dma,
5927 			    &pkt_data_cookie);
5928 			*cp = pkt_data_cookie;
5929 		}
5930 
5931 	}
5932 
5933 	return (FC_SUCCESS);
5934 
5935 fail:
5936 	if (bound) {
5937 		(void) ddi_dma_unbind_handle(fpkt->pkt_data_dma);
5938 	}
5939 
5940 	if (fpkt->pkt_data_dma) {
5941 		if (fpkt->pkt_data) {
5942 			ddi_dma_mem_free(&fpkt->pkt_data_acc);
5943 		}
5944 		ddi_dma_free_handle(&fpkt->pkt_data_dma);
5945 	}
5946 
5947 	if (nodma) {
5948 		if (fpkt->pkt_cmd) {
5949 			kmem_free(fpkt->pkt_cmd, fpkt->pkt_cmdlen);
5950 		}
5951 		if (fpkt->pkt_resp) {
5952 			kmem_free(fpkt->pkt_resp, fpkt->pkt_rsplen);
5953 		}
5954 	} else {
5955 		if (cmd_resp) {
5956 			fcp_free_cmd_resp(pptr, fpkt);
5957 		}
5958 	}
5959 
5960 	return (FC_NOMEM);
5961 }
5962 
5963 
5964 static void
5965 fcp_free_dma(struct fcp_port *pptr, struct fcp_ipkt *icmd)
5966 {
5967 	fc_packet_t *fpkt = icmd->ipkt_fpkt;
5968 
5969 	if (fpkt->pkt_data_dma) {
5970 		(void) ddi_dma_unbind_handle(fpkt->pkt_data_dma);
5971 		if (fpkt->pkt_data) {
5972 			ddi_dma_mem_free(&fpkt->pkt_data_acc);
5973 		}
5974 		ddi_dma_free_handle(&fpkt->pkt_data_dma);
5975 	}
5976 
5977 	if (icmd->ipkt_nodma) {
5978 		if (fpkt->pkt_cmd) {
5979 			kmem_free(fpkt->pkt_cmd, icmd->ipkt_cmdlen);
5980 		}
5981 		if (fpkt->pkt_resp) {
5982 			kmem_free(fpkt->pkt_resp, icmd->ipkt_resplen);
5983 		}
5984 	} else {
5985 		ASSERT(fpkt->pkt_resp_dma != NULL && fpkt->pkt_cmd_dma != NULL);
5986 
5987 		fcp_free_cmd_resp(pptr, fpkt);
5988 	}
5989 }
5990 
5991 /*
5992  *     Function: fcp_lookup_target
5993  *
5994  *  Description: Finds a target given a WWN.
5995  *
5996  *     Argument: *pptr	FCP port.
5997  *		 *wwn	World Wide Name of the device to look for.
5998  *
5999  * Return Value: NULL		No target found
6000  *		 Not NULL	Target structure
6001  *
6002  *	Context: Interrupt context.
6003  *		 The mutex pptr->port_mutex must be owned.
6004  */
6005 /* ARGSUSED */
6006 static struct fcp_tgt *
6007 fcp_lookup_target(struct fcp_port *pptr, uchar_t *wwn)
6008 {
6009 	int			hash;
6010 	struct fcp_tgt	*ptgt;
6011 
6012 	ASSERT(mutex_owned(&pptr->port_mutex));
6013 
6014 	hash = FCP_HASH(wwn);
6015 
6016 	for (ptgt = pptr->port_tgt_hash_table[hash]; ptgt != NULL;
6017 	    ptgt = ptgt->tgt_next) {
6018 		if (!(ptgt->tgt_state & FCP_TGT_ORPHAN) &&
6019 		    bcmp((caddr_t)wwn, (caddr_t)&ptgt->tgt_port_wwn.raw_wwn[0],
6020 		    sizeof (ptgt->tgt_port_wwn)) == 0) {
6021 			break;
6022 		}
6023 	}
6024 
6025 	return (ptgt);
6026 }
6027 
6028 
6029 /*
6030  * Find target structure given a port identifier
6031  */
6032 static struct fcp_tgt *
6033 fcp_get_target_by_did(struct fcp_port *pptr, uint32_t d_id)
6034 {
6035 	fc_portid_t		port_id;
6036 	la_wwn_t		pwwn;
6037 	struct fcp_tgt	*ptgt = NULL;
6038 
6039 	port_id.priv_lilp_posit = 0;
6040 	port_id.port_id = d_id;
6041 	if (fc_ulp_get_pwwn_by_did(pptr->port_fp_handle, port_id,
6042 	    &pwwn) == FC_SUCCESS) {
6043 		mutex_enter(&pptr->port_mutex);
6044 		ptgt = fcp_lookup_target(pptr, pwwn.raw_wwn);
6045 		mutex_exit(&pptr->port_mutex);
6046 	}
6047 
6048 	return (ptgt);
6049 }
6050 
6051 
6052 /*
6053  * the packet completion callback routine for info cmd pkts
6054  *
6055  * this means fpkt pts to a response to either a PLOGI or a PRLI
6056  *
6057  * if there is an error an attempt is made to call a routine to resend
6058  * the command that failed
6059  */
6060 static void
6061 fcp_icmd_callback(fc_packet_t *fpkt)
6062 {
6063 	struct fcp_ipkt	*icmd;
6064 	struct fcp_port	*pptr;
6065 	struct fcp_tgt	*ptgt;
6066 	struct la_els_prli	*prli;
6067 	struct la_els_prli	prli_s;
6068 	struct fcp_prli		*fprli;
6069 	struct fcp_lun	*plun;
6070 	int		free_pkt = 1;
6071 	int		rval;
6072 	ls_code_t	resp;
6073 	uchar_t		prli_acc = 0;
6074 	uint32_t	rscn_count = FC_INVALID_RSCN_COUNT;
6075 	int		lun0_newalloc;
6076 
6077 	icmd = (struct fcp_ipkt *)fpkt->pkt_ulp_private;
6078 
6079 	/* get ptrs to the port and target structs for the cmd */
6080 	pptr = icmd->ipkt_port;
6081 	ptgt = icmd->ipkt_tgt;
6082 
6083 	FCP_CP_IN(fpkt->pkt_resp, &resp, fpkt->pkt_resp_acc, sizeof (resp));
6084 
6085 	if (icmd->ipkt_opcode == LA_ELS_PRLI) {
6086 		FCP_CP_IN(fpkt->pkt_cmd, &prli_s, fpkt->pkt_cmd_acc,
6087 		    sizeof (prli_s));
6088 		prli_acc = (prli_s.ls_code == LA_ELS_ACC);
6089 	}
6090 
6091 	FCP_TRACE(fcp_logq, pptr->port_instbuf,
6092 	    fcp_trace, FCP_BUF_LEVEL_2, 0,
6093 	    "ELS (%x) callback state=0x%x reason=0x%x for %x",
6094 	    icmd->ipkt_opcode, fpkt->pkt_state, fpkt->pkt_reason,
6095 	    ptgt->tgt_d_id);
6096 
6097 	if ((fpkt->pkt_state == FC_PKT_SUCCESS) &&
6098 	    ((resp.ls_code == LA_ELS_ACC) || prli_acc)) {
6099 
6100 		mutex_enter(&ptgt->tgt_mutex);
6101 		if (ptgt->tgt_pd_handle == NULL) {
6102 			/*
6103 			 * in a fabric environment the port device handles
6104 			 * get created only after successful LOGIN into the
6105 			 * transport, so the transport makes this port
6106 			 * device (pd) handle available in this packet, so
6107 			 * save it now
6108 			 */
6109 			ASSERT(fpkt->pkt_pd != NULL);
6110 			ptgt->tgt_pd_handle = fpkt->pkt_pd;
6111 		}
6112 		mutex_exit(&ptgt->tgt_mutex);
6113 
6114 		/* which ELS cmd is this response for ?? */
6115 		switch (icmd->ipkt_opcode) {
6116 		case LA_ELS_PLOGI:
6117 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
6118 			    fcp_trace, FCP_BUF_LEVEL_5, 0,
6119 			    "PLOGI to d_id=0x%x succeeded, wwn=%08x%08x",
6120 			    ptgt->tgt_d_id,
6121 			    *((int *)&ptgt->tgt_port_wwn.raw_wwn[0]),
6122 			    *((int *)&ptgt->tgt_port_wwn.raw_wwn[4]));
6123 
6124 			FCP_TGT_TRACE(ptgt, icmd->ipkt_change_cnt,
6125 			    FCP_TGT_TRACE_15);
6126 
6127 			/* Note that we are not allocating a new icmd */
6128 			if (fcp_send_els(pptr, ptgt, icmd, LA_ELS_PRLI,
6129 			    icmd->ipkt_link_cnt, icmd->ipkt_change_cnt,
6130 			    icmd->ipkt_cause) != DDI_SUCCESS) {
6131 				FCP_TGT_TRACE(ptgt, icmd->ipkt_change_cnt,
6132 				    FCP_TGT_TRACE_16);
6133 				goto fail;
6134 			}
6135 			break;
6136 
6137 		case LA_ELS_PRLI:
6138 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
6139 			    fcp_trace, FCP_BUF_LEVEL_5, 0,
6140 			    "PRLI to d_id=0x%x succeeded", ptgt->tgt_d_id);
6141 
6142 			FCP_TGT_TRACE(ptgt, icmd->ipkt_change_cnt,
6143 			    FCP_TGT_TRACE_17);
6144 
6145 			prli = &prli_s;
6146 
6147 			FCP_CP_IN(fpkt->pkt_resp, prli, fpkt->pkt_resp_acc,
6148 			    sizeof (prli_s));
6149 
6150 			fprli = (struct fcp_prli *)prli->service_params;
6151 
6152 			mutex_enter(&ptgt->tgt_mutex);
6153 			ptgt->tgt_icap = fprli->initiator_fn;
6154 			ptgt->tgt_tcap = fprli->target_fn;
6155 			mutex_exit(&ptgt->tgt_mutex);
6156 
6157 			if ((fprli->type != 0x08) || (fprli->target_fn != 1)) {
6158 				/*
6159 				 * this FCP device does not support target mode
6160 				 */
6161 				FCP_TGT_TRACE(ptgt, icmd->ipkt_change_cnt,
6162 				    FCP_TGT_TRACE_18);
6163 				goto fail;
6164 			}
6165 			if (fprli->retry == 1) {
6166 				fc_ulp_disable_relogin(pptr->port_fp_handle,
6167 				    &ptgt->tgt_port_wwn);
6168 			}
6169 
6170 			/* target is no longer offline */
6171 			mutex_enter(&pptr->port_mutex);
6172 			mutex_enter(&ptgt->tgt_mutex);
6173 			if (!FCP_TGT_STATE_CHANGED(ptgt, icmd)) {
6174 				ptgt->tgt_state &= ~(FCP_TGT_OFFLINE |
6175 				    FCP_TGT_MARK);
6176 			} else {
6177 				FCP_TRACE(fcp_logq, pptr->port_instbuf,
6178 				    fcp_trace, FCP_BUF_LEVEL_2, 0,
6179 				    "fcp_icmd_callback,1: state change "
6180 				    " occured for D_ID=0x%x", ptgt->tgt_d_id);
6181 				mutex_exit(&ptgt->tgt_mutex);
6182 				mutex_exit(&pptr->port_mutex);
6183 				goto fail;
6184 			}
6185 			mutex_exit(&ptgt->tgt_mutex);
6186 			mutex_exit(&pptr->port_mutex);
6187 
6188 			/*
6189 			 * lun 0 should always respond to inquiry, so
6190 			 * get the LUN struct for LUN 0
6191 			 *
6192 			 * Currently we deal with first level of addressing.
6193 			 * If / when we start supporting 0x device types
6194 			 * (DTYPE_ARRAY_CTRL, i.e. array controllers)
6195 			 * this logic will need revisiting.
6196 			 */
6197 			lun0_newalloc = 0;
6198 			if ((plun = fcp_get_lun(ptgt, 0)) == NULL) {
6199 				/*
6200 				 * no LUN struct for LUN 0 yet exists,
6201 				 * so create one
6202 				 */
6203 				plun = fcp_alloc_lun(ptgt);
6204 				if (plun == NULL) {
6205 					fcp_log(CE_WARN, pptr->port_dip,
6206 					    "!Failed to allocate lun 0 for"
6207 					    " D_ID=%x", ptgt->tgt_d_id);
6208 					goto fail;
6209 				}
6210 				lun0_newalloc = 1;
6211 			}
6212 
6213 			/* fill in LUN info */
6214 			mutex_enter(&ptgt->tgt_mutex);
6215 			/*
6216 			 * consider lun 0 as device not connected if it is
6217 			 * offlined or newly allocated
6218 			 */
6219 			if ((plun->lun_state & FCP_LUN_OFFLINE) ||
6220 			    lun0_newalloc) {
6221 				plun->lun_state |= FCP_LUN_DEVICE_NOT_CONNECTED;
6222 			}
6223 			plun->lun_state |= (FCP_LUN_BUSY | FCP_LUN_MARK);
6224 			plun->lun_state &= ~FCP_LUN_OFFLINE;
6225 			ptgt->tgt_lun_cnt = 1;
6226 			ptgt->tgt_report_lun_cnt = 0;
6227 			mutex_exit(&ptgt->tgt_mutex);
6228 
6229 			/* Retrieve the rscn count (if a valid one exists) */
6230 			if (icmd->ipkt_fpkt->pkt_ulp_rscn_infop != NULL) {
6231 				rscn_count = ((fc_ulp_rscn_info_t *)
6232 				    (icmd->ipkt_fpkt->pkt_ulp_rscn_infop))
6233 				    ->ulp_rscn_count;
6234 			} else {
6235 				rscn_count = FC_INVALID_RSCN_COUNT;
6236 			}
6237 
6238 			/* send Report Lun request to target */
6239 			if (fcp_send_scsi(plun, SCMD_REPORT_LUN,
6240 			    sizeof (struct fcp_reportlun_resp),
6241 			    icmd->ipkt_link_cnt, icmd->ipkt_change_cnt,
6242 			    icmd->ipkt_cause, rscn_count) != DDI_SUCCESS) {
6243 				mutex_enter(&pptr->port_mutex);
6244 				if (!FCP_TGT_STATE_CHANGED(ptgt, icmd)) {
6245 					fcp_log(CE_WARN, pptr->port_dip,
6246 					    "!Failed to send REPORT LUN to"
6247 					    "  D_ID=%x", ptgt->tgt_d_id);
6248 				} else {
6249 					FCP_TRACE(fcp_logq,
6250 					    pptr->port_instbuf, fcp_trace,
6251 					    FCP_BUF_LEVEL_5, 0,
6252 					    "fcp_icmd_callback,2:state change"
6253 					    " occured for D_ID=0x%x",
6254 					    ptgt->tgt_d_id);
6255 				}
6256 				mutex_exit(&pptr->port_mutex);
6257 
6258 				FCP_TGT_TRACE(ptgt, icmd->ipkt_change_cnt,
6259 				    FCP_TGT_TRACE_19);
6260 
6261 				goto fail;
6262 			} else {
6263 				free_pkt = 0;
6264 				fcp_icmd_free(pptr, icmd);
6265 			}
6266 			break;
6267 
6268 		default:
6269 			fcp_log(CE_WARN, pptr->port_dip,
6270 			    "!fcp_icmd_callback Invalid opcode");
6271 			goto fail;
6272 		}
6273 
6274 		return;
6275 	}
6276 
6277 
6278 	/*
6279 	 * Other PLOGI failures are not retried as the
6280 	 * transport does it already
6281 	 */
6282 	if (icmd->ipkt_opcode != LA_ELS_PLOGI) {
6283 		if (fcp_is_retryable(icmd) &&
6284 		    icmd->ipkt_retries++ < FCP_MAX_RETRIES) {
6285 
6286 			if (FCP_MUST_RETRY(fpkt)) {
6287 				fcp_queue_ipkt(pptr, fpkt);
6288 				return;
6289 			}
6290 
6291 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
6292 			    fcp_trace, FCP_BUF_LEVEL_2, 0,
6293 			    "ELS PRLI is retried for d_id=0x%x, state=%x,"
6294 			    " reason= %x", ptgt->tgt_d_id, fpkt->pkt_state,
6295 			    fpkt->pkt_reason);
6296 
6297 			/*
6298 			 * Retry by recalling the routine that
6299 			 * originally queued this packet
6300 			 */
6301 			mutex_enter(&pptr->port_mutex);
6302 			if (!FCP_TGT_STATE_CHANGED(ptgt, icmd)) {
6303 				caddr_t msg;
6304 
6305 				mutex_exit(&pptr->port_mutex);
6306 
6307 				ASSERT(icmd->ipkt_opcode != LA_ELS_PLOGI);
6308 
6309 				if (fpkt->pkt_state == FC_PKT_TIMEOUT) {
6310 					fpkt->pkt_timeout +=
6311 					    FCP_TIMEOUT_DELTA;
6312 				}
6313 
6314 				rval = fc_ulp_issue_els(pptr->port_fp_handle,
6315 				    fpkt);
6316 				if (rval == FC_SUCCESS) {
6317 					return;
6318 				}
6319 
6320 				if (rval == FC_STATEC_BUSY ||
6321 				    rval == FC_OFFLINE) {
6322 					fcp_queue_ipkt(pptr, fpkt);
6323 					return;
6324 				}
6325 				(void) fc_ulp_error(rval, &msg);
6326 
6327 				fcp_log(CE_NOTE, pptr->port_dip,
6328 				    "!ELS 0x%x failed to d_id=0x%x;"
6329 				    " %s", icmd->ipkt_opcode,
6330 				    ptgt->tgt_d_id, msg);
6331 			} else {
6332 				FCP_TRACE(fcp_logq, pptr->port_instbuf,
6333 				    fcp_trace, FCP_BUF_LEVEL_2, 0,
6334 				    "fcp_icmd_callback,3: state change "
6335 				    " occured for D_ID=0x%x", ptgt->tgt_d_id);
6336 				mutex_exit(&pptr->port_mutex);
6337 			}
6338 		}
6339 	} else {
6340 		if (fcp_is_retryable(icmd) &&
6341 		    icmd->ipkt_retries++ < FCP_MAX_RETRIES) {
6342 			if (FCP_MUST_RETRY(fpkt)) {
6343 				fcp_queue_ipkt(pptr, fpkt);
6344 				return;
6345 			}
6346 		}
6347 		mutex_enter(&pptr->port_mutex);
6348 		if (!FCP_TGT_STATE_CHANGED(ptgt, icmd) &&
6349 		    fpkt->pkt_state != FC_PKT_PORT_OFFLINE) {
6350 			mutex_exit(&pptr->port_mutex);
6351 			fcp_print_error(fpkt);
6352 		} else {
6353 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
6354 			    fcp_trace, FCP_BUF_LEVEL_2, 0,
6355 			    "fcp_icmd_callback,4: state change occured"
6356 			    " for D_ID=0x%x", ptgt->tgt_d_id);
6357 			mutex_exit(&pptr->port_mutex);
6358 		}
6359 	}
6360 
6361 fail:
6362 	if (free_pkt) {
6363 		(void) fcp_call_finish_init(pptr, ptgt, icmd->ipkt_link_cnt,
6364 		    icmd->ipkt_change_cnt, icmd->ipkt_cause);
6365 		fcp_icmd_free(pptr, icmd);
6366 	}
6367 }
6368 
6369 
6370 /*
6371  * called internally to send an info cmd using the transport
6372  *
6373  * sends either an INQ or a REPORT_LUN
6374  *
6375  * when the packet is completed fcp_scsi_callback is called
6376  */
6377 static int
6378 fcp_send_scsi(struct fcp_lun *plun, uchar_t opcode, int alloc_len,
6379     int lcount, int tcount, int cause, uint32_t rscn_count)
6380 {
6381 	int			nodma;
6382 	struct fcp_ipkt		*icmd;
6383 	struct fcp_tgt		*ptgt;
6384 	struct fcp_port		*pptr;
6385 	fc_frame_hdr_t		*hp;
6386 	fc_packet_t		*fpkt;
6387 	struct fcp_cmd		fcp_cmd;
6388 	struct fcp_cmd		*fcmd;
6389 	union scsi_cdb		*scsi_cdb;
6390 
6391 	ASSERT(plun != NULL);
6392 
6393 	ptgt = plun->lun_tgt;
6394 	ASSERT(ptgt != NULL);
6395 
6396 	pptr = ptgt->tgt_port;
6397 	ASSERT(pptr != NULL);
6398 
6399 	FCP_TRACE(fcp_logq, pptr->port_instbuf,
6400 	    fcp_trace, FCP_BUF_LEVEL_5, 0,
6401 	    "fcp_send_scsi: d_id=0x%x opcode=0x%x", ptgt->tgt_d_id, opcode);
6402 
6403 	nodma = (pptr->port_fcp_dma == FC_NO_DVMA_SPACE) ? 1 : 0;
6404 
6405 	icmd = fcp_icmd_alloc(pptr, ptgt, sizeof (struct fcp_cmd),
6406 	    FCP_MAX_RSP_IU_SIZE, alloc_len, nodma, lcount, tcount, cause,
6407 	    rscn_count);
6408 
6409 	if (icmd == NULL) {
6410 		return (DDI_FAILURE);
6411 	}
6412 
6413 	fpkt = icmd->ipkt_fpkt;
6414 	fpkt->pkt_tran_flags = FC_TRAN_CLASS3 | FC_TRAN_INTR;
6415 	icmd->ipkt_retries = 0;
6416 	icmd->ipkt_opcode = opcode;
6417 	icmd->ipkt_lun = plun;
6418 
6419 	if (nodma) {
6420 		fcmd = (struct fcp_cmd *)fpkt->pkt_cmd;
6421 	} else {
6422 		fcmd = &fcp_cmd;
6423 	}
6424 	bzero(fcmd, sizeof (struct fcp_cmd));
6425 
6426 	fpkt->pkt_timeout = FCP_SCSI_CMD_TIMEOUT;
6427 
6428 	hp = &fpkt->pkt_cmd_fhdr;
6429 
6430 	hp->s_id = pptr->port_id;
6431 	hp->d_id = ptgt->tgt_d_id;
6432 	hp->r_ctl = R_CTL_COMMAND;
6433 	hp->type = FC_TYPE_SCSI_FCP;
6434 	hp->f_ctl = F_CTL_SEQ_INITIATIVE | F_CTL_FIRST_SEQ;
6435 	hp->rsvd = 0;
6436 	hp->seq_id = 0;
6437 	hp->seq_cnt = 0;
6438 	hp->ox_id = 0xffff;
6439 	hp->rx_id = 0xffff;
6440 	hp->ro = 0;
6441 
6442 	bcopy(&(plun->lun_addr), &(fcmd->fcp_ent_addr), FCP_LUN_SIZE);
6443 
6444 	/*
6445 	 * Request SCSI target for expedited processing
6446 	 */
6447 
6448 	/*
6449 	 * Set up for untagged queuing because we do not
6450 	 * know if the fibre device supports queuing.
6451 	 */
6452 	fcmd->fcp_cntl.cntl_reserved_0 = 0;
6453 	fcmd->fcp_cntl.cntl_reserved_1 = 0;
6454 	fcmd->fcp_cntl.cntl_reserved_2 = 0;
6455 	fcmd->fcp_cntl.cntl_reserved_3 = 0;
6456 	fcmd->fcp_cntl.cntl_reserved_4 = 0;
6457 	fcmd->fcp_cntl.cntl_qtype = FCP_QTYPE_UNTAGGED;
6458 	scsi_cdb = (union scsi_cdb *)fcmd->fcp_cdb;
6459 
6460 	switch (opcode) {
6461 	case SCMD_INQUIRY_PAGE83:
6462 		/*
6463 		 * Prepare to get the Inquiry VPD page 83 information
6464 		 */
6465 		fcmd->fcp_cntl.cntl_read_data = 1;
6466 		fcmd->fcp_cntl.cntl_write_data = 0;
6467 		fcmd->fcp_data_len = alloc_len;
6468 
6469 		fpkt->pkt_tran_type = FC_PKT_FCP_READ;
6470 		fpkt->pkt_comp = fcp_scsi_callback;
6471 
6472 		scsi_cdb->scc_cmd = SCMD_INQUIRY;
6473 		scsi_cdb->g0_addr2 = 0x01;
6474 		scsi_cdb->g0_addr1 = 0x83;
6475 		scsi_cdb->g0_count0 = (uchar_t)alloc_len;
6476 		break;
6477 
6478 	case SCMD_INQUIRY:
6479 		fcmd->fcp_cntl.cntl_read_data = 1;
6480 		fcmd->fcp_cntl.cntl_write_data = 0;
6481 		fcmd->fcp_data_len = alloc_len;
6482 
6483 		fpkt->pkt_tran_type = FC_PKT_FCP_READ;
6484 		fpkt->pkt_comp = fcp_scsi_callback;
6485 
6486 		scsi_cdb->scc_cmd = SCMD_INQUIRY;
6487 		scsi_cdb->g0_count0 = SUN_INQSIZE;
6488 		break;
6489 
6490 	case SCMD_REPORT_LUN: {
6491 		fc_portid_t	d_id;
6492 		opaque_t	fca_dev;
6493 
6494 		ASSERT(alloc_len >= 16);
6495 
6496 		d_id.priv_lilp_posit = 0;
6497 		d_id.port_id = ptgt->tgt_d_id;
6498 
6499 		fca_dev = fc_ulp_get_fca_device(pptr->port_fp_handle, d_id);
6500 
6501 		mutex_enter(&ptgt->tgt_mutex);
6502 		ptgt->tgt_fca_dev = fca_dev;
6503 		mutex_exit(&ptgt->tgt_mutex);
6504 
6505 		fcmd->fcp_cntl.cntl_read_data = 1;
6506 		fcmd->fcp_cntl.cntl_write_data = 0;
6507 		fcmd->fcp_data_len = alloc_len;
6508 
6509 		fpkt->pkt_tran_type = FC_PKT_FCP_READ;
6510 		fpkt->pkt_comp = fcp_scsi_callback;
6511 
6512 		scsi_cdb->scc_cmd = SCMD_REPORT_LUN;
6513 		scsi_cdb->scc5_count0 = alloc_len & 0xff;
6514 		scsi_cdb->scc5_count1 = (alloc_len >> 8) & 0xff;
6515 		scsi_cdb->scc5_count2 = (alloc_len >> 16) & 0xff;
6516 		scsi_cdb->scc5_count3 = (alloc_len >> 24) & 0xff;
6517 		break;
6518 	}
6519 
6520 	default:
6521 		fcp_log(CE_WARN, pptr->port_dip,
6522 		    "!fcp_send_scsi Invalid opcode");
6523 		break;
6524 	}
6525 
6526 	if (!nodma) {
6527 		FCP_CP_OUT((uint8_t *)fcmd, fpkt->pkt_cmd,
6528 		    fpkt->pkt_cmd_acc, sizeof (struct fcp_cmd));
6529 	}
6530 
6531 	mutex_enter(&pptr->port_mutex);
6532 	if (!FCP_TGT_STATE_CHANGED(ptgt, icmd)) {
6533 
6534 		mutex_exit(&pptr->port_mutex);
6535 		if (fcp_transport(pptr->port_fp_handle, fpkt, 1) !=
6536 		    FC_SUCCESS) {
6537 			fcp_icmd_free(pptr, icmd);
6538 			return (DDI_FAILURE);
6539 		}
6540 		return (DDI_SUCCESS);
6541 	} else {
6542 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
6543 		    fcp_trace, FCP_BUF_LEVEL_2, 0,
6544 		    "fcp_send_scsi,1: state change occured"
6545 		    " for D_ID=0x%x", ptgt->tgt_d_id);
6546 		mutex_exit(&pptr->port_mutex);
6547 		fcp_icmd_free(pptr, icmd);
6548 		return (DDI_FAILURE);
6549 	}
6550 }
6551 
6552 
6553 /*
6554  * called by fcp_scsi_callback to check to handle the case where
6555  * REPORT_LUN returns ILLEGAL REQUEST or a UNIT ATTENTION
6556  */
6557 static int
6558 fcp_check_reportlun(struct fcp_rsp *rsp, fc_packet_t *fpkt)
6559 {
6560 	uchar_t				rqlen;
6561 	int				rval = DDI_FAILURE;
6562 	struct scsi_extended_sense	sense_info, *sense;
6563 	struct fcp_ipkt		*icmd = (struct fcp_ipkt *)
6564 	    fpkt->pkt_ulp_private;
6565 	struct fcp_tgt		*ptgt = icmd->ipkt_tgt;
6566 	struct fcp_port		*pptr = ptgt->tgt_port;
6567 
6568 	ASSERT(icmd->ipkt_opcode == SCMD_REPORT_LUN);
6569 
6570 	if (rsp->fcp_u.fcp_status.scsi_status == STATUS_RESERVATION_CONFLICT) {
6571 		/*
6572 		 * SCSI-II Reserve Release support. Some older FC drives return
6573 		 * Reservation conflict for Report Luns command.
6574 		 */
6575 		if (icmd->ipkt_nodma) {
6576 			rsp->fcp_u.fcp_status.rsp_len_set = 0;
6577 			rsp->fcp_u.fcp_status.sense_len_set = 0;
6578 			rsp->fcp_u.fcp_status.scsi_status = STATUS_GOOD;
6579 		} else {
6580 			fcp_rsp_t	new_resp;
6581 
6582 			FCP_CP_IN(fpkt->pkt_resp, &new_resp,
6583 			    fpkt->pkt_resp_acc, sizeof (new_resp));
6584 
6585 			new_resp.fcp_u.fcp_status.rsp_len_set = 0;
6586 			new_resp.fcp_u.fcp_status.sense_len_set = 0;
6587 			new_resp.fcp_u.fcp_status.scsi_status = STATUS_GOOD;
6588 
6589 			FCP_CP_OUT(&new_resp, fpkt->pkt_resp,
6590 			    fpkt->pkt_resp_acc, sizeof (new_resp));
6591 		}
6592 
6593 		FCP_CP_OUT(fcp_dummy_lun, fpkt->pkt_data,
6594 		    fpkt->pkt_data_acc, sizeof (fcp_dummy_lun));
6595 
6596 		return (DDI_SUCCESS);
6597 	}
6598 
6599 	sense = &sense_info;
6600 	if (!rsp->fcp_u.fcp_status.sense_len_set) {
6601 		/* no need to continue if sense length is not set */
6602 		return (rval);
6603 	}
6604 
6605 	/* casting 64-bit integer to 8-bit */
6606 	rqlen = (uchar_t)min(rsp->fcp_sense_len,
6607 	    sizeof (struct scsi_extended_sense));
6608 
6609 	if (rqlen < 14) {
6610 		/* no need to continue if request length isn't long enough */
6611 		return (rval);
6612 	}
6613 
6614 	if (icmd->ipkt_nodma) {
6615 		/*
6616 		 * We can safely use fcp_response_len here since the
6617 		 * only path that calls fcp_check_reportlun,
6618 		 * fcp_scsi_callback, has already called
6619 		 * fcp_validate_fcp_response.
6620 		 */
6621 		sense = (struct scsi_extended_sense *)(fpkt->pkt_resp +
6622 		    sizeof (struct fcp_rsp) + rsp->fcp_response_len);
6623 	} else {
6624 		FCP_CP_IN(fpkt->pkt_resp + sizeof (struct fcp_rsp) +
6625 		    rsp->fcp_response_len, sense, fpkt->pkt_resp_acc,
6626 		    sizeof (struct scsi_extended_sense));
6627 	}
6628 
6629 	if (!FCP_SENSE_NO_LUN(sense)) {
6630 		mutex_enter(&ptgt->tgt_mutex);
6631 		/* clear the flag if any */
6632 		ptgt->tgt_state &= ~FCP_TGT_ILLREQ;
6633 		mutex_exit(&ptgt->tgt_mutex);
6634 	}
6635 
6636 	if ((sense->es_key == KEY_ILLEGAL_REQUEST) &&
6637 	    (sense->es_add_code == 0x20)) {
6638 		if (icmd->ipkt_nodma) {
6639 			rsp->fcp_u.fcp_status.rsp_len_set = 0;
6640 			rsp->fcp_u.fcp_status.sense_len_set = 0;
6641 			rsp->fcp_u.fcp_status.scsi_status = STATUS_GOOD;
6642 		} else {
6643 			fcp_rsp_t	new_resp;
6644 
6645 			FCP_CP_IN(fpkt->pkt_resp, &new_resp,
6646 			    fpkt->pkt_resp_acc, sizeof (new_resp));
6647 
6648 			new_resp.fcp_u.fcp_status.rsp_len_set = 0;
6649 			new_resp.fcp_u.fcp_status.sense_len_set = 0;
6650 			new_resp.fcp_u.fcp_status.scsi_status = STATUS_GOOD;
6651 
6652 			FCP_CP_OUT(&new_resp, fpkt->pkt_resp,
6653 			    fpkt->pkt_resp_acc, sizeof (new_resp));
6654 		}
6655 
6656 		FCP_CP_OUT(fcp_dummy_lun, fpkt->pkt_data,
6657 		    fpkt->pkt_data_acc, sizeof (fcp_dummy_lun));
6658 
6659 		return (DDI_SUCCESS);
6660 	}
6661 
6662 	/*
6663 	 * This is for the STK library which returns a check condition,
6664 	 * to indicate device is not ready, manual assistance needed.
6665 	 * This is to a report lun command when the door is open.
6666 	 */
6667 	if ((sense->es_key == KEY_NOT_READY) && (sense->es_add_code == 0x04)) {
6668 		if (icmd->ipkt_nodma) {
6669 			rsp->fcp_u.fcp_status.rsp_len_set = 0;
6670 			rsp->fcp_u.fcp_status.sense_len_set = 0;
6671 			rsp->fcp_u.fcp_status.scsi_status = STATUS_GOOD;
6672 		} else {
6673 			fcp_rsp_t	new_resp;
6674 
6675 			FCP_CP_IN(fpkt->pkt_resp, &new_resp,
6676 			    fpkt->pkt_resp_acc, sizeof (new_resp));
6677 
6678 			new_resp.fcp_u.fcp_status.rsp_len_set = 0;
6679 			new_resp.fcp_u.fcp_status.sense_len_set = 0;
6680 			new_resp.fcp_u.fcp_status.scsi_status = STATUS_GOOD;
6681 
6682 			FCP_CP_OUT(&new_resp, fpkt->pkt_resp,
6683 			    fpkt->pkt_resp_acc, sizeof (new_resp));
6684 		}
6685 
6686 		FCP_CP_OUT(fcp_dummy_lun, fpkt->pkt_data,
6687 		    fpkt->pkt_data_acc, sizeof (fcp_dummy_lun));
6688 
6689 		return (DDI_SUCCESS);
6690 	}
6691 
6692 	if ((FCP_SENSE_REPORTLUN_CHANGED(sense)) ||
6693 	    (FCP_SENSE_NO_LUN(sense))) {
6694 		mutex_enter(&ptgt->tgt_mutex);
6695 		if ((FCP_SENSE_NO_LUN(sense)) &&
6696 		    (ptgt->tgt_state & FCP_TGT_ILLREQ)) {
6697 			ptgt->tgt_state &= ~FCP_TGT_ILLREQ;
6698 			mutex_exit(&ptgt->tgt_mutex);
6699 			/*
6700 			 * reconfig was triggred by ILLEGAL REQUEST but
6701 			 * got ILLEGAL REQUEST again
6702 			 */
6703 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
6704 			    fcp_trace, FCP_BUF_LEVEL_3, 0,
6705 			    "!FCP: Unable to obtain Report Lun data"
6706 			    " target=%x", ptgt->tgt_d_id);
6707 		} else {
6708 			if (ptgt->tgt_tid == NULL) {
6709 				timeout_id_t	tid;
6710 				/*
6711 				 * REPORT LUN data has changed.	 Kick off
6712 				 * rediscovery
6713 				 */
6714 				tid = timeout(fcp_reconfigure_luns,
6715 				    (caddr_t)ptgt, (clock_t)drv_usectohz(1));
6716 
6717 				ptgt->tgt_tid = tid;
6718 				ptgt->tgt_state |= FCP_TGT_BUSY;
6719 			}
6720 			if (FCP_SENSE_NO_LUN(sense)) {
6721 				ptgt->tgt_state |= FCP_TGT_ILLREQ;
6722 			}
6723 			mutex_exit(&ptgt->tgt_mutex);
6724 			if (FCP_SENSE_REPORTLUN_CHANGED(sense)) {
6725 				FCP_TRACE(fcp_logq, pptr->port_instbuf,
6726 				    fcp_trace, FCP_BUF_LEVEL_3, 0,
6727 				    "!FCP:Report Lun Has Changed"
6728 				    " target=%x", ptgt->tgt_d_id);
6729 			} else if (FCP_SENSE_NO_LUN(sense)) {
6730 				FCP_TRACE(fcp_logq, pptr->port_instbuf,
6731 				    fcp_trace, FCP_BUF_LEVEL_3, 0,
6732 				    "!FCP:LU Not Supported"
6733 				    " target=%x", ptgt->tgt_d_id);
6734 			}
6735 		}
6736 		rval = DDI_SUCCESS;
6737 	}
6738 
6739 	FCP_TRACE(fcp_logq, pptr->port_instbuf,
6740 	    fcp_trace, FCP_BUF_LEVEL_5, 0,
6741 	    "D_ID=%x, sense=%x, status=%x",
6742 	    fpkt->pkt_cmd_fhdr.d_id, sense->es_key,
6743 	    rsp->fcp_u.fcp_status.scsi_status);
6744 
6745 	return (rval);
6746 }
6747 
6748 /*
6749  *     Function: fcp_scsi_callback
6750  *
6751  *  Description: This is the callback routine set by fcp_send_scsi() after
6752  *		 it calls fcp_icmd_alloc().  The SCSI command completed here
6753  *		 and autogenerated by FCP are:	REPORT_LUN, INQUIRY and
6754  *		 INQUIRY_PAGE83.
6755  *
6756  *     Argument: *fpkt	 FC packet used to convey the command
6757  *
6758  * Return Value: None
6759  */
6760 static void
6761 fcp_scsi_callback(fc_packet_t *fpkt)
6762 {
6763 	struct fcp_ipkt	*icmd = (struct fcp_ipkt *)
6764 	    fpkt->pkt_ulp_private;
6765 	struct fcp_rsp_info	fcp_rsp_err, *bep;
6766 	struct fcp_port	*pptr;
6767 	struct fcp_tgt	*ptgt;
6768 	struct fcp_lun	*plun;
6769 	struct fcp_rsp		response, *rsp;
6770 
6771 	if (icmd->ipkt_nodma) {
6772 		rsp = (struct fcp_rsp *)fpkt->pkt_resp;
6773 	} else {
6774 		rsp = &response;
6775 		FCP_CP_IN(fpkt->pkt_resp, rsp, fpkt->pkt_resp_acc,
6776 		    sizeof (struct fcp_rsp));
6777 	}
6778 
6779 	ptgt = icmd->ipkt_tgt;
6780 	pptr = ptgt->tgt_port;
6781 	plun = icmd->ipkt_lun;
6782 
6783 	FCP_TRACE(fcp_logq, pptr->port_instbuf,
6784 	    fcp_trace, FCP_BUF_LEVEL_2, 0,
6785 	    "SCSI callback state=0x%x for %x, op_code=0x%x, "
6786 	    "status=%x, lun num=%x",
6787 	    fpkt->pkt_state, ptgt->tgt_d_id, icmd->ipkt_opcode,
6788 	    rsp->fcp_u.fcp_status.scsi_status, plun->lun_num);
6789 
6790 	/*
6791 	 * Pre-init LUN GUID with NWWN if it is not a device that
6792 	 * supports multiple luns and we know it's not page83
6793 	 * compliant.  Although using a NWWN is not lun unique,
6794 	 * we will be fine since there is only one lun behind the taget
6795 	 * in this case.
6796 	 */
6797 	if ((plun->lun_guid_size == 0) &&
6798 	    (icmd->ipkt_opcode == SCMD_INQUIRY_PAGE83) &&
6799 	    (fcp_symmetric_device_probe(plun) == 0)) {
6800 
6801 		char ascii_wwn[FC_WWN_SIZE*2+1];
6802 		fcp_wwn_to_ascii(&ptgt->tgt_node_wwn.raw_wwn[0], ascii_wwn);
6803 		(void) fcp_copy_guid_2_lun_block(plun, ascii_wwn);
6804 	}
6805 
6806 	/*
6807 	 * Some old FC tapes and FC <-> SCSI bridge devices return overrun
6808 	 * when thay have more data than what is asked in CDB. An overrun
6809 	 * is really when FCP_DL is smaller than the data length in CDB.
6810 	 * In the case here we know that REPORT LUN command we formed within
6811 	 * this binary has correct FCP_DL. So this OVERRUN is due to bad device
6812 	 * behavior. In reality this is FC_SUCCESS.
6813 	 */
6814 	if ((fpkt->pkt_state != FC_PKT_SUCCESS) &&
6815 	    (fpkt->pkt_reason == FC_REASON_OVERRUN) &&
6816 	    (icmd->ipkt_opcode == SCMD_REPORT_LUN)) {
6817 		fpkt->pkt_state = FC_PKT_SUCCESS;
6818 	}
6819 
6820 	if (fpkt->pkt_state != FC_PKT_SUCCESS) {
6821 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
6822 		    fcp_trace, FCP_BUF_LEVEL_2, 0,
6823 		    "icmd failed with state=0x%x for %x", fpkt->pkt_state,
6824 		    ptgt->tgt_d_id);
6825 
6826 		if (fpkt->pkt_reason == FC_REASON_CRC_ERROR) {
6827 			/*
6828 			 * Inquiry VPD page command on A5K SES devices would
6829 			 * result in data CRC errors.
6830 			 */
6831 			if (icmd->ipkt_opcode == SCMD_INQUIRY_PAGE83) {
6832 				(void) fcp_handle_page83(fpkt, icmd, 1);
6833 				return;
6834 			}
6835 		}
6836 		if (fpkt->pkt_state == FC_PKT_TIMEOUT ||
6837 		    FCP_MUST_RETRY(fpkt)) {
6838 			fpkt->pkt_timeout += FCP_TIMEOUT_DELTA;
6839 			fcp_retry_scsi_cmd(fpkt);
6840 			return;
6841 		}
6842 
6843 		FCP_TGT_TRACE(ptgt, icmd->ipkt_change_cnt,
6844 		    FCP_TGT_TRACE_20);
6845 
6846 		mutex_enter(&pptr->port_mutex);
6847 		mutex_enter(&ptgt->tgt_mutex);
6848 		if (!FCP_STATE_CHANGED(pptr, ptgt, icmd)) {
6849 			mutex_exit(&ptgt->tgt_mutex);
6850 			mutex_exit(&pptr->port_mutex);
6851 			fcp_print_error(fpkt);
6852 		} else {
6853 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
6854 			    fcp_trace, FCP_BUF_LEVEL_2, 0,
6855 			    "fcp_scsi_callback,1: state change occured"
6856 			    " for D_ID=0x%x", ptgt->tgt_d_id);
6857 			mutex_exit(&ptgt->tgt_mutex);
6858 			mutex_exit(&pptr->port_mutex);
6859 		}
6860 		(void) fcp_call_finish_init(pptr, ptgt, icmd->ipkt_link_cnt,
6861 		    icmd->ipkt_change_cnt, icmd->ipkt_cause);
6862 		fcp_icmd_free(pptr, icmd);
6863 		return;
6864 	}
6865 
6866 	FCP_TGT_TRACE(ptgt, icmd->ipkt_change_cnt, FCP_TGT_TRACE_21);
6867 
6868 	mutex_enter(&pptr->port_mutex);
6869 	mutex_enter(&ptgt->tgt_mutex);
6870 	if (FCP_STATE_CHANGED(pptr, ptgt, icmd)) {
6871 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
6872 		    fcp_trace, FCP_BUF_LEVEL_2, 0,
6873 		    "fcp_scsi_callback,2: state change occured"
6874 		    " for D_ID=0x%x", ptgt->tgt_d_id);
6875 		mutex_exit(&ptgt->tgt_mutex);
6876 		mutex_exit(&pptr->port_mutex);
6877 		(void) fcp_call_finish_init(pptr, ptgt, icmd->ipkt_link_cnt,
6878 		    icmd->ipkt_change_cnt, icmd->ipkt_cause);
6879 		fcp_icmd_free(pptr, icmd);
6880 		return;
6881 	}
6882 	ASSERT((ptgt->tgt_state & FCP_TGT_MARK) == 0);
6883 
6884 	mutex_exit(&ptgt->tgt_mutex);
6885 	mutex_exit(&pptr->port_mutex);
6886 
6887 	if (icmd->ipkt_nodma) {
6888 		bep = (struct fcp_rsp_info *)(fpkt->pkt_resp +
6889 		    sizeof (struct fcp_rsp));
6890 	} else {
6891 		bep = &fcp_rsp_err;
6892 		FCP_CP_IN(fpkt->pkt_resp + sizeof (struct fcp_rsp), bep,
6893 		    fpkt->pkt_resp_acc, sizeof (struct fcp_rsp_info));
6894 	}
6895 
6896 	if (fcp_validate_fcp_response(rsp, pptr) != FC_SUCCESS) {
6897 		fcp_retry_scsi_cmd(fpkt);
6898 		return;
6899 	}
6900 
6901 	if (rsp->fcp_u.fcp_status.rsp_len_set && bep->rsp_code !=
6902 	    FCP_NO_FAILURE) {
6903 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
6904 		    fcp_trace, FCP_BUF_LEVEL_2, 0,
6905 		    "rsp_code=0x%x, rsp_len_set=0x%x",
6906 		    bep->rsp_code, rsp->fcp_u.fcp_status.rsp_len_set);
6907 		fcp_retry_scsi_cmd(fpkt);
6908 		return;
6909 	}
6910 
6911 	if (rsp->fcp_u.fcp_status.scsi_status == STATUS_QFULL ||
6912 	    rsp->fcp_u.fcp_status.scsi_status == STATUS_BUSY) {
6913 		fcp_queue_ipkt(pptr, fpkt);
6914 		return;
6915 	}
6916 
6917 	/*
6918 	 * Devices that do not support INQUIRY_PAGE83, return check condition
6919 	 * with illegal request as per SCSI spec.
6920 	 * Crossbridge is one such device and Daktari's SES node is another.
6921 	 * We want to ideally enumerate these devices as a non-mpxio devices.
6922 	 * SES nodes (Daktari only currently) are an exception to this.
6923 	 */
6924 	if ((icmd->ipkt_opcode == SCMD_INQUIRY_PAGE83) &&
6925 	    (rsp->fcp_u.fcp_status.scsi_status & STATUS_CHECK)) {
6926 
6927 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
6928 		    fcp_trace, FCP_BUF_LEVEL_3, 0,
6929 		    "INQUIRY_PAGE83 for d_id %x (dtype:0x%x) failed with "
6930 		    "check condition. May enumerate as non-mpxio device",
6931 		    ptgt->tgt_d_id, plun->lun_type);
6932 
6933 		/*
6934 		 * If we let Daktari's SES be enumerated as a non-mpxio
6935 		 * device, there will be a discrepency in that the other
6936 		 * internal FC disks will get enumerated as mpxio devices.
6937 		 * Applications like luxadm expect this to be consistent.
6938 		 *
6939 		 * So, we put in a hack here to check if this is an SES device
6940 		 * and handle it here.
6941 		 */
6942 		if (plun->lun_type == DTYPE_ESI) {
6943 			/*
6944 			 * Since, pkt_state is actually FC_PKT_SUCCESS
6945 			 * at this stage, we fake a failure here so that
6946 			 * fcp_handle_page83 will create a device path using
6947 			 * the WWN instead of the GUID which is not there anyway
6948 			 */
6949 			fpkt->pkt_state = FC_PKT_LOCAL_RJT;
6950 			(void) fcp_handle_page83(fpkt, icmd, 1);
6951 			return;
6952 		}
6953 
6954 		mutex_enter(&ptgt->tgt_mutex);
6955 		plun->lun_state &= ~(FCP_LUN_OFFLINE |
6956 		    FCP_LUN_MARK | FCP_LUN_BUSY);
6957 		mutex_exit(&ptgt->tgt_mutex);
6958 
6959 		(void) fcp_call_finish_init(pptr, ptgt,
6960 		    icmd->ipkt_link_cnt, icmd->ipkt_change_cnt,
6961 		    icmd->ipkt_cause);
6962 		fcp_icmd_free(pptr, icmd);
6963 		return;
6964 	}
6965 
6966 	if (rsp->fcp_u.fcp_status.scsi_status != STATUS_GOOD) {
6967 		int rval = DDI_FAILURE;
6968 
6969 		/*
6970 		 * handle cases where report lun isn't supported
6971 		 * by faking up our own REPORT_LUN response or
6972 		 * UNIT ATTENTION
6973 		 */
6974 		if (icmd->ipkt_opcode == SCMD_REPORT_LUN) {
6975 			rval = fcp_check_reportlun(rsp, fpkt);
6976 
6977 			/*
6978 			 * fcp_check_reportlun might have modified the
6979 			 * FCP response. Copy it in again to get an updated
6980 			 * FCP response
6981 			 */
6982 			if (rval == DDI_SUCCESS && icmd->ipkt_nodma == 0) {
6983 				rsp = &response;
6984 
6985 				FCP_CP_IN(fpkt->pkt_resp, rsp,
6986 				    fpkt->pkt_resp_acc,
6987 				    sizeof (struct fcp_rsp));
6988 			}
6989 		}
6990 
6991 		if (rsp->fcp_u.fcp_status.scsi_status != STATUS_GOOD) {
6992 			if (rval == DDI_SUCCESS) {
6993 				(void) fcp_call_finish_init(pptr, ptgt,
6994 				    icmd->ipkt_link_cnt, icmd->ipkt_change_cnt,
6995 				    icmd->ipkt_cause);
6996 				fcp_icmd_free(pptr, icmd);
6997 			} else {
6998 				fcp_retry_scsi_cmd(fpkt);
6999 			}
7000 
7001 			return;
7002 		}
7003 	} else {
7004 		if (icmd->ipkt_opcode == SCMD_REPORT_LUN) {
7005 			mutex_enter(&ptgt->tgt_mutex);
7006 			ptgt->tgt_state &= ~FCP_TGT_ILLREQ;
7007 			mutex_exit(&ptgt->tgt_mutex);
7008 		}
7009 	}
7010 
7011 	ASSERT(rsp->fcp_u.fcp_status.scsi_status == STATUS_GOOD);
7012 
7013 	(void) ddi_dma_sync(fpkt->pkt_data_dma, 0, 0, DDI_DMA_SYNC_FORCPU);
7014 
7015 	switch (icmd->ipkt_opcode) {
7016 	case SCMD_INQUIRY:
7017 		FCP_LUN_TRACE(plun, FCP_LUN_TRACE_1);
7018 		fcp_handle_inquiry(fpkt, icmd);
7019 		break;
7020 
7021 	case SCMD_REPORT_LUN:
7022 		FCP_TGT_TRACE(ptgt, icmd->ipkt_change_cnt,
7023 		    FCP_TGT_TRACE_22);
7024 		fcp_handle_reportlun(fpkt, icmd);
7025 		break;
7026 
7027 	case SCMD_INQUIRY_PAGE83:
7028 		FCP_LUN_TRACE(plun, FCP_LUN_TRACE_2);
7029 		(void) fcp_handle_page83(fpkt, icmd, 0);
7030 		break;
7031 
7032 	default:
7033 		fcp_log(CE_WARN, NULL, "!Invalid SCSI opcode");
7034 		(void) fcp_call_finish_init(pptr, ptgt, icmd->ipkt_link_cnt,
7035 		    icmd->ipkt_change_cnt, icmd->ipkt_cause);
7036 		fcp_icmd_free(pptr, icmd);
7037 		break;
7038 	}
7039 }
7040 
7041 
7042 static void
7043 fcp_retry_scsi_cmd(fc_packet_t *fpkt)
7044 {
7045 	struct fcp_ipkt	*icmd = (struct fcp_ipkt *)
7046 	    fpkt->pkt_ulp_private;
7047 	struct fcp_tgt	*ptgt = icmd->ipkt_tgt;
7048 	struct fcp_port	*pptr = ptgt->tgt_port;
7049 
7050 	if (icmd->ipkt_retries < FCP_MAX_RETRIES &&
7051 	    fcp_is_retryable(icmd)) {
7052 		mutex_enter(&pptr->port_mutex);
7053 		if (!FCP_TGT_STATE_CHANGED(ptgt, icmd)) {
7054 			mutex_exit(&pptr->port_mutex);
7055 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
7056 			    fcp_trace, FCP_BUF_LEVEL_3, 0,
7057 			    "Retrying %s to %x; state=%x, reason=%x",
7058 			    (icmd->ipkt_opcode == SCMD_REPORT_LUN) ?
7059 			    "Report LUN" : "INQUIRY", ptgt->tgt_d_id,
7060 			    fpkt->pkt_state, fpkt->pkt_reason);
7061 
7062 			fcp_queue_ipkt(pptr, fpkt);
7063 		} else {
7064 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
7065 			    fcp_trace, FCP_BUF_LEVEL_3, 0,
7066 			    "fcp_retry_scsi_cmd,1: state change occured"
7067 			    " for D_ID=0x%x", ptgt->tgt_d_id);
7068 			mutex_exit(&pptr->port_mutex);
7069 			(void) fcp_call_finish_init(pptr, ptgt,
7070 			    icmd->ipkt_link_cnt, icmd->ipkt_change_cnt,
7071 			    icmd->ipkt_cause);
7072 			fcp_icmd_free(pptr, icmd);
7073 		}
7074 	} else {
7075 		fcp_print_error(fpkt);
7076 		(void) fcp_call_finish_init(pptr, ptgt, icmd->ipkt_link_cnt,
7077 		    icmd->ipkt_change_cnt, icmd->ipkt_cause);
7078 		fcp_icmd_free(pptr, icmd);
7079 	}
7080 }
7081 
7082 /*
7083  *     Function: fcp_handle_page83
7084  *
7085  *  Description: Treats the response to INQUIRY_PAGE83.
7086  *
7087  *     Argument: *fpkt	FC packet used to convey the command.
7088  *		 *icmd	Original fcp_ipkt structure.
7089  *		 ignore_page83_data
7090  *			if it's 1, that means it's a special devices's
7091  *			page83 response, it should be enumerated under mpxio
7092  *
7093  * Return Value: None
7094  */
7095 static void
7096 fcp_handle_page83(fc_packet_t *fpkt, struct fcp_ipkt *icmd,
7097     int ignore_page83_data)
7098 {
7099 	struct fcp_port	*pptr;
7100 	struct fcp_lun	*plun;
7101 	struct fcp_tgt	*ptgt;
7102 	uchar_t			dev_id_page[SCMD_MAX_INQUIRY_PAGE83_SIZE];
7103 	int			fail = 0;
7104 	ddi_devid_t		devid;
7105 	char			*guid = NULL;
7106 	int			ret;
7107 
7108 	ASSERT(icmd != NULL && fpkt != NULL);
7109 
7110 	pptr = icmd->ipkt_port;
7111 	ptgt = icmd->ipkt_tgt;
7112 	plun = icmd->ipkt_lun;
7113 
7114 	if (fpkt->pkt_state == FC_PKT_SUCCESS) {
7115 		FCP_LUN_TRACE(plun, FCP_LUN_TRACE_7);
7116 
7117 		FCP_CP_IN(fpkt->pkt_data, dev_id_page, fpkt->pkt_data_acc,
7118 		    SCMD_MAX_INQUIRY_PAGE83_SIZE);
7119 
7120 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
7121 		    fcp_trace, FCP_BUF_LEVEL_5, 0,
7122 		    "fcp_handle_page83: port=%d, tgt D_ID=0x%x, "
7123 		    "dtype=0x%x, lun num=%x",
7124 		    pptr->port_instance, ptgt->tgt_d_id,
7125 		    dev_id_page[0], plun->lun_num);
7126 
7127 		ret = ddi_devid_scsi_encode(
7128 		    DEVID_SCSI_ENCODE_VERSION_LATEST,
7129 		    NULL,		/* driver name */
7130 		    (unsigned char *) &plun->lun_inq, /* standard inquiry */
7131 		    sizeof (plun->lun_inq), /* size of standard inquiry */
7132 		    NULL,		/* page 80 data */
7133 		    0,		/* page 80 len */
7134 		    dev_id_page,	/* page 83 data */
7135 		    SCMD_MAX_INQUIRY_PAGE83_SIZE, /* page 83 data len */
7136 		    &devid);
7137 
7138 		if (ret == DDI_SUCCESS) {
7139 
7140 			guid = ddi_devid_to_guid(devid);
7141 
7142 			if (guid) {
7143 				/*
7144 				 * Check our current guid.  If it's non null
7145 				 * and it has changed, we need to copy it into
7146 				 * lun_old_guid since we might still need it.
7147 				 */
7148 				if (plun->lun_guid &&
7149 				    strcmp(guid, plun->lun_guid)) {
7150 					unsigned int len;
7151 
7152 					/*
7153 					 * If the guid of the LUN changes,
7154 					 * reconfiguration should be triggered
7155 					 * to reflect the changes.
7156 					 * i.e. we should offline the LUN with
7157 					 * the old guid, and online the LUN with
7158 					 * the new guid.
7159 					 */
7160 					plun->lun_state |= FCP_LUN_CHANGED;
7161 
7162 					if (plun->lun_old_guid) {
7163 						kmem_free(plun->lun_old_guid,
7164 						    plun->lun_old_guid_size);
7165 					}
7166 
7167 					len = plun->lun_guid_size;
7168 					plun->lun_old_guid_size = len;
7169 
7170 					plun->lun_old_guid = kmem_zalloc(len,
7171 					    KM_NOSLEEP);
7172 
7173 					if (plun->lun_old_guid) {
7174 						/*
7175 						 * The alloc was successful then
7176 						 * let's do the copy.
7177 						 */
7178 						bcopy(plun->lun_guid,
7179 						    plun->lun_old_guid, len);
7180 					} else {
7181 						fail = 1;
7182 						plun->lun_old_guid_size = 0;
7183 					}
7184 				}
7185 				if (!fail) {
7186 					if (fcp_copy_guid_2_lun_block(
7187 					    plun, guid)) {
7188 						fail = 1;
7189 					}
7190 				}
7191 				ddi_devid_free_guid(guid);
7192 
7193 			} else {
7194 				FCP_TRACE(fcp_logq, pptr->port_instbuf,
7195 				    fcp_trace, FCP_BUF_LEVEL_2, 0,
7196 				    "fcp_handle_page83: unable to create "
7197 				    "GUID");
7198 
7199 				/* couldn't create good guid from devid */
7200 				fail = 1;
7201 			}
7202 			ddi_devid_free(devid);
7203 
7204 		} else if (ret == DDI_NOT_WELL_FORMED) {
7205 			/* NULL filled data for page 83 */
7206 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
7207 			    fcp_trace, FCP_BUF_LEVEL_2, 0,
7208 			    "fcp_handle_page83: retry GUID");
7209 
7210 			icmd->ipkt_retries = 0;
7211 			fcp_retry_scsi_cmd(fpkt);
7212 			return;
7213 		} else {
7214 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
7215 			    fcp_trace, FCP_BUF_LEVEL_2, 0,
7216 			    "fcp_handle_page83: bad ddi_devid_scsi_encode %x",
7217 			    ret);
7218 			/*
7219 			 * Since the page83 validation
7220 			 * introduced late, we are being
7221 			 * tolerant to the existing devices
7222 			 * that already found to be working
7223 			 * under mpxio, like A5200's SES device,
7224 			 * its page83 response will not be standard-compliant,
7225 			 * but we still want it to be enumerated under mpxio.
7226 			 */
7227 			if (fcp_symmetric_device_probe(plun) != 0) {
7228 				fail = 1;
7229 			}
7230 		}
7231 
7232 	} else {
7233 		/* bad packet state */
7234 		FCP_LUN_TRACE(plun, FCP_LUN_TRACE_8);
7235 
7236 		/*
7237 		 * For some special devices (A5K SES and Daktari's SES devices),
7238 		 * they should be enumerated under mpxio
7239 		 * or "luxadm dis" will fail
7240 		 */
7241 		if (ignore_page83_data) {
7242 			fail = 0;
7243 		} else {
7244 			fail = 1;
7245 		}
7246 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
7247 		    fcp_trace, FCP_BUF_LEVEL_2, 0,
7248 		    "!Devid page cmd failed. "
7249 		    "fpkt_state: %x fpkt_reason: %x",
7250 		    "ignore_page83: %d",
7251 		    fpkt->pkt_state, fpkt->pkt_reason,
7252 		    ignore_page83_data);
7253 	}
7254 
7255 	mutex_enter(&pptr->port_mutex);
7256 	mutex_enter(&plun->lun_mutex);
7257 	/*
7258 	 * If lun_cip is not NULL, then we needn't update lun_mpxio to avoid
7259 	 * mismatch between lun_cip and lun_mpxio.
7260 	 */
7261 	if (plun->lun_cip == NULL) {
7262 		/*
7263 		 * If we don't have a guid for this lun it's because we were
7264 		 * unable to glean one from the page 83 response.  Set the
7265 		 * control flag to 0 here to make sure that we don't attempt to
7266 		 * enumerate it under mpxio.
7267 		 */
7268 		if (fail || pptr->port_mpxio == 0) {
7269 			plun->lun_mpxio = 0;
7270 		} else {
7271 			plun->lun_mpxio = 1;
7272 		}
7273 	}
7274 	mutex_exit(&plun->lun_mutex);
7275 	mutex_exit(&pptr->port_mutex);
7276 
7277 	mutex_enter(&ptgt->tgt_mutex);
7278 	plun->lun_state &=
7279 	    ~(FCP_LUN_OFFLINE | FCP_LUN_MARK | FCP_LUN_BUSY);
7280 	mutex_exit(&ptgt->tgt_mutex);
7281 
7282 	(void) fcp_call_finish_init(pptr, ptgt, icmd->ipkt_link_cnt,
7283 	    icmd->ipkt_change_cnt, icmd->ipkt_cause);
7284 
7285 	fcp_icmd_free(pptr, icmd);
7286 }
7287 
7288 /*
7289  *     Function: fcp_handle_inquiry
7290  *
7291  *  Description: Called by fcp_scsi_callback to handle the response to an
7292  *		 INQUIRY request.
7293  *
7294  *     Argument: *fpkt	FC packet used to convey the command.
7295  *		 *icmd	Original fcp_ipkt structure.
7296  *
7297  * Return Value: None
7298  */
7299 static void
7300 fcp_handle_inquiry(fc_packet_t *fpkt, struct fcp_ipkt *icmd)
7301 {
7302 	struct fcp_port	*pptr;
7303 	struct fcp_lun	*plun;
7304 	struct fcp_tgt	*ptgt;
7305 	uchar_t		dtype;
7306 	uchar_t		pqual;
7307 	uint32_t	rscn_count = FC_INVALID_RSCN_COUNT;
7308 
7309 	ASSERT(icmd != NULL && fpkt != NULL);
7310 
7311 	pptr = icmd->ipkt_port;
7312 	ptgt = icmd->ipkt_tgt;
7313 	plun = icmd->ipkt_lun;
7314 
7315 	FCP_CP_IN(fpkt->pkt_data, &plun->lun_inq, fpkt->pkt_data_acc,
7316 	    sizeof (struct scsi_inquiry));
7317 
7318 	dtype = plun->lun_inq.inq_dtype & DTYPE_MASK;
7319 	pqual = plun->lun_inq.inq_dtype >> 5;
7320 
7321 	FCP_TRACE(fcp_logq, pptr->port_instbuf,
7322 	    fcp_trace, FCP_BUF_LEVEL_5, 0,
7323 	    "fcp_handle_inquiry: port=%d, tgt D_ID=0x%x, lun=0x%x, "
7324 	    "dtype=0x%x pqual: 0x%x", pptr->port_instance, ptgt->tgt_d_id,
7325 	    plun->lun_num, dtype, pqual);
7326 
7327 	if (pqual != 0) {
7328 		/*
7329 		 * Non-zero peripheral qualifier
7330 		 */
7331 		fcp_log(CE_CONT, pptr->port_dip,
7332 		    "!Target 0x%x lun 0x%x: Nonzero peripheral qualifier: "
7333 		    "Device type=0x%x Peripheral qual=0x%x\n",
7334 		    ptgt->tgt_d_id, plun->lun_num, dtype, pqual);
7335 
7336 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
7337 		    fcp_trace, FCP_BUF_LEVEL_5, 0,
7338 		    "!Target 0x%x lun 0x%x: Nonzero peripheral qualifier: "
7339 		    "Device type=0x%x Peripheral qual=0x%x\n",
7340 		    ptgt->tgt_d_id, plun->lun_num, dtype, pqual);
7341 
7342 		FCP_LUN_TRACE(plun, FCP_LUN_TRACE_3);
7343 
7344 		(void) fcp_call_finish_init(pptr, ptgt, icmd->ipkt_link_cnt,
7345 		    icmd->ipkt_change_cnt, icmd->ipkt_cause);
7346 		fcp_icmd_free(pptr, icmd);
7347 		return;
7348 	}
7349 
7350 	/*
7351 	 * If the device is already initialized, check the dtype
7352 	 * for a change. If it has changed then update the flags
7353 	 * so the create_luns will offline the old device and
7354 	 * create the new device. Refer to bug: 4764752
7355 	 */
7356 	if ((plun->lun_state & FCP_LUN_INIT) && dtype != plun->lun_type) {
7357 		plun->lun_state |= FCP_LUN_CHANGED;
7358 	}
7359 	plun->lun_type = plun->lun_inq.inq_dtype;
7360 
7361 	/*
7362 	 * This code is setting/initializing the throttling in the FCA
7363 	 * driver.
7364 	 */
7365 	mutex_enter(&pptr->port_mutex);
7366 	if (!pptr->port_notify) {
7367 		if (bcmp(plun->lun_inq.inq_pid, pid, strlen(pid)) == 0) {
7368 			uint32_t cmd = 0;
7369 			cmd = ((cmd & 0xFF | FC_NOTIFY_THROTTLE) |
7370 			    ((cmd & 0xFFFFFF00 >> 8) |
7371 			    FCP_SVE_THROTTLE << 8));
7372 			pptr->port_notify = 1;
7373 			mutex_exit(&pptr->port_mutex);
7374 			(void) fc_ulp_port_notify(pptr->port_fp_handle, cmd);
7375 			mutex_enter(&pptr->port_mutex);
7376 		}
7377 	}
7378 
7379 	if (FCP_TGT_STATE_CHANGED(ptgt, icmd)) {
7380 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
7381 		    fcp_trace, FCP_BUF_LEVEL_2, 0,
7382 		    "fcp_handle_inquiry,1:state change occured"
7383 		    " for D_ID=0x%x", ptgt->tgt_d_id);
7384 		mutex_exit(&pptr->port_mutex);
7385 
7386 		FCP_LUN_TRACE(plun, FCP_LUN_TRACE_5);
7387 		(void) fcp_call_finish_init(pptr, ptgt,
7388 		    icmd->ipkt_link_cnt, icmd->ipkt_change_cnt,
7389 		    icmd->ipkt_cause);
7390 		fcp_icmd_free(pptr, icmd);
7391 		return;
7392 	}
7393 	ASSERT((ptgt->tgt_state & FCP_TGT_MARK) == 0);
7394 	mutex_exit(&pptr->port_mutex);
7395 
7396 	/* Retrieve the rscn count (if a valid one exists) */
7397 	if (icmd->ipkt_fpkt->pkt_ulp_rscn_infop != NULL) {
7398 		rscn_count = ((fc_ulp_rscn_info_t *)
7399 		    (icmd->ipkt_fpkt->pkt_ulp_rscn_infop))->ulp_rscn_count;
7400 	} else {
7401 		rscn_count = FC_INVALID_RSCN_COUNT;
7402 	}
7403 
7404 	if (fcp_send_scsi(plun, SCMD_INQUIRY_PAGE83,
7405 	    SCMD_MAX_INQUIRY_PAGE83_SIZE,
7406 	    icmd->ipkt_link_cnt, icmd->ipkt_change_cnt,
7407 	    icmd->ipkt_cause, rscn_count) != DDI_SUCCESS) {
7408 		fcp_log(CE_WARN, NULL, "!failed to send page 83");
7409 		FCP_LUN_TRACE(plun, FCP_LUN_TRACE_6);
7410 		(void) fcp_call_finish_init(pptr, ptgt,
7411 		    icmd->ipkt_link_cnt, icmd->ipkt_change_cnt,
7412 		    icmd->ipkt_cause);
7413 	}
7414 
7415 	/*
7416 	 * Read Inquiry VPD Page 0x83 to uniquely
7417 	 * identify this logical unit.
7418 	 */
7419 	fcp_icmd_free(pptr, icmd);
7420 }
7421 
7422 /*
7423  *     Function: fcp_handle_reportlun
7424  *
7425  *  Description: Called by fcp_scsi_callback to handle the response to a
7426  *		 REPORT_LUN request.
7427  *
7428  *     Argument: *fpkt	FC packet used to convey the command.
7429  *		 *icmd	Original fcp_ipkt structure.
7430  *
7431  * Return Value: None
7432  */
7433 static void
7434 fcp_handle_reportlun(fc_packet_t *fpkt, struct fcp_ipkt *icmd)
7435 {
7436 	int				i;
7437 	int				nluns_claimed;
7438 	int				nluns_bufmax;
7439 	int				len;
7440 	uint16_t			lun_num;
7441 	uint32_t			rscn_count = FC_INVALID_RSCN_COUNT;
7442 	struct fcp_port			*pptr;
7443 	struct fcp_tgt			*ptgt;
7444 	struct fcp_lun			*plun;
7445 	struct fcp_reportlun_resp	*report_lun;
7446 
7447 	pptr = icmd->ipkt_port;
7448 	ptgt = icmd->ipkt_tgt;
7449 	len = fpkt->pkt_datalen;
7450 
7451 	if ((len < FCP_LUN_HEADER) ||
7452 	    ((report_lun = kmem_zalloc(len, KM_NOSLEEP)) == NULL)) {
7453 		(void) fcp_call_finish_init(pptr, ptgt, icmd->ipkt_link_cnt,
7454 		    icmd->ipkt_change_cnt, icmd->ipkt_cause);
7455 		fcp_icmd_free(pptr, icmd);
7456 		return;
7457 	}
7458 
7459 	FCP_CP_IN(fpkt->pkt_data, report_lun, fpkt->pkt_data_acc,
7460 	    fpkt->pkt_datalen);
7461 
7462 	FCP_TRACE(fcp_logq, pptr->port_instbuf,
7463 	    fcp_trace, FCP_BUF_LEVEL_5, 0,
7464 	    "fcp_handle_reportlun: port=%d, tgt D_ID=0x%x",
7465 	    pptr->port_instance, ptgt->tgt_d_id);
7466 
7467 	/*
7468 	 * Get the number of luns (which is supplied as LUNS * 8) the
7469 	 * device claims it has.
7470 	 */
7471 	nluns_claimed = BE_32(report_lun->num_lun) >> 3;
7472 
7473 	/*
7474 	 * Get the maximum number of luns the buffer submitted can hold.
7475 	 */
7476 	nluns_bufmax = (fpkt->pkt_datalen - FCP_LUN_HEADER) / FCP_LUN_SIZE;
7477 
7478 	/*
7479 	 * Due to limitations of certain hardware, we support only 16 bit LUNs
7480 	 */
7481 	if (nluns_claimed > FCP_MAX_LUNS_SUPPORTED) {
7482 		kmem_free(report_lun, len);
7483 
7484 		fcp_log(CE_NOTE, pptr->port_dip, "!Can not support"
7485 		    " 0x%x number of LUNs for target=%x", nluns_claimed,
7486 		    ptgt->tgt_d_id);
7487 
7488 		(void) fcp_call_finish_init(pptr, ptgt, icmd->ipkt_link_cnt,
7489 		    icmd->ipkt_change_cnt, icmd->ipkt_cause);
7490 		fcp_icmd_free(pptr, icmd);
7491 		return;
7492 	}
7493 
7494 	/*
7495 	 * If there are more LUNs than we have allocated memory for,
7496 	 * allocate more space and send down yet another report lun if
7497 	 * the maximum number of attempts hasn't been reached.
7498 	 */
7499 	mutex_enter(&ptgt->tgt_mutex);
7500 
7501 	if ((nluns_claimed > nluns_bufmax) &&
7502 	    (ptgt->tgt_report_lun_cnt < FCP_MAX_REPORTLUNS_ATTEMPTS)) {
7503 
7504 		struct fcp_lun *plun;
7505 
7506 		ptgt->tgt_report_lun_cnt++;
7507 		plun = ptgt->tgt_lun;
7508 		ASSERT(plun != NULL);
7509 		mutex_exit(&ptgt->tgt_mutex);
7510 
7511 		kmem_free(report_lun, len);
7512 
7513 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
7514 		    fcp_trace, FCP_BUF_LEVEL_5, 0,
7515 		    "!Dynamically discovered %d LUNs for D_ID=%x",
7516 		    nluns_claimed, ptgt->tgt_d_id);
7517 
7518 		/* Retrieve the rscn count (if a valid one exists) */
7519 		if (icmd->ipkt_fpkt->pkt_ulp_rscn_infop != NULL) {
7520 			rscn_count = ((fc_ulp_rscn_info_t *)
7521 			    (icmd->ipkt_fpkt->pkt_ulp_rscn_infop))->
7522 			    ulp_rscn_count;
7523 		} else {
7524 			rscn_count = FC_INVALID_RSCN_COUNT;
7525 		}
7526 
7527 		if (fcp_send_scsi(icmd->ipkt_lun, SCMD_REPORT_LUN,
7528 		    FCP_LUN_HEADER + (nluns_claimed * FCP_LUN_SIZE),
7529 		    icmd->ipkt_link_cnt, icmd->ipkt_change_cnt,
7530 		    icmd->ipkt_cause, rscn_count) != DDI_SUCCESS) {
7531 			(void) fcp_call_finish_init(pptr, ptgt,
7532 			    icmd->ipkt_link_cnt, icmd->ipkt_change_cnt,
7533 			    icmd->ipkt_cause);
7534 		}
7535 
7536 		fcp_icmd_free(pptr, icmd);
7537 		return;
7538 	}
7539 
7540 	if (nluns_claimed > nluns_bufmax) {
7541 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
7542 		    fcp_trace, FCP_BUF_LEVEL_5, 0,
7543 		    "Target=%x:%x:%x:%x:%x:%x:%x:%x"
7544 		    "	 Number of LUNs lost=%x",
7545 		    ptgt->tgt_port_wwn.raw_wwn[0],
7546 		    ptgt->tgt_port_wwn.raw_wwn[1],
7547 		    ptgt->tgt_port_wwn.raw_wwn[2],
7548 		    ptgt->tgt_port_wwn.raw_wwn[3],
7549 		    ptgt->tgt_port_wwn.raw_wwn[4],
7550 		    ptgt->tgt_port_wwn.raw_wwn[5],
7551 		    ptgt->tgt_port_wwn.raw_wwn[6],
7552 		    ptgt->tgt_port_wwn.raw_wwn[7],
7553 		    nluns_claimed - nluns_bufmax);
7554 
7555 		nluns_claimed = nluns_bufmax;
7556 	}
7557 	ptgt->tgt_lun_cnt = nluns_claimed;
7558 
7559 	/*
7560 	 * Identify missing LUNs and print warning messages
7561 	 */
7562 	for (plun = ptgt->tgt_lun; plun; plun = plun->lun_next) {
7563 		int offline;
7564 		int exists = 0;
7565 
7566 		offline = (plun->lun_state & FCP_LUN_OFFLINE) ? 1 : 0;
7567 
7568 		for (i = 0; i < nluns_claimed && exists == 0; i++) {
7569 			uchar_t		*lun_string;
7570 
7571 			lun_string = (uchar_t *)&(report_lun->lun_string[i]);
7572 
7573 			switch (lun_string[0] & 0xC0) {
7574 			case FCP_LUN_ADDRESSING:
7575 			case FCP_PD_ADDRESSING:
7576 				lun_num = ((lun_string[0] & 0x3F) << 8) |
7577 				    lun_string[1];
7578 				if (plun->lun_num == lun_num) {
7579 					exists++;
7580 					break;
7581 				}
7582 				break;
7583 
7584 			default:
7585 				break;
7586 			}
7587 		}
7588 
7589 		if (!exists && !offline) {
7590 			mutex_exit(&ptgt->tgt_mutex);
7591 
7592 			mutex_enter(&pptr->port_mutex);
7593 			mutex_enter(&ptgt->tgt_mutex);
7594 			if (!FCP_STATE_CHANGED(pptr, ptgt, icmd)) {
7595 				/*
7596 				 * set disappear flag when device was connected
7597 				 */
7598 				if (!(plun->lun_state &
7599 				    FCP_LUN_DEVICE_NOT_CONNECTED)) {
7600 					plun->lun_state |= FCP_LUN_DISAPPEARED;
7601 				}
7602 				mutex_exit(&ptgt->tgt_mutex);
7603 				mutex_exit(&pptr->port_mutex);
7604 				if (!(plun->lun_state &
7605 				    FCP_LUN_DEVICE_NOT_CONNECTED)) {
7606 					fcp_log(CE_NOTE, pptr->port_dip,
7607 					    "!Lun=%x for target=%x disappeared",
7608 					    plun->lun_num, ptgt->tgt_d_id);
7609 				}
7610 				mutex_enter(&ptgt->tgt_mutex);
7611 			} else {
7612 				FCP_TRACE(fcp_logq, pptr->port_instbuf,
7613 				    fcp_trace, FCP_BUF_LEVEL_5, 0,
7614 				    "fcp_handle_reportlun,1: state change"
7615 				    " occured for D_ID=0x%x", ptgt->tgt_d_id);
7616 				mutex_exit(&ptgt->tgt_mutex);
7617 				mutex_exit(&pptr->port_mutex);
7618 				kmem_free(report_lun, len);
7619 				(void) fcp_call_finish_init(pptr, ptgt,
7620 				    icmd->ipkt_link_cnt, icmd->ipkt_change_cnt,
7621 				    icmd->ipkt_cause);
7622 				fcp_icmd_free(pptr, icmd);
7623 				return;
7624 			}
7625 		} else if (exists) {
7626 			/*
7627 			 * clear FCP_LUN_DEVICE_NOT_CONNECTED when lun 0
7628 			 * actually exists in REPORT_LUN response
7629 			 */
7630 			if (plun->lun_state & FCP_LUN_DEVICE_NOT_CONNECTED) {
7631 				plun->lun_state &=
7632 				    ~FCP_LUN_DEVICE_NOT_CONNECTED;
7633 			}
7634 			if (offline || plun->lun_num == 0) {
7635 				if (plun->lun_state & FCP_LUN_DISAPPEARED)  {
7636 					plun->lun_state &= ~FCP_LUN_DISAPPEARED;
7637 					mutex_exit(&ptgt->tgt_mutex);
7638 					fcp_log(CE_NOTE, pptr->port_dip,
7639 					    "!Lun=%x for target=%x reappeared",
7640 					    plun->lun_num, ptgt->tgt_d_id);
7641 					mutex_enter(&ptgt->tgt_mutex);
7642 				}
7643 			}
7644 		}
7645 	}
7646 
7647 	ptgt->tgt_tmp_cnt = nluns_claimed ? nluns_claimed : 1;
7648 	mutex_exit(&ptgt->tgt_mutex);
7649 
7650 	FCP_TRACE(fcp_logq, pptr->port_instbuf,
7651 	    fcp_trace, FCP_BUF_LEVEL_5, 0,
7652 	    "fcp_handle_reportlun: port=%d, tgt D_ID=0x%x, %d LUN(s)",
7653 	    pptr->port_instance, ptgt->tgt_d_id, nluns_claimed);
7654 
7655 	/* scan each lun */
7656 	for (i = 0; i < nluns_claimed; i++) {
7657 		uchar_t	*lun_string;
7658 
7659 		lun_string = (uchar_t *)&(report_lun->lun_string[i]);
7660 
7661 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
7662 		    fcp_trace, FCP_BUF_LEVEL_5, 0,
7663 		    "handle_reportlun: d_id=%x, LUN ind=%d, LUN=%d,"
7664 		    " addr=0x%x", ptgt->tgt_d_id, i, lun_string[1],
7665 		    lun_string[0]);
7666 
7667 		switch (lun_string[0] & 0xC0) {
7668 		case FCP_LUN_ADDRESSING:
7669 		case FCP_PD_ADDRESSING:
7670 			lun_num = ((lun_string[0] & 0x3F) << 8) | lun_string[1];
7671 
7672 			/* We will skip masked LUNs because of the blacklist. */
7673 			if (fcp_lun_blacklist != NULL) {
7674 				mutex_enter(&ptgt->tgt_mutex);
7675 				if (fcp_should_mask(&ptgt->tgt_port_wwn,
7676 				    lun_num) == TRUE) {
7677 					ptgt->tgt_lun_cnt--;
7678 					mutex_exit(&ptgt->tgt_mutex);
7679 					break;
7680 				}
7681 				mutex_exit(&ptgt->tgt_mutex);
7682 			}
7683 
7684 			/* see if this LUN is already allocated */
7685 			if ((plun = fcp_get_lun(ptgt, lun_num)) == NULL) {
7686 				plun = fcp_alloc_lun(ptgt);
7687 				if (plun == NULL) {
7688 					fcp_log(CE_NOTE, pptr->port_dip,
7689 					    "!Lun allocation failed"
7690 					    " target=%x lun=%x",
7691 					    ptgt->tgt_d_id, lun_num);
7692 					break;
7693 				}
7694 			}
7695 
7696 			mutex_enter(&plun->lun_tgt->tgt_mutex);
7697 			/* convert to LUN */
7698 			plun->lun_addr.ent_addr_0 =
7699 			    BE_16(*(uint16_t *)&(lun_string[0]));
7700 			plun->lun_addr.ent_addr_1 =
7701 			    BE_16(*(uint16_t *)&(lun_string[2]));
7702 			plun->lun_addr.ent_addr_2 =
7703 			    BE_16(*(uint16_t *)&(lun_string[4]));
7704 			plun->lun_addr.ent_addr_3 =
7705 			    BE_16(*(uint16_t *)&(lun_string[6]));
7706 
7707 			plun->lun_num = lun_num;
7708 			plun->lun_state |= FCP_LUN_BUSY | FCP_LUN_MARK;
7709 			plun->lun_state &= ~FCP_LUN_OFFLINE;
7710 			mutex_exit(&plun->lun_tgt->tgt_mutex);
7711 
7712 			/* Retrieve the rscn count (if a valid one exists) */
7713 			if (icmd->ipkt_fpkt->pkt_ulp_rscn_infop != NULL) {
7714 				rscn_count = ((fc_ulp_rscn_info_t *)
7715 				    (icmd->ipkt_fpkt->pkt_ulp_rscn_infop))->
7716 				    ulp_rscn_count;
7717 			} else {
7718 				rscn_count = FC_INVALID_RSCN_COUNT;
7719 			}
7720 
7721 			if (fcp_send_scsi(plun, SCMD_INQUIRY, SUN_INQSIZE,
7722 			    icmd->ipkt_link_cnt, icmd->ipkt_change_cnt,
7723 			    icmd->ipkt_cause, rscn_count) != DDI_SUCCESS) {
7724 				mutex_enter(&pptr->port_mutex);
7725 				mutex_enter(&plun->lun_tgt->tgt_mutex);
7726 				if (!FCP_STATE_CHANGED(pptr, ptgt, icmd)) {
7727 					fcp_log(CE_NOTE, pptr->port_dip,
7728 					    "!failed to send INQUIRY"
7729 					    " target=%x lun=%x",
7730 					    ptgt->tgt_d_id, plun->lun_num);
7731 				} else {
7732 					FCP_TRACE(fcp_logq,
7733 					    pptr->port_instbuf, fcp_trace,
7734 					    FCP_BUF_LEVEL_5, 0,
7735 					    "fcp_handle_reportlun,2: state"
7736 					    " change occured for D_ID=0x%x",
7737 					    ptgt->tgt_d_id);
7738 				}
7739 				mutex_exit(&plun->lun_tgt->tgt_mutex);
7740 				mutex_exit(&pptr->port_mutex);
7741 			} else {
7742 				continue;
7743 			}
7744 			break;
7745 
7746 		case FCP_VOLUME_ADDRESSING:
7747 			/* FALLTHROUGH */
7748 		default:
7749 			fcp_log(CE_WARN, NULL,
7750 			    "!Unsupported LUN Addressing method %x "
7751 			    "in response to REPORT_LUN", lun_string[0]);
7752 			break;
7753 		}
7754 
7755 		/*
7756 		 * each time through this loop we should decrement
7757 		 * the tmp_cnt by one -- since we go through this loop
7758 		 * one time for each LUN, the tmp_cnt should never be <=0
7759 		 */
7760 		(void) fcp_call_finish_init(pptr, ptgt, icmd->ipkt_link_cnt,
7761 		    icmd->ipkt_change_cnt, icmd->ipkt_cause);
7762 	}
7763 
7764 	if (i == 0) {
7765 		fcp_log(CE_WARN, pptr->port_dip,
7766 		    "!FCP: target=%x reported NO Luns", ptgt->tgt_d_id);
7767 		(void) fcp_call_finish_init(pptr, ptgt, icmd->ipkt_link_cnt,
7768 		    icmd->ipkt_change_cnt, icmd->ipkt_cause);
7769 	}
7770 
7771 	kmem_free(report_lun, len);
7772 	fcp_icmd_free(pptr, icmd);
7773 }
7774 
7775 
7776 /*
7777  * called internally to return a LUN given a target and a LUN number
7778  */
7779 static struct fcp_lun *
7780 fcp_get_lun(struct fcp_tgt *ptgt, uint16_t lun_num)
7781 {
7782 	struct fcp_lun	*plun;
7783 
7784 	mutex_enter(&ptgt->tgt_mutex);
7785 	for (plun = ptgt->tgt_lun; plun != NULL; plun = plun->lun_next) {
7786 		if (plun->lun_num == lun_num) {
7787 			mutex_exit(&ptgt->tgt_mutex);
7788 			return (plun);
7789 		}
7790 	}
7791 	mutex_exit(&ptgt->tgt_mutex);
7792 
7793 	return (NULL);
7794 }
7795 
7796 
7797 /*
7798  * handle finishing one target for fcp_finish_init
7799  *
7800  * return true (non-zero) if we want finish_init to continue with the
7801  * next target
7802  *
7803  * called with the port mutex held
7804  */
7805 /*ARGSUSED*/
7806 static int
7807 fcp_finish_tgt(struct fcp_port *pptr, struct fcp_tgt *ptgt,
7808     int link_cnt, int tgt_cnt, int cause)
7809 {
7810 	int	rval = 1;
7811 	ASSERT(pptr != NULL);
7812 	ASSERT(ptgt != NULL);
7813 
7814 	FCP_TRACE(fcp_logq, pptr->port_instbuf,
7815 	    fcp_trace, FCP_BUF_LEVEL_5, 0,
7816 	    "finish_tgt: D_ID/state = 0x%x/0x%x", ptgt->tgt_d_id,
7817 	    ptgt->tgt_state);
7818 
7819 	ASSERT(mutex_owned(&pptr->port_mutex));
7820 
7821 	if ((pptr->port_link_cnt != link_cnt) ||
7822 	    (tgt_cnt && ptgt->tgt_change_cnt != tgt_cnt)) {
7823 		/*
7824 		 * oh oh -- another link reset or target change
7825 		 * must have occurred while we are in here
7826 		 */
7827 		FCP_TGT_TRACE(ptgt, tgt_cnt, FCP_TGT_TRACE_23);
7828 
7829 		return (0);
7830 	} else {
7831 		FCP_TGT_TRACE(ptgt, tgt_cnt, FCP_TGT_TRACE_24);
7832 	}
7833 
7834 	mutex_enter(&ptgt->tgt_mutex);
7835 
7836 	if (!(ptgt->tgt_state & FCP_TGT_OFFLINE)) {
7837 		/*
7838 		 * tgt is not offline -- is it marked (i.e. needs
7839 		 * to be offlined) ??
7840 		 */
7841 		if (ptgt->tgt_state & FCP_TGT_MARK) {
7842 			/*
7843 			 * this target not offline *and*
7844 			 * marked
7845 			 */
7846 			ptgt->tgt_state &= ~FCP_TGT_MARK;
7847 			rval = fcp_offline_target(pptr, ptgt, link_cnt,
7848 			    tgt_cnt, 0, 0);
7849 		} else {
7850 			ptgt->tgt_state &= ~FCP_TGT_BUSY;
7851 
7852 			/* create the LUNs */
7853 			if (ptgt->tgt_node_state != FCP_TGT_NODE_ON_DEMAND) {
7854 				ptgt->tgt_node_state = FCP_TGT_NODE_PRESENT;
7855 				fcp_create_luns(ptgt, link_cnt, tgt_cnt,
7856 				    cause);
7857 				ptgt->tgt_device_created = 1;
7858 			} else {
7859 				fcp_update_tgt_state(ptgt, FCP_RESET,
7860 				    FCP_LUN_BUSY);
7861 			}
7862 		}
7863 	}
7864 
7865 	mutex_exit(&ptgt->tgt_mutex);
7866 
7867 	return (rval);
7868 }
7869 
7870 
7871 /*
7872  * this routine is called to finish port initialization
7873  *
7874  * Each port has a "temp" counter -- when a state change happens (e.g.
7875  * port online), the temp count is set to the number of devices in the map.
7876  * Then, as each device gets "discovered", the temp counter is decremented
7877  * by one.  When this count reaches zero we know that all of the devices
7878  * in the map have been discovered (or an error has occurred), so we can
7879  * then finish initialization -- which is done by this routine (well, this
7880  * and fcp-finish_tgt())
7881  *
7882  * acquires and releases the global mutex
7883  *
7884  * called with the port mutex owned
7885  */
7886 static void
7887 fcp_finish_init(struct fcp_port *pptr)
7888 {
7889 #ifdef	DEBUG
7890 	bzero(pptr->port_finish_stack, sizeof (pptr->port_finish_stack));
7891 	pptr->port_finish_depth = getpcstack(pptr->port_finish_stack,
7892 	    FCP_STACK_DEPTH);
7893 #endif /* DEBUG */
7894 
7895 	ASSERT(mutex_owned(&pptr->port_mutex));
7896 
7897 	FCP_TRACE(fcp_logq, pptr->port_instbuf,
7898 	    fcp_trace, FCP_BUF_LEVEL_2, 0, "finish_init:"
7899 	    " entering; ipkt count=%d", pptr->port_ipkt_cnt);
7900 
7901 	if ((pptr->port_state & FCP_STATE_ONLINING) &&
7902 	    !(pptr->port_state & (FCP_STATE_SUSPENDED |
7903 	    FCP_STATE_DETACHING | FCP_STATE_POWER_DOWN))) {
7904 		pptr->port_state &= ~FCP_STATE_ONLINING;
7905 		pptr->port_state |= FCP_STATE_ONLINE;
7906 	}
7907 
7908 	/* Wake up threads waiting on config done */
7909 	cv_broadcast(&pptr->port_config_cv);
7910 }
7911 
7912 
7913 /*
7914  * called from fcp_finish_init to create the LUNs for a target
7915  *
7916  * called with the port mutex owned
7917  */
7918 static void
7919 fcp_create_luns(struct fcp_tgt *ptgt, int link_cnt, int tgt_cnt, int cause)
7920 {
7921 	struct fcp_lun	*plun;
7922 	struct fcp_port	*pptr;
7923 	child_info_t		*cip = NULL;
7924 
7925 	ASSERT(ptgt != NULL);
7926 	ASSERT(mutex_owned(&ptgt->tgt_mutex));
7927 
7928 	pptr = ptgt->tgt_port;
7929 
7930 	ASSERT(pptr != NULL);
7931 
7932 	/* scan all LUNs for this target */
7933 	for (plun = ptgt->tgt_lun; plun != NULL; plun = plun->lun_next) {
7934 		if (plun->lun_state & FCP_LUN_OFFLINE) {
7935 			continue;
7936 		}
7937 
7938 		if (plun->lun_state & FCP_LUN_MARK) {
7939 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
7940 			    fcp_trace, FCP_BUF_LEVEL_2, 0,
7941 			    "fcp_create_luns: offlining marked LUN!");
7942 			fcp_offline_lun(plun, link_cnt, tgt_cnt, 1, 0);
7943 			continue;
7944 		}
7945 
7946 		plun->lun_state &= ~FCP_LUN_BUSY;
7947 
7948 		/*
7949 		 * There are conditions in which FCP_LUN_INIT flag is cleared
7950 		 * but we have a valid plun->lun_cip. To cover this case also
7951 		 * CLEAR_BUSY whenever we have a valid lun_cip.
7952 		 */
7953 		if (plun->lun_mpxio && plun->lun_cip &&
7954 		    (!fcp_pass_to_hp(pptr, plun, plun->lun_cip,
7955 		    FCP_MPXIO_PATH_CLEAR_BUSY, link_cnt, tgt_cnt,
7956 		    0, 0))) {
7957 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
7958 			    fcp_trace, FCP_BUF_LEVEL_2, 0,
7959 			    "fcp_create_luns: enable lun %p failed!",
7960 			    plun);
7961 		}
7962 
7963 		if (plun->lun_state & FCP_LUN_INIT &&
7964 		    !(plun->lun_state & FCP_LUN_CHANGED)) {
7965 			continue;
7966 		}
7967 
7968 		if (cause == FCP_CAUSE_USER_CREATE) {
7969 			continue;
7970 		}
7971 
7972 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
7973 		    fcp_trace, FCP_BUF_LEVEL_6, 0,
7974 		    "create_luns: passing ONLINE elem to HP thread");
7975 
7976 		/*
7977 		 * If lun has changed, prepare for offlining the old path.
7978 		 * Do not offline the old path right now, since it may be
7979 		 * still opened.
7980 		 */
7981 		if (plun->lun_cip && (plun->lun_state & FCP_LUN_CHANGED)) {
7982 			fcp_prepare_offline_lun(plun, link_cnt, tgt_cnt);
7983 		}
7984 
7985 		/* pass an ONLINE element to the hotplug thread */
7986 		if (!fcp_pass_to_hp(pptr, plun, cip, FCP_ONLINE,
7987 		    link_cnt, tgt_cnt, NDI_ONLINE_ATTACH, 0)) {
7988 
7989 			/*
7990 			 * We can not synchronous attach (i.e pass
7991 			 * NDI_ONLINE_ATTACH) here as we might be
7992 			 * coming from an interrupt or callback
7993 			 * thread.
7994 			 */
7995 			if (!fcp_pass_to_hp(pptr, plun, cip, FCP_ONLINE,
7996 			    link_cnt, tgt_cnt, 0, 0)) {
7997 				fcp_log(CE_CONT, pptr->port_dip,
7998 				    "Can not ONLINE LUN; D_ID=%x, LUN=%x\n",
7999 				    plun->lun_tgt->tgt_d_id, plun->lun_num);
8000 			}
8001 		}
8002 	}
8003 }
8004 
8005 
8006 /*
8007  * function to online/offline devices
8008  */
8009 static int
8010 fcp_trigger_lun(struct fcp_lun *plun, child_info_t *cip, int old_mpxio,
8011     int online, int lcount, int tcount, int flags)
8012 {
8013 	int			rval = NDI_FAILURE;
8014 	int			circ;
8015 	child_info_t		*ccip;
8016 	struct fcp_port		*pptr = plun->lun_tgt->tgt_port;
8017 	int			is_mpxio = pptr->port_mpxio;
8018 	dev_info_t		*cdip, *pdip;
8019 	char			*devname;
8020 
8021 	if ((old_mpxio != 0) && (plun->lun_mpxio != old_mpxio)) {
8022 		/*
8023 		 * When this event gets serviced, lun_cip and lun_mpxio
8024 		 * has changed, so it should be invalidated now.
8025 		 */
8026 		FCP_TRACE(fcp_logq, pptr->port_instbuf, fcp_trace,
8027 		    FCP_BUF_LEVEL_2, 0, "fcp_trigger_lun: lun_mpxio changed: "
8028 		    "plun: %p, cip: %p, what:%d", plun, cip, online);
8029 		return (rval);
8030 	}
8031 
8032 	FCP_TRACE(fcp_logq, pptr->port_instbuf,
8033 	    fcp_trace, FCP_BUF_LEVEL_2, 0,
8034 	    "fcp_trigger_lun: plun=%p target=%x lun=%d cip=%p what=%x "
8035 	    "flags=%x mpxio=%x\n",
8036 	    plun, LUN_TGT->tgt_d_id, plun->lun_num, cip, online, flags,
8037 	    plun->lun_mpxio);
8038 
8039 	/*
8040 	 * lun_mpxio needs checking here because we can end up in a race
8041 	 * condition where this task has been dispatched while lun_mpxio is
8042 	 * set, but an earlier FCP_ONLINE task for the same LUN tried to
8043 	 * enable MPXIO for the LUN, but was unable to, and hence cleared
8044 	 * the flag. We rely on the serialization of the tasks here. We return
8045 	 * NDI_SUCCESS so any callers continue without reporting spurious
8046 	 * errors, and the still think we're an MPXIO LUN.
8047 	 */
8048 
8049 	if (online == FCP_MPXIO_PATH_CLEAR_BUSY ||
8050 	    online == FCP_MPXIO_PATH_SET_BUSY) {
8051 		if (plun->lun_mpxio) {
8052 			rval = fcp_update_mpxio_path(plun, cip, online);
8053 		} else {
8054 			rval = NDI_SUCCESS;
8055 		}
8056 		return (rval);
8057 	}
8058 
8059 	/*
8060 	 * Explicit devfs_clean() due to ndi_devi_offline() not
8061 	 * executing devfs_clean() if parent lock is held.
8062 	 */
8063 	ASSERT(!servicing_interrupt());
8064 	if (online == FCP_OFFLINE) {
8065 		if (plun->lun_mpxio == 0) {
8066 			if (plun->lun_cip == cip) {
8067 				cdip = DIP(plun->lun_cip);
8068 			} else {
8069 				cdip = DIP(cip);
8070 			}
8071 		} else if ((plun->lun_cip == cip) && plun->lun_cip) {
8072 			cdip = mdi_pi_get_client(PIP(plun->lun_cip));
8073 		} else if ((plun->lun_cip != cip) && cip) {
8074 			/*
8075 			 * This means a DTYPE/GUID change, we shall get the
8076 			 * dip of the old cip instead of the current lun_cip.
8077 			 */
8078 			cdip = mdi_pi_get_client(PIP(cip));
8079 		}
8080 		if (cdip) {
8081 			if (i_ddi_devi_attached(cdip)) {
8082 				pdip = ddi_get_parent(cdip);
8083 				devname = kmem_alloc(MAXNAMELEN + 1, KM_SLEEP);
8084 				ndi_devi_enter(pdip, &circ);
8085 				(void) ddi_deviname(cdip, devname);
8086 				ndi_devi_exit(pdip, circ);
8087 				/*
8088 				 * Release parent lock before calling
8089 				 * devfs_clean().
8090 				 */
8091 				rval = devfs_clean(pdip, devname + 1,
8092 				    DV_CLEAN_FORCE);
8093 				kmem_free(devname, MAXNAMELEN + 1);
8094 				/*
8095 				 * Return if devfs_clean() fails for
8096 				 * non-MPXIO case.
8097 				 * For MPXIO case, another path could be
8098 				 * offlined.
8099 				 */
8100 				if (rval && plun->lun_mpxio == 0) {
8101 					FCP_TRACE(fcp_logq, pptr->port_instbuf,
8102 					    fcp_trace, FCP_BUF_LEVEL_3, 0,
8103 					    "fcp_trigger_lun: devfs_clean "
8104 					    "failed rval=%x  dip=%p",
8105 					    rval, pdip);
8106 					return (NDI_FAILURE);
8107 				}
8108 			}
8109 		}
8110 	}
8111 
8112 	if (fc_ulp_busy_port(pptr->port_fp_handle) != 0) {
8113 		return (NDI_FAILURE);
8114 	}
8115 
8116 	if (is_mpxio) {
8117 		mdi_devi_enter(pptr->port_dip, &circ);
8118 	} else {
8119 		ndi_devi_enter(pptr->port_dip, &circ);
8120 	}
8121 
8122 	mutex_enter(&pptr->port_mutex);
8123 	mutex_enter(&plun->lun_mutex);
8124 
8125 	if (online == FCP_ONLINE) {
8126 		ccip = fcp_get_cip(plun, cip, lcount, tcount);
8127 		if (ccip == NULL) {
8128 			goto fail;
8129 		}
8130 	} else {
8131 		if (fcp_is_child_present(plun, cip) != FC_SUCCESS) {
8132 			goto fail;
8133 		}
8134 		ccip = cip;
8135 	}
8136 
8137 	if (online == FCP_ONLINE) {
8138 		rval = fcp_online_child(plun, ccip, lcount, tcount, flags,
8139 		    &circ);
8140 		fc_ulp_log_device_event(pptr->port_fp_handle,
8141 		    FC_ULP_DEVICE_ONLINE);
8142 	} else {
8143 		rval = fcp_offline_child(plun, ccip, lcount, tcount, flags,
8144 		    &circ);
8145 		fc_ulp_log_device_event(pptr->port_fp_handle,
8146 		    FC_ULP_DEVICE_OFFLINE);
8147 	}
8148 
8149 fail:	mutex_exit(&plun->lun_mutex);
8150 	mutex_exit(&pptr->port_mutex);
8151 
8152 	if (is_mpxio) {
8153 		mdi_devi_exit(pptr->port_dip, circ);
8154 	} else {
8155 		ndi_devi_exit(pptr->port_dip, circ);
8156 	}
8157 
8158 	fc_ulp_idle_port(pptr->port_fp_handle);
8159 
8160 	return (rval);
8161 }
8162 
8163 
8164 /*
8165  * take a target offline by taking all of its LUNs offline
8166  */
8167 /*ARGSUSED*/
8168 static int
8169 fcp_offline_target(struct fcp_port *pptr, struct fcp_tgt *ptgt,
8170     int link_cnt, int tgt_cnt, int nowait, int flags)
8171 {
8172 	struct fcp_tgt_elem	*elem;
8173 
8174 	ASSERT(mutex_owned(&pptr->port_mutex));
8175 	ASSERT(mutex_owned(&ptgt->tgt_mutex));
8176 
8177 	ASSERT(!(ptgt->tgt_state & FCP_TGT_OFFLINE));
8178 
8179 	if (link_cnt != pptr->port_link_cnt || (tgt_cnt && tgt_cnt !=
8180 	    ptgt->tgt_change_cnt)) {
8181 		mutex_exit(&ptgt->tgt_mutex);
8182 		FCP_TGT_TRACE(ptgt, tgt_cnt, FCP_TGT_TRACE_25);
8183 		mutex_enter(&ptgt->tgt_mutex);
8184 
8185 		return (0);
8186 	}
8187 
8188 	ptgt->tgt_pd_handle = NULL;
8189 	mutex_exit(&ptgt->tgt_mutex);
8190 	FCP_TGT_TRACE(ptgt, tgt_cnt, FCP_TGT_TRACE_26);
8191 	mutex_enter(&ptgt->tgt_mutex);
8192 
8193 	tgt_cnt = tgt_cnt ? tgt_cnt : ptgt->tgt_change_cnt;
8194 
8195 	if (ptgt->tgt_tcap &&
8196 	    (elem = kmem_zalloc(sizeof (*elem), KM_NOSLEEP)) != NULL) {
8197 		elem->flags = flags;
8198 		elem->time = fcp_watchdog_time;
8199 		if (nowait == 0) {
8200 			elem->time += fcp_offline_delay;
8201 		}
8202 		elem->ptgt = ptgt;
8203 		elem->link_cnt = link_cnt;
8204 		elem->tgt_cnt = tgt_cnt;
8205 		elem->next = pptr->port_offline_tgts;
8206 		pptr->port_offline_tgts = elem;
8207 	} else {
8208 		fcp_offline_target_now(pptr, ptgt, link_cnt, tgt_cnt, flags);
8209 	}
8210 
8211 	return (1);
8212 }
8213 
8214 
8215 static void
8216 fcp_offline_target_now(struct fcp_port *pptr, struct fcp_tgt *ptgt,
8217     int link_cnt, int tgt_cnt, int flags)
8218 {
8219 	ASSERT(mutex_owned(&pptr->port_mutex));
8220 	ASSERT(mutex_owned(&ptgt->tgt_mutex));
8221 
8222 	fc_ulp_enable_relogin(pptr->port_fp_handle, &ptgt->tgt_port_wwn);
8223 	ptgt->tgt_state = FCP_TGT_OFFLINE;
8224 	ptgt->tgt_pd_handle = NULL;
8225 	fcp_offline_tgt_luns(ptgt, link_cnt, tgt_cnt, flags);
8226 }
8227 
8228 
8229 static void
8230 fcp_offline_tgt_luns(struct fcp_tgt *ptgt, int link_cnt, int tgt_cnt,
8231     int flags)
8232 {
8233 	struct	fcp_lun	*plun;
8234 
8235 	ASSERT(mutex_owned(&ptgt->tgt_port->port_mutex));
8236 	ASSERT(mutex_owned(&ptgt->tgt_mutex));
8237 
8238 	for (plun = ptgt->tgt_lun; plun != NULL; plun = plun->lun_next) {
8239 		if (!(plun->lun_state & FCP_LUN_OFFLINE)) {
8240 			fcp_offline_lun(plun, link_cnt, tgt_cnt, 1, flags);
8241 		}
8242 	}
8243 }
8244 
8245 
8246 /*
8247  * take a LUN offline
8248  *
8249  * enters and leaves with the target mutex held, releasing it in the process
8250  *
8251  * allocates memory in non-sleep mode
8252  */
8253 static void
8254 fcp_offline_lun(struct fcp_lun *plun, int link_cnt, int tgt_cnt,
8255     int nowait, int flags)
8256 {
8257 	struct fcp_port	*pptr = plun->lun_tgt->tgt_port;
8258 	struct fcp_lun_elem	*elem;
8259 
8260 	ASSERT(plun != NULL);
8261 	ASSERT(mutex_owned(&LUN_TGT->tgt_mutex));
8262 
8263 	if (nowait) {
8264 		fcp_offline_lun_now(plun, link_cnt, tgt_cnt, flags);
8265 		return;
8266 	}
8267 
8268 	if ((elem = kmem_zalloc(sizeof (*elem), KM_NOSLEEP)) != NULL) {
8269 		elem->flags = flags;
8270 		elem->time = fcp_watchdog_time;
8271 		if (nowait == 0) {
8272 			elem->time += fcp_offline_delay;
8273 		}
8274 		elem->plun = plun;
8275 		elem->link_cnt = link_cnt;
8276 		elem->tgt_cnt = plun->lun_tgt->tgt_change_cnt;
8277 		elem->next = pptr->port_offline_luns;
8278 		pptr->port_offline_luns = elem;
8279 	} else {
8280 		fcp_offline_lun_now(plun, link_cnt, tgt_cnt, flags);
8281 	}
8282 }
8283 
8284 
8285 static void
8286 fcp_prepare_offline_lun(struct fcp_lun *plun, int link_cnt, int tgt_cnt)
8287 {
8288 	struct fcp_pkt	*head = NULL;
8289 
8290 	ASSERT(mutex_owned(&LUN_TGT->tgt_mutex));
8291 
8292 	mutex_exit(&LUN_TGT->tgt_mutex);
8293 
8294 	head = fcp_scan_commands(plun);
8295 	if (head != NULL) {
8296 		fcp_abort_commands(head, LUN_PORT);
8297 	}
8298 
8299 	mutex_enter(&LUN_TGT->tgt_mutex);
8300 
8301 	if (plun->lun_cip && plun->lun_mpxio) {
8302 		/*
8303 		 * Intimate MPxIO lun busy is cleared
8304 		 */
8305 		if (!fcp_pass_to_hp(LUN_PORT, plun, plun->lun_cip,
8306 		    FCP_MPXIO_PATH_CLEAR_BUSY, link_cnt, tgt_cnt,
8307 		    0, 0)) {
8308 			fcp_log(CE_NOTE, LUN_PORT->port_dip,
8309 			    "Can not ENABLE LUN; D_ID=%x, LUN=%x",
8310 			    LUN_TGT->tgt_d_id, plun->lun_num);
8311 		}
8312 		/*
8313 		 * Intimate MPxIO that the lun is now marked for offline
8314 		 */
8315 		mutex_exit(&LUN_TGT->tgt_mutex);
8316 		(void) mdi_pi_disable_path(PIP(plun->lun_cip), DRIVER_DISABLE);
8317 		mutex_enter(&LUN_TGT->tgt_mutex);
8318 	}
8319 }
8320 
8321 static void
8322 fcp_offline_lun_now(struct fcp_lun *plun, int link_cnt, int tgt_cnt,
8323     int flags)
8324 {
8325 	ASSERT(mutex_owned(&LUN_TGT->tgt_mutex));
8326 
8327 	mutex_exit(&LUN_TGT->tgt_mutex);
8328 	fcp_update_offline_flags(plun);
8329 	mutex_enter(&LUN_TGT->tgt_mutex);
8330 
8331 	fcp_prepare_offline_lun(plun, link_cnt, tgt_cnt);
8332 
8333 	FCP_TRACE(fcp_logq, LUN_PORT->port_instbuf,
8334 	    fcp_trace, FCP_BUF_LEVEL_4, 0,
8335 	    "offline_lun: passing OFFLINE elem to HP thread");
8336 
8337 	if (plun->lun_cip) {
8338 		fcp_log(CE_NOTE, LUN_PORT->port_dip,
8339 		    "!offlining lun=%x (trace=%x), target=%x (trace=%x)",
8340 		    plun->lun_num, plun->lun_trace, LUN_TGT->tgt_d_id,
8341 		    LUN_TGT->tgt_trace);
8342 
8343 		if (!fcp_pass_to_hp(LUN_PORT, plun, plun->lun_cip, FCP_OFFLINE,
8344 		    link_cnt, tgt_cnt, flags, 0)) {
8345 			fcp_log(CE_CONT, LUN_PORT->port_dip,
8346 			    "Can not OFFLINE LUN; D_ID=%x, LUN=%x\n",
8347 			    LUN_TGT->tgt_d_id, plun->lun_num);
8348 		}
8349 	}
8350 }
8351 
8352 static void
8353 fcp_scan_offline_luns(struct fcp_port *pptr)
8354 {
8355 	struct fcp_lun_elem	*elem;
8356 	struct fcp_lun_elem	*prev;
8357 	struct fcp_lun_elem	*next;
8358 
8359 	ASSERT(MUTEX_HELD(&pptr->port_mutex));
8360 
8361 	prev = NULL;
8362 	elem = pptr->port_offline_luns;
8363 	while (elem) {
8364 		next = elem->next;
8365 		if (elem->time <= fcp_watchdog_time) {
8366 			int			changed = 1;
8367 			struct fcp_tgt	*ptgt = elem->plun->lun_tgt;
8368 
8369 			mutex_enter(&ptgt->tgt_mutex);
8370 			if (pptr->port_link_cnt == elem->link_cnt &&
8371 			    ptgt->tgt_change_cnt == elem->tgt_cnt) {
8372 				changed = 0;
8373 			}
8374 
8375 			if (!changed &&
8376 			    !(elem->plun->lun_state & FCP_TGT_OFFLINE)) {
8377 				fcp_offline_lun_now(elem->plun,
8378 				    elem->link_cnt, elem->tgt_cnt, elem->flags);
8379 			}
8380 			mutex_exit(&ptgt->tgt_mutex);
8381 
8382 			kmem_free(elem, sizeof (*elem));
8383 
8384 			if (prev) {
8385 				prev->next = next;
8386 			} else {
8387 				pptr->port_offline_luns = next;
8388 			}
8389 		} else {
8390 			prev = elem;
8391 		}
8392 		elem = next;
8393 	}
8394 }
8395 
8396 
8397 static void
8398 fcp_scan_offline_tgts(struct fcp_port *pptr)
8399 {
8400 	struct fcp_tgt_elem	*elem;
8401 	struct fcp_tgt_elem	*prev;
8402 	struct fcp_tgt_elem	*next;
8403 
8404 	ASSERT(MUTEX_HELD(&pptr->port_mutex));
8405 
8406 	prev = NULL;
8407 	elem = pptr->port_offline_tgts;
8408 	while (elem) {
8409 		next = elem->next;
8410 		if (elem->time <= fcp_watchdog_time) {
8411 			int			changed = 1;
8412 			struct fcp_tgt	*ptgt = elem->ptgt;
8413 
8414 			if (ptgt->tgt_change_cnt == elem->tgt_cnt) {
8415 				changed = 0;
8416 			}
8417 
8418 			mutex_enter(&ptgt->tgt_mutex);
8419 			if (!changed && !(ptgt->tgt_state &
8420 			    FCP_TGT_OFFLINE)) {
8421 				fcp_offline_target_now(pptr,
8422 				    ptgt, elem->link_cnt, elem->tgt_cnt,
8423 				    elem->flags);
8424 			}
8425 			mutex_exit(&ptgt->tgt_mutex);
8426 
8427 			kmem_free(elem, sizeof (*elem));
8428 
8429 			if (prev) {
8430 				prev->next = next;
8431 			} else {
8432 				pptr->port_offline_tgts = next;
8433 			}
8434 		} else {
8435 			prev = elem;
8436 		}
8437 		elem = next;
8438 	}
8439 }
8440 
8441 
8442 static void
8443 fcp_update_offline_flags(struct fcp_lun *plun)
8444 {
8445 	struct fcp_port	*pptr = LUN_PORT;
8446 	ASSERT(plun != NULL);
8447 
8448 	mutex_enter(&LUN_TGT->tgt_mutex);
8449 	plun->lun_state |= FCP_LUN_OFFLINE;
8450 	plun->lun_state &= ~(FCP_LUN_INIT | FCP_LUN_BUSY | FCP_LUN_MARK);
8451 
8452 	mutex_enter(&plun->lun_mutex);
8453 	if (plun->lun_cip && plun->lun_state & FCP_SCSI_LUN_TGT_INIT) {
8454 		dev_info_t *cdip = NULL;
8455 
8456 		mutex_exit(&LUN_TGT->tgt_mutex);
8457 
8458 		if (plun->lun_mpxio == 0) {
8459 			cdip = DIP(plun->lun_cip);
8460 		} else if (plun->lun_cip) {
8461 			cdip = mdi_pi_get_client(PIP(plun->lun_cip));
8462 		}
8463 
8464 		mutex_exit(&plun->lun_mutex);
8465 		if (cdip) {
8466 			(void) ndi_event_retrieve_cookie(
8467 			    pptr->port_ndi_event_hdl, cdip, FCAL_REMOVE_EVENT,
8468 			    &fcp_remove_eid, NDI_EVENT_NOPASS);
8469 			(void) ndi_event_run_callbacks(
8470 			    pptr->port_ndi_event_hdl, cdip,
8471 			    fcp_remove_eid, NULL);
8472 		}
8473 	} else {
8474 		mutex_exit(&plun->lun_mutex);
8475 		mutex_exit(&LUN_TGT->tgt_mutex);
8476 	}
8477 }
8478 
8479 
8480 /*
8481  * Scan all of the command pkts for this port, moving pkts that
8482  * match our LUN onto our own list (headed by "head")
8483  */
8484 static struct fcp_pkt *
8485 fcp_scan_commands(struct fcp_lun *plun)
8486 {
8487 	struct fcp_port	*pptr = LUN_PORT;
8488 
8489 	struct fcp_pkt	*cmd = NULL;	/* pkt cmd ptr */
8490 	struct fcp_pkt	*ncmd = NULL;	/* next pkt ptr */
8491 	struct fcp_pkt	*pcmd = NULL;	/* the previous command */
8492 
8493 	struct fcp_pkt	*head = NULL;	/* head of our list */
8494 	struct fcp_pkt	*tail = NULL;	/* tail of our list */
8495 
8496 	int			cmds_found = 0;
8497 
8498 	mutex_enter(&pptr->port_pkt_mutex);
8499 	for (cmd = pptr->port_pkt_head; cmd != NULL; cmd = ncmd) {
8500 		struct fcp_lun *tlun =
8501 		    ADDR2LUN(&cmd->cmd_pkt->pkt_address);
8502 
8503 		ncmd = cmd->cmd_next;	/* set next command */
8504 
8505 		/*
8506 		 * if this pkt is for a different LUN  or the
8507 		 * command is sent down, skip it.
8508 		 */
8509 		if (tlun != plun || cmd->cmd_state == FCP_PKT_ISSUED ||
8510 		    (cmd->cmd_pkt->pkt_flags & FLAG_NOINTR)) {
8511 			pcmd = cmd;
8512 			continue;
8513 		}
8514 		cmds_found++;
8515 		if (pcmd != NULL) {
8516 			ASSERT(pptr->port_pkt_head != cmd);
8517 			pcmd->cmd_next = cmd->cmd_next;
8518 		} else {
8519 			ASSERT(cmd == pptr->port_pkt_head);
8520 			pptr->port_pkt_head = cmd->cmd_next;
8521 		}
8522 
8523 		if (cmd == pptr->port_pkt_tail) {
8524 			pptr->port_pkt_tail = pcmd;
8525 			if (pcmd) {
8526 				pcmd->cmd_next = NULL;
8527 			}
8528 		}
8529 
8530 		if (head == NULL) {
8531 			head = tail = cmd;
8532 		} else {
8533 			ASSERT(tail != NULL);
8534 
8535 			tail->cmd_next = cmd;
8536 			tail = cmd;
8537 		}
8538 		cmd->cmd_next = NULL;
8539 	}
8540 	mutex_exit(&pptr->port_pkt_mutex);
8541 
8542 	FCP_DTRACE(fcp_logq, pptr->port_instbuf,
8543 	    fcp_trace, FCP_BUF_LEVEL_8, 0,
8544 	    "scan commands: %d cmd(s) found", cmds_found);
8545 
8546 	return (head);
8547 }
8548 
8549 
8550 /*
8551  * Abort all the commands in the command queue
8552  */
8553 static void
8554 fcp_abort_commands(struct fcp_pkt *head, struct fcp_port *pptr)
8555 {
8556 	struct fcp_pkt	*cmd = NULL;	/* pkt cmd ptr */
8557 	struct	fcp_pkt	*ncmd = NULL;	/* next pkt ptr */
8558 
8559 	ASSERT(mutex_owned(&pptr->port_mutex));
8560 
8561 	/* scan through the pkts and invalid them */
8562 	for (cmd = head; cmd != NULL; cmd = ncmd) {
8563 		struct scsi_pkt *pkt = cmd->cmd_pkt;
8564 
8565 		ncmd = cmd->cmd_next;
8566 		ASSERT(pkt != NULL);
8567 
8568 		/*
8569 		 * The lun is going to be marked offline. Indicate
8570 		 * the target driver not to requeue or retry this command
8571 		 * as the device is going to be offlined pretty soon.
8572 		 */
8573 		pkt->pkt_reason = CMD_DEV_GONE;
8574 		pkt->pkt_statistics = 0;
8575 		pkt->pkt_state = 0;
8576 
8577 		/* reset cmd flags/state */
8578 		cmd->cmd_flags &= ~CFLAG_IN_QUEUE;
8579 		cmd->cmd_state = FCP_PKT_IDLE;
8580 
8581 		/*
8582 		 * ensure we have a packet completion routine,
8583 		 * then call it.
8584 		 */
8585 		ASSERT(pkt->pkt_comp != NULL);
8586 
8587 		mutex_exit(&pptr->port_mutex);
8588 		fcp_post_callback(cmd);
8589 		mutex_enter(&pptr->port_mutex);
8590 	}
8591 }
8592 
8593 
8594 /*
8595  * the pkt_comp callback for command packets
8596  */
8597 static void
8598 fcp_cmd_callback(fc_packet_t *fpkt)
8599 {
8600 	struct fcp_pkt *cmd = (struct fcp_pkt *)fpkt->pkt_ulp_private;
8601 	struct scsi_pkt *pkt = cmd->cmd_pkt;
8602 	struct fcp_port *pptr = ADDR2FCP(&pkt->pkt_address);
8603 
8604 	ASSERT(cmd->cmd_state != FCP_PKT_IDLE);
8605 
8606 	if (cmd->cmd_state == FCP_PKT_IDLE) {
8607 		cmn_err(CE_PANIC, "Packet already completed %p",
8608 		    (void *)cmd);
8609 	}
8610 
8611 	/*
8612 	 * Watch thread should be freeing the packet, ignore the pkt.
8613 	 */
8614 	if (cmd->cmd_state == FCP_PKT_ABORTING) {
8615 		fcp_log(CE_CONT, pptr->port_dip,
8616 		    "!FCP: Pkt completed while aborting\n");
8617 		return;
8618 	}
8619 	cmd->cmd_state = FCP_PKT_IDLE;
8620 
8621 	fcp_complete_pkt(fpkt);
8622 
8623 #ifdef	DEBUG
8624 	mutex_enter(&pptr->port_pkt_mutex);
8625 	pptr->port_npkts--;
8626 	mutex_exit(&pptr->port_pkt_mutex);
8627 #endif /* DEBUG */
8628 
8629 	fcp_post_callback(cmd);
8630 }
8631 
8632 
8633 static void
8634 fcp_complete_pkt(fc_packet_t *fpkt)
8635 {
8636 	int			error = 0;
8637 	struct fcp_pkt	*cmd = (struct fcp_pkt *)
8638 	    fpkt->pkt_ulp_private;
8639 	struct scsi_pkt		*pkt = cmd->cmd_pkt;
8640 	struct fcp_port		*pptr = ADDR2FCP(&pkt->pkt_address);
8641 	struct fcp_lun	*plun;
8642 	struct fcp_tgt	*ptgt;
8643 	struct fcp_rsp		*rsp;
8644 	struct scsi_address	save;
8645 
8646 #ifdef	DEBUG
8647 	save = pkt->pkt_address;
8648 #endif /* DEBUG */
8649 
8650 	rsp = (struct fcp_rsp *)cmd->cmd_fcp_rsp;
8651 
8652 	if (fpkt->pkt_state == FC_PKT_SUCCESS) {
8653 		if (pptr->port_fcp_dma != FC_NO_DVMA_SPACE) {
8654 			FCP_CP_IN(fpkt->pkt_resp, rsp, fpkt->pkt_resp_acc,
8655 			    sizeof (struct fcp_rsp));
8656 		}
8657 
8658 		pkt->pkt_state = STATE_GOT_BUS | STATE_GOT_TARGET |
8659 		    STATE_SENT_CMD | STATE_GOT_STATUS;
8660 
8661 		pkt->pkt_resid = 0;
8662 
8663 		if (cmd->cmd_pkt->pkt_numcookies) {
8664 			pkt->pkt_state |= STATE_XFERRED_DATA;
8665 			if (fpkt->pkt_data_resid) {
8666 				error++;
8667 			}
8668 		}
8669 
8670 		if ((pkt->pkt_scbp != NULL) && ((*(pkt->pkt_scbp) =
8671 		    rsp->fcp_u.fcp_status.scsi_status) != STATUS_GOOD)) {
8672 			/*
8673 			 * The next two checks make sure that if there
8674 			 * is no sense data or a valid response and
8675 			 * the command came back with check condition,
8676 			 * the command should be retried.
8677 			 */
8678 			if (!rsp->fcp_u.fcp_status.rsp_len_set &&
8679 			    !rsp->fcp_u.fcp_status.sense_len_set) {
8680 				pkt->pkt_state &= ~STATE_XFERRED_DATA;
8681 				pkt->pkt_resid = cmd->cmd_dmacount;
8682 			}
8683 		}
8684 
8685 		if ((error | rsp->fcp_u.i_fcp_status | rsp->fcp_resid) == 0) {
8686 			return;
8687 		}
8688 
8689 		plun = ADDR2LUN(&pkt->pkt_address);
8690 		ptgt = plun->lun_tgt;
8691 		ASSERT(ptgt != NULL);
8692 
8693 		/*
8694 		 * Update the transfer resid, if appropriate
8695 		 */
8696 		if (rsp->fcp_u.fcp_status.resid_over ||
8697 		    rsp->fcp_u.fcp_status.resid_under) {
8698 			pkt->pkt_resid = rsp->fcp_resid;
8699 		}
8700 
8701 		/*
8702 		 * First see if we got a FCP protocol error.
8703 		 */
8704 		if (rsp->fcp_u.fcp_status.rsp_len_set) {
8705 			struct fcp_rsp_info	*bep;
8706 			bep = (struct fcp_rsp_info *)(cmd->cmd_fcp_rsp +
8707 			    sizeof (struct fcp_rsp));
8708 
8709 			if (fcp_validate_fcp_response(rsp, pptr) !=
8710 			    FC_SUCCESS) {
8711 				pkt->pkt_reason = CMD_CMPLT;
8712 				*(pkt->pkt_scbp) = STATUS_CHECK;
8713 
8714 				fcp_log(CE_WARN, pptr->port_dip,
8715 				    "!SCSI command to d_id=0x%x lun=0x%x"
8716 				    " failed, Bad FCP response values:"
8717 				    " rsvd1=%x, rsvd2=%x, sts-rsvd1=%x,"
8718 				    " sts-rsvd2=%x, rsplen=%x, senselen=%x",
8719 				    ptgt->tgt_d_id, plun->lun_num,
8720 				    rsp->reserved_0, rsp->reserved_1,
8721 				    rsp->fcp_u.fcp_status.reserved_0,
8722 				    rsp->fcp_u.fcp_status.reserved_1,
8723 				    rsp->fcp_response_len, rsp->fcp_sense_len);
8724 
8725 				return;
8726 			}
8727 
8728 			if (pptr->port_fcp_dma != FC_NO_DVMA_SPACE) {
8729 				FCP_CP_IN(fpkt->pkt_resp +
8730 				    sizeof (struct fcp_rsp), bep,
8731 				    fpkt->pkt_resp_acc,
8732 				    sizeof (struct fcp_rsp_info));
8733 			}
8734 
8735 			if (bep->rsp_code != FCP_NO_FAILURE) {
8736 				child_info_t	*cip;
8737 
8738 				pkt->pkt_reason = CMD_TRAN_ERR;
8739 
8740 				mutex_enter(&plun->lun_mutex);
8741 				cip = plun->lun_cip;
8742 				mutex_exit(&plun->lun_mutex);
8743 
8744 				FCP_TRACE(fcp_logq, pptr->port_instbuf,
8745 				    fcp_trace, FCP_BUF_LEVEL_2, 0,
8746 				    "FCP response error on cmd=%p"
8747 				    " target=0x%x, cip=%p", cmd,
8748 				    ptgt->tgt_d_id, cip);
8749 			}
8750 		}
8751 
8752 		/*
8753 		 * See if we got a SCSI error with sense data
8754 		 */
8755 		if (rsp->fcp_u.fcp_status.sense_len_set) {
8756 			uchar_t				rqlen;
8757 			caddr_t				sense_from;
8758 			child_info_t			*cip;
8759 			timeout_id_t			tid;
8760 			struct scsi_arq_status		*arq;
8761 			struct scsi_extended_sense	*sense_to;
8762 
8763 			arq = (struct scsi_arq_status *)pkt->pkt_scbp;
8764 			sense_to = &arq->sts_sensedata;
8765 
8766 			rqlen = (uchar_t)min(rsp->fcp_sense_len,
8767 			    sizeof (struct scsi_extended_sense));
8768 
8769 			sense_from = (caddr_t)fpkt->pkt_resp +
8770 			    sizeof (struct fcp_rsp) + rsp->fcp_response_len;
8771 
8772 			if (fcp_validate_fcp_response(rsp, pptr) !=
8773 			    FC_SUCCESS) {
8774 				pkt->pkt_reason = CMD_CMPLT;
8775 				*(pkt->pkt_scbp) = STATUS_CHECK;
8776 
8777 				fcp_log(CE_WARN, pptr->port_dip,
8778 				    "!SCSI command to d_id=0x%x lun=0x%x"
8779 				    " failed, Bad FCP response values:"
8780 				    " rsvd1=%x, rsvd2=%x, sts-rsvd1=%x,"
8781 				    " sts-rsvd2=%x, rsplen=%x, senselen=%x",
8782 				    ptgt->tgt_d_id, plun->lun_num,
8783 				    rsp->reserved_0, rsp->reserved_1,
8784 				    rsp->fcp_u.fcp_status.reserved_0,
8785 				    rsp->fcp_u.fcp_status.reserved_1,
8786 				    rsp->fcp_response_len, rsp->fcp_sense_len);
8787 
8788 				return;
8789 			}
8790 
8791 			/*
8792 			 * copy in sense information
8793 			 */
8794 			if (pptr->port_fcp_dma != FC_NO_DVMA_SPACE) {
8795 				FCP_CP_IN(sense_from, sense_to,
8796 				    fpkt->pkt_resp_acc, rqlen);
8797 			} else {
8798 				bcopy(sense_from, sense_to, rqlen);
8799 			}
8800 
8801 			if ((FCP_SENSE_REPORTLUN_CHANGED(sense_to)) ||
8802 			    (FCP_SENSE_NO_LUN(sense_to))) {
8803 				mutex_enter(&ptgt->tgt_mutex);
8804 				if (ptgt->tgt_tid == NULL) {
8805 					/*
8806 					 * Kick off rediscovery
8807 					 */
8808 					tid = timeout(fcp_reconfigure_luns,
8809 					    (caddr_t)ptgt, drv_usectohz(1));
8810 
8811 					ptgt->tgt_tid = tid;
8812 					ptgt->tgt_state |= FCP_TGT_BUSY;
8813 				}
8814 				mutex_exit(&ptgt->tgt_mutex);
8815 				if (FCP_SENSE_REPORTLUN_CHANGED(sense_to)) {
8816 					FCP_TRACE(fcp_logq, pptr->port_instbuf,
8817 					    fcp_trace, FCP_BUF_LEVEL_3, 0,
8818 					    "!FCP: Report Lun Has Changed"
8819 					    " target=%x", ptgt->tgt_d_id);
8820 				} else if (FCP_SENSE_NO_LUN(sense_to)) {
8821 					FCP_TRACE(fcp_logq, pptr->port_instbuf,
8822 					    fcp_trace, FCP_BUF_LEVEL_3, 0,
8823 					    "!FCP: LU Not Supported"
8824 					    " target=%x", ptgt->tgt_d_id);
8825 				}
8826 			}
8827 			ASSERT(pkt->pkt_scbp != NULL);
8828 
8829 			pkt->pkt_state |= STATE_ARQ_DONE;
8830 
8831 			arq->sts_rqpkt_resid = SENSE_LENGTH - rqlen;
8832 
8833 			*((uchar_t *)&arq->sts_rqpkt_status) = STATUS_GOOD;
8834 			arq->sts_rqpkt_reason = 0;
8835 			arq->sts_rqpkt_statistics = 0;
8836 
8837 			arq->sts_rqpkt_state = STATE_GOT_BUS |
8838 			    STATE_GOT_TARGET | STATE_SENT_CMD |
8839 			    STATE_GOT_STATUS | STATE_ARQ_DONE |
8840 			    STATE_XFERRED_DATA;
8841 
8842 			mutex_enter(&plun->lun_mutex);
8843 			cip = plun->lun_cip;
8844 			mutex_exit(&plun->lun_mutex);
8845 
8846 			FCP_DTRACE(fcp_logq, pptr->port_instbuf,
8847 			    fcp_trace, FCP_BUF_LEVEL_8, 0,
8848 			    "SCSI Check condition on cmd=%p target=0x%x"
8849 			    " LUN=%p, cmd=%x SCSI status=%x, es key=%x"
8850 			    " ASC=%x ASCQ=%x", cmd, ptgt->tgt_d_id, cip,
8851 			    cmd->cmd_fcp_cmd.fcp_cdb[0],
8852 			    rsp->fcp_u.fcp_status.scsi_status,
8853 			    sense_to->es_key, sense_to->es_add_code,
8854 			    sense_to->es_qual_code);
8855 		}
8856 	} else {
8857 		plun = ADDR2LUN(&pkt->pkt_address);
8858 		ptgt = plun->lun_tgt;
8859 		ASSERT(ptgt != NULL);
8860 
8861 		/*
8862 		 * Work harder to translate errors into target driver
8863 		 * understandable ones. Note with despair that the target
8864 		 * drivers don't decode pkt_state and pkt_reason exhaustively
8865 		 * They resort to using the big hammer most often, which
8866 		 * may not get fixed in the life time of this driver.
8867 		 */
8868 		pkt->pkt_state = 0;
8869 		pkt->pkt_statistics = 0;
8870 
8871 		switch (fpkt->pkt_state) {
8872 		case FC_PKT_TRAN_ERROR:
8873 			switch (fpkt->pkt_reason) {
8874 			case FC_REASON_OVERRUN:
8875 				pkt->pkt_reason = CMD_CMD_OVR;
8876 				pkt->pkt_statistics |= STAT_ABORTED;
8877 				break;
8878 
8879 			case FC_REASON_XCHG_BSY: {
8880 				caddr_t ptr;
8881 
8882 				pkt->pkt_reason = CMD_CMPLT;	/* Lie */
8883 
8884 				ptr = (caddr_t)pkt->pkt_scbp;
8885 				if (ptr) {
8886 					*ptr = STATUS_BUSY;
8887 				}
8888 				break;
8889 			}
8890 
8891 			case FC_REASON_ABORTED:
8892 				pkt->pkt_reason = CMD_TRAN_ERR;
8893 				pkt->pkt_statistics |= STAT_ABORTED;
8894 				break;
8895 
8896 			case FC_REASON_ABORT_FAILED:
8897 				pkt->pkt_reason = CMD_ABORT_FAIL;
8898 				break;
8899 
8900 			case FC_REASON_NO_SEQ_INIT:
8901 			case FC_REASON_CRC_ERROR:
8902 				pkt->pkt_reason = CMD_TRAN_ERR;
8903 				pkt->pkt_statistics |= STAT_ABORTED;
8904 				break;
8905 			default:
8906 				pkt->pkt_reason = CMD_TRAN_ERR;
8907 				break;
8908 			}
8909 			break;
8910 
8911 		case FC_PKT_PORT_OFFLINE: {
8912 			dev_info_t	*cdip = NULL;
8913 			caddr_t		ptr;
8914 
8915 			if (fpkt->pkt_reason == FC_REASON_LOGIN_REQUIRED) {
8916 				FCP_DTRACE(fcp_logq, pptr->port_instbuf,
8917 				    fcp_trace, FCP_BUF_LEVEL_8, 0,
8918 				    "SCSI cmd; LOGIN REQUIRED from FCA for %x",
8919 				    ptgt->tgt_d_id);
8920 			}
8921 
8922 			mutex_enter(&plun->lun_mutex);
8923 			if (plun->lun_mpxio == 0) {
8924 				cdip = DIP(plun->lun_cip);
8925 			} else if (plun->lun_cip) {
8926 				cdip = mdi_pi_get_client(PIP(plun->lun_cip));
8927 			}
8928 
8929 			mutex_exit(&plun->lun_mutex);
8930 
8931 			if (cdip) {
8932 				(void) ndi_event_retrieve_cookie(
8933 				    pptr->port_ndi_event_hdl, cdip,
8934 				    FCAL_REMOVE_EVENT, &fcp_remove_eid,
8935 				    NDI_EVENT_NOPASS);
8936 				(void) ndi_event_run_callbacks(
8937 				    pptr->port_ndi_event_hdl, cdip,
8938 				    fcp_remove_eid, NULL);
8939 			}
8940 
8941 			/*
8942 			 * If the link goes off-line for a lip,
8943 			 * this will cause a error to the ST SG
8944 			 * SGEN drivers. By setting BUSY we will
8945 			 * give the drivers the chance to retry
8946 			 * before it blows of the job. ST will
8947 			 * remember how many times it has retried.
8948 			 */
8949 
8950 			if ((plun->lun_type == DTYPE_SEQUENTIAL) ||
8951 			    (plun->lun_type == DTYPE_CHANGER)) {
8952 				pkt->pkt_reason = CMD_CMPLT;	/* Lie */
8953 				ptr = (caddr_t)pkt->pkt_scbp;
8954 				if (ptr) {
8955 					*ptr = STATUS_BUSY;
8956 				}
8957 			} else {
8958 				pkt->pkt_reason = CMD_TRAN_ERR;
8959 				pkt->pkt_statistics |= STAT_BUS_RESET;
8960 			}
8961 			break;
8962 		}
8963 
8964 		case FC_PKT_TRAN_BSY:
8965 			/*
8966 			 * Use the ssd Qfull handling here.
8967 			 */
8968 			*pkt->pkt_scbp = STATUS_INTERMEDIATE;
8969 			pkt->pkt_state = STATE_GOT_BUS;
8970 			break;
8971 
8972 		case FC_PKT_TIMEOUT:
8973 			pkt->pkt_reason = CMD_TIMEOUT;
8974 			if (fpkt->pkt_reason == FC_REASON_ABORT_FAILED) {
8975 				pkt->pkt_statistics |= STAT_TIMEOUT;
8976 			} else {
8977 				pkt->pkt_statistics |= STAT_ABORTED;
8978 			}
8979 			break;
8980 
8981 		case FC_PKT_LOCAL_RJT:
8982 			switch (fpkt->pkt_reason) {
8983 			case FC_REASON_OFFLINE: {
8984 				dev_info_t	*cdip = NULL;
8985 
8986 				mutex_enter(&plun->lun_mutex);
8987 				if (plun->lun_mpxio == 0) {
8988 					cdip = DIP(plun->lun_cip);
8989 				} else if (plun->lun_cip) {
8990 					cdip = mdi_pi_get_client(
8991 					    PIP(plun->lun_cip));
8992 				}
8993 				mutex_exit(&plun->lun_mutex);
8994 
8995 				if (cdip) {
8996 					(void) ndi_event_retrieve_cookie(
8997 					    pptr->port_ndi_event_hdl, cdip,
8998 					    FCAL_REMOVE_EVENT,
8999 					    &fcp_remove_eid,
9000 					    NDI_EVENT_NOPASS);
9001 					(void) ndi_event_run_callbacks(
9002 					    pptr->port_ndi_event_hdl,
9003 					    cdip, fcp_remove_eid, NULL);
9004 				}
9005 
9006 				pkt->pkt_reason = CMD_TRAN_ERR;
9007 				pkt->pkt_statistics |= STAT_BUS_RESET;
9008 
9009 				break;
9010 			}
9011 
9012 			case FC_REASON_NOMEM:
9013 			case FC_REASON_QFULL: {
9014 				caddr_t ptr;
9015 
9016 				pkt->pkt_reason = CMD_CMPLT;	/* Lie */
9017 				ptr = (caddr_t)pkt->pkt_scbp;
9018 				if (ptr) {
9019 					*ptr = STATUS_BUSY;
9020 				}
9021 				break;
9022 			}
9023 
9024 			case FC_REASON_DMA_ERROR:
9025 				pkt->pkt_reason = CMD_DMA_DERR;
9026 				pkt->pkt_statistics |= STAT_ABORTED;
9027 				break;
9028 
9029 			case FC_REASON_CRC_ERROR:
9030 			case FC_REASON_UNDERRUN: {
9031 				uchar_t		status;
9032 				/*
9033 				 * Work around for Bugid: 4240945.
9034 				 * IB on A5k doesn't set the Underrun bit
9035 				 * in the fcp status, when it is transferring
9036 				 * less than requested amount of data. Work
9037 				 * around the ses problem to keep luxadm
9038 				 * happy till ibfirmware is fixed.
9039 				 */
9040 				if (pptr->port_fcp_dma != FC_NO_DVMA_SPACE) {
9041 					FCP_CP_IN(fpkt->pkt_resp, rsp,
9042 					    fpkt->pkt_resp_acc,
9043 					    sizeof (struct fcp_rsp));
9044 				}
9045 				status = rsp->fcp_u.fcp_status.scsi_status;
9046 				if (((plun->lun_type & DTYPE_MASK) ==
9047 				    DTYPE_ESI) && (status == STATUS_GOOD)) {
9048 					pkt->pkt_reason = CMD_CMPLT;
9049 					*pkt->pkt_scbp = status;
9050 					pkt->pkt_resid = 0;
9051 				} else {
9052 					pkt->pkt_reason = CMD_TRAN_ERR;
9053 					pkt->pkt_statistics |= STAT_ABORTED;
9054 				}
9055 				break;
9056 			}
9057 
9058 			case FC_REASON_NO_CONNECTION:
9059 			case FC_REASON_UNSUPPORTED:
9060 			case FC_REASON_ILLEGAL_REQ:
9061 			case FC_REASON_BAD_SID:
9062 			case FC_REASON_DIAG_BUSY:
9063 			case FC_REASON_FCAL_OPN_FAIL:
9064 			case FC_REASON_BAD_XID:
9065 			default:
9066 				pkt->pkt_reason = CMD_TRAN_ERR;
9067 				pkt->pkt_statistics |= STAT_ABORTED;
9068 				break;
9069 
9070 			}
9071 			break;
9072 
9073 		case FC_PKT_NPORT_RJT:
9074 		case FC_PKT_FABRIC_RJT:
9075 		case FC_PKT_NPORT_BSY:
9076 		case FC_PKT_FABRIC_BSY:
9077 		default:
9078 			FCP_DTRACE(fcp_logq, pptr->port_instbuf,
9079 			    fcp_trace, FCP_BUF_LEVEL_8, 0,
9080 			    "FC Status 0x%x, reason 0x%x",
9081 			    fpkt->pkt_state, fpkt->pkt_reason);
9082 			pkt->pkt_reason = CMD_TRAN_ERR;
9083 			pkt->pkt_statistics |= STAT_ABORTED;
9084 			break;
9085 		}
9086 
9087 		FCP_DTRACE(fcp_logq, pptr->port_instbuf,
9088 		    fcp_trace, FCP_BUF_LEVEL_9, 0,
9089 		    "!FC error on cmd=%p target=0x%x: pkt state=0x%x "
9090 		    " pkt reason=0x%x", cmd, ptgt->tgt_d_id, fpkt->pkt_state,
9091 		    fpkt->pkt_reason);
9092 	}
9093 
9094 	ASSERT(save.a_hba_tran == pkt->pkt_address.a_hba_tran);
9095 }
9096 
9097 
9098 static int
9099 fcp_validate_fcp_response(struct fcp_rsp *rsp, struct fcp_port *pptr)
9100 {
9101 	if (rsp->reserved_0 || rsp->reserved_1 ||
9102 	    rsp->fcp_u.fcp_status.reserved_0 ||
9103 	    rsp->fcp_u.fcp_status.reserved_1) {
9104 		/*
9105 		 * These reserved fields should ideally be zero. FCP-2 does say
9106 		 * that the recipient need not check for reserved fields to be
9107 		 * zero. If they are not zero, we will not make a fuss about it
9108 		 * - just log it (in debug to both trace buffer and messages
9109 		 * file and to trace buffer only in non-debug) and move on.
9110 		 *
9111 		 * Non-zero reserved fields were seen with minnows.
9112 		 *
9113 		 * qlc takes care of some of this but we cannot assume that all
9114 		 * FCAs will do so.
9115 		 */
9116 		FCP_TRACE(fcp_logq, pptr->port_instbuf, fcp_trace,
9117 		    FCP_BUF_LEVEL_5, 0,
9118 		    "Got fcp response packet with non-zero reserved fields "
9119 		    "rsp->reserved_0:0x%x, rsp_reserved_1:0x%x, "
9120 		    "status.reserved_0:0x%x, status.reserved_1:0x%x",
9121 		    rsp->reserved_0, rsp->reserved_1,
9122 		    rsp->fcp_u.fcp_status.reserved_0,
9123 		    rsp->fcp_u.fcp_status.reserved_1);
9124 	}
9125 
9126 	if (rsp->fcp_u.fcp_status.rsp_len_set && (rsp->fcp_response_len >
9127 	    (FCP_MAX_RSP_IU_SIZE - sizeof (struct fcp_rsp)))) {
9128 		return (FC_FAILURE);
9129 	}
9130 
9131 	if (rsp->fcp_u.fcp_status.sense_len_set && rsp->fcp_sense_len >
9132 	    (FCP_MAX_RSP_IU_SIZE - rsp->fcp_response_len -
9133 	    sizeof (struct fcp_rsp))) {
9134 		return (FC_FAILURE);
9135 	}
9136 
9137 	return (FC_SUCCESS);
9138 }
9139 
9140 
9141 /*
9142  * This is called when there is a change the in device state. The case we're
9143  * handling here is, if the d_id s does not match, offline this tgt and online
9144  * a new tgt with the new d_id.	 called from fcp_handle_devices with
9145  * port_mutex held.
9146  */
9147 static int
9148 fcp_device_changed(struct fcp_port *pptr, struct fcp_tgt *ptgt,
9149     fc_portmap_t *map_entry, int link_cnt, int tgt_cnt, int cause)
9150 {
9151 	ASSERT(mutex_owned(&pptr->port_mutex));
9152 
9153 	FCP_TRACE(fcp_logq, pptr->port_instbuf,
9154 	    fcp_trace, FCP_BUF_LEVEL_3, 0,
9155 	    "Starting fcp_device_changed...");
9156 
9157 	/*
9158 	 * The two cases where the port_device_changed is called is
9159 	 * either it changes it's d_id or it's hard address.
9160 	 */
9161 	if ((ptgt->tgt_d_id != map_entry->map_did.port_id) ||
9162 	    (FC_TOP_EXTERNAL(pptr->port_topology) &&
9163 	    (ptgt->tgt_hard_addr != map_entry->map_hard_addr.hard_addr))) {
9164 
9165 		/* offline this target */
9166 		mutex_enter(&ptgt->tgt_mutex);
9167 		if (!(ptgt->tgt_state & FCP_TGT_OFFLINE)) {
9168 			(void) fcp_offline_target(pptr, ptgt, link_cnt,
9169 			    0, 1, NDI_DEVI_REMOVE);
9170 		}
9171 		mutex_exit(&ptgt->tgt_mutex);
9172 
9173 		fcp_log(CE_NOTE, pptr->port_dip,
9174 		    "Change in target properties: Old D_ID=%x New D_ID=%x"
9175 		    " Old HA=%x New HA=%x", ptgt->tgt_d_id,
9176 		    map_entry->map_did.port_id, ptgt->tgt_hard_addr,
9177 		    map_entry->map_hard_addr.hard_addr);
9178 	}
9179 
9180 	return (fcp_handle_mapflags(pptr, ptgt, map_entry,
9181 	    link_cnt, tgt_cnt, cause));
9182 }
9183 
9184 /*
9185  *     Function: fcp_alloc_lun
9186  *
9187  *  Description: Creates a new lun structure and adds it to the list
9188  *		 of luns of the target.
9189  *
9190  *     Argument: ptgt		Target the lun will belong to.
9191  *
9192  * Return Value: NULL		Failed
9193  *		 Not NULL	Succeeded
9194  *
9195  *	Context: Kernel context
9196  */
9197 static struct fcp_lun *
9198 fcp_alloc_lun(struct fcp_tgt *ptgt)
9199 {
9200 	struct fcp_lun *plun;
9201 
9202 	plun = kmem_zalloc(sizeof (struct fcp_lun), KM_NOSLEEP);
9203 	if (plun != NULL) {
9204 		/*
9205 		 * Initialize the mutex before putting in the target list
9206 		 * especially before releasing the target mutex.
9207 		 */
9208 		mutex_init(&plun->lun_mutex, NULL, MUTEX_DRIVER, NULL);
9209 		plun->lun_tgt = ptgt;
9210 
9211 		mutex_enter(&ptgt->tgt_mutex);
9212 		plun->lun_next = ptgt->tgt_lun;
9213 		ptgt->tgt_lun = plun;
9214 		plun->lun_old_guid = NULL;
9215 		plun->lun_old_guid_size = 0;
9216 		mutex_exit(&ptgt->tgt_mutex);
9217 	}
9218 
9219 	return (plun);
9220 }
9221 
9222 /*
9223  *     Function: fcp_dealloc_lun
9224  *
9225  *  Description: Frees the LUN structure passed by the caller.
9226  *
9227  *     Argument: plun		LUN structure to free.
9228  *
9229  * Return Value: None
9230  *
9231  *	Context: Kernel context.
9232  */
9233 static void
9234 fcp_dealloc_lun(struct fcp_lun *plun)
9235 {
9236 	mutex_enter(&plun->lun_mutex);
9237 	if (plun->lun_cip) {
9238 		fcp_remove_child(plun);
9239 	}
9240 	mutex_exit(&plun->lun_mutex);
9241 
9242 	mutex_destroy(&plun->lun_mutex);
9243 	if (plun->lun_guid) {
9244 		kmem_free(plun->lun_guid, plun->lun_guid_size);
9245 	}
9246 	if (plun->lun_old_guid) {
9247 		kmem_free(plun->lun_old_guid, plun->lun_old_guid_size);
9248 	}
9249 	kmem_free(plun, sizeof (*plun));
9250 }
9251 
9252 /*
9253  *     Function: fcp_alloc_tgt
9254  *
9255  *  Description: Creates a new target structure and adds it to the port
9256  *		 hash list.
9257  *
9258  *     Argument: pptr		fcp port structure
9259  *		 *map_entry	entry describing the target to create
9260  *		 link_cnt	Link state change counter
9261  *
9262  * Return Value: NULL		Failed
9263  *		 Not NULL	Succeeded
9264  *
9265  *	Context: Kernel context.
9266  */
9267 static struct fcp_tgt *
9268 fcp_alloc_tgt(struct fcp_port *pptr, fc_portmap_t *map_entry, int link_cnt)
9269 {
9270 	int			hash;
9271 	uchar_t			*wwn;
9272 	struct fcp_tgt	*ptgt;
9273 
9274 	ptgt = kmem_zalloc(sizeof (*ptgt), KM_NOSLEEP);
9275 	if (ptgt != NULL) {
9276 		mutex_enter(&pptr->port_mutex);
9277 		if (link_cnt != pptr->port_link_cnt) {
9278 			/*
9279 			 * oh oh -- another link reset
9280 			 * in progress -- give up
9281 			 */
9282 			mutex_exit(&pptr->port_mutex);
9283 			kmem_free(ptgt, sizeof (*ptgt));
9284 			ptgt = NULL;
9285 		} else {
9286 			/*
9287 			 * initialize the mutex before putting in the port
9288 			 * wwn list, especially before releasing the port
9289 			 * mutex.
9290 			 */
9291 			mutex_init(&ptgt->tgt_mutex, NULL, MUTEX_DRIVER, NULL);
9292 
9293 			/* add new target entry to the port's hash list */
9294 			wwn = (uchar_t *)&map_entry->map_pwwn;
9295 			hash = FCP_HASH(wwn);
9296 
9297 			ptgt->tgt_next = pptr->port_tgt_hash_table[hash];
9298 			pptr->port_tgt_hash_table[hash] = ptgt;
9299 
9300 			/* save cross-ptr */
9301 			ptgt->tgt_port = pptr;
9302 
9303 			ptgt->tgt_change_cnt = 1;
9304 
9305 			/* initialize the target manual_config_only flag */
9306 			if (fcp_enable_auto_configuration) {
9307 				ptgt->tgt_manual_config_only = 0;
9308 			} else {
9309 				ptgt->tgt_manual_config_only = 1;
9310 			}
9311 
9312 			mutex_exit(&pptr->port_mutex);
9313 		}
9314 	}
9315 
9316 	return (ptgt);
9317 }
9318 
9319 /*
9320  *     Function: fcp_dealloc_tgt
9321  *
9322  *  Description: Frees the target structure passed by the caller.
9323  *
9324  *     Argument: ptgt		Target structure to free.
9325  *
9326  * Return Value: None
9327  *
9328  *	Context: Kernel context.
9329  */
9330 static void
9331 fcp_dealloc_tgt(struct fcp_tgt *ptgt)
9332 {
9333 	mutex_destroy(&ptgt->tgt_mutex);
9334 	kmem_free(ptgt, sizeof (*ptgt));
9335 }
9336 
9337 
9338 /*
9339  * Handle STATUS_QFULL and STATUS_BUSY by performing delayed retry
9340  *
9341  *	Device discovery commands will not be retried for-ever as
9342  *	this will have repercussions on other devices that need to
9343  *	be submitted to the hotplug thread. After a quick glance
9344  *	at the SCSI-3 spec, it was found that the spec doesn't
9345  *	mandate a forever retry, rather recommends a delayed retry.
9346  *
9347  *	Since Photon IB is single threaded, STATUS_BUSY is common
9348  *	in a 4+initiator environment. Make sure the total time
9349  *	spent on retries (including command timeout) does not
9350  *	60 seconds
9351  */
9352 static void
9353 fcp_queue_ipkt(struct fcp_port *pptr, fc_packet_t *fpkt)
9354 {
9355 	struct fcp_ipkt *icmd = (struct fcp_ipkt *)fpkt->pkt_ulp_private;
9356 	struct fcp_tgt *ptgt = icmd->ipkt_tgt;
9357 
9358 	mutex_enter(&pptr->port_mutex);
9359 	mutex_enter(&ptgt->tgt_mutex);
9360 	if (FCP_STATE_CHANGED(pptr, ptgt, icmd)) {
9361 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
9362 		    fcp_trace, FCP_BUF_LEVEL_2, 0,
9363 		    "fcp_queue_ipkt,1:state change occured"
9364 		    " for D_ID=0x%x", ptgt->tgt_d_id);
9365 		mutex_exit(&ptgt->tgt_mutex);
9366 		mutex_exit(&pptr->port_mutex);
9367 		(void) fcp_call_finish_init(pptr, ptgt, icmd->ipkt_link_cnt,
9368 		    icmd->ipkt_change_cnt, icmd->ipkt_cause);
9369 		fcp_icmd_free(pptr, icmd);
9370 		return;
9371 	}
9372 	mutex_exit(&ptgt->tgt_mutex);
9373 
9374 	icmd->ipkt_restart = fcp_watchdog_time + icmd->ipkt_retries++;
9375 
9376 	if (pptr->port_ipkt_list != NULL) {
9377 		/* add pkt to front of doubly-linked list */
9378 		pptr->port_ipkt_list->ipkt_prev = icmd;
9379 		icmd->ipkt_next = pptr->port_ipkt_list;
9380 		pptr->port_ipkt_list = icmd;
9381 		icmd->ipkt_prev = NULL;
9382 	} else {
9383 		/* this is the first/only pkt on the list */
9384 		pptr->port_ipkt_list = icmd;
9385 		icmd->ipkt_next = NULL;
9386 		icmd->ipkt_prev = NULL;
9387 	}
9388 	mutex_exit(&pptr->port_mutex);
9389 }
9390 
9391 /*
9392  *     Function: fcp_transport
9393  *
9394  *  Description: This function submits the Fibre Channel packet to the transort
9395  *		 layer by calling fc_ulp_transport().  If fc_ulp_transport()
9396  *		 fails the submission, the treatment depends on the value of
9397  *		 the variable internal.
9398  *
9399  *     Argument: port_handle	fp/fctl port handle.
9400  *		 *fpkt		Packet to submit to the transport layer.
9401  *		 internal	Not zero when it's an internal packet.
9402  *
9403  * Return Value: FC_TRAN_BUSY
9404  *		 FC_STATEC_BUSY
9405  *		 FC_OFFLINE
9406  *		 FC_LOGINREQ
9407  *		 FC_DEVICE_BUSY
9408  *		 FC_SUCCESS
9409  */
9410 static int
9411 fcp_transport(opaque_t port_handle, fc_packet_t *fpkt, int internal)
9412 {
9413 	int	rval;
9414 
9415 	rval = fc_ulp_transport(port_handle, fpkt);
9416 	if (rval == FC_SUCCESS) {
9417 		return (rval);
9418 	}
9419 
9420 	/*
9421 	 * LUN isn't marked BUSY or OFFLINE, so we got here to transport
9422 	 * a command, if the underlying modules see that there is a state
9423 	 * change, or if a port is OFFLINE, that means, that state change
9424 	 * hasn't reached FCP yet, so re-queue the command for deferred
9425 	 * submission.
9426 	 */
9427 	if ((rval == FC_STATEC_BUSY) || (rval == FC_OFFLINE) ||
9428 	    (rval == FC_LOGINREQ) || (rval == FC_DEVICE_BUSY) ||
9429 	    (rval == FC_DEVICE_BUSY_NEW_RSCN) || (rval == FC_TRAN_BUSY)) {
9430 		/*
9431 		 * Defer packet re-submission. Life hang is possible on
9432 		 * internal commands if the port driver sends FC_STATEC_BUSY
9433 		 * for ever, but that shouldn't happen in a good environment.
9434 		 * Limiting re-transport for internal commands is probably a
9435 		 * good idea..
9436 		 * A race condition can happen when a port sees barrage of
9437 		 * link transitions offline to online. If the FCTL has
9438 		 * returned FC_STATEC_BUSY or FC_OFFLINE then none of the
9439 		 * internal commands should be queued to do the discovery.
9440 		 * The race condition is when an online comes and FCP starts
9441 		 * its internal discovery and the link goes offline. It is
9442 		 * possible that the statec_callback has not reached FCP
9443 		 * and FCP is carrying on with its internal discovery.
9444 		 * FC_STATEC_BUSY or FC_OFFLINE will be the first indication
9445 		 * that the link has gone offline. At this point FCP should
9446 		 * drop all the internal commands and wait for the
9447 		 * statec_callback. It will be facilitated by incrementing
9448 		 * port_link_cnt.
9449 		 *
9450 		 * For external commands, the (FC)pkt_timeout is decremented
9451 		 * by the QUEUE Delay added by our driver, Care is taken to
9452 		 * ensure that it doesn't become zero (zero means no timeout)
9453 		 * If the time expires right inside driver queue itself,
9454 		 * the watch thread will return it to the original caller
9455 		 * indicating that the command has timed-out.
9456 		 */
9457 		if (internal) {
9458 			char			*op;
9459 			struct fcp_ipkt	*icmd;
9460 
9461 			icmd = (struct fcp_ipkt *)fpkt->pkt_ulp_private;
9462 			switch (icmd->ipkt_opcode) {
9463 			case SCMD_REPORT_LUN:
9464 				op = "REPORT LUN";
9465 				break;
9466 
9467 			case SCMD_INQUIRY:
9468 				op = "INQUIRY";
9469 				break;
9470 
9471 			case SCMD_INQUIRY_PAGE83:
9472 				op = "INQUIRY-83";
9473 				break;
9474 
9475 			default:
9476 				op = "Internal SCSI COMMAND";
9477 				break;
9478 			}
9479 
9480 			if (fcp_handle_ipkt_errors(icmd->ipkt_port,
9481 			    icmd->ipkt_tgt, icmd, rval, op) == DDI_SUCCESS) {
9482 				rval = FC_SUCCESS;
9483 			}
9484 		} else {
9485 			struct fcp_pkt *cmd;
9486 			struct fcp_port *pptr;
9487 
9488 			cmd = (struct fcp_pkt *)fpkt->pkt_ulp_private;
9489 			cmd->cmd_state = FCP_PKT_IDLE;
9490 			pptr = ADDR2FCP(&cmd->cmd_pkt->pkt_address);
9491 
9492 			if (cmd->cmd_pkt->pkt_flags & FLAG_NOQUEUE) {
9493 				FCP_DTRACE(fcp_logq, pptr->port_instbuf,
9494 				    fcp_trace, FCP_BUF_LEVEL_9, 0,
9495 				    "fcp_transport: xport busy for pkt %p",
9496 				    cmd->cmd_pkt);
9497 				rval = FC_TRAN_BUSY;
9498 			} else {
9499 				fcp_queue_pkt(pptr, cmd);
9500 				rval = FC_SUCCESS;
9501 			}
9502 		}
9503 	}
9504 
9505 	return (rval);
9506 }
9507 
9508 /*VARARGS3*/
9509 static void
9510 fcp_log(int level, dev_info_t *dip, const char *fmt, ...)
9511 {
9512 	char		buf[256];
9513 	va_list		ap;
9514 
9515 	if (dip == NULL) {
9516 		dip = fcp_global_dip;
9517 	}
9518 
9519 	va_start(ap, fmt);
9520 	(void) vsprintf(buf, fmt, ap);
9521 	va_end(ap);
9522 
9523 	scsi_log(dip, "fcp", level, buf);
9524 }
9525 
9526 /*
9527  * This function retries NS registry of FC4 type.
9528  * It assumes that fcp_mutex is held.
9529  * The function does nothing if topology is not fabric
9530  * So, the topology has to be set before this function can be called
9531  */
9532 static void
9533 fcp_retry_ns_registry(struct fcp_port *pptr, uint32_t s_id)
9534 {
9535 	int	rval;
9536 
9537 	ASSERT(MUTEX_HELD(&pptr->port_mutex));
9538 
9539 	if (((pptr->port_state & FCP_STATE_NS_REG_FAILED) == 0) ||
9540 	    ((pptr->port_topology != FC_TOP_FABRIC) &&
9541 	    (pptr->port_topology != FC_TOP_PUBLIC_LOOP))) {
9542 		if (pptr->port_state & FCP_STATE_NS_REG_FAILED) {
9543 			pptr->port_state &= ~FCP_STATE_NS_REG_FAILED;
9544 		}
9545 		return;
9546 	}
9547 	mutex_exit(&pptr->port_mutex);
9548 	rval = fcp_do_ns_registry(pptr, s_id);
9549 	mutex_enter(&pptr->port_mutex);
9550 
9551 	if (rval == 0) {
9552 		/* Registry successful. Reset flag */
9553 		pptr->port_state &= ~(FCP_STATE_NS_REG_FAILED);
9554 	}
9555 }
9556 
9557 /*
9558  * This function registers the ULP with the switch by calling transport i/f
9559  */
9560 static int
9561 fcp_do_ns_registry(struct fcp_port *pptr, uint32_t s_id)
9562 {
9563 	fc_ns_cmd_t		ns_cmd;
9564 	ns_rfc_type_t		rfc;
9565 	uint32_t		types[8];
9566 
9567 	/*
9568 	 * Prepare the Name server structure to
9569 	 * register with the transport in case of
9570 	 * Fabric configuration.
9571 	 */
9572 	bzero(&rfc, sizeof (rfc));
9573 	bzero(types, sizeof (types));
9574 
9575 	types[FC4_TYPE_WORD_POS(FC_TYPE_SCSI_FCP)] =
9576 	    (1 << FC4_TYPE_BIT_POS(FC_TYPE_SCSI_FCP));
9577 
9578 	rfc.rfc_port_id.port_id = s_id;
9579 	bcopy(types, rfc.rfc_types, sizeof (types));
9580 
9581 	ns_cmd.ns_flags = 0;
9582 	ns_cmd.ns_cmd = NS_RFT_ID;
9583 	ns_cmd.ns_req_len = sizeof (rfc);
9584 	ns_cmd.ns_req_payload = (caddr_t)&rfc;
9585 	ns_cmd.ns_resp_len = 0;
9586 	ns_cmd.ns_resp_payload = NULL;
9587 
9588 	/*
9589 	 * Perform the Name Server Registration for SCSI_FCP FC4 Type.
9590 	 */
9591 	if (fc_ulp_port_ns(pptr->port_fp_handle, NULL, &ns_cmd)) {
9592 		fcp_log(CE_WARN, pptr->port_dip,
9593 		    "!ns_registry: failed name server registration");
9594 		return (1);
9595 	}
9596 
9597 	return (0);
9598 }
9599 
9600 /*
9601  *     Function: fcp_handle_port_attach
9602  *
9603  *  Description: This function is called from fcp_port_attach() to attach a
9604  *		 new port. This routine does the following:
9605  *
9606  *		1) Allocates an fcp_port structure and initializes it.
9607  *		2) Tries to register the new FC-4 (FCP) capablity with the name
9608  *		   server.
9609  *		3) Kicks off the enumeration of the targets/luns visible
9610  *		   through this new port.  That is done by calling
9611  *		   fcp_statec_callback() if the port is online.
9612  *
9613  *     Argument: ulph		fp/fctl port handle.
9614  *		 *pinfo		Port information.
9615  *		 s_id		Port ID.
9616  *		 instance	Device instance number for the local port
9617  *				(returned by ddi_get_instance()).
9618  *
9619  * Return Value: DDI_SUCCESS
9620  *		 DDI_FAILURE
9621  *
9622  *	Context: User and Kernel context.
9623  */
9624 /*ARGSUSED*/
9625 int
9626 fcp_handle_port_attach(opaque_t ulph, fc_ulp_port_info_t *pinfo,
9627     uint32_t s_id, int instance)
9628 {
9629 	int			res = DDI_FAILURE;
9630 	scsi_hba_tran_t		*tran;
9631 	int			mutex_initted = FALSE;
9632 	int			hba_attached = FALSE;
9633 	int			soft_state_linked = FALSE;
9634 	int			event_bind = FALSE;
9635 	struct fcp_port		*pptr;
9636 	fc_portmap_t		*tmp_list = NULL;
9637 	uint32_t		max_cnt, alloc_cnt;
9638 	uchar_t			*boot_wwn = NULL;
9639 	uint_t			nbytes;
9640 	int			manual_cfg;
9641 
9642 	/*
9643 	 * this port instance attaching for the first time (or after
9644 	 * being detached before)
9645 	 */
9646 	FCP_TRACE(fcp_logq, "fcp", fcp_trace,
9647 	    FCP_BUF_LEVEL_3, 0, "port attach: for port %d", instance);
9648 
9649 	if (ddi_soft_state_zalloc(fcp_softstate, instance) != DDI_SUCCESS) {
9650 		cmn_err(CE_WARN, "fcp: Softstate struct alloc failed"
9651 		    "parent dip: %p; instance: %d", (void *)pinfo->port_dip,
9652 		    instance);
9653 		return (res);
9654 	}
9655 
9656 	if ((pptr = ddi_get_soft_state(fcp_softstate, instance)) == NULL) {
9657 		/* this shouldn't happen */
9658 		ddi_soft_state_free(fcp_softstate, instance);
9659 		cmn_err(CE_WARN, "fcp: bad soft state");
9660 		return (res);
9661 	}
9662 
9663 	(void) sprintf(pptr->port_instbuf, "fcp(%d)", instance);
9664 
9665 	/*
9666 	 * Make a copy of ulp_port_info as fctl allocates
9667 	 * a temp struct.
9668 	 */
9669 	(void) fcp_cp_pinfo(pptr, pinfo);
9670 
9671 	/*
9672 	 * Check for manual_configuration_only property.
9673 	 * Enable manual configurtion if the property is
9674 	 * set to 1, otherwise disable manual configuration.
9675 	 */
9676 	if ((manual_cfg = ddi_prop_get_int(DDI_DEV_T_ANY, pptr->port_dip,
9677 	    DDI_PROP_NOTPROM | DDI_PROP_DONTPASS,
9678 	    MANUAL_CFG_ONLY,
9679 	    -1)) != -1) {
9680 		if (manual_cfg == 1) {
9681 			char	*pathname;
9682 			pathname = kmem_zalloc(MAXPATHLEN, KM_SLEEP);
9683 			(void) ddi_pathname(pptr->port_dip, pathname);
9684 			cmn_err(CE_NOTE,
9685 			    "%s (%s%d) %s is enabled via %s.conf.",
9686 			    pathname,
9687 			    ddi_driver_name(pptr->port_dip),
9688 			    ddi_get_instance(pptr->port_dip),
9689 			    MANUAL_CFG_ONLY,
9690 			    ddi_driver_name(pptr->port_dip));
9691 			fcp_enable_auto_configuration = 0;
9692 			kmem_free(pathname, MAXPATHLEN);
9693 		}
9694 	}
9695 	_NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(pptr->port_link_cnt))
9696 	pptr->port_link_cnt = 1;
9697 	_NOTE(NOW_VISIBLE_TO_OTHER_THREADS(pptr->port_link_cnt))
9698 	pptr->port_id = s_id;
9699 	pptr->port_instance = instance;
9700 	_NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(pptr->port_state))
9701 	pptr->port_state = FCP_STATE_INIT;
9702 	_NOTE(NOW_VISIBLE_TO_OTHER_THREADS(pptr->port_state))
9703 
9704 	pptr->port_dmacookie_sz = (pptr->port_data_dma_attr.dma_attr_sgllen *
9705 	    sizeof (ddi_dma_cookie_t));
9706 
9707 	/*
9708 	 * The two mutexes of fcp_port are initialized.	 The variable
9709 	 * mutex_initted is incremented to remember that fact.	That variable
9710 	 * is checked when the routine fails and the mutexes have to be
9711 	 * destroyed.
9712 	 */
9713 	mutex_init(&pptr->port_mutex, NULL, MUTEX_DRIVER, NULL);
9714 	mutex_init(&pptr->port_pkt_mutex, NULL, MUTEX_DRIVER, NULL);
9715 	mutex_initted++;
9716 
9717 	/*
9718 	 * The SCSI tran structure is allocate and initialized now.
9719 	 */
9720 	if ((tran = scsi_hba_tran_alloc(pptr->port_dip, 0)) == NULL) {
9721 		fcp_log(CE_WARN, pptr->port_dip,
9722 		    "!fcp%d: scsi_hba_tran_alloc failed", instance);
9723 		goto fail;
9724 	}
9725 
9726 	/* link in the transport structure then fill it in */
9727 	pptr->port_tran = tran;
9728 	tran->tran_hba_private		= pptr;
9729 	tran->tran_tgt_init		= fcp_scsi_tgt_init;
9730 	tran->tran_tgt_probe		= NULL;
9731 	tran->tran_tgt_free		= fcp_scsi_tgt_free;
9732 	tran->tran_start		= fcp_scsi_start;
9733 	tran->tran_reset		= fcp_scsi_reset;
9734 	tran->tran_abort		= fcp_scsi_abort;
9735 	tran->tran_getcap		= fcp_scsi_getcap;
9736 	tran->tran_setcap		= fcp_scsi_setcap;
9737 	tran->tran_init_pkt		= NULL;
9738 	tran->tran_destroy_pkt		= NULL;
9739 	tran->tran_dmafree		= NULL;
9740 	tran->tran_sync_pkt		= NULL;
9741 	tran->tran_reset_notify		= fcp_scsi_reset_notify;
9742 	tran->tran_get_bus_addr		= fcp_scsi_get_bus_addr;
9743 	tran->tran_get_name		= fcp_scsi_get_name;
9744 	tran->tran_clear_aca		= NULL;
9745 	tran->tran_clear_task_set	= NULL;
9746 	tran->tran_terminate_task	= NULL;
9747 	tran->tran_get_eventcookie	= fcp_scsi_bus_get_eventcookie;
9748 	tran->tran_add_eventcall	= fcp_scsi_bus_add_eventcall;
9749 	tran->tran_remove_eventcall	= fcp_scsi_bus_remove_eventcall;
9750 	tran->tran_post_event		= fcp_scsi_bus_post_event;
9751 	tran->tran_quiesce		= NULL;
9752 	tran->tran_unquiesce		= NULL;
9753 	tran->tran_bus_reset		= NULL;
9754 	tran->tran_bus_config		= fcp_scsi_bus_config;
9755 	tran->tran_bus_unconfig		= fcp_scsi_bus_unconfig;
9756 	tran->tran_bus_power		= NULL;
9757 	tran->tran_interconnect_type	= INTERCONNECT_FABRIC;
9758 
9759 	tran->tran_pkt_constructor	= fcp_kmem_cache_constructor;
9760 	tran->tran_pkt_destructor	= fcp_kmem_cache_destructor;
9761 	tran->tran_setup_pkt		= fcp_pkt_setup;
9762 	tran->tran_teardown_pkt		= fcp_pkt_teardown;
9763 	tran->tran_hba_len		= pptr->port_priv_pkt_len +
9764 	    sizeof (struct fcp_pkt) + pptr->port_dmacookie_sz;
9765 
9766 	/*
9767 	 * Allocate an ndi event handle
9768 	 */
9769 	pptr->port_ndi_event_defs = (ndi_event_definition_t *)
9770 	    kmem_zalloc(sizeof (fcp_ndi_event_defs), KM_SLEEP);
9771 
9772 	bcopy(fcp_ndi_event_defs, pptr->port_ndi_event_defs,
9773 	    sizeof (fcp_ndi_event_defs));
9774 
9775 	(void) ndi_event_alloc_hdl(pptr->port_dip, NULL,
9776 	    &pptr->port_ndi_event_hdl, NDI_SLEEP);
9777 
9778 	pptr->port_ndi_events.ndi_events_version = NDI_EVENTS_REV1;
9779 	pptr->port_ndi_events.ndi_n_events = FCP_N_NDI_EVENTS;
9780 	pptr->port_ndi_events.ndi_event_defs = pptr->port_ndi_event_defs;
9781 
9782 	if (DEVI_IS_ATTACHING(pptr->port_dip) &&
9783 	    (ndi_event_bind_set(pptr->port_ndi_event_hdl,
9784 	    &pptr->port_ndi_events, NDI_SLEEP) != NDI_SUCCESS)) {
9785 		goto fail;
9786 	}
9787 	event_bind++;	/* Checked in fail case */
9788 
9789 	if (scsi_hba_attach_setup(pptr->port_dip, &pptr->port_data_dma_attr,
9790 	    tran, SCSI_HBA_ADDR_COMPLEX | SCSI_HBA_TRAN_SCB)
9791 	    != DDI_SUCCESS) {
9792 		fcp_log(CE_WARN, pptr->port_dip,
9793 		    "!fcp%d: scsi_hba_attach_setup failed", instance);
9794 		goto fail;
9795 	}
9796 	hba_attached++;	/* Checked in fail case */
9797 
9798 	pptr->port_mpxio = 0;
9799 	if (mdi_phci_register(MDI_HCI_CLASS_SCSI, pptr->port_dip, 0) ==
9800 	    MDI_SUCCESS) {
9801 		pptr->port_mpxio++;
9802 	}
9803 
9804 	/*
9805 	 * The following code is putting the new port structure in the global
9806 	 * list of ports and, if it is the first port to attach, it start the
9807 	 * fcp_watchdog_tick.
9808 	 *
9809 	 * Why put this new port in the global before we are done attaching it?
9810 	 * We are actually making the structure globally known before we are
9811 	 * done attaching it.  The reason for that is: because of the code that
9812 	 * follows.  At this point the resources to handle the port are
9813 	 * allocated.  This function is now going to do the following:
9814 	 *
9815 	 *   1) It is going to try to register with the name server advertizing
9816 	 *	the new FCP capability of the port.
9817 	 *   2) It is going to play the role of the fp/fctl layer by building
9818 	 *	a list of worlwide names reachable through this port and call
9819 	 *	itself on fcp_statec_callback().  That requires the port to
9820 	 *	be part of the global list.
9821 	 */
9822 	mutex_enter(&fcp_global_mutex);
9823 	if (fcp_port_head == NULL) {
9824 		fcp_read_blacklist(pinfo->port_dip, &fcp_lun_blacklist);
9825 	}
9826 	pptr->port_next = fcp_port_head;
9827 	fcp_port_head = pptr;
9828 	soft_state_linked++;
9829 
9830 	if (fcp_watchdog_init++ == 0) {
9831 		fcp_watchdog_tick = fcp_watchdog_timeout *
9832 		    drv_usectohz(1000000);
9833 		fcp_watchdog_id = timeout(fcp_watch, NULL,
9834 		    fcp_watchdog_tick);
9835 	}
9836 	mutex_exit(&fcp_global_mutex);
9837 
9838 	/*
9839 	 * Here an attempt is made to register with the name server, the new
9840 	 * FCP capability.  That is done using an RTF_ID to the name server.
9841 	 * It is done synchronously.  The function fcp_do_ns_registry()
9842 	 * doesn't return till the name server responded.
9843 	 * On failures, just ignore it for now and it will get retried during
9844 	 * state change callbacks. We'll set a flag to show this failure
9845 	 */
9846 	if (fcp_do_ns_registry(pptr, s_id)) {
9847 		mutex_enter(&pptr->port_mutex);
9848 		pptr->port_state |= FCP_STATE_NS_REG_FAILED;
9849 		mutex_exit(&pptr->port_mutex);
9850 	} else {
9851 		mutex_enter(&pptr->port_mutex);
9852 		pptr->port_state &= ~(FCP_STATE_NS_REG_FAILED);
9853 		mutex_exit(&pptr->port_mutex);
9854 	}
9855 
9856 	/*
9857 	 * Lookup for boot WWN property
9858 	 */
9859 	if (modrootloaded != 1) {
9860 		if ((ddi_prop_lookup_byte_array(DDI_DEV_T_ANY,
9861 		    ddi_get_parent(pinfo->port_dip),
9862 		    DDI_PROP_DONTPASS, OBP_BOOT_WWN,
9863 		    &boot_wwn, &nbytes) == DDI_PROP_SUCCESS) &&
9864 		    (nbytes == FC_WWN_SIZE)) {
9865 			bcopy(boot_wwn, pptr->port_boot_wwn, FC_WWN_SIZE);
9866 		}
9867 		if (boot_wwn) {
9868 			ddi_prop_free(boot_wwn);
9869 		}
9870 	}
9871 
9872 	/*
9873 	 * Handle various topologies and link states.
9874 	 */
9875 	switch (FC_PORT_STATE_MASK(pptr->port_phys_state)) {
9876 	case FC_STATE_OFFLINE:
9877 
9878 		/*
9879 		 * we're attaching a port where the link is offline
9880 		 *
9881 		 * Wait for ONLINE, at which time a state
9882 		 * change will cause a statec_callback
9883 		 *
9884 		 * in the mean time, do not do anything
9885 		 */
9886 		res = DDI_SUCCESS;
9887 		pptr->port_state |= FCP_STATE_OFFLINE;
9888 		break;
9889 
9890 	case FC_STATE_ONLINE: {
9891 		if (pptr->port_topology == FC_TOP_UNKNOWN) {
9892 			(void) fcp_linkreset(pptr, NULL, KM_NOSLEEP);
9893 			res = DDI_SUCCESS;
9894 			break;
9895 		}
9896 		/*
9897 		 * discover devices and create nodes (a private
9898 		 * loop or point-to-point)
9899 		 */
9900 		ASSERT(pptr->port_topology != FC_TOP_UNKNOWN);
9901 
9902 		/*
9903 		 * At this point we are going to build a list of all the ports
9904 		 * that	can be reached through this local port.	 It looks like
9905 		 * we cannot handle more than FCP_MAX_DEVICES per local port
9906 		 * (128).
9907 		 */
9908 		if ((tmp_list = (fc_portmap_t *)kmem_zalloc(
9909 		    sizeof (fc_portmap_t) * FCP_MAX_DEVICES,
9910 		    KM_NOSLEEP)) == NULL) {
9911 			fcp_log(CE_WARN, pptr->port_dip,
9912 			    "!fcp%d: failed to allocate portmap",
9913 			    instance);
9914 			goto fail;
9915 		}
9916 
9917 		/*
9918 		 * fc_ulp_getportmap() is going to provide us with the list of
9919 		 * remote ports in the buffer we just allocated.  The way the
9920 		 * list is going to be retrieved depends on the topology.
9921 		 * However, if we are connected to a Fabric, a name server
9922 		 * request may be sent to get the list of FCP capable ports.
9923 		 * It should be noted that is the case the request is
9924 		 * synchronous.	 This means we are stuck here till the name
9925 		 * server replies.  A lot of things can change during that time
9926 		 * and including, may be, being called on
9927 		 * fcp_statec_callback() for different reasons. I'm not sure
9928 		 * the code can handle that.
9929 		 */
9930 		max_cnt = FCP_MAX_DEVICES;
9931 		alloc_cnt = FCP_MAX_DEVICES;
9932 		if ((res = fc_ulp_getportmap(pptr->port_fp_handle,
9933 		    &tmp_list, &max_cnt, FC_ULP_PLOGI_PRESERVE)) !=
9934 		    FC_SUCCESS) {
9935 			caddr_t msg;
9936 
9937 			(void) fc_ulp_error(res, &msg);
9938 
9939 			/*
9940 			 * this	 just means the transport is
9941 			 * busy perhaps building a portmap so,
9942 			 * for now, succeed this port attach
9943 			 * when the transport has a new map,
9944 			 * it'll send us a state change then
9945 			 */
9946 			fcp_log(CE_WARN, pptr->port_dip,
9947 			    "!failed to get port map : %s", msg);
9948 
9949 			res = DDI_SUCCESS;
9950 			break;	/* go return result */
9951 		}
9952 		if (max_cnt > alloc_cnt) {
9953 			alloc_cnt = max_cnt;
9954 		}
9955 
9956 		/*
9957 		 * We are now going to call fcp_statec_callback() ourselves.
9958 		 * By issuing this call we are trying to kick off the enumera-
9959 		 * tion process.
9960 		 */
9961 		/*
9962 		 * let the state change callback do the SCSI device
9963 		 * discovery and create the devinfos
9964 		 */
9965 		fcp_statec_callback(ulph, pptr->port_fp_handle,
9966 		    pptr->port_phys_state, pptr->port_topology, tmp_list,
9967 		    max_cnt, pptr->port_id);
9968 
9969 		res = DDI_SUCCESS;
9970 		break;
9971 	}
9972 
9973 	default:
9974 		/* unknown port state */
9975 		fcp_log(CE_WARN, pptr->port_dip,
9976 		    "!fcp%d: invalid port state at attach=0x%x",
9977 		    instance, pptr->port_phys_state);
9978 
9979 		mutex_enter(&pptr->port_mutex);
9980 		pptr->port_phys_state = FCP_STATE_OFFLINE;
9981 		mutex_exit(&pptr->port_mutex);
9982 
9983 		res = DDI_SUCCESS;
9984 		break;
9985 	}
9986 
9987 	/* free temp list if used */
9988 	if (tmp_list != NULL) {
9989 		kmem_free(tmp_list, sizeof (fc_portmap_t) * alloc_cnt);
9990 	}
9991 
9992 	/* note the attach time */
9993 	pptr->port_attach_time = lbolt64;
9994 
9995 	/* all done */
9996 	return (res);
9997 
9998 	/* a failure we have to clean up after */
9999 fail:
10000 	fcp_log(CE_WARN, pptr->port_dip, "!failed to attach to port");
10001 
10002 	if (soft_state_linked) {
10003 		/* remove this fcp_port from the linked list */
10004 		(void) fcp_soft_state_unlink(pptr);
10005 	}
10006 
10007 	/* unbind and free event set */
10008 	if (pptr->port_ndi_event_hdl) {
10009 		if (event_bind) {
10010 			(void) ndi_event_unbind_set(pptr->port_ndi_event_hdl,
10011 			    &pptr->port_ndi_events, NDI_SLEEP);
10012 		}
10013 		(void) ndi_event_free_hdl(pptr->port_ndi_event_hdl);
10014 	}
10015 
10016 	if (pptr->port_ndi_event_defs) {
10017 		(void) kmem_free(pptr->port_ndi_event_defs,
10018 		    sizeof (fcp_ndi_event_defs));
10019 	}
10020 
10021 	/*
10022 	 * Clean up mpxio stuff
10023 	 */
10024 	if (pptr->port_mpxio) {
10025 		(void) mdi_phci_unregister(pptr->port_dip, 0);
10026 		pptr->port_mpxio--;
10027 	}
10028 
10029 	/* undo SCSI HBA setup */
10030 	if (hba_attached) {
10031 		(void) scsi_hba_detach(pptr->port_dip);
10032 	}
10033 	if (pptr->port_tran != NULL) {
10034 		scsi_hba_tran_free(pptr->port_tran);
10035 	}
10036 
10037 	mutex_enter(&fcp_global_mutex);
10038 
10039 	/*
10040 	 * We check soft_state_linked, because it is incremented right before
10041 	 * we call increment fcp_watchdog_init.	 Therefore, we know if
10042 	 * soft_state_linked is still FALSE, we do not want to decrement
10043 	 * fcp_watchdog_init or possibly call untimeout.
10044 	 */
10045 
10046 	if (soft_state_linked) {
10047 		if (--fcp_watchdog_init == 0) {
10048 			timeout_id_t	tid = fcp_watchdog_id;
10049 
10050 			mutex_exit(&fcp_global_mutex);
10051 			(void) untimeout(tid);
10052 		} else {
10053 			mutex_exit(&fcp_global_mutex);
10054 		}
10055 	} else {
10056 		mutex_exit(&fcp_global_mutex);
10057 	}
10058 
10059 	if (mutex_initted) {
10060 		mutex_destroy(&pptr->port_mutex);
10061 		mutex_destroy(&pptr->port_pkt_mutex);
10062 	}
10063 
10064 	if (tmp_list != NULL) {
10065 		kmem_free(tmp_list, sizeof (fc_portmap_t) * alloc_cnt);
10066 	}
10067 
10068 	/* this makes pptr invalid */
10069 	ddi_soft_state_free(fcp_softstate, instance);
10070 
10071 	return (DDI_FAILURE);
10072 }
10073 
10074 
10075 static int
10076 fcp_handle_port_detach(struct fcp_port *pptr, int flag, int instance)
10077 {
10078 	int count = 0;
10079 
10080 	mutex_enter(&pptr->port_mutex);
10081 
10082 	/*
10083 	 * if the port is powered down or suspended, nothing else
10084 	 * to do; just return.
10085 	 */
10086 	if (flag != FCP_STATE_DETACHING) {
10087 		if (pptr->port_state & (FCP_STATE_POWER_DOWN |
10088 		    FCP_STATE_SUSPENDED)) {
10089 			pptr->port_state |= flag;
10090 			mutex_exit(&pptr->port_mutex);
10091 			return (FC_SUCCESS);
10092 		}
10093 	}
10094 
10095 	if (pptr->port_state & FCP_STATE_IN_MDI) {
10096 		mutex_exit(&pptr->port_mutex);
10097 		return (FC_FAILURE);
10098 	}
10099 
10100 	FCP_TRACE(fcp_logq, pptr->port_instbuf,
10101 	    fcp_trace, FCP_BUF_LEVEL_2, 0,
10102 	    "fcp_handle_port_detach: port is detaching");
10103 
10104 	pptr->port_state |= flag;
10105 
10106 	/*
10107 	 * Wait for any ongoing reconfig/ipkt to complete, that
10108 	 * ensures the freeing to targets/luns is safe.
10109 	 * No more ref to this port should happen from statec/ioctl
10110 	 * after that as it was removed from the global port list.
10111 	 */
10112 	while (pptr->port_tmp_cnt || pptr->port_ipkt_cnt ||
10113 	    (pptr->port_state & FCP_STATE_IN_WATCHDOG)) {
10114 		/*
10115 		 * Let's give sufficient time for reconfig/ipkt
10116 		 * to complete.
10117 		 */
10118 		if (count++ >= FCP_ICMD_DEADLINE) {
10119 			break;
10120 		}
10121 		mutex_exit(&pptr->port_mutex);
10122 		delay(drv_usectohz(1000000));
10123 		mutex_enter(&pptr->port_mutex);
10124 	}
10125 
10126 	/*
10127 	 * if the driver is still busy then fail to
10128 	 * suspend/power down.
10129 	 */
10130 	if (pptr->port_tmp_cnt || pptr->port_ipkt_cnt ||
10131 	    (pptr->port_state & FCP_STATE_IN_WATCHDOG)) {
10132 		pptr->port_state &= ~flag;
10133 		mutex_exit(&pptr->port_mutex);
10134 		return (FC_FAILURE);
10135 	}
10136 
10137 	if (flag == FCP_STATE_DETACHING) {
10138 		pptr = fcp_soft_state_unlink(pptr);
10139 		ASSERT(pptr != NULL);
10140 	}
10141 
10142 	pptr->port_link_cnt++;
10143 	pptr->port_state |= FCP_STATE_OFFLINE;
10144 	pptr->port_state &= ~(FCP_STATE_ONLINING | FCP_STATE_ONLINE);
10145 
10146 	fcp_update_state(pptr, (FCP_LUN_BUSY | FCP_LUN_MARK),
10147 	    FCP_CAUSE_LINK_DOWN);
10148 	mutex_exit(&pptr->port_mutex);
10149 
10150 	/* kill watch dog timer if we're the last */
10151 	mutex_enter(&fcp_global_mutex);
10152 	if (--fcp_watchdog_init == 0) {
10153 		timeout_id_t	tid = fcp_watchdog_id;
10154 		mutex_exit(&fcp_global_mutex);
10155 		(void) untimeout(tid);
10156 	} else {
10157 		mutex_exit(&fcp_global_mutex);
10158 	}
10159 
10160 	/* clean up the port structures */
10161 	if (flag == FCP_STATE_DETACHING) {
10162 		fcp_cleanup_port(pptr, instance);
10163 	}
10164 
10165 	return (FC_SUCCESS);
10166 }
10167 
10168 
10169 static void
10170 fcp_cleanup_port(struct fcp_port *pptr, int instance)
10171 {
10172 	ASSERT(pptr != NULL);
10173 
10174 	/* unbind and free event set */
10175 	if (pptr->port_ndi_event_hdl) {
10176 		(void) ndi_event_unbind_set(pptr->port_ndi_event_hdl,
10177 		    &pptr->port_ndi_events, NDI_SLEEP);
10178 		(void) ndi_event_free_hdl(pptr->port_ndi_event_hdl);
10179 	}
10180 
10181 	if (pptr->port_ndi_event_defs) {
10182 		(void) kmem_free(pptr->port_ndi_event_defs,
10183 		    sizeof (fcp_ndi_event_defs));
10184 	}
10185 
10186 	/* free the lun/target structures and devinfos */
10187 	fcp_free_targets(pptr);
10188 
10189 	/*
10190 	 * Clean up mpxio stuff
10191 	 */
10192 	if (pptr->port_mpxio) {
10193 		(void) mdi_phci_unregister(pptr->port_dip, 0);
10194 		pptr->port_mpxio--;
10195 	}
10196 
10197 	/* clean up SCSA stuff */
10198 	(void) scsi_hba_detach(pptr->port_dip);
10199 	if (pptr->port_tran != NULL) {
10200 		scsi_hba_tran_free(pptr->port_tran);
10201 	}
10202 
10203 #ifdef	KSTATS_CODE
10204 	/* clean up kstats */
10205 	if (pptr->fcp_ksp != NULL) {
10206 		kstat_delete(pptr->fcp_ksp);
10207 	}
10208 #endif
10209 
10210 	/* clean up soft state mutexes/condition variables */
10211 	mutex_destroy(&pptr->port_mutex);
10212 	mutex_destroy(&pptr->port_pkt_mutex);
10213 
10214 	/* all done with soft state */
10215 	ddi_soft_state_free(fcp_softstate, instance);
10216 }
10217 
10218 /*
10219  *     Function: fcp_kmem_cache_constructor
10220  *
10221  *  Description: This function allocates and initializes the resources required
10222  *		 to build a scsi_pkt structure the target driver.  The result
10223  *		 of the allocation and initialization will be cached in the
10224  *		 memory cache.	As DMA resources may be allocated here, that
10225  *		 means DMA resources will be tied up in the cache manager.
10226  *		 This is a tradeoff that has been made for performance reasons.
10227  *
10228  *     Argument: *buf		Memory to preinitialize.
10229  *		 *arg		FCP port structure (fcp_port).
10230  *		 kmflags	Value passed to kmem_cache_alloc() and
10231  *				propagated to the constructor.
10232  *
10233  * Return Value: 0	Allocation/Initialization was successful.
10234  *		 -1	Allocation or Initialization failed.
10235  *
10236  *
10237  * If the returned value is 0, the buffer is initialized like this:
10238  *
10239  *		    +================================+
10240  *	     +----> |	      struct scsi_pkt	     |
10241  *	     |	    |				     |
10242  *	     | +--- | pkt_ha_private		     |
10243  *	     | |    |				     |
10244  *	     | |    +================================+
10245  *	     | |
10246  *	     | |    +================================+
10247  *	     | +--> |	    struct fcp_pkt	     | <---------+
10248  *	     |	    |				     |		 |
10249  *	     +----- | cmd_pkt			     |		 |
10250  *		    |			  cmd_fp_pkt | ---+	 |
10251  *	  +-------->| cmd_fcp_rsp[]		     |	  |	 |
10252  *	  |    +--->| cmd_fcp_cmd[]		     |	  |	 |
10253  *	  |    |    |--------------------------------|	  |	 |
10254  *	  |    |    |	      struct fc_packet	     | <--+	 |
10255  *	  |    |    |				     |		 |
10256  *	  |    |    |		     pkt_ulp_private | ----------+
10257  *	  |    |    |		     pkt_fca_private | -----+
10258  *	  |    |    |		     pkt_data_cookie | ---+ |
10259  *	  |    |    | pkt_cmdlen		     |	  | |
10260  *	  |    |(a) | pkt_rsplen		     |	  | |
10261  *	  |    +----| .......... pkt_cmd ........... | ---|-|---------------+
10262  *	  |	(b) |		      pkt_cmd_cookie | ---|-|----------+    |
10263  *	  +---------| .......... pkt_resp .......... | ---|-|------+   |    |
10264  *		    |		     pkt_resp_cookie | ---|-|--+   |   |    |
10265  *		    | pkt_cmd_dma		     |	  | |  |   |   |    |
10266  *		    | pkt_cmd_acc		     |	  | |  |   |   |    |
10267  *		    +================================+	  | |  |   |   |    |
10268  *		    |	      dma_cookies	     | <--+ |  |   |   |    |
10269  *		    |				     |	    |  |   |   |    |
10270  *		    +================================+	    |  |   |   |    |
10271  *		    |	      fca_private	     | <----+  |   |   |    |
10272  *		    |				     |	       |   |   |    |
10273  *		    +================================+	       |   |   |    |
10274  *							       |   |   |    |
10275  *							       |   |   |    |
10276  *		    +================================+	 (d)   |   |   |    |
10277  *		    |	     fcp_resp cookies	     | <-------+   |   |    |
10278  *		    |				     |		   |   |    |
10279  *		    +================================+		   |   |    |
10280  *								   |   |    |
10281  *		    +================================+	 (d)	   |   |    |
10282  *		    |		fcp_resp	     | <-----------+   |    |
10283  *		    |	(DMA resources associated)   |		       |    |
10284  *		    +================================+		       |    |
10285  *								       |    |
10286  *								       |    |
10287  *								       |    |
10288  *		    +================================+	 (c)	       |    |
10289  *		    |	     fcp_cmd cookies	     | <---------------+    |
10290  *		    |				     |			    |
10291  *		    +================================+			    |
10292  *									    |
10293  *		    +================================+	 (c)		    |
10294  *		    |		 fcp_cmd	     | <--------------------+
10295  *		    |	(DMA resources associated)   |
10296  *		    +================================+
10297  *
10298  * (a) Only if DMA is NOT used for the FCP_CMD buffer.
10299  * (b) Only if DMA is NOT used for the FCP_RESP buffer
10300  * (c) Only if DMA is used for the FCP_CMD buffer.
10301  * (d) Only if DMA is used for the FCP_RESP buffer
10302  */
10303 static int
10304 fcp_kmem_cache_constructor(struct scsi_pkt *pkt, scsi_hba_tran_t *tran,
10305     int kmflags)
10306 {
10307 	struct fcp_pkt	*cmd;
10308 	struct fcp_port	*pptr;
10309 	fc_packet_t	*fpkt;
10310 
10311 	pptr = (struct fcp_port *)tran->tran_hba_private;
10312 	cmd = (struct fcp_pkt *)pkt->pkt_ha_private;
10313 	bzero(cmd, tran->tran_hba_len);
10314 
10315 	cmd->cmd_pkt = pkt;
10316 	pkt->pkt_cdbp = cmd->cmd_fcp_cmd.fcp_cdb;
10317 	fpkt = (fc_packet_t *)&cmd->cmd_fc_packet;
10318 	cmd->cmd_fp_pkt = fpkt;
10319 
10320 	cmd->cmd_pkt->pkt_ha_private = (opaque_t)cmd;
10321 	cmd->cmd_fp_pkt->pkt_ulp_private = (opaque_t)cmd;
10322 	cmd->cmd_fp_pkt->pkt_fca_private = (opaque_t)((caddr_t)cmd +
10323 	    sizeof (struct fcp_pkt) + pptr->port_dmacookie_sz);
10324 
10325 	fpkt->pkt_data_cookie = (ddi_dma_cookie_t *)((caddr_t)cmd +
10326 	    sizeof (struct fcp_pkt));
10327 
10328 	fpkt->pkt_cmdlen = sizeof (struct fcp_cmd);
10329 	fpkt->pkt_rsplen = FCP_MAX_RSP_IU_SIZE;
10330 
10331 	if (pptr->port_fcp_dma == FC_NO_DVMA_SPACE) {
10332 		/*
10333 		 * The underlying HBA doesn't want to DMA the fcp_cmd or
10334 		 * fcp_resp.  The transfer of information will be done by
10335 		 * bcopy.
10336 		 * The naming of the flags (that is actually a value) is
10337 		 * unfortunate.	 FC_NO_DVMA_SPACE doesn't mean "NO VIRTUAL
10338 		 * DMA" but instead "NO DMA".
10339 		 */
10340 		fpkt->pkt_resp_acc = fpkt->pkt_cmd_acc = NULL;
10341 		fpkt->pkt_cmd = (caddr_t)&cmd->cmd_fcp_cmd;
10342 		fpkt->pkt_resp = cmd->cmd_fcp_rsp;
10343 	} else {
10344 		/*
10345 		 * The underlying HBA will dma the fcp_cmd buffer and fcp_resp
10346 		 * buffer.  A buffer is allocated for each one the ddi_dma_*
10347 		 * interfaces.
10348 		 */
10349 		if (fcp_alloc_cmd_resp(pptr, fpkt, kmflags) != FC_SUCCESS) {
10350 			return (-1);
10351 		}
10352 	}
10353 
10354 	return (0);
10355 }
10356 
10357 /*
10358  *     Function: fcp_kmem_cache_destructor
10359  *
10360  *  Description: Called by the destructor of the cache managed by SCSA.
10361  *		 All the resources pre-allocated in fcp_pkt_constructor
10362  *		 and the data also pre-initialized in fcp_pkt_constructor
10363  *		 are freed and uninitialized here.
10364  *
10365  *     Argument: *buf		Memory to uninitialize.
10366  *		 *arg		FCP port structure (fcp_port).
10367  *
10368  * Return Value: None
10369  *
10370  *	Context: kernel
10371  */
10372 static void
10373 fcp_kmem_cache_destructor(struct scsi_pkt *pkt, scsi_hba_tran_t *tran)
10374 {
10375 	struct fcp_pkt	*cmd;
10376 	struct fcp_port	*pptr;
10377 
10378 	pptr = (struct fcp_port *)(tran->tran_hba_private);
10379 	cmd = pkt->pkt_ha_private;
10380 
10381 	if (pptr->port_fcp_dma != FC_NO_DVMA_SPACE) {
10382 		/*
10383 		 * If DMA was used to transfer the FCP_CMD and FCP_RESP, the
10384 		 * buffer and DMA resources allocated to do so are released.
10385 		 */
10386 		fcp_free_cmd_resp(pptr, cmd->cmd_fp_pkt);
10387 	}
10388 }
10389 
10390 /*
10391  *     Function: fcp_alloc_cmd_resp
10392  *
10393  *  Description: This function allocated an FCP_CMD and FCP_RESP buffer that
10394  *		 will be DMAed by the HBA.  The buffer is allocated applying
10395  *		 the DMA requirements for the HBA.  The buffers allocated will
10396  *		 also be bound.	 DMA resources are allocated in the process.
10397  *		 They will be released by fcp_free_cmd_resp().
10398  *
10399  *     Argument: *pptr	FCP port.
10400  *		 *fpkt	fc packet for which the cmd and resp packet should be
10401  *			allocated.
10402  *		 flags	Allocation flags.
10403  *
10404  * Return Value: FC_FAILURE
10405  *		 FC_SUCCESS
10406  *
10407  *	Context: User or Kernel context only if flags == KM_SLEEP.
10408  *		 Interrupt context if the KM_SLEEP is not specified.
10409  */
10410 static int
10411 fcp_alloc_cmd_resp(struct fcp_port *pptr, fc_packet_t *fpkt, int flags)
10412 {
10413 	int			rval;
10414 	int			cmd_len;
10415 	int			resp_len;
10416 	ulong_t			real_len;
10417 	int			(*cb) (caddr_t);
10418 	ddi_dma_cookie_t	pkt_cookie;
10419 	ddi_dma_cookie_t	*cp;
10420 	uint32_t		cnt;
10421 
10422 	cb = (flags == KM_SLEEP) ? DDI_DMA_SLEEP : DDI_DMA_DONTWAIT;
10423 
10424 	cmd_len = fpkt->pkt_cmdlen;
10425 	resp_len = fpkt->pkt_rsplen;
10426 
10427 	ASSERT(fpkt->pkt_cmd_dma == NULL);
10428 
10429 	/* Allocation of a DMA handle used in subsequent calls. */
10430 	if (ddi_dma_alloc_handle(pptr->port_dip, &pptr->port_cmd_dma_attr,
10431 	    cb, NULL, &fpkt->pkt_cmd_dma) != DDI_SUCCESS) {
10432 		return (FC_FAILURE);
10433 	}
10434 
10435 	/* A buffer is allocated that satisfies the DMA requirements. */
10436 	rval = ddi_dma_mem_alloc(fpkt->pkt_cmd_dma, cmd_len,
10437 	    &pptr->port_dma_acc_attr, DDI_DMA_CONSISTENT, cb, NULL,
10438 	    (caddr_t *)&fpkt->pkt_cmd, &real_len, &fpkt->pkt_cmd_acc);
10439 
10440 	if (rval != DDI_SUCCESS) {
10441 		ddi_dma_free_handle(&fpkt->pkt_cmd_dma);
10442 		return (FC_FAILURE);
10443 	}
10444 
10445 	if (real_len < cmd_len) {
10446 		ddi_dma_mem_free(&fpkt->pkt_cmd_acc);
10447 		ddi_dma_free_handle(&fpkt->pkt_cmd_dma);
10448 		return (FC_FAILURE);
10449 	}
10450 
10451 	/* The buffer allocated is DMA bound. */
10452 	rval = ddi_dma_addr_bind_handle(fpkt->pkt_cmd_dma, NULL,
10453 	    fpkt->pkt_cmd, real_len, DDI_DMA_WRITE | DDI_DMA_CONSISTENT,
10454 	    cb, NULL, &pkt_cookie, &fpkt->pkt_cmd_cookie_cnt);
10455 
10456 	if (rval != DDI_DMA_MAPPED) {
10457 		ddi_dma_mem_free(&fpkt->pkt_cmd_acc);
10458 		ddi_dma_free_handle(&fpkt->pkt_cmd_dma);
10459 		return (FC_FAILURE);
10460 	}
10461 
10462 	if (fpkt->pkt_cmd_cookie_cnt >
10463 	    pptr->port_cmd_dma_attr.dma_attr_sgllen) {
10464 		(void) ddi_dma_unbind_handle(fpkt->pkt_cmd_dma);
10465 		ddi_dma_mem_free(&fpkt->pkt_cmd_acc);
10466 		ddi_dma_free_handle(&fpkt->pkt_cmd_dma);
10467 		return (FC_FAILURE);
10468 	}
10469 
10470 	ASSERT(fpkt->pkt_cmd_cookie_cnt != 0);
10471 
10472 	/*
10473 	 * The buffer where the scatter/gather list is going to be built is
10474 	 * allocated.
10475 	 */
10476 	cp = fpkt->pkt_cmd_cookie = (ddi_dma_cookie_t *)kmem_alloc(
10477 	    fpkt->pkt_cmd_cookie_cnt * sizeof (pkt_cookie),
10478 	    KM_NOSLEEP);
10479 
10480 	if (cp == NULL) {
10481 		(void) ddi_dma_unbind_handle(fpkt->pkt_cmd_dma);
10482 		ddi_dma_mem_free(&fpkt->pkt_cmd_acc);
10483 		ddi_dma_free_handle(&fpkt->pkt_cmd_dma);
10484 		return (FC_FAILURE);
10485 	}
10486 
10487 	/*
10488 	 * The scatter/gather list for the buffer we just allocated is built
10489 	 * here.
10490 	 */
10491 	*cp = pkt_cookie;
10492 	cp++;
10493 
10494 	for (cnt = 1; cnt < fpkt->pkt_cmd_cookie_cnt; cnt++, cp++) {
10495 		ddi_dma_nextcookie(fpkt->pkt_cmd_dma,
10496 		    &pkt_cookie);
10497 		*cp = pkt_cookie;
10498 	}
10499 
10500 	ASSERT(fpkt->pkt_resp_dma == NULL);
10501 	if (ddi_dma_alloc_handle(pptr->port_dip, &pptr->port_resp_dma_attr,
10502 	    cb, NULL, &fpkt->pkt_resp_dma) != DDI_SUCCESS) {
10503 		(void) ddi_dma_unbind_handle(fpkt->pkt_cmd_dma);
10504 		ddi_dma_mem_free(&fpkt->pkt_cmd_acc);
10505 		ddi_dma_free_handle(&fpkt->pkt_cmd_dma);
10506 		return (FC_FAILURE);
10507 	}
10508 
10509 	rval = ddi_dma_mem_alloc(fpkt->pkt_resp_dma, resp_len,
10510 	    &pptr->port_dma_acc_attr, DDI_DMA_CONSISTENT, cb, NULL,
10511 	    (caddr_t *)&fpkt->pkt_resp, &real_len,
10512 	    &fpkt->pkt_resp_acc);
10513 
10514 	if (rval != DDI_SUCCESS) {
10515 		ddi_dma_free_handle(&fpkt->pkt_resp_dma);
10516 		(void) ddi_dma_unbind_handle(fpkt->pkt_cmd_dma);
10517 		ddi_dma_mem_free(&fpkt->pkt_cmd_acc);
10518 		ddi_dma_free_handle(&fpkt->pkt_cmd_dma);
10519 		kmem_free(fpkt->pkt_cmd_cookie,
10520 		    fpkt->pkt_cmd_cookie_cnt * sizeof (pkt_cookie));
10521 		return (FC_FAILURE);
10522 	}
10523 
10524 	if (real_len < resp_len) {
10525 		ddi_dma_mem_free(&fpkt->pkt_resp_acc);
10526 		ddi_dma_free_handle(&fpkt->pkt_resp_dma);
10527 		(void) ddi_dma_unbind_handle(fpkt->pkt_cmd_dma);
10528 		ddi_dma_mem_free(&fpkt->pkt_cmd_acc);
10529 		ddi_dma_free_handle(&fpkt->pkt_cmd_dma);
10530 		kmem_free(fpkt->pkt_cmd_cookie,
10531 		    fpkt->pkt_cmd_cookie_cnt * sizeof (pkt_cookie));
10532 		return (FC_FAILURE);
10533 	}
10534 
10535 	rval = ddi_dma_addr_bind_handle(fpkt->pkt_resp_dma, NULL,
10536 	    fpkt->pkt_resp, real_len, DDI_DMA_READ | DDI_DMA_CONSISTENT,
10537 	    cb, NULL, &pkt_cookie, &fpkt->pkt_resp_cookie_cnt);
10538 
10539 	if (rval != DDI_DMA_MAPPED) {
10540 		ddi_dma_mem_free(&fpkt->pkt_resp_acc);
10541 		ddi_dma_free_handle(&fpkt->pkt_resp_dma);
10542 		(void) ddi_dma_unbind_handle(fpkt->pkt_cmd_dma);
10543 		ddi_dma_mem_free(&fpkt->pkt_cmd_acc);
10544 		ddi_dma_free_handle(&fpkt->pkt_cmd_dma);
10545 		kmem_free(fpkt->pkt_cmd_cookie,
10546 		    fpkt->pkt_cmd_cookie_cnt * sizeof (pkt_cookie));
10547 		return (FC_FAILURE);
10548 	}
10549 
10550 	if (fpkt->pkt_resp_cookie_cnt >
10551 	    pptr->port_resp_dma_attr.dma_attr_sgllen) {
10552 		ddi_dma_mem_free(&fpkt->pkt_resp_acc);
10553 		ddi_dma_free_handle(&fpkt->pkt_resp_dma);
10554 		(void) ddi_dma_unbind_handle(fpkt->pkt_cmd_dma);
10555 		ddi_dma_mem_free(&fpkt->pkt_cmd_acc);
10556 		ddi_dma_free_handle(&fpkt->pkt_cmd_dma);
10557 		kmem_free(fpkt->pkt_cmd_cookie,
10558 		    fpkt->pkt_cmd_cookie_cnt * sizeof (pkt_cookie));
10559 		return (FC_FAILURE);
10560 	}
10561 
10562 	ASSERT(fpkt->pkt_resp_cookie_cnt != 0);
10563 
10564 	cp = fpkt->pkt_resp_cookie = (ddi_dma_cookie_t *)kmem_alloc(
10565 	    fpkt->pkt_resp_cookie_cnt * sizeof (pkt_cookie),
10566 	    KM_NOSLEEP);
10567 
10568 	if (cp == NULL) {
10569 		ddi_dma_mem_free(&fpkt->pkt_resp_acc);
10570 		ddi_dma_free_handle(&fpkt->pkt_resp_dma);
10571 		(void) ddi_dma_unbind_handle(fpkt->pkt_cmd_dma);
10572 		ddi_dma_mem_free(&fpkt->pkt_cmd_acc);
10573 		ddi_dma_free_handle(&fpkt->pkt_cmd_dma);
10574 		kmem_free(fpkt->pkt_cmd_cookie,
10575 		    fpkt->pkt_cmd_cookie_cnt * sizeof (pkt_cookie));
10576 		return (FC_FAILURE);
10577 	}
10578 
10579 	*cp = pkt_cookie;
10580 	cp++;
10581 
10582 	for (cnt = 1; cnt < fpkt->pkt_resp_cookie_cnt; cnt++, cp++) {
10583 		ddi_dma_nextcookie(fpkt->pkt_resp_dma,
10584 		    &pkt_cookie);
10585 		*cp = pkt_cookie;
10586 	}
10587 
10588 	return (FC_SUCCESS);
10589 }
10590 
10591 /*
10592  *     Function: fcp_free_cmd_resp
10593  *
10594  *  Description: This function releases the FCP_CMD and FCP_RESP buffer
10595  *		 allocated by fcp_alloc_cmd_resp() and all the resources
10596  *		 associated with them.	That includes the DMA resources and the
10597  *		 buffer allocated for the cookies of each one of them.
10598  *
10599  *     Argument: *pptr		FCP port context.
10600  *		 *fpkt		fc packet containing the cmd and resp packet
10601  *				to be released.
10602  *
10603  * Return Value: None
10604  *
10605  *	Context: Interrupt, User and Kernel context.
10606  */
10607 /* ARGSUSED */
10608 static void
10609 fcp_free_cmd_resp(struct fcp_port *pptr, fc_packet_t *fpkt)
10610 {
10611 	ASSERT(fpkt->pkt_resp_dma != NULL && fpkt->pkt_cmd_dma != NULL);
10612 
10613 	if (fpkt->pkt_resp_dma) {
10614 		(void) ddi_dma_unbind_handle(fpkt->pkt_resp_dma);
10615 		ddi_dma_mem_free(&fpkt->pkt_resp_acc);
10616 		ddi_dma_free_handle(&fpkt->pkt_resp_dma);
10617 	}
10618 
10619 	if (fpkt->pkt_resp_cookie) {
10620 		kmem_free(fpkt->pkt_resp_cookie,
10621 		    fpkt->pkt_resp_cookie_cnt * sizeof (ddi_dma_cookie_t));
10622 		fpkt->pkt_resp_cookie = NULL;
10623 	}
10624 
10625 	if (fpkt->pkt_cmd_dma) {
10626 		(void) ddi_dma_unbind_handle(fpkt->pkt_cmd_dma);
10627 		ddi_dma_mem_free(&fpkt->pkt_cmd_acc);
10628 		ddi_dma_free_handle(&fpkt->pkt_cmd_dma);
10629 	}
10630 
10631 	if (fpkt->pkt_cmd_cookie) {
10632 		kmem_free(fpkt->pkt_cmd_cookie,
10633 		    fpkt->pkt_cmd_cookie_cnt * sizeof (ddi_dma_cookie_t));
10634 		fpkt->pkt_cmd_cookie = NULL;
10635 	}
10636 }
10637 
10638 
10639 /*
10640  * called by the transport to do our own target initialization
10641  *
10642  * can acquire and release the global mutex
10643  */
10644 /* ARGSUSED */
10645 static int
10646 fcp_phys_tgt_init(dev_info_t *hba_dip, dev_info_t *tgt_dip,
10647     scsi_hba_tran_t *hba_tran, struct scsi_device *sd)
10648 {
10649 	int			*words;
10650 	uchar_t			*bytes;
10651 	uint_t			nbytes;
10652 	uint_t			nwords;
10653 	struct fcp_tgt	*ptgt;
10654 	struct fcp_lun	*plun;
10655 	struct fcp_port	*pptr = (struct fcp_port *)
10656 	    hba_tran->tran_hba_private;
10657 
10658 	ASSERT(pptr != NULL);
10659 
10660 	FCP_DTRACE(fcp_logq, pptr->port_instbuf, fcp_trace,
10661 	    FCP_BUF_LEVEL_8, 0,
10662 	    "fcp_phys_tgt_init: called for %s (instance %d)",
10663 	    ddi_get_name(tgt_dip), ddi_get_instance(tgt_dip));
10664 
10665 	/* get our port WWN property */
10666 	bytes = NULL;
10667 	if ((ddi_prop_lookup_byte_array(DDI_DEV_T_ANY, tgt_dip,
10668 	    DDI_PROP_DONTPASS | DDI_PROP_NOTPROM, PORT_WWN_PROP, &bytes,
10669 	    &nbytes) != DDI_PROP_SUCCESS) || nbytes != FC_WWN_SIZE) {
10670 		/* no port WWN property */
10671 		FCP_DTRACE(fcp_logq, pptr->port_instbuf, fcp_trace,
10672 		    FCP_BUF_LEVEL_8, 0,
10673 		    "fcp_phys_tgt_init: Returning DDI_NOT_WELL_FORMED"
10674 		    " for %s (instance %d): bytes=%p nbytes=%x",
10675 		    ddi_get_name(tgt_dip), ddi_get_instance(tgt_dip), bytes,
10676 		    nbytes);
10677 
10678 		if (bytes != NULL) {
10679 			ddi_prop_free(bytes);
10680 		}
10681 
10682 		return (DDI_NOT_WELL_FORMED);
10683 	}
10684 
10685 	words = NULL;
10686 	if (ddi_prop_lookup_int_array(DDI_DEV_T_ANY, tgt_dip,
10687 	    DDI_PROP_DONTPASS | DDI_PROP_NOTPROM,
10688 	    LUN_PROP, &words, &nwords) != DDI_PROP_SUCCESS) {
10689 		ASSERT(bytes != NULL);
10690 
10691 		FCP_DTRACE(fcp_logq, pptr->port_instbuf, fcp_trace,
10692 		    FCP_BUF_LEVEL_8, 0,
10693 		    "fcp_phys_tgt_init: Returning DDI_FAILURE:lun"
10694 		    " for %s (instance %d)", ddi_get_name(tgt_dip),
10695 		    ddi_get_instance(tgt_dip));
10696 
10697 		ddi_prop_free(bytes);
10698 
10699 		return (DDI_NOT_WELL_FORMED);
10700 	}
10701 
10702 	if (nwords == 0) {
10703 		ddi_prop_free(bytes);
10704 		ddi_prop_free(words);
10705 		return (DDI_NOT_WELL_FORMED);
10706 	}
10707 
10708 	ASSERT(bytes != NULL && words != NULL);
10709 
10710 	mutex_enter(&pptr->port_mutex);
10711 	if ((plun = fcp_lookup_lun(pptr, bytes, *words)) == NULL) {
10712 		mutex_exit(&pptr->port_mutex);
10713 		FCP_DTRACE(fcp_logq, pptr->port_instbuf, fcp_trace,
10714 		    FCP_BUF_LEVEL_8, 0,
10715 		    "fcp_phys_tgt_init: Returning DDI_FAILURE: No Lun"
10716 		    " for %s (instance %d)", ddi_get_name(tgt_dip),
10717 		    ddi_get_instance(tgt_dip));
10718 
10719 		ddi_prop_free(bytes);
10720 		ddi_prop_free(words);
10721 
10722 		return (DDI_FAILURE);
10723 	}
10724 
10725 	ASSERT(bcmp(plun->lun_tgt->tgt_port_wwn.raw_wwn, bytes,
10726 	    FC_WWN_SIZE) == 0);
10727 	ASSERT(plun->lun_num == (uint16_t)*words);
10728 
10729 	ddi_prop_free(bytes);
10730 	ddi_prop_free(words);
10731 
10732 	ptgt = plun->lun_tgt;
10733 
10734 	mutex_enter(&ptgt->tgt_mutex);
10735 	plun->lun_tgt_count++;
10736 	scsi_device_hba_private_set(sd, plun);
10737 	plun->lun_state |= FCP_SCSI_LUN_TGT_INIT;
10738 	plun->lun_tran = hba_tran;
10739 	mutex_exit(&ptgt->tgt_mutex);
10740 	mutex_exit(&pptr->port_mutex);
10741 
10742 	return (DDI_SUCCESS);
10743 }
10744 
10745 /*ARGSUSED*/
10746 static int
10747 fcp_virt_tgt_init(dev_info_t *hba_dip, dev_info_t *tgt_dip,
10748     scsi_hba_tran_t *hba_tran, struct scsi_device *sd)
10749 {
10750 	int			words;
10751 	uchar_t			*bytes;
10752 	uint_t			nbytes;
10753 	struct fcp_tgt	*ptgt;
10754 	struct fcp_lun	*plun;
10755 	struct fcp_port	*pptr = (struct fcp_port *)
10756 	    hba_tran->tran_hba_private;
10757 	child_info_t		*cip;
10758 
10759 	ASSERT(pptr != NULL);
10760 
10761 	FCP_DTRACE(fcp_logq, pptr->port_instbuf,
10762 	    fcp_trace, FCP_BUF_LEVEL_8, 0,
10763 	    "fcp_virt_tgt_init: called for %s (instance %d) (hba_dip %p),"
10764 	    " (tgt_dip %p)", ddi_get_name(tgt_dip),
10765 	    ddi_get_instance(tgt_dip), hba_dip, tgt_dip);
10766 
10767 	cip = (child_info_t *)sd->sd_pathinfo;
10768 	if (cip == NULL) {
10769 		FCP_DTRACE(fcp_logq, pptr->port_instbuf,
10770 		    fcp_trace, FCP_BUF_LEVEL_8, 0,
10771 		    "fcp_virt_tgt_init: Returning DDI_NOT_WELL_FORMED"
10772 		    " for %s (instance %d)", ddi_get_name(tgt_dip),
10773 		    ddi_get_instance(tgt_dip));
10774 
10775 		return (DDI_NOT_WELL_FORMED);
10776 	}
10777 
10778 	/* get our port WWN property */
10779 	bytes = NULL;
10780 	if ((mdi_prop_lookup_byte_array(PIP(cip), PORT_WWN_PROP, &bytes,
10781 	    &nbytes) != DDI_PROP_SUCCESS) || nbytes != FC_WWN_SIZE) {
10782 		if (bytes) {
10783 			(void) mdi_prop_free(bytes);
10784 		}
10785 		return (DDI_NOT_WELL_FORMED);
10786 	}
10787 
10788 	words = 0;
10789 	if (mdi_prop_lookup_int(PIP(cip), LUN_PROP, &words) !=
10790 	    DDI_PROP_SUCCESS) {
10791 		ASSERT(bytes != NULL);
10792 
10793 		FCP_DTRACE(fcp_logq, pptr->port_instbuf,
10794 		    fcp_trace, FCP_BUF_LEVEL_8, 0,
10795 		    "fcp_virt_tgt_init: Returning DDI_FAILURE:lun"
10796 		    " for %s (instance %d)", ddi_get_name(tgt_dip),
10797 		    ddi_get_instance(tgt_dip));
10798 
10799 		(void) mdi_prop_free(bytes);
10800 		return (DDI_NOT_WELL_FORMED);
10801 	}
10802 
10803 	ASSERT(bytes != NULL);
10804 
10805 	mutex_enter(&pptr->port_mutex);
10806 	if ((plun = fcp_lookup_lun(pptr, bytes, words)) == NULL) {
10807 		mutex_exit(&pptr->port_mutex);
10808 		FCP_DTRACE(fcp_logq, pptr->port_instbuf,
10809 		    fcp_trace, FCP_BUF_LEVEL_8, 0,
10810 		    "fcp_virt_tgt_init: Returning DDI_FAILURE: No Lun"
10811 		    " for %s (instance %d)", ddi_get_name(tgt_dip),
10812 		    ddi_get_instance(tgt_dip));
10813 
10814 		(void) mdi_prop_free(bytes);
10815 		(void) mdi_prop_free(&words);
10816 
10817 		return (DDI_FAILURE);
10818 	}
10819 
10820 	ASSERT(bcmp(plun->lun_tgt->tgt_port_wwn.raw_wwn, bytes,
10821 	    FC_WWN_SIZE) == 0);
10822 	ASSERT(plun->lun_num == (uint16_t)words);
10823 
10824 	(void) mdi_prop_free(bytes);
10825 	(void) mdi_prop_free(&words);
10826 
10827 	ptgt = plun->lun_tgt;
10828 
10829 	mutex_enter(&ptgt->tgt_mutex);
10830 	plun->lun_tgt_count++;
10831 	scsi_device_hba_private_set(sd, plun);
10832 	plun->lun_state |= FCP_SCSI_LUN_TGT_INIT;
10833 	plun->lun_tran = hba_tran;
10834 	mutex_exit(&ptgt->tgt_mutex);
10835 	mutex_exit(&pptr->port_mutex);
10836 
10837 	return (DDI_SUCCESS);
10838 }
10839 
10840 
10841 /*
10842  * called by the transport to do our own target initialization
10843  *
10844  * can acquire and release the global mutex
10845  */
10846 /* ARGSUSED */
10847 static int
10848 fcp_scsi_tgt_init(dev_info_t *hba_dip, dev_info_t *tgt_dip,
10849     scsi_hba_tran_t *hba_tran, struct scsi_device *sd)
10850 {
10851 	struct fcp_port	*pptr = (struct fcp_port *)
10852 	    hba_tran->tran_hba_private;
10853 	int			rval;
10854 
10855 	ASSERT(pptr != NULL);
10856 
10857 	/*
10858 	 * Child node is getting initialized.  Look at the mpxio component
10859 	 * type on the child device to see if this device is mpxio managed
10860 	 * or not.
10861 	 */
10862 	if (mdi_component_is_client(tgt_dip, NULL) == MDI_SUCCESS) {
10863 		rval = fcp_virt_tgt_init(hba_dip, tgt_dip, hba_tran, sd);
10864 	} else {
10865 		rval = fcp_phys_tgt_init(hba_dip, tgt_dip, hba_tran, sd);
10866 	}
10867 
10868 	return (rval);
10869 }
10870 
10871 
10872 /* ARGSUSED */
10873 static void
10874 fcp_scsi_tgt_free(dev_info_t *hba_dip, dev_info_t *tgt_dip,
10875     scsi_hba_tran_t *hba_tran, struct scsi_device *sd)
10876 {
10877 	struct fcp_lun	*plun = scsi_device_hba_private_get(sd);
10878 	struct fcp_tgt	*ptgt;
10879 
10880 	FCP_DTRACE(fcp_logq, LUN_PORT->port_instbuf,
10881 	    fcp_trace, FCP_BUF_LEVEL_8, 0,
10882 	    "fcp_scsi_tgt_free: called for tran %s%d, dev %s%d",
10883 	    ddi_get_name(hba_dip), ddi_get_instance(hba_dip),
10884 	    ddi_get_name(tgt_dip), ddi_get_instance(tgt_dip));
10885 
10886 	if (plun == NULL) {
10887 		return;
10888 	}
10889 	ptgt = plun->lun_tgt;
10890 
10891 	ASSERT(ptgt != NULL);
10892 
10893 	mutex_enter(&ptgt->tgt_mutex);
10894 	ASSERT(plun->lun_tgt_count > 0);
10895 
10896 	if (--plun->lun_tgt_count == 0) {
10897 		plun->lun_state &= ~FCP_SCSI_LUN_TGT_INIT;
10898 	}
10899 	plun->lun_tran = NULL;
10900 	mutex_exit(&ptgt->tgt_mutex);
10901 }
10902 
10903 /*
10904  *     Function: fcp_scsi_start
10905  *
10906  *  Description: This function is called by the target driver to request a
10907  *		 command to be sent.
10908  *
10909  *     Argument: *ap		SCSI address of the device.
10910  *		 *pkt		SCSI packet containing the cmd to send.
10911  *
10912  * Return Value: TRAN_ACCEPT
10913  *		 TRAN_BUSY
10914  *		 TRAN_BADPKT
10915  *		 TRAN_FATAL_ERROR
10916  */
10917 static int
10918 fcp_scsi_start(struct scsi_address *ap, struct scsi_pkt *pkt)
10919 {
10920 	struct fcp_port	*pptr = ADDR2FCP(ap);
10921 	struct fcp_lun	*plun = ADDR2LUN(ap);
10922 	struct fcp_pkt	*cmd = PKT2CMD(pkt);
10923 	struct fcp_tgt	*ptgt = plun->lun_tgt;
10924 	int			rval;
10925 
10926 	/* ensure command isn't already issued */
10927 	ASSERT(cmd->cmd_state != FCP_PKT_ISSUED);
10928 
10929 	FCP_DTRACE(fcp_logq, pptr->port_instbuf,
10930 	    fcp_trace, FCP_BUF_LEVEL_9, 0,
10931 	    "fcp_transport Invoked for %x", plun->lun_tgt->tgt_d_id);
10932 
10933 	/*
10934 	 * It is strange that we enter the fcp_port mutex and the target
10935 	 * mutex to check the lun state (which has a mutex of its own).
10936 	 */
10937 	mutex_enter(&pptr->port_mutex);
10938 	mutex_enter(&ptgt->tgt_mutex);
10939 
10940 	/*
10941 	 * If the device is offline and is not in the process of coming
10942 	 * online, fail the request.
10943 	 */
10944 
10945 	if ((plun->lun_state & FCP_LUN_OFFLINE) &&
10946 	    !(plun->lun_state & FCP_LUN_ONLINING)) {
10947 		mutex_exit(&ptgt->tgt_mutex);
10948 		mutex_exit(&pptr->port_mutex);
10949 
10950 		if (cmd->cmd_fp_pkt->pkt_pd == NULL) {
10951 			pkt->pkt_reason = CMD_DEV_GONE;
10952 		}
10953 
10954 		return (TRAN_FATAL_ERROR);
10955 	}
10956 	cmd->cmd_fp_pkt->pkt_timeout = pkt->pkt_time;
10957 
10958 	/*
10959 	 * If we are suspended, kernel is trying to dump, so don't
10960 	 * block, fail or defer requests - send them down right away.
10961 	 * NOTE: If we are in panic (i.e. trying to dump), we can't
10962 	 * assume we have been suspended.  There is hardware such as
10963 	 * the v880 that doesn't do PM.	 Thus, the check for
10964 	 * ddi_in_panic.
10965 	 *
10966 	 * If FCP_STATE_IN_CB_DEVC is set, devices are in the process
10967 	 * of changing.	 So, if we can queue the packet, do it.	 Eventually,
10968 	 * either the device will have gone away or changed and we can fail
10969 	 * the request, or we can proceed if the device didn't change.
10970 	 *
10971 	 * If the pd in the target or the packet is NULL it's probably
10972 	 * because the device has gone away, we allow the request to be
10973 	 * put on the internal queue here in case the device comes back within
10974 	 * the offline timeout. fctl will fix up the pd's if the tgt_pd_handle
10975 	 * has gone NULL, while fcp deals cases where pkt_pd is NULL. pkt_pd
10976 	 * could be NULL because the device was disappearing during or since
10977 	 * packet initialization.
10978 	 */
10979 
10980 	if (((plun->lun_state & FCP_LUN_BUSY) && (!(pptr->port_state &
10981 	    FCP_STATE_SUSPENDED)) && !ddi_in_panic()) ||
10982 	    (pptr->port_state & (FCP_STATE_ONLINING | FCP_STATE_IN_CB_DEVC)) ||
10983 	    (ptgt->tgt_pd_handle == NULL) ||
10984 	    (cmd->cmd_fp_pkt->pkt_pd == NULL)) {
10985 		/*
10986 		 * If ((LUN is busy AND
10987 		 *	LUN not suspended AND
10988 		 *	The system is not in panic state) OR
10989 		 *	(The port is coming up))
10990 		 *
10991 		 * We check to see if the any of the flags FLAG_NOINTR or
10992 		 * FLAG_NOQUEUE is set.	 If one of them is set the value
10993 		 * returned will be TRAN_BUSY.	If not, the request is queued.
10994 		 */
10995 		mutex_exit(&ptgt->tgt_mutex);
10996 		mutex_exit(&pptr->port_mutex);
10997 
10998 		/* see if using interrupts is allowed (so queueing'll work) */
10999 		if (pkt->pkt_flags & FLAG_NOINTR) {
11000 			pkt->pkt_resid = 0;
11001 			return (TRAN_BUSY);
11002 		}
11003 		if (pkt->pkt_flags & FLAG_NOQUEUE) {
11004 			FCP_DTRACE(fcp_logq, pptr->port_instbuf,
11005 			    fcp_trace, FCP_BUF_LEVEL_9, 0,
11006 			    "fcp_scsi_start: lun busy for pkt %p", pkt);
11007 			return (TRAN_BUSY);
11008 		}
11009 #ifdef	DEBUG
11010 		mutex_enter(&pptr->port_pkt_mutex);
11011 		pptr->port_npkts++;
11012 		mutex_exit(&pptr->port_pkt_mutex);
11013 #endif /* DEBUG */
11014 
11015 		/* got queue up the pkt for later */
11016 		fcp_queue_pkt(pptr, cmd);
11017 		return (TRAN_ACCEPT);
11018 	}
11019 	cmd->cmd_state = FCP_PKT_ISSUED;
11020 
11021 	mutex_exit(&ptgt->tgt_mutex);
11022 	mutex_exit(&pptr->port_mutex);
11023 
11024 	/*
11025 	 * Now that we released the mutexes, what was protected by them can
11026 	 * change.
11027 	 */
11028 
11029 	/*
11030 	 * If there is a reconfiguration in progress, wait for it to complete.
11031 	 */
11032 	fcp_reconfig_wait(pptr);
11033 
11034 	cmd->cmd_timeout = pkt->pkt_time ? fcp_watchdog_time +
11035 	    pkt->pkt_time : 0;
11036 
11037 	/* prepare the packet */
11038 
11039 	fcp_prepare_pkt(pptr, cmd, plun);
11040 
11041 	if (cmd->cmd_pkt->pkt_time) {
11042 		cmd->cmd_fp_pkt->pkt_timeout = cmd->cmd_pkt->pkt_time;
11043 	} else {
11044 		cmd->cmd_fp_pkt->pkt_timeout = 5 * 60 * 60;
11045 	}
11046 
11047 	/*
11048 	 * if interrupts aren't allowed (e.g. at dump time) then we'll
11049 	 * have to do polled I/O
11050 	 */
11051 	if (pkt->pkt_flags & FLAG_NOINTR) {
11052 		cmd->cmd_state &= ~FCP_PKT_ISSUED;
11053 		return (fcp_dopoll(pptr, cmd));
11054 	}
11055 
11056 #ifdef	DEBUG
11057 	mutex_enter(&pptr->port_pkt_mutex);
11058 	pptr->port_npkts++;
11059 	mutex_exit(&pptr->port_pkt_mutex);
11060 #endif /* DEBUG */
11061 
11062 	rval = fcp_transport(pptr->port_fp_handle, cmd->cmd_fp_pkt, 0);
11063 	if (rval == FC_SUCCESS) {
11064 		FCP_DTRACE(fcp_logq, pptr->port_instbuf,
11065 		    fcp_trace, FCP_BUF_LEVEL_9, 0,
11066 		    "fcp_transport success for %x", plun->lun_tgt->tgt_d_id);
11067 		return (TRAN_ACCEPT);
11068 	}
11069 
11070 	cmd->cmd_state = FCP_PKT_IDLE;
11071 
11072 #ifdef	DEBUG
11073 	mutex_enter(&pptr->port_pkt_mutex);
11074 	pptr->port_npkts--;
11075 	mutex_exit(&pptr->port_pkt_mutex);
11076 #endif /* DEBUG */
11077 
11078 	/*
11079 	 * For lack of clearer definitions, choose
11080 	 * between TRAN_BUSY and TRAN_FATAL_ERROR.
11081 	 */
11082 
11083 	if (rval == FC_TRAN_BUSY) {
11084 		pkt->pkt_resid = 0;
11085 		rval = TRAN_BUSY;
11086 	} else {
11087 		mutex_enter(&ptgt->tgt_mutex);
11088 		if (plun->lun_state & FCP_LUN_OFFLINE) {
11089 			child_info_t	*cip;
11090 
11091 			mutex_enter(&plun->lun_mutex);
11092 			cip = plun->lun_cip;
11093 			mutex_exit(&plun->lun_mutex);
11094 
11095 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
11096 			    fcp_trace, FCP_BUF_LEVEL_6, 0,
11097 			    "fcp_transport failed 2 for %x: %x; dip=%p",
11098 			    plun->lun_tgt->tgt_d_id, rval, cip);
11099 
11100 			rval = TRAN_FATAL_ERROR;
11101 		} else {
11102 			if (pkt->pkt_flags & FLAG_NOQUEUE) {
11103 				FCP_DTRACE(fcp_logq, pptr->port_instbuf,
11104 				    fcp_trace, FCP_BUF_LEVEL_9, 0,
11105 				    "fcp_scsi_start: FC_BUSY for pkt %p",
11106 				    pkt);
11107 				rval = TRAN_BUSY;
11108 			} else {
11109 				rval = TRAN_ACCEPT;
11110 				fcp_queue_pkt(pptr, cmd);
11111 			}
11112 		}
11113 		mutex_exit(&ptgt->tgt_mutex);
11114 	}
11115 
11116 	return (rval);
11117 }
11118 
11119 /*
11120  * called by the transport to abort a packet
11121  */
11122 /*ARGSUSED*/
11123 static int
11124 fcp_scsi_abort(struct scsi_address *ap, struct scsi_pkt *pkt)
11125 {
11126 	int tgt_cnt;
11127 	struct fcp_port		*pptr = ADDR2FCP(ap);
11128 	struct fcp_lun	*plun = ADDR2LUN(ap);
11129 	struct fcp_tgt	*ptgt = plun->lun_tgt;
11130 
11131 	if (pkt == NULL) {
11132 		if (ptgt) {
11133 			mutex_enter(&ptgt->tgt_mutex);
11134 			tgt_cnt = ptgt->tgt_change_cnt;
11135 			mutex_exit(&ptgt->tgt_mutex);
11136 			fcp_abort_all(pptr, ptgt, plun, tgt_cnt);
11137 			return (TRUE);
11138 		}
11139 	}
11140 	return (FALSE);
11141 }
11142 
11143 
11144 /*
11145  * Perform reset
11146  */
11147 int
11148 fcp_scsi_reset(struct scsi_address *ap, int level)
11149 {
11150 	int			rval = 0;
11151 	struct fcp_port		*pptr = ADDR2FCP(ap);
11152 	struct fcp_lun	*plun = ADDR2LUN(ap);
11153 	struct fcp_tgt	*ptgt = plun->lun_tgt;
11154 
11155 	if (level == RESET_ALL) {
11156 		if (fcp_linkreset(pptr, ap, KM_NOSLEEP) == FC_SUCCESS) {
11157 			rval = 1;
11158 		}
11159 	} else if (level == RESET_TARGET || level == RESET_LUN) {
11160 		/*
11161 		 * If we are in the middle of discovery, return
11162 		 * SUCCESS as this target will be rediscovered
11163 		 * anyway
11164 		 */
11165 		mutex_enter(&ptgt->tgt_mutex);
11166 		if (ptgt->tgt_state & (FCP_TGT_OFFLINE | FCP_TGT_BUSY)) {
11167 			mutex_exit(&ptgt->tgt_mutex);
11168 			return (1);
11169 		}
11170 		mutex_exit(&ptgt->tgt_mutex);
11171 
11172 		if (fcp_reset_target(ap, level) == FC_SUCCESS) {
11173 			rval = 1;
11174 		}
11175 	}
11176 	return (rval);
11177 }
11178 
11179 
11180 /*
11181  * called by the framework to get a SCSI capability
11182  */
11183 static int
11184 fcp_scsi_getcap(struct scsi_address *ap, char *cap, int whom)
11185 {
11186 	return (fcp_commoncap(ap, cap, 0, whom, 0));
11187 }
11188 
11189 
11190 /*
11191  * called by the framework to set a SCSI capability
11192  */
11193 static int
11194 fcp_scsi_setcap(struct scsi_address *ap, char *cap, int value, int whom)
11195 {
11196 	return (fcp_commoncap(ap, cap, value, whom, 1));
11197 }
11198 
11199 /*
11200  *     Function: fcp_pkt_setup
11201  *
11202  *  Description: This function sets up the scsi_pkt structure passed by the
11203  *		 caller. This function assumes fcp_pkt_constructor has been
11204  *		 called previously for the packet passed by the caller.	 If
11205  *		 successful this call will have the following results:
11206  *
11207  *		   - The resources needed that will be constant through out
11208  *		     the whole transaction are allocated.
11209  *		   - The fields that will be constant through out the whole
11210  *		     transaction are initialized.
11211  *		   - The scsi packet will be linked to the LUN structure
11212  *		     addressed by the transaction.
11213  *
11214  *     Argument:
11215  *		 *pkt		Pointer to a scsi_pkt structure.
11216  *		 callback
11217  *		 arg
11218  *
11219  * Return Value: 0	Success
11220  *		 !0	Failure
11221  *
11222  *	Context: Kernel context or interrupt context
11223  */
11224 /* ARGSUSED */
11225 static int
11226 fcp_pkt_setup(struct scsi_pkt *pkt,
11227     int (*callback)(caddr_t arg),
11228     caddr_t arg)
11229 {
11230 	struct fcp_pkt	*cmd;
11231 	struct fcp_port	*pptr;
11232 	struct fcp_lun	*plun;
11233 	struct fcp_tgt	*ptgt;
11234 	int		kf;
11235 	fc_packet_t	*fpkt;
11236 	fc_frame_hdr_t	*hp;
11237 
11238 	pptr = ADDR2FCP(&pkt->pkt_address);
11239 	plun = ADDR2LUN(&pkt->pkt_address);
11240 	ptgt = plun->lun_tgt;
11241 
11242 	cmd = (struct fcp_pkt *)pkt->pkt_ha_private;
11243 	fpkt = cmd->cmd_fp_pkt;
11244 
11245 	/*
11246 	 * this request is for dma allocation only
11247 	 */
11248 	/*
11249 	 * First step of fcp_scsi_init_pkt: pkt allocation
11250 	 * We determine if the caller is willing to wait for the
11251 	 * resources.
11252 	 */
11253 	kf = (callback == SLEEP_FUNC) ? KM_SLEEP: KM_NOSLEEP;
11254 
11255 	/*
11256 	 * Selective zeroing of the pkt.
11257 	 */
11258 	cmd->cmd_back = NULL;
11259 	cmd->cmd_next = NULL;
11260 
11261 	/*
11262 	 * Zero out fcp command
11263 	 */
11264 	bzero(&cmd->cmd_fcp_cmd, sizeof (cmd->cmd_fcp_cmd));
11265 
11266 	cmd->cmd_state = FCP_PKT_IDLE;
11267 
11268 	fpkt = cmd->cmd_fp_pkt;
11269 	fpkt->pkt_data_acc = NULL;
11270 
11271 	mutex_enter(&ptgt->tgt_mutex);
11272 	fpkt->pkt_pd = ptgt->tgt_pd_handle;
11273 
11274 	if (fc_ulp_init_packet(pptr->port_fp_handle, fpkt, kf)
11275 	    != FC_SUCCESS) {
11276 		mutex_exit(&ptgt->tgt_mutex);
11277 		return (-1);
11278 	}
11279 
11280 	mutex_exit(&ptgt->tgt_mutex);
11281 
11282 	/* Fill in the Fabric Channel Header */
11283 	hp = &fpkt->pkt_cmd_fhdr;
11284 	hp->r_ctl = R_CTL_COMMAND;
11285 	hp->rsvd = 0;
11286 	hp->type = FC_TYPE_SCSI_FCP;
11287 	hp->f_ctl = F_CTL_SEQ_INITIATIVE | F_CTL_FIRST_SEQ;
11288 	hp->seq_id = 0;
11289 	hp->df_ctl  = 0;
11290 	hp->seq_cnt = 0;
11291 	hp->ox_id = 0xffff;
11292 	hp->rx_id = 0xffff;
11293 	hp->ro = 0;
11294 
11295 	/*
11296 	 * A doubly linked list (cmd_forw, cmd_back) is built
11297 	 * out of every allocated packet on a per-lun basis
11298 	 *
11299 	 * The packets are maintained in the list so as to satisfy
11300 	 * scsi_abort() requests. At present (which is unlikely to
11301 	 * change in the future) nobody performs a real scsi_abort
11302 	 * in the SCSI target drivers (as they don't keep the packets
11303 	 * after doing scsi_transport - so they don't know how to
11304 	 * abort a packet other than sending a NULL to abort all
11305 	 * outstanding packets)
11306 	 */
11307 	mutex_enter(&plun->lun_mutex);
11308 	if ((cmd->cmd_forw = plun->lun_pkt_head) != NULL) {
11309 		plun->lun_pkt_head->cmd_back = cmd;
11310 	} else {
11311 		plun->lun_pkt_tail = cmd;
11312 	}
11313 	plun->lun_pkt_head = cmd;
11314 	mutex_exit(&plun->lun_mutex);
11315 	return (0);
11316 }
11317 
11318 /*
11319  *     Function: fcp_pkt_teardown
11320  *
11321  *  Description: This function releases a scsi_pkt structure and all the
11322  *		 resources attached to it.
11323  *
11324  *     Argument: *pkt		Pointer to a scsi_pkt structure.
11325  *
11326  * Return Value: None
11327  *
11328  *	Context: User, Kernel or Interrupt context.
11329  */
11330 static void
11331 fcp_pkt_teardown(struct scsi_pkt *pkt)
11332 {
11333 	struct fcp_port	*pptr = ADDR2FCP(&pkt->pkt_address);
11334 	struct fcp_lun	*plun = ADDR2LUN(&pkt->pkt_address);
11335 	struct fcp_pkt	*cmd = (struct fcp_pkt *)pkt->pkt_ha_private;
11336 
11337 	/*
11338 	 * Remove the packet from the per-lun list
11339 	 */
11340 	mutex_enter(&plun->lun_mutex);
11341 	if (cmd->cmd_back) {
11342 		ASSERT(cmd != plun->lun_pkt_head);
11343 		cmd->cmd_back->cmd_forw = cmd->cmd_forw;
11344 	} else {
11345 		ASSERT(cmd == plun->lun_pkt_head);
11346 		plun->lun_pkt_head = cmd->cmd_forw;
11347 	}
11348 
11349 	if (cmd->cmd_forw) {
11350 		cmd->cmd_forw->cmd_back = cmd->cmd_back;
11351 	} else {
11352 		ASSERT(cmd == plun->lun_pkt_tail);
11353 		plun->lun_pkt_tail = cmd->cmd_back;
11354 	}
11355 
11356 	mutex_exit(&plun->lun_mutex);
11357 
11358 	(void) fc_ulp_uninit_packet(pptr->port_fp_handle, cmd->cmd_fp_pkt);
11359 }
11360 
11361 /*
11362  * Routine for reset notification setup, to register or cancel.
11363  * This function is called by SCSA
11364  */
11365 /*ARGSUSED*/
11366 static int
11367 fcp_scsi_reset_notify(struct scsi_address *ap, int flag,
11368     void (*callback)(caddr_t), caddr_t arg)
11369 {
11370 	struct fcp_port *pptr = ADDR2FCP(ap);
11371 
11372 	return (scsi_hba_reset_notify_setup(ap, flag, callback, arg,
11373 	    &pptr->port_mutex, &pptr->port_reset_notify_listf));
11374 }
11375 
11376 
11377 static int
11378 fcp_scsi_bus_get_eventcookie(dev_info_t *dip, dev_info_t *rdip, char *name,
11379     ddi_eventcookie_t *event_cookiep)
11380 {
11381 	struct fcp_port *pptr = fcp_dip2port(dip);
11382 
11383 	if (pptr == NULL) {
11384 		return (DDI_FAILURE);
11385 	}
11386 
11387 	return (ndi_event_retrieve_cookie(pptr->port_ndi_event_hdl, rdip, name,
11388 	    event_cookiep, NDI_EVENT_NOPASS));
11389 }
11390 
11391 
11392 static int
11393 fcp_scsi_bus_add_eventcall(dev_info_t *dip, dev_info_t *rdip,
11394     ddi_eventcookie_t eventid, void (*callback)(), void *arg,
11395     ddi_callback_id_t *cb_id)
11396 {
11397 	struct fcp_port *pptr = fcp_dip2port(dip);
11398 
11399 	if (pptr == NULL) {
11400 		return (DDI_FAILURE);
11401 	}
11402 
11403 	return (ndi_event_add_callback(pptr->port_ndi_event_hdl, rdip,
11404 	    eventid, callback, arg, NDI_SLEEP, cb_id));
11405 }
11406 
11407 
11408 static int
11409 fcp_scsi_bus_remove_eventcall(dev_info_t *dip, ddi_callback_id_t cb_id)
11410 {
11411 
11412 	struct fcp_port *pptr = fcp_dip2port(dip);
11413 
11414 	if (pptr == NULL) {
11415 		return (DDI_FAILURE);
11416 	}
11417 	return (ndi_event_remove_callback(pptr->port_ndi_event_hdl, cb_id));
11418 }
11419 
11420 
11421 /*
11422  * called by the transport to post an event
11423  */
11424 static int
11425 fcp_scsi_bus_post_event(dev_info_t *dip, dev_info_t *rdip,
11426     ddi_eventcookie_t eventid, void *impldata)
11427 {
11428 	struct fcp_port *pptr = fcp_dip2port(dip);
11429 
11430 	if (pptr == NULL) {
11431 		return (DDI_FAILURE);
11432 	}
11433 
11434 	return (ndi_event_run_callbacks(pptr->port_ndi_event_hdl, rdip,
11435 	    eventid, impldata));
11436 }
11437 
11438 
11439 /*
11440  * A target in in many cases in Fibre Channel has a one to one relation
11441  * with a port identifier (which is also known as D_ID and also as AL_PA
11442  * in private Loop) On Fibre Channel-to-SCSI bridge boxes a target reset
11443  * will most likely result in resetting all LUNs (which means a reset will
11444  * occur on all the SCSI devices connected at the other end of the bridge)
11445  * That is the latest favorite topic for discussion, for, one can debate as
11446  * hot as one likes and come up with arguably a best solution to one's
11447  * satisfaction
11448  *
11449  * To stay on track and not digress much, here are the problems stated
11450  * briefly:
11451  *
11452  *	SCSA doesn't define RESET_LUN, It defines RESET_TARGET, but the
11453  *	target drivers use RESET_TARGET even if their instance is on a
11454  *	LUN. Doesn't that sound a bit broken ?
11455  *
11456  *	FCP SCSI (the current spec) only defines RESET TARGET in the
11457  *	control fields of an FCP_CMND structure. It should have been
11458  *	fixed right there, giving flexibility to the initiators to
11459  *	minimize havoc that could be caused by resetting a target.
11460  */
11461 static int
11462 fcp_reset_target(struct scsi_address *ap, int level)
11463 {
11464 	int			rval = FC_FAILURE;
11465 	char			lun_id[25];
11466 	struct fcp_port		*pptr = ADDR2FCP(ap);
11467 	struct fcp_lun	*plun = ADDR2LUN(ap);
11468 	struct fcp_tgt	*ptgt = plun->lun_tgt;
11469 	struct scsi_pkt		*pkt;
11470 	struct fcp_pkt	*cmd;
11471 	struct fcp_rsp		*rsp;
11472 	uint32_t		tgt_cnt;
11473 	struct fcp_rsp_info	*rsp_info;
11474 	struct fcp_reset_elem	*p;
11475 	int			bval;
11476 
11477 	if ((p = kmem_alloc(sizeof (struct fcp_reset_elem),
11478 	    KM_NOSLEEP)) == NULL) {
11479 		return (rval);
11480 	}
11481 
11482 	mutex_enter(&ptgt->tgt_mutex);
11483 	if (level == RESET_TARGET) {
11484 		if (ptgt->tgt_state & (FCP_TGT_OFFLINE | FCP_TGT_BUSY)) {
11485 			mutex_exit(&ptgt->tgt_mutex);
11486 			kmem_free(p, sizeof (struct fcp_reset_elem));
11487 			return (rval);
11488 		}
11489 		fcp_update_tgt_state(ptgt, FCP_SET, FCP_LUN_BUSY);
11490 		(void) strcpy(lun_id, " ");
11491 	} else {
11492 		if (plun->lun_state & (FCP_LUN_OFFLINE | FCP_LUN_BUSY)) {
11493 			mutex_exit(&ptgt->tgt_mutex);
11494 			kmem_free(p, sizeof (struct fcp_reset_elem));
11495 			return (rval);
11496 		}
11497 		fcp_update_lun_state(plun, FCP_SET, FCP_LUN_BUSY);
11498 
11499 		(void) sprintf(lun_id, ", LUN=%d", plun->lun_num);
11500 	}
11501 	tgt_cnt = ptgt->tgt_change_cnt;
11502 
11503 	mutex_exit(&ptgt->tgt_mutex);
11504 
11505 	if ((pkt = scsi_init_pkt(ap, NULL, NULL, 0, 0,
11506 	    0, 0, NULL, 0)) == NULL) {
11507 		kmem_free(p, sizeof (struct fcp_reset_elem));
11508 		mutex_enter(&ptgt->tgt_mutex);
11509 		fcp_update_tgt_state(ptgt, FCP_RESET, FCP_LUN_BUSY);
11510 		mutex_exit(&ptgt->tgt_mutex);
11511 		return (rval);
11512 	}
11513 	pkt->pkt_time = FCP_POLL_TIMEOUT;
11514 
11515 	/* fill in cmd part of packet */
11516 	cmd = PKT2CMD(pkt);
11517 	if (level == RESET_TARGET) {
11518 		cmd->cmd_fcp_cmd.fcp_cntl.cntl_reset_tgt = 1;
11519 	} else {
11520 		cmd->cmd_fcp_cmd.fcp_cntl.cntl_reset_lun = 1;
11521 	}
11522 	cmd->cmd_fp_pkt->pkt_comp = NULL;
11523 	cmd->cmd_pkt->pkt_flags |= FLAG_NOINTR;
11524 
11525 	/* prepare a packet for transport */
11526 	fcp_prepare_pkt(pptr, cmd, plun);
11527 
11528 	if (cmd->cmd_pkt->pkt_time) {
11529 		cmd->cmd_fp_pkt->pkt_timeout = cmd->cmd_pkt->pkt_time;
11530 	} else {
11531 		cmd->cmd_fp_pkt->pkt_timeout = 5 * 60 * 60;
11532 	}
11533 
11534 	(void) fc_ulp_busy_port(pptr->port_fp_handle);
11535 	bval = fcp_dopoll(pptr, cmd);
11536 	fc_ulp_idle_port(pptr->port_fp_handle);
11537 
11538 	/* submit the packet */
11539 	if (bval == TRAN_ACCEPT) {
11540 		int error = 3;
11541 
11542 		rsp = (struct fcp_rsp *)cmd->cmd_fcp_rsp;
11543 		rsp_info = (struct fcp_rsp_info *)(cmd->cmd_fcp_rsp +
11544 		    sizeof (struct fcp_rsp));
11545 
11546 		if (rsp->fcp_u.fcp_status.rsp_len_set) {
11547 			if (fcp_validate_fcp_response(rsp, pptr) ==
11548 			    FC_SUCCESS) {
11549 				if (pptr->port_fcp_dma != FC_NO_DVMA_SPACE) {
11550 					FCP_CP_IN(cmd->cmd_fp_pkt->pkt_resp +
11551 					    sizeof (struct fcp_rsp), rsp_info,
11552 					    cmd->cmd_fp_pkt->pkt_resp_acc,
11553 					    sizeof (struct fcp_rsp_info));
11554 				}
11555 				if (rsp_info->rsp_code == FCP_NO_FAILURE) {
11556 					rval = FC_SUCCESS;
11557 					error = 0;
11558 				} else {
11559 					error = 1;
11560 				}
11561 			} else {
11562 				error = 2;
11563 			}
11564 		}
11565 
11566 		switch (error) {
11567 		case 0:
11568 			fcp_log(CE_WARN, pptr->port_dip,
11569 			    "!FCP: WWN 0x%08x%08x %s reset successfully",
11570 			    *((int *)&ptgt->tgt_port_wwn.raw_wwn[0]),
11571 			    *((int *)&ptgt->tgt_port_wwn.raw_wwn[4]), lun_id);
11572 			break;
11573 
11574 		case 1:
11575 			fcp_log(CE_WARN, pptr->port_dip,
11576 			    "!FCP: Reset to WWN	 0x%08x%08x %s failed,"
11577 			    " response code=%x",
11578 			    *((int *)&ptgt->tgt_port_wwn.raw_wwn[0]),
11579 			    *((int *)&ptgt->tgt_port_wwn.raw_wwn[4]), lun_id,
11580 			    rsp_info->rsp_code);
11581 			break;
11582 
11583 		case 2:
11584 			fcp_log(CE_WARN, pptr->port_dip,
11585 			    "!FCP: Reset to WWN 0x%08x%08x %s failed,"
11586 			    " Bad FCP response values: rsvd1=%x,"
11587 			    " rsvd2=%x, sts-rsvd1=%x, sts-rsvd2=%x,"
11588 			    " rsplen=%x, senselen=%x",
11589 			    *((int *)&ptgt->tgt_port_wwn.raw_wwn[0]),
11590 			    *((int *)&ptgt->tgt_port_wwn.raw_wwn[4]), lun_id,
11591 			    rsp->reserved_0, rsp->reserved_1,
11592 			    rsp->fcp_u.fcp_status.reserved_0,
11593 			    rsp->fcp_u.fcp_status.reserved_1,
11594 			    rsp->fcp_response_len, rsp->fcp_sense_len);
11595 			break;
11596 
11597 		default:
11598 			fcp_log(CE_WARN, pptr->port_dip,
11599 			    "!FCP: Reset to WWN	 0x%08x%08x %s failed",
11600 			    *((int *)&ptgt->tgt_port_wwn.raw_wwn[0]),
11601 			    *((int *)&ptgt->tgt_port_wwn.raw_wwn[4]), lun_id);
11602 			break;
11603 		}
11604 	}
11605 	scsi_destroy_pkt(pkt);
11606 
11607 	if (rval == FC_FAILURE) {
11608 		mutex_enter(&ptgt->tgt_mutex);
11609 		if (level == RESET_TARGET) {
11610 			fcp_update_tgt_state(ptgt, FCP_RESET, FCP_LUN_BUSY);
11611 		} else {
11612 			fcp_update_lun_state(plun, FCP_RESET, FCP_LUN_BUSY);
11613 		}
11614 		mutex_exit(&ptgt->tgt_mutex);
11615 		kmem_free(p, sizeof (struct fcp_reset_elem));
11616 		return (rval);
11617 	}
11618 
11619 	mutex_enter(&pptr->port_mutex);
11620 	if (level == RESET_TARGET) {
11621 		p->tgt = ptgt;
11622 		p->lun = NULL;
11623 	} else {
11624 		p->tgt = NULL;
11625 		p->lun = plun;
11626 	}
11627 	p->tgt = ptgt;
11628 	p->tgt_cnt = tgt_cnt;
11629 	p->timeout = fcp_watchdog_time + FCP_RESET_DELAY;
11630 	p->next = pptr->port_reset_list;
11631 	pptr->port_reset_list = p;
11632 
11633 	FCP_TRACE(fcp_logq, pptr->port_instbuf,
11634 	    fcp_trace, FCP_BUF_LEVEL_3, 0,
11635 	    "Notify ssd of the reset to reinstate the reservations");
11636 
11637 	scsi_hba_reset_notify_callback(&pptr->port_mutex,
11638 	    &pptr->port_reset_notify_listf);
11639 
11640 	mutex_exit(&pptr->port_mutex);
11641 
11642 	return (rval);
11643 }
11644 
11645 
11646 /*
11647  * called by fcp_getcap and fcp_setcap to get and set (respectively)
11648  * SCSI capabilities
11649  */
11650 /* ARGSUSED */
11651 static int
11652 fcp_commoncap(struct scsi_address *ap, char *cap,
11653     int val, int tgtonly, int doset)
11654 {
11655 	struct fcp_port		*pptr = ADDR2FCP(ap);
11656 	struct fcp_lun	*plun = ADDR2LUN(ap);
11657 	struct fcp_tgt	*ptgt = plun->lun_tgt;
11658 	int			cidx;
11659 	int			rval = FALSE;
11660 
11661 	if (cap == (char *)0) {
11662 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
11663 		    fcp_trace, FCP_BUF_LEVEL_3, 0,
11664 		    "fcp_commoncap: invalid arg");
11665 		return (rval);
11666 	}
11667 
11668 	if ((cidx = scsi_hba_lookup_capstr(cap)) == -1) {
11669 		return (UNDEFINED);
11670 	}
11671 
11672 	/*
11673 	 * Process setcap request.
11674 	 */
11675 	if (doset) {
11676 		/*
11677 		 * At present, we can only set binary (0/1) values
11678 		 */
11679 		switch (cidx) {
11680 		case SCSI_CAP_ARQ:
11681 			if (val == 0) {
11682 				rval = FALSE;
11683 			} else {
11684 				rval = TRUE;
11685 			}
11686 			break;
11687 
11688 		case SCSI_CAP_LUN_RESET:
11689 			if (val) {
11690 				plun->lun_cap |= FCP_LUN_CAP_RESET;
11691 			} else {
11692 				plun->lun_cap &= ~FCP_LUN_CAP_RESET;
11693 			}
11694 			rval = TRUE;
11695 			break;
11696 
11697 		case SCSI_CAP_SECTOR_SIZE:
11698 			rval = TRUE;
11699 			break;
11700 		default:
11701 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
11702 			    fcp_trace, FCP_BUF_LEVEL_4, 0,
11703 			    "fcp_setcap: unsupported %d", cidx);
11704 			rval = UNDEFINED;
11705 			break;
11706 		}
11707 
11708 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
11709 		    fcp_trace, FCP_BUF_LEVEL_5, 0,
11710 		    "set cap: cap=%s, val/tgtonly/doset/rval = "
11711 		    "0x%x/0x%x/0x%x/%d",
11712 		    cap, val, tgtonly, doset, rval);
11713 
11714 	} else {
11715 		/*
11716 		 * Process getcap request.
11717 		 */
11718 		switch (cidx) {
11719 		case SCSI_CAP_DMA_MAX:
11720 			rval = (int)pptr->port_data_dma_attr.dma_attr_maxxfer;
11721 
11722 			/*
11723 			 * Need to make an adjustment qlc is uint_t 64
11724 			 * st is int, so we will make the adjustment here
11725 			 * being as nobody wants to touch this.
11726 			 * It still leaves the max single block length
11727 			 * of 2 gig. This should last .
11728 			 */
11729 
11730 			if (rval == -1) {
11731 				rval = MAX_INT_DMA;
11732 			}
11733 
11734 			break;
11735 
11736 		case SCSI_CAP_INITIATOR_ID:
11737 			rval = pptr->port_id;
11738 			break;
11739 
11740 		case SCSI_CAP_ARQ:
11741 		case SCSI_CAP_RESET_NOTIFICATION:
11742 		case SCSI_CAP_TAGGED_QING:
11743 			rval = TRUE;
11744 			break;
11745 
11746 		case SCSI_CAP_SCSI_VERSION:
11747 			rval = 3;
11748 			break;
11749 
11750 		case SCSI_CAP_INTERCONNECT_TYPE:
11751 			if (FC_TOP_EXTERNAL(pptr->port_topology) ||
11752 			    (ptgt->tgt_hard_addr == 0)) {
11753 				rval = INTERCONNECT_FABRIC;
11754 			} else {
11755 				rval = INTERCONNECT_FIBRE;
11756 			}
11757 			break;
11758 
11759 		case SCSI_CAP_LUN_RESET:
11760 			rval = ((plun->lun_cap & FCP_LUN_CAP_RESET) != 0) ?
11761 			    TRUE : FALSE;
11762 			break;
11763 
11764 		default:
11765 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
11766 			    fcp_trace, FCP_BUF_LEVEL_4, 0,
11767 			    "fcp_getcap: unsupported %d", cidx);
11768 			rval = UNDEFINED;
11769 			break;
11770 		}
11771 
11772 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
11773 		    fcp_trace, FCP_BUF_LEVEL_8, 0,
11774 		    "get cap: cap=%s, val/tgtonly/doset/rval = "
11775 		    "0x%x/0x%x/0x%x/%d",
11776 		    cap, val, tgtonly, doset, rval);
11777 	}
11778 
11779 	return (rval);
11780 }
11781 
11782 /*
11783  * called by the transport to get the port-wwn and lun
11784  * properties of this device, and to create a "name" based on them
11785  *
11786  * these properties don't exist on sun4m
11787  *
11788  * return 1 for success else return 0
11789  */
11790 /* ARGSUSED */
11791 static int
11792 fcp_scsi_get_name(struct scsi_device *sd, char *name, int len)
11793 {
11794 	int			i;
11795 	int			*lun;
11796 	int			numChars;
11797 	uint_t			nlun;
11798 	uint_t			count;
11799 	uint_t			nbytes;
11800 	uchar_t			*bytes;
11801 	uint16_t		lun_num;
11802 	uint32_t		tgt_id;
11803 	char			**conf_wwn;
11804 	char			tbuf[(FC_WWN_SIZE << 1) + 1];
11805 	uchar_t			barray[FC_WWN_SIZE];
11806 	dev_info_t		*tgt_dip;
11807 	struct fcp_tgt	*ptgt;
11808 	struct fcp_port	*pptr;
11809 	struct fcp_lun	*plun;
11810 
11811 	ASSERT(sd != NULL);
11812 	ASSERT(name != NULL);
11813 
11814 	tgt_dip = sd->sd_dev;
11815 	pptr = ddi_get_soft_state(fcp_softstate,
11816 	    ddi_get_instance(ddi_get_parent(tgt_dip)));
11817 	if (pptr == NULL) {
11818 		return (0);
11819 	}
11820 
11821 	ASSERT(tgt_dip != NULL);
11822 
11823 	if (ddi_prop_lookup_int_array(DDI_DEV_T_ANY, sd->sd_dev,
11824 	    DDI_PROP_DONTPASS | DDI_PROP_NOTPROM,
11825 	    LUN_PROP, &lun, &nlun) != DDI_SUCCESS) {
11826 		name[0] = '\0';
11827 		return (0);
11828 	}
11829 
11830 	if (nlun == 0) {
11831 		ddi_prop_free(lun);
11832 		return (0);
11833 	}
11834 
11835 	lun_num = lun[0];
11836 	ddi_prop_free(lun);
11837 
11838 	/*
11839 	 * Lookup for .conf WWN property
11840 	 */
11841 	if (ddi_prop_lookup_string_array(DDI_DEV_T_ANY, tgt_dip,
11842 	    DDI_PROP_DONTPASS | DDI_PROP_NOTPROM, CONF_WWN_PROP,
11843 	    &conf_wwn, &count) == DDI_PROP_SUCCESS) {
11844 		ASSERT(count >= 1);
11845 
11846 		fcp_ascii_to_wwn(conf_wwn[0], barray, FC_WWN_SIZE);
11847 		ddi_prop_free(conf_wwn);
11848 		mutex_enter(&pptr->port_mutex);
11849 		if ((plun = fcp_lookup_lun(pptr, barray, lun_num)) == NULL) {
11850 			mutex_exit(&pptr->port_mutex);
11851 			return (0);
11852 		}
11853 		ptgt = plun->lun_tgt;
11854 		mutex_exit(&pptr->port_mutex);
11855 
11856 		(void) ndi_prop_update_byte_array(DDI_DEV_T_NONE,
11857 		    tgt_dip, PORT_WWN_PROP, barray, FC_WWN_SIZE);
11858 
11859 		if (!FC_TOP_EXTERNAL(pptr->port_topology) &&
11860 		    ptgt->tgt_hard_addr != 0) {
11861 			tgt_id = (uint32_t)fcp_alpa_to_switch[
11862 			    ptgt->tgt_hard_addr];
11863 		} else {
11864 			tgt_id = ptgt->tgt_d_id;
11865 		}
11866 
11867 		(void) ndi_prop_update_int(DDI_DEV_T_NONE, tgt_dip,
11868 		    TARGET_PROP, tgt_id);
11869 	}
11870 
11871 	/* get the our port-wwn property */
11872 	bytes = NULL;
11873 	if ((ddi_prop_lookup_byte_array(DDI_DEV_T_ANY, tgt_dip,
11874 	    DDI_PROP_DONTPASS | DDI_PROP_NOTPROM, PORT_WWN_PROP, &bytes,
11875 	    &nbytes) != DDI_PROP_SUCCESS) || nbytes != FC_WWN_SIZE) {
11876 		if (bytes != NULL) {
11877 			ddi_prop_free(bytes);
11878 		}
11879 		return (0);
11880 	}
11881 
11882 	for (i = 0; i < FC_WWN_SIZE; i++) {
11883 		(void) sprintf(&tbuf[i << 1], "%02x", *(bytes + i));
11884 	}
11885 
11886 	/* Stick in the address of the form "wWWN,LUN" */
11887 	numChars = snprintf(name, len, "w%s,%x", tbuf, lun_num);
11888 
11889 	ASSERT(numChars < len);
11890 	if (numChars >= len) {
11891 		fcp_log(CE_WARN, pptr->port_dip,
11892 		    "!fcp_scsi_get_name: "
11893 		    "name parameter length too small, it needs to be %d",
11894 		    numChars+1);
11895 	}
11896 
11897 	ddi_prop_free(bytes);
11898 
11899 	return (1);
11900 }
11901 
11902 
11903 /*
11904  * called by the transport to get the SCSI target id value, returning
11905  * it in "name"
11906  *
11907  * this isn't needed/used on sun4m
11908  *
11909  * return 1 for success else return 0
11910  */
11911 /* ARGSUSED */
11912 static int
11913 fcp_scsi_get_bus_addr(struct scsi_device *sd, char *name, int len)
11914 {
11915 	struct fcp_lun	*plun = ADDR2LUN(&sd->sd_address);
11916 	struct fcp_tgt	*ptgt;
11917 	int    numChars;
11918 
11919 	if (plun == NULL) {
11920 		return (0);
11921 	}
11922 
11923 	if ((ptgt = plun->lun_tgt) == NULL) {
11924 		return (0);
11925 	}
11926 
11927 	numChars = snprintf(name, len, "%x", ptgt->tgt_d_id);
11928 
11929 	ASSERT(numChars < len);
11930 	if (numChars >= len) {
11931 		fcp_log(CE_WARN, NULL,
11932 		    "!fcp_scsi_get_bus_addr: "
11933 		    "name parameter length too small, it needs to be %d",
11934 		    numChars+1);
11935 	}
11936 
11937 	return (1);
11938 }
11939 
11940 
11941 /*
11942  * called internally to reset the link where the specified port lives
11943  */
11944 static int
11945 fcp_linkreset(struct fcp_port *pptr, struct scsi_address *ap, int sleep)
11946 {
11947 	la_wwn_t		wwn;
11948 	struct fcp_lun	*plun;
11949 	struct fcp_tgt	*ptgt;
11950 
11951 	/* disable restart of lip if we're suspended */
11952 	mutex_enter(&pptr->port_mutex);
11953 
11954 	if (pptr->port_state & (FCP_STATE_SUSPENDED |
11955 	    FCP_STATE_POWER_DOWN)) {
11956 		mutex_exit(&pptr->port_mutex);
11957 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
11958 		    fcp_trace, FCP_BUF_LEVEL_2, 0,
11959 		    "fcp_linkreset, fcp%d: link reset "
11960 		    "disabled due to DDI_SUSPEND",
11961 		    ddi_get_instance(pptr->port_dip));
11962 		return (FC_FAILURE);
11963 	}
11964 
11965 	if (pptr->port_state & (FCP_STATE_OFFLINE | FCP_STATE_ONLINING)) {
11966 		mutex_exit(&pptr->port_mutex);
11967 		return (FC_SUCCESS);
11968 	}
11969 
11970 	FCP_DTRACE(fcp_logq, pptr->port_instbuf,
11971 	    fcp_trace, FCP_BUF_LEVEL_8, 0, "Forcing link reset");
11972 
11973 	/*
11974 	 * If ap == NULL assume local link reset.
11975 	 */
11976 	if (FC_TOP_EXTERNAL(pptr->port_topology) && (ap != NULL)) {
11977 		plun = ADDR2LUN(ap);
11978 		ptgt = plun->lun_tgt;
11979 		bcopy(&ptgt->tgt_port_wwn.raw_wwn[0], &wwn, sizeof (wwn));
11980 	} else {
11981 		bzero((caddr_t)&wwn, sizeof (wwn));
11982 	}
11983 	mutex_exit(&pptr->port_mutex);
11984 
11985 	return (fc_ulp_linkreset(pptr->port_fp_handle, &wwn, sleep));
11986 }
11987 
11988 
11989 /*
11990  * called from fcp_port_attach() to resume a port
11991  * return DDI_* success/failure status
11992  * acquires and releases the global mutex
11993  * acquires and releases the port mutex
11994  */
11995 /*ARGSUSED*/
11996 
11997 static int
11998 fcp_handle_port_resume(opaque_t ulph, fc_ulp_port_info_t *pinfo,
11999     uint32_t s_id, fc_attach_cmd_t cmd, int instance)
12000 {
12001 	int			res = DDI_FAILURE; /* default result */
12002 	struct fcp_port	*pptr;		/* port state ptr */
12003 	uint32_t		alloc_cnt;
12004 	uint32_t		max_cnt;
12005 	fc_portmap_t		*tmp_list = NULL;
12006 
12007 	FCP_DTRACE(fcp_logq, "fcp", fcp_trace,
12008 	    FCP_BUF_LEVEL_8, 0, "port resume: for port %d",
12009 	    instance);
12010 
12011 	if ((pptr = ddi_get_soft_state(fcp_softstate, instance)) == NULL) {
12012 		cmn_err(CE_WARN, "fcp: bad soft state");
12013 		return (res);
12014 	}
12015 
12016 	mutex_enter(&pptr->port_mutex);
12017 	switch (cmd) {
12018 	case FC_CMD_RESUME:
12019 		ASSERT((pptr->port_state & FCP_STATE_POWER_DOWN) == 0);
12020 		pptr->port_state &= ~FCP_STATE_SUSPENDED;
12021 		break;
12022 
12023 	case FC_CMD_POWER_UP:
12024 		/*
12025 		 * If the port is DDI_SUSPENded, defer rediscovery
12026 		 * until DDI_RESUME occurs
12027 		 */
12028 		if (pptr->port_state & FCP_STATE_SUSPENDED) {
12029 			pptr->port_state &= ~FCP_STATE_POWER_DOWN;
12030 			mutex_exit(&pptr->port_mutex);
12031 			return (DDI_SUCCESS);
12032 		}
12033 		pptr->port_state &= ~FCP_STATE_POWER_DOWN;
12034 	}
12035 	pptr->port_id = s_id;
12036 	pptr->port_state = FCP_STATE_INIT;
12037 	mutex_exit(&pptr->port_mutex);
12038 
12039 	/*
12040 	 * Make a copy of ulp_port_info as fctl allocates
12041 	 * a temp struct.
12042 	 */
12043 	(void) fcp_cp_pinfo(pptr, pinfo);
12044 
12045 	mutex_enter(&fcp_global_mutex);
12046 	if (fcp_watchdog_init++ == 0) {
12047 		fcp_watchdog_tick = fcp_watchdog_timeout *
12048 		    drv_usectohz(1000000);
12049 		fcp_watchdog_id = timeout(fcp_watch,
12050 		    NULL, fcp_watchdog_tick);
12051 	}
12052 	mutex_exit(&fcp_global_mutex);
12053 
12054 	/*
12055 	 * Handle various topologies and link states.
12056 	 */
12057 	switch (FC_PORT_STATE_MASK(pptr->port_phys_state)) {
12058 	case FC_STATE_OFFLINE:
12059 		/*
12060 		 * Wait for ONLINE, at which time a state
12061 		 * change will cause a statec_callback
12062 		 */
12063 		res = DDI_SUCCESS;
12064 		break;
12065 
12066 	case FC_STATE_ONLINE:
12067 
12068 		if (pptr->port_topology == FC_TOP_UNKNOWN) {
12069 			(void) fcp_linkreset(pptr, NULL, KM_NOSLEEP);
12070 			res = DDI_SUCCESS;
12071 			break;
12072 		}
12073 
12074 		if (FC_TOP_EXTERNAL(pptr->port_topology) &&
12075 		    !fcp_enable_auto_configuration) {
12076 			tmp_list = fcp_construct_map(pptr, &alloc_cnt);
12077 			if (tmp_list == NULL) {
12078 				if (!alloc_cnt) {
12079 					res = DDI_SUCCESS;
12080 				}
12081 				break;
12082 			}
12083 			max_cnt = alloc_cnt;
12084 		} else {
12085 			ASSERT(pptr->port_topology != FC_TOP_UNKNOWN);
12086 
12087 			alloc_cnt = FCP_MAX_DEVICES;
12088 
12089 			if ((tmp_list = (fc_portmap_t *)kmem_zalloc(
12090 			    (sizeof (fc_portmap_t)) * alloc_cnt,
12091 			    KM_NOSLEEP)) == NULL) {
12092 				fcp_log(CE_WARN, pptr->port_dip,
12093 				    "!fcp%d: failed to allocate portmap",
12094 				    instance);
12095 				break;
12096 			}
12097 
12098 			max_cnt = alloc_cnt;
12099 			if ((res = fc_ulp_getportmap(pptr->port_fp_handle,
12100 			    &tmp_list, &max_cnt, FC_ULP_PLOGI_PRESERVE)) !=
12101 			    FC_SUCCESS) {
12102 				caddr_t msg;
12103 
12104 				(void) fc_ulp_error(res, &msg);
12105 
12106 				FCP_TRACE(fcp_logq, pptr->port_instbuf,
12107 				    fcp_trace, FCP_BUF_LEVEL_2, 0,
12108 				    "resume failed getportmap: reason=0x%x",
12109 				    res);
12110 
12111 				fcp_log(CE_WARN, pptr->port_dip,
12112 				    "!failed to get port map : %s", msg);
12113 				break;
12114 			}
12115 			if (max_cnt > alloc_cnt) {
12116 				alloc_cnt = max_cnt;
12117 			}
12118 		}
12119 
12120 		/*
12121 		 * do the SCSI device discovery and create
12122 		 * the devinfos
12123 		 */
12124 		fcp_statec_callback(ulph, pptr->port_fp_handle,
12125 		    pptr->port_phys_state, pptr->port_topology, tmp_list,
12126 		    max_cnt, pptr->port_id);
12127 
12128 		res = DDI_SUCCESS;
12129 		break;
12130 
12131 	default:
12132 		fcp_log(CE_WARN, pptr->port_dip,
12133 		    "!fcp%d: invalid port state at attach=0x%x",
12134 		    instance, pptr->port_phys_state);
12135 
12136 		mutex_enter(&pptr->port_mutex);
12137 		pptr->port_phys_state = FCP_STATE_OFFLINE;
12138 		mutex_exit(&pptr->port_mutex);
12139 		res = DDI_SUCCESS;
12140 
12141 		break;
12142 	}
12143 
12144 	if (tmp_list != NULL) {
12145 		kmem_free(tmp_list, sizeof (fc_portmap_t) * alloc_cnt);
12146 	}
12147 
12148 	return (res);
12149 }
12150 
12151 
12152 static void
12153 fcp_cp_pinfo(struct fcp_port *pptr, fc_ulp_port_info_t *pinfo)
12154 {
12155 	pptr->port_fp_modlinkage = *pinfo->port_linkage;
12156 	pptr->port_dip = pinfo->port_dip;
12157 	pptr->port_fp_handle = pinfo->port_handle;
12158 	pptr->port_data_dma_attr = *pinfo->port_data_dma_attr;
12159 	pptr->port_cmd_dma_attr = *pinfo->port_cmd_dma_attr;
12160 	pptr->port_resp_dma_attr = *pinfo->port_resp_dma_attr;
12161 	pptr->port_dma_acc_attr = *pinfo->port_acc_attr;
12162 	pptr->port_priv_pkt_len = pinfo->port_fca_pkt_size;
12163 	pptr->port_max_exch = pinfo->port_fca_max_exch;
12164 	pptr->port_phys_state = pinfo->port_state;
12165 	pptr->port_topology = pinfo->port_flags;
12166 	pptr->port_reset_action = pinfo->port_reset_action;
12167 	pptr->port_cmds_dma_flags = pinfo->port_dma_behavior;
12168 	pptr->port_fcp_dma = pinfo->port_fcp_dma;
12169 	bcopy(&pinfo->port_nwwn, &pptr->port_nwwn, sizeof (la_wwn_t));
12170 	bcopy(&pinfo->port_pwwn, &pptr->port_pwwn, sizeof (la_wwn_t));
12171 }
12172 
12173 /*
12174  * If the elements wait field is set to 1 then
12175  * another thread is waiting for the operation to complete. Once
12176  * it is complete, the waiting thread is signaled and the element is
12177  * freed by the waiting thread. If the elements wait field is set to 0
12178  * the element is freed.
12179  */
12180 static void
12181 fcp_process_elem(struct fcp_hp_elem *elem, int result)
12182 {
12183 	ASSERT(elem != NULL);
12184 	mutex_enter(&elem->mutex);
12185 	elem->result = result;
12186 	if (elem->wait) {
12187 		elem->wait = 0;
12188 		cv_signal(&elem->cv);
12189 		mutex_exit(&elem->mutex);
12190 	} else {
12191 		mutex_exit(&elem->mutex);
12192 		cv_destroy(&elem->cv);
12193 		mutex_destroy(&elem->mutex);
12194 		kmem_free(elem, sizeof (struct fcp_hp_elem));
12195 	}
12196 }
12197 
12198 /*
12199  * This function is invoked from the taskq thread to allocate
12200  * devinfo nodes and to online/offline them.
12201  */
12202 static void
12203 fcp_hp_task(void *arg)
12204 {
12205 	struct fcp_hp_elem	*elem = (struct fcp_hp_elem *)arg;
12206 	struct fcp_lun	*plun = elem->lun;
12207 	struct fcp_port		*pptr = elem->port;
12208 	int			result;
12209 
12210 	ASSERT(elem->what == FCP_ONLINE ||
12211 	    elem->what == FCP_OFFLINE ||
12212 	    elem->what == FCP_MPXIO_PATH_CLEAR_BUSY ||
12213 	    elem->what == FCP_MPXIO_PATH_SET_BUSY);
12214 
12215 	mutex_enter(&pptr->port_mutex);
12216 	mutex_enter(&plun->lun_mutex);
12217 	if (((elem->what == FCP_ONLINE || elem->what == FCP_OFFLINE) &&
12218 	    plun->lun_event_count != elem->event_cnt) ||
12219 	    pptr->port_state & (FCP_STATE_SUSPENDED |
12220 	    FCP_STATE_DETACHING | FCP_STATE_POWER_DOWN)) {
12221 		mutex_exit(&plun->lun_mutex);
12222 		mutex_exit(&pptr->port_mutex);
12223 		fcp_process_elem(elem, NDI_FAILURE);
12224 		return;
12225 	}
12226 	mutex_exit(&plun->lun_mutex);
12227 	mutex_exit(&pptr->port_mutex);
12228 
12229 	result = fcp_trigger_lun(plun, elem->cip, elem->old_lun_mpxio,
12230 	    elem->what, elem->link_cnt, elem->tgt_cnt, elem->flags);
12231 	fcp_process_elem(elem, result);
12232 }
12233 
12234 
12235 static child_info_t *
12236 fcp_get_cip(struct fcp_lun *plun, child_info_t *cip, int lcount,
12237     int tcount)
12238 {
12239 	ASSERT(MUTEX_HELD(&plun->lun_mutex));
12240 
12241 	if (fcp_is_child_present(plun, cip) == FC_FAILURE) {
12242 		struct fcp_port *pptr = plun->lun_tgt->tgt_port;
12243 
12244 		ASSERT(MUTEX_HELD(&pptr->port_mutex));
12245 		/*
12246 		 * Child has not been created yet. Create the child device
12247 		 * based on the per-Lun flags.
12248 		 */
12249 		if (pptr->port_mpxio == 0 || plun->lun_mpxio == 0) {
12250 			plun->lun_cip =
12251 			    CIP(fcp_create_dip(plun, lcount, tcount));
12252 			plun->lun_mpxio = 0;
12253 		} else {
12254 			plun->lun_cip =
12255 			    CIP(fcp_create_pip(plun, lcount, tcount));
12256 			plun->lun_mpxio = 1;
12257 		}
12258 	} else {
12259 		plun->lun_cip = cip;
12260 	}
12261 
12262 	return (plun->lun_cip);
12263 }
12264 
12265 
12266 static int
12267 fcp_is_dip_present(struct fcp_lun *plun, dev_info_t *cdip)
12268 {
12269 	int		rval = FC_FAILURE;
12270 	dev_info_t	*pdip;
12271 	struct dev_info	*dip;
12272 	int		circular;
12273 
12274 	ASSERT(MUTEX_HELD(&plun->lun_mutex));
12275 
12276 	pdip = plun->lun_tgt->tgt_port->port_dip;
12277 
12278 	if (plun->lun_cip == NULL) {
12279 		FCP_TRACE(fcp_logq, LUN_PORT->port_instbuf,
12280 		    fcp_trace, FCP_BUF_LEVEL_3, 0,
12281 		    "fcp_is_dip_present: plun->lun_cip is NULL: "
12282 		    "plun: %p lun state: %x num: %d target state: %x",
12283 		    plun, plun->lun_state, plun->lun_num,
12284 		    plun->lun_tgt->tgt_port->port_state);
12285 		return (rval);
12286 	}
12287 	ndi_devi_enter(pdip, &circular);
12288 	dip = DEVI(pdip)->devi_child;
12289 	while (dip) {
12290 		if (dip == DEVI(cdip)) {
12291 			rval = FC_SUCCESS;
12292 			break;
12293 		}
12294 		dip = dip->devi_sibling;
12295 	}
12296 	ndi_devi_exit(pdip, circular);
12297 	return (rval);
12298 }
12299 
12300 static int
12301 fcp_is_child_present(struct fcp_lun *plun, child_info_t *cip)
12302 {
12303 	int		rval = FC_FAILURE;
12304 
12305 	ASSERT(plun != NULL);
12306 	ASSERT(MUTEX_HELD(&plun->lun_mutex));
12307 
12308 	if (plun->lun_mpxio == 0) {
12309 		rval = fcp_is_dip_present(plun, DIP(cip));
12310 	} else {
12311 		rval = fcp_is_pip_present(plun, PIP(cip));
12312 	}
12313 
12314 	return (rval);
12315 }
12316 
12317 /*
12318  *     Function: fcp_create_dip
12319  *
12320  *  Description: Creates a dev_info_t structure for the LUN specified by the
12321  *		 caller.
12322  *
12323  *     Argument: plun		Lun structure
12324  *		 link_cnt	Link state count.
12325  *		 tgt_cnt	Target state change count.
12326  *
12327  * Return Value: NULL if it failed
12328  *		 dev_info_t structure address if it succeeded
12329  *
12330  *	Context: Kernel context
12331  */
12332 static dev_info_t *
12333 fcp_create_dip(struct fcp_lun *plun, int link_cnt, int tgt_cnt)
12334 {
12335 	int			failure = 0;
12336 	uint32_t		tgt_id;
12337 	uint64_t		sam_lun;
12338 	struct fcp_tgt	*ptgt = plun->lun_tgt;
12339 	struct fcp_port	*pptr = ptgt->tgt_port;
12340 	dev_info_t		*pdip = pptr->port_dip;
12341 	dev_info_t		*cdip = NULL;
12342 	dev_info_t		*old_dip = DIP(plun->lun_cip);
12343 	char			*nname = NULL;
12344 	char			**compatible = NULL;
12345 	int			ncompatible;
12346 	char			*scsi_binding_set;
12347 	char			t_pwwn[17];
12348 
12349 	ASSERT(MUTEX_HELD(&plun->lun_mutex));
12350 	ASSERT(MUTEX_HELD(&pptr->port_mutex));
12351 
12352 	/* get the 'scsi-binding-set' property */
12353 	if (ddi_prop_lookup_string(DDI_DEV_T_ANY, pdip,
12354 	    DDI_PROP_NOTPROM | DDI_PROP_DONTPASS, "scsi-binding-set",
12355 	    &scsi_binding_set) != DDI_PROP_SUCCESS) {
12356 		scsi_binding_set = NULL;
12357 	}
12358 
12359 	/* determine the node name and compatible */
12360 	scsi_hba_nodename_compatible_get(&plun->lun_inq, scsi_binding_set,
12361 	    plun->lun_inq.inq_dtype, NULL, &nname, &compatible, &ncompatible);
12362 	if (scsi_binding_set) {
12363 		ddi_prop_free(scsi_binding_set);
12364 	}
12365 
12366 	if (nname == NULL) {
12367 #ifdef	DEBUG
12368 		cmn_err(CE_WARN, "%s%d: no driver for "
12369 		    "device @w%02x%02x%02x%02x%02x%02x%02x%02x,%d:"
12370 		    "	 compatible: %s",
12371 		    ddi_driver_name(pdip), ddi_get_instance(pdip),
12372 		    ptgt->tgt_port_wwn.raw_wwn[0],
12373 		    ptgt->tgt_port_wwn.raw_wwn[1],
12374 		    ptgt->tgt_port_wwn.raw_wwn[2],
12375 		    ptgt->tgt_port_wwn.raw_wwn[3],
12376 		    ptgt->tgt_port_wwn.raw_wwn[4],
12377 		    ptgt->tgt_port_wwn.raw_wwn[5],
12378 		    ptgt->tgt_port_wwn.raw_wwn[6],
12379 		    ptgt->tgt_port_wwn.raw_wwn[7], plun->lun_num,
12380 		    *compatible);
12381 #endif	/* DEBUG */
12382 		failure++;
12383 		goto end_of_fcp_create_dip;
12384 	}
12385 
12386 	cdip = fcp_find_existing_dip(plun, pdip, nname);
12387 
12388 	/*
12389 	 * if the old_dip does not match the cdip, that means there is
12390 	 * some property change. since we'll be using the cdip, we need
12391 	 * to offline the old_dip. If the state contains FCP_LUN_CHANGED
12392 	 * then the dtype for the device has been updated. Offline the
12393 	 * the old device and create a new device with the new device type
12394 	 * Refer to bug: 4764752
12395 	 */
12396 	if (old_dip && (cdip != old_dip ||
12397 	    plun->lun_state & FCP_LUN_CHANGED)) {
12398 		plun->lun_state &= ~(FCP_LUN_INIT);
12399 		mutex_exit(&plun->lun_mutex);
12400 		mutex_exit(&pptr->port_mutex);
12401 
12402 		mutex_enter(&ptgt->tgt_mutex);
12403 		(void) fcp_pass_to_hp(pptr, plun, CIP(old_dip), FCP_OFFLINE,
12404 		    link_cnt, tgt_cnt, NDI_DEVI_REMOVE, 0);
12405 		mutex_exit(&ptgt->tgt_mutex);
12406 
12407 #ifdef DEBUG
12408 		if (cdip != NULL) {
12409 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
12410 			    fcp_trace, FCP_BUF_LEVEL_2, 0,
12411 			    "Old dip=%p; New dip=%p don't match", old_dip,
12412 			    cdip);
12413 		} else {
12414 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
12415 			    fcp_trace, FCP_BUF_LEVEL_2, 0,
12416 			    "Old dip=%p; New dip=NULL don't match", old_dip);
12417 		}
12418 #endif
12419 
12420 		mutex_enter(&pptr->port_mutex);
12421 		mutex_enter(&plun->lun_mutex);
12422 	}
12423 
12424 	if (cdip == NULL || plun->lun_state & FCP_LUN_CHANGED) {
12425 		plun->lun_state &= ~(FCP_LUN_CHANGED);
12426 		if (ndi_devi_alloc(pptr->port_dip, nname,
12427 		    DEVI_SID_NODEID, &cdip) != NDI_SUCCESS) {
12428 			failure++;
12429 			goto end_of_fcp_create_dip;
12430 		}
12431 	}
12432 
12433 	/*
12434 	 * Previously all the properties for the devinfo were destroyed here
12435 	 * with a call to ndi_prop_remove_all(). Since this may cause loss of
12436 	 * the devid property (and other properties established by the target
12437 	 * driver or framework) which the code does not always recreate, this
12438 	 * call was removed.
12439 	 * This opens a theoretical possibility that we may return with a
12440 	 * stale devid on the node if the scsi entity behind the fibre channel
12441 	 * lun has changed.
12442 	 */
12443 
12444 	/* decorate the node with compatible */
12445 	if (ndi_prop_update_string_array(DDI_DEV_T_NONE, cdip,
12446 	    "compatible", compatible, ncompatible) != DDI_PROP_SUCCESS) {
12447 		failure++;
12448 		goto end_of_fcp_create_dip;
12449 	}
12450 
12451 	if (ndi_prop_update_byte_array(DDI_DEV_T_NONE, cdip, NODE_WWN_PROP,
12452 	    ptgt->tgt_node_wwn.raw_wwn, FC_WWN_SIZE) != DDI_PROP_SUCCESS) {
12453 		failure++;
12454 		goto end_of_fcp_create_dip;
12455 	}
12456 
12457 	if (ndi_prop_update_byte_array(DDI_DEV_T_NONE, cdip, PORT_WWN_PROP,
12458 	    ptgt->tgt_port_wwn.raw_wwn, FC_WWN_SIZE) != DDI_PROP_SUCCESS) {
12459 		failure++;
12460 		goto end_of_fcp_create_dip;
12461 	}
12462 
12463 	fcp_wwn_to_ascii(ptgt->tgt_port_wwn.raw_wwn, t_pwwn);
12464 	t_pwwn[16] = '\0';
12465 	if (ndi_prop_update_string(DDI_DEV_T_NONE, cdip, TGT_PORT_PROP, t_pwwn)
12466 	    != DDI_PROP_SUCCESS) {
12467 		failure++;
12468 		goto end_of_fcp_create_dip;
12469 	}
12470 
12471 	/*
12472 	 * If there is no hard address - We might have to deal with
12473 	 * that by using WWN - Having said that it is important to
12474 	 * recognize this problem early so ssd can be informed of
12475 	 * the right interconnect type.
12476 	 */
12477 	if (!FC_TOP_EXTERNAL(pptr->port_topology) && ptgt->tgt_hard_addr != 0) {
12478 		tgt_id = (uint32_t)fcp_alpa_to_switch[ptgt->tgt_hard_addr];
12479 	} else {
12480 		tgt_id = ptgt->tgt_d_id;
12481 	}
12482 
12483 	if (ndi_prop_update_int(DDI_DEV_T_NONE, cdip, TARGET_PROP,
12484 	    tgt_id) != DDI_PROP_SUCCESS) {
12485 		failure++;
12486 		goto end_of_fcp_create_dip;
12487 	}
12488 
12489 	if (ndi_prop_update_int(DDI_DEV_T_NONE, cdip, LUN_PROP,
12490 	    (int)plun->lun_num) != DDI_PROP_SUCCESS) {
12491 		failure++;
12492 		goto end_of_fcp_create_dip;
12493 	}
12494 	bcopy(&plun->lun_addr, &sam_lun, FCP_LUN_SIZE);
12495 	if (ndi_prop_update_int64(DDI_DEV_T_NONE, cdip, SAM_LUN_PROP,
12496 	    sam_lun) != DDI_PROP_SUCCESS) {
12497 		failure++;
12498 		goto end_of_fcp_create_dip;
12499 	}
12500 
12501 end_of_fcp_create_dip:
12502 	scsi_hba_nodename_compatible_free(nname, compatible);
12503 
12504 	if (cdip != NULL && failure) {
12505 		(void) ndi_prop_remove_all(cdip);
12506 		(void) ndi_devi_free(cdip);
12507 		cdip = NULL;
12508 	}
12509 
12510 	return (cdip);
12511 }
12512 
12513 /*
12514  *     Function: fcp_create_pip
12515  *
12516  *  Description: Creates a Path Id for the LUN specified by the caller.
12517  *
12518  *     Argument: plun		Lun structure
12519  *		 link_cnt	Link state count.
12520  *		 tgt_cnt	Target state count.
12521  *
12522  * Return Value: NULL if it failed
12523  *		 mdi_pathinfo_t structure address if it succeeded
12524  *
12525  *	Context: Kernel context
12526  */
12527 static mdi_pathinfo_t *
12528 fcp_create_pip(struct fcp_lun *plun, int lcount, int tcount)
12529 {
12530 	int			i;
12531 	char			buf[MAXNAMELEN];
12532 	char			uaddr[MAXNAMELEN];
12533 	int			failure = 0;
12534 	uint32_t		tgt_id;
12535 	uint64_t		sam_lun;
12536 	struct fcp_tgt	*ptgt = plun->lun_tgt;
12537 	struct fcp_port	*pptr = ptgt->tgt_port;
12538 	dev_info_t		*pdip = pptr->port_dip;
12539 	mdi_pathinfo_t		*pip = NULL;
12540 	mdi_pathinfo_t		*old_pip = PIP(plun->lun_cip);
12541 	char			*nname = NULL;
12542 	char			**compatible = NULL;
12543 	int			ncompatible;
12544 	char			*scsi_binding_set;
12545 	char			t_pwwn[17];
12546 
12547 	ASSERT(MUTEX_HELD(&plun->lun_mutex));
12548 	ASSERT(MUTEX_HELD(&pptr->port_mutex));
12549 
12550 	scsi_binding_set = "vhci";
12551 
12552 	/* determine the node name and compatible */
12553 	scsi_hba_nodename_compatible_get(&plun->lun_inq, scsi_binding_set,
12554 	    plun->lun_inq.inq_dtype, NULL, &nname, &compatible, &ncompatible);
12555 
12556 	if (nname == NULL) {
12557 #ifdef	DEBUG
12558 		cmn_err(CE_WARN, "fcp_create_dip: %s%d: no driver for "
12559 		    "device @w%02x%02x%02x%02x%02x%02x%02x%02x,%d:"
12560 		    "	 compatible: %s",
12561 		    ddi_driver_name(pdip), ddi_get_instance(pdip),
12562 		    ptgt->tgt_port_wwn.raw_wwn[0],
12563 		    ptgt->tgt_port_wwn.raw_wwn[1],
12564 		    ptgt->tgt_port_wwn.raw_wwn[2],
12565 		    ptgt->tgt_port_wwn.raw_wwn[3],
12566 		    ptgt->tgt_port_wwn.raw_wwn[4],
12567 		    ptgt->tgt_port_wwn.raw_wwn[5],
12568 		    ptgt->tgt_port_wwn.raw_wwn[6],
12569 		    ptgt->tgt_port_wwn.raw_wwn[7], plun->lun_num,
12570 		    *compatible);
12571 #endif	/* DEBUG */
12572 		failure++;
12573 		goto end_of_fcp_create_pip;
12574 	}
12575 
12576 	pip = fcp_find_existing_pip(plun, pdip);
12577 
12578 	/*
12579 	 * if the old_dip does not match the cdip, that means there is
12580 	 * some property change. since we'll be using the cdip, we need
12581 	 * to offline the old_dip. If the state contains FCP_LUN_CHANGED
12582 	 * then the dtype for the device has been updated. Offline the
12583 	 * the old device and create a new device with the new device type
12584 	 * Refer to bug: 4764752
12585 	 */
12586 	if (old_pip && (pip != old_pip ||
12587 	    plun->lun_state & FCP_LUN_CHANGED)) {
12588 		plun->lun_state &= ~(FCP_LUN_INIT);
12589 		mutex_exit(&plun->lun_mutex);
12590 		mutex_exit(&pptr->port_mutex);
12591 
12592 		mutex_enter(&ptgt->tgt_mutex);
12593 		(void) fcp_pass_to_hp(pptr, plun, CIP(old_pip),
12594 		    FCP_OFFLINE, lcount, tcount,
12595 		    NDI_DEVI_REMOVE, 0);
12596 		mutex_exit(&ptgt->tgt_mutex);
12597 
12598 		if (pip != NULL) {
12599 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
12600 			    fcp_trace, FCP_BUF_LEVEL_2, 0,
12601 			    "Old pip=%p; New pip=%p don't match",
12602 			    old_pip, pip);
12603 		} else {
12604 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
12605 			    fcp_trace, FCP_BUF_LEVEL_2, 0,
12606 			    "Old pip=%p; New pip=NULL don't match",
12607 			    old_pip);
12608 		}
12609 
12610 		mutex_enter(&pptr->port_mutex);
12611 		mutex_enter(&plun->lun_mutex);
12612 	}
12613 
12614 	/*
12615 	 * Since FC_WWN_SIZE is 8 bytes and its not like the
12616 	 * lun_guid_size which is dependent on the target, I don't
12617 	 * believe the same trancation happens here UNLESS the standards
12618 	 * change the FC_WWN_SIZE value to something larger than
12619 	 * MAXNAMELEN(currently 255 bytes).
12620 	 */
12621 
12622 	for (i = 0; i < FC_WWN_SIZE; i++) {
12623 		(void) sprintf(&buf[i << 1], "%02x",
12624 		    ptgt->tgt_port_wwn.raw_wwn[i]);
12625 	}
12626 
12627 	(void) snprintf(uaddr, MAXNAMELEN, "w%s,%x",
12628 	    buf, plun->lun_num);
12629 
12630 	if (pip == NULL || plun->lun_state & FCP_LUN_CHANGED) {
12631 		/*
12632 		 * Release the locks before calling into
12633 		 * mdi_pi_alloc_compatible() since this can result in a
12634 		 * callback into fcp which can result in a deadlock
12635 		 * (see bug # 4870272).
12636 		 *
12637 		 * Basically, what we are trying to avoid is the scenario where
12638 		 * one thread does ndi_devi_enter() and tries to grab
12639 		 * fcp_mutex and another does it the other way round.
12640 		 *
12641 		 * But before we do that, make sure that nobody releases the
12642 		 * port in the meantime. We can do this by setting a flag.
12643 		 */
12644 		plun->lun_state &= ~(FCP_LUN_CHANGED);
12645 		pptr->port_state |= FCP_STATE_IN_MDI;
12646 		mutex_exit(&plun->lun_mutex);
12647 		mutex_exit(&pptr->port_mutex);
12648 		if (mdi_pi_alloc_compatible(pdip, nname, plun->lun_guid,
12649 		    uaddr, compatible, ncompatible, 0, &pip) != MDI_SUCCESS) {
12650 			fcp_log(CE_WARN, pptr->port_dip,
12651 			    "!path alloc failed:0x%x", plun);
12652 			mutex_enter(&pptr->port_mutex);
12653 			mutex_enter(&plun->lun_mutex);
12654 			pptr->port_state &= ~FCP_STATE_IN_MDI;
12655 			failure++;
12656 			goto end_of_fcp_create_pip;
12657 		}
12658 		mutex_enter(&pptr->port_mutex);
12659 		mutex_enter(&plun->lun_mutex);
12660 		pptr->port_state &= ~FCP_STATE_IN_MDI;
12661 	} else {
12662 		(void) mdi_prop_remove(pip, NULL);
12663 	}
12664 
12665 	mdi_pi_set_phci_private(pip, (caddr_t)plun);
12666 
12667 	if (mdi_prop_update_byte_array(pip, NODE_WWN_PROP,
12668 	    ptgt->tgt_node_wwn.raw_wwn, FC_WWN_SIZE)
12669 	    != DDI_PROP_SUCCESS) {
12670 		failure++;
12671 		goto end_of_fcp_create_pip;
12672 	}
12673 
12674 	if (mdi_prop_update_byte_array(pip, PORT_WWN_PROP,
12675 	    ptgt->tgt_port_wwn.raw_wwn, FC_WWN_SIZE)
12676 	    != DDI_PROP_SUCCESS) {
12677 		failure++;
12678 		goto end_of_fcp_create_pip;
12679 	}
12680 
12681 	fcp_wwn_to_ascii(ptgt->tgt_port_wwn.raw_wwn, t_pwwn);
12682 	t_pwwn[16] = '\0';
12683 	if (mdi_prop_update_string(pip, TGT_PORT_PROP, t_pwwn)
12684 	    != DDI_PROP_SUCCESS) {
12685 		failure++;
12686 		goto end_of_fcp_create_pip;
12687 	}
12688 
12689 	/*
12690 	 * If there is no hard address - We might have to deal with
12691 	 * that by using WWN - Having said that it is important to
12692 	 * recognize this problem early so ssd can be informed of
12693 	 * the right interconnect type.
12694 	 */
12695 	if (!FC_TOP_EXTERNAL(pptr->port_topology) &&
12696 	    ptgt->tgt_hard_addr != 0) {
12697 		tgt_id = (uint32_t)
12698 		    fcp_alpa_to_switch[ptgt->tgt_hard_addr];
12699 	} else {
12700 		tgt_id = ptgt->tgt_d_id;
12701 	}
12702 
12703 	if (mdi_prop_update_int(pip, TARGET_PROP, tgt_id)
12704 	    != DDI_PROP_SUCCESS) {
12705 		failure++;
12706 		goto end_of_fcp_create_pip;
12707 	}
12708 
12709 	if (mdi_prop_update_int(pip, LUN_PROP, (int)plun->lun_num)
12710 	    != DDI_PROP_SUCCESS) {
12711 		failure++;
12712 		goto end_of_fcp_create_pip;
12713 	}
12714 	bcopy(&plun->lun_addr, &sam_lun, FCP_LUN_SIZE);
12715 	if (mdi_prop_update_int64(pip, SAM_LUN_PROP, sam_lun)
12716 	    != DDI_PROP_SUCCESS) {
12717 		failure++;
12718 		goto end_of_fcp_create_pip;
12719 	}
12720 
12721 end_of_fcp_create_pip:
12722 	scsi_hba_nodename_compatible_free(nname, compatible);
12723 
12724 	if (pip != NULL && failure) {
12725 		(void) mdi_prop_remove(pip, NULL);
12726 		mutex_exit(&plun->lun_mutex);
12727 		mutex_exit(&pptr->port_mutex);
12728 		(void) mdi_pi_free(pip, 0);
12729 		mutex_enter(&pptr->port_mutex);
12730 		mutex_enter(&plun->lun_mutex);
12731 		pip = NULL;
12732 	}
12733 
12734 	return (pip);
12735 }
12736 
12737 static dev_info_t *
12738 fcp_find_existing_dip(struct fcp_lun *plun, dev_info_t *pdip, caddr_t name)
12739 {
12740 	uint_t			nbytes;
12741 	uchar_t			*bytes;
12742 	uint_t			nwords;
12743 	uint32_t		tgt_id;
12744 	int			*words;
12745 	dev_info_t		*cdip;
12746 	dev_info_t		*ndip;
12747 	struct fcp_tgt	*ptgt = plun->lun_tgt;
12748 	struct fcp_port	*pptr = ptgt->tgt_port;
12749 	int			circular;
12750 
12751 	ndi_devi_enter(pdip, &circular);
12752 
12753 	ndip = (dev_info_t *)DEVI(pdip)->devi_child;
12754 	while ((cdip = ndip) != NULL) {
12755 		ndip = (dev_info_t *)DEVI(cdip)->devi_sibling;
12756 
12757 		if (strcmp(DEVI(cdip)->devi_node_name, name)) {
12758 			continue;
12759 		}
12760 
12761 		if (ddi_prop_lookup_byte_array(DDI_DEV_T_ANY, cdip,
12762 		    DDI_PROP_DONTPASS | DDI_PROP_NOTPROM, NODE_WWN_PROP, &bytes,
12763 		    &nbytes) != DDI_PROP_SUCCESS) {
12764 			continue;
12765 		}
12766 
12767 		if (nbytes != FC_WWN_SIZE || bytes == NULL) {
12768 			if (bytes != NULL) {
12769 				ddi_prop_free(bytes);
12770 			}
12771 			continue;
12772 		}
12773 		ASSERT(bytes != NULL);
12774 
12775 		if (bcmp(bytes, ptgt->tgt_node_wwn.raw_wwn, nbytes) != 0) {
12776 			ddi_prop_free(bytes);
12777 			continue;
12778 		}
12779 
12780 		ddi_prop_free(bytes);
12781 
12782 		if (ddi_prop_lookup_byte_array(DDI_DEV_T_ANY, cdip,
12783 		    DDI_PROP_DONTPASS | DDI_PROP_NOTPROM, PORT_WWN_PROP, &bytes,
12784 		    &nbytes) != DDI_PROP_SUCCESS) {
12785 			continue;
12786 		}
12787 
12788 		if (nbytes != FC_WWN_SIZE || bytes == NULL) {
12789 			if (bytes != NULL) {
12790 				ddi_prop_free(bytes);
12791 			}
12792 			continue;
12793 		}
12794 		ASSERT(bytes != NULL);
12795 
12796 		if (bcmp(bytes, ptgt->tgt_port_wwn.raw_wwn, nbytes) != 0) {
12797 			ddi_prop_free(bytes);
12798 			continue;
12799 		}
12800 
12801 		ddi_prop_free(bytes);
12802 
12803 		if (ddi_prop_lookup_int_array(DDI_DEV_T_ANY, cdip,
12804 		    DDI_PROP_DONTPASS | DDI_PROP_NOTPROM, TARGET_PROP, &words,
12805 		    &nwords) != DDI_PROP_SUCCESS) {
12806 			continue;
12807 		}
12808 
12809 		if (nwords != 1 || words == NULL) {
12810 			if (words != NULL) {
12811 				ddi_prop_free(words);
12812 			}
12813 			continue;
12814 		}
12815 		ASSERT(words != NULL);
12816 
12817 		/*
12818 		 * If there is no hard address - We might have to deal with
12819 		 * that by using WWN - Having said that it is important to
12820 		 * recognize this problem early so ssd can be informed of
12821 		 * the right interconnect type.
12822 		 */
12823 		if (!FC_TOP_EXTERNAL(pptr->port_topology) &&
12824 		    ptgt->tgt_hard_addr != 0) {
12825 			tgt_id =
12826 			    (uint32_t)fcp_alpa_to_switch[ptgt->tgt_hard_addr];
12827 		} else {
12828 			tgt_id = ptgt->tgt_d_id;
12829 		}
12830 
12831 		if (tgt_id != (uint32_t)*words) {
12832 			ddi_prop_free(words);
12833 			continue;
12834 		}
12835 		ddi_prop_free(words);
12836 
12837 		if (ddi_prop_lookup_int_array(DDI_DEV_T_ANY, cdip,
12838 		    DDI_PROP_DONTPASS | DDI_PROP_NOTPROM, LUN_PROP, &words,
12839 		    &nwords) != DDI_PROP_SUCCESS) {
12840 			continue;
12841 		}
12842 
12843 		if (nwords != 1 || words == NULL) {
12844 			if (words != NULL) {
12845 				ddi_prop_free(words);
12846 			}
12847 			continue;
12848 		}
12849 		ASSERT(words != NULL);
12850 
12851 		if (plun->lun_num == (uint16_t)*words) {
12852 			ddi_prop_free(words);
12853 			break;
12854 		}
12855 		ddi_prop_free(words);
12856 	}
12857 	ndi_devi_exit(pdip, circular);
12858 
12859 	return (cdip);
12860 }
12861 
12862 
12863 static int
12864 fcp_is_pip_present(struct fcp_lun *plun, mdi_pathinfo_t *pip)
12865 {
12866 	dev_info_t	*pdip;
12867 	char		buf[MAXNAMELEN];
12868 	char		uaddr[MAXNAMELEN];
12869 	int		rval = FC_FAILURE;
12870 
12871 	ASSERT(MUTEX_HELD(&plun->lun_mutex));
12872 
12873 	pdip = plun->lun_tgt->tgt_port->port_dip;
12874 
12875 	/*
12876 	 * Check if pip (and not plun->lun_cip) is NULL. plun->lun_cip can be
12877 	 * non-NULL even when the LUN is not there as in the case when a LUN is
12878 	 * configured and then deleted on the device end (for T3/T4 case). In
12879 	 * such cases, pip will be NULL.
12880 	 *
12881 	 * If the device generates an RSCN, it will end up getting offlined when
12882 	 * it disappeared and a new LUN will get created when it is rediscovered
12883 	 * on the device. If we check for lun_cip here, the LUN will not end
12884 	 * up getting onlined since this function will end up returning a
12885 	 * FC_SUCCESS.
12886 	 *
12887 	 * The behavior is different on other devices. For instance, on a HDS,
12888 	 * there was no RSCN generated by the device but the next I/O generated
12889 	 * a check condition and rediscovery got triggered that way. So, in
12890 	 * such cases, this path will not be exercised
12891 	 */
12892 	if (pip == NULL) {
12893 		FCP_TRACE(fcp_logq, LUN_PORT->port_instbuf,
12894 		    fcp_trace, FCP_BUF_LEVEL_4, 0,
12895 		    "fcp_is_pip_present: plun->lun_cip is NULL: "
12896 		    "plun: %p lun state: %x num: %d target state: %x",
12897 		    plun, plun->lun_state, plun->lun_num,
12898 		    plun->lun_tgt->tgt_port->port_state);
12899 		return (rval);
12900 	}
12901 
12902 	fcp_wwn_to_ascii(plun->lun_tgt->tgt_port_wwn.raw_wwn, buf);
12903 
12904 	(void) snprintf(uaddr, MAXNAMELEN, "w%s,%x", buf, plun->lun_num);
12905 
12906 	if (plun->lun_old_guid) {
12907 		if (mdi_pi_find(pdip, plun->lun_old_guid, uaddr) == pip) {
12908 			rval = FC_SUCCESS;
12909 		}
12910 	} else {
12911 		if (mdi_pi_find(pdip, plun->lun_guid, uaddr) == pip) {
12912 			rval = FC_SUCCESS;
12913 		}
12914 	}
12915 	return (rval);
12916 }
12917 
12918 static mdi_pathinfo_t *
12919 fcp_find_existing_pip(struct fcp_lun *plun, dev_info_t *pdip)
12920 {
12921 	char			buf[MAXNAMELEN];
12922 	char			uaddr[MAXNAMELEN];
12923 	mdi_pathinfo_t		*pip;
12924 	struct fcp_tgt	*ptgt = plun->lun_tgt;
12925 	struct fcp_port	*pptr = ptgt->tgt_port;
12926 
12927 	ASSERT(MUTEX_HELD(&pptr->port_mutex));
12928 
12929 	fcp_wwn_to_ascii(ptgt->tgt_port_wwn.raw_wwn, buf);
12930 	(void) snprintf(uaddr, MAXNAMELEN, "w%s,%x", buf, plun->lun_num);
12931 
12932 	pip = mdi_pi_find(pdip, plun->lun_guid, uaddr);
12933 
12934 	return (pip);
12935 }
12936 
12937 
12938 static int
12939 fcp_online_child(struct fcp_lun *plun, child_info_t *cip, int lcount,
12940     int tcount, int flags, int *circ)
12941 {
12942 	int			rval;
12943 	struct fcp_port		*pptr = plun->lun_tgt->tgt_port;
12944 	struct fcp_tgt	*ptgt = plun->lun_tgt;
12945 	dev_info_t		*cdip = NULL;
12946 
12947 	ASSERT(MUTEX_HELD(&pptr->port_mutex));
12948 	ASSERT(MUTEX_HELD(&plun->lun_mutex));
12949 
12950 	if (plun->lun_cip == NULL) {
12951 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
12952 		    fcp_trace, FCP_BUF_LEVEL_3, 0,
12953 		    "fcp_online_child: plun->lun_cip is NULL: "
12954 		    "plun: %p state: %x num: %d target state: %x",
12955 		    plun, plun->lun_state, plun->lun_num,
12956 		    plun->lun_tgt->tgt_port->port_state);
12957 		return (NDI_FAILURE);
12958 	}
12959 again:
12960 	if (plun->lun_mpxio == 0) {
12961 		cdip = DIP(cip);
12962 		mutex_exit(&plun->lun_mutex);
12963 		mutex_exit(&pptr->port_mutex);
12964 
12965 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
12966 		    fcp_trace, FCP_BUF_LEVEL_3, 0,
12967 		    "!Invoking ndi_devi_online for %s: target=%x lun=%x",
12968 		    ddi_get_name(cdip), ptgt->tgt_d_id, plun->lun_num);
12969 
12970 		/*
12971 		 * We could check for FCP_LUN_INIT here but chances
12972 		 * of getting here when it's already in FCP_LUN_INIT
12973 		 * is rare and a duplicate ndi_devi_online wouldn't
12974 		 * hurt either (as the node would already have been
12975 		 * in CF2)
12976 		 */
12977 		if (!i_ddi_devi_attached(ddi_get_parent(cdip))) {
12978 			rval = ndi_devi_bind_driver(cdip, flags);
12979 		} else {
12980 			rval = ndi_devi_online(cdip, flags);
12981 		}
12982 		/*
12983 		 * We log the message into trace buffer if the device
12984 		 * is "ses" and into syslog for any other device
12985 		 * type. This is to prevent the ndi_devi_online failure
12986 		 * message that appears for V880/A5K ses devices.
12987 		 */
12988 		if (rval == NDI_SUCCESS) {
12989 			mutex_enter(&ptgt->tgt_mutex);
12990 			plun->lun_state |= FCP_LUN_INIT;
12991 			mutex_exit(&ptgt->tgt_mutex);
12992 		} else if (strncmp(ddi_node_name(cdip), "ses", 3) != 0) {
12993 			fcp_log(CE_NOTE, pptr->port_dip,
12994 			    "!ndi_devi_online:"
12995 			    " failed for %s: target=%x lun=%x %x",
12996 			    ddi_get_name(cdip), ptgt->tgt_d_id,
12997 			    plun->lun_num, rval);
12998 		} else {
12999 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
13000 			    fcp_trace, FCP_BUF_LEVEL_3, 0,
13001 			    " !ndi_devi_online:"
13002 			    " failed for %s: target=%x lun=%x %x",
13003 			    ddi_get_name(cdip), ptgt->tgt_d_id,
13004 			    plun->lun_num, rval);
13005 		}
13006 	} else {
13007 		cdip = mdi_pi_get_client(PIP(cip));
13008 		mutex_exit(&plun->lun_mutex);
13009 		mutex_exit(&pptr->port_mutex);
13010 
13011 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
13012 		    fcp_trace, FCP_BUF_LEVEL_3, 0,
13013 		    "!Invoking mdi_pi_online for %s: target=%x lun=%x",
13014 		    ddi_get_name(cdip), ptgt->tgt_d_id, plun->lun_num);
13015 
13016 		/*
13017 		 * Hold path and exit phci to avoid deadlock with power
13018 		 * management code during mdi_pi_online.
13019 		 */
13020 		mdi_hold_path(PIP(cip));
13021 		mdi_devi_exit_phci(pptr->port_dip, *circ);
13022 
13023 		rval = mdi_pi_online(PIP(cip), flags);
13024 
13025 		mdi_devi_enter_phci(pptr->port_dip, circ);
13026 		mdi_rele_path(PIP(cip));
13027 
13028 		if (rval == MDI_SUCCESS) {
13029 			mutex_enter(&ptgt->tgt_mutex);
13030 			plun->lun_state |= FCP_LUN_INIT;
13031 			mutex_exit(&ptgt->tgt_mutex);
13032 
13033 			/*
13034 			 * Clear MPxIO path permanent disable in case
13035 			 * fcp hotplug dropped the offline event.
13036 			 */
13037 			(void) mdi_pi_enable_path(PIP(cip), DRIVER_DISABLE);
13038 
13039 		} else if (rval == MDI_NOT_SUPPORTED) {
13040 			child_info_t	*old_cip = cip;
13041 
13042 			/*
13043 			 * MPxIO does not support this device yet.
13044 			 * Enumerate in legacy mode.
13045 			 */
13046 			mutex_enter(&pptr->port_mutex);
13047 			mutex_enter(&plun->lun_mutex);
13048 			plun->lun_mpxio = 0;
13049 			plun->lun_cip = NULL;
13050 			cdip = fcp_create_dip(plun, lcount, tcount);
13051 			plun->lun_cip = cip = CIP(cdip);
13052 			if (cip == NULL) {
13053 				fcp_log(CE_WARN, pptr->port_dip,
13054 				    "!fcp_online_child: "
13055 				    "Create devinfo failed for LU=%p", plun);
13056 				mutex_exit(&plun->lun_mutex);
13057 
13058 				mutex_enter(&ptgt->tgt_mutex);
13059 				plun->lun_state |= FCP_LUN_OFFLINE;
13060 				mutex_exit(&ptgt->tgt_mutex);
13061 
13062 				mutex_exit(&pptr->port_mutex);
13063 
13064 				/*
13065 				 * free the mdi_pathinfo node
13066 				 */
13067 				(void) mdi_pi_free(PIP(old_cip), 0);
13068 			} else {
13069 				FCP_TRACE(fcp_logq, pptr->port_instbuf,
13070 				    fcp_trace, FCP_BUF_LEVEL_3, 0,
13071 				    "fcp_online_child: creating devinfo "
13072 				    "node 0x%p for plun 0x%p",
13073 				    cip, plun);
13074 				mutex_exit(&plun->lun_mutex);
13075 				mutex_exit(&pptr->port_mutex);
13076 				/*
13077 				 * free the mdi_pathinfo node
13078 				 */
13079 				(void) mdi_pi_free(PIP(old_cip), 0);
13080 				mutex_enter(&pptr->port_mutex);
13081 				mutex_enter(&plun->lun_mutex);
13082 				goto again;
13083 			}
13084 		} else {
13085 			if (cdip) {
13086 				fcp_log(CE_NOTE, pptr->port_dip,
13087 				    "!fcp_online_child: mdi_pi_online:"
13088 				    " failed for %s: target=%x lun=%x %x",
13089 				    ddi_get_name(cdip), ptgt->tgt_d_id,
13090 				    plun->lun_num, rval);
13091 			}
13092 		}
13093 		rval = (rval == MDI_SUCCESS) ? NDI_SUCCESS : NDI_FAILURE;
13094 	}
13095 
13096 	if (rval == NDI_SUCCESS) {
13097 		if (cdip) {
13098 			(void) ndi_event_retrieve_cookie(
13099 			    pptr->port_ndi_event_hdl, cdip, FCAL_INSERT_EVENT,
13100 			    &fcp_insert_eid, NDI_EVENT_NOPASS);
13101 			(void) ndi_event_run_callbacks(pptr->port_ndi_event_hdl,
13102 			    cdip, fcp_insert_eid, NULL);
13103 		}
13104 	}
13105 	mutex_enter(&pptr->port_mutex);
13106 	mutex_enter(&plun->lun_mutex);
13107 	return (rval);
13108 }
13109 
13110 /* ARGSUSED */
13111 static int
13112 fcp_offline_child(struct fcp_lun *plun, child_info_t *cip, int lcount,
13113     int tcount, int flags, int *circ)
13114 {
13115 	int rval;
13116 	struct fcp_port		*pptr = plun->lun_tgt->tgt_port;
13117 	struct fcp_tgt	*ptgt = plun->lun_tgt;
13118 	dev_info_t		*cdip;
13119 
13120 	ASSERT(MUTEX_HELD(&plun->lun_mutex));
13121 	ASSERT(MUTEX_HELD(&pptr->port_mutex));
13122 
13123 	if (plun->lun_cip == NULL) {
13124 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
13125 		    fcp_trace, FCP_BUF_LEVEL_3, 0,
13126 		    "fcp_offline_child: plun->lun_cip is NULL: "
13127 		    "plun: %p lun state: %x num: %d target state: %x",
13128 		    plun, plun->lun_state, plun->lun_num,
13129 		    plun->lun_tgt->tgt_port->port_state);
13130 		return (NDI_FAILURE);
13131 	}
13132 
13133 	if (plun->lun_mpxio == 0) {
13134 		cdip = DIP(cip);
13135 		mutex_exit(&plun->lun_mutex);
13136 		mutex_exit(&pptr->port_mutex);
13137 		rval = ndi_devi_offline(DIP(cip), flags);
13138 		if (rval != NDI_SUCCESS) {
13139 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
13140 			    fcp_trace, FCP_BUF_LEVEL_3, 0,
13141 			    "fcp_offline_child: ndi_devi_offline failed "
13142 			    "rval=%x cip=%p", rval, cip);
13143 		}
13144 	} else {
13145 		cdip = mdi_pi_get_client(PIP(cip));
13146 		mutex_exit(&plun->lun_mutex);
13147 		mutex_exit(&pptr->port_mutex);
13148 
13149 		/*
13150 		 * Exit phci to avoid deadlock with power management code
13151 		 * during mdi_pi_offline
13152 		 */
13153 		mdi_hold_path(PIP(cip));
13154 		mdi_devi_exit_phci(pptr->port_dip, *circ);
13155 
13156 		rval = mdi_pi_offline(PIP(cip), flags);
13157 
13158 		mdi_devi_enter_phci(pptr->port_dip, circ);
13159 		mdi_rele_path(PIP(cip));
13160 
13161 		if (rval == MDI_SUCCESS) {
13162 			/*
13163 			 * Clear MPxIO path permanent disable as the path is
13164 			 * already offlined.
13165 			 */
13166 			(void) mdi_pi_enable_path(PIP(cip), DRIVER_DISABLE);
13167 
13168 			if (flags & NDI_DEVI_REMOVE) {
13169 				(void) mdi_pi_free(PIP(cip), 0);
13170 			}
13171 		} else {
13172 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
13173 			    fcp_trace, FCP_BUF_LEVEL_3, 0,
13174 			    "fcp_offline_child: mdi_pi_offline failed "
13175 			    "rval=%x cip=%p", rval, cip);
13176 		}
13177 		rval = (rval == MDI_SUCCESS) ? NDI_SUCCESS : NDI_FAILURE;
13178 	}
13179 
13180 	mutex_enter(&ptgt->tgt_mutex);
13181 	plun->lun_state &= ~FCP_LUN_INIT;
13182 	mutex_exit(&ptgt->tgt_mutex);
13183 
13184 	mutex_enter(&pptr->port_mutex);
13185 	mutex_enter(&plun->lun_mutex);
13186 
13187 	if (rval == NDI_SUCCESS) {
13188 		cdip = NULL;
13189 		if (flags & NDI_DEVI_REMOVE) {
13190 			/*
13191 			 * If the guid of the LUN changes, lun_cip will not
13192 			 * equal to cip, and after offlining the LUN with the
13193 			 * old guid, we should keep lun_cip since it's the cip
13194 			 * of the LUN with the new guid.
13195 			 * Otherwise remove our reference to child node.
13196 			 */
13197 			if (plun->lun_cip == cip) {
13198 				plun->lun_cip = NULL;
13199 			}
13200 			if (plun->lun_old_guid) {
13201 				kmem_free(plun->lun_old_guid,
13202 				    plun->lun_old_guid_size);
13203 				plun->lun_old_guid = NULL;
13204 				plun->lun_old_guid_size = 0;
13205 			}
13206 		}
13207 	}
13208 
13209 	if (cdip) {
13210 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
13211 		    fcp_trace, FCP_BUF_LEVEL_3, 0, "!%s failed for %s:"
13212 		    " target=%x lun=%x", "ndi_offline",
13213 		    ddi_get_name(cdip), ptgt->tgt_d_id, plun->lun_num);
13214 	}
13215 
13216 	return (rval);
13217 }
13218 
13219 static void
13220 fcp_remove_child(struct fcp_lun *plun)
13221 {
13222 	ASSERT(MUTEX_HELD(&plun->lun_mutex));
13223 
13224 	if (fcp_is_child_present(plun, plun->lun_cip) == FC_SUCCESS) {
13225 		if (plun->lun_mpxio == 0) {
13226 			(void) ndi_prop_remove_all(DIP(plun->lun_cip));
13227 			(void) ndi_devi_free(DIP(plun->lun_cip));
13228 		} else {
13229 			mutex_exit(&plun->lun_mutex);
13230 			mutex_exit(&plun->lun_tgt->tgt_mutex);
13231 			mutex_exit(&plun->lun_tgt->tgt_port->port_mutex);
13232 			FCP_TRACE(fcp_logq,
13233 			    plun->lun_tgt->tgt_port->port_instbuf,
13234 			    fcp_trace, FCP_BUF_LEVEL_3, 0,
13235 			    "lun=%p pip freed %p", plun, plun->lun_cip);
13236 			(void) mdi_prop_remove(PIP(plun->lun_cip), NULL);
13237 			(void) mdi_pi_free(PIP(plun->lun_cip), 0);
13238 			mutex_enter(&plun->lun_tgt->tgt_port->port_mutex);
13239 			mutex_enter(&plun->lun_tgt->tgt_mutex);
13240 			mutex_enter(&plun->lun_mutex);
13241 		}
13242 	}
13243 
13244 	plun->lun_cip = NULL;
13245 }
13246 
13247 /*
13248  * called when a timeout occurs
13249  *
13250  * can be scheduled during an attach or resume (if not already running)
13251  *
13252  * one timeout is set up for all ports
13253  *
13254  * acquires and releases the global mutex
13255  */
13256 /*ARGSUSED*/
13257 static void
13258 fcp_watch(void *arg)
13259 {
13260 	struct fcp_port	*pptr;
13261 	struct fcp_ipkt	*icmd;
13262 	struct fcp_ipkt	*nicmd;
13263 	struct fcp_pkt	*cmd;
13264 	struct fcp_pkt	*ncmd;
13265 	struct fcp_pkt	*tail;
13266 	struct fcp_pkt	*pcmd;
13267 	struct fcp_pkt	*save_head;
13268 	struct fcp_port	*save_port;
13269 
13270 	/* increment global watchdog time */
13271 	fcp_watchdog_time += fcp_watchdog_timeout;
13272 
13273 	mutex_enter(&fcp_global_mutex);
13274 
13275 	/* scan each port in our list */
13276 	for (pptr = fcp_port_head; pptr != NULL; pptr = pptr->port_next) {
13277 		save_port = fcp_port_head;
13278 		pptr->port_state |= FCP_STATE_IN_WATCHDOG;
13279 		mutex_exit(&fcp_global_mutex);
13280 
13281 		mutex_enter(&pptr->port_mutex);
13282 		if (pptr->port_ipkt_list == NULL &&
13283 		    (pptr->port_state & (FCP_STATE_SUSPENDED |
13284 		    FCP_STATE_DETACHING | FCP_STATE_POWER_DOWN))) {
13285 			pptr->port_state &= ~FCP_STATE_IN_WATCHDOG;
13286 			mutex_exit(&pptr->port_mutex);
13287 			mutex_enter(&fcp_global_mutex);
13288 			goto end_of_watchdog;
13289 		}
13290 
13291 		/*
13292 		 * We check if a list of targets need to be offlined.
13293 		 */
13294 		if (pptr->port_offline_tgts) {
13295 			fcp_scan_offline_tgts(pptr);
13296 		}
13297 
13298 		/*
13299 		 * We check if a list of luns need to be offlined.
13300 		 */
13301 		if (pptr->port_offline_luns) {
13302 			fcp_scan_offline_luns(pptr);
13303 		}
13304 
13305 		/*
13306 		 * We check if a list of targets or luns need to be reset.
13307 		 */
13308 		if (pptr->port_reset_list) {
13309 			fcp_check_reset_delay(pptr);
13310 		}
13311 
13312 		mutex_exit(&pptr->port_mutex);
13313 
13314 		/*
13315 		 * This is where the pending commands (pkt) are checked for
13316 		 * timeout.
13317 		 */
13318 		mutex_enter(&pptr->port_pkt_mutex);
13319 		tail = pptr->port_pkt_tail;
13320 
13321 		for (pcmd = NULL, cmd = pptr->port_pkt_head;
13322 		    cmd != NULL; cmd = ncmd) {
13323 			ncmd = cmd->cmd_next;
13324 			/*
13325 			 * If a command is in this queue the bit CFLAG_IN_QUEUE
13326 			 * must be set.
13327 			 */
13328 			ASSERT(cmd->cmd_flags & CFLAG_IN_QUEUE);
13329 			/*
13330 			 * FCP_INVALID_TIMEOUT will be set for those
13331 			 * command that need to be failed. Mostly those
13332 			 * cmds that could not be queued down for the
13333 			 * "timeout" value. cmd->cmd_timeout is used
13334 			 * to try and requeue the command regularly.
13335 			 */
13336 			if (cmd->cmd_timeout >= fcp_watchdog_time) {
13337 				/*
13338 				 * This command hasn't timed out yet.  Let's
13339 				 * go to the next one.
13340 				 */
13341 				pcmd = cmd;
13342 				goto end_of_loop;
13343 			}
13344 
13345 			if (cmd == pptr->port_pkt_head) {
13346 				ASSERT(pcmd == NULL);
13347 				pptr->port_pkt_head = cmd->cmd_next;
13348 			} else {
13349 				ASSERT(pcmd != NULL);
13350 				pcmd->cmd_next = cmd->cmd_next;
13351 			}
13352 
13353 			if (cmd == pptr->port_pkt_tail) {
13354 				ASSERT(cmd->cmd_next == NULL);
13355 				pptr->port_pkt_tail = pcmd;
13356 				if (pcmd) {
13357 					pcmd->cmd_next = NULL;
13358 				}
13359 			}
13360 			cmd->cmd_next = NULL;
13361 
13362 			/*
13363 			 * save the current head before dropping the
13364 			 * mutex - If the head doesn't remain the
13365 			 * same after re acquiring the mutex, just
13366 			 * bail out and revisit on next tick.
13367 			 *
13368 			 * PS: The tail pointer can change as the commands
13369 			 * get requeued after failure to retransport
13370 			 */
13371 			save_head = pptr->port_pkt_head;
13372 			mutex_exit(&pptr->port_pkt_mutex);
13373 
13374 			if (cmd->cmd_fp_pkt->pkt_timeout ==
13375 			    FCP_INVALID_TIMEOUT) {
13376 				struct scsi_pkt		*pkt = cmd->cmd_pkt;
13377 				struct fcp_lun	*plun;
13378 				struct fcp_tgt	*ptgt;
13379 
13380 				plun = ADDR2LUN(&pkt->pkt_address);
13381 				ptgt = plun->lun_tgt;
13382 
13383 				FCP_TRACE(fcp_logq, pptr->port_instbuf,
13384 				    fcp_trace, FCP_BUF_LEVEL_2, 0,
13385 				    "SCSI cmd 0x%x to D_ID=%x timed out",
13386 				    pkt->pkt_cdbp[0], ptgt->tgt_d_id);
13387 
13388 				cmd->cmd_state == FCP_PKT_ABORTING ?
13389 				    fcp_fail_cmd(cmd, CMD_RESET,
13390 				    STAT_DEV_RESET) : fcp_fail_cmd(cmd,
13391 				    CMD_TIMEOUT, STAT_ABORTED);
13392 			} else {
13393 				fcp_retransport_cmd(pptr, cmd);
13394 			}
13395 			mutex_enter(&pptr->port_pkt_mutex);
13396 			if (save_head && save_head != pptr->port_pkt_head) {
13397 				/*
13398 				 * Looks like linked list got changed (mostly
13399 				 * happens when an an OFFLINE LUN code starts
13400 				 * returning overflow queue commands in
13401 				 * parallel. So bail out and revisit during
13402 				 * next tick
13403 				 */
13404 				break;
13405 			}
13406 		end_of_loop:
13407 			/*
13408 			 * Scan only upto the previously known tail pointer
13409 			 * to avoid excessive processing - lots of new packets
13410 			 * could have been added to the tail or the old ones
13411 			 * re-queued.
13412 			 */
13413 			if (cmd == tail) {
13414 				break;
13415 			}
13416 		}
13417 		mutex_exit(&pptr->port_pkt_mutex);
13418 
13419 		mutex_enter(&pptr->port_mutex);
13420 		for (icmd = pptr->port_ipkt_list; icmd != NULL; icmd = nicmd) {
13421 			struct fcp_tgt *ptgt = icmd->ipkt_tgt;
13422 
13423 			nicmd = icmd->ipkt_next;
13424 			if ((icmd->ipkt_restart != 0) &&
13425 			    (icmd->ipkt_restart >= fcp_watchdog_time)) {
13426 				/* packet has not timed out */
13427 				continue;
13428 			}
13429 
13430 			/* time for packet re-transport */
13431 			if (icmd == pptr->port_ipkt_list) {
13432 				pptr->port_ipkt_list = icmd->ipkt_next;
13433 				if (pptr->port_ipkt_list) {
13434 					pptr->port_ipkt_list->ipkt_prev =
13435 					    NULL;
13436 				}
13437 			} else {
13438 				icmd->ipkt_prev->ipkt_next = icmd->ipkt_next;
13439 				if (icmd->ipkt_next) {
13440 					icmd->ipkt_next->ipkt_prev =
13441 					    icmd->ipkt_prev;
13442 				}
13443 			}
13444 			icmd->ipkt_next = NULL;
13445 			icmd->ipkt_prev = NULL;
13446 			mutex_exit(&pptr->port_mutex);
13447 
13448 			if (fcp_is_retryable(icmd)) {
13449 				fc_ulp_rscn_info_t *rscnp =
13450 				    (fc_ulp_rscn_info_t *)icmd->ipkt_fpkt->
13451 				    pkt_ulp_rscn_infop;
13452 
13453 				FCP_TRACE(fcp_logq, pptr->port_instbuf,
13454 				    fcp_trace, FCP_BUF_LEVEL_2, 0,
13455 				    "%x to D_ID=%x Retrying..",
13456 				    icmd->ipkt_opcode,
13457 				    icmd->ipkt_fpkt->pkt_cmd_fhdr.d_id);
13458 
13459 				/*
13460 				 * Update the RSCN count in the packet
13461 				 * before resending.
13462 				 */
13463 
13464 				if (rscnp != NULL) {
13465 					rscnp->ulp_rscn_count =
13466 					    fc_ulp_get_rscn_count(pptr->
13467 					    port_fp_handle);
13468 				}
13469 
13470 				mutex_enter(&pptr->port_mutex);
13471 				mutex_enter(&ptgt->tgt_mutex);
13472 				if (!FCP_STATE_CHANGED(pptr, ptgt, icmd)) {
13473 					mutex_exit(&ptgt->tgt_mutex);
13474 					mutex_exit(&pptr->port_mutex);
13475 					switch (icmd->ipkt_opcode) {
13476 						int rval;
13477 					case LA_ELS_PLOGI:
13478 						if ((rval = fc_ulp_login(
13479 						    pptr->port_fp_handle,
13480 						    &icmd->ipkt_fpkt, 1)) ==
13481 						    FC_SUCCESS) {
13482 							mutex_enter(
13483 							    &pptr->port_mutex);
13484 							continue;
13485 						}
13486 						if (fcp_handle_ipkt_errors(
13487 						    pptr, ptgt, icmd, rval,
13488 						    "PLOGI") == DDI_SUCCESS) {
13489 							mutex_enter(
13490 							    &pptr->port_mutex);
13491 							continue;
13492 						}
13493 						break;
13494 
13495 					case LA_ELS_PRLI:
13496 						if ((rval = fc_ulp_issue_els(
13497 						    pptr->port_fp_handle,
13498 						    icmd->ipkt_fpkt)) ==
13499 						    FC_SUCCESS) {
13500 							mutex_enter(
13501 							    &pptr->port_mutex);
13502 							continue;
13503 						}
13504 						if (fcp_handle_ipkt_errors(
13505 						    pptr, ptgt, icmd, rval,
13506 						    "PRLI") == DDI_SUCCESS) {
13507 							mutex_enter(
13508 							    &pptr->port_mutex);
13509 							continue;
13510 						}
13511 						break;
13512 
13513 					default:
13514 						if ((rval = fcp_transport(
13515 						    pptr->port_fp_handle,
13516 						    icmd->ipkt_fpkt, 1)) ==
13517 						    FC_SUCCESS) {
13518 							mutex_enter(
13519 							    &pptr->port_mutex);
13520 							continue;
13521 						}
13522 						if (fcp_handle_ipkt_errors(
13523 						    pptr, ptgt, icmd, rval,
13524 						    "PRLI") == DDI_SUCCESS) {
13525 							mutex_enter(
13526 							    &pptr->port_mutex);
13527 							continue;
13528 						}
13529 						break;
13530 					}
13531 				} else {
13532 					mutex_exit(&ptgt->tgt_mutex);
13533 					mutex_exit(&pptr->port_mutex);
13534 				}
13535 			} else {
13536 				fcp_print_error(icmd->ipkt_fpkt);
13537 			}
13538 
13539 			(void) fcp_call_finish_init(pptr, ptgt,
13540 			    icmd->ipkt_link_cnt, icmd->ipkt_change_cnt,
13541 			    icmd->ipkt_cause);
13542 			fcp_icmd_free(pptr, icmd);
13543 			mutex_enter(&pptr->port_mutex);
13544 		}
13545 
13546 		pptr->port_state &= ~FCP_STATE_IN_WATCHDOG;
13547 		mutex_exit(&pptr->port_mutex);
13548 		mutex_enter(&fcp_global_mutex);
13549 
13550 	end_of_watchdog:
13551 		/*
13552 		 * Bail out early before getting into trouble
13553 		 */
13554 		if (save_port != fcp_port_head) {
13555 			break;
13556 		}
13557 	}
13558 
13559 	if (fcp_watchdog_init > 0) {
13560 		/* reschedule timeout to go again */
13561 		fcp_watchdog_id =
13562 		    timeout(fcp_watch, NULL, fcp_watchdog_tick);
13563 	}
13564 	mutex_exit(&fcp_global_mutex);
13565 }
13566 
13567 
13568 static void
13569 fcp_check_reset_delay(struct fcp_port *pptr)
13570 {
13571 	uint32_t		tgt_cnt;
13572 	int			level;
13573 	struct fcp_tgt	*ptgt;
13574 	struct fcp_lun	*plun;
13575 	struct fcp_reset_elem *cur = NULL;
13576 	struct fcp_reset_elem *next = NULL;
13577 	struct fcp_reset_elem *prev = NULL;
13578 
13579 	ASSERT(mutex_owned(&pptr->port_mutex));
13580 
13581 	next = pptr->port_reset_list;
13582 	while ((cur = next) != NULL) {
13583 		next = cur->next;
13584 
13585 		if (cur->timeout < fcp_watchdog_time) {
13586 			prev = cur;
13587 			continue;
13588 		}
13589 
13590 		ptgt = cur->tgt;
13591 		plun = cur->lun;
13592 		tgt_cnt = cur->tgt_cnt;
13593 
13594 		if (ptgt) {
13595 			level = RESET_TARGET;
13596 		} else {
13597 			ASSERT(plun != NULL);
13598 			level = RESET_LUN;
13599 			ptgt = plun->lun_tgt;
13600 		}
13601 		if (prev) {
13602 			prev->next = next;
13603 		} else {
13604 			/*
13605 			 * Because we drop port mutex while doing aborts for
13606 			 * packets, we can't rely on reset_list pointing to
13607 			 * our head
13608 			 */
13609 			if (cur == pptr->port_reset_list) {
13610 				pptr->port_reset_list = next;
13611 			} else {
13612 				struct fcp_reset_elem *which;
13613 
13614 				which = pptr->port_reset_list;
13615 				while (which && which->next != cur) {
13616 					which = which->next;
13617 				}
13618 				ASSERT(which != NULL);
13619 
13620 				which->next = next;
13621 				prev = which;
13622 			}
13623 		}
13624 
13625 		kmem_free(cur, sizeof (*cur));
13626 
13627 		if (tgt_cnt == ptgt->tgt_change_cnt) {
13628 			mutex_enter(&ptgt->tgt_mutex);
13629 			if (level == RESET_TARGET) {
13630 				fcp_update_tgt_state(ptgt,
13631 				    FCP_RESET, FCP_LUN_BUSY);
13632 			} else {
13633 				fcp_update_lun_state(plun,
13634 				    FCP_RESET, FCP_LUN_BUSY);
13635 			}
13636 			mutex_exit(&ptgt->tgt_mutex);
13637 
13638 			mutex_exit(&pptr->port_mutex);
13639 			fcp_abort_all(pptr, ptgt, plun, tgt_cnt);
13640 			mutex_enter(&pptr->port_mutex);
13641 		}
13642 	}
13643 }
13644 
13645 
13646 static void
13647 fcp_abort_all(struct fcp_port *pptr, struct fcp_tgt *ttgt,
13648     struct fcp_lun *rlun, int tgt_cnt)
13649 {
13650 	int			rval;
13651 	struct fcp_lun	*tlun, *nlun;
13652 	struct fcp_pkt	*pcmd = NULL, *ncmd = NULL,
13653 	    *cmd = NULL, *head = NULL,
13654 	    *tail = NULL;
13655 
13656 	mutex_enter(&pptr->port_pkt_mutex);
13657 	for (cmd = pptr->port_pkt_head; cmd != NULL; cmd = ncmd) {
13658 		struct fcp_lun *plun = ADDR2LUN(&cmd->cmd_pkt->pkt_address);
13659 		struct fcp_tgt *ptgt = plun->lun_tgt;
13660 
13661 		ncmd = cmd->cmd_next;
13662 
13663 		if (ptgt != ttgt && plun != rlun) {
13664 			pcmd = cmd;
13665 			continue;
13666 		}
13667 
13668 		if (pcmd != NULL) {
13669 			ASSERT(pptr->port_pkt_head != cmd);
13670 			pcmd->cmd_next = ncmd;
13671 		} else {
13672 			ASSERT(cmd == pptr->port_pkt_head);
13673 			pptr->port_pkt_head = ncmd;
13674 		}
13675 		if (pptr->port_pkt_tail == cmd) {
13676 			ASSERT(cmd->cmd_next == NULL);
13677 			pptr->port_pkt_tail = pcmd;
13678 			if (pcmd != NULL) {
13679 				pcmd->cmd_next = NULL;
13680 			}
13681 		}
13682 
13683 		if (head == NULL) {
13684 			head = tail = cmd;
13685 		} else {
13686 			ASSERT(tail != NULL);
13687 			tail->cmd_next = cmd;
13688 			tail = cmd;
13689 		}
13690 		cmd->cmd_next = NULL;
13691 	}
13692 	mutex_exit(&pptr->port_pkt_mutex);
13693 
13694 	for (cmd = head; cmd != NULL; cmd = ncmd) {
13695 		struct scsi_pkt *pkt = cmd->cmd_pkt;
13696 
13697 		ncmd = cmd->cmd_next;
13698 		ASSERT(pkt != NULL);
13699 
13700 		mutex_enter(&pptr->port_mutex);
13701 		if (ttgt->tgt_change_cnt == tgt_cnt) {
13702 			mutex_exit(&pptr->port_mutex);
13703 			cmd->cmd_flags &= ~CFLAG_IN_QUEUE;
13704 			pkt->pkt_reason = CMD_RESET;
13705 			pkt->pkt_statistics |= STAT_DEV_RESET;
13706 			cmd->cmd_state = FCP_PKT_IDLE;
13707 			fcp_post_callback(cmd);
13708 		} else {
13709 			mutex_exit(&pptr->port_mutex);
13710 		}
13711 	}
13712 
13713 	/*
13714 	 * If the FCA will return all the commands in its queue then our
13715 	 * work is easy, just return.
13716 	 */
13717 
13718 	if (pptr->port_reset_action == FC_RESET_RETURN_ALL) {
13719 		return;
13720 	}
13721 
13722 	/*
13723 	 * For RESET_LUN get hold of target pointer
13724 	 */
13725 	if (ttgt == NULL) {
13726 		ASSERT(rlun != NULL);
13727 
13728 		ttgt = rlun->lun_tgt;
13729 
13730 		ASSERT(ttgt != NULL);
13731 	}
13732 
13733 	/*
13734 	 * There are some severe race conditions here.
13735 	 * While we are trying to abort the pkt, it might be completing
13736 	 * so mark it aborted and if the abort does not succeed then
13737 	 * handle it in the watch thread.
13738 	 */
13739 	mutex_enter(&ttgt->tgt_mutex);
13740 	nlun = ttgt->tgt_lun;
13741 	mutex_exit(&ttgt->tgt_mutex);
13742 	while ((tlun = nlun) != NULL) {
13743 		int restart = 0;
13744 		if (rlun && rlun != tlun) {
13745 			mutex_enter(&ttgt->tgt_mutex);
13746 			nlun = tlun->lun_next;
13747 			mutex_exit(&ttgt->tgt_mutex);
13748 			continue;
13749 		}
13750 		mutex_enter(&tlun->lun_mutex);
13751 		cmd = tlun->lun_pkt_head;
13752 		while (cmd != NULL) {
13753 			if (cmd->cmd_state == FCP_PKT_ISSUED) {
13754 				struct scsi_pkt *pkt;
13755 
13756 				restart = 1;
13757 				cmd->cmd_state = FCP_PKT_ABORTING;
13758 				mutex_exit(&tlun->lun_mutex);
13759 				rval = fc_ulp_abort(pptr->port_fp_handle,
13760 				    cmd->cmd_fp_pkt, KM_SLEEP);
13761 				if (rval == FC_SUCCESS) {
13762 					pkt = cmd->cmd_pkt;
13763 					pkt->pkt_reason = CMD_RESET;
13764 					pkt->pkt_statistics |= STAT_DEV_RESET;
13765 					cmd->cmd_state = FCP_PKT_IDLE;
13766 					fcp_post_callback(cmd);
13767 				} else {
13768 					caddr_t msg;
13769 
13770 					(void) fc_ulp_error(rval, &msg);
13771 
13772 					/*
13773 					 * This part is tricky. The abort
13774 					 * failed and now the command could
13775 					 * be completing.  The cmd_state ==
13776 					 * FCP_PKT_ABORTING should save
13777 					 * us in fcp_cmd_callback. If we
13778 					 * are already aborting ignore the
13779 					 * command in fcp_cmd_callback.
13780 					 * Here we leave this packet for 20
13781 					 * sec to be aborted in the
13782 					 * fcp_watch thread.
13783 					 */
13784 					fcp_log(CE_WARN, pptr->port_dip,
13785 					    "!Abort failed after reset %s",
13786 					    msg);
13787 
13788 					cmd->cmd_timeout =
13789 					    fcp_watchdog_time +
13790 					    cmd->cmd_pkt->pkt_time +
13791 					    FCP_FAILED_DELAY;
13792 
13793 					cmd->cmd_fp_pkt->pkt_timeout =
13794 					    FCP_INVALID_TIMEOUT;
13795 					/*
13796 					 * This is a hack, cmd is put in the
13797 					 * overflow queue so that it can be
13798 					 * timed out finally
13799 					 */
13800 					cmd->cmd_flags |= CFLAG_IN_QUEUE;
13801 
13802 					mutex_enter(&pptr->port_pkt_mutex);
13803 					if (pptr->port_pkt_head) {
13804 						ASSERT(pptr->port_pkt_tail
13805 						    != NULL);
13806 						pptr->port_pkt_tail->cmd_next
13807 						    = cmd;
13808 						pptr->port_pkt_tail = cmd;
13809 					} else {
13810 						ASSERT(pptr->port_pkt_tail
13811 						    == NULL);
13812 						pptr->port_pkt_head =
13813 						    pptr->port_pkt_tail
13814 						    = cmd;
13815 					}
13816 					cmd->cmd_next = NULL;
13817 					mutex_exit(&pptr->port_pkt_mutex);
13818 				}
13819 				mutex_enter(&tlun->lun_mutex);
13820 				cmd = tlun->lun_pkt_head;
13821 			} else {
13822 				cmd = cmd->cmd_forw;
13823 			}
13824 		}
13825 		mutex_exit(&tlun->lun_mutex);
13826 
13827 		mutex_enter(&ttgt->tgt_mutex);
13828 		restart == 1 ? (nlun = ttgt->tgt_lun) : (nlun = tlun->lun_next);
13829 		mutex_exit(&ttgt->tgt_mutex);
13830 
13831 		mutex_enter(&pptr->port_mutex);
13832 		if (tgt_cnt != ttgt->tgt_change_cnt) {
13833 			mutex_exit(&pptr->port_mutex);
13834 			return;
13835 		} else {
13836 			mutex_exit(&pptr->port_mutex);
13837 		}
13838 	}
13839 }
13840 
13841 
13842 /*
13843  * unlink the soft state, returning the soft state found (if any)
13844  *
13845  * acquires and releases the global mutex
13846  */
13847 struct fcp_port *
13848 fcp_soft_state_unlink(struct fcp_port *pptr)
13849 {
13850 	struct fcp_port	*hptr;		/* ptr index */
13851 	struct fcp_port	*tptr;		/* prev hptr */
13852 
13853 	mutex_enter(&fcp_global_mutex);
13854 	for (hptr = fcp_port_head, tptr = NULL;
13855 	    hptr != NULL;
13856 	    tptr = hptr, hptr = hptr->port_next) {
13857 		if (hptr == pptr) {
13858 			/* we found a match -- remove this item */
13859 			if (tptr == NULL) {
13860 				/* we're at the head of the list */
13861 				fcp_port_head = hptr->port_next;
13862 			} else {
13863 				tptr->port_next = hptr->port_next;
13864 			}
13865 			break;			/* success */
13866 		}
13867 	}
13868 	if (fcp_port_head == NULL) {
13869 		fcp_cleanup_blacklist(&fcp_lun_blacklist);
13870 	}
13871 	mutex_exit(&fcp_global_mutex);
13872 	return (hptr);
13873 }
13874 
13875 
13876 /*
13877  * called by fcp_scsi_hba_tgt_init to find a LUN given a
13878  * WWN and a LUN number
13879  */
13880 /* ARGSUSED */
13881 static struct fcp_lun *
13882 fcp_lookup_lun(struct fcp_port *pptr, uchar_t *wwn, uint16_t lun)
13883 {
13884 	int hash;
13885 	struct fcp_tgt *ptgt;
13886 	struct fcp_lun *plun;
13887 
13888 	ASSERT(mutex_owned(&pptr->port_mutex));
13889 
13890 	hash = FCP_HASH(wwn);
13891 	for (ptgt = pptr->port_tgt_hash_table[hash]; ptgt != NULL;
13892 	    ptgt = ptgt->tgt_next) {
13893 		if (bcmp((caddr_t)wwn, (caddr_t)&ptgt->tgt_port_wwn.raw_wwn[0],
13894 		    sizeof (ptgt->tgt_port_wwn)) == 0) {
13895 			mutex_enter(&ptgt->tgt_mutex);
13896 			for (plun = ptgt->tgt_lun;
13897 			    plun != NULL;
13898 			    plun = plun->lun_next) {
13899 				if (plun->lun_num == lun) {
13900 					mutex_exit(&ptgt->tgt_mutex);
13901 					return (plun);
13902 				}
13903 			}
13904 			mutex_exit(&ptgt->tgt_mutex);
13905 			return (NULL);
13906 		}
13907 	}
13908 	return (NULL);
13909 }
13910 
13911 /*
13912  *     Function: fcp_prepare_pkt
13913  *
13914  *  Description: This function prepares the SCSI cmd pkt, passed by the caller,
13915  *		 for fcp_start(). It binds the data or partially maps it.
13916  *		 Builds the FCP header and starts the initialization of the
13917  *		 Fibre Channel header.
13918  *
13919  *     Argument: *pptr		FCP port.
13920  *		 *cmd		FCP packet.
13921  *		 *plun		LUN the command will be sent to.
13922  *
13923  *	Context: User, Kernel and Interrupt context.
13924  */
13925 static void
13926 fcp_prepare_pkt(struct fcp_port *pptr, struct fcp_pkt *cmd,
13927     struct fcp_lun *plun)
13928 {
13929 	fc_packet_t		*fpkt = cmd->cmd_fp_pkt;
13930 	struct fcp_tgt		*ptgt = plun->lun_tgt;
13931 	struct fcp_cmd		*fcmd = &cmd->cmd_fcp_cmd;
13932 
13933 	ASSERT(cmd->cmd_pkt->pkt_comp ||
13934 	    (cmd->cmd_pkt->pkt_flags & FLAG_NOINTR));
13935 
13936 	if (cmd->cmd_pkt->pkt_numcookies) {
13937 		if (cmd->cmd_pkt->pkt_dma_flags & DDI_DMA_READ) {
13938 			fcmd->fcp_cntl.cntl_read_data = 1;
13939 			fcmd->fcp_cntl.cntl_write_data = 0;
13940 			fpkt->pkt_tran_type = FC_PKT_FCP_READ;
13941 		} else {
13942 			fcmd->fcp_cntl.cntl_read_data = 0;
13943 			fcmd->fcp_cntl.cntl_write_data = 1;
13944 			fpkt->pkt_tran_type = FC_PKT_FCP_WRITE;
13945 		}
13946 
13947 		fpkt->pkt_data_cookie = cmd->cmd_pkt->pkt_cookies;
13948 
13949 		fpkt->pkt_data_cookie_cnt = cmd->cmd_pkt->pkt_numcookies;
13950 		ASSERT(fpkt->pkt_data_cookie_cnt <=
13951 		    pptr->port_data_dma_attr.dma_attr_sgllen);
13952 
13953 		cmd->cmd_dmacount = cmd->cmd_pkt->pkt_dma_len;
13954 
13955 		/* FCA needs pkt_datalen to be set */
13956 		fpkt->pkt_datalen = cmd->cmd_dmacount;
13957 		fcmd->fcp_data_len = cmd->cmd_dmacount;
13958 	} else {
13959 		fcmd->fcp_cntl.cntl_read_data = 0;
13960 		fcmd->fcp_cntl.cntl_write_data = 0;
13961 		fpkt->pkt_tran_type = FC_PKT_EXCHANGE;
13962 		fpkt->pkt_datalen = 0;
13963 		fcmd->fcp_data_len = 0;
13964 	}
13965 
13966 	/* set up the Tagged Queuing type */
13967 	if (cmd->cmd_pkt->pkt_flags & FLAG_HTAG) {
13968 		fcmd->fcp_cntl.cntl_qtype = FCP_QTYPE_HEAD_OF_Q;
13969 	} else if (cmd->cmd_pkt->pkt_flags & FLAG_OTAG) {
13970 		fcmd->fcp_cntl.cntl_qtype = FCP_QTYPE_ORDERED;
13971 	} else if (cmd->cmd_pkt->pkt_flags & FLAG_STAG) {
13972 		fcmd->fcp_cntl.cntl_qtype = FCP_QTYPE_SIMPLE;
13973 	} else {
13974 		fcmd->fcp_cntl.cntl_qtype = FCP_QTYPE_UNTAGGED;
13975 	}
13976 
13977 	fcmd->fcp_ent_addr = plun->lun_addr;
13978 
13979 	if (pptr->port_fcp_dma != FC_NO_DVMA_SPACE) {
13980 		FCP_CP_OUT((uint8_t *)fcmd, fpkt->pkt_cmd,
13981 		    fpkt->pkt_cmd_acc, sizeof (struct fcp_cmd));
13982 	} else {
13983 		ASSERT(fpkt->pkt_cmd_dma == NULL && fpkt->pkt_resp_dma == NULL);
13984 	}
13985 
13986 	cmd->cmd_pkt->pkt_reason = CMD_CMPLT;
13987 	cmd->cmd_pkt->pkt_state = 0;
13988 	cmd->cmd_pkt->pkt_statistics = 0;
13989 	cmd->cmd_pkt->pkt_resid = 0;
13990 
13991 	cmd->cmd_fp_pkt->pkt_data_dma = cmd->cmd_pkt->pkt_handle;
13992 
13993 	if (cmd->cmd_pkt->pkt_flags & FLAG_NOINTR) {
13994 		fpkt->pkt_tran_flags = (FC_TRAN_CLASS3 | FC_TRAN_NO_INTR);
13995 		fpkt->pkt_comp = NULL;
13996 	} else {
13997 		fpkt->pkt_tran_flags = (FC_TRAN_CLASS3 | FC_TRAN_INTR);
13998 		if (cmd->cmd_pkt->pkt_flags & FLAG_IMMEDIATE_CB) {
13999 			fpkt->pkt_tran_flags |= FC_TRAN_IMMEDIATE_CB;
14000 		}
14001 		fpkt->pkt_comp = fcp_cmd_callback;
14002 	}
14003 
14004 	mutex_enter(&pptr->port_mutex);
14005 	if (pptr->port_state & FCP_STATE_SUSPENDED) {
14006 		fpkt->pkt_tran_flags |= FC_TRAN_DUMPING;
14007 	}
14008 	mutex_exit(&pptr->port_mutex);
14009 
14010 	fpkt->pkt_cmd_fhdr.d_id = ptgt->tgt_d_id;
14011 	fpkt->pkt_cmd_fhdr.s_id = pptr->port_id;
14012 
14013 	/*
14014 	 * Save a few kernel cycles here
14015 	 */
14016 #ifndef	__lock_lint
14017 	fpkt->pkt_fca_device = ptgt->tgt_fca_dev;
14018 #endif /* __lock_lint */
14019 }
14020 
14021 static void
14022 fcp_post_callback(struct fcp_pkt *cmd)
14023 {
14024 	if (cmd->cmd_pkt->pkt_comp) {
14025 		(*cmd->cmd_pkt->pkt_comp) (cmd->cmd_pkt);
14026 	}
14027 }
14028 
14029 
14030 /*
14031  * called to do polled I/O by fcp_start()
14032  *
14033  * return a transport status value, i.e. TRAN_ACCECPT for success
14034  */
14035 static int
14036 fcp_dopoll(struct fcp_port *pptr, struct fcp_pkt *cmd)
14037 {
14038 	int	rval;
14039 
14040 #ifdef	DEBUG
14041 	mutex_enter(&pptr->port_pkt_mutex);
14042 	pptr->port_npkts++;
14043 	mutex_exit(&pptr->port_pkt_mutex);
14044 #endif /* DEBUG */
14045 
14046 	if (cmd->cmd_fp_pkt->pkt_timeout) {
14047 		cmd->cmd_fp_pkt->pkt_timeout = cmd->cmd_pkt->pkt_time;
14048 	} else {
14049 		cmd->cmd_fp_pkt->pkt_timeout = FCP_POLL_TIMEOUT;
14050 	}
14051 
14052 	ASSERT(cmd->cmd_fp_pkt->pkt_comp == NULL);
14053 
14054 	cmd->cmd_state = FCP_PKT_ISSUED;
14055 
14056 	rval = fc_ulp_transport(pptr->port_fp_handle, cmd->cmd_fp_pkt);
14057 
14058 #ifdef	DEBUG
14059 	mutex_enter(&pptr->port_pkt_mutex);
14060 	pptr->port_npkts--;
14061 	mutex_exit(&pptr->port_pkt_mutex);
14062 #endif /* DEBUG */
14063 
14064 	cmd->cmd_state = FCP_PKT_IDLE;
14065 
14066 	switch (rval) {
14067 	case FC_SUCCESS:
14068 		if (cmd->cmd_fp_pkt->pkt_state == FC_PKT_SUCCESS) {
14069 			fcp_complete_pkt(cmd->cmd_fp_pkt);
14070 			rval = TRAN_ACCEPT;
14071 		} else {
14072 			rval = TRAN_FATAL_ERROR;
14073 		}
14074 		break;
14075 
14076 	case FC_TRAN_BUSY:
14077 		rval = TRAN_BUSY;
14078 		cmd->cmd_pkt->pkt_resid = 0;
14079 		break;
14080 
14081 	case FC_BADPACKET:
14082 		rval = TRAN_BADPKT;
14083 		break;
14084 
14085 	default:
14086 		rval = TRAN_FATAL_ERROR;
14087 		break;
14088 	}
14089 
14090 	return (rval);
14091 }
14092 
14093 
14094 /*
14095  * called by some of the following transport-called routines to convert
14096  * a supplied dip ptr to a port struct ptr (i.e. to the soft state)
14097  */
14098 static struct fcp_port *
14099 fcp_dip2port(dev_info_t *dip)
14100 {
14101 	int	instance;
14102 
14103 	instance = ddi_get_instance(dip);
14104 	return (ddi_get_soft_state(fcp_softstate, instance));
14105 }
14106 
14107 
14108 /*
14109  * called internally to return a LUN given a dip
14110  */
14111 struct fcp_lun *
14112 fcp_get_lun_from_cip(struct fcp_port *pptr, child_info_t *cip)
14113 {
14114 	struct fcp_tgt *ptgt;
14115 	struct fcp_lun *plun;
14116 	int i;
14117 
14118 
14119 	ASSERT(mutex_owned(&pptr->port_mutex));
14120 
14121 	for (i = 0; i < FCP_NUM_HASH; i++) {
14122 		for (ptgt = pptr->port_tgt_hash_table[i];
14123 		    ptgt != NULL;
14124 		    ptgt = ptgt->tgt_next) {
14125 			mutex_enter(&ptgt->tgt_mutex);
14126 			for (plun = ptgt->tgt_lun; plun != NULL;
14127 			    plun = plun->lun_next) {
14128 				mutex_enter(&plun->lun_mutex);
14129 				if (plun->lun_cip == cip) {
14130 					mutex_exit(&plun->lun_mutex);
14131 					mutex_exit(&ptgt->tgt_mutex);
14132 					return (plun); /* match found */
14133 				}
14134 				mutex_exit(&plun->lun_mutex);
14135 			}
14136 			mutex_exit(&ptgt->tgt_mutex);
14137 		}
14138 	}
14139 	return (NULL);				/* no LUN found */
14140 }
14141 
14142 /*
14143  * pass an element to the hotplug list, kick the hotplug thread
14144  * and wait for the element to get processed by the hotplug thread.
14145  * on return the element is freed.
14146  *
14147  * return zero success and non-zero on failure
14148  *
14149  * acquires/releases the target mutex
14150  *
14151  */
14152 static int
14153 fcp_pass_to_hp_and_wait(struct fcp_port *pptr, struct fcp_lun *plun,
14154     child_info_t *cip, int what, int link_cnt, int tgt_cnt, int flags)
14155 {
14156 	struct fcp_hp_elem	*elem;
14157 	int			rval;
14158 
14159 	mutex_enter(&plun->lun_tgt->tgt_mutex);
14160 	if ((elem = fcp_pass_to_hp(pptr, plun, cip,
14161 	    what, link_cnt, tgt_cnt, flags, 1)) == NULL) {
14162 		mutex_exit(&plun->lun_tgt->tgt_mutex);
14163 		fcp_log(CE_CONT, pptr->port_dip,
14164 		    "Can not pass_to_hp: what: %d; D_ID=%x, LUN=%x\n",
14165 		    what, plun->lun_tgt->tgt_d_id, plun->lun_num);
14166 		return (NDI_FAILURE);
14167 	}
14168 	mutex_exit(&plun->lun_tgt->tgt_mutex);
14169 	mutex_enter(&elem->mutex);
14170 	if (elem->wait) {
14171 		while (elem->wait) {
14172 			cv_wait(&elem->cv, &elem->mutex);
14173 		}
14174 	}
14175 	rval = (elem->result);
14176 	mutex_exit(&elem->mutex);
14177 	mutex_destroy(&elem->mutex);
14178 	cv_destroy(&elem->cv);
14179 	kmem_free(elem, sizeof (struct fcp_hp_elem));
14180 	return (rval);
14181 }
14182 
14183 /*
14184  * pass an element to the hotplug list, and then
14185  * kick the hotplug thread
14186  *
14187  * return Boolean success, i.e. non-zero if all goes well, else zero on error
14188  *
14189  * acquires/releases the hotplug mutex
14190  *
14191  * called with the target mutex owned
14192  *
14193  * memory acquired in NOSLEEP mode
14194  * NOTE: if wait is set to 1 then the caller is responsible for waiting on
14195  *	 for the hp daemon to process the request and is responsible for
14196  *	 freeing the element
14197  */
14198 static struct fcp_hp_elem *
14199 fcp_pass_to_hp(struct fcp_port *pptr, struct fcp_lun *plun,
14200     child_info_t *cip, int what, int link_cnt, int tgt_cnt, int flags, int wait)
14201 {
14202 	struct fcp_hp_elem	*elem;
14203 	dev_info_t *pdip;
14204 
14205 	ASSERT(pptr != NULL);
14206 	ASSERT(plun != NULL);
14207 	ASSERT(plun->lun_tgt != NULL);
14208 	ASSERT(mutex_owned(&plun->lun_tgt->tgt_mutex));
14209 
14210 	/* create space for a hotplug element */
14211 	if ((elem = kmem_zalloc(sizeof (struct fcp_hp_elem), KM_NOSLEEP))
14212 	    == NULL) {
14213 		fcp_log(CE_WARN, NULL,
14214 		    "!can't allocate memory for hotplug element");
14215 		return (NULL);
14216 	}
14217 
14218 	/* fill in hotplug element */
14219 	elem->port = pptr;
14220 	elem->lun = plun;
14221 	elem->cip = cip;
14222 	elem->old_lun_mpxio = plun->lun_mpxio;
14223 	elem->what = what;
14224 	elem->flags = flags;
14225 	elem->link_cnt = link_cnt;
14226 	elem->tgt_cnt = tgt_cnt;
14227 	elem->wait = wait;
14228 	mutex_init(&elem->mutex, NULL, MUTEX_DRIVER, NULL);
14229 	cv_init(&elem->cv, NULL, CV_DRIVER, NULL);
14230 
14231 	/* schedule the hotplug task */
14232 	pdip = pptr->port_dip;
14233 	mutex_enter(&plun->lun_mutex);
14234 	if (elem->what == FCP_ONLINE || elem->what == FCP_OFFLINE) {
14235 		plun->lun_event_count++;
14236 		elem->event_cnt = plun->lun_event_count;
14237 	}
14238 	mutex_exit(&plun->lun_mutex);
14239 	if (taskq_dispatch(DEVI(pdip)->devi_taskq, fcp_hp_task,
14240 	    (void *)elem, KM_NOSLEEP) == NULL) {
14241 		mutex_enter(&plun->lun_mutex);
14242 		if (elem->what == FCP_ONLINE || elem->what == FCP_OFFLINE) {
14243 			plun->lun_event_count--;
14244 		}
14245 		mutex_exit(&plun->lun_mutex);
14246 		kmem_free(elem, sizeof (*elem));
14247 		return (0);
14248 	}
14249 
14250 	return (elem);
14251 }
14252 
14253 
14254 static void
14255 fcp_retransport_cmd(struct fcp_port *pptr, struct fcp_pkt *cmd)
14256 {
14257 	int			rval;
14258 	struct scsi_address	*ap;
14259 	struct fcp_lun	*plun;
14260 	struct fcp_tgt	*ptgt;
14261 	fc_packet_t	*fpkt;
14262 
14263 	ap = &cmd->cmd_pkt->pkt_address;
14264 	plun = ADDR2LUN(ap);
14265 	ptgt = plun->lun_tgt;
14266 
14267 	ASSERT(cmd->cmd_flags & CFLAG_IN_QUEUE);
14268 
14269 	cmd->cmd_state = FCP_PKT_IDLE;
14270 
14271 	mutex_enter(&pptr->port_mutex);
14272 	mutex_enter(&ptgt->tgt_mutex);
14273 	if (((plun->lun_state & (FCP_LUN_BUSY | FCP_LUN_OFFLINE)) == 0) &&
14274 	    (!(pptr->port_state & FCP_STATE_ONLINING))) {
14275 		fc_ulp_rscn_info_t *rscnp;
14276 
14277 		cmd->cmd_state = FCP_PKT_ISSUED;
14278 
14279 		/*
14280 		 * It is possible for pkt_pd to be NULL if tgt_pd_handle was
14281 		 * originally NULL, hence we try to set it to the pd pointed
14282 		 * to by the SCSI device we're trying to get to.
14283 		 */
14284 
14285 		fpkt = cmd->cmd_fp_pkt;
14286 		if ((fpkt->pkt_pd == NULL) && (ptgt->tgt_pd_handle != NULL)) {
14287 			fpkt->pkt_pd = ptgt->tgt_pd_handle;
14288 			/*
14289 			 * We need to notify the transport that we now have a
14290 			 * reference to the remote port handle.
14291 			 */
14292 			fc_ulp_hold_remote_port(ptgt->tgt_pd_handle);
14293 		}
14294 
14295 		mutex_exit(&ptgt->tgt_mutex);
14296 		mutex_exit(&pptr->port_mutex);
14297 
14298 		ASSERT((cmd->cmd_pkt->pkt_flags & FLAG_NOINTR) == 0);
14299 
14300 		/* prepare the packet */
14301 
14302 		fcp_prepare_pkt(pptr, cmd, plun);
14303 
14304 		rscnp = (fc_ulp_rscn_info_t *)cmd->cmd_fp_pkt->
14305 		    pkt_ulp_rscn_infop;
14306 
14307 		cmd->cmd_timeout = cmd->cmd_pkt->pkt_time ?
14308 		    fcp_watchdog_time + cmd->cmd_pkt->pkt_time : 0;
14309 
14310 		if (rscnp != NULL) {
14311 			rscnp->ulp_rscn_count =
14312 			    fc_ulp_get_rscn_count(pptr->
14313 			    port_fp_handle);
14314 		}
14315 
14316 		rval = fcp_transport(pptr->port_fp_handle,
14317 		    cmd->cmd_fp_pkt, 0);
14318 
14319 		if (rval == FC_SUCCESS) {
14320 			return;
14321 		}
14322 		cmd->cmd_state &= ~FCP_PKT_ISSUED;
14323 	} else {
14324 		mutex_exit(&ptgt->tgt_mutex);
14325 		mutex_exit(&pptr->port_mutex);
14326 	}
14327 
14328 	fcp_queue_pkt(pptr, cmd);
14329 }
14330 
14331 
14332 static void
14333 fcp_fail_cmd(struct fcp_pkt *cmd, uchar_t reason, uint_t statistics)
14334 {
14335 	ASSERT(cmd->cmd_flags & CFLAG_IN_QUEUE);
14336 
14337 	cmd->cmd_flags &= ~CFLAG_IN_QUEUE;
14338 	cmd->cmd_state = FCP_PKT_IDLE;
14339 
14340 	cmd->cmd_pkt->pkt_reason = reason;
14341 	cmd->cmd_pkt->pkt_state = 0;
14342 	cmd->cmd_pkt->pkt_statistics = statistics;
14343 
14344 	fcp_post_callback(cmd);
14345 }
14346 
14347 /*
14348  *     Function: fcp_queue_pkt
14349  *
14350  *  Description: This function queues the packet passed by the caller into
14351  *		 the list of packets of the FCP port.
14352  *
14353  *     Argument: *pptr		FCP port.
14354  *		 *cmd		FCP packet to queue.
14355  *
14356  * Return Value: None
14357  *
14358  *	Context: User, Kernel and Interrupt context.
14359  */
14360 static void
14361 fcp_queue_pkt(struct fcp_port *pptr, struct fcp_pkt *cmd)
14362 {
14363 	ASSERT((cmd->cmd_pkt->pkt_flags & FLAG_NOQUEUE) == NULL);
14364 
14365 	mutex_enter(&pptr->port_pkt_mutex);
14366 	cmd->cmd_flags |= CFLAG_IN_QUEUE;
14367 	ASSERT(cmd->cmd_state != FCP_PKT_ISSUED);
14368 	cmd->cmd_timeout = fcp_watchdog_time + FCP_QUEUE_DELAY;
14369 
14370 	/*
14371 	 * zero pkt_time means hang around for ever
14372 	 */
14373 	if (cmd->cmd_pkt->pkt_time) {
14374 		if (cmd->cmd_fp_pkt->pkt_timeout > FCP_QUEUE_DELAY) {
14375 			cmd->cmd_fp_pkt->pkt_timeout -= FCP_QUEUE_DELAY;
14376 		} else {
14377 			/*
14378 			 * Indicate the watch thread to fail the
14379 			 * command by setting it to highest value
14380 			 */
14381 			cmd->cmd_timeout = fcp_watchdog_time;
14382 			cmd->cmd_fp_pkt->pkt_timeout = FCP_INVALID_TIMEOUT;
14383 		}
14384 	}
14385 
14386 	if (pptr->port_pkt_head) {
14387 		ASSERT(pptr->port_pkt_tail != NULL);
14388 
14389 		pptr->port_pkt_tail->cmd_next = cmd;
14390 		pptr->port_pkt_tail = cmd;
14391 	} else {
14392 		ASSERT(pptr->port_pkt_tail == NULL);
14393 
14394 		pptr->port_pkt_head = pptr->port_pkt_tail = cmd;
14395 	}
14396 	cmd->cmd_next = NULL;
14397 	mutex_exit(&pptr->port_pkt_mutex);
14398 }
14399 
14400 /*
14401  *     Function: fcp_update_targets
14402  *
14403  *  Description: This function applies the specified change of state to all
14404  *		 the targets listed.  The operation applied is 'set'.
14405  *
14406  *     Argument: *pptr		FCP port.
14407  *		 *dev_list	Array of fc_portmap_t structures.
14408  *		 count		Length of dev_list.
14409  *		 state		State bits to update.
14410  *		 cause		Reason for the update.
14411  *
14412  * Return Value: None
14413  *
14414  *	Context: User, Kernel and Interrupt context.
14415  *		 The mutex pptr->port_mutex must be held.
14416  */
14417 static void
14418 fcp_update_targets(struct fcp_port *pptr, fc_portmap_t *dev_list,
14419     uint32_t count, uint32_t state, int cause)
14420 {
14421 	fc_portmap_t		*map_entry;
14422 	struct fcp_tgt	*ptgt;
14423 
14424 	ASSERT(MUTEX_HELD(&pptr->port_mutex));
14425 
14426 	while (count--) {
14427 		map_entry = &(dev_list[count]);
14428 		ptgt = fcp_lookup_target(pptr,
14429 		    (uchar_t *)&(map_entry->map_pwwn));
14430 		if (ptgt == NULL) {
14431 			continue;
14432 		}
14433 
14434 		mutex_enter(&ptgt->tgt_mutex);
14435 		ptgt->tgt_trace = 0;
14436 		ptgt->tgt_change_cnt++;
14437 		ptgt->tgt_statec_cause = cause;
14438 		ptgt->tgt_tmp_cnt = 1;
14439 		fcp_update_tgt_state(ptgt, FCP_SET, state);
14440 		mutex_exit(&ptgt->tgt_mutex);
14441 	}
14442 }
14443 
14444 static int
14445 fcp_call_finish_init(struct fcp_port *pptr, struct fcp_tgt *ptgt,
14446     int lcount, int tcount, int cause)
14447 {
14448 	int rval;
14449 
14450 	mutex_enter(&pptr->port_mutex);
14451 	rval = fcp_call_finish_init_held(pptr, ptgt, lcount, tcount, cause);
14452 	mutex_exit(&pptr->port_mutex);
14453 
14454 	return (rval);
14455 }
14456 
14457 
14458 static int
14459 fcp_call_finish_init_held(struct fcp_port *pptr, struct fcp_tgt *ptgt,
14460     int lcount, int tcount, int cause)
14461 {
14462 	int	finish_init = 0;
14463 	int	finish_tgt = 0;
14464 	int	do_finish_init = 0;
14465 	int	rval = FCP_NO_CHANGE;
14466 
14467 	if (cause == FCP_CAUSE_LINK_CHANGE ||
14468 	    cause == FCP_CAUSE_LINK_DOWN) {
14469 		do_finish_init = 1;
14470 	}
14471 
14472 	if (ptgt != NULL) {
14473 		FCP_TRACE(fcp_logq, pptr->port_instbuf, fcp_trace,
14474 		    FCP_BUF_LEVEL_2, 0,
14475 		    "link_cnt: %d,%d; tgt_cnt: %d,%d; tmp_cnt: %d,%d;"
14476 		    " cause = %d, d_id = 0x%x, tgt_done = %d",
14477 		    pptr->port_link_cnt, lcount, ptgt->tgt_change_cnt, tcount,
14478 		    pptr->port_tmp_cnt, ptgt->tgt_tmp_cnt, cause,
14479 		    ptgt->tgt_d_id, ptgt->tgt_done);
14480 
14481 		mutex_enter(&ptgt->tgt_mutex);
14482 
14483 		if (tcount && (ptgt->tgt_change_cnt != tcount)) {
14484 			rval = FCP_DEV_CHANGE;
14485 			if (do_finish_init && ptgt->tgt_done == 0) {
14486 				ptgt->tgt_done++;
14487 				finish_init = 1;
14488 			}
14489 		} else {
14490 			if (--ptgt->tgt_tmp_cnt <= 0) {
14491 				ptgt->tgt_tmp_cnt = 0;
14492 				finish_tgt = 1;
14493 
14494 				if (do_finish_init) {
14495 					finish_init = 1;
14496 				}
14497 			}
14498 		}
14499 		mutex_exit(&ptgt->tgt_mutex);
14500 	} else {
14501 		FCP_TRACE(fcp_logq, pptr->port_instbuf, fcp_trace,
14502 		    FCP_BUF_LEVEL_2, 0,
14503 		    "Call Finish Init for NO target");
14504 
14505 		if (do_finish_init) {
14506 			finish_init = 1;
14507 		}
14508 	}
14509 
14510 	if (finish_tgt) {
14511 		ASSERT(ptgt != NULL);
14512 
14513 		mutex_enter(&ptgt->tgt_mutex);
14514 #ifdef	DEBUG
14515 		bzero(ptgt->tgt_tmp_cnt_stack,
14516 		    sizeof (ptgt->tgt_tmp_cnt_stack));
14517 
14518 		ptgt->tgt_tmp_cnt_depth = getpcstack(ptgt->tgt_tmp_cnt_stack,
14519 		    FCP_STACK_DEPTH);
14520 #endif /* DEBUG */
14521 		mutex_exit(&ptgt->tgt_mutex);
14522 
14523 		(void) fcp_finish_tgt(pptr, ptgt, lcount, tcount, cause);
14524 	}
14525 
14526 	if (finish_init && lcount == pptr->port_link_cnt) {
14527 		ASSERT(pptr->port_tmp_cnt > 0);
14528 		if (--pptr->port_tmp_cnt == 0) {
14529 			fcp_finish_init(pptr);
14530 		}
14531 	} else if (lcount != pptr->port_link_cnt) {
14532 		FCP_TRACE(fcp_logq, pptr->port_instbuf,
14533 		    fcp_trace, FCP_BUF_LEVEL_2, 0,
14534 		    "fcp_call_finish_init_held,1: state change occured"
14535 		    " for D_ID=0x%x", (ptgt) ? ptgt->tgt_d_id : 0);
14536 	}
14537 
14538 	return (rval);
14539 }
14540 
14541 
14542 static void
14543 fcp_reconfigure_luns(void * tgt_handle)
14544 {
14545 	uint32_t		dev_cnt;
14546 	fc_portmap_t		*devlist;
14547 	struct fcp_tgt	*ptgt = (struct fcp_tgt *)tgt_handle;
14548 	struct fcp_port		*pptr = ptgt->tgt_port;
14549 
14550 	/*
14551 	 * If the timer that fires this off got canceled too late, the
14552 	 * target could have been destroyed.
14553 	 */
14554 
14555 	if (ptgt->tgt_tid == NULL) {
14556 		return;
14557 	}
14558 
14559 	devlist = kmem_zalloc(sizeof (*devlist), KM_NOSLEEP);
14560 	if (devlist == NULL) {
14561 		fcp_log(CE_WARN, pptr->port_dip,
14562 		    "!fcp%d: failed to allocate for portmap",
14563 		    pptr->port_instance);
14564 		return;
14565 	}
14566 
14567 	dev_cnt = 1;
14568 	devlist->map_pd = ptgt->tgt_pd_handle;
14569 	devlist->map_hard_addr.hard_addr = ptgt->tgt_hard_addr;
14570 	devlist->map_did.port_id = ptgt->tgt_d_id;
14571 
14572 	bcopy(&ptgt->tgt_node_wwn.raw_wwn[0], &devlist->map_nwwn, FC_WWN_SIZE);
14573 	bcopy(&ptgt->tgt_port_wwn.raw_wwn[0], &devlist->map_pwwn, FC_WWN_SIZE);
14574 
14575 	devlist->map_state = PORT_DEVICE_LOGGED_IN;
14576 	devlist->map_type = PORT_DEVICE_NEW;
14577 	devlist->map_flags = 0;
14578 
14579 	fcp_statec_callback(NULL, pptr->port_fp_handle, FC_STATE_DEVICE_CHANGE,
14580 	    pptr->port_topology, devlist, dev_cnt, pptr->port_id);
14581 
14582 	/*
14583 	 * Clear the tgt_tid after no more references to
14584 	 * the fcp_tgt
14585 	 */
14586 	mutex_enter(&ptgt->tgt_mutex);
14587 	ptgt->tgt_tid = NULL;
14588 	mutex_exit(&ptgt->tgt_mutex);
14589 
14590 	kmem_free(devlist, sizeof (*devlist));
14591 }
14592 
14593 
14594 static void
14595 fcp_free_targets(struct fcp_port *pptr)
14596 {
14597 	int			i;
14598 	struct fcp_tgt	*ptgt;
14599 
14600 	mutex_enter(&pptr->port_mutex);
14601 	for (i = 0; i < FCP_NUM_HASH; i++) {
14602 		ptgt = pptr->port_tgt_hash_table[i];
14603 		while (ptgt != NULL) {
14604 			struct fcp_tgt *next_tgt = ptgt->tgt_next;
14605 
14606 			fcp_free_target(ptgt);
14607 			ptgt = next_tgt;
14608 		}
14609 	}
14610 	mutex_exit(&pptr->port_mutex);
14611 }
14612 
14613 
14614 static void
14615 fcp_free_target(struct fcp_tgt *ptgt)
14616 {
14617 	struct fcp_lun	*plun;
14618 	timeout_id_t		tid;
14619 
14620 	mutex_enter(&ptgt->tgt_mutex);
14621 	tid = ptgt->tgt_tid;
14622 
14623 	/*
14624 	 * Cancel any pending timeouts for this target.
14625 	 */
14626 
14627 	if (tid != NULL) {
14628 		/*
14629 		 * Set tgt_tid to NULL first to avoid a race in the callback.
14630 		 * If tgt_tid is NULL, the callback will simply return.
14631 		 */
14632 		ptgt->tgt_tid = NULL;
14633 		mutex_exit(&ptgt->tgt_mutex);
14634 		(void) untimeout(tid);
14635 		mutex_enter(&ptgt->tgt_mutex);
14636 	}
14637 
14638 	plun = ptgt->tgt_lun;
14639 	while (plun != NULL) {
14640 		struct fcp_lun *next_lun = plun->lun_next;
14641 
14642 		fcp_dealloc_lun(plun);
14643 		plun = next_lun;
14644 	}
14645 
14646 	mutex_exit(&ptgt->tgt_mutex);
14647 	fcp_dealloc_tgt(ptgt);
14648 }
14649 
14650 /*
14651  *     Function: fcp_is_retryable
14652  *
14653  *  Description: Indicates if the internal packet is retryable.
14654  *
14655  *     Argument: *icmd		FCP internal packet.
14656  *
14657  * Return Value: 0	Not retryable
14658  *		 1	Retryable
14659  *
14660  *	Context: User, Kernel and Interrupt context
14661  */
14662 static int
14663 fcp_is_retryable(struct fcp_ipkt *icmd)
14664 {
14665 	if (icmd->ipkt_port->port_state & (FCP_STATE_SUSPENDED |
14666 	    FCP_STATE_DETACHING | FCP_STATE_POWER_DOWN)) {
14667 		return (0);
14668 	}
14669 
14670 	return (((fcp_watchdog_time + icmd->ipkt_fpkt->pkt_timeout) <
14671 	    icmd->ipkt_port->port_deadline) ? 1 : 0);
14672 }
14673 
14674 /*
14675  *     Function: fcp_create_on_demand
14676  *
14677  *     Argument: *pptr		FCP port.
14678  *		 *pwwn		Port WWN.
14679  *
14680  * Return Value: 0	Success
14681  *		 EIO
14682  *		 ENOMEM
14683  *		 EBUSY
14684  *		 EINVAL
14685  *
14686  *	Context: User and Kernel context
14687  */
14688 static int
14689 fcp_create_on_demand(struct fcp_port *pptr, uchar_t *pwwn)
14690 {
14691 	int			wait_ms;
14692 	int			tcount;
14693 	int			lcount;
14694 	int			ret;
14695 	int			error;
14696 	int			rval = EIO;
14697 	int			ntries;
14698 	fc_portmap_t		*devlist;
14699 	opaque_t		pd;
14700 	struct fcp_lun		*plun;
14701 	struct fcp_tgt		*ptgt;
14702 	int			old_manual = 0;
14703 
14704 	/* Allocates the fc_portmap_t structure. */
14705 	devlist = kmem_zalloc(sizeof (*devlist), KM_SLEEP);
14706 
14707 	/*
14708 	 * If FC_INVALID_RSCN_COUNT is non-zero, we will have to init as shown
14709 	 * in the commented statement below:
14710 	 *
14711 	 * devlist->map_rscn_info.ulp_rscn_count = FC_INVALID_RSCN_COUNT;
14712 	 *
14713 	 * Below, the deadline for the discovery process is set.
14714 	 */
14715 	mutex_enter(&pptr->port_mutex);
14716 	pptr->port_deadline = fcp_watchdog_time + FCP_ICMD_DEADLINE;
14717 	mutex_exit(&pptr->port_mutex);
14718 
14719 	/*
14720 	 * We try to find the remote port based on the WWN provided by the
14721 	 * caller.  We actually ask fp/fctl if it has it.
14722 	 */
14723 	pd = fc_ulp_get_remote_port(pptr->port_fp_handle,
14724 	    (la_wwn_t *)pwwn, &error, 1);
14725 
14726 	if (pd == NULL) {
14727 		kmem_free(devlist, sizeof (*devlist));
14728 		return (rval);
14729 	}
14730 
14731 	/*
14732 	 * The remote port was found.  We ask fp/fctl to update our
14733 	 * fc_portmap_t structure.
14734 	 */
14735 	ret = fc_ulp_pwwn_to_portmap(pptr->port_fp_handle,
14736 	    (la_wwn_t *)pwwn, devlist);
14737 	if (ret != FC_SUCCESS) {
14738 		kmem_free(devlist, sizeof (*devlist));
14739 		return (rval);
14740 	}
14741 
14742 	/*
14743 	 * The map flag field is set to indicates that the creation is being
14744 	 * done at the user request (Ioclt probably luxadm or cfgadm).
14745 	 */
14746 	devlist->map_type = PORT_DEVICE_USER_CREATE;
14747 
14748 	mutex_enter(&pptr->port_mutex);
14749 
14750 	/*
14751 	 * We check to see if fcp already has a target that describes the
14752 	 * device being created.  If not it is created.
14753 	 */
14754 	ptgt = fcp_lookup_target(pptr, pwwn);
14755 	if (ptgt == NULL) {
14756 		lcount = pptr->port_link_cnt;
14757 		mutex_exit(&pptr->port_mutex);
14758 
14759 		ptgt = fcp_alloc_tgt(pptr, devlist, lcount);
14760 		if (ptgt == NULL) {
14761 			fcp_log(CE_WARN, pptr->port_dip,
14762 			    "!FC target allocation failed");
14763 			return (ENOMEM);
14764 		}
14765 
14766 		mutex_enter(&pptr->port_mutex);
14767 	}
14768 
14769 	mutex_enter(&ptgt->tgt_mutex);
14770 	ptgt->tgt_statec_cause = FCP_CAUSE_USER_CREATE;
14771 	ptgt->tgt_tmp_cnt = 1;
14772 	ptgt->tgt_device_created = 0;
14773 	/*
14774 	 * If fabric and auto config is set but the target was
14775 	 * manually unconfigured then reset to the manual_config_only to
14776 	 * 0 so the device will get configured.
14777 	 */
14778 	if (FC_TOP_EXTERNAL(pptr->port_topology) &&
14779 	    fcp_enable_auto_configuration &&
14780 	    ptgt->tgt_manual_config_only == 1) {
14781 		old_manual = 1;
14782 		ptgt->tgt_manual_config_only = 0;
14783 	}
14784 	mutex_exit(&ptgt->tgt_mutex);
14785 
14786 	fcp_update_targets(pptr, devlist, 1,
14787 	    FCP_LUN_BUSY | FCP_LUN_MARK, FCP_CAUSE_USER_CREATE);
14788 
14789 	lcount = pptr->port_link_cnt;
14790 	tcount = ptgt->tgt_change_cnt;
14791 
14792 	if (fcp_handle_mapflags(pptr, ptgt, devlist, lcount,
14793 	    tcount, FCP_CAUSE_USER_CREATE) == TRUE) {
14794 		if (FC_TOP_EXTERNAL(pptr->port_topology) &&
14795 		    fcp_enable_auto_configuration && old_manual) {
14796 			mutex_enter(&ptgt->tgt_mutex);
14797 			ptgt->tgt_manual_config_only = 1;
14798 			mutex_exit(&ptgt->tgt_mutex);
14799 		}
14800 
14801 		if (pptr->port_link_cnt != lcount ||
14802 		    ptgt->tgt_change_cnt != tcount) {
14803 			rval = EBUSY;
14804 		}
14805 		mutex_exit(&pptr->port_mutex);
14806 
14807 		FCP_TRACE(fcp_logq, pptr->port_instbuf, fcp_trace,
14808 		    FCP_BUF_LEVEL_3, 0,
14809 		    "fcp_create_on_demand: mapflags ptgt=%x, "
14810 		    "lcount=%x::port_link_cnt=%x, "
14811 		    "tcount=%x: tgt_change_cnt=%x, rval=%x",
14812 		    ptgt, lcount, pptr->port_link_cnt,
14813 		    tcount, ptgt->tgt_change_cnt, rval);
14814 		return (rval);
14815 	}
14816 
14817 	/*
14818 	 * Due to lack of synchronization mechanisms, we perform
14819 	 * periodic monitoring of our request; Because requests
14820 	 * get dropped when another one supercedes (either because
14821 	 * of a link change or a target change), it is difficult to
14822 	 * provide a clean synchronization mechanism (such as a
14823 	 * semaphore or a conditional variable) without exhaustively
14824 	 * rewriting the mainline discovery code of this driver.
14825 	 */
14826 	wait_ms = 500;
14827 
14828 	ntries = fcp_max_target_retries;
14829 
14830 	FCP_TRACE(fcp_logq, pptr->port_instbuf, fcp_trace,
14831 	    FCP_BUF_LEVEL_3, 0,
14832 	    "fcp_create_on_demand(1): ntries=%x, ptgt=%x, "
14833 	    "lcount=%x::port_link_cnt=%x, "
14834 	    "tcount=%x::tgt_change_cnt=%x, rval=%x, tgt_device_created=%x "
14835 	    "tgt_tmp_cnt =%x",
14836 	    ntries, ptgt, lcount, pptr->port_link_cnt,
14837 	    tcount, ptgt->tgt_change_cnt, rval, ptgt->tgt_device_created,
14838 	    ptgt->tgt_tmp_cnt);
14839 
14840 	mutex_enter(&ptgt->tgt_mutex);
14841 	while (ntries-- != 0 && pptr->port_link_cnt == lcount &&
14842 	    ptgt->tgt_change_cnt == tcount && ptgt->tgt_device_created == 0) {
14843 		mutex_exit(&ptgt->tgt_mutex);
14844 		mutex_exit(&pptr->port_mutex);
14845 
14846 		delay(drv_usectohz(wait_ms * 1000));
14847 
14848 		mutex_enter(&pptr->port_mutex);
14849 		mutex_enter(&ptgt->tgt_mutex);
14850 	}
14851 
14852 
14853 	if (pptr->port_link_cnt != lcount || ptgt->tgt_change_cnt != tcount) {
14854 		rval = EBUSY;
14855 	} else {
14856 		if (ptgt->tgt_tmp_cnt == 0 && ptgt->tgt_node_state ==
14857 		    FCP_TGT_NODE_PRESENT) {
14858 			rval = 0;
14859 		}
14860 	}
14861 
14862 	FCP_TRACE(fcp_logq, pptr->port_instbuf, fcp_trace,
14863 	    FCP_BUF_LEVEL_3, 0,
14864 	    "fcp_create_on_demand(2): ntries=%x, ptgt=%x, "
14865 	    "lcount=%x::port_link_cnt=%x, "
14866 	    "tcount=%x::tgt_change_cnt=%x, rval=%x, tgt_device_created=%x "
14867 	    "tgt_tmp_cnt =%x",
14868 	    ntries, ptgt, lcount, pptr->port_link_cnt,
14869 	    tcount, ptgt->tgt_change_cnt, rval, ptgt->tgt_device_created,
14870 	    ptgt->tgt_tmp_cnt);
14871 
14872 	if (rval) {
14873 		if (FC_TOP_EXTERNAL(pptr->port_topology) &&
14874 		    fcp_enable_auto_configuration && old_manual) {
14875 			ptgt->tgt_manual_config_only = 1;
14876 		}
14877 		mutex_exit(&ptgt->tgt_mutex);
14878 		mutex_exit(&pptr->port_mutex);
14879 		kmem_free(devlist, sizeof (*devlist));
14880 
14881 		FCP_TRACE(fcp_logq, pptr->port_instbuf, fcp_trace,
14882 		    FCP_BUF_LEVEL_3, 0,
14883 		    "fcp_create_on_demand(3): ntries=%x, ptgt=%x, "
14884 		    "lcount=%x::port_link_cnt=%x, "
14885 		    "tcount=%x::tgt_change_cnt=%x, rval=%x, "
14886 		    "tgt_device_created=%x, tgt D_ID=%x",
14887 		    ntries, ptgt, lcount, pptr->port_link_cnt,
14888 		    tcount, ptgt->tgt_change_cnt, rval,
14889 		    ptgt->tgt_device_created, ptgt->tgt_d_id);
14890 		return (rval);
14891 	}
14892 
14893 	if ((plun = ptgt->tgt_lun) != NULL) {
14894 		tcount = plun->lun_tgt->tgt_change_cnt;
14895 	} else {
14896 		rval = EINVAL;
14897 	}
14898 	lcount = pptr->port_link_cnt;
14899 
14900 	/*
14901 	 * Configuring the target with no LUNs will fail. We
14902 	 * should reset the node state so that it is not
14903 	 * automatically configured when the LUNs are added
14904 	 * to this target.
14905 	 */
14906 	if (ptgt->tgt_lun_cnt == 0) {
14907 		ptgt->tgt_node_state = FCP_TGT_NODE_NONE;
14908 	}
14909 	mutex_exit(&ptgt->tgt_mutex);
14910 	mutex_exit(&pptr->port_mutex);
14911 
14912 	while (plun) {
14913 		child_info_t	*cip;
14914 
14915 		mutex_enter(&plun->lun_mutex);
14916 		cip = plun->lun_cip;
14917 		mutex_exit(&plun->lun_mutex);
14918 
14919 		mutex_enter(&ptgt->tgt_mutex);
14920 		if (!(plun->lun_state & FCP_LUN_OFFLINE)) {
14921 			mutex_exit(&ptgt->tgt_mutex);
14922 
14923 			rval = fcp_pass_to_hp_and_wait(pptr, plun, cip,
14924 			    FCP_ONLINE, lcount, tcount,
14925 			    NDI_ONLINE_ATTACH);
14926 			if (rval != NDI_SUCCESS) {
14927 				FCP_TRACE(fcp_logq,
14928 				    pptr->port_instbuf, fcp_trace,
14929 				    FCP_BUF_LEVEL_3, 0,
14930 				    "fcp_create_on_demand: "
14931 				    "pass_to_hp_and_wait failed "
14932 				    "rval=%x", rval);
14933 				rval = EIO;
14934 			} else {
14935 				mutex_enter(&LUN_TGT->tgt_mutex);
14936 				plun->lun_state &= ~(FCP_LUN_OFFLINE |
14937 				    FCP_LUN_BUSY);
14938 				mutex_exit(&LUN_TGT->tgt_mutex);
14939 			}
14940 			mutex_enter(&ptgt->tgt_mutex);
14941 		}
14942 
14943 		plun = plun->lun_next;
14944 		mutex_exit(&ptgt->tgt_mutex);
14945 	}
14946 
14947 	kmem_free(devlist, sizeof (*devlist));
14948 
14949 	if (FC_TOP_EXTERNAL(pptr->port_topology) &&
14950 	    fcp_enable_auto_configuration && old_manual) {
14951 		mutex_enter(&ptgt->tgt_mutex);
14952 		/* if successful then set manual to 0 */
14953 		if (rval == 0) {
14954 			ptgt->tgt_manual_config_only = 0;
14955 		} else {
14956 			/* reset to 1 so the user has to do the config */
14957 			ptgt->tgt_manual_config_only = 1;
14958 		}
14959 		mutex_exit(&ptgt->tgt_mutex);
14960 	}
14961 
14962 	return (rval);
14963 }
14964 
14965 
14966 static void
14967 fcp_ascii_to_wwn(caddr_t string, uchar_t bytes[], unsigned int byte_len)
14968 {
14969 	int		count;
14970 	uchar_t		byte;
14971 
14972 	count = 0;
14973 	while (*string) {
14974 		byte = FCP_ATOB(*string); string++;
14975 		byte = byte << 4 | FCP_ATOB(*string); string++;
14976 		bytes[count++] = byte;
14977 
14978 		if (count >= byte_len) {
14979 			break;
14980 		}
14981 	}
14982 }
14983 
14984 static void
14985 fcp_wwn_to_ascii(uchar_t wwn[], char *string)
14986 {
14987 	int		i;
14988 
14989 	for (i = 0; i < FC_WWN_SIZE; i++) {
14990 		(void) sprintf(string + (i * 2),
14991 		    "%02x", wwn[i]);
14992 	}
14993 
14994 }
14995 
14996 static void
14997 fcp_print_error(fc_packet_t *fpkt)
14998 {
14999 	struct fcp_ipkt	*icmd = (struct fcp_ipkt *)
15000 	    fpkt->pkt_ulp_private;
15001 	struct fcp_port	*pptr;
15002 	struct fcp_tgt	*ptgt;
15003 	struct fcp_lun	*plun;
15004 	caddr_t			buf;
15005 	int			scsi_cmd = 0;
15006 
15007 	ptgt = icmd->ipkt_tgt;
15008 	plun = icmd->ipkt_lun;
15009 	pptr = ptgt->tgt_port;
15010 
15011 	buf = kmem_zalloc(256, KM_NOSLEEP);
15012 	if (buf == NULL) {
15013 		return;
15014 	}
15015 
15016 	switch (icmd->ipkt_opcode) {
15017 	case SCMD_REPORT_LUN:
15018 		(void) sprintf(buf, "!REPORT LUN to D_ID=0x%%x"
15019 		    " lun=0x%%x failed");
15020 		scsi_cmd++;
15021 		break;
15022 
15023 	case SCMD_INQUIRY_PAGE83:
15024 		(void) sprintf(buf, "!INQUIRY-83 to D_ID=0x%%x"
15025 		    " lun=0x%%x failed");
15026 		scsi_cmd++;
15027 		break;
15028 
15029 	case SCMD_INQUIRY:
15030 		(void) sprintf(buf, "!INQUIRY to D_ID=0x%%x"
15031 		    " lun=0x%%x failed");
15032 		scsi_cmd++;
15033 		break;
15034 
15035 	case LA_ELS_PLOGI:
15036 		(void) sprintf(buf, "!PLOGI to D_ID=0x%%x failed");
15037 		break;
15038 
15039 	case LA_ELS_PRLI:
15040 		(void) sprintf(buf, "!PRLI to D_ID=0x%%x failed");
15041 		break;
15042 	}
15043 
15044 	if (scsi_cmd && fpkt->pkt_state == FC_PKT_SUCCESS) {
15045 		struct fcp_rsp		response, *rsp;
15046 		uchar_t			asc, ascq;
15047 		caddr_t			sense_key = NULL;
15048 		struct fcp_rsp_info	fcp_rsp_err, *bep;
15049 
15050 		if (icmd->ipkt_nodma) {
15051 			rsp = (struct fcp_rsp *)fpkt->pkt_resp;
15052 			bep = (struct fcp_rsp_info *)((caddr_t)rsp +
15053 			    sizeof (struct fcp_rsp));
15054 		} else {
15055 			rsp = &response;
15056 			bep = &fcp_rsp_err;
15057 
15058 			FCP_CP_IN(fpkt->pkt_resp, rsp, fpkt->pkt_resp_acc,
15059 			    sizeof (struct fcp_rsp));
15060 
15061 			FCP_CP_IN(fpkt->pkt_resp + sizeof (struct fcp_rsp),
15062 			    bep, fpkt->pkt_resp_acc,
15063 			    sizeof (struct fcp_rsp_info));
15064 		}
15065 
15066 
15067 		if (fcp_validate_fcp_response(rsp, pptr) != FC_SUCCESS) {
15068 			(void) sprintf(buf + strlen(buf),
15069 			    " : Bad FCP response values rsvd1=%%x, rsvd2=%%x,"
15070 			    " sts-rsvd1=%%x, sts-rsvd2=%%x, rsplen=%%x,"
15071 			    " senselen=%%x. Giving up");
15072 
15073 			fcp_log(CE_WARN, pptr->port_dip, buf,
15074 			    ptgt->tgt_d_id, plun->lun_num, rsp->reserved_0,
15075 			    rsp->reserved_1, rsp->fcp_u.fcp_status.reserved_0,
15076 			    rsp->fcp_u.fcp_status.reserved_1,
15077 			    rsp->fcp_response_len, rsp->fcp_sense_len);
15078 
15079 			kmem_free(buf, 256);
15080 			return;
15081 		}
15082 
15083 		if (rsp->fcp_u.fcp_status.rsp_len_set &&
15084 		    bep->rsp_code != FCP_NO_FAILURE) {
15085 			(void) sprintf(buf + strlen(buf),
15086 			    " FCP Response code = 0x%x", bep->rsp_code);
15087 		}
15088 
15089 		if (rsp->fcp_u.fcp_status.scsi_status & STATUS_CHECK) {
15090 			struct scsi_extended_sense sense_info, *sense_ptr;
15091 
15092 			if (icmd->ipkt_nodma) {
15093 				sense_ptr = (struct scsi_extended_sense *)
15094 				    ((caddr_t)fpkt->pkt_resp +
15095 				    sizeof (struct fcp_rsp) +
15096 				    rsp->fcp_response_len);
15097 			} else {
15098 				sense_ptr = &sense_info;
15099 
15100 				FCP_CP_IN(fpkt->pkt_resp +
15101 				    sizeof (struct fcp_rsp) +
15102 				    rsp->fcp_response_len, &sense_info,
15103 				    fpkt->pkt_resp_acc,
15104 				    sizeof (struct scsi_extended_sense));
15105 			}
15106 
15107 			if (sense_ptr->es_key < NUM_SENSE_KEYS +
15108 			    NUM_IMPL_SENSE_KEYS) {
15109 				sense_key = sense_keys[sense_ptr->es_key];
15110 			} else {
15111 				sense_key = "Undefined";
15112 			}
15113 
15114 			asc = sense_ptr->es_add_code;
15115 			ascq = sense_ptr->es_qual_code;
15116 
15117 			(void) sprintf(buf + strlen(buf),
15118 			    ": sense key=%%s, ASC=%%x," " ASCQ=%%x."
15119 			    " Giving up");
15120 
15121 			fcp_log(CE_WARN, pptr->port_dip, buf,
15122 			    ptgt->tgt_d_id, plun->lun_num, sense_key,
15123 			    asc, ascq);
15124 		} else {
15125 			(void) sprintf(buf + strlen(buf),
15126 			    " : SCSI status=%%x. Giving up");
15127 
15128 			fcp_log(CE_WARN, pptr->port_dip, buf,
15129 			    ptgt->tgt_d_id, plun->lun_num,
15130 			    rsp->fcp_u.fcp_status.scsi_status);
15131 		}
15132 	} else {
15133 		caddr_t state, reason, action, expln;
15134 
15135 		(void) fc_ulp_pkt_error(fpkt, &state, &reason,
15136 		    &action, &expln);
15137 
15138 		(void) sprintf(buf + strlen(buf), ": State:%%s,"
15139 		    " Reason:%%s. Giving up");
15140 
15141 		if (scsi_cmd) {
15142 			fcp_log(CE_WARN, pptr->port_dip, buf,
15143 			    ptgt->tgt_d_id, plun->lun_num, state, reason);
15144 		} else {
15145 			fcp_log(CE_WARN, pptr->port_dip, buf,
15146 			    ptgt->tgt_d_id, state, reason);
15147 		}
15148 	}
15149 
15150 	kmem_free(buf, 256);
15151 }
15152 
15153 
15154 static int
15155 fcp_handle_ipkt_errors(struct fcp_port *pptr, struct fcp_tgt *ptgt,
15156     struct fcp_ipkt *icmd, int rval, caddr_t op)
15157 {
15158 	int	ret = DDI_FAILURE;
15159 	char	*error;
15160 
15161 	switch (rval) {
15162 	case FC_DEVICE_BUSY_NEW_RSCN:
15163 		/*
15164 		 * This means that there was a new RSCN that the transport
15165 		 * knows about (which the ULP *may* know about too) but the
15166 		 * pkt that was sent down was related to an older RSCN. So, we
15167 		 * are just going to reset the retry count and deadline and
15168 		 * continue to retry. The idea is that transport is currently
15169 		 * working on the new RSCN and will soon let the ULPs know
15170 		 * about it and when it does the existing logic will kick in
15171 		 * where it will change the tcount to indicate that something
15172 		 * changed on the target. So, rediscovery will start and there
15173 		 * will not be an infinite retry.
15174 		 *
15175 		 * For a full flow of how the RSCN info is transferred back and
15176 		 * forth, see fp.c
15177 		 */
15178 		icmd->ipkt_retries = 0;
15179 		icmd->ipkt_port->port_deadline = fcp_watchdog_time +
15180 		    FCP_ICMD_DEADLINE;
15181 
15182 		FCP_TRACE(fcp_logq, pptr->port_instbuf, fcp_trace,
15183 		    FCP_BUF_LEVEL_3, 0,
15184 		    "fcp_handle_ipkt_errors: rval=%x  for D_ID=%x",
15185 		    rval, ptgt->tgt_d_id);
15186 		/* FALLTHROUGH */
15187 
15188 	case FC_STATEC_BUSY:
15189 	case FC_DEVICE_BUSY:
15190 	case FC_PBUSY:
15191 	case FC_FBUSY:
15192 	case FC_TRAN_BUSY:
15193 	case FC_OFFLINE:
15194 		FCP_TRACE(fcp_logq, pptr->port_instbuf, fcp_trace,
15195 		    FCP_BUF_LEVEL_3, 0,
15196 		    "fcp_handle_ipkt_errors: rval=%x  for D_ID=%x",
15197 		    rval, ptgt->tgt_d_id);
15198 		if (icmd->ipkt_retries < FCP_MAX_RETRIES &&
15199 		    fcp_is_retryable(icmd)) {
15200 			fcp_queue_ipkt(pptr, icmd->ipkt_fpkt);
15201 			ret = DDI_SUCCESS;
15202 		}
15203 		break;
15204 
15205 	case FC_LOGINREQ:
15206 		/*
15207 		 * FC_LOGINREQ used to be handled just like all the cases
15208 		 * above. It has been changed to handled a PRLI that fails
15209 		 * with FC_LOGINREQ different than other ipkts that fail
15210 		 * with FC_LOGINREQ. If a PRLI fails with FC_LOGINREQ it is
15211 		 * a simple matter to turn it into a PLOGI instead, so that's
15212 		 * exactly what we do here.
15213 		 */
15214 		if (icmd->ipkt_opcode == LA_ELS_PRLI) {
15215 			ret = fcp_send_els(icmd->ipkt_port, icmd->ipkt_tgt,
15216 			    icmd, LA_ELS_PLOGI, icmd->ipkt_link_cnt,
15217 			    icmd->ipkt_change_cnt, icmd->ipkt_cause);
15218 		} else {
15219 			FCP_TRACE(fcp_logq, pptr->port_instbuf, fcp_trace,
15220 			    FCP_BUF_LEVEL_3, 0,
15221 			    "fcp_handle_ipkt_errors: rval=%x  for D_ID=%x",
15222 			    rval, ptgt->tgt_d_id);
15223 			if (icmd->ipkt_retries < FCP_MAX_RETRIES &&
15224 			    fcp_is_retryable(icmd)) {
15225 				fcp_queue_ipkt(pptr, icmd->ipkt_fpkt);
15226 				ret = DDI_SUCCESS;
15227 			}
15228 		}
15229 		break;
15230 
15231 	default:
15232 		mutex_enter(&pptr->port_mutex);
15233 		mutex_enter(&ptgt->tgt_mutex);
15234 		if (!FCP_STATE_CHANGED(pptr, ptgt, icmd)) {
15235 			mutex_exit(&ptgt->tgt_mutex);
15236 			mutex_exit(&pptr->port_mutex);
15237 
15238 			(void) fc_ulp_error(rval, &error);
15239 			fcp_log(CE_WARN, pptr->port_dip,
15240 			    "!Failed to send %s to D_ID=%x error=%s",
15241 			    op, ptgt->tgt_d_id, error);
15242 		} else {
15243 			FCP_TRACE(fcp_logq, pptr->port_instbuf,
15244 			    fcp_trace, FCP_BUF_LEVEL_2, 0,
15245 			    "fcp_handle_ipkt_errors,1: state change occured"
15246 			    " for D_ID=0x%x", ptgt->tgt_d_id);
15247 			mutex_exit(&ptgt->tgt_mutex);
15248 			mutex_exit(&pptr->port_mutex);
15249 		}
15250 		break;
15251 	}
15252 
15253 	return (ret);
15254 }
15255 
15256 
15257 /*
15258  * Check of outstanding commands on any LUN for this target
15259  */
15260 static int
15261 fcp_outstanding_lun_cmds(struct fcp_tgt *ptgt)
15262 {
15263 	struct	fcp_lun	*plun;
15264 	struct	fcp_pkt	*cmd;
15265 
15266 	for (plun = ptgt->tgt_lun; plun != NULL; plun = plun->lun_next) {
15267 		mutex_enter(&plun->lun_mutex);
15268 		for (cmd = plun->lun_pkt_head; cmd != NULL;
15269 		    cmd = cmd->cmd_forw) {
15270 			if (cmd->cmd_state == FCP_PKT_ISSUED) {
15271 				mutex_exit(&plun->lun_mutex);
15272 				return (FC_SUCCESS);
15273 			}
15274 		}
15275 		mutex_exit(&plun->lun_mutex);
15276 	}
15277 
15278 	return (FC_FAILURE);
15279 }
15280 
15281 static fc_portmap_t *
15282 fcp_construct_map(struct fcp_port *pptr, uint32_t *dev_cnt)
15283 {
15284 	int			i;
15285 	fc_portmap_t		*devlist;
15286 	fc_portmap_t		*devptr = NULL;
15287 	struct fcp_tgt	*ptgt;
15288 
15289 	mutex_enter(&pptr->port_mutex);
15290 	for (i = 0, *dev_cnt = 0; i < FCP_NUM_HASH; i++) {
15291 		for (ptgt = pptr->port_tgt_hash_table[i]; ptgt != NULL;
15292 		    ptgt = ptgt->tgt_next) {
15293 			if (!(ptgt->tgt_state & FCP_TGT_ORPHAN)) {
15294 				++*dev_cnt;
15295 			}
15296 		}
15297 	}
15298 
15299 	devptr = devlist = kmem_zalloc(sizeof (*devlist) * *dev_cnt,
15300 	    KM_NOSLEEP);
15301 	if (devlist == NULL) {
15302 		mutex_exit(&pptr->port_mutex);
15303 		fcp_log(CE_WARN, pptr->port_dip,
15304 		    "!fcp%d: failed to allocate for portmap for construct map",
15305 		    pptr->port_instance);
15306 		return (devptr);
15307 	}
15308 
15309 	for (i = 0; i < FCP_NUM_HASH; i++) {
15310 		for (ptgt = pptr->port_tgt_hash_table[i]; ptgt != NULL;
15311 		    ptgt = ptgt->tgt_next) {
15312 			if (!(ptgt->tgt_state & FCP_TGT_ORPHAN)) {
15313 				int ret;
15314 
15315 				ret = fc_ulp_pwwn_to_portmap(
15316 				    pptr->port_fp_handle,
15317 				    (la_wwn_t *)&ptgt->tgt_port_wwn.raw_wwn[0],
15318 				    devlist);
15319 
15320 				if (ret == FC_SUCCESS) {
15321 					devlist++;
15322 					continue;
15323 				}
15324 
15325 				devlist->map_pd = NULL;
15326 				devlist->map_did.port_id = ptgt->tgt_d_id;
15327 				devlist->map_hard_addr.hard_addr =
15328 				    ptgt->tgt_hard_addr;
15329 
15330 				devlist->map_state = PORT_DEVICE_INVALID;
15331 				devlist->map_type = PORT_DEVICE_OLD;
15332 
15333 				bcopy(&ptgt->tgt_node_wwn.raw_wwn[0],
15334 				    &devlist->map_nwwn, FC_WWN_SIZE);
15335 
15336 				bcopy(&ptgt->tgt_port_wwn.raw_wwn[0],
15337 				    &devlist->map_pwwn, FC_WWN_SIZE);
15338 
15339 				devlist++;
15340 			}
15341 		}
15342 	}
15343 
15344 	mutex_exit(&pptr->port_mutex);
15345 
15346 	return (devptr);
15347 }
15348 /*
15349  * Inimate MPxIO that the lun is busy and cannot accept regular IO
15350  */
15351 static void
15352 fcp_update_mpxio_path_verifybusy(struct fcp_port *pptr)
15353 {
15354 	int i;
15355 	struct fcp_tgt	*ptgt;
15356 	struct fcp_lun	*plun;
15357 
15358 	for (i = 0; i < FCP_NUM_HASH; i++) {
15359 		for (ptgt = pptr->port_tgt_hash_table[i]; ptgt != NULL;
15360 		    ptgt = ptgt->tgt_next) {
15361 			mutex_enter(&ptgt->tgt_mutex);
15362 			for (plun = ptgt->tgt_lun; plun != NULL;
15363 			    plun = plun->lun_next) {
15364 				if (plun->lun_mpxio &&
15365 				    plun->lun_state & FCP_LUN_BUSY) {
15366 					if (!fcp_pass_to_hp(pptr, plun,
15367 					    plun->lun_cip,
15368 					    FCP_MPXIO_PATH_SET_BUSY,
15369 					    pptr->port_link_cnt,
15370 					    ptgt->tgt_change_cnt, 0, 0)) {
15371 						FCP_TRACE(fcp_logq,
15372 						    pptr->port_instbuf,
15373 						    fcp_trace,
15374 						    FCP_BUF_LEVEL_2, 0,
15375 						    "path_verifybusy: "
15376 						    "disable lun %p failed!",
15377 						    plun);
15378 					}
15379 				}
15380 			}
15381 			mutex_exit(&ptgt->tgt_mutex);
15382 		}
15383 	}
15384 }
15385 
15386 static int
15387 fcp_update_mpxio_path(struct fcp_lun *plun, child_info_t *cip, int what)
15388 {
15389 	dev_info_t		*cdip = NULL;
15390 	dev_info_t		*pdip = NULL;
15391 
15392 	ASSERT(plun);
15393 
15394 	mutex_enter(&plun->lun_mutex);
15395 	if (fcp_is_child_present(plun, cip) == FC_FAILURE) {
15396 		mutex_exit(&plun->lun_mutex);
15397 		return (NDI_FAILURE);
15398 	}
15399 	mutex_exit(&plun->lun_mutex);
15400 	cdip = mdi_pi_get_client(PIP(cip));
15401 	pdip = mdi_pi_get_phci(PIP(cip));
15402 
15403 	ASSERT(cdip != NULL);
15404 	ASSERT(pdip != NULL);
15405 
15406 	if (what == FCP_MPXIO_PATH_CLEAR_BUSY) {
15407 		/* LUN ready for IO */
15408 		(void) mdi_pi_enable_path(PIP(cip), DRIVER_DISABLE_TRANSIENT);
15409 	} else {
15410 		/* LUN busy to accept IO */
15411 		(void) mdi_pi_disable_path(PIP(cip), DRIVER_DISABLE_TRANSIENT);
15412 	}
15413 	return (NDI_SUCCESS);
15414 }
15415 
15416 /*
15417  * Caller must free the returned string of MAXPATHLEN len
15418  * If the device is offline (-1 instance number) NULL
15419  * will be returned.
15420  */
15421 static char *
15422 fcp_get_lun_path(struct fcp_lun *plun) {
15423 	dev_info_t	*dip = NULL;
15424 	char	*path = NULL;
15425 	if (plun == NULL) {
15426 		return (NULL);
15427 	}
15428 	if (plun->lun_mpxio == 0) {
15429 		dip = DIP(plun->lun_cip);
15430 	} else {
15431 		dip = mdi_pi_get_client(PIP(plun->lun_cip));
15432 	}
15433 	if (dip == NULL) {
15434 		return (NULL);
15435 	}
15436 	if (ddi_get_instance(dip) < 0) {
15437 		return (NULL);
15438 	}
15439 	path = kmem_alloc(MAXPATHLEN, KM_SLEEP);
15440 	if (path == NULL) {
15441 		return (NULL);
15442 	}
15443 
15444 	(void) ddi_pathname(dip, path);
15445 	/*
15446 	 * In reality, the user wants a fully valid path (one they can open)
15447 	 * but this string is lacking the mount point, and the minor node.
15448 	 * It would be nice if we could "figure these out" somehow
15449 	 * and fill them in.  Otherwise, the userland code has to understand
15450 	 * driver specific details of which minor node is the "best" or
15451 	 * "right" one to expose.  (Ex: which slice is the whole disk, or
15452 	 * which tape doesn't rewind)
15453 	 */
15454 	return (path);
15455 }
15456 
15457 static int
15458 fcp_scsi_bus_config(dev_info_t *parent, uint_t flag,
15459     ddi_bus_config_op_t op, void *arg, dev_info_t **childp)
15460 {
15461 	int64_t reset_delay;
15462 	int rval, retry = 0;
15463 	struct fcp_port *pptr = fcp_dip2port(parent);
15464 
15465 	reset_delay = (int64_t)(USEC_TO_TICK(FCP_INIT_WAIT_TIMEOUT)) -
15466 	    (lbolt64 - pptr->port_attach_time);
15467 	if (reset_delay < 0) {
15468 		reset_delay = 0;
15469 	}
15470 
15471 	if (fcp_bus_config_debug) {
15472 		flag |= NDI_DEVI_DEBUG;
15473 	}
15474 
15475 	switch (op) {
15476 	case BUS_CONFIG_ONE:
15477 		/*
15478 		 * Retry the command since we need to ensure
15479 		 * the fabric devices are available for root
15480 		 */
15481 		while (retry++ < fcp_max_bus_config_retries) {
15482 			rval =	(ndi_busop_bus_config(parent,
15483 			    flag | NDI_MDI_FALLBACK, op,
15484 			    arg, childp, (clock_t)reset_delay));
15485 			if (rval == 0) {
15486 				return (rval);
15487 			}
15488 		}
15489 
15490 		/*
15491 		 * drain taskq to make sure nodes are created and then
15492 		 * try again.
15493 		 */
15494 		taskq_wait(DEVI(parent)->devi_taskq);
15495 		return (ndi_busop_bus_config(parent, flag | NDI_MDI_FALLBACK,
15496 		    op, arg, childp, 0));
15497 
15498 	case BUS_CONFIG_DRIVER:
15499 	case BUS_CONFIG_ALL: {
15500 		/*
15501 		 * delay till all devices report in (port_tmp_cnt == 0)
15502 		 * or FCP_INIT_WAIT_TIMEOUT
15503 		 */
15504 		mutex_enter(&pptr->port_mutex);
15505 		while ((reset_delay > 0) && pptr->port_tmp_cnt) {
15506 			(void) cv_timedwait(&pptr->port_config_cv,
15507 			    &pptr->port_mutex,
15508 			    ddi_get_lbolt() + (clock_t)reset_delay);
15509 			reset_delay =
15510 			    (int64_t)(USEC_TO_TICK(FCP_INIT_WAIT_TIMEOUT)) -
15511 			    (lbolt64 - pptr->port_attach_time);
15512 		}
15513 		mutex_exit(&pptr->port_mutex);
15514 		/* drain taskq to make sure nodes are created */
15515 		taskq_wait(DEVI(parent)->devi_taskq);
15516 		return (ndi_busop_bus_config(parent, flag, op,
15517 		    arg, childp, 0));
15518 	}
15519 
15520 	default:
15521 		return (NDI_FAILURE);
15522 	}
15523 	/*NOTREACHED*/
15524 }
15525 
15526 static int
15527 fcp_scsi_bus_unconfig(dev_info_t *parent, uint_t flag,
15528     ddi_bus_config_op_t op, void *arg)
15529 {
15530 	if (fcp_bus_config_debug) {
15531 		flag |= NDI_DEVI_DEBUG;
15532 	}
15533 
15534 	return (ndi_busop_bus_unconfig(parent, flag, op, arg));
15535 }
15536 
15537 
15538 /*
15539  * Routine to copy GUID into the lun structure.
15540  * returns 0 if copy was successful and 1 if encountered a
15541  * failure and did not copy the guid.
15542  */
15543 static int
15544 fcp_copy_guid_2_lun_block(struct fcp_lun *plun, char *guidp)
15545 {
15546 
15547 	int retval = 0;
15548 
15549 	/* add one for the null terminator */
15550 	const unsigned int len = strlen(guidp) + 1;
15551 
15552 	if ((guidp == NULL) || (plun == NULL)) {
15553 		return (1);
15554 	}
15555 
15556 	/*
15557 	 * if the plun->lun_guid already has been allocated,
15558 	 * then check the size. if the size is exact, reuse
15559 	 * it....if not free it an allocate the required size.
15560 	 * The reallocation should NOT typically happen
15561 	 * unless the GUIDs reported changes between passes.
15562 	 * We free up and alloc again even if the
15563 	 * size was more than required. This is due to the
15564 	 * fact that the field lun_guid_size - serves
15565 	 * dual role of indicating the size of the wwn
15566 	 * size and ALSO the allocation size.
15567 	 */
15568 	if (plun->lun_guid) {
15569 		if (plun->lun_guid_size != len) {
15570 			/*
15571 			 * free the allocated memory and
15572 			 * initialize the field
15573 			 * lun_guid_size to 0.
15574 			 */
15575 			kmem_free(plun->lun_guid, plun->lun_guid_size);
15576 			plun->lun_guid = NULL;
15577 			plun->lun_guid_size = 0;
15578 		}
15579 	}
15580 	/*
15581 	 * alloc only if not already done.
15582 	 */
15583 	if (plun->lun_guid == NULL) {
15584 		plun->lun_guid = kmem_zalloc(len, KM_NOSLEEP);
15585 		if (plun->lun_guid == NULL) {
15586 			cmn_err(CE_WARN, "fcp_copy_guid_2_lun_block:"
15587 			    "Unable to allocate"
15588 			    "Memory for GUID!!! size %d", len);
15589 			retval = 1;
15590 		} else {
15591 			plun->lun_guid_size = len;
15592 		}
15593 	}
15594 	if (plun->lun_guid) {
15595 		/*
15596 		 * now copy the GUID
15597 		 */
15598 		bcopy(guidp, plun->lun_guid, plun->lun_guid_size);
15599 	}
15600 	return (retval);
15601 }
15602 
15603 /*
15604  * fcp_reconfig_wait
15605  *
15606  * Wait for a rediscovery/reconfiguration to complete before continuing.
15607  */
15608 
15609 static void
15610 fcp_reconfig_wait(struct fcp_port *pptr)
15611 {
15612 	clock_t		reconfig_start, wait_timeout;
15613 
15614 	/*
15615 	 * Quick check.	 If pptr->port_tmp_cnt is 0, there is no
15616 	 * reconfiguration in progress.
15617 	 */
15618 
15619 	mutex_enter(&pptr->port_mutex);
15620 	if (pptr->port_tmp_cnt == 0) {
15621 		mutex_exit(&pptr->port_mutex);
15622 		return;
15623 	}
15624 	mutex_exit(&pptr->port_mutex);
15625 
15626 	/*
15627 	 * If we cause a reconfig by raising power, delay until all devices
15628 	 * report in (port_tmp_cnt returns to 0)
15629 	 */
15630 
15631 	reconfig_start = ddi_get_lbolt();
15632 	wait_timeout = drv_usectohz(FCP_INIT_WAIT_TIMEOUT);
15633 
15634 	mutex_enter(&pptr->port_mutex);
15635 
15636 	while (((ddi_get_lbolt() - reconfig_start) < wait_timeout) &&
15637 	    pptr->port_tmp_cnt) {
15638 
15639 		(void) cv_timedwait(&pptr->port_config_cv, &pptr->port_mutex,
15640 		    reconfig_start + wait_timeout);
15641 	}
15642 
15643 	mutex_exit(&pptr->port_mutex);
15644 
15645 	/*
15646 	 * Even if fcp_tmp_count isn't 0, continue without error.  The port
15647 	 * we want may still be ok.  If not, it will error out later
15648 	 */
15649 }
15650 
15651 /*
15652  * Read masking info from fp.conf and construct the global fcp_lun_blacklist.
15653  * We rely on the fcp_global_mutex to provide protection against changes to
15654  * the fcp_lun_blacklist.
15655  *
15656  * You can describe a list of target port WWNs and LUN numbers which will
15657  * not be configured. LUN numbers will be interpreted as decimal. White
15658  * spaces and ',' can be used in the list of LUN numbers.
15659  *
15660  * To prevent LUNs 1 and 2 from being configured for target
15661  * port 510000f010fd92a1 and target port 510000e012079df1, set:
15662  *
15663  * pwwn-lun-blacklist=
15664  * "510000f010fd92a1,1,2",
15665  * "510000e012079df1,1,2";
15666  */
15667 static void
15668 fcp_read_blacklist(dev_info_t *dip,
15669     struct fcp_black_list_entry **pplun_blacklist) {
15670 	char **prop_array	= NULL;
15671 	char *curr_pwwn		= NULL;
15672 	char *curr_lun		= NULL;
15673 	uint32_t prop_item	= 0;
15674 	int idx			= 0;
15675 	int len			= 0;
15676 
15677 	ASSERT(mutex_owned(&fcp_global_mutex));
15678 	if (ddi_prop_lookup_string_array(DDI_DEV_T_ANY, dip,
15679 	    DDI_PROP_DONTPASS | DDI_PROP_NOTPROM,
15680 	    LUN_BLACKLIST_PROP, &prop_array, &prop_item) != DDI_PROP_SUCCESS) {
15681 		return;
15682 	}
15683 
15684 	for (idx = 0; idx < prop_item; idx++) {
15685 
15686 		curr_pwwn = prop_array[idx];
15687 		while (*curr_pwwn == ' ') {
15688 			curr_pwwn++;
15689 		}
15690 		if (strlen(curr_pwwn) <= (sizeof (la_wwn_t) * 2 + 1)) {
15691 			fcp_log(CE_WARN, NULL, "Invalid WWN %s in the blacklist"
15692 			    ", please check.", curr_pwwn);
15693 			continue;
15694 		}
15695 		if ((*(curr_pwwn + sizeof (la_wwn_t) * 2) != ' ') &&
15696 		    (*(curr_pwwn + sizeof (la_wwn_t) * 2) != ',')) {
15697 			fcp_log(CE_WARN, NULL, "Invalid WWN %s in the blacklist"
15698 			    ", please check.", curr_pwwn);
15699 			continue;
15700 		}
15701 		for (len = 0; len < sizeof (la_wwn_t) * 2; len++) {
15702 			if (isxdigit(curr_pwwn[len]) != TRUE) {
15703 				fcp_log(CE_WARN, NULL, "Invalid WWN %s in the "
15704 				    "blacklist, please check.", curr_pwwn);
15705 				break;
15706 			}
15707 		}
15708 		if (len != sizeof (la_wwn_t) * 2) {
15709 			continue;
15710 		}
15711 
15712 		curr_lun = curr_pwwn + sizeof (la_wwn_t) * 2 + 1;
15713 		*(curr_lun - 1) = '\0';
15714 		fcp_mask_pwwn_lun(curr_pwwn, curr_lun, pplun_blacklist);
15715 	}
15716 
15717 	ddi_prop_free(prop_array);
15718 }
15719 
15720 /*
15721  * Get the masking info about one remote target port designated by wwn.
15722  * Lun ids could be separated by ',' or white spaces.
15723  */
15724 static void
15725 fcp_mask_pwwn_lun(char *curr_pwwn, char *curr_lun,
15726     struct fcp_black_list_entry **pplun_blacklist) {
15727 	int		idx			= 0;
15728 	uint32_t	offset			= 0;
15729 	unsigned long	lun_id			= 0;
15730 	char		lunid_buf[16];
15731 	char		*pend			= NULL;
15732 	int		illegal_digit		= 0;
15733 
15734 	while (offset < strlen(curr_lun)) {
15735 		while ((curr_lun[offset + idx] != ',') &&
15736 		    (curr_lun[offset + idx] != '\0') &&
15737 		    (curr_lun[offset + idx] != ' ')) {
15738 			if (isdigit(curr_lun[offset + idx]) == 0) {
15739 				illegal_digit++;
15740 			}
15741 			idx++;
15742 		}
15743 		if (illegal_digit > 0) {
15744 			offset += (idx+1);	/* To the start of next lun */
15745 			idx = 0;
15746 			illegal_digit = 0;
15747 			fcp_log(CE_WARN, NULL, "Invalid LUN %s for WWN %s in "
15748 			    "the blacklist, please check digits.",
15749 			    curr_lun, curr_pwwn);
15750 			continue;
15751 		}
15752 		if (idx >= (sizeof (lunid_buf) / sizeof (lunid_buf[0]))) {
15753 			fcp_log(CE_WARN, NULL, "Invalid LUN %s for WWN %s in "
15754 			    "the blacklist, please check the length of LUN#.",
15755 			    curr_lun, curr_pwwn);
15756 			break;
15757 		}
15758 		if (idx == 0) {	/* ignore ' ' or ',' or '\0' */
15759 		    offset++;
15760 		    continue;
15761 		}
15762 
15763 		bcopy(curr_lun + offset, lunid_buf, idx);
15764 		lunid_buf[idx] = '\0';
15765 		if (ddi_strtoul(lunid_buf, &pend, 10, &lun_id) == 0) {
15766 			fcp_add_one_mask(curr_pwwn, lun_id, pplun_blacklist);
15767 		} else {
15768 			fcp_log(CE_WARN, NULL, "Invalid LUN %s for WWN %s in "
15769 			    "the blacklist, please check %s.",
15770 			    curr_lun, curr_pwwn, lunid_buf);
15771 		}
15772 		offset += (idx+1);	/* To the start of next lun */
15773 		idx = 0;
15774 	}
15775 }
15776 
15777 /*
15778  * Add one masking record
15779  */
15780 static void
15781 fcp_add_one_mask(char *curr_pwwn, uint32_t lun_id,
15782     struct fcp_black_list_entry **pplun_blacklist) {
15783 	struct fcp_black_list_entry	*tmp_entry	= *pplun_blacklist;
15784 	struct fcp_black_list_entry	*new_entry	= NULL;
15785 	la_wwn_t			wwn;
15786 
15787 	fcp_ascii_to_wwn(curr_pwwn, wwn.raw_wwn, sizeof (la_wwn_t));
15788 	while (tmp_entry) {
15789 		if ((bcmp(&tmp_entry->wwn, &wwn,
15790 		    sizeof (la_wwn_t)) == 0) && (tmp_entry->lun == lun_id)) {
15791 			return;
15792 		}
15793 
15794 		tmp_entry = tmp_entry->next;
15795 	}
15796 
15797 	/* add to black list */
15798 	new_entry = (struct fcp_black_list_entry *)kmem_zalloc
15799 	    (sizeof (struct fcp_black_list_entry), KM_SLEEP);
15800 	bcopy(&wwn, &new_entry->wwn, sizeof (la_wwn_t));
15801 	new_entry->lun = lun_id;
15802 	new_entry->masked = 0;
15803 	new_entry->next = *pplun_blacklist;
15804 	*pplun_blacklist = new_entry;
15805 }
15806 
15807 /*
15808  * Check if we should mask the specified lun of this fcp_tgt
15809  */
15810 static int
15811 fcp_should_mask(la_wwn_t *wwn, uint32_t lun_id) {
15812 	struct fcp_black_list_entry *remote_port;
15813 
15814 	remote_port = fcp_lun_blacklist;
15815 	while (remote_port != NULL) {
15816 		if (bcmp(wwn, &remote_port->wwn, sizeof (la_wwn_t)) == 0) {
15817 			if (remote_port->lun == lun_id) {
15818 				remote_port->masked++;
15819 				if (remote_port->masked == 1) {
15820 					fcp_log(CE_NOTE, NULL, "LUN %d of port "
15821 					    "%02x%02x%02x%02x%02x%02x%02x%02x "
15822 					    "is masked due to black listing.\n",
15823 					    lun_id, wwn->raw_wwn[0],
15824 					    wwn->raw_wwn[1], wwn->raw_wwn[2],
15825 					    wwn->raw_wwn[3], wwn->raw_wwn[4],
15826 					    wwn->raw_wwn[5], wwn->raw_wwn[6],
15827 					    wwn->raw_wwn[7]);
15828 				}
15829 				return (TRUE);
15830 			}
15831 		}
15832 		remote_port = remote_port->next;
15833 	}
15834 	return (FALSE);
15835 }
15836 
15837 /*
15838  * Release all allocated resources
15839  */
15840 static void
15841 fcp_cleanup_blacklist(struct fcp_black_list_entry **pplun_blacklist) {
15842 	struct fcp_black_list_entry	*tmp_entry	= *pplun_blacklist;
15843 	struct fcp_black_list_entry	*current_entry	= NULL;
15844 
15845 	ASSERT(mutex_owned(&fcp_global_mutex));
15846 	/*
15847 	 * Traverse all luns
15848 	 */
15849 	while (tmp_entry) {
15850 		current_entry = tmp_entry;
15851 		tmp_entry = tmp_entry->next;
15852 		kmem_free(current_entry, sizeof (struct fcp_black_list_entry));
15853 	}
15854 	*pplun_blacklist = NULL;
15855 }
15856