1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 /* 22 * Copyright 2008 Sun Microsystems, Inc. All rights reserved. 23 * Use is subject to license terms. 24 */ 25 26 /* 27 * SunOS 5.x Multithreaded STREAMS DLPI FCIP Module 28 * This is a pseudo driver module to handle encapsulation of IP and ARP 29 * datagrams over FibreChannel interfaces. FCIP is a cloneable STREAMS 30 * driver module which interfaces with IP/ARP using DLPI. This module 31 * is a Style-2 DLS provider. 32 * 33 * The implementation of this module is based on RFC 2625 which gives 34 * details on the encapsulation of IP/ARP data over FibreChannel. 35 * The fcip module needs to resolve an IP address to a port address before 36 * sending data to a destination port. A FC device port has 2 addresses 37 * associated with it: A 8 byte World Wide unique Port Name and a 3 byte 38 * volatile Port number or Port_ID. 39 * 40 * The mapping between a IP address and the World Wide Port Name is handled 41 * by the ARP layer since the IP over FC draft requires the MAC address to 42 * be the least significant six bytes of the WorldWide Port Names. The 43 * fcip module however needs to identify the destination port uniquely when 44 * the destination FC device has multiple FC ports. 45 * 46 * The FC layer mapping between the World Wide Port Name and the Port_ID 47 * will be handled through the use of a fabric name server or through the 48 * use of the FARP ELS command as described in the draft. Since the Port_IDs 49 * are volatile, the mapping between the World Wide Port Name and Port_IDs 50 * must be maintained and validated before use each time a datagram 51 * needs to be sent to the destination ports. The FC transport module 52 * informs the fcip module of all changes to states of ports on the 53 * fabric through registered callbacks. This enables the fcip module 54 * to maintain the WW_PN to Port_ID mappings current. 55 * 56 * For details on how this module interfaces with the FibreChannel Transport 57 * modules, refer to PSARC/1997/385. Chapter 3 of the FibreChannel Transport 58 * Programming guide details the APIs between ULPs and the Transport. 59 * 60 * Now for some Caveats: 61 * 62 * RFC 2625 requires that a FibreChannel Port name (the Port WWN) have 63 * the NAA bits set to '0001' indicating a IEEE 48bit address which 64 * corresponds to a ULA (Universal LAN MAC address). But with FibreChannel 65 * adapters containing 2 or more ports, IEEE naming cannot identify the 66 * ports on an adapter uniquely so we will in the first implementation 67 * be operating only on Port 0 of each adapter. 68 */ 69 70 #include <sys/types.h> 71 #include <sys/errno.h> 72 #include <sys/debug.h> 73 #include <sys/time.h> 74 #include <sys/sysmacros.h> 75 #include <sys/systm.h> 76 #include <sys/user.h> 77 #include <sys/stropts.h> 78 #include <sys/stream.h> 79 #include <sys/strlog.h> 80 #include <sys/strsubr.h> 81 #include <sys/cmn_err.h> 82 #include <sys/cpu.h> 83 #include <sys/kmem.h> 84 #include <sys/conf.h> 85 #include <sys/ddi.h> 86 #include <sys/sunddi.h> 87 #include <sys/ksynch.h> 88 #include <sys/stat.h> 89 #include <sys/kstat.h> 90 #include <sys/vtrace.h> 91 #include <sys/strsun.h> 92 #include <sys/varargs.h> 93 #include <sys/modctl.h> 94 #include <sys/thread.h> 95 #include <sys/var.h> 96 #include <sys/proc.h> 97 #include <inet/common.h> 98 #include <netinet/ip6.h> 99 #include <inet/ip.h> 100 #include <inet/arp.h> 101 #include <inet/mi.h> 102 #include <inet/nd.h> 103 #include <sys/dlpi.h> 104 #include <sys/ethernet.h> 105 #include <sys/file.h> 106 #include <sys/syslog.h> 107 #include <sys/disp.h> 108 #include <sys/taskq.h> 109 110 /* 111 * Leadville includes 112 */ 113 114 #include <sys/fibre-channel/fc.h> 115 #include <sys/fibre-channel/impl/fc_ulpif.h> 116 #include <sys/fibre-channel/ulp/fcip.h> 117 118 /* 119 * TNF Probe/trace facility include 120 */ 121 #if defined(lint) || defined(FCIP_TNF_ENABLED) 122 #include <sys/tnf_probe.h> 123 #endif 124 125 #define FCIP_ESBALLOC 126 127 /* 128 * Function prototypes 129 */ 130 131 /* standard loadable modules entry points */ 132 static int fcip_attach(dev_info_t *, ddi_attach_cmd_t); 133 static int fcip_detach(dev_info_t *, ddi_detach_cmd_t); 134 static void fcip_dodetach(struct fcipstr *slp); 135 static int fcip_getinfo(dev_info_t *dip, ddi_info_cmd_t cmd, 136 void *arg, void **result); 137 138 139 /* streams specific */ 140 static void fcip_setipq(struct fcip *fptr); 141 static int fcip_wput(queue_t *, mblk_t *); 142 static int fcip_wsrv(queue_t *); 143 static void fcip_proto(queue_t *, mblk_t *); 144 static void fcip_ioctl(queue_t *, mblk_t *); 145 static int fcip_open(queue_t *wq, dev_t *devp, int flag, 146 int sflag, cred_t *credp); 147 static int fcip_close(queue_t *rq, int flag, int otyp, cred_t *credp); 148 static int fcip_start(queue_t *wq, mblk_t *mp, struct fcip *fptr, 149 struct fcip_dest *fdestp, int flags); 150 static void fcip_sendup(struct fcip *fptr, mblk_t *mp, 151 struct fcipstr *(*acceptfunc)()); 152 static struct fcipstr *fcip_accept(struct fcipstr *slp, struct fcip *fptr, 153 int type, la_wwn_t *dhostp); 154 static mblk_t *fcip_addudind(struct fcip *fptr, mblk_t *mp, 155 fcph_network_hdr_t *nhdr, int type); 156 static int fcip_setup_mac_addr(struct fcip *fptr); 157 static void fcip_kstat_init(struct fcip *fptr); 158 static int fcip_stat_update(kstat_t *, int); 159 160 161 /* dlpi specific */ 162 static void fcip_spareq(queue_t *wq, mblk_t *mp); 163 static void fcip_pareq(queue_t *wq, mblk_t *mp); 164 static void fcip_ubreq(queue_t *wq, mblk_t *mp); 165 static void fcip_breq(queue_t *wq, mblk_t *mp); 166 static void fcip_dreq(queue_t *wq, mblk_t *mp); 167 static void fcip_areq(queue_t *wq, mblk_t *mp); 168 static void fcip_udreq(queue_t *wq, mblk_t *mp); 169 static void fcip_ireq(queue_t *wq, mblk_t *mp); 170 static void fcip_dl_ioc_hdr_info(queue_t *wq, mblk_t *mp); 171 172 173 /* solaris sundry, DR/CPR etc */ 174 static int fcip_cache_constructor(void *buf, void *arg, int size); 175 static void fcip_cache_destructor(void *buf, void *size); 176 static int fcip_handle_suspend(fcip_port_info_t *fport, fc_detach_cmd_t cmd); 177 static int fcip_handle_resume(fcip_port_info_t *fport, 178 fc_ulp_port_info_t *port_info, fc_attach_cmd_t cmd); 179 static fcip_port_info_t *fcip_softstate_free(fcip_port_info_t *fport); 180 static int fcip_port_attach_handler(struct fcip *fptr); 181 182 183 /* 184 * ulp - transport interface function prototypes 185 */ 186 static int fcip_port_attach(opaque_t ulp_handle, fc_ulp_port_info_t *, 187 fc_attach_cmd_t cmd, uint32_t sid); 188 static int fcip_port_detach(opaque_t ulp_handle, fc_ulp_port_info_t *, 189 fc_detach_cmd_t cmd); 190 static int fcip_port_ioctl(opaque_t ulp_handle, opaque_t port_handle, 191 dev_t dev, int cmd, intptr_t data, int mode, cred_t *credp, int *rval, 192 uint32_t claimed); 193 static void fcip_statec_cb(opaque_t ulp_handle, opaque_t phandle, 194 uint32_t port_state, uint32_t port_top, fc_portmap_t changelist[], 195 uint32_t listlen, uint32_t sid); 196 static int fcip_els_cb(opaque_t ulp_handle, opaque_t phandle, 197 fc_unsol_buf_t *buf, uint32_t claimed); 198 static int fcip_data_cb(opaque_t ulp_handle, opaque_t phandle, 199 fc_unsol_buf_t *payload, uint32_t claimed); 200 201 202 /* Routing table specific */ 203 static void fcip_handle_topology(struct fcip *fptr); 204 static int fcip_init_port(struct fcip *fptr); 205 struct fcip_routing_table *fcip_lookup_rtable(struct fcip *fptr, 206 la_wwn_t *pwwn, int matchflag); 207 static void fcip_rt_update(struct fcip *fptr, fc_portmap_t *devlist, 208 uint32_t listlen); 209 static void fcip_rt_flush(struct fcip *fptr); 210 static void fcip_rte_remove_deferred(void *arg); 211 static int fcip_do_plogi(struct fcip *fptr, struct fcip_routing_table *frp); 212 213 214 /* dest table specific */ 215 static struct fcip_dest *fcip_get_dest(struct fcip *fptr, 216 la_wwn_t *dlphys); 217 static struct fcip_dest *fcip_add_dest(struct fcip *fptr, 218 struct fcip_routing_table *frp); 219 static int fcip_dest_add_broadcast_entry(struct fcip *fptr, int new_flag); 220 static uint32_t fcip_get_broadcast_did(struct fcip *fptr); 221 static void fcip_cleanup_dest(struct fcip *fptr); 222 223 224 /* helper functions */ 225 static fcip_port_info_t *fcip_get_port(opaque_t phandle); 226 static int fcip_wwn_compare(la_wwn_t *wwn1, la_wwn_t *wwn2, int flag); 227 static void fcip_ether_to_str(struct ether_addr *e, caddr_t s); 228 static int fcip_port_get_num_pkts(struct fcip *fptr); 229 static int fcip_check_port_busy(struct fcip *fptr); 230 static void fcip_check_remove_minor_node(void); 231 static int fcip_set_wwn(la_wwn_t *pwwn); 232 static int fcip_plogi_in_progress(struct fcip *fptr); 233 static int fcip_check_port_exists(struct fcip *fptr); 234 static int fcip_is_supported_fc_topology(int fc_topology); 235 236 237 /* pkt specific */ 238 static fcip_pkt_t *fcip_pkt_alloc(struct fcip *fptr, mblk_t *bp, 239 int flags, int datalen); 240 static void fcip_pkt_free(struct fcip_pkt *fcip_pkt, int flags); 241 static fcip_pkt_t *fcip_ipkt_alloc(struct fcip *fptr, int cmdlen, 242 int resplen, opaque_t pd, int flags); 243 static void fcip_ipkt_free(fcip_pkt_t *fcip_pkt); 244 static void fcip_ipkt_callback(fc_packet_t *fc_pkt); 245 static void fcip_free_pkt_dma(fcip_pkt_t *fcip_pkt); 246 static void fcip_pkt_callback(fc_packet_t *fc_pkt); 247 static void fcip_init_unicast_pkt(fcip_pkt_t *fcip_pkt, fc_portid_t sid, 248 fc_portid_t did, void (*comp) ()); 249 static int fcip_transport(fcip_pkt_t *fcip_pkt); 250 static void fcip_pkt_timeout(void *arg); 251 static void fcip_timeout(void *arg); 252 static void fcip_fdestp_enqueue_pkt(struct fcip_dest *fdestp, 253 fcip_pkt_t *fcip_pkt); 254 static int fcip_fdestp_dequeue_pkt(struct fcip_dest *fdestp, 255 fcip_pkt_t *fcip_pkt); 256 static int fcip_sendup_constructor(void *buf, void *arg, int flags); 257 static void fcip_sendup_thr(void *arg); 258 static int fcip_sendup_alloc_enque(struct fcip *ftpr, mblk_t *mp, 259 struct fcipstr *(*f)()); 260 261 /* 262 * zero copy inbound data handling 263 */ 264 #ifdef FCIP_ESBALLOC 265 static void fcip_ubfree(char *arg); 266 #endif /* FCIP_ESBALLOC */ 267 268 #if !defined(FCIP_ESBALLOC) 269 static void *fcip_allocb(size_t size, uint_t pri); 270 #endif 271 272 273 /* FCIP FARP support functions */ 274 static struct fcip_dest *fcip_do_farp(struct fcip *fptr, la_wwn_t *pwwn, 275 char *ip_addr, size_t ip_addr_len, int flags); 276 static void fcip_init_broadcast_pkt(fcip_pkt_t *fcip_pkt, void (*comp) (), 277 int is_els); 278 static int fcip_handle_farp_request(struct fcip *fptr, la_els_farp_t *fcmd); 279 static int fcip_handle_farp_response(struct fcip *fptr, la_els_farp_t *fcmd); 280 static void fcip_cache_arp_broadcast(struct fcip *ftpr, fc_unsol_buf_t *buf); 281 static void fcip_port_ns(void *arg); 282 283 #ifdef DEBUG 284 285 #include <sys/debug.h> 286 287 #define FCIP_DEBUG_DEFAULT 0x1 288 #define FCIP_DEBUG_ATTACH 0x2 289 #define FCIP_DEBUG_INIT 0x4 290 #define FCIP_DEBUG_DETACH 0x8 291 #define FCIP_DEBUG_DLPI 0x10 292 #define FCIP_DEBUG_ELS 0x20 293 #define FCIP_DEBUG_DOWNSTREAM 0x40 294 #define FCIP_DEBUG_UPSTREAM 0x80 295 #define FCIP_DEBUG_MISC 0x100 296 297 #define FCIP_DEBUG_STARTUP (FCIP_DEBUG_ATTACH|FCIP_DEBUG_INIT) 298 #define FCIP_DEBUG_DATAOUT (FCIP_DEBUG_DLPI|FCIP_DEBUG_DOWNSTREAM) 299 #define FCIP_DEBUG_DATAIN (FCIP_DEBUG_ELS|FCIP_DEBUG_UPSTREAM) 300 301 static int fcip_debug = FCIP_DEBUG_DEFAULT; 302 303 #define FCIP_DEBUG(level, args) \ 304 if (fcip_debug & (level)) cmn_err args; 305 306 #else /* DEBUG */ 307 308 #define FCIP_DEBUG(level, args) /* do nothing */ 309 310 #endif /* DEBUG */ 311 312 #define KIOIP KSTAT_INTR_PTR(fcip->fcip_intrstats) 313 314 /* 315 * Endian independent ethernet to WWN copy 316 */ 317 #define ether_to_wwn(E, W) \ 318 bzero((void *)(W), sizeof (la_wwn_t)); \ 319 bcopy((void *)(E), (void *)&((W)->raw_wwn[2]), ETHERADDRL); \ 320 (W)->raw_wwn[0] |= 0x10 321 322 /* 323 * wwn_to_ether : Endian independent, copies a WWN to struct ether_addr. 324 * The args to the macro are pointers to WWN and ether_addr structures 325 */ 326 #define wwn_to_ether(W, E) \ 327 bcopy((void *)&((W)->raw_wwn[2]), (void *)E, ETHERADDRL) 328 329 /* 330 * The module_info structure contains identification and limit values. 331 * All queues associated with a certain driver share the same module_info 332 * structures. This structure defines the characteristics of that driver/ 333 * module's queues. The module name must be unique. The max and min packet 334 * sizes limit the no. of characters in M_DATA messages. The Hi and Lo 335 * water marks are for flow control when a module has a service procedure. 336 */ 337 static struct module_info fcipminfo = { 338 FCIPIDNUM, /* mi_idnum : Module ID num */ 339 FCIPNAME, /* mi_idname: Module Name */ 340 FCIPMINPSZ, /* mi_minpsz: Min packet size */ 341 FCIPMAXPSZ, /* mi_maxpsz: Max packet size */ 342 FCIPHIWAT, /* mi_hiwat : High water mark */ 343 FCIPLOWAT /* mi_lowat : Low water mark */ 344 }; 345 346 /* 347 * The qinit structres contain the module put, service. open and close 348 * procedure pointers. All modules and drivers with the same streamtab 349 * file (i.e same fmodsw or cdevsw entry points) point to the same 350 * upstream (read) and downstream (write) qinit structs. 351 */ 352 static struct qinit fcip_rinit = { 353 NULL, /* qi_putp */ 354 NULL, /* qi_srvp */ 355 fcip_open, /* qi_qopen */ 356 fcip_close, /* qi_qclose */ 357 NULL, /* qi_qadmin */ 358 &fcipminfo, /* qi_minfo */ 359 NULL /* qi_mstat */ 360 }; 361 362 static struct qinit fcip_winit = { 363 fcip_wput, /* qi_putp */ 364 fcip_wsrv, /* qi_srvp */ 365 NULL, /* qi_qopen */ 366 NULL, /* qi_qclose */ 367 NULL, /* qi_qadmin */ 368 &fcipminfo, /* qi_minfo */ 369 NULL /* qi_mstat */ 370 }; 371 372 /* 373 * streamtab contains pointers to the read and write qinit structures 374 */ 375 376 static struct streamtab fcip_info = { 377 &fcip_rinit, /* st_rdinit */ 378 &fcip_winit, /* st_wrinit */ 379 NULL, /* st_muxrinit */ 380 NULL, /* st_muxwrinit */ 381 }; 382 383 static struct cb_ops fcip_cb_ops = { 384 nodev, /* open */ 385 nodev, /* close */ 386 nodev, /* strategy */ 387 nodev, /* print */ 388 nodev, /* dump */ 389 nodev, /* read */ 390 nodev, /* write */ 391 nodev, /* ioctl */ 392 nodev, /* devmap */ 393 nodev, /* mmap */ 394 nodev, /* segmap */ 395 nochpoll, /* poll */ 396 ddi_prop_op, /* cb_prop_op */ 397 &fcip_info, /* streamtab */ 398 D_MP | D_HOTPLUG, /* Driver compatibility flag */ 399 CB_REV, /* rev */ 400 nodev, /* int (*cb_aread)() */ 401 nodev /* int (*cb_awrite)() */ 402 }; 403 404 /* 405 * autoconfiguration routines. 406 */ 407 static struct dev_ops fcip_ops = { 408 DEVO_REV, /* devo_rev, */ 409 0, /* refcnt */ 410 fcip_getinfo, /* info */ 411 nulldev, /* identify */ 412 nulldev, /* probe */ 413 fcip_attach, /* attach */ 414 fcip_detach, /* detach */ 415 nodev, /* RESET */ 416 &fcip_cb_ops, /* driver operations */ 417 NULL, /* bus operations */ 418 ddi_power /* power management */ 419 }; 420 421 #define FCIP_VERSION "1.61" 422 #define FCIP_NAME "SunFC FCIP v" FCIP_VERSION 423 424 #define PORT_DRIVER "fp" 425 426 #define GETSTRUCT(struct, number) \ 427 ((struct *)kmem_zalloc((size_t)(sizeof (struct) * (number)), \ 428 KM_SLEEP)) 429 430 static struct modldrv modldrv = { 431 &mod_driverops, /* Type of module - driver */ 432 FCIP_NAME, /* Name of module */ 433 &fcip_ops, /* driver ops */ 434 }; 435 436 static struct modlinkage modlinkage = { 437 MODREV_1, (void *)&modldrv, NULL 438 }; 439 440 441 /* 442 * Now for some global statics 443 */ 444 static uint32_t fcip_ub_nbufs = FCIP_UB_NBUFS; 445 static uint32_t fcip_ub_size = FCIP_UB_SIZE; 446 static int fcip_pkt_ttl_ticks = FCIP_PKT_TTL; 447 static int fcip_tick_incr = 1; 448 static int fcip_wait_cmds = FCIP_WAIT_CMDS; 449 static int fcip_num_attaching = 0; 450 static int fcip_port_attach_pending = 0; 451 static int fcip_create_nodes_on_demand = 1; /* keep it similar to fcp */ 452 static int fcip_cache_on_arp_broadcast = 0; 453 static int fcip_farp_supported = 0; 454 static int fcip_minor_node_created = 0; 455 456 /* 457 * Supported FCAs 458 */ 459 #define QLC_PORT_1_ID_BITS 0x100 460 #define QLC_PORT_2_ID_BITS 0x101 461 #define QLC_PORT_NAA 0x2 462 #define QLC_MODULE_NAME "qlc" 463 #define IS_QLC_PORT(port_dip) \ 464 (strcmp(ddi_driver_name(ddi_get_parent((port_dip))),\ 465 QLC_MODULE_NAME) == 0) 466 467 468 /* 469 * fcip softstate structures head. 470 */ 471 472 static void *fcip_softp = NULL; 473 474 /* 475 * linked list of active (inuse) driver streams 476 */ 477 478 static int fcip_num_instances = 0; 479 static dev_info_t *fcip_module_dip = (dev_info_t *)0; 480 481 482 /* 483 * Ethernet broadcast address: Broadcast addressing in IP over fibre 484 * channel should be the IEEE ULA (also the low 6 bytes of the Port WWN). 485 * 486 * The broadcast addressing varies for differing topologies a node may be in: 487 * - On a private loop the ARP broadcast is a class 3 sequence sent 488 * using OPNfr (Open Broadcast Replicate primitive) followed by 489 * the ARP frame to D_ID 0xFFFFFF 490 * 491 * - On a public Loop the broadcast sequence is sent to AL_PA 0x00 492 * (no OPNfr primitive). 493 * 494 * - For direct attach and point to point topologies we just send 495 * the frame to D_ID 0xFFFFFF 496 * 497 * For public loop the handling would probably be different - for now 498 * I'll just declare this struct - It can be deleted if not necessary. 499 * 500 */ 501 502 503 /* 504 * DL_INFO_ACK template for the fcip module. The dl_info_ack_t structure is 505 * returned as a part of an DL_INFO_ACK message which is a M_PCPROTO message 506 * returned in response to a DL_INFO_REQ message sent to us from a DLS user 507 * Let us fake an ether header as much as possible. 508 * 509 * dl_addr_length is the Provider's DLSAP addr which is SAP addr + 510 * Physical addr of the provider. We set this to 511 * ushort_t + sizeof (la_wwn_t) for Fibre Channel ports. 512 * dl_mac_type Lets just use DL_ETHER - we can try using DL_IPFC, a new 513 * dlpi.h define later. 514 * dl_sap_length -2 indicating the SAP address follows the Physical addr 515 * component in the DLSAP addr. 516 * dl_service_mode: DLCLDS - connectionless data link service. 517 * 518 */ 519 520 static dl_info_ack_t fcip_infoack = { 521 DL_INFO_ACK, /* dl_primitive */ 522 FCIPMTU, /* dl_max_sdu */ 523 0, /* dl_min_sdu */ 524 FCIPADDRL, /* dl_addr_length */ 525 DL_ETHER, /* dl_mac_type */ 526 0, /* dl_reserved */ 527 0, /* dl_current_state */ 528 -2, /* dl_sap_length */ 529 DL_CLDLS, /* dl_service_mode */ 530 0, /* dl_qos_length */ 531 0, /* dl_qos_offset */ 532 0, /* dl_range_length */ 533 0, /* dl_range_offset */ 534 DL_STYLE2, /* dl_provider_style */ 535 sizeof (dl_info_ack_t), /* dl_addr_offset */ 536 DL_VERSION_2, /* dl_version */ 537 ETHERADDRL, /* dl_brdcst_addr_length */ 538 sizeof (dl_info_ack_t) + FCIPADDRL, /* dl_brdcst_addr_offset */ 539 0 /* dl_growth */ 540 }; 541 542 /* 543 * FCIP broadcast address definition. 544 */ 545 static struct ether_addr fcipnhbroadcastaddr = { 546 0xff, 0xff, 0xff, 0xff, 0xff, 0xff 547 }; 548 549 /* 550 * RFC2625 requires the broadcast ARP address in the ARP data payload to 551 * be set to 0x00 00 00 00 00 00 for ARP broadcast packets 552 */ 553 static struct ether_addr fcip_arpbroadcast_addr = { 554 0x00, 0x00, 0x00, 0x00, 0x00, 0x00 555 }; 556 557 558 #define ether_bcopy(src, dest) bcopy((src), (dest), ETHERADDRL); 559 560 /* 561 * global kernel locks 562 */ 563 static kcondvar_t fcip_global_cv; 564 static kmutex_t fcip_global_mutex; 565 566 /* 567 * fctl external defines 568 */ 569 extern int fc_ulp_add(fc_ulp_modinfo_t *); 570 571 /* 572 * fctl data structures 573 */ 574 575 #define FCIP_REV 0x07 576 577 /* linked list of port info structures */ 578 static fcip_port_info_t *fcip_port_head = NULL; 579 580 /* linked list of fcip structures */ 581 static struct fcipstr *fcipstrup = NULL; 582 static krwlock_t fcipstruplock; 583 584 585 /* 586 * Module information structure. This structure gives the FC Transport modules 587 * information about an ULP that registers with it. 588 */ 589 static fc_ulp_modinfo_t fcip_modinfo = { 590 0, /* for xref checks? */ 591 FCTL_ULP_MODREV_4, /* FCIP revision */ 592 FC_TYPE_IS8802_SNAP, /* type 5 for SNAP encapsulated datagrams */ 593 FCIP_NAME, /* module name as in the modldrv struct */ 594 0x0, /* get all statec callbacks for now */ 595 fcip_port_attach, /* port attach callback */ 596 fcip_port_detach, /* port detach callback */ 597 fcip_port_ioctl, /* port ioctl callback */ 598 fcip_els_cb, /* els callback */ 599 fcip_data_cb, /* data callback */ 600 fcip_statec_cb /* state change callback */ 601 }; 602 603 604 /* 605 * Solaris 9 and up, the /kernel/drv/fp.conf file will have the following entry 606 * 607 * ddi-forceattach=1; 608 * 609 * This will ensure that fp is loaded at bootup. No additional checks are needed 610 */ 611 int 612 _init(void) 613 { 614 int rval; 615 616 FCIP_TNF_LOAD(); 617 618 /* 619 * Initialize the mutexs used by port attach and other callbacks. 620 * The transport can call back into our port_attach_callback 621 * routine even before _init() completes and bad things can happen. 622 */ 623 mutex_init(&fcip_global_mutex, NULL, MUTEX_DRIVER, NULL); 624 cv_init(&fcip_global_cv, NULL, CV_DRIVER, NULL); 625 rw_init(&fcipstruplock, NULL, RW_DRIVER, NULL); 626 627 mutex_enter(&fcip_global_mutex); 628 fcip_port_attach_pending = 1; 629 mutex_exit(&fcip_global_mutex); 630 631 /* 632 * Now attempt to register fcip with the transport. 633 * If fc_ulp_add fails, fcip module will not be loaded. 634 */ 635 rval = fc_ulp_add(&fcip_modinfo); 636 if (rval != FC_SUCCESS) { 637 mutex_destroy(&fcip_global_mutex); 638 cv_destroy(&fcip_global_cv); 639 rw_destroy(&fcipstruplock); 640 switch (rval) { 641 case FC_ULP_SAMEMODULE: 642 FCIP_DEBUG(FCIP_DEBUG_DEFAULT, (CE_WARN, 643 "!fcip: module is already registered with" 644 " transport")); 645 rval = EEXIST; 646 break; 647 case FC_ULP_SAMETYPE: 648 FCIP_DEBUG(FCIP_DEBUG_DEFAULT, (CE_WARN, 649 "!fcip: Another module of the same ULP type 0x%x" 650 " is already registered with the transport", 651 fcip_modinfo.ulp_type)); 652 rval = EEXIST; 653 break; 654 case FC_BADULP: 655 FCIP_DEBUG(FCIP_DEBUG_DEFAULT, (CE_WARN, 656 "!fcip: Current fcip version 0x%x does not match" 657 " fctl version", 658 fcip_modinfo.ulp_rev)); 659 rval = ENODEV; 660 break; 661 default: 662 FCIP_DEBUG(FCIP_DEBUG_DEFAULT, (CE_WARN, 663 "!fcip: fc_ulp_add failed with status 0x%x", rval)); 664 rval = ENODEV; 665 break; 666 } 667 FCIP_TNF_UNLOAD(&modlinkage); 668 return (rval); 669 } 670 671 if ((rval = ddi_soft_state_init(&fcip_softp, sizeof (struct fcip), 672 FCIP_NUM_INSTANCES)) != 0) { 673 mutex_destroy(&fcip_global_mutex); 674 cv_destroy(&fcip_global_cv); 675 rw_destroy(&fcipstruplock); 676 (void) fc_ulp_remove(&fcip_modinfo); 677 FCIP_TNF_UNLOAD(&modlinkage); 678 return (rval); 679 } 680 681 if ((rval = mod_install(&modlinkage)) != 0) { 682 FCIP_TNF_UNLOAD(&modlinkage); 683 (void) fc_ulp_remove(&fcip_modinfo); 684 mutex_destroy(&fcip_global_mutex); 685 cv_destroy(&fcip_global_cv); 686 rw_destroy(&fcipstruplock); 687 ddi_soft_state_fini(&fcip_softp); 688 } 689 return (rval); 690 } 691 692 /* 693 * Unload the port driver if this was the only ULP loaded and then 694 * deregister with the transport. 695 */ 696 int 697 _fini(void) 698 { 699 int rval; 700 int rval1; 701 702 /* 703 * Do not permit the module to be unloaded before a port 704 * attach callback has happened. 705 */ 706 mutex_enter(&fcip_global_mutex); 707 if (fcip_num_attaching || fcip_port_attach_pending) { 708 mutex_exit(&fcip_global_mutex); 709 return (EBUSY); 710 } 711 mutex_exit(&fcip_global_mutex); 712 713 if ((rval = mod_remove(&modlinkage)) != 0) { 714 return (rval); 715 } 716 717 /* 718 * unregister with the transport layer 719 */ 720 rval1 = fc_ulp_remove(&fcip_modinfo); 721 722 /* 723 * If the ULP was not registered with the transport, init should 724 * have failed. If transport has no knowledge of our existence 725 * we should simply bail out and succeed 726 */ 727 #ifdef DEBUG 728 if (rval1 == FC_BADULP) { 729 FCIP_DEBUG(FCIP_DEBUG_DEFAULT, (CE_WARN, 730 "fcip: ULP was never registered with the transport")); 731 rval = ENODEV; 732 } else if (rval1 == FC_BADTYPE) { 733 FCIP_DEBUG(FCIP_DEBUG_DEFAULT, (CE_WARN, 734 "fcip: No ULP of this type 0x%x was registered with " 735 "transport", fcip_modinfo.ulp_type)); 736 rval = ENODEV; 737 } 738 #endif /* DEBUG */ 739 740 mutex_destroy(&fcip_global_mutex); 741 rw_destroy(&fcipstruplock); 742 cv_destroy(&fcip_global_cv); 743 ddi_soft_state_fini(&fcip_softp); 744 745 FCIP_TNF_UNLOAD(&modlinkage); 746 747 return (rval); 748 } 749 750 /* 751 * Info about this loadable module 752 */ 753 int 754 _info(struct modinfo *modinfop) 755 { 756 return (mod_info(&modlinkage, modinfop)); 757 } 758 759 /* 760 * The port attach callback is invoked by the port driver when a FCA 761 * port comes online and binds with the transport layer. The transport 762 * then callsback into all ULP modules registered with it. The Port attach 763 * call back will also provide the ULP module with the Port's WWN and S_ID 764 */ 765 /* ARGSUSED */ 766 static int 767 fcip_port_attach(opaque_t ulp_handle, fc_ulp_port_info_t *port_info, 768 fc_attach_cmd_t cmd, uint32_t sid) 769 { 770 int rval = FC_FAILURE; 771 int instance; 772 struct fcip *fptr; 773 fcip_port_info_t *fport = NULL; 774 fcip_port_info_t *cur_fport; 775 fc_portid_t src_id; 776 777 switch (cmd) { 778 case FC_CMD_ATTACH: { 779 la_wwn_t *ww_pn = NULL; 780 /* 781 * It was determined that, as per spec, the lower 48 bits of 782 * the port-WWN will always be unique. This will make the MAC 783 * address (i.e the lower 48 bits of the WWN), that IP/ARP 784 * depend on, unique too. Hence we should be able to remove the 785 * restriction of attaching to only one of the ports of 786 * multi port FCAs. 787 * 788 * Earlier, fcip used to attach only to qlc module and fail 789 * silently for attach failures resulting from unknown FCAs or 790 * unsupported FCA ports. Now, we'll do no such checks. 791 */ 792 ww_pn = &port_info->port_pwwn; 793 794 FCIP_TNF_PROBE_2((fcip_port_attach, "fcip io", /* CSTYLED */, 795 tnf_string, msg, "port id bits", 796 tnf_opaque, nport_id, ww_pn->w.nport_id)); 797 FCIP_DEBUG(FCIP_DEBUG_ATTACH, (CE_NOTE, 798 "port id bits: 0x%x", ww_pn->w.nport_id)); 799 /* 800 * A port has come online 801 */ 802 mutex_enter(&fcip_global_mutex); 803 fcip_num_instances++; 804 fcip_num_attaching++; 805 806 if (fcip_port_head == NULL) { 807 /* OK to sleep here ? */ 808 fport = kmem_zalloc(sizeof (fcip_port_info_t), 809 KM_NOSLEEP); 810 if (fport == NULL) { 811 fcip_num_instances--; 812 fcip_num_attaching--; 813 ASSERT(fcip_num_attaching >= 0); 814 mutex_exit(&fcip_global_mutex); 815 rval = FC_FAILURE; 816 cmn_err(CE_WARN, "!fcip(%d): port attach " 817 "failed: alloc failed", 818 ddi_get_instance(port_info->port_dip)); 819 goto done; 820 } 821 fcip_port_head = fport; 822 } else { 823 /* 824 * traverse the port list and also check for 825 * duplicate port attaches - Nothing wrong in being 826 * paranoid Heh Heh. 827 */ 828 cur_fport = fcip_port_head; 829 while (cur_fport != NULL) { 830 if (cur_fport->fcipp_handle == 831 port_info->port_handle) { 832 fcip_num_instances--; 833 fcip_num_attaching--; 834 ASSERT(fcip_num_attaching >= 0); 835 mutex_exit(&fcip_global_mutex); 836 FCIP_DEBUG(FCIP_DEBUG_ATTACH, (CE_WARN, 837 "!fcip(%d): port already " 838 "attached!!", ddi_get_instance( 839 port_info->port_dip))); 840 rval = FC_FAILURE; 841 goto done; 842 } 843 cur_fport = cur_fport->fcipp_next; 844 } 845 fport = kmem_zalloc(sizeof (fcip_port_info_t), 846 KM_NOSLEEP); 847 if (fport == NULL) { 848 rval = FC_FAILURE; 849 fcip_num_instances--; 850 fcip_num_attaching--; 851 ASSERT(fcip_num_attaching >= 0); 852 mutex_exit(&fcip_global_mutex); 853 cmn_err(CE_WARN, "!fcip(%d): port attach " 854 "failed: alloc failed", 855 ddi_get_instance(port_info->port_dip)); 856 goto done; 857 } 858 fport->fcipp_next = fcip_port_head; 859 fcip_port_head = fport; 860 } 861 862 mutex_exit(&fcip_global_mutex); 863 864 /* 865 * now fill in the details about the port itself 866 */ 867 fport->fcipp_linkage = *port_info->port_linkage; 868 fport->fcipp_handle = port_info->port_handle; 869 fport->fcipp_dip = port_info->port_dip; 870 fport->fcipp_topology = port_info->port_flags; 871 fport->fcipp_pstate = port_info->port_state; 872 fport->fcipp_naa = port_info->port_pwwn.w.naa_id; 873 bcopy(&port_info->port_pwwn, &fport->fcipp_pwwn, 874 sizeof (la_wwn_t)); 875 bcopy(&port_info->port_nwwn, &fport->fcipp_nwwn, 876 sizeof (la_wwn_t)); 877 fport->fcipp_fca_pkt_size = port_info->port_fca_pkt_size; 878 fport->fcipp_cmd_dma_attr = *port_info->port_cmd_dma_attr; 879 fport->fcipp_resp_dma_attr = *port_info->port_resp_dma_attr; 880 fport->fcipp_fca_acc_attr = *port_info->port_acc_attr; 881 src_id.port_id = sid; 882 src_id.priv_lilp_posit = 0; 883 fport->fcipp_sid = src_id; 884 885 /* 886 * allocate soft state for this instance 887 */ 888 instance = ddi_get_instance(fport->fcipp_dip); 889 if (ddi_soft_state_zalloc(fcip_softp, 890 instance) != DDI_SUCCESS) { 891 rval = FC_FAILURE; 892 cmn_err(CE_WARN, "!fcip(%d): port attach failed: " 893 "soft state alloc failed", instance); 894 goto failure; 895 } 896 897 fptr = ddi_get_soft_state(fcip_softp, instance); 898 899 if (fptr == NULL) { 900 rval = FC_FAILURE; 901 cmn_err(CE_WARN, "!fcip(%d): port attach failed: " 902 "failure to get soft state", instance); 903 goto failure; 904 } 905 906 /* 907 * initialize all mutexes and locks required for this module 908 */ 909 mutex_init(&fptr->fcip_mutex, NULL, MUTEX_DRIVER, NULL); 910 mutex_init(&fptr->fcip_ub_mutex, NULL, MUTEX_DRIVER, NULL); 911 mutex_init(&fptr->fcip_rt_mutex, NULL, MUTEX_DRIVER, NULL); 912 mutex_init(&fptr->fcip_dest_mutex, NULL, MUTEX_DRIVER, NULL); 913 mutex_init(&fptr->fcip_sendup_mutex, NULL, MUTEX_DRIVER, NULL); 914 cv_init(&fptr->fcip_farp_cv, NULL, CV_DRIVER, NULL); 915 cv_init(&fptr->fcip_sendup_cv, NULL, CV_DRIVER, NULL); 916 cv_init(&fptr->fcip_ub_cv, NULL, CV_DRIVER, NULL); 917 918 mutex_enter(&fptr->fcip_mutex); 919 920 fptr->fcip_dip = fport->fcipp_dip; /* parent's dip */ 921 fptr->fcip_instance = instance; 922 fptr->fcip_ub_upstream = 0; 923 924 if (FC_PORT_STATE_MASK(port_info->port_state) == 925 FC_STATE_ONLINE) { 926 fptr->fcip_port_state = FCIP_PORT_ONLINE; 927 if (fptr->fcip_flags & FCIP_LINK_DOWN) { 928 fptr->fcip_flags &= ~FCIP_LINK_DOWN; 929 } 930 } else { 931 fptr->fcip_port_state = FCIP_PORT_OFFLINE; 932 } 933 934 fptr->fcip_flags |= FCIP_ATTACHING; 935 fptr->fcip_port_info = fport; 936 937 /* 938 * Extract our MAC addr from our port's WWN. The lower 48 939 * bits will be our MAC address 940 */ 941 wwn_to_ether(&fport->fcipp_nwwn, &fptr->fcip_macaddr); 942 943 fport->fcipp_fcip = fptr; 944 945 FCIP_DEBUG(FCIP_DEBUG_ATTACH, 946 (CE_NOTE, "fcipdest : 0x%lx, rtable : 0x%lx", 947 (long)(sizeof (fptr->fcip_dest)), 948 (long)(sizeof (fptr->fcip_rtable)))); 949 950 bzero(fptr->fcip_dest, sizeof (fptr->fcip_dest)); 951 bzero(fptr->fcip_rtable, sizeof (fptr->fcip_rtable)); 952 953 /* 954 * create a taskq to handle sundry jobs for the driver 955 * This way we can have jobs run in parallel 956 */ 957 fptr->fcip_tq = taskq_create("fcip_tasks", 958 FCIP_NUM_THREADS, MINCLSYSPRI, FCIP_MIN_TASKS, 959 FCIP_MAX_TASKS, TASKQ_PREPOPULATE); 960 961 mutex_exit(&fptr->fcip_mutex); 962 963 /* 964 * create a separate thread to handle all unsolicited 965 * callback handling. This is because unsolicited_callback 966 * can happen from an interrupt context and the upstream 967 * modules can put new messages right back in the same 968 * thread context. This usually works fine, but sometimes 969 * we may have to block to obtain the dest struct entries 970 * for some remote ports. 971 */ 972 mutex_enter(&fptr->fcip_sendup_mutex); 973 if (thread_create(NULL, DEFAULTSTKSZ, 974 (void (*)())fcip_sendup_thr, (caddr_t)fptr, 0, &p0, 975 TS_RUN, minclsyspri) == NULL) { 976 mutex_exit(&fptr->fcip_sendup_mutex); 977 cmn_err(CE_WARN, 978 "!unable to create fcip sendup thread for " 979 " instance: 0x%x", instance); 980 rval = FC_FAILURE; 981 goto done; 982 } 983 fptr->fcip_sendup_thr_initted = 1; 984 fptr->fcip_sendup_head = fptr->fcip_sendup_tail = NULL; 985 mutex_exit(&fptr->fcip_sendup_mutex); 986 987 988 /* Let the attach handler do the rest */ 989 if (fcip_port_attach_handler(fptr) != FC_SUCCESS) { 990 /* 991 * We have already cleaned up so return 992 */ 993 rval = FC_FAILURE; 994 cmn_err(CE_WARN, "!fcip(%d): port attach failed", 995 instance); 996 goto done; 997 } 998 999 FCIP_DEBUG(FCIP_DEBUG_ATTACH, (CE_CONT, 1000 "!fcip attach for port instance (0x%x) successful", 1001 instance)); 1002 1003 rval = FC_SUCCESS; 1004 goto done; 1005 } 1006 case FC_CMD_POWER_UP: 1007 /* FALLTHROUGH */ 1008 case FC_CMD_RESUME: 1009 mutex_enter(&fcip_global_mutex); 1010 fport = fcip_port_head; 1011 while (fport != NULL) { 1012 if (fport->fcipp_handle == port_info->port_handle) { 1013 break; 1014 } 1015 fport = fport->fcipp_next; 1016 } 1017 if (fport == NULL) { 1018 rval = FC_SUCCESS; 1019 mutex_exit(&fcip_global_mutex); 1020 goto done; 1021 } 1022 rval = fcip_handle_resume(fport, port_info, cmd); 1023 mutex_exit(&fcip_global_mutex); 1024 goto done; 1025 1026 default: 1027 FCIP_TNF_PROBE_2((fcip_port_attach, "fcip io", /* CSTYLED */, 1028 tnf_string, msg, "unknown command type", 1029 tnf_uint, cmd, cmd)); 1030 FCIP_DEBUG(FCIP_DEBUG_ATTACH, (CE_WARN, 1031 "unknown cmd type 0x%x in port_attach", cmd)); 1032 rval = FC_FAILURE; 1033 goto done; 1034 } 1035 1036 failure: 1037 if (fport) { 1038 mutex_enter(&fcip_global_mutex); 1039 fcip_num_attaching--; 1040 ASSERT(fcip_num_attaching >= 0); 1041 (void) fcip_softstate_free(fport); 1042 fcip_port_attach_pending = 0; 1043 mutex_exit(&fcip_global_mutex); 1044 } 1045 return (rval); 1046 1047 done: 1048 mutex_enter(&fcip_global_mutex); 1049 fcip_port_attach_pending = 0; 1050 mutex_exit(&fcip_global_mutex); 1051 return (rval); 1052 } 1053 1054 /* 1055 * fcip_port_attach_handler : Completes the port attach operation after 1056 * the ulp_port_attach routine has completed its ground work. The job 1057 * of this function among other things is to obtain and handle topology 1058 * specifics, initialize a port, setup broadcast address entries in 1059 * the fcip tables etc. This routine cleans up behind itself on failures. 1060 * Returns FC_SUCCESS or FC_FAILURE. 1061 */ 1062 static int 1063 fcip_port_attach_handler(struct fcip *fptr) 1064 { 1065 fcip_port_info_t *fport = fptr->fcip_port_info; 1066 int rval = FC_FAILURE; 1067 1068 ASSERT(fport != NULL); 1069 1070 mutex_enter(&fcip_global_mutex); 1071 1072 FCIP_DEBUG(FCIP_DEBUG_ATTACH, (CE_NOTE, 1073 "fcip module dip: %p instance: %d", 1074 (void *)fcip_module_dip, ddi_get_instance(fptr->fcip_dip))); 1075 1076 if (fcip_module_dip == NULL) { 1077 clock_t fcip_lbolt; 1078 1079 fcip_lbolt = ddi_get_lbolt(); 1080 /* 1081 * we need to use the fcip devinfo for creating 1082 * the clone device node, but the fcip attach 1083 * (from its conf file entry claiming to be a 1084 * child of pseudo) may not have happened yet. 1085 * wait here for 10 seconds and fail port attach 1086 * if the fcip devinfo is not attached yet 1087 */ 1088 fcip_lbolt += drv_usectohz(FCIP_INIT_DELAY); 1089 1090 FCIP_DEBUG(FCIP_DEBUG_ATTACH, 1091 (CE_WARN, "cv_timedwait lbolt %lx", fcip_lbolt)); 1092 1093 (void) cv_timedwait(&fcip_global_cv, &fcip_global_mutex, 1094 fcip_lbolt); 1095 1096 if (fcip_module_dip == NULL) { 1097 mutex_exit(&fcip_global_mutex); 1098 1099 FCIP_DEBUG(FCIP_DEBUG_ATTACH, (CE_WARN, 1100 "fcip attach did not happen")); 1101 goto port_attach_cleanup; 1102 } 1103 } 1104 1105 if ((!fcip_minor_node_created) && 1106 fcip_is_supported_fc_topology(fport->fcipp_topology)) { 1107 /* 1108 * Checking for same topologies which are considered valid 1109 * by fcip_handle_topology(). Dont create a minor node if 1110 * nothing is hanging off the FC port. 1111 */ 1112 if (ddi_create_minor_node(fcip_module_dip, "fcip", S_IFCHR, 1113 ddi_get_instance(fptr->fcip_dip), DDI_PSEUDO, 1114 CLONE_DEV) == DDI_FAILURE) { 1115 mutex_exit(&fcip_global_mutex); 1116 FCIP_DEBUG(FCIP_DEBUG_ATTACH, (CE_WARN, 1117 "failed to create minor node for fcip(%d)", 1118 ddi_get_instance(fptr->fcip_dip))); 1119 goto port_attach_cleanup; 1120 } 1121 fcip_minor_node_created++; 1122 } 1123 mutex_exit(&fcip_global_mutex); 1124 1125 /* 1126 * initialize port for traffic 1127 */ 1128 if (fcip_init_port(fptr) != FC_SUCCESS) { 1129 /* fcip_init_port has already cleaned up its stuff */ 1130 1131 mutex_enter(&fcip_global_mutex); 1132 1133 if ((fcip_num_instances == 1) && 1134 (fcip_minor_node_created == 1)) { 1135 /* Remove minor node iff this is the last instance */ 1136 ddi_remove_minor_node(fcip_module_dip, NULL); 1137 } 1138 1139 mutex_exit(&fcip_global_mutex); 1140 1141 goto port_attach_cleanup; 1142 } 1143 1144 mutex_enter(&fptr->fcip_mutex); 1145 fptr->fcip_flags &= ~FCIP_ATTACHING; 1146 fptr->fcip_flags |= FCIP_INITED; 1147 fptr->fcip_timeout_ticks = 0; 1148 1149 /* 1150 * start the timeout threads 1151 */ 1152 fptr->fcip_timeout_id = timeout(fcip_timeout, fptr, 1153 drv_usectohz(1000000)); 1154 1155 mutex_exit(&fptr->fcip_mutex); 1156 mutex_enter(&fcip_global_mutex); 1157 fcip_num_attaching--; 1158 ASSERT(fcip_num_attaching >= 0); 1159 mutex_exit(&fcip_global_mutex); 1160 rval = FC_SUCCESS; 1161 return (rval); 1162 1163 port_attach_cleanup: 1164 mutex_enter(&fcip_global_mutex); 1165 (void) fcip_softstate_free(fport); 1166 fcip_num_attaching--; 1167 ASSERT(fcip_num_attaching >= 0); 1168 mutex_exit(&fcip_global_mutex); 1169 rval = FC_FAILURE; 1170 return (rval); 1171 } 1172 1173 1174 /* 1175 * Handler for DDI_RESUME operations. Port must be ready to restart IP 1176 * traffic on resume 1177 */ 1178 static int 1179 fcip_handle_resume(fcip_port_info_t *fport, fc_ulp_port_info_t *port_info, 1180 fc_attach_cmd_t cmd) 1181 { 1182 int rval = FC_SUCCESS; 1183 struct fcip *fptr = fport->fcipp_fcip; 1184 struct fcipstr *tslp; 1185 int index; 1186 1187 1188 ASSERT(fptr != NULL); 1189 1190 mutex_enter(&fptr->fcip_mutex); 1191 1192 if (cmd == FC_CMD_POWER_UP) { 1193 fptr->fcip_flags &= ~(FCIP_POWER_DOWN); 1194 if (fptr->fcip_flags & FCIP_SUSPENDED) { 1195 mutex_exit(&fptr->fcip_mutex); 1196 return (FC_SUCCESS); 1197 } 1198 } else if (cmd == FC_CMD_RESUME) { 1199 fptr->fcip_flags &= ~(FCIP_SUSPENDED); 1200 } else { 1201 mutex_exit(&fptr->fcip_mutex); 1202 return (FC_FAILURE); 1203 } 1204 1205 /* 1206 * set the current port state and topology 1207 */ 1208 fport->fcipp_topology = port_info->port_flags; 1209 fport->fcipp_pstate = port_info->port_state; 1210 1211 rw_enter(&fcipstruplock, RW_READER); 1212 for (tslp = fcipstrup; tslp; tslp = tslp->sl_nextp) { 1213 if (tslp->sl_fcip == fptr) { 1214 break; 1215 } 1216 } 1217 rw_exit(&fcipstruplock); 1218 1219 /* 1220 * No active streams on this port 1221 */ 1222 if (tslp == NULL) { 1223 rval = FC_SUCCESS; 1224 goto done; 1225 } 1226 1227 mutex_enter(&fptr->fcip_rt_mutex); 1228 for (index = 0; index < FCIP_RT_HASH_ELEMS; index++) { 1229 struct fcip_routing_table *frp; 1230 1231 frp = fptr->fcip_rtable[index]; 1232 while (frp) { 1233 uint32_t did; 1234 /* 1235 * Mark the broadcast RTE available again. It 1236 * was marked SUSPENDED during SUSPEND. 1237 */ 1238 did = fcip_get_broadcast_did(fptr); 1239 if (frp->fcipr_d_id.port_id == did) { 1240 frp->fcipr_state = 0; 1241 index = FCIP_RT_HASH_ELEMS; 1242 break; 1243 } 1244 frp = frp->fcipr_next; 1245 } 1246 } 1247 mutex_exit(&fptr->fcip_rt_mutex); 1248 1249 /* 1250 * fcip_handle_topology will update the port entries in the 1251 * routing table. 1252 * fcip_handle_topology also takes care of resetting the 1253 * fcipr_state field in the routing table structure. The entries 1254 * were set to RT_INVALID during suspend. 1255 */ 1256 fcip_handle_topology(fptr); 1257 1258 done: 1259 /* 1260 * Restart the timeout thread 1261 */ 1262 fptr->fcip_timeout_id = timeout(fcip_timeout, fptr, 1263 drv_usectohz(1000000)); 1264 mutex_exit(&fptr->fcip_mutex); 1265 return (rval); 1266 } 1267 1268 1269 /* 1270 * Insert a destination port entry into the routing table for 1271 * this port 1272 */ 1273 static void 1274 fcip_rt_update(struct fcip *fptr, fc_portmap_t *devlist, uint32_t listlen) 1275 { 1276 struct fcip_routing_table *frp; 1277 fcip_port_info_t *fport = fptr->fcip_port_info; 1278 int hash_bucket, i; 1279 fc_portmap_t *pmap; 1280 char wwn_buf[20]; 1281 1282 FCIP_TNF_PROBE_2((fcip_rt_update, "fcip io", /* CSTYLED */, 1283 tnf_string, msg, "enter", 1284 tnf_int, listlen, listlen)); 1285 1286 ASSERT(!mutex_owned(&fptr->fcip_mutex)); 1287 mutex_enter(&fptr->fcip_rt_mutex); 1288 1289 for (i = 0; i < listlen; i++) { 1290 pmap = &(devlist[i]); 1291 1292 frp = fcip_lookup_rtable(fptr, &(pmap->map_pwwn), 1293 FCIP_COMPARE_PWWN); 1294 /* 1295 * If an entry for a port in the devlist exists in the 1296 * in the per port routing table, make sure the data 1297 * is current. We need to do this irrespective of the 1298 * underlying port topology. 1299 */ 1300 switch (pmap->map_type) { 1301 /* FALLTHROUGH */ 1302 case PORT_DEVICE_NOCHANGE: 1303 /* FALLTHROUGH */ 1304 case PORT_DEVICE_USER_LOGIN: 1305 /* FALLTHROUGH */ 1306 case PORT_DEVICE_CHANGED: 1307 /* FALLTHROUGH */ 1308 case PORT_DEVICE_NEW: 1309 if (frp == NULL) { 1310 goto add_new_entry; 1311 } else if (frp) { 1312 goto update_entry; 1313 } else { 1314 continue; 1315 } 1316 1317 case PORT_DEVICE_OLD: 1318 /* FALLTHROUGH */ 1319 case PORT_DEVICE_USER_LOGOUT: 1320 /* 1321 * Mark entry for removal from Routing Table if 1322 * one exists. Let the timeout thread actually 1323 * remove the entry after we've given up hopes 1324 * of the port ever showing up. 1325 */ 1326 if (frp) { 1327 uint32_t did; 1328 1329 /* 1330 * Mark the routing table as invalid to bail 1331 * the packets early that are in transit 1332 */ 1333 did = fptr->fcip_broadcast_did; 1334 if (frp->fcipr_d_id.port_id != did) { 1335 frp->fcipr_pd = NULL; 1336 frp->fcipr_state = FCIP_RT_INVALID; 1337 frp->fcipr_invalid_timeout = 1338 fptr->fcip_timeout_ticks + 1339 FCIP_RTE_TIMEOUT; 1340 } 1341 } 1342 continue; 1343 1344 default: 1345 FCIP_DEBUG(FCIP_DEBUG_INIT, (CE_WARN, 1346 "unknown map flags in rt_update")); 1347 continue; 1348 } 1349 add_new_entry: 1350 ASSERT(frp == NULL); 1351 hash_bucket = FCIP_RT_HASH(pmap->map_pwwn.raw_wwn); 1352 1353 ASSERT(hash_bucket < FCIP_RT_HASH_ELEMS); 1354 1355 FCIP_TNF_PROBE_2((fcip_rt_update, "cfip io", /* CSTYLED */, 1356 tnf_string, msg, 1357 "add new entry", 1358 tnf_int, hashbucket, hash_bucket)); 1359 1360 frp = (struct fcip_routing_table *) 1361 kmem_zalloc(sizeof (struct fcip_routing_table), KM_SLEEP); 1362 /* insert at beginning of hash bucket */ 1363 frp->fcipr_next = fptr->fcip_rtable[hash_bucket]; 1364 fptr->fcip_rtable[hash_bucket] = frp; 1365 fc_wwn_to_str(&pmap->map_pwwn, wwn_buf); 1366 FCIP_DEBUG(FCIP_DEBUG_ATTACH, (CE_NOTE, 1367 "added entry for pwwn %s and d_id 0x%x", 1368 wwn_buf, pmap->map_did.port_id)); 1369 update_entry: 1370 bcopy((void *)&pmap->map_pwwn, 1371 (void *)&frp->fcipr_pwwn, sizeof (la_wwn_t)); 1372 bcopy((void *)&pmap->map_nwwn, (void *)&frp->fcipr_nwwn, 1373 sizeof (la_wwn_t)); 1374 frp->fcipr_d_id = pmap->map_did; 1375 frp->fcipr_state = pmap->map_state; 1376 frp->fcipr_pd = pmap->map_pd; 1377 1378 /* 1379 * If there is no pd for a destination port that is not 1380 * a broadcast entry, the port is pretty much unusable - so 1381 * mark the port for removal so we can try adding back the 1382 * entry again. 1383 */ 1384 if ((frp->fcipr_pd == NULL) && 1385 (frp->fcipr_d_id.port_id != fptr->fcip_broadcast_did)) { 1386 frp->fcipr_state = PORT_DEVICE_INVALID; 1387 frp->fcipr_invalid_timeout = fptr->fcip_timeout_ticks + 1388 (FCIP_RTE_TIMEOUT / 2); 1389 } 1390 frp->fcipr_fca_dev = 1391 fc_ulp_get_fca_device(fport->fcipp_handle, pmap->map_did); 1392 1393 /* 1394 * login to the remote port. Don't worry about 1395 * plogi failures for now 1396 */ 1397 if (pmap->map_pd != NULL) { 1398 (void) fcip_do_plogi(fptr, frp); 1399 } else if (FC_TOP_EXTERNAL(fport->fcipp_topology)) { 1400 fc_wwn_to_str(&frp->fcipr_pwwn, wwn_buf); 1401 FCIP_DEBUG(FCIP_DEBUG_MISC, (CE_NOTE, 1402 "logging into pwwn %s, d_id 0x%x", 1403 wwn_buf, frp->fcipr_d_id.port_id)); 1404 (void) fcip_do_plogi(fptr, frp); 1405 } 1406 1407 FCIP_TNF_BYTE_ARRAY(fcip_rt_update, "fcip io", "detail", 1408 "new wwn in rt", pwwn, 1409 &frp->fcipr_pwwn, sizeof (la_wwn_t)); 1410 } 1411 mutex_exit(&fptr->fcip_rt_mutex); 1412 } 1413 1414 1415 /* 1416 * return a matching routing table entry for a given fcip instance 1417 */ 1418 struct fcip_routing_table * 1419 fcip_lookup_rtable(struct fcip *fptr, la_wwn_t *wwn, int matchflag) 1420 { 1421 struct fcip_routing_table *frp = NULL; 1422 int hash_bucket; 1423 1424 1425 FCIP_TNF_PROBE_1((fcip_lookup_rtable, "fcip io", /* CSTYLED */, 1426 tnf_string, msg, "enter")); 1427 FCIP_TNF_BYTE_ARRAY(fcip_lookup_rtable, "fcip io", "detail", 1428 "rtable lookup for", wwn, 1429 &wwn->raw_wwn, sizeof (la_wwn_t)); 1430 FCIP_TNF_PROBE_2((fcip_lookup_rtable, "fcip io", /* CSTYLED */, 1431 tnf_string, msg, "match by", 1432 tnf_int, matchflag, matchflag)); 1433 1434 ASSERT(mutex_owned(&fptr->fcip_rt_mutex)); 1435 1436 hash_bucket = FCIP_RT_HASH(wwn->raw_wwn); 1437 frp = fptr->fcip_rtable[hash_bucket]; 1438 while (frp != NULL) { 1439 1440 FCIP_TNF_BYTE_ARRAY(fcip_lookup_rtable, "fcip io", "detail", 1441 "rtable entry", nwwn, 1442 &(frp->fcipr_nwwn.raw_wwn), sizeof (la_wwn_t)); 1443 1444 if (fcip_wwn_compare(&frp->fcipr_pwwn, wwn, matchflag) == 0) { 1445 break; 1446 } 1447 1448 frp = frp->fcipr_next; 1449 } 1450 FCIP_TNF_PROBE_2((fcip_lookup_rtable, "fcip io", /* CSTYLED */, 1451 tnf_string, msg, "lookup result", 1452 tnf_opaque, frp, frp)); 1453 return (frp); 1454 } 1455 1456 /* 1457 * Attach of fcip under pseudo. The actual setup of the interface 1458 * actually happens in fcip_port_attach on a callback from the 1459 * transport. The port_attach callback however can proceed only 1460 * after the devinfo for fcip has been created under pseudo 1461 */ 1462 static int 1463 fcip_attach(dev_info_t *dip, ddi_attach_cmd_t cmd) 1464 { 1465 switch ((int)cmd) { 1466 1467 case DDI_ATTACH: { 1468 ASSERT(fcip_module_dip == NULL); 1469 fcip_module_dip = dip; 1470 1471 /* 1472 * this call originates as a result of fcip's conf 1473 * file entry and will result in a fcip instance being 1474 * a child of pseudo. We should ensure here that the port 1475 * driver (fp) has been loaded and initted since we would 1476 * never get a port attach callback without fp being loaded. 1477 * If we are unable to succesfully load and initalize fp - 1478 * just fail this attach. 1479 */ 1480 mutex_enter(&fcip_global_mutex); 1481 1482 FCIP_DEBUG(FCIP_DEBUG_ATTACH, 1483 (CE_WARN, "global cv - signaling")); 1484 1485 cv_signal(&fcip_global_cv); 1486 1487 FCIP_DEBUG(FCIP_DEBUG_ATTACH, 1488 (CE_WARN, "global cv - signaled")); 1489 mutex_exit(&fcip_global_mutex); 1490 return (DDI_SUCCESS); 1491 } 1492 case DDI_RESUME: 1493 /* 1494 * Resume appears trickier 1495 */ 1496 return (DDI_SUCCESS); 1497 default: 1498 return (DDI_FAILURE); 1499 } 1500 } 1501 1502 1503 /* 1504 * The detach entry point to permit unloading fcip. We make sure 1505 * there are no active streams before we proceed with the detach 1506 */ 1507 /* ARGSUSED */ 1508 static int 1509 fcip_detach(dev_info_t *dip, ddi_detach_cmd_t cmd) 1510 { 1511 struct fcip *fptr; 1512 fcip_port_info_t *fport; 1513 int detached; 1514 1515 switch (cmd) { 1516 case DDI_DETACH: { 1517 /* 1518 * If we got here, any active streams should have been 1519 * unplumbed but check anyway 1520 */ 1521 mutex_enter(&fcip_global_mutex); 1522 if (fcipstrup != NULL) { 1523 mutex_exit(&fcip_global_mutex); 1524 return (DDI_FAILURE); 1525 } 1526 1527 if (fcip_port_head != NULL) { 1528 /* 1529 * Check to see if we have unattached/unbound 1530 * ports. If all the ports are unattached/unbound go 1531 * ahead and unregister with the transport 1532 */ 1533 fport = fcip_port_head; 1534 while (fport != NULL) { 1535 fptr = fport->fcipp_fcip; 1536 if (fptr == NULL) { 1537 continue; 1538 } 1539 mutex_enter(&fptr->fcip_mutex); 1540 fptr->fcip_flags |= FCIP_DETACHING; 1541 if (fptr->fcip_ipq || 1542 fptr->fcip_flags & (FCIP_IN_TIMEOUT | 1543 FCIP_IN_CALLBACK | FCIP_ATTACHING | 1544 FCIP_SUSPENDED | FCIP_POWER_DOWN | 1545 FCIP_REG_INPROGRESS)) { 1546 FCIP_TNF_PROBE_1((fcip_detach, 1547 "fcip io", /* CSTYLED */, 1548 tnf_string, msg, 1549 "fcip instance busy")); 1550 1551 mutex_exit(&fptr->fcip_mutex); 1552 FCIP_DEBUG(FCIP_DEBUG_DETACH, (CE_WARN, 1553 "fcip instance busy")); 1554 break; 1555 } 1556 /* 1557 * Check for any outstanding pkts. If yes 1558 * fail the detach 1559 */ 1560 mutex_enter(&fptr->fcip_dest_mutex); 1561 if (fcip_port_get_num_pkts(fptr) > 0) { 1562 mutex_exit(&fptr->fcip_dest_mutex); 1563 mutex_exit(&fptr->fcip_mutex); 1564 FCIP_DEBUG(FCIP_DEBUG_DETACH, (CE_WARN, 1565 "fcip instance busy - pkts " 1566 "pending")); 1567 break; 1568 } 1569 mutex_exit(&fptr->fcip_dest_mutex); 1570 1571 mutex_enter(&fptr->fcip_rt_mutex); 1572 if (fcip_plogi_in_progress(fptr)) { 1573 mutex_exit(&fptr->fcip_rt_mutex); 1574 mutex_exit(&fptr->fcip_mutex); 1575 FCIP_DEBUG(FCIP_DEBUG_DETACH, (CE_WARN, 1576 "fcip instance busy - plogi in " 1577 "progress")); 1578 break; 1579 } 1580 mutex_exit(&fptr->fcip_rt_mutex); 1581 1582 mutex_exit(&fptr->fcip_mutex); 1583 fport = fport->fcipp_next; 1584 } 1585 /* 1586 * if fport is non NULL - we have active ports 1587 */ 1588 if (fport != NULL) { 1589 /* 1590 * Remove the DETACHING flags on the ports 1591 */ 1592 fport = fcip_port_head; 1593 while (fport != NULL) { 1594 fptr = fport->fcipp_fcip; 1595 mutex_enter(&fptr->fcip_mutex); 1596 fptr->fcip_flags &= ~(FCIP_DETACHING); 1597 mutex_exit(&fptr->fcip_mutex); 1598 fport = fport->fcipp_next; 1599 } 1600 mutex_exit(&fcip_global_mutex); 1601 return (DDI_FAILURE); 1602 } 1603 } 1604 1605 /* 1606 * free up all softstate structures 1607 */ 1608 fport = fcip_port_head; 1609 while (fport != NULL) { 1610 detached = 1; 1611 1612 fptr = fport->fcipp_fcip; 1613 if (fptr) { 1614 mutex_enter(&fptr->fcip_mutex); 1615 /* 1616 * Check to see if somebody beat us to the 1617 * punch 1618 */ 1619 detached = fptr->fcip_flags & FCIP_DETACHED; 1620 fptr->fcip_flags &= ~(FCIP_DETACHING); 1621 fptr->fcip_flags |= FCIP_DETACHED; 1622 mutex_exit(&fptr->fcip_mutex); 1623 } 1624 1625 if (!detached) { 1626 fport = fcip_softstate_free(fport); 1627 } else { 1628 /* 1629 * If the port was marked as detached 1630 * but it was still in the list, that 1631 * means another thread has marked it 1632 * but we got in while it released the 1633 * fcip_global_mutex in softstate_free. 1634 * Given that, we're still safe to use 1635 * fport->fcipp_next to find out what 1636 * the next port on the list is. 1637 */ 1638 fport = fport->fcipp_next; 1639 } 1640 1641 FCIP_DEBUG(FCIP_DEBUG_DETACH, 1642 (CE_NOTE, "detaching port")); 1643 1644 FCIP_TNF_PROBE_1((fcip_detach, 1645 "fcip io", /* CSTYLED */, tnf_string, 1646 msg, "detaching port")); 1647 } 1648 1649 /* 1650 * If we haven't removed all the port structures, we 1651 * aren't yet ready to be detached. 1652 */ 1653 if (fcip_port_head != NULL) { 1654 mutex_exit(&fcip_global_mutex); 1655 return (DDI_FAILURE); 1656 } 1657 1658 fcip_num_instances = 0; 1659 mutex_exit(&fcip_global_mutex); 1660 fcip_module_dip = NULL; 1661 return (DDI_SUCCESS); 1662 } 1663 case DDI_SUSPEND: 1664 return (DDI_SUCCESS); 1665 default: 1666 return (DDI_FAILURE); 1667 } 1668 } 1669 1670 /* 1671 * The port_detach callback is called from the transport when a 1672 * FC port is being removed from the transport's control. This routine 1673 * provides fcip with an opportunity to cleanup all activities and 1674 * structures on the port marked for removal. 1675 */ 1676 /* ARGSUSED */ 1677 static int 1678 fcip_port_detach(opaque_t ulp_handle, fc_ulp_port_info_t *port_info, 1679 fc_detach_cmd_t cmd) 1680 { 1681 int rval = FC_FAILURE; 1682 fcip_port_info_t *fport; 1683 struct fcip *fptr; 1684 struct fcipstr *strp; 1685 1686 switch (cmd) { 1687 case FC_CMD_DETACH: { 1688 mutex_enter(&fcip_global_mutex); 1689 1690 if (fcip_port_head == NULL) { 1691 /* 1692 * we are all done but our fini has not been 1693 * called yet!! Let's hope we have no active 1694 * fcip instances here. - strange secnario but 1695 * no harm in having this return a success. 1696 */ 1697 fcip_check_remove_minor_node(); 1698 1699 mutex_exit(&fcip_global_mutex); 1700 return (FC_SUCCESS); 1701 } else { 1702 /* 1703 * traverse the port list 1704 */ 1705 fport = fcip_port_head; 1706 while (fport != NULL) { 1707 if (fport->fcipp_handle == 1708 port_info->port_handle) { 1709 fptr = fport->fcipp_fcip; 1710 1711 /* 1712 * Fail the port detach if there is 1713 * still an attached, bound stream on 1714 * this interface. 1715 */ 1716 1717 rw_enter(&fcipstruplock, RW_READER); 1718 1719 for (strp = fcipstrup; strp != NULL; 1720 strp = strp->sl_nextp) { 1721 if (strp->sl_fcip == fptr) { 1722 rw_exit(&fcipstruplock); 1723 mutex_exit( 1724 &fcip_global_mutex); 1725 return (FC_FAILURE); 1726 } 1727 } 1728 1729 rw_exit(&fcipstruplock); 1730 1731 /* 1732 * fail port detach if we are in 1733 * the middle of a deferred port attach 1734 * or if the port has outstanding pkts 1735 */ 1736 if (fptr != NULL) { 1737 mutex_enter(&fptr->fcip_mutex); 1738 if (fcip_check_port_busy 1739 (fptr) || 1740 (fptr->fcip_flags & 1741 FCIP_DETACHED)) { 1742 mutex_exit( 1743 &fptr->fcip_mutex); 1744 mutex_exit( 1745 &fcip_global_mutex); 1746 return (FC_FAILURE); 1747 } 1748 1749 fptr->fcip_flags |= 1750 FCIP_DETACHED; 1751 mutex_exit(&fptr->fcip_mutex); 1752 } 1753 (void) fcip_softstate_free(fport); 1754 1755 fcip_check_remove_minor_node(); 1756 mutex_exit(&fcip_global_mutex); 1757 return (FC_SUCCESS); 1758 } 1759 fport = fport->fcipp_next; 1760 } 1761 ASSERT(fport == NULL); 1762 } 1763 mutex_exit(&fcip_global_mutex); 1764 break; 1765 } 1766 case FC_CMD_POWER_DOWN: 1767 /* FALLTHROUGH */ 1768 case FC_CMD_SUSPEND: 1769 mutex_enter(&fcip_global_mutex); 1770 fport = fcip_port_head; 1771 while (fport != NULL) { 1772 if (fport->fcipp_handle == port_info->port_handle) { 1773 break; 1774 } 1775 fport = fport->fcipp_next; 1776 } 1777 if (fport == NULL) { 1778 mutex_exit(&fcip_global_mutex); 1779 break; 1780 } 1781 rval = fcip_handle_suspend(fport, cmd); 1782 mutex_exit(&fcip_global_mutex); 1783 break; 1784 default: 1785 FCIP_DEBUG(FCIP_DEBUG_DETACH, 1786 (CE_WARN, "unknown port detach command!!")); 1787 break; 1788 } 1789 return (rval); 1790 } 1791 1792 1793 /* 1794 * Returns 0 if the port is not busy, else returns non zero. 1795 */ 1796 static int 1797 fcip_check_port_busy(struct fcip *fptr) 1798 { 1799 int rval = 0, num_pkts = 0; 1800 1801 ASSERT(fptr != NULL); 1802 ASSERT(MUTEX_HELD(&fptr->fcip_mutex)); 1803 1804 mutex_enter(&fptr->fcip_dest_mutex); 1805 1806 if (fptr->fcip_flags & FCIP_PORT_BUSY || 1807 ((num_pkts = fcip_port_get_num_pkts(fptr)) > 0) || 1808 fptr->fcip_num_ipkts_pending) { 1809 rval = 1; 1810 FCIP_DEBUG(FCIP_DEBUG_DETACH, 1811 (CE_NOTE, "!fcip_check_port_busy: port is busy " 1812 "fcip_flags: 0x%x, num_pkts: 0x%x, ipkts_pending: 0x%lx!", 1813 fptr->fcip_flags, num_pkts, fptr->fcip_num_ipkts_pending)); 1814 } 1815 1816 mutex_exit(&fptr->fcip_dest_mutex); 1817 return (rval); 1818 } 1819 1820 /* 1821 * Helper routine to remove fcip's minor node 1822 * There is one minor node per system and it should be removed if there are no 1823 * other fcip instances (which has a 1:1 mapping for fp instances) present 1824 */ 1825 static void 1826 fcip_check_remove_minor_node(void) 1827 { 1828 ASSERT(MUTEX_HELD(&fcip_global_mutex)); 1829 1830 /* 1831 * If there are no more fcip (fp) instances, remove the 1832 * minor node for fcip. 1833 * Reset fcip_minor_node_created to invalidate it. 1834 */ 1835 if (fcip_num_instances == 0 && (fcip_module_dip != NULL)) { 1836 ddi_remove_minor_node(fcip_module_dip, NULL); 1837 fcip_minor_node_created = 0; 1838 } 1839 } 1840 1841 /* 1842 * This routine permits the suspend operation during a CPR/System 1843 * power management operation. The routine basically quiesces I/Os 1844 * on all active interfaces 1845 */ 1846 static int 1847 fcip_handle_suspend(fcip_port_info_t *fport, fc_detach_cmd_t cmd) 1848 { 1849 struct fcip *fptr = fport->fcipp_fcip; 1850 timeout_id_t tid; 1851 int index; 1852 int tryagain = 0; 1853 int count; 1854 struct fcipstr *tslp; 1855 1856 1857 ASSERT(fptr != NULL); 1858 mutex_enter(&fptr->fcip_mutex); 1859 1860 /* 1861 * Fail if we are in the middle of a callback. Don't use delay during 1862 * suspend since clock intrs are not available so busy wait 1863 */ 1864 count = 0; 1865 while (count++ < 15 && 1866 ((fptr->fcip_flags & FCIP_IN_CALLBACK) || 1867 (fptr->fcip_flags & FCIP_IN_TIMEOUT))) { 1868 mutex_exit(&fptr->fcip_mutex); 1869 drv_usecwait(1000000); 1870 mutex_enter(&fptr->fcip_mutex); 1871 } 1872 1873 if (fptr->fcip_flags & FCIP_IN_CALLBACK || 1874 fptr->fcip_flags & FCIP_IN_TIMEOUT) { 1875 mutex_exit(&fptr->fcip_mutex); 1876 return (FC_FAILURE); 1877 } 1878 1879 if (cmd == FC_CMD_POWER_DOWN) { 1880 if (fptr->fcip_flags & FCIP_SUSPENDED) { 1881 fptr->fcip_flags |= FCIP_POWER_DOWN; 1882 mutex_exit(&fptr->fcip_mutex); 1883 goto success; 1884 } else { 1885 fptr->fcip_flags |= FCIP_POWER_DOWN; 1886 } 1887 } else if (cmd == FC_CMD_SUSPEND) { 1888 fptr->fcip_flags |= FCIP_SUSPENDED; 1889 } else { 1890 mutex_exit(&fptr->fcip_mutex); 1891 return (FC_FAILURE); 1892 } 1893 1894 mutex_exit(&fptr->fcip_mutex); 1895 /* 1896 * If no streams are plumbed - its the easiest case - Just 1897 * bail out without having to do much 1898 */ 1899 1900 rw_enter(&fcipstruplock, RW_READER); 1901 for (tslp = fcipstrup; tslp; tslp = tslp->sl_nextp) { 1902 if (tslp->sl_fcip == fptr) { 1903 break; 1904 } 1905 } 1906 rw_exit(&fcipstruplock); 1907 1908 /* 1909 * No active streams on this port 1910 */ 1911 if (tslp == NULL) { 1912 goto success; 1913 } 1914 1915 /* 1916 * Walk through each Routing table structure and check if 1917 * the destination table has any outstanding commands. If yes 1918 * wait for the commands to drain. Since we go through each 1919 * routing table entry in succession, it may be wise to wait 1920 * only a few seconds for each entry. 1921 */ 1922 mutex_enter(&fptr->fcip_rt_mutex); 1923 while (!tryagain) { 1924 1925 tryagain = 0; 1926 for (index = 0; index < FCIP_RT_HASH_ELEMS; index++) { 1927 struct fcip_routing_table *frp; 1928 struct fcip_dest *fdestp; 1929 la_wwn_t *pwwn; 1930 int hash_bucket; 1931 1932 frp = fptr->fcip_rtable[index]; 1933 while (frp) { 1934 /* 1935 * Mark the routing table as SUSPENDED. Even 1936 * mark the broadcast entry SUSPENDED to 1937 * prevent any ARP or other broadcasts. We 1938 * can reset the state of the broadcast 1939 * RTE when we resume. 1940 */ 1941 frp->fcipr_state = FCIP_RT_SUSPENDED; 1942 pwwn = &frp->fcipr_pwwn; 1943 1944 /* 1945 * Get hold of destination pointer 1946 */ 1947 mutex_enter(&fptr->fcip_dest_mutex); 1948 1949 hash_bucket = FCIP_DEST_HASH(pwwn->raw_wwn); 1950 ASSERT(hash_bucket < FCIP_DEST_HASH_ELEMS); 1951 1952 fdestp = fptr->fcip_dest[hash_bucket]; 1953 while (fdestp != NULL) { 1954 mutex_enter(&fdestp->fcipd_mutex); 1955 if (fdestp->fcipd_rtable) { 1956 if (fcip_wwn_compare(pwwn, 1957 &fdestp->fcipd_pwwn, 1958 FCIP_COMPARE_PWWN) == 0) { 1959 mutex_exit( 1960 &fdestp->fcipd_mutex); 1961 break; 1962 } 1963 } 1964 mutex_exit(&fdestp->fcipd_mutex); 1965 fdestp = fdestp->fcipd_next; 1966 } 1967 1968 mutex_exit(&fptr->fcip_dest_mutex); 1969 if (fdestp == NULL) { 1970 frp = frp->fcipr_next; 1971 continue; 1972 } 1973 1974 /* 1975 * Wait for fcip_wait_cmds seconds for 1976 * the commands to drain. 1977 */ 1978 count = 0; 1979 mutex_enter(&fdestp->fcipd_mutex); 1980 while (fdestp->fcipd_ncmds && 1981 count < fcip_wait_cmds) { 1982 mutex_exit(&fdestp->fcipd_mutex); 1983 mutex_exit(&fptr->fcip_rt_mutex); 1984 drv_usecwait(1000000); 1985 mutex_enter(&fptr->fcip_rt_mutex); 1986 mutex_enter(&fdestp->fcipd_mutex); 1987 count++; 1988 } 1989 /* 1990 * Check if we were able to drain all cmds 1991 * successfully. Else continue with other 1992 * ports and try during the second pass 1993 */ 1994 if (fdestp->fcipd_ncmds) { 1995 tryagain++; 1996 } 1997 mutex_exit(&fdestp->fcipd_mutex); 1998 1999 frp = frp->fcipr_next; 2000 } 2001 } 2002 if (tryagain == 0) { 2003 break; 2004 } 2005 } 2006 mutex_exit(&fptr->fcip_rt_mutex); 2007 2008 if (tryagain) { 2009 mutex_enter(&fptr->fcip_mutex); 2010 fptr->fcip_flags &= ~(FCIP_SUSPENDED | FCIP_POWER_DOWN); 2011 mutex_exit(&fptr->fcip_mutex); 2012 return (FC_FAILURE); 2013 } 2014 2015 success: 2016 mutex_enter(&fptr->fcip_mutex); 2017 tid = fptr->fcip_timeout_id; 2018 fptr->fcip_timeout_id = NULL; 2019 mutex_exit(&fptr->fcip_mutex); 2020 2021 (void) untimeout(tid); 2022 2023 return (FC_SUCCESS); 2024 } 2025 2026 /* 2027 * the getinfo(9E) entry point 2028 */ 2029 /* ARGSUSED */ 2030 static int 2031 fcip_getinfo(dev_info_t *dip, ddi_info_cmd_t cmd, void *arg, void **result) 2032 { 2033 int rval = DDI_FAILURE; 2034 2035 switch (cmd) { 2036 case DDI_INFO_DEVT2DEVINFO: 2037 *result = fcip_module_dip; 2038 if (*result) 2039 rval = DDI_SUCCESS; 2040 break; 2041 2042 case DDI_INFO_DEVT2INSTANCE: 2043 *result = (void *)0; 2044 rval = DDI_SUCCESS; 2045 break; 2046 default: 2047 break; 2048 } 2049 2050 return (rval); 2051 } 2052 2053 /* 2054 * called from fcip_attach to initialize kstats for the link 2055 */ 2056 /* ARGSUSED */ 2057 static void 2058 fcip_kstat_init(struct fcip *fptr) 2059 { 2060 int instance; 2061 char buf[16]; 2062 struct fcipstat *fcipstatp; 2063 2064 ASSERT(mutex_owned(&fptr->fcip_mutex)); 2065 2066 instance = ddi_get_instance(fptr->fcip_dip); 2067 (void) sprintf(buf, "fcip%d", instance); 2068 2069 #ifdef kstat 2070 fptr->fcip_kstatp = kstat_create("fcip", instance, buf, "net", 2071 KSTAT_TYPE_NAMED, 2072 (sizeof (struct fcipstat)/ sizeof (kstat_named_t)), 2073 KSTAT_FLAG_PERSISTENT); 2074 #else 2075 fptr->fcip_kstatp = kstat_create("fcip", instance, buf, "net", 2076 KSTAT_TYPE_NAMED, 2077 (sizeof (struct fcipstat)/ sizeof (kstat_named_t)), 0); 2078 #endif 2079 if (fptr->fcip_kstatp == NULL) { 2080 FCIP_DEBUG(FCIP_DEBUG_INIT, (CE_WARN, "kstat created failed")); 2081 return; 2082 } 2083 2084 fcipstatp = (struct fcipstat *)fptr->fcip_kstatp->ks_data; 2085 kstat_named_init(&fcipstatp->fcips_ipackets, "ipackets", 2086 KSTAT_DATA_ULONG); 2087 kstat_named_init(&fcipstatp->fcips_ierrors, "ierrors", 2088 KSTAT_DATA_ULONG); 2089 kstat_named_init(&fcipstatp->fcips_opackets, "opackets", 2090 KSTAT_DATA_ULONG); 2091 kstat_named_init(&fcipstatp->fcips_oerrors, "oerrors", 2092 KSTAT_DATA_ULONG); 2093 kstat_named_init(&fcipstatp->fcips_collisions, "collisions", 2094 KSTAT_DATA_ULONG); 2095 kstat_named_init(&fcipstatp->fcips_nocanput, "nocanput", 2096 KSTAT_DATA_ULONG); 2097 kstat_named_init(&fcipstatp->fcips_allocbfail, "allocbfail", 2098 KSTAT_DATA_ULONG); 2099 2100 kstat_named_init(&fcipstatp->fcips_defer, "defer", 2101 KSTAT_DATA_ULONG); 2102 kstat_named_init(&fcipstatp->fcips_fram, "fram", 2103 KSTAT_DATA_ULONG); 2104 kstat_named_init(&fcipstatp->fcips_crc, "crc", 2105 KSTAT_DATA_ULONG); 2106 kstat_named_init(&fcipstatp->fcips_oflo, "oflo", 2107 KSTAT_DATA_ULONG); 2108 kstat_named_init(&fcipstatp->fcips_uflo, "uflo", 2109 KSTAT_DATA_ULONG); 2110 kstat_named_init(&fcipstatp->fcips_missed, "missed", 2111 KSTAT_DATA_ULONG); 2112 kstat_named_init(&fcipstatp->fcips_tlcol, "tlcol", 2113 KSTAT_DATA_ULONG); 2114 kstat_named_init(&fcipstatp->fcips_trtry, "trtry", 2115 KSTAT_DATA_ULONG); 2116 kstat_named_init(&fcipstatp->fcips_tnocar, "tnocar", 2117 KSTAT_DATA_ULONG); 2118 kstat_named_init(&fcipstatp->fcips_inits, "inits", 2119 KSTAT_DATA_ULONG); 2120 kstat_named_init(&fcipstatp->fcips_notbufs, "notbufs", 2121 KSTAT_DATA_ULONG); 2122 kstat_named_init(&fcipstatp->fcips_norbufs, "norbufs", 2123 KSTAT_DATA_ULONG); 2124 kstat_named_init(&fcipstatp->fcips_allocbfail, "allocbfail", 2125 KSTAT_DATA_ULONG); 2126 2127 /* 2128 * required by kstat for MIB II objects(RFC 1213) 2129 */ 2130 kstat_named_init(&fcipstatp->fcips_rcvbytes, "fcips_rcvbytes", 2131 KSTAT_DATA_ULONG); /* # octets received */ 2132 /* MIB - ifInOctets */ 2133 kstat_named_init(&fcipstatp->fcips_xmtbytes, "fcips_xmtbytes", 2134 KSTAT_DATA_ULONG); /* # octets xmitted */ 2135 /* MIB - ifOutOctets */ 2136 kstat_named_init(&fcipstatp->fcips_multircv, "fcips_multircv", 2137 KSTAT_DATA_ULONG); /* # multicast packets */ 2138 /* delivered to upper layer */ 2139 /* MIB - ifInNUcastPkts */ 2140 kstat_named_init(&fcipstatp->fcips_multixmt, "fcips_multixmt", 2141 KSTAT_DATA_ULONG); /* # multicast packets */ 2142 /* requested to be sent */ 2143 /* MIB - ifOutNUcastPkts */ 2144 kstat_named_init(&fcipstatp->fcips_brdcstrcv, "fcips_brdcstrcv", 2145 KSTAT_DATA_ULONG); /* # broadcast packets */ 2146 /* delivered to upper layer */ 2147 /* MIB - ifInNUcastPkts */ 2148 kstat_named_init(&fcipstatp->fcips_brdcstxmt, "fcips_brdcstxmt", 2149 KSTAT_DATA_ULONG); /* # broadcast packets */ 2150 /* requested to be sent */ 2151 /* MIB - ifOutNUcastPkts */ 2152 kstat_named_init(&fcipstatp->fcips_norcvbuf, "fcips_norcvbuf", 2153 KSTAT_DATA_ULONG); /* # rcv packets discarded */ 2154 /* MIB - ifInDiscards */ 2155 kstat_named_init(&fcipstatp->fcips_noxmtbuf, "fcips_noxmtbuf", 2156 KSTAT_DATA_ULONG); /* # xmt packets discarded */ 2157 2158 fptr->fcip_kstatp->ks_update = fcip_stat_update; 2159 fptr->fcip_kstatp->ks_private = (void *) fptr; 2160 kstat_install(fptr->fcip_kstatp); 2161 } 2162 2163 /* 2164 * Update the defined kstats for netstat et al to use 2165 */ 2166 /* ARGSUSED */ 2167 static int 2168 fcip_stat_update(kstat_t *fcip_statp, int val) 2169 { 2170 struct fcipstat *fcipstatp; 2171 struct fcip *fptr; 2172 2173 fptr = (struct fcip *)fcip_statp->ks_private; 2174 fcipstatp = (struct fcipstat *)fcip_statp->ks_data; 2175 2176 if (val == KSTAT_WRITE) { 2177 fptr->fcip_ipackets = fcipstatp->fcips_ipackets.value.ul; 2178 fptr->fcip_ierrors = fcipstatp->fcips_ierrors.value.ul; 2179 fptr->fcip_opackets = fcipstatp->fcips_opackets.value.ul; 2180 fptr->fcip_oerrors = fcipstatp->fcips_oerrors.value.ul; 2181 fptr->fcip_collisions = fcipstatp->fcips_collisions.value.ul; 2182 fptr->fcip_defer = fcipstatp->fcips_defer.value.ul; 2183 fptr->fcip_fram = fcipstatp->fcips_fram.value.ul; 2184 fptr->fcip_crc = fcipstatp->fcips_crc.value.ul; 2185 fptr->fcip_oflo = fcipstatp->fcips_oflo.value.ul; 2186 fptr->fcip_uflo = fcipstatp->fcips_uflo.value.ul; 2187 fptr->fcip_missed = fcipstatp->fcips_missed.value.ul; 2188 fptr->fcip_tlcol = fcipstatp->fcips_tlcol.value.ul; 2189 fptr->fcip_trtry = fcipstatp->fcips_trtry.value.ul; 2190 fptr->fcip_tnocar = fcipstatp->fcips_tnocar.value.ul; 2191 fptr->fcip_inits = fcipstatp->fcips_inits.value.ul; 2192 fptr->fcip_notbufs = fcipstatp->fcips_notbufs.value.ul; 2193 fptr->fcip_norbufs = fcipstatp->fcips_norbufs.value.ul; 2194 fptr->fcip_nocanput = fcipstatp->fcips_nocanput.value.ul; 2195 fptr->fcip_allocbfail = fcipstatp->fcips_allocbfail.value.ul; 2196 fptr->fcip_rcvbytes = fcipstatp->fcips_rcvbytes.value.ul; 2197 fptr->fcip_xmtbytes = fcipstatp->fcips_xmtbytes.value.ul; 2198 fptr->fcip_multircv = fcipstatp->fcips_multircv.value.ul; 2199 fptr->fcip_multixmt = fcipstatp->fcips_multixmt.value.ul; 2200 fptr->fcip_brdcstrcv = fcipstatp->fcips_brdcstrcv.value.ul; 2201 fptr->fcip_norcvbuf = fcipstatp->fcips_norcvbuf.value.ul; 2202 fptr->fcip_noxmtbuf = fcipstatp->fcips_noxmtbuf.value.ul; 2203 fptr->fcip_allocbfail = fcipstatp->fcips_allocbfail.value.ul; 2204 fptr->fcip_allocbfail = fcipstatp->fcips_allocbfail.value.ul; 2205 fptr->fcip_allocbfail = fcipstatp->fcips_allocbfail.value.ul; 2206 fptr->fcip_allocbfail = fcipstatp->fcips_allocbfail.value.ul; 2207 fptr->fcip_allocbfail = fcipstatp->fcips_allocbfail.value.ul; 2208 fptr->fcip_allocbfail = fcipstatp->fcips_allocbfail.value.ul; 2209 fptr->fcip_allocbfail = fcipstatp->fcips_allocbfail.value.ul; 2210 fptr->fcip_allocbfail = fcipstatp->fcips_allocbfail.value.ul; 2211 2212 } else { 2213 fcipstatp->fcips_ipackets.value.ul = fptr->fcip_ipackets; 2214 fcipstatp->fcips_ierrors.value.ul = fptr->fcip_ierrors; 2215 fcipstatp->fcips_opackets.value.ul = fptr->fcip_opackets; 2216 fcipstatp->fcips_oerrors.value.ul = fptr->fcip_oerrors; 2217 fcipstatp->fcips_collisions.value.ul = fptr->fcip_collisions; 2218 fcipstatp->fcips_nocanput.value.ul = fptr->fcip_nocanput; 2219 fcipstatp->fcips_allocbfail.value.ul = fptr->fcip_allocbfail; 2220 fcipstatp->fcips_defer.value.ul = fptr->fcip_defer; 2221 fcipstatp->fcips_fram.value.ul = fptr->fcip_fram; 2222 fcipstatp->fcips_crc.value.ul = fptr->fcip_crc; 2223 fcipstatp->fcips_oflo.value.ul = fptr->fcip_oflo; 2224 fcipstatp->fcips_uflo.value.ul = fptr->fcip_uflo; 2225 fcipstatp->fcips_missed.value.ul = fptr->fcip_missed; 2226 fcipstatp->fcips_tlcol.value.ul = fptr->fcip_tlcol; 2227 fcipstatp->fcips_trtry.value.ul = fptr->fcip_trtry; 2228 fcipstatp->fcips_tnocar.value.ul = fptr->fcip_tnocar; 2229 fcipstatp->fcips_inits.value.ul = fptr->fcip_inits; 2230 fcipstatp->fcips_norbufs.value.ul = fptr->fcip_norbufs; 2231 fcipstatp->fcips_notbufs.value.ul = fptr->fcip_notbufs; 2232 fcipstatp->fcips_rcvbytes.value.ul = fptr->fcip_rcvbytes; 2233 fcipstatp->fcips_xmtbytes.value.ul = fptr->fcip_xmtbytes; 2234 fcipstatp->fcips_multircv.value.ul = fptr->fcip_multircv; 2235 fcipstatp->fcips_multixmt.value.ul = fptr->fcip_multixmt; 2236 fcipstatp->fcips_brdcstrcv.value.ul = fptr->fcip_brdcstrcv; 2237 fcipstatp->fcips_brdcstxmt.value.ul = fptr->fcip_brdcstxmt; 2238 fcipstatp->fcips_norcvbuf.value.ul = fptr->fcip_norcvbuf; 2239 fcipstatp->fcips_noxmtbuf.value.ul = fptr->fcip_noxmtbuf; 2240 2241 } 2242 return (0); 2243 } 2244 2245 2246 /* 2247 * fcip_statec_cb: handles all required state change callback notifications 2248 * it receives from the transport 2249 */ 2250 /* ARGSUSED */ 2251 static void 2252 fcip_statec_cb(opaque_t ulp_handle, opaque_t phandle, 2253 uint32_t port_state, uint32_t port_top, fc_portmap_t changelist[], 2254 uint32_t listlen, uint32_t sid) 2255 { 2256 fcip_port_info_t *fport; 2257 struct fcip *fptr; 2258 struct fcipstr *slp; 2259 queue_t *wrq; 2260 int instance; 2261 int index; 2262 struct fcip_routing_table *frtp; 2263 2264 fport = fcip_get_port(phandle); 2265 2266 if (fport == NULL) { 2267 return; 2268 } 2269 2270 fptr = fport->fcipp_fcip; 2271 ASSERT(fptr != NULL); 2272 2273 if (fptr == NULL) { 2274 return; 2275 } 2276 2277 instance = ddi_get_instance(fport->fcipp_dip); 2278 2279 FCIP_TNF_PROBE_4((fcip_statec_cb, "fcip io", /* CSTYLED */, 2280 tnf_string, msg, "state change callback", 2281 tnf_uint, instance, instance, 2282 tnf_uint, S_ID, sid, 2283 tnf_int, count, listlen)); 2284 FCIP_DEBUG(FCIP_DEBUG_ELS, 2285 (CE_NOTE, "fcip%d, state change callback: state:0x%x, " 2286 "S_ID:0x%x, count:0x%x", instance, port_state, sid, listlen)); 2287 2288 mutex_enter(&fptr->fcip_mutex); 2289 2290 if ((fptr->fcip_flags & (FCIP_DETACHING | FCIP_DETACHED)) || 2291 (fptr->fcip_flags & (FCIP_SUSPENDED | FCIP_POWER_DOWN))) { 2292 mutex_exit(&fptr->fcip_mutex); 2293 return; 2294 } 2295 2296 /* 2297 * set fcip flags to indicate we are in the middle of a 2298 * state change callback so we can wait till the statechange 2299 * is handled before succeeding/failing the SUSPEND/POWER DOWN. 2300 */ 2301 fptr->fcip_flags |= FCIP_IN_SC_CB; 2302 2303 fport->fcipp_pstate = port_state; 2304 2305 /* 2306 * Check if topology changed. If Yes - Modify the broadcast 2307 * RTE entries to understand the new broadcast D_IDs 2308 */ 2309 if (fport->fcipp_topology != port_top && 2310 (port_top != FC_TOP_UNKNOWN)) { 2311 /* REMOVE later */ 2312 FCIP_DEBUG(FCIP_DEBUG_ELS, (CE_NOTE, 2313 "topology changed: Old topology: 0x%x New topology 0x%x", 2314 fport->fcipp_topology, port_top)); 2315 /* 2316 * If topology changed - attempt a rediscovery of 2317 * devices. Helps specially in Fabric/Public loops 2318 * and if on_demand_node_creation is disabled 2319 */ 2320 fport->fcipp_topology = port_top; 2321 fcip_handle_topology(fptr); 2322 } 2323 2324 mutex_exit(&fptr->fcip_mutex); 2325 2326 switch (FC_PORT_STATE_MASK(port_state)) { 2327 case FC_STATE_ONLINE: 2328 /* FALLTHROUGH */ 2329 case FC_STATE_LIP: 2330 /* FALLTHROUGH */ 2331 case FC_STATE_LIP_LBIT_SET: 2332 2333 /* 2334 * nothing to do here actually other than if we 2335 * were actually logged onto a port in the devlist 2336 * (which indicates active communication between 2337 * the host port and the port in the changelist). 2338 * If however we are in a private loop or point to 2339 * point mode, we need to check for any IP capable 2340 * ports and update our routing table. 2341 */ 2342 switch (port_top) { 2343 case FC_TOP_FABRIC: 2344 /* 2345 * This indicates a fabric port with a NameServer. 2346 * Check the devlist to see if we are in active 2347 * communication with a port on the devlist. 2348 */ 2349 FCIP_DEBUG(FCIP_DEBUG_ELS, (CE_NOTE, 2350 "Statec_cb: fabric topology")); 2351 fcip_rt_update(fptr, changelist, listlen); 2352 break; 2353 case FC_TOP_NO_NS: 2354 /* 2355 * No nameserver - so treat it like a Private loop 2356 * or point to point topology and get a map of 2357 * devices on the link and get IP capable ports to 2358 * to update the routing table. 2359 */ 2360 FCIP_DEBUG(FCIP_DEBUG_ELS, 2361 (CE_NOTE, "Statec_cb: NO_NS topology")); 2362 /* FALLTHROUGH */ 2363 case FC_TOP_PRIVATE_LOOP: 2364 FCIP_DEBUG(FCIP_DEBUG_ELS, (CE_NOTE, 2365 "Statec_cb: Pvt_Loop topology")); 2366 /* FALLTHROUGH */ 2367 case FC_TOP_PT_PT: 2368 /* 2369 * call get_port_map() and update routing table 2370 */ 2371 fcip_rt_update(fptr, changelist, listlen); 2372 break; 2373 default: 2374 FCIP_DEBUG(FCIP_DEBUG_ELS, 2375 (CE_NOTE, "Statec_cb: Unknown topology")); 2376 } 2377 2378 /* 2379 * We should now enable the Queues and permit I/Os 2380 * to flow through downstream. The update of routing 2381 * table should have flushed out any port entries that 2382 * don't exist or are not available after the state change 2383 */ 2384 mutex_enter(&fptr->fcip_mutex); 2385 fptr->fcip_port_state = FCIP_PORT_ONLINE; 2386 if (fptr->fcip_flags & FCIP_LINK_DOWN) { 2387 fptr->fcip_flags &= ~FCIP_LINK_DOWN; 2388 } 2389 mutex_exit(&fptr->fcip_mutex); 2390 2391 /* 2392 * Enable write queues 2393 */ 2394 rw_enter(&fcipstruplock, RW_READER); 2395 for (slp = fcipstrup; slp != NULL; slp = slp->sl_nextp) { 2396 if (slp && slp->sl_fcip == fptr) { 2397 wrq = WR(slp->sl_rq); 2398 if (wrq->q_flag & QFULL) { 2399 qenable(wrq); 2400 } 2401 } 2402 } 2403 rw_exit(&fcipstruplock); 2404 break; 2405 case FC_STATE_OFFLINE: 2406 /* 2407 * mark the port_state OFFLINE and wait for it to 2408 * become online. Any new messages in this state will 2409 * simply be queued back up. If the port does not 2410 * come online in a short while, we can begin failing 2411 * messages and flush the routing table 2412 */ 2413 mutex_enter(&fptr->fcip_mutex); 2414 fptr->fcip_mark_offline = fptr->fcip_timeout_ticks + 2415 FCIP_OFFLINE_TIMEOUT; 2416 fptr->fcip_port_state = FCIP_PORT_OFFLINE; 2417 mutex_exit(&fptr->fcip_mutex); 2418 2419 /* 2420 * Mark all Routing table entries as invalid to prevent 2421 * any commands from trickling through to ports that 2422 * have disappeared from under us 2423 */ 2424 mutex_enter(&fptr->fcip_rt_mutex); 2425 for (index = 0; index < FCIP_RT_HASH_ELEMS; index++) { 2426 frtp = fptr->fcip_rtable[index]; 2427 while (frtp) { 2428 frtp->fcipr_state = PORT_DEVICE_INVALID; 2429 frtp = frtp->fcipr_next; 2430 } 2431 } 2432 mutex_exit(&fptr->fcip_rt_mutex); 2433 2434 break; 2435 2436 case FC_STATE_RESET_REQUESTED: 2437 /* 2438 * Release all Unsolicited buffers back to transport/FCA. 2439 * This also means the port state is marked offline - so 2440 * we may have to do what OFFLINE state requires us to do. 2441 * Care must be taken to wait for any active unsolicited 2442 * buffer with the other Streams modules - so wait for 2443 * a freeb if the unsolicited buffer is passed back all 2444 * the way upstream. 2445 */ 2446 mutex_enter(&fptr->fcip_mutex); 2447 2448 #ifdef FCIP_ESBALLOC 2449 while (fptr->fcip_ub_upstream) { 2450 cv_wait(&fptr->fcip_ub_cv, &fptr->fcip_mutex); 2451 } 2452 #endif /* FCIP_ESBALLOC */ 2453 2454 fptr->fcip_mark_offline = fptr->fcip_timeout_ticks + 2455 FCIP_OFFLINE_TIMEOUT; 2456 fptr->fcip_port_state = FCIP_PORT_OFFLINE; 2457 mutex_exit(&fptr->fcip_mutex); 2458 break; 2459 2460 case FC_STATE_DEVICE_CHANGE: 2461 if (listlen) { 2462 fcip_rt_update(fptr, changelist, listlen); 2463 } 2464 break; 2465 case FC_STATE_RESET: 2466 /* 2467 * Not much to do I guess - wait for port to become 2468 * ONLINE. If the port doesn't become online in a short 2469 * while, the upper layers abort any request themselves. 2470 * We can just putback the messages in the streams queues 2471 * if the link is offline 2472 */ 2473 break; 2474 } 2475 mutex_enter(&fptr->fcip_mutex); 2476 fptr->fcip_flags &= ~(FCIP_IN_SC_CB); 2477 mutex_exit(&fptr->fcip_mutex); 2478 } 2479 2480 /* 2481 * Given a port handle, return the fcip_port_info structure corresponding 2482 * to that port handle. The transport allocates and communicates with 2483 * ULPs using port handles 2484 */ 2485 static fcip_port_info_t * 2486 fcip_get_port(opaque_t phandle) 2487 { 2488 fcip_port_info_t *fport; 2489 2490 ASSERT(phandle != NULL); 2491 2492 mutex_enter(&fcip_global_mutex); 2493 fport = fcip_port_head; 2494 2495 while (fport != NULL) { 2496 if (fport->fcipp_handle == phandle) { 2497 /* found */ 2498 break; 2499 } 2500 fport = fport->fcipp_next; 2501 } 2502 2503 mutex_exit(&fcip_global_mutex); 2504 2505 return (fport); 2506 } 2507 2508 /* 2509 * Handle inbound ELS requests received by the transport. We are only 2510 * intereseted in FARP/InARP mostly. 2511 */ 2512 /* ARGSUSED */ 2513 static int 2514 fcip_els_cb(opaque_t ulp_handle, opaque_t phandle, 2515 fc_unsol_buf_t *buf, uint32_t claimed) 2516 { 2517 fcip_port_info_t *fport; 2518 struct fcip *fptr; 2519 int instance; 2520 uchar_t r_ctl; 2521 uchar_t ls_code; 2522 la_els_farp_t farp_cmd; 2523 la_els_farp_t *fcmd; 2524 int rval = FC_UNCLAIMED; 2525 2526 fport = fcip_get_port(phandle); 2527 if (fport == NULL) { 2528 return (FC_UNCLAIMED); 2529 } 2530 2531 fptr = fport->fcipp_fcip; 2532 ASSERT(fptr != NULL); 2533 if (fptr == NULL) { 2534 return (FC_UNCLAIMED); 2535 } 2536 2537 instance = ddi_get_instance(fport->fcipp_dip); 2538 2539 mutex_enter(&fptr->fcip_mutex); 2540 if ((fptr->fcip_flags & (FCIP_DETACHING | FCIP_DETACHED)) || 2541 (fptr->fcip_flags & (FCIP_SUSPENDED | FCIP_POWER_DOWN))) { 2542 mutex_exit(&fptr->fcip_mutex); 2543 return (FC_UNCLAIMED); 2544 } 2545 2546 /* 2547 * set fcip flags to indicate we are in the middle of a 2548 * ELS callback so we can wait till the statechange 2549 * is handled before succeeding/failing the SUSPEND/POWER DOWN. 2550 */ 2551 fptr->fcip_flags |= FCIP_IN_ELS_CB; 2552 mutex_exit(&fptr->fcip_mutex); 2553 2554 FCIP_TNF_PROBE_2((fcip_els_cb, "fcip io", /* CSTYLED */, 2555 tnf_string, msg, "ELS callback", 2556 tnf_uint, instance, instance)); 2557 2558 FCIP_DEBUG(FCIP_DEBUG_ELS, 2559 (CE_NOTE, "fcip%d, ELS callback , ", instance)); 2560 2561 r_ctl = buf->ub_frame.r_ctl; 2562 switch (r_ctl & R_CTL_ROUTING) { 2563 case R_CTL_EXTENDED_SVC: 2564 if (r_ctl == R_CTL_ELS_REQ) { 2565 ls_code = buf->ub_buffer[0]; 2566 if (ls_code == LA_ELS_FARP_REQ) { 2567 /* 2568 * Inbound FARP broadcast request 2569 */ 2570 if (buf->ub_bufsize != sizeof (la_els_farp_t)) { 2571 FCIP_DEBUG(FCIP_DEBUG_ELS, (CE_WARN, 2572 "Invalid FARP req buffer size " 2573 "expected 0x%lx, got 0x%x", 2574 (long)(sizeof (la_els_farp_t)), 2575 buf->ub_bufsize)); 2576 rval = FC_UNCLAIMED; 2577 goto els_cb_done; 2578 } 2579 fcmd = (la_els_farp_t *)buf; 2580 if (fcip_wwn_compare(&fcmd->resp_nwwn, 2581 &fport->fcipp_nwwn, 2582 FCIP_COMPARE_NWWN) != 0) { 2583 rval = FC_UNCLAIMED; 2584 goto els_cb_done; 2585 } 2586 /* 2587 * copy the FARP request and release the 2588 * unsolicited buffer 2589 */ 2590 fcmd = &farp_cmd; 2591 bcopy((void *)buf, (void *)fcmd, 2592 sizeof (la_els_farp_t)); 2593 (void) fc_ulp_ubrelease(fport->fcipp_handle, 1, 2594 &buf->ub_token); 2595 2596 if (fcip_farp_supported && 2597 fcip_handle_farp_request(fptr, fcmd) == 2598 FC_SUCCESS) { 2599 /* 2600 * We successfully sent out a FARP 2601 * reply to the requesting port 2602 */ 2603 rval = FC_SUCCESS; 2604 goto els_cb_done; 2605 } else { 2606 rval = FC_UNCLAIMED; 2607 goto els_cb_done; 2608 } 2609 } 2610 } else if (r_ctl == R_CTL_ELS_RSP) { 2611 ls_code = buf->ub_buffer[0]; 2612 if (ls_code == LA_ELS_FARP_REPLY) { 2613 /* 2614 * We received a REPLY to our FARP request 2615 */ 2616 if (buf->ub_bufsize != sizeof (la_els_farp_t)) { 2617 FCIP_DEBUG(FCIP_DEBUG_ELS, (CE_WARN, 2618 "Invalid FARP req buffer size " 2619 "expected 0x%lx, got 0x%x", 2620 (long)(sizeof (la_els_farp_t)), 2621 buf->ub_bufsize)); 2622 rval = FC_UNCLAIMED; 2623 goto els_cb_done; 2624 } 2625 fcmd = &farp_cmd; 2626 bcopy((void *)buf, (void *)fcmd, 2627 sizeof (la_els_farp_t)); 2628 (void) fc_ulp_ubrelease(fport->fcipp_handle, 1, 2629 &buf->ub_token); 2630 if (fcip_farp_supported && 2631 fcip_handle_farp_response(fptr, fcmd) == 2632 FC_SUCCESS) { 2633 FCIP_DEBUG(FCIP_DEBUG_ELS, (CE_NOTE, 2634 "Successfully recevied a FARP " 2635 "response")); 2636 mutex_enter(&fptr->fcip_mutex); 2637 fptr->fcip_farp_rsp_flag = 1; 2638 cv_signal(&fptr->fcip_farp_cv); 2639 mutex_exit(&fptr->fcip_mutex); 2640 rval = FC_SUCCESS; 2641 goto els_cb_done; 2642 } else { 2643 FCIP_DEBUG(FCIP_DEBUG_ELS, (CE_WARN, 2644 "Unable to handle a FARP response " 2645 "receive")); 2646 rval = FC_UNCLAIMED; 2647 goto els_cb_done; 2648 } 2649 } 2650 } 2651 break; 2652 default: 2653 break; 2654 } 2655 els_cb_done: 2656 mutex_enter(&fptr->fcip_mutex); 2657 fptr->fcip_flags &= ~(FCIP_IN_ELS_CB); 2658 mutex_exit(&fptr->fcip_mutex); 2659 return (rval); 2660 } 2661 2662 2663 /* 2664 * Handle inbound FARP requests 2665 */ 2666 static int 2667 fcip_handle_farp_request(struct fcip *fptr, la_els_farp_t *fcmd) 2668 { 2669 fcip_pkt_t *fcip_pkt; 2670 fc_packet_t *fc_pkt; 2671 fcip_port_info_t *fport = fptr->fcip_port_info; 2672 int rval = FC_FAILURE; 2673 opaque_t fca_dev; 2674 fc_portmap_t map; 2675 struct fcip_routing_table *frp; 2676 struct fcip_dest *fdestp; 2677 2678 /* 2679 * Add an entry for the remote port into our routing and destination 2680 * tables. 2681 */ 2682 map.map_did = fcmd->req_id; 2683 map.map_hard_addr.hard_addr = fcmd->req_id.port_id; 2684 map.map_state = PORT_DEVICE_VALID; 2685 map.map_type = PORT_DEVICE_NEW; 2686 map.map_flags = 0; 2687 map.map_pd = NULL; 2688 bcopy((void *)&fcmd->req_pwwn, (void *)&map.map_pwwn, 2689 sizeof (la_wwn_t)); 2690 bcopy((void *)&fcmd->req_nwwn, (void *)&map.map_nwwn, 2691 sizeof (la_wwn_t)); 2692 fcip_rt_update(fptr, &map, 1); 2693 mutex_enter(&fptr->fcip_rt_mutex); 2694 frp = fcip_lookup_rtable(fptr, &fcmd->req_pwwn, FCIP_COMPARE_NWWN); 2695 mutex_exit(&fptr->fcip_rt_mutex); 2696 2697 fdestp = fcip_add_dest(fptr, frp); 2698 2699 fcip_pkt = fcip_ipkt_alloc(fptr, sizeof (la_els_farp_t), 2700 sizeof (la_els_farp_t), NULL, KM_SLEEP); 2701 if (fcip_pkt == NULL) { 2702 rval = FC_FAILURE; 2703 goto farp_done; 2704 } 2705 /* 2706 * Fill in our port's PWWN and NWWN 2707 */ 2708 fcmd->resp_pwwn = fport->fcipp_pwwn; 2709 fcmd->resp_nwwn = fport->fcipp_nwwn; 2710 2711 fcip_init_unicast_pkt(fcip_pkt, fport->fcipp_sid, 2712 fcmd->req_id, NULL); 2713 2714 fca_dev = 2715 fc_ulp_get_fca_device(fport->fcipp_handle, fcmd->req_id); 2716 fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 2717 fc_pkt->pkt_cmd_fhdr.r_ctl = R_CTL_ELS_RSP; 2718 fc_pkt->pkt_fca_device = fca_dev; 2719 fcip_pkt->fcip_pkt_dest = fdestp; 2720 2721 /* 2722 * Attempt a PLOGI again 2723 */ 2724 if (fcmd->resp_flags & FARP_INIT_P_LOGI) { 2725 if (fcip_do_plogi(fptr, frp) != FC_SUCCESS) { 2726 /* 2727 * Login to the remote port failed. There is no 2728 * point continuing with the FARP request further 2729 * so bail out here. 2730 */ 2731 frp->fcipr_state = PORT_DEVICE_INVALID; 2732 rval = FC_FAILURE; 2733 goto farp_done; 2734 } 2735 } 2736 2737 FCIP_CP_OUT(fcmd, fc_pkt->pkt_cmd, fc_pkt->pkt_cmd_acc, 2738 sizeof (la_els_farp_t)); 2739 2740 rval = fc_ulp_issue_els(fport->fcipp_handle, fc_pkt); 2741 if (rval != FC_SUCCESS) { 2742 FCIP_TNF_PROBE_2((fcip_handle_farp_request, "fcip io", 2743 /* CSTYLED */, tnf_string, msg, 2744 "fcip_transport of farp reply failed", 2745 tnf_uint, rval, rval)); 2746 FCIP_DEBUG(FCIP_DEBUG_ELS, (CE_WARN, 2747 "fcip_transport of farp reply failed 0x%x", rval)); 2748 } 2749 2750 farp_done: 2751 return (rval); 2752 } 2753 2754 2755 /* 2756 * Handle FARP responses to our FARP requests. When we receive a FARP 2757 * reply, we need to add the entry for the Port that replied into our 2758 * routing and destination hash tables. It is possible that the remote 2759 * port did not login into us (FARP responses can be received without 2760 * a PLOGI) 2761 */ 2762 static int 2763 fcip_handle_farp_response(struct fcip *fptr, la_els_farp_t *fcmd) 2764 { 2765 int rval = FC_FAILURE; 2766 fc_portmap_t map; 2767 struct fcip_routing_table *frp; 2768 struct fcip_dest *fdestp; 2769 2770 /* 2771 * Add an entry for the remote port into our routing and destination 2772 * tables. 2773 */ 2774 map.map_did = fcmd->dest_id; 2775 map.map_hard_addr.hard_addr = fcmd->dest_id.port_id; 2776 map.map_state = PORT_DEVICE_VALID; 2777 map.map_type = PORT_DEVICE_NEW; 2778 map.map_flags = 0; 2779 map.map_pd = NULL; 2780 bcopy((void *)&fcmd->resp_pwwn, (void *)&map.map_pwwn, 2781 sizeof (la_wwn_t)); 2782 bcopy((void *)&fcmd->resp_nwwn, (void *)&map.map_nwwn, 2783 sizeof (la_wwn_t)); 2784 fcip_rt_update(fptr, &map, 1); 2785 mutex_enter(&fptr->fcip_rt_mutex); 2786 frp = fcip_lookup_rtable(fptr, &fcmd->resp_pwwn, FCIP_COMPARE_NWWN); 2787 mutex_exit(&fptr->fcip_rt_mutex); 2788 2789 fdestp = fcip_add_dest(fptr, frp); 2790 2791 if (fdestp != NULL) { 2792 rval = FC_SUCCESS; 2793 } 2794 return (rval); 2795 } 2796 2797 2798 #define FCIP_HDRS_LENGTH \ 2799 sizeof (fcph_network_hdr_t)+sizeof (llc_snap_hdr_t)+sizeof (ipha_t) 2800 2801 /* 2802 * fcip_data_cb is the heart of most IP operations. This routine is called 2803 * by the transport when any unsolicited IP data arrives at a port (which 2804 * is almost all IP data). This routine then strips off the Network header 2805 * from the payload (after authenticating the received payload ofcourse), 2806 * creates a message blk and sends the data upstream. You will see ugly 2807 * #defines because of problems with using esballoc() as opposed to 2808 * allocb to prevent an extra copy of data. We should probably move to 2809 * esballoc entirely when the MTU eventually will be larger than 1500 bytes 2810 * since copies will get more expensive then. At 1500 byte MTUs, there is 2811 * no noticable difference between using allocb and esballoc. The other 2812 * caveat is that the qlc firmware still cannot tell us accurately the 2813 * no. of valid bytes in the unsol buffer it DMA'ed so we have to resort 2814 * to looking into the IP header and hoping that the no. of bytes speficified 2815 * in the header was actually received. 2816 */ 2817 /* ARGSUSED */ 2818 static int 2819 fcip_data_cb(opaque_t ulp_handle, opaque_t phandle, 2820 fc_unsol_buf_t *buf, uint32_t claimed) 2821 { 2822 fcip_port_info_t *fport; 2823 struct fcip *fptr; 2824 fcph_network_hdr_t *nhdr; 2825 llc_snap_hdr_t *snaphdr; 2826 mblk_t *bp; 2827 uint32_t len; 2828 uint32_t hdrlen; 2829 ushort_t type; 2830 ipha_t *iphdr; 2831 int rval; 2832 2833 #ifdef FCIP_ESBALLOC 2834 frtn_t *free_ubuf; 2835 struct fcip_esballoc_arg *fesb_argp; 2836 #endif /* FCIP_ESBALLOC */ 2837 2838 fport = fcip_get_port(phandle); 2839 if (fport == NULL) { 2840 return (FC_UNCLAIMED); 2841 } 2842 2843 fptr = fport->fcipp_fcip; 2844 ASSERT(fptr != NULL); 2845 2846 if (fptr == NULL) { 2847 return (FC_UNCLAIMED); 2848 } 2849 2850 mutex_enter(&fptr->fcip_mutex); 2851 if ((fptr->fcip_flags & (FCIP_DETACHING | FCIP_DETACHED)) || 2852 (fptr->fcip_flags & (FCIP_SUSPENDED | FCIP_POWER_DOWN))) { 2853 mutex_exit(&fptr->fcip_mutex); 2854 rval = FC_UNCLAIMED; 2855 goto data_cb_done; 2856 } 2857 2858 /* 2859 * set fcip flags to indicate we are in the middle of a 2860 * data callback so we can wait till the statechange 2861 * is handled before succeeding/failing the SUSPEND/POWER DOWN. 2862 */ 2863 fptr->fcip_flags |= FCIP_IN_DATA_CB; 2864 mutex_exit(&fptr->fcip_mutex); 2865 2866 FCIP_TNF_PROBE_2((fcip_data_cb, "fcip io", /* CSTYLED */, 2867 tnf_string, msg, "data callback", 2868 tnf_int, instance, ddi_get_instance(fport->fcipp_dip))); 2869 FCIP_DEBUG(FCIP_DEBUG_UPSTREAM, 2870 (CE_NOTE, "fcip%d, data callback", 2871 ddi_get_instance(fport->fcipp_dip))); 2872 2873 /* 2874 * get to the network and snap headers in the payload 2875 */ 2876 nhdr = (fcph_network_hdr_t *)buf->ub_buffer; 2877 snaphdr = (llc_snap_hdr_t *)(buf->ub_buffer + 2878 sizeof (fcph_network_hdr_t)); 2879 2880 hdrlen = sizeof (fcph_network_hdr_t) + sizeof (llc_snap_hdr_t); 2881 2882 /* 2883 * get the IP header to obtain the no. of bytes we need to read 2884 * off from the unsol buffer. This obviously is because not all 2885 * data fills up the unsol buffer completely and the firmware 2886 * doesn't tell us how many valid bytes are in there as well 2887 */ 2888 iphdr = (ipha_t *)(buf->ub_buffer + hdrlen); 2889 snaphdr->pid = BE_16(snaphdr->pid); 2890 type = snaphdr->pid; 2891 2892 FCIP_DEBUG(FCIP_DEBUG_UPSTREAM, 2893 (CE_CONT, "SNAPHDR: dsap %x, ssap %x, ctrl %x\n", 2894 snaphdr->dsap, snaphdr->ssap, snaphdr->ctrl)); 2895 2896 FCIP_DEBUG(FCIP_DEBUG_UPSTREAM, 2897 (CE_CONT, "oui[0] 0x%x oui[1] 0x%x oui[2] 0x%x pid 0x%x\n", 2898 snaphdr->oui[0], snaphdr->oui[1], snaphdr->oui[2], snaphdr->pid)); 2899 2900 /* Authneticate, Authenticate */ 2901 if (type == ETHERTYPE_IP) { 2902 len = hdrlen + BE_16(iphdr->ipha_length); 2903 } else if (type == ETHERTYPE_ARP) { 2904 len = hdrlen + 28; 2905 } else { 2906 len = buf->ub_bufsize; 2907 } 2908 2909 FCIP_DEBUG(FCIP_DEBUG_UPSTREAM, 2910 (CE_CONT, "effective packet length is %d bytes.\n", len)); 2911 2912 if (len < hdrlen || len > FCIP_UB_SIZE) { 2913 FCIP_DEBUG(FCIP_DEBUG_UPSTREAM, 2914 (CE_NOTE, "Incorrect buffer size %d bytes", len)); 2915 rval = FC_UNCLAIMED; 2916 goto data_cb_done; 2917 } 2918 2919 if (buf->ub_frame.type != FC_TYPE_IS8802_SNAP) { 2920 FCIP_DEBUG(FCIP_DEBUG_UPSTREAM, (CE_NOTE, "Not IP/ARP data")); 2921 rval = FC_UNCLAIMED; 2922 goto data_cb_done; 2923 } 2924 2925 FCIP_DEBUG(FCIP_DEBUG_UPSTREAM, (CE_NOTE, "checking wwn")); 2926 2927 if ((fcip_wwn_compare(&nhdr->net_dest_addr, &fport->fcipp_pwwn, 2928 FCIP_COMPARE_NWWN) != 0) && 2929 (!IS_BROADCAST_ADDR(&nhdr->net_dest_addr))) { 2930 rval = FC_UNCLAIMED; 2931 goto data_cb_done; 2932 } else if (fcip_cache_on_arp_broadcast && 2933 IS_BROADCAST_ADDR(&nhdr->net_dest_addr)) { 2934 fcip_cache_arp_broadcast(fptr, buf); 2935 } 2936 2937 FCIP_DEBUG(FCIP_DEBUG_UPSTREAM, (CE_NOTE, "Allocate streams block")); 2938 2939 /* 2940 * Using esballoc instead of allocb should be faster, atleast at 2941 * larger MTUs than 1500 bytes. Someday we'll get there :) 2942 */ 2943 #if defined(FCIP_ESBALLOC) 2944 /* 2945 * allocate memory for the frtn function arg. The Function 2946 * (fcip_ubfree) arg is a struct fcip_esballoc_arg type 2947 * which contains pointers to the unsol buffer and the 2948 * opaque port handle for releasing the unsol buffer back to 2949 * the FCA for reuse 2950 */ 2951 fesb_argp = (struct fcip_esballoc_arg *) 2952 kmem_zalloc(sizeof (struct fcip_esballoc_arg), KM_NOSLEEP); 2953 2954 if (fesb_argp == NULL) { 2955 FCIP_DEBUG(FCIP_DEBUG_UPSTREAM, 2956 (CE_WARN, "esballoc of mblk failed in data_cb")); 2957 rval = FC_UNCLAIMED; 2958 goto data_cb_done; 2959 } 2960 /* 2961 * Check with KM_NOSLEEP 2962 */ 2963 free_ubuf = (frtn_t *)kmem_zalloc(sizeof (frtn_t), KM_NOSLEEP); 2964 if (free_ubuf == NULL) { 2965 kmem_free(fesb_argp, sizeof (struct fcip_esballoc_arg)); 2966 FCIP_DEBUG(FCIP_DEBUG_UPSTREAM, 2967 (CE_WARN, "esballoc of mblk failed in data_cb")); 2968 rval = FC_UNCLAIMED; 2969 goto data_cb_done; 2970 } 2971 2972 fesb_argp->frtnp = free_ubuf; 2973 fesb_argp->buf = buf; 2974 fesb_argp->phandle = phandle; 2975 free_ubuf->free_func = fcip_ubfree; 2976 free_ubuf->free_arg = (char *)fesb_argp; 2977 if ((bp = (mblk_t *)esballoc((unsigned char *)buf->ub_buffer, 2978 len, BPRI_MED, free_ubuf)) == NULL) { 2979 kmem_free(fesb_argp, sizeof (struct fcip_esballoc_arg)); 2980 kmem_free(free_ubuf, sizeof (frtn_t)); 2981 FCIP_DEBUG(FCIP_DEBUG_UPSTREAM, 2982 (CE_WARN, "esballoc of mblk failed in data_cb")); 2983 rval = FC_UNCLAIMED; 2984 goto data_cb_done; 2985 } 2986 #elif !defined(FCIP_ESBALLOC) 2987 /* 2988 * allocate streams mblk and copy the contents of the 2989 * unsolicited buffer into this newly alloc'ed mblk 2990 */ 2991 if ((bp = (mblk_t *)fcip_allocb((size_t)len, BPRI_LO)) == NULL) { 2992 FCIP_DEBUG(FCIP_DEBUG_UPSTREAM, 2993 (CE_WARN, "alloc of mblk failed in data_cb")); 2994 rval = FC_UNCLAIMED; 2995 goto data_cb_done; 2996 } 2997 2998 /* 2999 * Unsolicited buffers handed up to us from the FCA must be 3000 * endian clean so just bcopy the data into our mblk. Else 3001 * we may have to either copy the data byte by byte or 3002 * use the ddi_rep_get* routines to do the copy for us. 3003 */ 3004 bcopy(buf->ub_buffer, bp->b_rptr, len); 3005 3006 /* 3007 * for esballoc'ed mblks - free the UB in the frtn function 3008 * along with the memory allocated for the function arg. 3009 * for allocb'ed mblk - release the unsolicited buffer here 3010 */ 3011 (void) fc_ulp_ubrelease(phandle, 1, &buf->ub_token); 3012 3013 #endif /* FCIP_ESBALLOC */ 3014 3015 bp->b_wptr = bp->b_rptr + len; 3016 fptr->fcip_ipackets++; 3017 3018 if (type == ETHERTYPE_IP) { 3019 mutex_enter(&fptr->fcip_mutex); 3020 fptr->fcip_ub_upstream++; 3021 mutex_exit(&fptr->fcip_mutex); 3022 bp->b_rptr += hdrlen; 3023 3024 /* 3025 * Check if ipq is valid in the sendup thread 3026 */ 3027 if (fcip_sendup_alloc_enque(fptr, bp, NULL) != FC_SUCCESS) { 3028 freemsg(bp); 3029 } 3030 } else { 3031 /* 3032 * We won't get ethernet 802.3 packets in FCIP but we may get 3033 * types other than ETHERTYPE_IP, such as ETHERTYPE_ARP. Let 3034 * fcip_sendup() do the matching. 3035 */ 3036 mutex_enter(&fptr->fcip_mutex); 3037 fptr->fcip_ub_upstream++; 3038 mutex_exit(&fptr->fcip_mutex); 3039 if (fcip_sendup_alloc_enque(fptr, bp, 3040 fcip_accept) != FC_SUCCESS) { 3041 freemsg(bp); 3042 } 3043 } 3044 3045 rval = FC_SUCCESS; 3046 3047 /* 3048 * Unset fcip_flags to indicate we are out of callback and return 3049 */ 3050 data_cb_done: 3051 mutex_enter(&fptr->fcip_mutex); 3052 fptr->fcip_flags &= ~(FCIP_IN_DATA_CB); 3053 mutex_exit(&fptr->fcip_mutex); 3054 return (rval); 3055 } 3056 3057 #if !defined(FCIP_ESBALLOC) 3058 /* 3059 * Allocate a message block for the inbound data to be sent upstream. 3060 */ 3061 static void * 3062 fcip_allocb(size_t size, uint_t pri) 3063 { 3064 mblk_t *mp; 3065 3066 if ((mp = allocb(size, pri)) == NULL) { 3067 return (NULL); 3068 } 3069 return (mp); 3070 } 3071 3072 #endif 3073 3074 /* 3075 * This helper routine kmem cache alloc's a sendup element for enquing 3076 * into the sendup list for callbacks upstream from the dedicated sendup 3077 * thread. We enque the msg buf into the sendup list and cv_signal the 3078 * sendup thread to finish the callback for us. 3079 */ 3080 static int 3081 fcip_sendup_alloc_enque(struct fcip *fptr, mblk_t *mp, struct fcipstr *(*f)()) 3082 { 3083 struct fcip_sendup_elem *msg_elem; 3084 int rval = FC_FAILURE; 3085 3086 FCIP_TNF_PROBE_1((fcip_sendup_alloc_enque, "fcip io", /* CSTYLED */, 3087 tnf_string, msg, "sendup msg enque")); 3088 msg_elem = kmem_cache_alloc(fptr->fcip_sendup_cache, KM_NOSLEEP); 3089 if (msg_elem == NULL) { 3090 /* drop pkt to floor - update stats */ 3091 rval = FC_FAILURE; 3092 goto sendup_alloc_done; 3093 } 3094 msg_elem->fcipsu_mp = mp; 3095 msg_elem->fcipsu_func = f; 3096 3097 mutex_enter(&fptr->fcip_sendup_mutex); 3098 if (fptr->fcip_sendup_head == NULL) { 3099 fptr->fcip_sendup_head = fptr->fcip_sendup_tail = msg_elem; 3100 } else { 3101 fptr->fcip_sendup_tail->fcipsu_next = msg_elem; 3102 fptr->fcip_sendup_tail = msg_elem; 3103 } 3104 fptr->fcip_sendup_cnt++; 3105 cv_signal(&fptr->fcip_sendup_cv); 3106 mutex_exit(&fptr->fcip_sendup_mutex); 3107 rval = FC_SUCCESS; 3108 3109 sendup_alloc_done: 3110 return (rval); 3111 } 3112 3113 /* 3114 * One of the ways of performing the WWN to D_ID mapping required for 3115 * IPFC data is to cache the unsolicited ARP broadcast messages received 3116 * and update the routing table to add entry for the destination port 3117 * if we are the intended recipient of the ARP broadcast message. This is 3118 * one of the methods recommended in the rfc to obtain the WWN to D_ID mapping 3119 * but is not typically used unless enabled. The driver prefers to use the 3120 * nameserver/lilp map to obtain this mapping. 3121 */ 3122 static void 3123 fcip_cache_arp_broadcast(struct fcip *fptr, fc_unsol_buf_t *buf) 3124 { 3125 fcip_port_info_t *fport; 3126 fcph_network_hdr_t *nhdr; 3127 struct fcip_routing_table *frp; 3128 fc_portmap_t map; 3129 3130 fport = fptr->fcip_port_info; 3131 if (fport == NULL) { 3132 return; 3133 } 3134 ASSERT(fport != NULL); 3135 3136 nhdr = (fcph_network_hdr_t *)buf->ub_buffer; 3137 3138 mutex_enter(&fptr->fcip_rt_mutex); 3139 frp = fcip_lookup_rtable(fptr, &nhdr->net_src_addr, FCIP_COMPARE_NWWN); 3140 mutex_exit(&fptr->fcip_rt_mutex); 3141 if (frp == NULL) { 3142 map.map_did.port_id = buf->ub_frame.s_id; 3143 map.map_hard_addr.hard_addr = buf->ub_frame.s_id; 3144 map.map_state = PORT_DEVICE_VALID; 3145 map.map_type = PORT_DEVICE_NEW; 3146 map.map_flags = 0; 3147 map.map_pd = NULL; 3148 bcopy((void *)&nhdr->net_src_addr, (void *)&map.map_pwwn, 3149 sizeof (la_wwn_t)); 3150 bcopy((void *)&nhdr->net_src_addr, (void *)&map.map_nwwn, 3151 sizeof (la_wwn_t)); 3152 fcip_rt_update(fptr, &map, 1); 3153 mutex_enter(&fptr->fcip_rt_mutex); 3154 frp = fcip_lookup_rtable(fptr, &nhdr->net_src_addr, 3155 FCIP_COMPARE_NWWN); 3156 mutex_exit(&fptr->fcip_rt_mutex); 3157 3158 (void) fcip_add_dest(fptr, frp); 3159 } 3160 3161 } 3162 3163 /* 3164 * This is a dedicated thread to do callbacks from fcip's data callback 3165 * routines into the modules upstream. The reason for this thread is 3166 * the data callback function can be called from an interrupt context and 3167 * the upstream modules *can* make calls downstream in the same thread 3168 * context. If the call is to a fabric port which is not yet in our 3169 * routing tables, we may have to query the nameserver/fabric for the 3170 * MAC addr to Port_ID mapping which may be blocking calls. 3171 */ 3172 static void 3173 fcip_sendup_thr(void *arg) 3174 { 3175 struct fcip *fptr = (struct fcip *)arg; 3176 struct fcip_sendup_elem *msg_elem; 3177 queue_t *ip4q = NULL; 3178 3179 CALLB_CPR_INIT(&fptr->fcip_cpr_info, &fptr->fcip_sendup_mutex, 3180 callb_generic_cpr, "fcip_sendup_thr"); 3181 3182 mutex_enter(&fptr->fcip_sendup_mutex); 3183 for (;;) { 3184 3185 while (fptr->fcip_sendup_thr_initted && 3186 fptr->fcip_sendup_head == NULL) { 3187 CALLB_CPR_SAFE_BEGIN(&fptr->fcip_cpr_info); 3188 cv_wait(&fptr->fcip_sendup_cv, 3189 &fptr->fcip_sendup_mutex); 3190 CALLB_CPR_SAFE_END(&fptr->fcip_cpr_info, 3191 &fptr->fcip_sendup_mutex); 3192 } 3193 3194 if (fptr->fcip_sendup_thr_initted == 0) { 3195 break; 3196 } 3197 3198 FCIP_TNF_PROBE_1((fcip_sendup_thr, "fcip io", /* CSTYLED */, 3199 tnf_string, msg, "fcip sendup thr - new msg")); 3200 3201 msg_elem = fptr->fcip_sendup_head; 3202 fptr->fcip_sendup_head = msg_elem->fcipsu_next; 3203 msg_elem->fcipsu_next = NULL; 3204 mutex_exit(&fptr->fcip_sendup_mutex); 3205 3206 if (msg_elem->fcipsu_func == NULL) { 3207 /* 3208 * Message for ipq. Check to see if the ipq is 3209 * is still valid. Since the thread is asynchronous, 3210 * there could have been a close on the stream 3211 */ 3212 mutex_enter(&fptr->fcip_mutex); 3213 if (fptr->fcip_ipq && canputnext(fptr->fcip_ipq)) { 3214 ip4q = fptr->fcip_ipq; 3215 mutex_exit(&fptr->fcip_mutex); 3216 putnext(ip4q, msg_elem->fcipsu_mp); 3217 } else { 3218 mutex_exit(&fptr->fcip_mutex); 3219 freemsg(msg_elem->fcipsu_mp); 3220 } 3221 } else { 3222 fcip_sendup(fptr, msg_elem->fcipsu_mp, 3223 msg_elem->fcipsu_func); 3224 } 3225 3226 #if !defined(FCIP_ESBALLOC) 3227 /* 3228 * for allocb'ed mblk - decrement upstream count here 3229 */ 3230 mutex_enter(&fptr->fcip_mutex); 3231 ASSERT(fptr->fcip_ub_upstream > 0); 3232 fptr->fcip_ub_upstream--; 3233 mutex_exit(&fptr->fcip_mutex); 3234 #endif /* FCIP_ESBALLOC */ 3235 3236 kmem_cache_free(fptr->fcip_sendup_cache, (void *)msg_elem); 3237 mutex_enter(&fptr->fcip_sendup_mutex); 3238 fptr->fcip_sendup_cnt--; 3239 } 3240 3241 3242 #ifndef __lock_lint 3243 CALLB_CPR_EXIT(&fptr->fcip_cpr_info); 3244 #else 3245 mutex_exit(&fptr->fcip_sendup_mutex); 3246 #endif /* __lock_lint */ 3247 3248 /* Wake up fcip detach thread by the end */ 3249 cv_signal(&fptr->fcip_sendup_cv); 3250 3251 thread_exit(); 3252 } 3253 3254 #ifdef FCIP_ESBALLOC 3255 3256 /* 3257 * called from the stream head when it is done using an unsolicited buffer. 3258 * We release this buffer then to the FCA for reuse. 3259 */ 3260 static void 3261 fcip_ubfree(char *arg) 3262 { 3263 struct fcip_esballoc_arg *fesb_argp = (struct fcip_esballoc_arg *)arg; 3264 fc_unsol_buf_t *ubuf; 3265 frtn_t *frtnp; 3266 fcip_port_info_t *fport; 3267 struct fcip *fptr; 3268 3269 3270 fport = fcip_get_port(fesb_argp->phandle); 3271 fptr = fport->fcipp_fcip; 3272 3273 ASSERT(fesb_argp != NULL); 3274 ubuf = fesb_argp->buf; 3275 frtnp = fesb_argp->frtnp; 3276 3277 3278 FCIP_DEBUG(FCIP_DEBUG_UPSTREAM, 3279 (CE_WARN, "freeing ubuf after esballoc in fcip_ubfree")); 3280 (void) fc_ulp_ubrelease(fesb_argp->phandle, 1, &ubuf->ub_token); 3281 3282 mutex_enter(&fptr->fcip_mutex); 3283 ASSERT(fptr->fcip_ub_upstream > 0); 3284 fptr->fcip_ub_upstream--; 3285 cv_signal(&fptr->fcip_ub_cv); 3286 mutex_exit(&fptr->fcip_mutex); 3287 3288 kmem_free(frtnp, sizeof (frtn_t)); 3289 kmem_free(fesb_argp, sizeof (struct fcip_esballoc_arg)); 3290 } 3291 3292 #endif /* FCIP_ESBALLOC */ 3293 3294 /* 3295 * handle data other than that of type ETHERTYPE_IP and send it on its 3296 * way upstream to the right streams module to handle 3297 */ 3298 static void 3299 fcip_sendup(struct fcip *fptr, mblk_t *mp, struct fcipstr *(*acceptfunc)()) 3300 { 3301 struct fcipstr *slp, *nslp; 3302 la_wwn_t *dhostp; 3303 mblk_t *nmp; 3304 uint32_t isgroupaddr; 3305 int type; 3306 uint32_t hdrlen; 3307 fcph_network_hdr_t *nhdr; 3308 llc_snap_hdr_t *snaphdr; 3309 3310 FCIP_TNF_PROBE_1((fcip_sendup, "fcip io", /* CSTYLED */, 3311 tnf_string, msg, "fcip sendup")); 3312 nhdr = (fcph_network_hdr_t *)mp->b_rptr; 3313 snaphdr = 3314 (llc_snap_hdr_t *)(mp->b_rptr + sizeof (fcph_network_hdr_t)); 3315 dhostp = &nhdr->net_dest_addr; 3316 type = snaphdr->pid; 3317 hdrlen = sizeof (fcph_network_hdr_t) + sizeof (llc_snap_hdr_t); 3318 3319 /* No group address with fibre channel */ 3320 isgroupaddr = 0; 3321 3322 /* 3323 * While holding a reader lock on the linked list of streams structures, 3324 * attempt to match the address criteria for each stream 3325 * and pass up the raw M_DATA ("fastpath") or a DL_UNITDATA_IND. 3326 */ 3327 3328 rw_enter(&fcipstruplock, RW_READER); 3329 3330 if ((slp = (*acceptfunc)(fcipstrup, fptr, type, dhostp)) == NULL) { 3331 rw_exit(&fcipstruplock); 3332 FCIP_TNF_PROBE_1((fcip_sendup, "fcip io", /* CSTYLED */, 3333 tnf_string, msg, "fcip sendup - no slp")); 3334 freemsg(mp); 3335 return; 3336 } 3337 3338 /* 3339 * Loop on matching open streams until (*acceptfunc)() returns NULL. 3340 */ 3341 for (; nslp = (*acceptfunc)(slp->sl_nextp, fptr, type, dhostp); 3342 slp = nslp) { 3343 if (canputnext(slp->sl_rq)) { 3344 if (nmp = dupmsg(mp)) { 3345 if ((slp->sl_flags & FCIP_SLFAST) && 3346 !isgroupaddr) { 3347 nmp->b_rptr += hdrlen; 3348 putnext(slp->sl_rq, nmp); 3349 } else if (slp->sl_flags & FCIP_SLRAW) { 3350 /* No headers when FCIP_SLRAW is set */ 3351 putnext(slp->sl_rq, nmp); 3352 } else if ((nmp = fcip_addudind(fptr, nmp, 3353 nhdr, type))) { 3354 putnext(slp->sl_rq, nmp); 3355 } 3356 } 3357 } 3358 } 3359 3360 /* 3361 * Do the last one. 3362 */ 3363 if (canputnext(slp->sl_rq)) { 3364 if (slp->sl_flags & FCIP_SLFAST) { 3365 mp->b_rptr += hdrlen; 3366 putnext(slp->sl_rq, mp); 3367 } else if (slp->sl_flags & FCIP_SLRAW) { 3368 putnext(slp->sl_rq, mp); 3369 } else if ((mp = fcip_addudind(fptr, mp, nhdr, type))) { 3370 putnext(slp->sl_rq, mp); 3371 } 3372 } else { 3373 freemsg(mp); 3374 } 3375 FCIP_TNF_PROBE_1((fcip_sendup, "fcip io", /* CSTYLED */, 3376 tnf_string, msg, "fcip sendup done")); 3377 3378 rw_exit(&fcipstruplock); 3379 } 3380 3381 /* 3382 * Match the stream based on type and wwn if necessary. 3383 * Destination wwn dhostp is passed to this routine is reserved 3384 * for future usage. We don't need to use it right now since port 3385 * to fcip instance mapping is unique and wwn is already validated when 3386 * packet comes to fcip. 3387 */ 3388 /* ARGSUSED */ 3389 static struct fcipstr * 3390 fcip_accept(struct fcipstr *slp, struct fcip *fptr, int type, la_wwn_t *dhostp) 3391 { 3392 t_uscalar_t sap; 3393 3394 FCIP_TNF_PROBE_1((fcip_accept, "fcip io", /* CSTYLED */, 3395 tnf_string, msg, "fcip accept")); 3396 3397 for (; slp; slp = slp->sl_nextp) { 3398 sap = slp->sl_sap; 3399 FCIP_DEBUG(FCIP_DEBUG_UPSTREAM, (CE_CONT, 3400 "fcip_accept: checking next sap = %x, type = %x", 3401 sap, type)); 3402 3403 if ((slp->sl_fcip == fptr) && (type == sap)) { 3404 return (slp); 3405 } 3406 } 3407 return (NULL); 3408 } 3409 3410 /* 3411 * Handle DL_UNITDATA_IND messages 3412 */ 3413 static mblk_t * 3414 fcip_addudind(struct fcip *fptr, mblk_t *mp, fcph_network_hdr_t *nhdr, 3415 int type) 3416 { 3417 dl_unitdata_ind_t *dludindp; 3418 struct fcipdladdr *dlap; 3419 mblk_t *nmp; 3420 int size; 3421 uint32_t hdrlen; 3422 struct ether_addr src_addr; 3423 struct ether_addr dest_addr; 3424 3425 3426 hdrlen = (sizeof (llc_snap_hdr_t) + sizeof (fcph_network_hdr_t)); 3427 mp->b_rptr += hdrlen; 3428 3429 FCIP_TNF_PROBE_1((fcip_addudind, "fcip io", /* CSTYLED */, 3430 tnf_string, msg, "fcip addudind")); 3431 3432 /* 3433 * Allocate an M_PROTO mblk for the DL_UNITDATA_IND. 3434 */ 3435 size = sizeof (dl_unitdata_ind_t) + FCIPADDRL + FCIPADDRL; 3436 if ((nmp = allocb(size, BPRI_LO)) == NULL) { 3437 fptr->fcip_allocbfail++; 3438 freemsg(mp); 3439 return (NULL); 3440 } 3441 DB_TYPE(nmp) = M_PROTO; 3442 nmp->b_wptr = nmp->b_datap->db_lim; 3443 nmp->b_rptr = nmp->b_wptr - size; 3444 3445 /* 3446 * Construct a DL_UNITDATA_IND primitive. 3447 */ 3448 dludindp = (dl_unitdata_ind_t *)nmp->b_rptr; 3449 dludindp->dl_primitive = DL_UNITDATA_IND; 3450 dludindp->dl_dest_addr_length = FCIPADDRL; 3451 dludindp->dl_dest_addr_offset = sizeof (dl_unitdata_ind_t); 3452 dludindp->dl_src_addr_length = FCIPADDRL; 3453 dludindp->dl_src_addr_offset = sizeof (dl_unitdata_ind_t) + FCIPADDRL; 3454 dludindp->dl_group_address = 0; /* not DL_MULTI */ 3455 3456 dlap = (struct fcipdladdr *)(nmp->b_rptr + sizeof (dl_unitdata_ind_t)); 3457 wwn_to_ether(&nhdr->net_dest_addr, &dest_addr); 3458 ether_bcopy(&dest_addr, &dlap->dl_phys); 3459 dlap->dl_sap = (uint16_t)type; 3460 3461 dlap = (struct fcipdladdr *)(nmp->b_rptr + sizeof (dl_unitdata_ind_t) 3462 + FCIPADDRL); 3463 wwn_to_ether(&nhdr->net_src_addr, &src_addr); 3464 ether_bcopy(&src_addr, &dlap->dl_phys); 3465 dlap->dl_sap = (uint16_t)type; 3466 3467 /* 3468 * Link the M_PROTO and M_DATA together. 3469 */ 3470 nmp->b_cont = mp; 3471 return (nmp); 3472 } 3473 3474 3475 /* 3476 * The open routine. For clone opens, we return the next available minor 3477 * no. for the stream to use 3478 */ 3479 /* ARGSUSED */ 3480 static int 3481 fcip_open(queue_t *rq, dev_t *devp, int flag, int sflag, cred_t *credp) 3482 { 3483 struct fcipstr *slp; 3484 struct fcipstr **prevslp; 3485 minor_t minor; 3486 3487 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, (CE_NOTE, "in fcip_open")); 3488 FCIP_TNF_PROBE_1((fcip_open, "fcip io", /* CSTYLED */, 3489 tnf_string, msg, "enter")); 3490 /* 3491 * We need to ensure that the port driver is loaded before 3492 * we proceed 3493 */ 3494 if (ddi_hold_installed_driver(ddi_name_to_major(PORT_DRIVER)) == NULL) { 3495 /* no port driver instances found */ 3496 FCIP_DEBUG(FCIP_DEBUG_STARTUP, (CE_WARN, 3497 "!ddi_hold_installed_driver of fp failed\n")); 3498 return (ENXIO); 3499 } 3500 /* serialize opens */ 3501 rw_enter(&fcipstruplock, RW_WRITER); 3502 3503 prevslp = &fcipstrup; 3504 if (sflag == CLONEOPEN) { 3505 minor = 0; 3506 for (; (slp = *prevslp) != NULL; prevslp = &slp->sl_nextp) { 3507 if (minor < slp->sl_minor) { 3508 break; 3509 } 3510 minor ++; 3511 } 3512 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, (CE_NOTE, 3513 "getmajor returns 0x%x", getmajor(*devp))); 3514 *devp = makedevice(getmajor(*devp), minor); 3515 } else { 3516 minor = getminor(*devp); 3517 } 3518 3519 /* 3520 * check if our qp's private area is already initialized. If yes 3521 * the stream is already open - just return 3522 */ 3523 if (rq->q_ptr) { 3524 goto done; 3525 } 3526 3527 slp = GETSTRUCT(struct fcipstr, 1); 3528 slp->sl_minor = minor; 3529 slp->sl_rq = rq; 3530 slp->sl_sap = 0; 3531 slp->sl_flags = 0; 3532 slp->sl_state = DL_UNATTACHED; 3533 slp->sl_fcip = NULL; 3534 3535 mutex_init(&slp->sl_lock, NULL, MUTEX_DRIVER, NULL); 3536 3537 /* 3538 * link this new stream entry into list of active streams 3539 */ 3540 slp->sl_nextp = *prevslp; 3541 *prevslp = slp; 3542 3543 rq->q_ptr = WR(rq)->q_ptr = (char *)slp; 3544 3545 /* 3546 * Disable automatic enabling of our write service procedures 3547 * we need to control this explicitly. This will prevent 3548 * anyone scheduling of our write service procedures. 3549 */ 3550 noenable(WR(rq)); 3551 3552 done: 3553 rw_exit(&fcipstruplock); 3554 /* 3555 * enable our put and service routines on the read side 3556 */ 3557 qprocson(rq); 3558 3559 /* 3560 * There is only one instance of fcip (instance = 0) 3561 * for multiple instances of hardware 3562 */ 3563 (void) qassociate(rq, 0); /* don't allow drcompat to be pushed */ 3564 return (0); 3565 } 3566 3567 /* 3568 * close an opened stream. The minor no. will then be available for 3569 * future opens. 3570 */ 3571 /* ARGSUSED */ 3572 static int 3573 fcip_close(queue_t *rq, int flag, int otyp, cred_t *credp) 3574 { 3575 struct fcipstr *slp; 3576 struct fcipstr **prevslp; 3577 3578 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, (CE_NOTE, "in fcip_close")); 3579 FCIP_TNF_PROBE_1((fcip_close, "fcip io", /* CSTYLED */, 3580 tnf_string, msg, "enter")); 3581 ASSERT(rq); 3582 /* we should also have the active stream pointer in q_ptr */ 3583 ASSERT(rq->q_ptr); 3584 3585 ddi_rele_driver(ddi_name_to_major(PORT_DRIVER)); 3586 /* 3587 * disable our put and service procedures. We had enabled them 3588 * on open 3589 */ 3590 qprocsoff(rq); 3591 slp = (struct fcipstr *)rq->q_ptr; 3592 3593 /* 3594 * Implicitly detach stream a stream from an interface. 3595 */ 3596 if (slp->sl_fcip) { 3597 fcip_dodetach(slp); 3598 } 3599 3600 (void) qassociate(rq, -1); /* undo association in open */ 3601 3602 rw_enter(&fcipstruplock, RW_WRITER); 3603 3604 /* 3605 * unlink this stream from the active stream list and free it 3606 */ 3607 for (prevslp = &fcipstrup; (slp = *prevslp) != NULL; 3608 prevslp = &slp->sl_nextp) { 3609 if (slp == (struct fcipstr *)rq->q_ptr) { 3610 break; 3611 } 3612 } 3613 3614 /* we should have found slp */ 3615 ASSERT(slp); 3616 3617 *prevslp = slp->sl_nextp; 3618 mutex_destroy(&slp->sl_lock); 3619 kmem_free(slp, sizeof (struct fcipstr)); 3620 rq->q_ptr = WR(rq)->q_ptr = NULL; 3621 3622 rw_exit(&fcipstruplock); 3623 return (0); 3624 } 3625 3626 /* 3627 * This is not an extension of the DDI_DETACH request. This routine 3628 * only detaches a stream from an interface 3629 */ 3630 static void 3631 fcip_dodetach(struct fcipstr *slp) 3632 { 3633 struct fcipstr *tslp; 3634 struct fcip *fptr; 3635 3636 FCIP_DEBUG(FCIP_DEBUG_DETACH, (CE_NOTE, "in fcip_dodetach")); 3637 FCIP_TNF_PROBE_1((fcip_dodetach, "fcip io", /* CSTYLED */, 3638 tnf_string, msg, "enter")); 3639 ASSERT(slp->sl_fcip != NULL); 3640 3641 fptr = slp->sl_fcip; 3642 slp->sl_fcip = NULL; 3643 3644 /* 3645 * we don't support promiscuous mode currently but check 3646 * for and disable any promiscuous mode operation 3647 */ 3648 if (slp->sl_flags & SLALLPHYS) { 3649 slp->sl_flags &= ~SLALLPHYS; 3650 } 3651 3652 /* 3653 * disable ALLMULTI mode if all mulitcast addr are ON 3654 */ 3655 if (slp->sl_flags & SLALLMULTI) { 3656 slp->sl_flags &= ~SLALLMULTI; 3657 } 3658 3659 /* 3660 * we are most likely going to perform multicast by 3661 * broadcasting to the well known addr (D_ID) 0xFFFFFF or 3662 * ALPA 0x00 in case of public loops 3663 */ 3664 3665 3666 /* 3667 * detach unit from device structure. 3668 */ 3669 for (tslp = fcipstrup; tslp != NULL; tslp = tslp->sl_nextp) { 3670 if (tslp->sl_fcip == fptr) { 3671 break; 3672 } 3673 } 3674 if (tslp == NULL) { 3675 FCIP_DEBUG(FCIP_DEBUG_DETACH, (CE_WARN, 3676 "fcip_dodeatch - active stream struct not found")); 3677 3678 /* unregister with Fabric nameserver?? */ 3679 } 3680 slp->sl_state = DL_UNATTACHED; 3681 3682 fcip_setipq(fptr); 3683 } 3684 3685 3686 /* 3687 * Set or clear device ipq pointer. 3688 * Walk thru all the streams on this device, if a ETHERTYPE_IP 3689 * stream is found, assign device ipq to its sl_rq. 3690 */ 3691 static void 3692 fcip_setipq(struct fcip *fptr) 3693 { 3694 struct fcipstr *slp; 3695 int ok = 1; 3696 queue_t *ipq = NULL; 3697 3698 FCIP_DEBUG(FCIP_DEBUG_INIT, (CE_NOTE, "entered fcip_setipq")); 3699 3700 rw_enter(&fcipstruplock, RW_READER); 3701 3702 for (slp = fcipstrup; slp != NULL; slp = slp->sl_nextp) { 3703 if (slp->sl_fcip == fptr) { 3704 if (slp->sl_flags & (SLALLPHYS|SLALLSAP)) { 3705 ok = 0; 3706 } 3707 if (slp->sl_sap == ETHERTYPE_IP) { 3708 if (ipq == NULL) { 3709 ipq = slp->sl_rq; 3710 } else { 3711 ok = 0; 3712 } 3713 } 3714 } 3715 } 3716 3717 rw_exit(&fcipstruplock); 3718 3719 if (fcip_check_port_exists(fptr)) { 3720 /* fptr passed to us is stale */ 3721 return; 3722 } 3723 3724 mutex_enter(&fptr->fcip_mutex); 3725 if (ok) { 3726 fptr->fcip_ipq = ipq; 3727 } else { 3728 fptr->fcip_ipq = NULL; 3729 } 3730 mutex_exit(&fptr->fcip_mutex); 3731 } 3732 3733 3734 /* ARGSUSED */ 3735 static void 3736 fcip_ioctl(queue_t *wq, mblk_t *mp) 3737 { 3738 struct iocblk *iocp = (struct iocblk *)mp->b_rptr; 3739 struct fcipstr *slp = (struct fcipstr *)wq->q_ptr; 3740 3741 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 3742 (CE_NOTE, "in fcip ioctl : %d", iocp->ioc_cmd)); 3743 FCIP_TNF_PROBE_1((fcip_ioctl, "fcip io", /* CSTYLED */, 3744 tnf_string, msg, "enter")); 3745 3746 switch (iocp->ioc_cmd) { 3747 case DLIOCRAW: 3748 slp->sl_flags |= FCIP_SLRAW; 3749 miocack(wq, mp, 0, 0); 3750 break; 3751 3752 case DL_IOC_HDR_INFO: 3753 fcip_dl_ioc_hdr_info(wq, mp); 3754 break; 3755 3756 default: 3757 miocnak(wq, mp, 0, EINVAL); 3758 break; 3759 } 3760 } 3761 3762 /* 3763 * The streams 'Put' routine. 3764 */ 3765 /* ARGSUSED */ 3766 static int 3767 fcip_wput(queue_t *wq, mblk_t *mp) 3768 { 3769 struct fcipstr *slp = (struct fcipstr *)wq->q_ptr; 3770 struct fcip *fptr; 3771 struct fcip_dest *fdestp; 3772 fcph_network_hdr_t *headerp; 3773 3774 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 3775 (CE_NOTE, "in fcip_wput :: type:%x", DB_TYPE(mp))); 3776 3777 switch (DB_TYPE(mp)) { 3778 case M_DATA: { 3779 3780 fptr = slp->sl_fcip; 3781 3782 if (((slp->sl_flags & (FCIP_SLFAST|FCIP_SLRAW)) == 0) || 3783 (slp->sl_state != DL_IDLE) || 3784 (fptr == NULL)) { 3785 /* 3786 * set error in the message block and send a reply 3787 * back upstream. Sun's merror routine does this 3788 * for us more cleanly. 3789 */ 3790 merror(wq, mp, EPROTO); 3791 break; 3792 } 3793 3794 /* 3795 * if any messages are already enqueued or if the interface 3796 * is in promiscuous mode, causing the packets to loop back 3797 * up, then enqueue the message. Otherwise just transmit 3798 * the message. putq() puts the message on fcip's 3799 * write queue and qenable() puts the queue (wq) on 3800 * the list of queues to be called by the streams scheduler. 3801 */ 3802 if (wq->q_first) { 3803 (void) putq(wq, mp); 3804 fptr->fcip_wantw = 1; 3805 qenable(wq); 3806 } else if (fptr->fcip_flags & FCIP_PROMISC) { 3807 /* 3808 * Promiscous mode not supported but add this code in 3809 * case it will be supported in future. 3810 */ 3811 (void) putq(wq, mp); 3812 qenable(wq); 3813 } else { 3814 3815 headerp = (fcph_network_hdr_t *)mp->b_rptr; 3816 fdestp = fcip_get_dest(fptr, &headerp->net_dest_addr); 3817 3818 if (fdestp == NULL) { 3819 merror(wq, mp, EPROTO); 3820 break; 3821 } 3822 3823 ASSERT(fdestp != NULL); 3824 3825 (void) fcip_start(wq, mp, fptr, fdestp, KM_SLEEP); 3826 } 3827 break; 3828 } 3829 case M_PROTO: 3830 case M_PCPROTO: 3831 /* 3832 * to prevent recursive calls into fcip_proto 3833 * (PROTO and PCPROTO messages are handled by fcip_proto) 3834 * let the service procedure handle these messages by 3835 * calling putq here. 3836 */ 3837 (void) putq(wq, mp); 3838 qenable(wq); 3839 break; 3840 3841 case M_IOCTL: 3842 fcip_ioctl(wq, mp); 3843 break; 3844 3845 case M_FLUSH: 3846 if (*mp->b_rptr & FLUSHW) { 3847 flushq(wq, FLUSHALL); 3848 *mp->b_rptr &= ~FLUSHW; 3849 } 3850 /* 3851 * we have both FLUSHW and FLUSHR set with FLUSHRW 3852 */ 3853 if (*mp->b_rptr & FLUSHR) { 3854 /* 3855 * send msg back upstream. qreply() takes care 3856 * of using the RD(wq) queue on its reply 3857 */ 3858 qreply(wq, mp); 3859 } else { 3860 freemsg(mp); 3861 } 3862 break; 3863 3864 default: 3865 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 3866 (CE_NOTE, "default msg type: %x", DB_TYPE(mp))); 3867 freemsg(mp); 3868 break; 3869 } 3870 return (0); 3871 } 3872 3873 3874 /* 3875 * Handle M_PROTO and M_PCPROTO messages 3876 */ 3877 /* ARGSUSED */ 3878 static void 3879 fcip_proto(queue_t *wq, mblk_t *mp) 3880 { 3881 union DL_primitives *dlp; 3882 struct fcipstr *slp; 3883 t_uscalar_t prim; 3884 3885 slp = (struct fcipstr *)wq->q_ptr; 3886 dlp = (union DL_primitives *)mp->b_rptr; 3887 prim = dlp->dl_primitive; /* the DLPI command */ 3888 3889 FCIP_TNF_PROBE_5((fcip_proto, "fcip io", /* CSTYLED */, 3890 tnf_string, msg, "enter", 3891 tnf_opaque, wq, wq, 3892 tnf_opaque, mp, mp, 3893 tnf_opaque, MP_DB_TYPE, DB_TYPE(mp), 3894 tnf_opaque, dl_primitive, dlp->dl_primitive)); 3895 3896 FCIP_DEBUG(FCIP_DEBUG_INIT, (CE_NOTE, "dl_primitve : %x", prim)); 3897 3898 mutex_enter(&slp->sl_lock); 3899 3900 switch (prim) { 3901 case DL_UNITDATA_REQ: 3902 FCIP_TNF_PROBE_1((fcip_proto, "fcip io", /* CSTYLED */, 3903 tnf_string, msg, "unit data request")); 3904 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_NOTE, "unit data request")); 3905 fcip_udreq(wq, mp); 3906 break; 3907 3908 case DL_ATTACH_REQ: 3909 FCIP_TNF_PROBE_1((fcip_proto, "fcip io", /* CSTYLED */, 3910 tnf_string, msg, "Attach request")); 3911 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_NOTE, "Attach request")); 3912 fcip_areq(wq, mp); 3913 break; 3914 3915 case DL_DETACH_REQ: 3916 FCIP_TNF_PROBE_1((fcip_proto, "fcip io", /* CSTYLED */, 3917 tnf_string, msg, "Detach request")); 3918 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_NOTE, "Detach request")); 3919 fcip_dreq(wq, mp); 3920 break; 3921 3922 case DL_BIND_REQ: 3923 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_NOTE, "Bind request")); 3924 FCIP_TNF_PROBE_1((fcip_proto, "fcip io", /* CSTYLED */, 3925 tnf_string, msg, "Bind request")); 3926 fcip_breq(wq, mp); 3927 break; 3928 3929 case DL_UNBIND_REQ: 3930 FCIP_TNF_PROBE_1((fcip_proto, "fcip io", /* CSTYLED */, 3931 tnf_string, msg, "unbind request")); 3932 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_NOTE, "unbind request")); 3933 fcip_ubreq(wq, mp); 3934 break; 3935 3936 case DL_INFO_REQ: 3937 FCIP_TNF_PROBE_1((fcip_proto, "fcip io", /* CSTYLED */, 3938 tnf_string, msg, "Info request")); 3939 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_NOTE, "Info request")); 3940 fcip_ireq(wq, mp); 3941 break; 3942 3943 case DL_SET_PHYS_ADDR_REQ: 3944 FCIP_TNF_PROBE_1((fcip_proto, "fcip io", /* CSTYLED */, 3945 tnf_string, msg, "set phy addr request")); 3946 FCIP_DEBUG(FCIP_DEBUG_DLPI, 3947 (CE_NOTE, "set phy addr request")); 3948 fcip_spareq(wq, mp); 3949 break; 3950 3951 case DL_PHYS_ADDR_REQ: 3952 FCIP_TNF_PROBE_1((fcip_proto, "fcip io", /* CSTYLED */, 3953 tnf_string, msg, "phy addr request")); 3954 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_NOTE, "phy addr request")); 3955 fcip_pareq(wq, mp); 3956 break; 3957 3958 case DL_ENABMULTI_REQ: 3959 FCIP_TNF_PROBE_1((fcip_proto, "fcip io", /* CSTYLED */, 3960 tnf_string, msg, "Enable Multicast request")); 3961 FCIP_DEBUG(FCIP_DEBUG_DLPI, 3962 (CE_NOTE, "Enable Multicast request")); 3963 dlerrorack(wq, mp, prim, DL_UNSUPPORTED, 0); 3964 break; 3965 3966 case DL_DISABMULTI_REQ: 3967 FCIP_TNF_PROBE_1((fcip_proto, "fcip io", /* CSTYLED */, 3968 tnf_string, msg, "Disable Multicast request")); 3969 FCIP_DEBUG(FCIP_DEBUG_DLPI, 3970 (CE_NOTE, "Disable Multicast request")); 3971 dlerrorack(wq, mp, prim, DL_UNSUPPORTED, 0); 3972 break; 3973 3974 case DL_PROMISCON_REQ: 3975 FCIP_TNF_PROBE_1((fcip_proto, "fcip io", /* CSTYLED */, 3976 tnf_string, msg, "Promiscuous mode ON request")); 3977 FCIP_DEBUG(FCIP_DEBUG_DLPI, 3978 (CE_NOTE, "Promiscuous mode ON request")); 3979 dlerrorack(wq, mp, prim, DL_UNSUPPORTED, 0); 3980 break; 3981 3982 case DL_PROMISCOFF_REQ: 3983 FCIP_TNF_PROBE_1((fcip_proto, "fcip io", /* CSTYLED */, 3984 tnf_string, msg, "Promiscuous mode OFF request")); 3985 FCIP_DEBUG(FCIP_DEBUG_DLPI, 3986 (CE_NOTE, "Promiscuous mode OFF request")); 3987 dlerrorack(wq, mp, prim, DL_UNSUPPORTED, 0); 3988 break; 3989 3990 default: 3991 FCIP_TNF_PROBE_1((fcip_proto, "fcip io", /* CSTYLED */, 3992 tnf_string, msg, "Unsupported request")); 3993 dlerrorack(wq, mp, prim, DL_UNSUPPORTED, 0); 3994 break; 3995 } 3996 mutex_exit(&slp->sl_lock); 3997 } 3998 3999 /* 4000 * Always enqueue M_PROTO and M_PCPROTO messages pn the wq and M_DATA 4001 * messages sometimes. Processing of M_PROTO and M_PCPROTO messages 4002 * require us to hold fcip's internal locks across (upstream) putnext 4003 * calls. Specifically fcip_intr could hold fcip_intrlock and fcipstruplock 4004 * when it calls putnext(). That thread could loop back around to call 4005 * fcip_wput and eventually fcip_init() to cause a recursive mutex panic 4006 * 4007 * M_DATA messages are enqueued only if we are out of xmit resources. Once 4008 * the transmit resources are available the service procedure is enabled 4009 * and an attempt is made to xmit all messages on the wq. 4010 */ 4011 /* ARGSUSED */ 4012 static int 4013 fcip_wsrv(queue_t *wq) 4014 { 4015 mblk_t *mp; 4016 struct fcipstr *slp; 4017 struct fcip *fptr; 4018 struct fcip_dest *fdestp; 4019 fcph_network_hdr_t *headerp; 4020 4021 slp = (struct fcipstr *)wq->q_ptr; 4022 fptr = slp->sl_fcip; 4023 4024 FCIP_TNF_PROBE_2((fcip_wsrv, "fcip io", /* CSTYLED */, 4025 tnf_string, msg, "enter", 4026 tnf_opaque, wq, wq)); 4027 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, (CE_NOTE, "fcip wsrv")); 4028 4029 while (mp = getq(wq)) { 4030 switch (DB_TYPE(mp)) { 4031 case M_DATA: 4032 if (fptr && mp) { 4033 headerp = (fcph_network_hdr_t *)mp->b_rptr; 4034 fdestp = fcip_get_dest(fptr, 4035 &headerp->net_dest_addr); 4036 if (fdestp == NULL) { 4037 freemsg(mp); 4038 goto done; 4039 } 4040 if (fcip_start(wq, mp, fptr, fdestp, 4041 KM_SLEEP)) { 4042 goto done; 4043 } 4044 } else { 4045 freemsg(mp); 4046 } 4047 break; 4048 4049 case M_PROTO: 4050 case M_PCPROTO: 4051 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 4052 (CE_NOTE, "PROT msg in wsrv")); 4053 fcip_proto(wq, mp); 4054 break; 4055 default: 4056 break; 4057 } 4058 } 4059 done: 4060 return (0); 4061 } 4062 4063 4064 /* 4065 * This routine is called from fcip_wsrv to send a message downstream 4066 * on the fibre towards its destination. This routine performs the 4067 * actual WWN to D_ID mapping by looking up the routing and destination 4068 * tables. 4069 */ 4070 /* ARGSUSED */ 4071 static int 4072 fcip_start(queue_t *wq, mblk_t *mp, struct fcip *fptr, 4073 struct fcip_dest *fdestp, int flags) 4074 { 4075 int rval; 4076 int free; 4077 fcip_pkt_t *fcip_pkt; 4078 fc_packet_t *fc_pkt; 4079 fcip_port_info_t *fport = fptr->fcip_port_info; 4080 size_t datalen; 4081 4082 FCIP_TNF_PROBE_4((fcip_start, "fcip io", /* CSTYLED */, 4083 tnf_string, msg, "enter", tnf_opaque, wq, wq, 4084 tnf_opaque, mp, mp, 4085 tnf_opaque, MP_DB_TYPE, DB_TYPE(mp))); 4086 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, (CE_NOTE, "in fcipstart")); 4087 4088 ASSERT(fdestp != NULL); 4089 4090 /* 4091 * Only return if port has gone offline and not come back online 4092 * in a while 4093 */ 4094 if (fptr->fcip_flags & FCIP_LINK_DOWN) { 4095 freemsg(mp); 4096 return (0); 4097 } 4098 4099 /* 4100 * The message block coming in here already has the network and 4101 * llc_snap hdr stuffed in 4102 */ 4103 /* 4104 * Traditionally ethernet drivers at sun handle 3 cases here - 4105 * 1. messages with one mblk 4106 * 2. messages with 2 mblks 4107 * 3. messages with >2 mblks 4108 * For now lets handle all the 3 cases in a single case where we 4109 * put them together in one mblk that has all the data 4110 */ 4111 4112 if (mp->b_cont != NULL) { 4113 if (!pullupmsg(mp, -1)) { 4114 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 4115 (CE_WARN, "failed to concat message")); 4116 freemsg(mp); 4117 return (1); 4118 } 4119 } 4120 4121 datalen = msgsize(mp); 4122 4123 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, (CE_NOTE, 4124 "msgsize with nhdr & llcsnap hdr in fcip_pkt_alloc 0x%lx", 4125 datalen)); 4126 4127 /* 4128 * We cannot have requests larger than FCIPMTU+Headers 4129 */ 4130 if (datalen > (FCIPMTU + sizeof (llc_snap_hdr_t) + 4131 sizeof (fcph_network_hdr_t))) { 4132 freemsg(mp); 4133 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, (CE_NOTE, 4134 "fcip_pkt_alloc: datalen is larger than " 4135 "max possible size.")); 4136 return (1); 4137 } 4138 4139 fcip_pkt = fcip_pkt_alloc(fptr, mp, flags, datalen); 4140 if (fcip_pkt == NULL) { 4141 (void) putbq(wq, mp); 4142 return (1); 4143 } 4144 4145 fcip_pkt->fcip_pkt_mp = mp; 4146 fcip_pkt->fcip_pkt_wq = wq; 4147 fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 4148 4149 mutex_enter(&fdestp->fcipd_mutex); 4150 /* 4151 * If the device dynamically disappeared, just fail the request. 4152 */ 4153 if (fdestp->fcipd_rtable == NULL) { 4154 mutex_exit(&fdestp->fcipd_mutex); 4155 fcip_pkt_free(fcip_pkt, 1); 4156 return (1); 4157 } 4158 4159 /* 4160 * Now that we've assigned pkt_pd, we can call fc_ulp_init_packet 4161 */ 4162 4163 fc_pkt->pkt_pd = fdestp->fcipd_pd; 4164 4165 if (fc_ulp_init_packet((opaque_t)fport->fcipp_handle, 4166 fc_pkt, flags) != FC_SUCCESS) { 4167 mutex_exit(&fdestp->fcipd_mutex); 4168 fcip_pkt_free(fcip_pkt, 1); 4169 return (1); 4170 } 4171 4172 fcip_fdestp_enqueue_pkt(fdestp, fcip_pkt); 4173 fcip_pkt->fcip_pkt_dest = fdestp; 4174 fc_pkt->pkt_fca_device = fdestp->fcipd_fca_dev; 4175 4176 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, (CE_NOTE, 4177 "setting cmdlen to 0x%x: rsp 0x%x : data 0x%x", 4178 fc_pkt->pkt_cmdlen, fc_pkt->pkt_rsplen, fc_pkt->pkt_datalen)); 4179 4180 fcip_init_unicast_pkt(fcip_pkt, fport->fcipp_sid, 4181 fdestp->fcipd_did, fcip_pkt_callback); 4182 4183 fdestp->fcipd_ncmds++; 4184 4185 mutex_exit(&fdestp->fcipd_mutex); 4186 if ((rval = fcip_transport(fcip_pkt)) == FC_SUCCESS) { 4187 fptr->fcip_opackets++; 4188 return (0); 4189 } 4190 4191 free = (rval == FC_STATEC_BUSY || rval == FC_OFFLINE || 4192 rval == FC_TRAN_BUSY) ? 0 : 1; 4193 4194 mutex_enter(&fdestp->fcipd_mutex); 4195 rval = fcip_fdestp_dequeue_pkt(fdestp, fcip_pkt); 4196 4197 if (!rval) { 4198 fcip_pkt = NULL; 4199 } else { 4200 fdestp->fcipd_ncmds--; 4201 } 4202 mutex_exit(&fdestp->fcipd_mutex); 4203 4204 if (fcip_pkt != NULL) { 4205 fcip_pkt_free(fcip_pkt, free); 4206 } 4207 4208 if (!free) { 4209 (void) putbq(wq, mp); 4210 } 4211 4212 return (1); 4213 } 4214 4215 4216 /* 4217 * This routine enqueus a packet marked to be issued to the 4218 * transport in the dest structure. This enables us to timeout any 4219 * request stuck with the FCA/transport for long periods of time 4220 * without a response. fcip_pkt_timeout will attempt to clean up 4221 * any packets hung in this state of limbo. 4222 */ 4223 static void 4224 fcip_fdestp_enqueue_pkt(struct fcip_dest *fdestp, fcip_pkt_t *fcip_pkt) 4225 { 4226 ASSERT(mutex_owned(&fdestp->fcipd_mutex)); 4227 FCIP_TNF_PROBE_1((fcip_fdestp_enqueue_pkt, "fcip io", /* CSTYLED */, 4228 tnf_string, msg, "destp enq pkt")); 4229 4230 /* 4231 * Just hang it off the head of packet list 4232 */ 4233 fcip_pkt->fcip_pkt_next = fdestp->fcipd_head; 4234 fcip_pkt->fcip_pkt_prev = NULL; 4235 fcip_pkt->fcip_pkt_flags |= FCIP_PKT_IN_LIST; 4236 4237 if (fdestp->fcipd_head != NULL) { 4238 ASSERT(fdestp->fcipd_head->fcip_pkt_prev == NULL); 4239 fdestp->fcipd_head->fcip_pkt_prev = fcip_pkt; 4240 } 4241 4242 fdestp->fcipd_head = fcip_pkt; 4243 } 4244 4245 4246 /* 4247 * dequeues any packets after the transport/FCA tells us it has 4248 * been successfully sent on its way. Ofcourse it doesn't mean that 4249 * the packet will actually reach its destination but its atleast 4250 * a step closer in that direction 4251 */ 4252 static int 4253 fcip_fdestp_dequeue_pkt(struct fcip_dest *fdestp, fcip_pkt_t *fcip_pkt) 4254 { 4255 fcip_pkt_t *fcipd_pkt; 4256 4257 ASSERT(mutex_owned(&fdestp->fcipd_mutex)); 4258 if (fcip_pkt->fcip_pkt_flags & FCIP_PKT_IN_TIMEOUT) { 4259 fcipd_pkt = fdestp->fcipd_head; 4260 while (fcipd_pkt) { 4261 if (fcipd_pkt == fcip_pkt) { 4262 fcip_pkt_t *pptr = NULL; 4263 4264 if (fcipd_pkt == fdestp->fcipd_head) { 4265 ASSERT(fcipd_pkt->fcip_pkt_prev == 4266 NULL); 4267 fdestp->fcipd_head = 4268 fcipd_pkt->fcip_pkt_next; 4269 } else { 4270 pptr = fcipd_pkt->fcip_pkt_prev; 4271 ASSERT(pptr != NULL); 4272 pptr->fcip_pkt_next = 4273 fcipd_pkt->fcip_pkt_next; 4274 } 4275 if (fcipd_pkt->fcip_pkt_next) { 4276 pptr = fcipd_pkt->fcip_pkt_next; 4277 pptr->fcip_pkt_prev = 4278 fcipd_pkt->fcip_pkt_prev; 4279 } 4280 fcip_pkt->fcip_pkt_flags &= ~FCIP_PKT_IN_LIST; 4281 break; 4282 } 4283 fcipd_pkt = fcipd_pkt->fcip_pkt_next; 4284 } 4285 } else { 4286 if (fcip_pkt->fcip_pkt_prev == NULL) { 4287 ASSERT(fdestp->fcipd_head == fcip_pkt); 4288 fdestp->fcipd_head = fcip_pkt->fcip_pkt_next; 4289 } else { 4290 fcip_pkt->fcip_pkt_prev->fcip_pkt_next = 4291 fcip_pkt->fcip_pkt_next; 4292 } 4293 4294 if (fcip_pkt->fcip_pkt_next) { 4295 fcip_pkt->fcip_pkt_next->fcip_pkt_prev = 4296 fcip_pkt->fcip_pkt_prev; 4297 } 4298 4299 fcipd_pkt = fcip_pkt; 4300 fcip_pkt->fcip_pkt_flags &= ~FCIP_PKT_IN_LIST; 4301 } 4302 4303 return (fcipd_pkt == fcip_pkt); 4304 } 4305 4306 /* 4307 * The transport routine - this is the routine that actually calls 4308 * into the FCA driver (through the transport ofcourse) to transmit a 4309 * datagram on the fibre. The dest struct assoicated with the port to 4310 * which the data is intended is already bound to the packet, this routine 4311 * only takes care of marking the packet a broadcast packet if it is 4312 * intended to be a broadcast request. This permits the transport to send 4313 * the packet down on the wire even if it doesn't have an entry for the 4314 * D_ID in its d_id hash tables. 4315 */ 4316 static int 4317 fcip_transport(fcip_pkt_t *fcip_pkt) 4318 { 4319 struct fcip *fptr; 4320 fc_packet_t *fc_pkt; 4321 fcip_port_info_t *fport; 4322 struct fcip_dest *fdestp; 4323 uint32_t did; 4324 int rval = FC_FAILURE; 4325 struct fcip_routing_table *frp = NULL; 4326 4327 FCIP_TNF_PROBE_1((fcip_transport, "fcip io", /* CSTYLED */, 4328 tnf_string, msg, "enter")); 4329 4330 fptr = fcip_pkt->fcip_pkt_fptr; 4331 fport = fptr->fcip_port_info; 4332 fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 4333 fdestp = fcip_pkt->fcip_pkt_dest; 4334 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, (CE_WARN, "fcip_transport called")); 4335 4336 did = fptr->fcip_broadcast_did; 4337 if (fc_pkt->pkt_cmd_fhdr.d_id == did && 4338 fc_pkt->pkt_tran_type != FC_PKT_BROADCAST) { 4339 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 4340 (CE_NOTE, "trantype set to BROADCAST")); 4341 fc_pkt->pkt_tran_type = FC_PKT_BROADCAST; 4342 } 4343 4344 mutex_enter(&fptr->fcip_mutex); 4345 if ((fc_pkt->pkt_tran_type != FC_PKT_BROADCAST) && 4346 (fc_pkt->pkt_pd == NULL)) { 4347 mutex_exit(&fptr->fcip_mutex); 4348 FCIP_TNF_PROBE_1((fcip_transport, "fcip io", /* CSTYLED */, 4349 tnf_string, msg, "fcip transport no pd")); 4350 return (rval); 4351 } else if (fptr->fcip_port_state == FCIP_PORT_OFFLINE) { 4352 mutex_exit(&fptr->fcip_mutex); 4353 FCIP_TNF_PROBE_1((fcip_transport, "fcip io", /* CSTYLED */, 4354 tnf_string, msg, "fcip transport port offline")); 4355 return (FC_TRAN_BUSY); 4356 } 4357 mutex_exit(&fptr->fcip_mutex); 4358 4359 if (fdestp) { 4360 struct fcip_routing_table *frp; 4361 4362 frp = fdestp->fcipd_rtable; 4363 mutex_enter(&fptr->fcip_rt_mutex); 4364 mutex_enter(&fdestp->fcipd_mutex); 4365 if (fc_pkt->pkt_pd != NULL) { 4366 if ((frp == NULL) || 4367 (frp && FCIP_RTE_UNAVAIL(frp->fcipr_state))) { 4368 mutex_exit(&fdestp->fcipd_mutex); 4369 mutex_exit(&fptr->fcip_rt_mutex); 4370 if (frp && 4371 (frp->fcipr_state == FCIP_RT_INVALID)) { 4372 FCIP_TNF_PROBE_1((fcip_transport, 4373 "fcip io", /* CSTYLED */, 4374 tnf_string, msg, 4375 "fcip transport - TRANBUSY")); 4376 return (FC_TRAN_BUSY); 4377 } else { 4378 FCIP_TNF_PROBE_1((fcip_transport, 4379 "fcip io", /* CSTYLED */, 4380 tnf_string, msg, 4381 "fcip transport: frp unavailable")); 4382 return (rval); 4383 } 4384 } 4385 } 4386 mutex_exit(&fdestp->fcipd_mutex); 4387 mutex_exit(&fptr->fcip_rt_mutex); 4388 ASSERT(fcip_pkt->fcip_pkt_flags & FCIP_PKT_IN_LIST); 4389 } 4390 4391 /* Explicitly invalidate this field till fcip decides to use it */ 4392 fc_pkt->pkt_ulp_rscn_infop = NULL; 4393 4394 rval = fc_ulp_transport(fport->fcipp_handle, fc_pkt); 4395 if (rval == FC_STATEC_BUSY || rval == FC_OFFLINE) { 4396 /* 4397 * Need to queue up the command for retry 4398 */ 4399 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 4400 (CE_WARN, "ulp_transport failed: 0x%x", rval)); 4401 } else if (rval == FC_LOGINREQ && (frp != NULL)) { 4402 (void) fcip_do_plogi(fptr, frp); 4403 } else if (rval == FC_BADPACKET && (frp != NULL)) { 4404 /* 4405 * There is a distinct possiblity in our scheme of things 4406 * that we have a routing table entry with a NULL pd struct. 4407 * Mark the routing table entry for removal if it is not a 4408 * broadcast entry 4409 */ 4410 if ((frp->fcipr_d_id.port_id != 0x0) && 4411 (frp->fcipr_d_id.port_id != 0xffffff)) { 4412 mutex_enter(&fptr->fcip_rt_mutex); 4413 frp->fcipr_pd = NULL; 4414 frp->fcipr_state = PORT_DEVICE_INVALID; 4415 mutex_exit(&fptr->fcip_rt_mutex); 4416 } 4417 } 4418 4419 FCIP_TNF_PROBE_1((fcip_transport, "fcip io", /* CSTYLED */, 4420 tnf_string, msg, "fcip transport done")); 4421 return (rval); 4422 } 4423 4424 /* 4425 * Call back routine. Called by the FCA/transport when the messages 4426 * has been put onto the wire towards its intended destination. We can 4427 * now free the fc_packet associated with the message 4428 */ 4429 static void 4430 fcip_pkt_callback(fc_packet_t *fc_pkt) 4431 { 4432 int rval; 4433 fcip_pkt_t *fcip_pkt; 4434 struct fcip_dest *fdestp; 4435 4436 fcip_pkt = (fcip_pkt_t *)fc_pkt->pkt_ulp_private; 4437 fdestp = fcip_pkt->fcip_pkt_dest; 4438 4439 /* 4440 * take the lock early so that we don't have a race condition 4441 * with fcip_timeout 4442 * 4443 * fdestp->fcipd_mutex isn't really intended to lock per 4444 * packet struct - see bug 5105592 for permanent solution 4445 */ 4446 mutex_enter(&fdestp->fcipd_mutex); 4447 4448 fcip_pkt->fcip_pkt_flags |= FCIP_PKT_RETURNED; 4449 fcip_pkt->fcip_pkt_flags &= ~FCIP_PKT_IN_ABORT; 4450 if (fcip_pkt->fcip_pkt_flags & FCIP_PKT_IN_TIMEOUT) { 4451 mutex_exit(&fdestp->fcipd_mutex); 4452 return; 4453 } 4454 4455 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, (CE_NOTE, "pkt callback")); 4456 4457 ASSERT(fdestp->fcipd_rtable != NULL); 4458 ASSERT(fcip_pkt->fcip_pkt_flags & FCIP_PKT_IN_LIST); 4459 rval = fcip_fdestp_dequeue_pkt(fdestp, fcip_pkt); 4460 fdestp->fcipd_ncmds--; 4461 mutex_exit(&fdestp->fcipd_mutex); 4462 4463 if (rval) { 4464 fcip_pkt_free(fcip_pkt, 1); 4465 } 4466 4467 FCIP_TNF_PROBE_1((fcip_pkt_callback, "fcip io", /* CSTYLED */, 4468 tnf_string, msg, "pkt callback done")); 4469 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, (CE_NOTE, "pkt callback done")); 4470 } 4471 4472 /* 4473 * Return 1 if the topology is supported, else return 0. 4474 * Topology support is consistent with what the whole 4475 * stack supports together. 4476 */ 4477 static int 4478 fcip_is_supported_fc_topology(int fc_topology) 4479 { 4480 switch (fc_topology) { 4481 4482 case FC_TOP_PRIVATE_LOOP : 4483 case FC_TOP_PUBLIC_LOOP : 4484 case FC_TOP_FABRIC : 4485 case FC_TOP_NO_NS : 4486 return (1); 4487 default : 4488 return (0); 4489 } 4490 } 4491 4492 /* 4493 * handle any topology specific initializations here 4494 * this routine must be called while holding fcip_mutex 4495 */ 4496 /* ARGSUSED */ 4497 static void 4498 fcip_handle_topology(struct fcip *fptr) 4499 { 4500 4501 fcip_port_info_t *fport = fptr->fcip_port_info; 4502 4503 ASSERT(mutex_owned(&fptr->fcip_mutex)); 4504 4505 /* 4506 * Since we know the port's topology - handle topology 4507 * specific details here. In Point to Point and Private Loop 4508 * topologies - we would probably not have a name server 4509 */ 4510 4511 FCIP_TNF_PROBE_3((fcip_handle_topology, "fcip io", /* CSTYLED */, 4512 tnf_string, msg, "enter", 4513 tnf_uint, port_state, fport->fcipp_pstate, 4514 tnf_uint, topology, fport->fcipp_topology)); 4515 FCIP_DEBUG(FCIP_DEBUG_INIT, (CE_NOTE, "port state: %x, topology %x", 4516 fport->fcipp_pstate, fport->fcipp_topology)); 4517 4518 fptr->fcip_broadcast_did = fcip_get_broadcast_did(fptr); 4519 mutex_exit(&fptr->fcip_mutex); 4520 (void) fcip_dest_add_broadcast_entry(fptr, 0); 4521 mutex_enter(&fptr->fcip_mutex); 4522 4523 if (!fcip_is_supported_fc_topology(fport->fcipp_topology)) { 4524 FCIP_DEBUG(FCIP_DEBUG_INIT, 4525 (CE_WARN, "fcip(0x%x): Unsupported port topology (0x%x)", 4526 fptr->fcip_instance, fport->fcipp_topology)); 4527 return; 4528 } 4529 4530 switch (fport->fcipp_topology) { 4531 case FC_TOP_PRIVATE_LOOP: { 4532 4533 fc_portmap_t *port_map; 4534 uint32_t listlen, alloclen; 4535 /* 4536 * we may have to maintain routing. Get a list of 4537 * all devices on this port that the transport layer is 4538 * aware of. Check if any of them is a IS8802 type port, 4539 * if yes get its WWN and DID mapping and cache it in 4540 * the purport routing table. Since there is no 4541 * State Change notification for private loop/point_point 4542 * topologies - this table may not be accurate. The static 4543 * routing table is updated on a state change callback. 4544 */ 4545 FCIP_DEBUG(FCIP_DEBUG_INIT, (CE_WARN, "port state valid!!")); 4546 fptr->fcip_port_state = FCIP_PORT_ONLINE; 4547 listlen = alloclen = FCIP_MAX_PORTS; 4548 port_map = (fc_portmap_t *) 4549 kmem_zalloc((FCIP_MAX_PORTS * sizeof (fc_portmap_t)), 4550 KM_SLEEP); 4551 if (fc_ulp_getportmap(fport->fcipp_handle, &port_map, 4552 &listlen, FC_ULP_PLOGI_PRESERVE) == FC_SUCCESS) { 4553 mutex_exit(&fptr->fcip_mutex); 4554 fcip_rt_update(fptr, port_map, listlen); 4555 mutex_enter(&fptr->fcip_mutex); 4556 } 4557 if (listlen > alloclen) { 4558 alloclen = listlen; 4559 } 4560 kmem_free(port_map, (alloclen * sizeof (fc_portmap_t))); 4561 /* 4562 * Now fall through and register with the transport 4563 * that this port is IP capable 4564 */ 4565 } 4566 /* FALLTHROUGH */ 4567 case FC_TOP_NO_NS: 4568 /* 4569 * If we don't have a nameserver, lets wait until we 4570 * have to send out a packet to a remote port and then 4571 * try and discover the port using ARP/FARP. 4572 */ 4573 /* FALLTHROUGH */ 4574 case FC_TOP_PUBLIC_LOOP: 4575 case FC_TOP_FABRIC: { 4576 fc_portmap_t *port_map; 4577 uint32_t listlen, alloclen; 4578 4579 /* FC_TYPE of 0x05 goes to word 0, LSB */ 4580 fptr->fcip_port_state = FCIP_PORT_ONLINE; 4581 4582 if (!(fptr->fcip_flags & FCIP_REG_INPROGRESS)) { 4583 fptr->fcip_flags |= FCIP_REG_INPROGRESS; 4584 if (taskq_dispatch(fptr->fcip_tq, fcip_port_ns, 4585 fptr, KM_NOSLEEP) == 0) { 4586 fptr->fcip_flags &= ~FCIP_REG_INPROGRESS; 4587 } 4588 } 4589 4590 /* 4591 * If fcip_create_nodes_on_demand is overridden to force 4592 * discovery of all nodes in Fabric/Public loop topologies 4593 * we need to query for and obtain all nodes and log into 4594 * them as with private loop devices 4595 */ 4596 if (!fcip_create_nodes_on_demand) { 4597 fptr->fcip_port_state = FCIP_PORT_ONLINE; 4598 listlen = alloclen = FCIP_MAX_PORTS; 4599 port_map = (fc_portmap_t *) 4600 kmem_zalloc((FCIP_MAX_PORTS * 4601 sizeof (fc_portmap_t)), KM_SLEEP); 4602 if (fc_ulp_getportmap(fport->fcipp_handle, &port_map, 4603 &listlen, FC_ULP_PLOGI_PRESERVE) == FC_SUCCESS) { 4604 mutex_exit(&fptr->fcip_mutex); 4605 fcip_rt_update(fptr, port_map, listlen); 4606 mutex_enter(&fptr->fcip_mutex); 4607 } 4608 if (listlen > alloclen) { 4609 alloclen = listlen; 4610 } 4611 kmem_free(port_map, 4612 (alloclen * sizeof (fc_portmap_t))); 4613 } 4614 break; 4615 } 4616 4617 default: 4618 break; 4619 } 4620 } 4621 4622 static void 4623 fcip_port_ns(void *arg) 4624 { 4625 struct fcip *fptr = (struct fcip *)arg; 4626 fcip_port_info_t *fport = fptr->fcip_port_info; 4627 fc_ns_cmd_t ns_cmd; 4628 uint32_t types[8]; 4629 ns_rfc_type_t rfc; 4630 4631 mutex_enter(&fptr->fcip_mutex); 4632 if ((fptr->fcip_flags & (FCIP_DETACHING | FCIP_DETACHED)) || 4633 (fptr->fcip_flags & (FCIP_SUSPENDED | FCIP_POWER_DOWN))) { 4634 fptr->fcip_flags &= ~FCIP_REG_INPROGRESS; 4635 mutex_exit(&fptr->fcip_mutex); 4636 return; 4637 } 4638 mutex_exit(&fptr->fcip_mutex); 4639 4640 /* 4641 * Prepare the Name server structure to 4642 * register with the transport in case of 4643 * Fabric configuration. 4644 */ 4645 bzero(&rfc, sizeof (rfc)); 4646 bzero(types, sizeof (types)); 4647 4648 types[FC4_TYPE_WORD_POS(FC_TYPE_IS8802_SNAP)] = (1 << 4649 FC4_TYPE_BIT_POS(FC_TYPE_IS8802_SNAP)); 4650 4651 rfc.rfc_port_id.port_id = fport->fcipp_sid.port_id; 4652 bcopy(types, rfc.rfc_types, sizeof (types)); 4653 4654 ns_cmd.ns_flags = 0; 4655 ns_cmd.ns_cmd = NS_RFT_ID; 4656 ns_cmd.ns_req_len = sizeof (rfc); 4657 ns_cmd.ns_req_payload = (caddr_t)&rfc; 4658 ns_cmd.ns_resp_len = 0; 4659 ns_cmd.ns_resp_payload = NULL; 4660 4661 /* 4662 * Perform the Name Server Registration for FC IS8802_SNAP Type. 4663 * We don't expect a reply for registering port type 4664 */ 4665 (void) fc_ulp_port_ns(fptr->fcip_port_info->fcipp_handle, 4666 (opaque_t)0, &ns_cmd); 4667 4668 mutex_enter(&fptr->fcip_mutex); 4669 fptr->fcip_flags &= ~FCIP_REG_INPROGRESS; 4670 mutex_exit(&fptr->fcip_mutex); 4671 } 4672 4673 /* 4674 * setup this instance of fcip. This routine inits kstats, allocates 4675 * unsolicited buffers, determines' this port's siblings and handles 4676 * topology specific details which includes registering with the name 4677 * server and also setting up the routing table for this port for 4678 * private loops and point to point topologies 4679 */ 4680 static int 4681 fcip_init_port(struct fcip *fptr) 4682 { 4683 int rval = FC_SUCCESS; 4684 fcip_port_info_t *fport = fptr->fcip_port_info; 4685 static char buf[64]; 4686 size_t tok_buf_size; 4687 4688 ASSERT(fport != NULL); 4689 4690 FCIP_TNF_PROBE_1((fcip_init_port, "fcip io", /* CSTYLED */, 4691 tnf_string, msg, "enter")); 4692 mutex_enter(&fptr->fcip_mutex); 4693 4694 /* 4695 * setup mac address for this port. Don't be too worried if 4696 * the WWN is zero, there is probably nothing attached to 4697 * to the port. There is no point allocating unsolicited buffers 4698 * for an unused port so return success if we don't have a MAC 4699 * address. Do the port init on a state change notification. 4700 */ 4701 if (fcip_setup_mac_addr(fptr) == FCIP_INVALID_WWN) { 4702 fptr->fcip_port_state = FCIP_PORT_OFFLINE; 4703 rval = FC_SUCCESS; 4704 goto done; 4705 } 4706 4707 /* 4708 * clear routing table hash list for this port 4709 */ 4710 fcip_rt_flush(fptr); 4711 4712 /* 4713 * init kstats for this instance 4714 */ 4715 fcip_kstat_init(fptr); 4716 4717 /* 4718 * Allocate unsolicited buffers 4719 */ 4720 fptr->fcip_ub_nbufs = fcip_ub_nbufs; 4721 tok_buf_size = sizeof (*fptr->fcip_ub_tokens) * fcip_ub_nbufs; 4722 4723 FCIP_TNF_PROBE_2((fcip_init_port, "fcip io", /* CSTYLED */, 4724 tnf_string, msg, "debug", 4725 tnf_int, tokBufsize, tok_buf_size)); 4726 4727 FCIP_DEBUG(FCIP_DEBUG_INIT, 4728 (CE_WARN, "tokBufsize: 0x%lx", tok_buf_size)); 4729 4730 fptr->fcip_ub_tokens = kmem_zalloc(tok_buf_size, KM_SLEEP); 4731 4732 if (fptr->fcip_ub_tokens == NULL) { 4733 rval = FC_FAILURE; 4734 FCIP_DEBUG(FCIP_DEBUG_INIT, 4735 (CE_WARN, "fcip(%d): failed to allocate unsol buf", 4736 fptr->fcip_instance)); 4737 goto done; 4738 } 4739 rval = fc_ulp_uballoc(fport->fcipp_handle, &fptr->fcip_ub_nbufs, 4740 fcip_ub_size, FC_TYPE_IS8802_SNAP, fptr->fcip_ub_tokens); 4741 4742 if (rval != FC_SUCCESS) { 4743 FCIP_DEBUG(FCIP_DEBUG_INIT, 4744 (CE_WARN, "fcip(%d): fc_ulp_uballoc failed with 0x%x!!", 4745 fptr->fcip_instance, rval)); 4746 } 4747 4748 switch (rval) { 4749 case FC_SUCCESS: 4750 break; 4751 4752 case FC_OFFLINE: 4753 fptr->fcip_port_state = FCIP_PORT_OFFLINE; 4754 rval = FC_FAILURE; 4755 goto done; 4756 4757 case FC_UB_ERROR: 4758 FCIP_TNF_PROBE_1((fcip_init_port, "fcip io", /* CSTYLED */, 4759 tnf_string, msg, "invalid ub alloc request")); 4760 FCIP_DEBUG(FCIP_DEBUG_INIT, 4761 (CE_WARN, "invalid ub alloc request !!")); 4762 rval = FC_FAILURE; 4763 goto done; 4764 4765 case FC_FAILURE: 4766 /* 4767 * requested bytes could not be alloced 4768 */ 4769 if (fptr->fcip_ub_nbufs != fcip_ub_nbufs) { 4770 cmn_err(CE_WARN, 4771 "!fcip(0x%x): Failed to alloc unsolicited bufs", 4772 ddi_get_instance(fport->fcipp_dip)); 4773 rval = FC_FAILURE; 4774 goto done; 4775 } 4776 break; 4777 4778 default: 4779 rval = FC_FAILURE; 4780 break; 4781 } 4782 4783 /* 4784 * Preallocate a Cache of fcip packets for transmit and receive 4785 * We don't want to be holding on to unsolicited buffers while 4786 * we transmit the message upstream 4787 */ 4788 FCIP_DEBUG(FCIP_DEBUG_INIT, (CE_NOTE, "allocating fcip_pkt cache")); 4789 4790 (void) sprintf(buf, "fcip%d_cache", fptr->fcip_instance); 4791 fptr->fcip_xmit_cache = kmem_cache_create(buf, 4792 (fport->fcipp_fca_pkt_size + sizeof (fcip_pkt_t)), 4793 8, fcip_cache_constructor, fcip_cache_destructor, 4794 NULL, (void *)fport, NULL, 0); 4795 4796 (void) sprintf(buf, "fcip%d_sendup_cache", fptr->fcip_instance); 4797 fptr->fcip_sendup_cache = kmem_cache_create(buf, 4798 sizeof (struct fcip_sendup_elem), 4799 8, fcip_sendup_constructor, NULL, NULL, (void *)fport, NULL, 0); 4800 4801 if (fptr->fcip_xmit_cache == NULL) { 4802 FCIP_TNF_PROBE_2((fcip_init_port, "fcip io", /* CSTYLED */, 4803 tnf_string, msg, "unable to allocate xmit cache", 4804 tnf_int, instance, fptr->fcip_instance)); 4805 FCIP_DEBUG(FCIP_DEBUG_INIT, 4806 (CE_WARN, "fcip%d unable to allocate xmit cache", 4807 fptr->fcip_instance)); 4808 rval = FC_FAILURE; 4809 goto done; 4810 } 4811 4812 /* 4813 * We may need to handle routing tables for point to point and 4814 * fcal topologies and register with NameServer for Fabric 4815 * topologies. 4816 */ 4817 fcip_handle_topology(fptr); 4818 mutex_exit(&fptr->fcip_mutex); 4819 if (fcip_dest_add_broadcast_entry(fptr, 1) != FC_SUCCESS) { 4820 FCIP_DEBUG(FCIP_DEBUG_INIT, 4821 (CE_WARN, "fcip(0x%x):add broadcast entry failed!!", 4822 fptr->fcip_instance)); 4823 mutex_enter(&fptr->fcip_mutex); 4824 rval = FC_FAILURE; 4825 goto done; 4826 } 4827 4828 rval = FC_SUCCESS; 4829 return (rval); 4830 4831 done: 4832 /* 4833 * we don't always come here from port_attach - so cleanup 4834 * anything done in the init_port routine 4835 */ 4836 if (fptr->fcip_kstatp) { 4837 kstat_delete(fptr->fcip_kstatp); 4838 fptr->fcip_kstatp = NULL; 4839 } 4840 4841 if (fptr->fcip_xmit_cache) { 4842 kmem_cache_destroy(fptr->fcip_xmit_cache); 4843 fptr->fcip_xmit_cache = NULL; 4844 } 4845 4846 if (fptr->fcip_sendup_cache) { 4847 kmem_cache_destroy(fptr->fcip_sendup_cache); 4848 fptr->fcip_sendup_cache = NULL; 4849 } 4850 4851 /* release unsolicited buffers */ 4852 if (fptr->fcip_ub_tokens) { 4853 uint64_t *tokens = fptr->fcip_ub_tokens; 4854 fptr->fcip_ub_tokens = NULL; 4855 4856 mutex_exit(&fptr->fcip_mutex); 4857 (void) fc_ulp_ubfree(fport->fcipp_handle, fptr->fcip_ub_nbufs, 4858 tokens); 4859 kmem_free(tokens, tok_buf_size); 4860 4861 } else { 4862 mutex_exit(&fptr->fcip_mutex); 4863 } 4864 4865 return (rval); 4866 } 4867 4868 /* 4869 * Sets up a port's MAC address from its WWN 4870 */ 4871 static int 4872 fcip_setup_mac_addr(struct fcip *fptr) 4873 { 4874 fcip_port_info_t *fport = fptr->fcip_port_info; 4875 4876 ASSERT(mutex_owned(&fptr->fcip_mutex)); 4877 4878 fptr->fcip_addrflags = 0; 4879 4880 /* 4881 * we cannot choose a MAC address for our interface - we have 4882 * to live with whatever node WWN we get (minus the top two 4883 * MSbytes for the MAC address) from the transport layer. We will 4884 * treat the WWN as our factory MAC address. 4885 */ 4886 4887 if ((fport->fcipp_nwwn.w.wwn_hi != 0) || 4888 (fport->fcipp_nwwn.w.wwn_lo != 0)) { 4889 char etherstr[ETHERSTRL]; 4890 4891 wwn_to_ether(&fport->fcipp_nwwn, &fptr->fcip_macaddr); 4892 fcip_ether_to_str(&fptr->fcip_macaddr, etherstr); 4893 FCIP_DEBUG(FCIP_DEBUG_INIT, 4894 (CE_NOTE, "setupmacaddr ouraddr %s", etherstr)); 4895 4896 fptr->fcip_addrflags = (FCIP_FACTADDR_PRESENT | 4897 FCIP_FACTADDR_USE); 4898 } else { 4899 /* 4900 * No WWN - just return failure - there's not much 4901 * we can do since we cannot set the WWN. 4902 */ 4903 FCIP_DEBUG(FCIP_DEBUG_INIT, 4904 (CE_WARN, "Port does not have a valid WWN")); 4905 return (FCIP_INVALID_WWN); 4906 } 4907 return (FC_SUCCESS); 4908 } 4909 4910 4911 /* 4912 * flush routing table entries 4913 */ 4914 static void 4915 fcip_rt_flush(struct fcip *fptr) 4916 { 4917 int index; 4918 4919 mutex_enter(&fptr->fcip_rt_mutex); 4920 for (index = 0; index < FCIP_RT_HASH_ELEMS; index++) { 4921 struct fcip_routing_table *frtp, *frtp_next; 4922 frtp = fptr->fcip_rtable[index]; 4923 while (frtp) { 4924 frtp_next = frtp->fcipr_next; 4925 kmem_free(frtp, sizeof (struct fcip_routing_table)); 4926 frtp = frtp_next; 4927 } 4928 fptr->fcip_rtable[index] = NULL; 4929 } 4930 mutex_exit(&fptr->fcip_rt_mutex); 4931 } 4932 4933 /* 4934 * Free up the fcip softstate and all allocated resources for the 4935 * fcip instance assoicated with a given port driver instance 4936 * 4937 * Given that the list of structures pointed to by fcip_port_head, 4938 * this function is called from multiple sources, and the 4939 * fcip_global_mutex that protects fcip_port_head must be dropped, 4940 * our best solution is to return a value that indicates the next 4941 * port in the list. This way the caller doesn't need to worry 4942 * about the race condition where he saves off a pointer to the 4943 * next structure in the list and by the time this routine returns, 4944 * that next structure has already been freed. 4945 */ 4946 static fcip_port_info_t * 4947 fcip_softstate_free(fcip_port_info_t *fport) 4948 { 4949 struct fcip *fptr = NULL; 4950 int instance; 4951 timeout_id_t tid; 4952 opaque_t phandle = NULL; 4953 fcip_port_info_t *prev_fport, *cur_fport, *next_fport = NULL; 4954 4955 ASSERT(MUTEX_HELD(&fcip_global_mutex)); 4956 4957 if (fport) { 4958 phandle = fport->fcipp_handle; 4959 fptr = fport->fcipp_fcip; 4960 } else { 4961 return (next_fport); 4962 } 4963 4964 if (fptr) { 4965 mutex_enter(&fptr->fcip_mutex); 4966 instance = ddi_get_instance(fptr->fcip_dip); 4967 4968 /* 4969 * dismantle timeout thread for this instance of fcip 4970 */ 4971 tid = fptr->fcip_timeout_id; 4972 fptr->fcip_timeout_id = NULL; 4973 4974 mutex_exit(&fptr->fcip_mutex); 4975 (void) untimeout(tid); 4976 mutex_enter(&fptr->fcip_mutex); 4977 4978 ASSERT(fcip_num_instances >= 0); 4979 fcip_num_instances--; 4980 4981 /* 4982 * stop sendup thread 4983 */ 4984 mutex_enter(&fptr->fcip_sendup_mutex); 4985 if (fptr->fcip_sendup_thr_initted) { 4986 fptr->fcip_sendup_thr_initted = 0; 4987 cv_signal(&fptr->fcip_sendup_cv); 4988 cv_wait(&fptr->fcip_sendup_cv, 4989 &fptr->fcip_sendup_mutex); 4990 } 4991 ASSERT(fptr->fcip_sendup_head == NULL); 4992 fptr->fcip_sendup_head = fptr->fcip_sendup_tail = NULL; 4993 mutex_exit(&fptr->fcip_sendup_mutex); 4994 4995 /* 4996 * dismantle taskq 4997 */ 4998 if (fptr->fcip_tq) { 4999 taskq_t *tq = fptr->fcip_tq; 5000 5001 fptr->fcip_tq = NULL; 5002 5003 mutex_exit(&fptr->fcip_mutex); 5004 taskq_destroy(tq); 5005 mutex_enter(&fptr->fcip_mutex); 5006 } 5007 5008 if (fptr->fcip_kstatp) { 5009 kstat_delete(fptr->fcip_kstatp); 5010 fptr->fcip_kstatp = NULL; 5011 } 5012 5013 /* flush the routing table entries */ 5014 fcip_rt_flush(fptr); 5015 5016 if (fptr->fcip_xmit_cache) { 5017 kmem_cache_destroy(fptr->fcip_xmit_cache); 5018 fptr->fcip_xmit_cache = NULL; 5019 } 5020 5021 if (fptr->fcip_sendup_cache) { 5022 kmem_cache_destroy(fptr->fcip_sendup_cache); 5023 fptr->fcip_sendup_cache = NULL; 5024 } 5025 5026 fcip_cleanup_dest(fptr); 5027 5028 /* release unsolicited buffers */ 5029 if (fptr->fcip_ub_tokens) { 5030 uint64_t *tokens = fptr->fcip_ub_tokens; 5031 5032 fptr->fcip_ub_tokens = NULL; 5033 mutex_exit(&fptr->fcip_mutex); 5034 if (phandle) { 5035 /* 5036 * release the global mutex here to 5037 * permit any data pending callbacks to 5038 * complete. Else we will deadlock in the 5039 * FCA waiting for all unsol buffers to be 5040 * returned. 5041 */ 5042 mutex_exit(&fcip_global_mutex); 5043 (void) fc_ulp_ubfree(phandle, 5044 fptr->fcip_ub_nbufs, tokens); 5045 mutex_enter(&fcip_global_mutex); 5046 } 5047 kmem_free(tokens, (sizeof (*tokens) * fcip_ub_nbufs)); 5048 } else { 5049 mutex_exit(&fptr->fcip_mutex); 5050 } 5051 5052 mutex_destroy(&fptr->fcip_mutex); 5053 mutex_destroy(&fptr->fcip_ub_mutex); 5054 mutex_destroy(&fptr->fcip_rt_mutex); 5055 mutex_destroy(&fptr->fcip_dest_mutex); 5056 mutex_destroy(&fptr->fcip_sendup_mutex); 5057 cv_destroy(&fptr->fcip_farp_cv); 5058 cv_destroy(&fptr->fcip_sendup_cv); 5059 cv_destroy(&fptr->fcip_ub_cv); 5060 5061 ddi_soft_state_free(fcip_softp, instance); 5062 } 5063 5064 /* 5065 * Now dequeue the fcip_port_info from the port list 5066 */ 5067 cur_fport = fcip_port_head; 5068 prev_fport = NULL; 5069 while (cur_fport != NULL) { 5070 if (cur_fport == fport) { 5071 break; 5072 } 5073 prev_fport = cur_fport; 5074 cur_fport = cur_fport->fcipp_next; 5075 } 5076 5077 /* 5078 * Assert that we found a port in our port list 5079 */ 5080 ASSERT(cur_fport == fport); 5081 5082 if (prev_fport) { 5083 /* 5084 * Not the first port in the port list 5085 */ 5086 prev_fport->fcipp_next = fport->fcipp_next; 5087 } else { 5088 /* 5089 * first port 5090 */ 5091 fcip_port_head = fport->fcipp_next; 5092 } 5093 next_fport = fport->fcipp_next; 5094 kmem_free(fport, sizeof (fcip_port_info_t)); 5095 5096 return (next_fport); 5097 } 5098 5099 5100 /* 5101 * This is called by transport for any ioctl operations performed 5102 * on the devctl or other transport minor nodes. It is currently 5103 * unused for fcip 5104 */ 5105 /* ARGSUSED */ 5106 static int 5107 fcip_port_ioctl(opaque_t ulp_handle, opaque_t port_handle, dev_t dev, 5108 int cmd, intptr_t data, int mode, cred_t *credp, int *rval, 5109 uint32_t claimed) 5110 { 5111 return (FC_UNCLAIMED); 5112 } 5113 5114 /* 5115 * DL_INFO_REQ - returns information about the DLPI stream to the DLS user 5116 * requesting information about this interface 5117 */ 5118 static void 5119 fcip_ireq(queue_t *wq, mblk_t *mp) 5120 { 5121 struct fcipstr *slp; 5122 struct fcip *fptr; 5123 dl_info_ack_t *dlip; 5124 struct fcipdladdr *dlap; 5125 la_wwn_t *ep; 5126 int size; 5127 char etherstr[ETHERSTRL]; 5128 5129 slp = (struct fcipstr *)wq->q_ptr; 5130 5131 fptr = slp->sl_fcip; 5132 5133 FCIP_DEBUG(FCIP_DEBUG_DLPI, 5134 (CE_NOTE, "fcip_ireq: info request req rcvd")); 5135 5136 FCIP_TNF_PROBE_1((fcip_ireq, "fcip io", /* CSTYLED */, 5137 tnf_string, msg, "fcip ireq entered")); 5138 5139 if (MBLKL(mp) < DL_INFO_REQ_SIZE) { 5140 dlerrorack(wq, mp, DL_INFO_REQ, DL_BADPRIM, 0); 5141 return; 5142 } 5143 5144 /* 5145 * Exchange current message for a DL_INFO_ACK 5146 */ 5147 size = sizeof (dl_info_ack_t) + FCIPADDRL + ETHERADDRL; 5148 if ((mp = mexchange(wq, mp, size, M_PCPROTO, DL_INFO_ACK)) == NULL) { 5149 return; 5150 } 5151 5152 /* 5153 * FILL in the DL_INFO_ACK fields and reply 5154 */ 5155 dlip = (dl_info_ack_t *)mp->b_rptr; 5156 *dlip = fcip_infoack; 5157 dlip->dl_current_state = slp->sl_state; 5158 dlap = (struct fcipdladdr *)(mp->b_rptr + dlip->dl_addr_offset); 5159 dlap->dl_sap = slp->sl_sap; 5160 5161 5162 if (fptr) { 5163 fcip_ether_to_str(&fptr->fcip_macaddr, etherstr); 5164 FCIP_DEBUG(FCIP_DEBUG_DLPI, 5165 (CE_NOTE, "ireq - our mac: %s", etherstr)); 5166 ether_bcopy(&fptr->fcip_macaddr, &dlap->dl_phys); 5167 } else { 5168 bzero((caddr_t)&dlap->dl_phys, ETHERADDRL); 5169 } 5170 5171 ep = (la_wwn_t *)(mp->b_rptr + dlip->dl_brdcst_addr_offset); 5172 ether_bcopy(&fcip_arpbroadcast_addr, ep); 5173 5174 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_NOTE, "sending back info req..")); 5175 qreply(wq, mp); 5176 } 5177 5178 5179 /* 5180 * To handle DL_UNITDATA_REQ requests. 5181 */ 5182 5183 static void 5184 fcip_udreq(queue_t *wq, mblk_t *mp) 5185 { 5186 struct fcipstr *slp; 5187 struct fcip *fptr; 5188 fcip_port_info_t *fport; 5189 dl_unitdata_req_t *dludp; 5190 mblk_t *nmp; 5191 struct fcipdladdr *dlap; 5192 fcph_network_hdr_t *headerp; 5193 llc_snap_hdr_t *lsnap; 5194 t_uscalar_t off, len; 5195 struct fcip_dest *fdestp; 5196 la_wwn_t wwn; 5197 int hdr_size; 5198 5199 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_NOTE, "inside fcip_udreq")); 5200 5201 FCIP_TNF_PROBE_1((fcip_udreq, "fcip io", /* CSTYLED */, 5202 tnf_string, msg, "fcip udreq entered")); 5203 5204 slp = (struct fcipstr *)wq->q_ptr; 5205 5206 if (slp->sl_state != DL_IDLE) { 5207 dlerrorack(wq, mp, DL_UNITDATA_REQ, DL_OUTSTATE, 0); 5208 return; 5209 } 5210 5211 fptr = slp->sl_fcip; 5212 5213 if (fptr == NULL) { 5214 dlerrorack(wq, mp, DL_UNITDATA_REQ, DL_OUTSTATE, 0); 5215 return; 5216 } 5217 5218 fport = fptr->fcip_port_info; 5219 5220 dludp = (dl_unitdata_req_t *)mp->b_rptr; 5221 off = dludp->dl_dest_addr_offset; 5222 len = dludp->dl_dest_addr_length; 5223 5224 /* 5225 * Validate destination address format 5226 */ 5227 if (!MBLKIN(mp, off, len) || (len != FCIPADDRL)) { 5228 dluderrorind(wq, mp, (mp->b_rptr + off), len, DL_BADADDR, 0); 5229 return; 5230 } 5231 5232 /* 5233 * Error if no M_DATA follows 5234 */ 5235 nmp = mp->b_cont; 5236 if (nmp == NULL) { 5237 dluderrorind(wq, mp, (mp->b_rptr + off), len, DL_BADDATA, 0); 5238 return; 5239 } 5240 dlap = (struct fcipdladdr *)(mp->b_rptr + off); 5241 5242 /* 5243 * Now get the destination structure for the remote NPORT 5244 */ 5245 ether_to_wwn(&dlap->dl_phys, &wwn); 5246 fdestp = fcip_get_dest(fptr, &wwn); 5247 5248 if (fdestp == NULL) { 5249 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_NOTE, 5250 "udreq - couldn't find dest struct for remote port"); 5251 dluderrorind(wq, mp, (mp->b_rptr + off), len, DL_BADDATA, 0)); 5252 return; 5253 } 5254 5255 /* 5256 * Network header + SAP 5257 */ 5258 hdr_size = sizeof (fcph_network_hdr_t) + sizeof (llc_snap_hdr_t); 5259 5260 /* DB_REF gives the no. of msgs pointing to this block */ 5261 if ((DB_REF(nmp) == 1) && 5262 (MBLKHEAD(nmp) >= hdr_size) && 5263 (((uintptr_t)mp->b_rptr & 0x1) == 0)) { 5264 la_wwn_t wwn; 5265 nmp->b_rptr -= hdr_size; 5266 5267 /* first put the network header */ 5268 headerp = (fcph_network_hdr_t *)nmp->b_rptr; 5269 if (ether_cmp(&dlap->dl_phys, &fcip_arpbroadcast_addr) == 0) { 5270 ether_to_wwn(&fcipnhbroadcastaddr, &wwn); 5271 } else { 5272 ether_to_wwn(&dlap->dl_phys, &wwn); 5273 } 5274 bcopy(&wwn, &headerp->net_dest_addr, sizeof (la_wwn_t)); 5275 ether_to_wwn(&fptr->fcip_macaddr, &wwn); 5276 bcopy(&wwn, &headerp->net_src_addr, sizeof (la_wwn_t)); 5277 5278 /* Now the snap header */ 5279 lsnap = (llc_snap_hdr_t *)(nmp->b_rptr + 5280 sizeof (fcph_network_hdr_t)); 5281 lsnap->dsap = 0xAA; 5282 lsnap->ssap = 0xAA; 5283 lsnap->ctrl = 0x03; 5284 lsnap->oui[0] = 0x00; 5285 lsnap->oui[1] = 0x00; /* 80 */ 5286 lsnap->oui[2] = 0x00; /* C2 */ 5287 lsnap->pid = BE_16((dlap->dl_sap)); 5288 5289 freeb(mp); 5290 mp = nmp; 5291 5292 } else { 5293 la_wwn_t wwn; 5294 5295 DB_TYPE(mp) = M_DATA; 5296 headerp = (fcph_network_hdr_t *)mp->b_rptr; 5297 5298 /* 5299 * Only fill in the low 48bits of WWN for now - we can 5300 * fill in the NAA_ID after we find the port in the 5301 * routing tables 5302 */ 5303 if (ether_cmp(&dlap->dl_phys, &fcip_arpbroadcast_addr) == 0) { 5304 ether_to_wwn(&fcipnhbroadcastaddr, &wwn); 5305 } else { 5306 ether_to_wwn(&dlap->dl_phys, &wwn); 5307 } 5308 bcopy(&wwn, &headerp->net_dest_addr, sizeof (la_wwn_t)); 5309 /* need to send our PWWN */ 5310 bcopy(&fport->fcipp_pwwn, &headerp->net_src_addr, 5311 sizeof (la_wwn_t)); 5312 5313 lsnap = (llc_snap_hdr_t *)(nmp->b_rptr + 5314 sizeof (fcph_network_hdr_t)); 5315 lsnap->dsap = 0xAA; 5316 lsnap->ssap = 0xAA; 5317 lsnap->ctrl = 0x03; 5318 lsnap->oui[0] = 0x00; 5319 lsnap->oui[1] = 0x00; 5320 lsnap->oui[2] = 0x00; 5321 lsnap->pid = BE_16(dlap->dl_sap); 5322 5323 mp->b_wptr = mp->b_rptr + hdr_size; 5324 } 5325 5326 /* 5327 * Ethernet drivers have a lot of gunk here to put the Type 5328 * information (for Ethernet encapsulation (RFC 894) or the 5329 * Length (for 802.2/802.3) - I guess we'll just ignore that 5330 * here. 5331 */ 5332 5333 /* 5334 * Start the I/O on this port. If fcip_start failed for some reason 5335 * we call putbq in fcip_start so we don't need to check the 5336 * return value from fcip_start 5337 */ 5338 (void) fcip_start(wq, mp, fptr, fdestp, KM_SLEEP); 5339 } 5340 5341 /* 5342 * DL_ATTACH_REQ: attaches a PPA with a stream. ATTACH requets are needed 5343 * for style 2 DLS providers to identify the physical medium through which 5344 * the streams communication will happen 5345 */ 5346 static void 5347 fcip_areq(queue_t *wq, mblk_t *mp) 5348 { 5349 struct fcipstr *slp; 5350 union DL_primitives *dlp; 5351 fcip_port_info_t *fport; 5352 struct fcip *fptr; 5353 int ppa; 5354 5355 slp = (struct fcipstr *)wq->q_ptr; 5356 dlp = (union DL_primitives *)mp->b_rptr; 5357 5358 if (MBLKL(mp) < DL_ATTACH_REQ_SIZE) { 5359 dlerrorack(wq, mp, DL_ATTACH_REQ, DL_BADPRIM, 0); 5360 return; 5361 } 5362 5363 if (slp->sl_state != DL_UNATTACHED) { 5364 dlerrorack(wq, mp, DL_ATTACH_REQ, DL_OUTSTATE, 0); 5365 return; 5366 } 5367 5368 ppa = dlp->attach_req.dl_ppa; 5369 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_NOTE, "attach req: ppa %x", ppa)); 5370 5371 /* 5372 * check if the PPA is valid 5373 */ 5374 5375 mutex_enter(&fcip_global_mutex); 5376 5377 for (fport = fcip_port_head; fport; fport = fport->fcipp_next) { 5378 if ((fptr = fport->fcipp_fcip) == NULL) { 5379 continue; 5380 } 5381 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_NOTE, "ppa %x, inst %x", ppa, 5382 ddi_get_instance(fptr->fcip_dip))); 5383 5384 if (ppa == ddi_get_instance(fptr->fcip_dip)) { 5385 FCIP_DEBUG(FCIP_DEBUG_DLPI, 5386 (CE_NOTE, "ppa found %x", ppa)); 5387 break; 5388 } 5389 } 5390 5391 if (fport == NULL) { 5392 FCIP_DEBUG(FCIP_DEBUG_DLPI, 5393 (CE_NOTE, "dlerrorack coz fport==NULL")); 5394 5395 mutex_exit(&fcip_global_mutex); 5396 5397 if (fc_ulp_get_port_handle(ppa) == NULL) { 5398 dlerrorack(wq, mp, DL_ATTACH_REQ, DL_BADPPA, 0); 5399 return; 5400 } 5401 5402 /* 5403 * Wait for Port attach callback to trigger. If port_detach 5404 * got in while we were waiting, then ddi_get_soft_state 5405 * will return NULL, and we'll return error. 5406 */ 5407 5408 delay(drv_usectohz(FCIP_INIT_DELAY)); 5409 mutex_enter(&fcip_global_mutex); 5410 5411 fptr = ddi_get_soft_state(fcip_softp, ppa); 5412 if (fptr == NULL) { 5413 mutex_exit(&fcip_global_mutex); 5414 dlerrorack(wq, mp, DL_ATTACH_REQ, DL_BADPPA, 0); 5415 return; 5416 } 5417 } 5418 5419 /* 5420 * set link to device and update our state 5421 */ 5422 slp->sl_fcip = fptr; 5423 slp->sl_state = DL_UNBOUND; 5424 5425 mutex_exit(&fcip_global_mutex); 5426 5427 #ifdef DEBUG 5428 mutex_enter(&fptr->fcip_mutex); 5429 if (fptr->fcip_flags & FCIP_LINK_DOWN) { 5430 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_WARN, "port not online yet")); 5431 } 5432 mutex_exit(&fptr->fcip_mutex); 5433 #endif 5434 5435 dlokack(wq, mp, DL_ATTACH_REQ); 5436 } 5437 5438 5439 /* 5440 * DL_DETACH request - detaches a PPA from a stream 5441 */ 5442 static void 5443 fcip_dreq(queue_t *wq, mblk_t *mp) 5444 { 5445 struct fcipstr *slp; 5446 5447 slp = (struct fcipstr *)wq->q_ptr; 5448 5449 if (MBLKL(mp) < DL_DETACH_REQ_SIZE) { 5450 dlerrorack(wq, mp, DL_DETACH_REQ, DL_BADPRIM, 0); 5451 return; 5452 } 5453 5454 if (slp->sl_state != DL_UNBOUND) { 5455 dlerrorack(wq, mp, DL_DETACH_REQ, DL_OUTSTATE, 0); 5456 return; 5457 } 5458 5459 fcip_dodetach(slp); 5460 dlokack(wq, mp, DL_DETACH_REQ); 5461 } 5462 5463 /* 5464 * DL_BIND request: requests a DLS provider to bind a DLSAP to the stream. 5465 * DLS users communicate with a physical interface through DLSAPs. Multiple 5466 * DLSAPs can be bound to the same stream (PPA) 5467 */ 5468 static void 5469 fcip_breq(queue_t *wq, mblk_t *mp) 5470 { 5471 struct fcipstr *slp; 5472 union DL_primitives *dlp; 5473 struct fcip *fptr; 5474 struct fcipdladdr fcipaddr; 5475 t_uscalar_t sap; 5476 int xidtest; 5477 5478 slp = (struct fcipstr *)wq->q_ptr; 5479 5480 if (MBLKL(mp) < DL_BIND_REQ_SIZE) { 5481 dlerrorack(wq, mp, DL_BIND_REQ, DL_BADPRIM, 0); 5482 return; 5483 } 5484 5485 if (slp->sl_state != DL_UNBOUND) { 5486 dlerrorack(wq, mp, DL_BIND_REQ, DL_OUTSTATE, 0); 5487 return; 5488 } 5489 5490 dlp = (union DL_primitives *)mp->b_rptr; 5491 fptr = slp->sl_fcip; 5492 5493 if (fptr == NULL) { 5494 dlerrorack(wq, mp, DL_BIND_REQ, DL_OUTSTATE, 0); 5495 return; 5496 } 5497 5498 sap = dlp->bind_req.dl_sap; 5499 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_NOTE, "fcip_breq - sap: %x", sap)); 5500 xidtest = dlp->bind_req.dl_xidtest_flg; 5501 5502 if (xidtest) { 5503 dlerrorack(wq, mp, DL_BIND_REQ, DL_NOAUTO, 0); 5504 return; 5505 } 5506 5507 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_NOTE, "DLBIND: sap : %x", sap)); 5508 5509 if (sap > ETHERTYPE_MAX) { 5510 dlerrorack(wq, mp, dlp->dl_primitive, DL_BADSAP, 0); 5511 return; 5512 } 5513 /* 5514 * save SAP for this stream and change the link state 5515 */ 5516 slp->sl_sap = sap; 5517 slp->sl_state = DL_IDLE; 5518 5519 fcipaddr.dl_sap = sap; 5520 ether_bcopy(&fptr->fcip_macaddr, &fcipaddr.dl_phys); 5521 dlbindack(wq, mp, sap, &fcipaddr, FCIPADDRL, 0, 0); 5522 5523 fcip_setipq(fptr); 5524 } 5525 5526 /* 5527 * DL_UNBIND request to unbind a previously bound DLSAP, from this stream 5528 */ 5529 static void 5530 fcip_ubreq(queue_t *wq, mblk_t *mp) 5531 { 5532 struct fcipstr *slp; 5533 5534 slp = (struct fcipstr *)wq->q_ptr; 5535 5536 if (MBLKL(mp) < DL_UNBIND_REQ_SIZE) { 5537 dlerrorack(wq, mp, DL_UNBIND_REQ, DL_BADPRIM, 0); 5538 return; 5539 } 5540 5541 if (slp->sl_state != DL_IDLE) { 5542 dlerrorack(wq, mp, DL_UNBIND_REQ, DL_OUTSTATE, 0); 5543 return; 5544 } 5545 5546 slp->sl_state = DL_UNBOUND; 5547 slp->sl_sap = 0; 5548 5549 (void) putnextctl1(RD(wq), M_FLUSH, FLUSHRW); 5550 dlokack(wq, mp, DL_UNBIND_REQ); 5551 5552 fcip_setipq(slp->sl_fcip); 5553 } 5554 5555 /* 5556 * Return our physical address 5557 */ 5558 static void 5559 fcip_pareq(queue_t *wq, mblk_t *mp) 5560 { 5561 struct fcipstr *slp; 5562 union DL_primitives *dlp; 5563 int type; 5564 struct fcip *fptr; 5565 fcip_port_info_t *fport; 5566 struct ether_addr addr; 5567 5568 slp = (struct fcipstr *)wq->q_ptr; 5569 5570 if (MBLKL(mp) < DL_PHYS_ADDR_REQ_SIZE) { 5571 dlerrorack(wq, mp, DL_PHYS_ADDR_REQ, DL_BADPRIM, 0); 5572 return; 5573 } 5574 5575 dlp = (union DL_primitives *)mp->b_rptr; 5576 type = dlp->physaddr_req.dl_addr_type; 5577 fptr = slp->sl_fcip; 5578 5579 if (fptr == NULL) { 5580 dlerrorack(wq, mp, DL_PHYS_ADDR_REQ, DL_OUTSTATE, 0); 5581 return; 5582 } 5583 5584 fport = fptr->fcip_port_info; 5585 5586 switch (type) { 5587 case DL_FACT_PHYS_ADDR: 5588 FCIP_DEBUG(FCIP_DEBUG_DLPI, 5589 (CE_NOTE, "returning factory phys addr")); 5590 wwn_to_ether(&fport->fcipp_pwwn, &addr); 5591 break; 5592 5593 case DL_CURR_PHYS_ADDR: 5594 FCIP_DEBUG(FCIP_DEBUG_DLPI, 5595 (CE_NOTE, "returning current phys addr")); 5596 ether_bcopy(&fptr->fcip_macaddr, &addr); 5597 break; 5598 5599 default: 5600 FCIP_DEBUG(FCIP_DEBUG_DLPI, 5601 (CE_NOTE, "Not known cmd type in phys addr")); 5602 dlerrorack(wq, mp, DL_PHYS_ADDR_REQ, DL_NOTSUPPORTED, 0); 5603 return; 5604 } 5605 dlphysaddrack(wq, mp, &addr, ETHERADDRL); 5606 } 5607 5608 /* 5609 * Set physical address DLPI request 5610 */ 5611 static void 5612 fcip_spareq(queue_t *wq, mblk_t *mp) 5613 { 5614 struct fcipstr *slp; 5615 union DL_primitives *dlp; 5616 t_uscalar_t off, len; 5617 struct ether_addr *addrp; 5618 la_wwn_t wwn; 5619 struct fcip *fptr; 5620 fc_ns_cmd_t fcip_ns_cmd; 5621 5622 slp = (struct fcipstr *)wq->q_ptr; 5623 5624 if (MBLKL(mp) < DL_SET_PHYS_ADDR_REQ_SIZE) { 5625 dlerrorack(wq, mp, DL_SET_PHYS_ADDR_REQ, DL_BADPRIM, 0); 5626 return; 5627 } 5628 5629 dlp = (union DL_primitives *)mp->b_rptr; 5630 len = dlp->set_physaddr_req.dl_addr_length; 5631 off = dlp->set_physaddr_req.dl_addr_offset; 5632 5633 if (!MBLKIN(mp, off, len)) { 5634 dlerrorack(wq, mp, DL_SET_PHYS_ADDR_REQ, DL_BADPRIM, 0); 5635 return; 5636 } 5637 5638 addrp = (struct ether_addr *)(mp->b_rptr + off); 5639 5640 /* 5641 * If the length of physical address is not correct or address 5642 * specified is a broadcast address or multicast addr - 5643 * return an error. 5644 */ 5645 if ((len != ETHERADDRL) || 5646 ((addrp->ether_addr_octet[0] & 01) == 1) || 5647 (ether_cmp(addrp, &fcip_arpbroadcast_addr) == 0)) { 5648 dlerrorack(wq, mp, DL_SET_PHYS_ADDR_REQ, DL_BADADDR, 0); 5649 return; 5650 } 5651 5652 /* 5653 * check if a stream is attached to this device. Else return an error 5654 */ 5655 if ((fptr = slp->sl_fcip) == NULL) { 5656 dlerrorack(wq, mp, DL_SET_PHYS_ADDR_REQ, DL_OUTSTATE, 0); 5657 return; 5658 } 5659 5660 /* 5661 * set the new interface local address. We request the transport 5662 * layer to change the Port WWN for this device - return an error 5663 * if we don't succeed. 5664 */ 5665 5666 ether_to_wwn(addrp, &wwn); 5667 if (fcip_set_wwn(&wwn) == FC_SUCCESS) { 5668 FCIP_DEBUG(FCIP_DEBUG_DLPI, 5669 (CE_WARN, "WWN changed in spareq")); 5670 } else { 5671 dlerrorack(wq, mp, DL_SET_PHYS_ADDR_REQ, DL_BADADDR, 0); 5672 } 5673 5674 /* 5675 * register The new Port WWN and Node WWN with the transport 5676 * and Nameserver. Hope the transport ensures all current I/O 5677 * has stopped before actually attempting to register a new 5678 * port and Node WWN else we are hosed. Maybe a Link reset 5679 * will get everyone's attention. 5680 */ 5681 fcip_ns_cmd.ns_flags = 0; 5682 fcip_ns_cmd.ns_cmd = NS_RPN_ID; 5683 fcip_ns_cmd.ns_req_len = sizeof (la_wwn_t); 5684 fcip_ns_cmd.ns_req_payload = (caddr_t)&wwn.raw_wwn[0]; 5685 fcip_ns_cmd.ns_resp_len = 0; 5686 fcip_ns_cmd.ns_resp_payload = (caddr_t)0; 5687 if (fc_ulp_port_ns(fptr->fcip_port_info->fcipp_handle, 5688 (opaque_t)0, &fcip_ns_cmd) != FC_SUCCESS) { 5689 FCIP_DEBUG(FCIP_DEBUG_DLPI, 5690 (CE_WARN, "setting Port WWN failed")); 5691 dlerrorack(wq, mp, DL_SET_PHYS_ADDR_REQ, DL_BADPRIM, 0); 5692 return; 5693 } 5694 5695 dlokack(wq, mp, DL_SET_PHYS_ADDR_REQ); 5696 } 5697 5698 /* 5699 * change our port's WWN if permitted by hardware 5700 */ 5701 /* ARGSUSED */ 5702 static int 5703 fcip_set_wwn(la_wwn_t *pwwn) 5704 { 5705 /* 5706 * We're usually not allowed to change the WWN of adapters 5707 * but some adapters do permit us to change the WWN - don't 5708 * permit setting of WWNs (yet?) - This behavior could be 5709 * modified if needed 5710 */ 5711 return (FC_FAILURE); 5712 } 5713 5714 5715 /* 5716 * This routine fills in the header for fastpath data requests. What this 5717 * does in simple terms is, instead of sending all data through the Unitdata 5718 * request dlpi code paths (which will then append the protocol specific 5719 * header - network and snap headers in our case), the upper layers issue 5720 * a M_IOCTL with a DL_IOC_HDR_INFO request and ask the streams endpoint 5721 * driver to give the header it needs appended and the upper layer 5722 * allocates and fills in the header and calls our put routine 5723 */ 5724 static void 5725 fcip_dl_ioc_hdr_info(queue_t *wq, mblk_t *mp) 5726 { 5727 mblk_t *nmp; 5728 struct fcipstr *slp; 5729 struct fcipdladdr *dlap; 5730 dl_unitdata_req_t *dlup; 5731 fcph_network_hdr_t *headerp; 5732 la_wwn_t wwn; 5733 llc_snap_hdr_t *lsnap; 5734 struct fcip *fptr; 5735 fcip_port_info_t *fport; 5736 t_uscalar_t off, len; 5737 size_t hdrlen; 5738 int error; 5739 5740 slp = (struct fcipstr *)wq->q_ptr; 5741 fptr = slp->sl_fcip; 5742 if (fptr == NULL) { 5743 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 5744 (CE_NOTE, "dliochdr : returns EINVAL1")); 5745 miocnak(wq, mp, 0, EINVAL); 5746 return; 5747 } 5748 5749 error = miocpullup(mp, sizeof (dl_unitdata_req_t) + FCIPADDRL); 5750 if (error != 0) { 5751 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 5752 (CE_NOTE, "dliochdr : returns %d", error)); 5753 miocnak(wq, mp, 0, error); 5754 return; 5755 } 5756 5757 fport = fptr->fcip_port_info; 5758 5759 /* 5760 * check if the DL_UNITDATA_REQ destination addr has valid offset 5761 * and length values 5762 */ 5763 dlup = (dl_unitdata_req_t *)mp->b_cont->b_rptr; 5764 off = dlup->dl_dest_addr_offset; 5765 len = dlup->dl_dest_addr_length; 5766 if (dlup->dl_primitive != DL_UNITDATA_REQ || 5767 !MBLKIN(mp->b_cont, off, len) || (len != FCIPADDRL)) { 5768 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 5769 (CE_NOTE, "dliochdr : returns EINVAL2")); 5770 miocnak(wq, mp, 0, EINVAL); 5771 return; 5772 } 5773 5774 dlap = (struct fcipdladdr *)(mp->b_cont->b_rptr + off); 5775 5776 /* 5777 * Allocate a new mblk to hold the ether header 5778 */ 5779 5780 /* 5781 * setup space for network header 5782 */ 5783 hdrlen = (sizeof (llc_snap_hdr_t) + sizeof (fcph_network_hdr_t)); 5784 if ((nmp = allocb(hdrlen, BPRI_MED)) == NULL) { 5785 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 5786 (CE_NOTE, "dliochdr : returns ENOMEM")); 5787 miocnak(wq, mp, 0, ENOMEM); 5788 return; 5789 } 5790 nmp->b_wptr += hdrlen; 5791 5792 /* 5793 * Fill in the Network Hdr and LLC SNAP header; 5794 */ 5795 headerp = (fcph_network_hdr_t *)nmp->b_rptr; 5796 /* 5797 * just fill in the Node WWN here - we can fill in the NAA_ID when 5798 * we search the routing table 5799 */ 5800 if (ether_cmp(&dlap->dl_phys, &fcip_arpbroadcast_addr) == 0) { 5801 ether_to_wwn(&fcipnhbroadcastaddr, &wwn); 5802 } else { 5803 ether_to_wwn(&dlap->dl_phys, &wwn); 5804 } 5805 bcopy(&wwn, &headerp->net_dest_addr, sizeof (la_wwn_t)); 5806 bcopy(&fport->fcipp_pwwn, &headerp->net_src_addr, sizeof (la_wwn_t)); 5807 lsnap = (llc_snap_hdr_t *)(nmp->b_rptr + sizeof (fcph_network_hdr_t)); 5808 lsnap->dsap = 0xAA; 5809 lsnap->ssap = 0xAA; 5810 lsnap->ctrl = 0x03; 5811 lsnap->oui[0] = 0x00; 5812 lsnap->oui[1] = 0x00; 5813 lsnap->oui[2] = 0x00; 5814 lsnap->pid = BE_16(dlap->dl_sap); 5815 5816 /* 5817 * Link new mblk in after the "request" mblks. 5818 */ 5819 linkb(mp, nmp); 5820 5821 slp->sl_flags |= FCIP_SLFAST; 5822 5823 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 5824 (CE_NOTE, "dliochdr : returns success ")); 5825 miocack(wq, mp, msgsize(mp->b_cont), 0); 5826 } 5827 5828 5829 /* 5830 * Establish a kmem cache for fcip packets 5831 */ 5832 static int 5833 fcip_cache_constructor(void *buf, void *arg, int flags) 5834 { 5835 fcip_pkt_t *fcip_pkt = buf; 5836 fc_packet_t *fc_pkt; 5837 fcip_port_info_t *fport = (fcip_port_info_t *)arg; 5838 int (*cb) (caddr_t); 5839 struct fcip *fptr; 5840 5841 cb = (flags == KM_SLEEP) ? DDI_DMA_SLEEP : DDI_DMA_DONTWAIT; 5842 5843 ASSERT(fport != NULL); 5844 5845 fptr = fport->fcipp_fcip; 5846 5847 /* 5848 * we allocated space for our private area at the end of the 5849 * fc packet. Make sure we point to it correctly. Ideally we 5850 * should just push fc_packet_private to the beginning or end 5851 * of the fc_packet structure 5852 */ 5853 fcip_pkt->fcip_pkt_next = NULL; 5854 fcip_pkt->fcip_pkt_prev = NULL; 5855 fcip_pkt->fcip_pkt_dest = NULL; 5856 fcip_pkt->fcip_pkt_state = 0; 5857 fcip_pkt->fcip_pkt_reason = 0; 5858 fcip_pkt->fcip_pkt_flags = 0; 5859 fcip_pkt->fcip_pkt_fptr = fptr; 5860 fcip_pkt->fcip_pkt_dma_flags = 0; 5861 5862 fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 5863 fc_pkt->pkt_ulp_rscn_infop = NULL; 5864 5865 /* 5866 * We use pkt_cmd_dma for OUTBOUND requests. We don't expect 5867 * any responses for outbound IP data so no need to setup 5868 * response or data dma handles. 5869 */ 5870 if (ddi_dma_alloc_handle(fport->fcipp_dip, 5871 &fport->fcipp_cmd_dma_attr, cb, NULL, 5872 &fc_pkt->pkt_cmd_dma) != DDI_SUCCESS) { 5873 return (FCIP_FAILURE); 5874 } 5875 5876 fc_pkt->pkt_cmd_acc = fc_pkt->pkt_resp_acc = NULL; 5877 fc_pkt->pkt_fca_private = (opaque_t)((caddr_t)buf + 5878 sizeof (fcip_pkt_t)); 5879 fc_pkt->pkt_ulp_private = (opaque_t)fcip_pkt; 5880 5881 fc_pkt->pkt_cmd_cookie_cnt = fc_pkt->pkt_resp_cookie_cnt = 5882 fc_pkt->pkt_data_cookie_cnt = 0; 5883 fc_pkt->pkt_cmd_cookie = fc_pkt->pkt_resp_cookie = 5884 fc_pkt->pkt_data_cookie = NULL; 5885 5886 return (FCIP_SUCCESS); 5887 } 5888 5889 /* 5890 * destroy the fcip kmem cache 5891 */ 5892 static void 5893 fcip_cache_destructor(void *buf, void *arg) 5894 { 5895 fcip_pkt_t *fcip_pkt = (fcip_pkt_t *)buf; 5896 fc_packet_t *fc_pkt; 5897 fcip_port_info_t *fport = (fcip_port_info_t *)arg; 5898 struct fcip *fptr; 5899 5900 ASSERT(fport != NULL); 5901 5902 fptr = fport->fcipp_fcip; 5903 5904 ASSERT(fptr == fcip_pkt->fcip_pkt_fptr); 5905 fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 5906 5907 if (fc_pkt->pkt_cmd_dma) { 5908 ddi_dma_free_handle(&fc_pkt->pkt_cmd_dma); 5909 } 5910 } 5911 5912 /* 5913 * the fcip destination structure is hashed on Node WWN assuming 5914 * a NAA_ID of 0x1 (IEEE) 5915 */ 5916 static struct fcip_dest * 5917 fcip_get_dest(struct fcip *fptr, la_wwn_t *pwwn) 5918 { 5919 struct fcip_dest *fdestp = NULL; 5920 fcip_port_info_t *fport; 5921 int hash_bucket; 5922 opaque_t pd; 5923 int rval; 5924 struct fcip_routing_table *frp; 5925 la_wwn_t twwn; 5926 uint32_t *twwnp = (uint32_t *)&twwn; 5927 5928 hash_bucket = FCIP_DEST_HASH(pwwn->raw_wwn); 5929 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 5930 (CE_NOTE, "get dest hashbucket : 0x%x", hash_bucket)); 5931 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 5932 (CE_NOTE, "0x%x 0x%x 0x%x 0x%x 0x%x 0x%x", 5933 pwwn->raw_wwn[2], pwwn->raw_wwn[3], pwwn->raw_wwn[4], 5934 pwwn->raw_wwn[5], pwwn->raw_wwn[6], pwwn->raw_wwn[7])); 5935 5936 ASSERT(hash_bucket < FCIP_DEST_HASH_ELEMS); 5937 5938 if (fcip_check_port_exists(fptr)) { 5939 /* fptr is stale, return fdestp */ 5940 return (fdestp); 5941 } 5942 fport = fptr->fcip_port_info; 5943 5944 /* 5945 * First check if we have active I/Os going on with the 5946 * destination port (an entry would exist in fcip_dest hash table) 5947 */ 5948 mutex_enter(&fptr->fcip_dest_mutex); 5949 fdestp = fptr->fcip_dest[hash_bucket]; 5950 while (fdestp != NULL) { 5951 mutex_enter(&fdestp->fcipd_mutex); 5952 if (fdestp->fcipd_rtable) { 5953 if (fcip_wwn_compare(pwwn, &fdestp->fcipd_pwwn, 5954 FCIP_COMPARE_NWWN) == 0) { 5955 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 5956 (CE_NOTE, "found fdestp")); 5957 mutex_exit(&fdestp->fcipd_mutex); 5958 mutex_exit(&fptr->fcip_dest_mutex); 5959 return (fdestp); 5960 } 5961 } 5962 mutex_exit(&fdestp->fcipd_mutex); 5963 fdestp = fdestp->fcipd_next; 5964 } 5965 mutex_exit(&fptr->fcip_dest_mutex); 5966 5967 /* 5968 * We did not find the destination port information in our 5969 * active port list so search for an entry in our routing 5970 * table. 5971 */ 5972 mutex_enter(&fptr->fcip_rt_mutex); 5973 frp = fcip_lookup_rtable(fptr, pwwn, FCIP_COMPARE_NWWN); 5974 mutex_exit(&fptr->fcip_rt_mutex); 5975 5976 if (frp == NULL || (frp && (!FCIP_RTE_UNAVAIL(frp->fcipr_state)) && 5977 frp->fcipr_state != PORT_DEVICE_LOGGED_IN) || 5978 (frp && frp->fcipr_pd == NULL)) { 5979 /* 5980 * No entry for the destination port in our routing 5981 * table too. First query the transport to see if it 5982 * already has structures for the destination port in 5983 * its hash tables. This must be done for all topologies 5984 * since we could have retired entries in the hash tables 5985 * which may have to be re-added without a statechange 5986 * callback happening. Its better to try and get an entry 5987 * for the destination port rather than simply failing a 5988 * request though it may be an overkill in private loop 5989 * topologies. 5990 * If a entry for the remote port exists in the transport's 5991 * hash tables, we are fine and can add the entry to our 5992 * routing and dest hash lists, Else for fabric configs we 5993 * query the nameserver if one exists or issue FARP ELS. 5994 */ 5995 5996 /* 5997 * We need to do a PortName based Nameserver 5998 * query operation. So get the right PortWWN 5999 * for the adapter. 6000 */ 6001 bcopy(pwwn, &twwn, sizeof (la_wwn_t)); 6002 6003 /* 6004 * Try IEEE Name (Format 1) first, this is the default and 6005 * Emulex uses this format. 6006 */ 6007 pd = fc_ulp_get_remote_port(fport->fcipp_handle, 6008 &twwn, &rval, 1); 6009 6010 if (rval != FC_SUCCESS) { 6011 /* 6012 * If IEEE Name (Format 1) query failed, try IEEE 6013 * Extended Name (Format 2) which Qlogic uses. 6014 * And try port 1 on Qlogic FC-HBA first. 6015 * Note: On x86, we need to byte swap the 32-bit 6016 * word first, after the modification, swap it back. 6017 */ 6018 *twwnp = BE_32(*twwnp); 6019 twwn.w.nport_id = QLC_PORT_1_ID_BITS; 6020 twwn.w.naa_id = QLC_PORT_NAA; 6021 *twwnp = BE_32(*twwnp); 6022 pd = fc_ulp_get_remote_port(fport->fcipp_handle, 6023 &twwn, &rval, 1); 6024 } 6025 6026 if (rval != FC_SUCCESS) { 6027 /* If still failed, try port 2 on Qlogic FC-HBA. */ 6028 *twwnp = BE_32(*twwnp); 6029 twwn.w.nport_id = QLC_PORT_2_ID_BITS; 6030 *twwnp = BE_32(*twwnp); 6031 pd = fc_ulp_get_remote_port(fport->fcipp_handle, 6032 &twwn, &rval, 1); 6033 } 6034 6035 if (rval == FC_SUCCESS) { 6036 fc_portmap_t map; 6037 /* 6038 * Add the newly found destination structure 6039 * to our routing table. Create a map with 6040 * the device we found. We could ask the 6041 * transport to give us the list of all 6042 * devices connected to our port but we 6043 * probably don't need to know all the devices 6044 * so let us just constuct a list with only 6045 * one device instead. 6046 */ 6047 6048 fc_ulp_copy_portmap(&map, pd); 6049 fcip_rt_update(fptr, &map, 1); 6050 6051 mutex_enter(&fptr->fcip_rt_mutex); 6052 frp = fcip_lookup_rtable(fptr, pwwn, 6053 FCIP_COMPARE_NWWN); 6054 mutex_exit(&fptr->fcip_rt_mutex); 6055 6056 fdestp = fcip_add_dest(fptr, frp); 6057 } else if (fcip_farp_supported && 6058 (FC_TOP_EXTERNAL(fport->fcipp_topology) || 6059 (fport->fcipp_topology == FC_TOP_PT_PT))) { 6060 /* 6061 * The Name server request failed so 6062 * issue an FARP 6063 */ 6064 fdestp = fcip_do_farp(fptr, pwwn, NULL, 6065 0, 0); 6066 } else { 6067 fdestp = NULL; 6068 } 6069 } else if (frp && frp->fcipr_state == PORT_DEVICE_LOGGED_IN) { 6070 /* 6071 * Prepare a dest structure to return to caller 6072 */ 6073 fdestp = fcip_add_dest(fptr, frp); 6074 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 6075 (CE_NOTE, "in fcip get dest non fabric")); 6076 } 6077 return (fdestp); 6078 } 6079 6080 6081 /* 6082 * Endian clean WWN compare. 6083 * Returns 0 if they compare OK, else return non zero value. 6084 * flag can be bitwise OR of FCIP_COMPARE_NWWN, FCIP_COMPARE_PWWN, 6085 * FCIP_COMPARE_BROADCAST. 6086 */ 6087 static int 6088 fcip_wwn_compare(la_wwn_t *wwn1, la_wwn_t *wwn2, int flag) 6089 { 6090 int rval = 0; 6091 if ((wwn1->raw_wwn[2] != wwn2->raw_wwn[2]) || 6092 (wwn1->raw_wwn[3] != wwn2->raw_wwn[3]) || 6093 (wwn1->raw_wwn[4] != wwn2->raw_wwn[4]) || 6094 (wwn1->raw_wwn[5] != wwn2->raw_wwn[5]) || 6095 (wwn1->raw_wwn[6] != wwn2->raw_wwn[6]) || 6096 (wwn1->raw_wwn[7] != wwn2->raw_wwn[7])) { 6097 rval = 1; 6098 } else if ((flag == FCIP_COMPARE_PWWN) && 6099 (((wwn1->raw_wwn[0] & 0xf0) != (wwn2->raw_wwn[0] & 0xf0)) || 6100 (wwn1->raw_wwn[1] != wwn2->raw_wwn[1]))) { 6101 rval = 1; 6102 } 6103 return (rval); 6104 } 6105 6106 6107 /* 6108 * Add an entry for a remote port in the dest hash table. Dest hash table 6109 * has entries for ports in the routing hash table with which we decide 6110 * to establish IP communication with. The no. of entries in the dest hash 6111 * table must always be less than or equal to the entries in the routing 6112 * hash table. Every entry in the dest hash table ofcourse must have a 6113 * corresponding entry in the routing hash table 6114 */ 6115 static struct fcip_dest * 6116 fcip_add_dest(struct fcip *fptr, struct fcip_routing_table *frp) 6117 { 6118 struct fcip_dest *fdestp = NULL; 6119 la_wwn_t *pwwn; 6120 int hash_bucket; 6121 struct fcip_dest *fdest_new; 6122 6123 if (frp == NULL) { 6124 return (fdestp); 6125 } 6126 6127 pwwn = &frp->fcipr_pwwn; 6128 mutex_enter(&fptr->fcip_dest_mutex); 6129 hash_bucket = FCIP_DEST_HASH(pwwn->raw_wwn); 6130 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 6131 (CE_NOTE, "add dest hash_bucket: 0x%x", hash_bucket)); 6132 6133 ASSERT(hash_bucket < FCIP_DEST_HASH_ELEMS); 6134 6135 fdestp = fptr->fcip_dest[hash_bucket]; 6136 while (fdestp != NULL) { 6137 mutex_enter(&fdestp->fcipd_mutex); 6138 if (fdestp->fcipd_rtable) { 6139 if (fcip_wwn_compare(pwwn, &fdestp->fcipd_pwwn, 6140 FCIP_COMPARE_PWWN) == 0) { 6141 mutex_exit(&fdestp->fcipd_mutex); 6142 mutex_exit(&fptr->fcip_dest_mutex); 6143 return (fdestp); 6144 } 6145 } 6146 mutex_exit(&fdestp->fcipd_mutex); 6147 fdestp = fdestp->fcipd_next; 6148 } 6149 6150 ASSERT(fdestp == NULL); 6151 6152 fdest_new = (struct fcip_dest *) 6153 kmem_zalloc(sizeof (struct fcip_dest), KM_SLEEP); 6154 6155 mutex_init(&fdest_new->fcipd_mutex, NULL, MUTEX_DRIVER, NULL); 6156 fdest_new->fcipd_next = fptr->fcip_dest[hash_bucket]; 6157 fdest_new->fcipd_refcnt = 0; 6158 fdest_new->fcipd_rtable = frp; 6159 fdest_new->fcipd_ncmds = 0; 6160 fptr->fcip_dest[hash_bucket] = fdest_new; 6161 fdest_new->fcipd_flags = FCIP_PORT_NOTLOGGED; 6162 6163 mutex_exit(&fptr->fcip_dest_mutex); 6164 return (fdest_new); 6165 } 6166 6167 /* 6168 * Cleanup the dest hash table and remove all entries 6169 */ 6170 static void 6171 fcip_cleanup_dest(struct fcip *fptr) 6172 { 6173 struct fcip_dest *fdestp = NULL; 6174 struct fcip_dest *fdest_delp = NULL; 6175 int i; 6176 6177 mutex_enter(&fptr->fcip_dest_mutex); 6178 6179 for (i = 0; i < FCIP_DEST_HASH_ELEMS; i++) { 6180 fdestp = fptr->fcip_dest[i]; 6181 while (fdestp != NULL) { 6182 mutex_destroy(&fdestp->fcipd_mutex); 6183 fdest_delp = fdestp; 6184 fdestp = fdestp->fcipd_next; 6185 kmem_free(fdest_delp, sizeof (struct fcip_dest)); 6186 fptr->fcip_dest[i] = NULL; 6187 } 6188 } 6189 mutex_exit(&fptr->fcip_dest_mutex); 6190 } 6191 6192 6193 /* 6194 * Send FARP requests for Fabric ports when we don't have the port 6195 * we wish to talk to in our routing hash table. FARP is specially required 6196 * to talk to FC switches for inband switch management. Most FC switches 6197 * today have a switch FC IP address for IP over FC inband switch management 6198 * but the WWN and Port_ID for this traffic is not available through the 6199 * Nameservers since the switch themeselves are transparent. 6200 */ 6201 /* ARGSUSED */ 6202 static struct fcip_dest * 6203 fcip_do_farp(struct fcip *fptr, la_wwn_t *pwwn, char *ip_addr, 6204 size_t ip_addr_len, int flags) 6205 { 6206 fcip_pkt_t *fcip_pkt; 6207 fc_packet_t *fc_pkt; 6208 fcip_port_info_t *fport = fptr->fcip_port_info; 6209 la_els_farp_t farp_cmd; 6210 la_els_farp_t *fcmd; 6211 struct fcip_dest *fdestp = NULL; 6212 int rval; 6213 clock_t farp_lbolt; 6214 la_wwn_t broadcast_wwn; 6215 struct fcip_dest *bdestp; 6216 struct fcip_routing_table *frp; 6217 6218 bdestp = fcip_get_dest(fptr, &broadcast_wwn); 6219 6220 if (bdestp == NULL) { 6221 return (fdestp); 6222 } 6223 6224 fcip_pkt = fcip_ipkt_alloc(fptr, sizeof (la_els_farp_t), 6225 sizeof (la_els_farp_t), bdestp->fcipd_pd, KM_SLEEP); 6226 6227 if (fcip_pkt == NULL) { 6228 return (fdestp); 6229 } 6230 6231 fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 6232 ether_to_wwn(&fcip_arpbroadcast_addr, &broadcast_wwn); 6233 6234 mutex_enter(&bdestp->fcipd_mutex); 6235 if (bdestp->fcipd_rtable == NULL) { 6236 mutex_exit(&bdestp->fcipd_mutex); 6237 fcip_ipkt_free(fcip_pkt); 6238 return (fdestp); 6239 } 6240 6241 fcip_pkt->fcip_pkt_dest = bdestp; 6242 fc_pkt->pkt_fca_device = bdestp->fcipd_fca_dev; 6243 6244 bdestp->fcipd_ncmds++; 6245 mutex_exit(&bdestp->fcipd_mutex); 6246 6247 fcip_init_broadcast_pkt(fcip_pkt, NULL, 1); 6248 fcip_pkt->fcip_pkt_flags |= FCIP_PKT_IN_LIST; 6249 6250 /* 6251 * Now initialize the FARP payload itself 6252 */ 6253 fcmd = &farp_cmd; 6254 fcmd->ls_code.ls_code = LA_ELS_FARP_REQ; 6255 fcmd->ls_code.mbz = 0; 6256 /* 6257 * for now just match the Port WWN since the other match addr 6258 * code points are optional. We can explore matching the IP address 6259 * if needed 6260 */ 6261 if (ip_addr) { 6262 fcmd->match_addr = FARP_MATCH_WW_PN_IPv4; 6263 } else { 6264 fcmd->match_addr = FARP_MATCH_WW_PN; 6265 } 6266 6267 /* 6268 * Request the responder port to log into us - that way 6269 * the Transport is aware of the remote port when we create 6270 * an entry for it in our tables 6271 */ 6272 fcmd->resp_flags = FARP_INIT_REPLY | FARP_INIT_P_LOGI; 6273 fcmd->req_id = fport->fcipp_sid; 6274 fcmd->dest_id.port_id = fc_pkt->pkt_cmd_fhdr.d_id; 6275 bcopy(&fport->fcipp_pwwn, &fcmd->req_pwwn, sizeof (la_wwn_t)); 6276 bcopy(&fport->fcipp_nwwn, &fcmd->req_nwwn, sizeof (la_wwn_t)); 6277 bcopy(pwwn, &fcmd->resp_pwwn, sizeof (la_wwn_t)); 6278 /* 6279 * copy in source IP address if we get to know it 6280 */ 6281 if (ip_addr) { 6282 bcopy(ip_addr, fcmd->resp_ip, ip_addr_len); 6283 } 6284 6285 fc_pkt->pkt_cmdlen = sizeof (la_els_farp_t); 6286 fc_pkt->pkt_rsplen = sizeof (la_els_farp_t); 6287 fc_pkt->pkt_tran_type = FC_PKT_EXCHANGE; 6288 fc_pkt->pkt_ulp_private = (opaque_t)fcip_pkt; 6289 6290 /* 6291 * Endian safe copy 6292 */ 6293 FCIP_CP_OUT(fcmd, fc_pkt->pkt_cmd, fc_pkt->pkt_cmd_acc, 6294 sizeof (la_els_farp_t)); 6295 6296 /* 6297 * send the packet in polled mode. 6298 */ 6299 rval = fc_ulp_issue_els(fport->fcipp_handle, fc_pkt); 6300 if (rval != FC_SUCCESS) { 6301 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, (CE_WARN, 6302 "fcip_transport of farp pkt failed 0x%x", rval)); 6303 fcip_pkt->fcip_pkt_flags &= ~FCIP_PKT_IN_LIST; 6304 fcip_ipkt_free(fcip_pkt); 6305 6306 mutex_enter(&bdestp->fcipd_mutex); 6307 bdestp->fcipd_ncmds--; 6308 mutex_exit(&bdestp->fcipd_mutex); 6309 6310 return (fdestp); 6311 } 6312 6313 farp_lbolt = ddi_get_lbolt(); 6314 farp_lbolt += drv_usectohz(FCIP_FARP_TIMEOUT); 6315 6316 mutex_enter(&fptr->fcip_mutex); 6317 fptr->fcip_farp_rsp_flag = 0; 6318 while (!fptr->fcip_farp_rsp_flag) { 6319 if (cv_timedwait(&fptr->fcip_farp_cv, &fptr->fcip_mutex, 6320 farp_lbolt) == -1) { 6321 /* 6322 * No FARP response from any destination port 6323 * so bail out. 6324 */ 6325 fptr->fcip_farp_rsp_flag = 1; 6326 } else { 6327 /* 6328 * We received a FARP response - check to see if the 6329 * response was in reply to our FARP request. 6330 */ 6331 6332 mutex_enter(&fptr->fcip_rt_mutex); 6333 frp = fcip_lookup_rtable(fptr, pwwn, FCIP_COMPARE_NWWN); 6334 mutex_exit(&fptr->fcip_rt_mutex); 6335 6336 if ((frp != NULL) && 6337 !FCIP_RTE_UNAVAIL(frp->fcipr_state)) { 6338 fdestp = fcip_get_dest(fptr, pwwn); 6339 } else { 6340 /* 6341 * Not our FARP response so go back and wait 6342 * again till FARP_TIMEOUT expires 6343 */ 6344 fptr->fcip_farp_rsp_flag = 0; 6345 } 6346 } 6347 } 6348 mutex_exit(&fptr->fcip_mutex); 6349 6350 fcip_pkt->fcip_pkt_flags |= FCIP_PKT_IN_LIST; 6351 fcip_ipkt_free(fcip_pkt); 6352 mutex_enter(&bdestp->fcipd_mutex); 6353 bdestp->fcipd_ncmds--; 6354 mutex_exit(&bdestp->fcipd_mutex); 6355 return (fdestp); 6356 } 6357 6358 6359 6360 /* 6361 * Helper routine to PLOGI to a remote port we wish to talk to. 6362 * This may not be required since the port driver does logins anyway, 6363 * but this can be required in fabric cases since FARP requests/responses 6364 * don't require you to be logged in? 6365 */ 6366 6367 /* ARGSUSED */ 6368 static int 6369 fcip_do_plogi(struct fcip *fptr, struct fcip_routing_table *frp) 6370 { 6371 fcip_pkt_t *fcip_pkt; 6372 fc_packet_t *fc_pkt; 6373 fcip_port_info_t *fport = fptr->fcip_port_info; 6374 la_els_logi_t logi; 6375 int rval; 6376 fc_frame_hdr_t *fr_hdr; 6377 6378 /* 6379 * Don't bother to login for broadcast RTE entries 6380 */ 6381 if ((frp->fcipr_d_id.port_id == 0x0) || 6382 (frp->fcipr_d_id.port_id == 0xffffff)) { 6383 return (FC_FAILURE); 6384 } 6385 6386 /* 6387 * We shouldn't pound in too many logins here 6388 * 6389 */ 6390 if (frp->fcipr_state == FCIP_RT_LOGIN_PROGRESS || 6391 frp->fcipr_state == PORT_DEVICE_LOGGED_IN) { 6392 return (FC_SUCCESS); 6393 } 6394 6395 fcip_pkt = fcip_ipkt_alloc(fptr, sizeof (la_els_logi_t), 6396 sizeof (la_els_logi_t), frp->fcipr_pd, KM_SLEEP); 6397 6398 if (fcip_pkt == NULL) { 6399 return (FC_FAILURE); 6400 } 6401 6402 /* 6403 * Update back pointer for login state update 6404 */ 6405 fcip_pkt->fcip_pkt_frp = frp; 6406 frp->fcipr_state = FCIP_RT_LOGIN_PROGRESS; 6407 6408 fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 6409 6410 /* 6411 * Initialize frame header for ELS 6412 */ 6413 fr_hdr = &fc_pkt->pkt_cmd_fhdr; 6414 fr_hdr->r_ctl = R_CTL_ELS_REQ; 6415 fr_hdr->type = FC_TYPE_EXTENDED_LS; 6416 fr_hdr->f_ctl = F_CTL_SEQ_INITIATIVE | F_CTL_FIRST_SEQ; 6417 fr_hdr->df_ctl = 0; 6418 fr_hdr->s_id = fport->fcipp_sid.port_id; 6419 fr_hdr->d_id = frp->fcipr_d_id.port_id; 6420 fr_hdr->seq_cnt = 0; 6421 fr_hdr->ox_id = 0xffff; 6422 fr_hdr->rx_id = 0xffff; 6423 fr_hdr->ro = 0; 6424 6425 fc_pkt->pkt_rsplen = sizeof (la_els_logi_t); 6426 fc_pkt->pkt_comp = fcip_ipkt_callback; 6427 fc_pkt->pkt_tran_type = FC_PKT_EXCHANGE; 6428 fc_pkt->pkt_timeout = 10; /* 10 seconds */ 6429 fcip_pkt->fcip_pkt_ttl = fptr->fcip_timeout_ticks + fc_pkt->pkt_timeout; 6430 fc_pkt->pkt_ulp_private = (opaque_t)fcip_pkt; 6431 6432 /* 6433 * Everybody does class 3, so let's just set it. If the transport 6434 * knows better, it will deal with the class appropriately. 6435 */ 6436 6437 fc_pkt->pkt_tran_flags = FC_TRAN_INTR | FC_TRAN_CLASS3; 6438 6439 /* 6440 * we need only fill in the ls_code and the cmd frame header 6441 */ 6442 bzero((void *)&logi, sizeof (la_els_logi_t)); 6443 logi.ls_code.ls_code = LA_ELS_PLOGI; 6444 logi.ls_code.mbz = 0; 6445 6446 FCIP_CP_OUT((uint8_t *)&logi, fc_pkt->pkt_cmd, fc_pkt->pkt_cmd_acc, 6447 sizeof (la_els_logi_t)); 6448 6449 rval = fc_ulp_login(fport->fcipp_handle, &fc_pkt, 1); 6450 if (rval != FC_SUCCESS) { 6451 cmn_err(CE_WARN, 6452 "!fc_ulp_login failed for d_id: 0x%x, rval: 0x%x", 6453 frp->fcipr_d_id.port_id, rval); 6454 fcip_ipkt_free(fcip_pkt); 6455 } 6456 return (rval); 6457 } 6458 6459 /* 6460 * The packet callback routine - called from the transport/FCA after 6461 * it is done DMA'ing/sending out the packet contents on the wire so 6462 * that the alloc'ed packet can be freed 6463 */ 6464 static void 6465 fcip_ipkt_callback(fc_packet_t *fc_pkt) 6466 { 6467 ls_code_t logi_req; 6468 ls_code_t logi_resp; 6469 fcip_pkt_t *fcip_pkt; 6470 fc_frame_hdr_t *fr_hdr; 6471 struct fcip *fptr; 6472 fcip_port_info_t *fport; 6473 struct fcip_routing_table *frp; 6474 6475 fr_hdr = &fc_pkt->pkt_cmd_fhdr; 6476 6477 FCIP_CP_IN(fc_pkt->pkt_resp, (uint8_t *)&logi_resp, 6478 fc_pkt->pkt_resp_acc, sizeof (logi_resp)); 6479 6480 FCIP_CP_IN(fc_pkt->pkt_cmd, (uint8_t *)&logi_req, fc_pkt->pkt_cmd_acc, 6481 sizeof (logi_req)); 6482 6483 fcip_pkt = (fcip_pkt_t *)fc_pkt->pkt_ulp_private; 6484 frp = fcip_pkt->fcip_pkt_frp; 6485 fptr = fcip_pkt->fcip_pkt_fptr; 6486 fport = fptr->fcip_port_info; 6487 6488 ASSERT(logi_req.ls_code == LA_ELS_PLOGI); 6489 6490 if (fc_pkt->pkt_state != FC_PKT_SUCCESS || 6491 logi_resp.ls_code != LA_ELS_ACC) { 6492 /* EMPTY */ 6493 6494 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, (CE_WARN, 6495 "opcode : 0x%x to d_id: 0x%x failed", 6496 logi_req.ls_code, fr_hdr->d_id)); 6497 6498 mutex_enter(&fptr->fcip_rt_mutex); 6499 frp->fcipr_state = PORT_DEVICE_INVALID; 6500 frp->fcipr_invalid_timeout = fptr->fcip_timeout_ticks + 6501 (FCIP_RTE_TIMEOUT / 2); 6502 mutex_exit(&fptr->fcip_rt_mutex); 6503 } else { 6504 fc_portid_t d_id; 6505 6506 d_id.port_id = fr_hdr->d_id; 6507 d_id.priv_lilp_posit = 0; 6508 6509 /* 6510 * Update PLOGI results; FCA Handle, and Port device handles 6511 */ 6512 mutex_enter(&fptr->fcip_rt_mutex); 6513 frp->fcipr_pd = fc_pkt->pkt_pd; 6514 frp->fcipr_fca_dev = 6515 fc_ulp_get_fca_device(fport->fcipp_handle, d_id); 6516 frp->fcipr_state = PORT_DEVICE_LOGGED_IN; 6517 mutex_exit(&fptr->fcip_rt_mutex); 6518 } 6519 6520 fcip_ipkt_free(fcip_pkt); 6521 } 6522 6523 6524 /* 6525 * pkt_alloc routine for outbound IP datagrams. The cache constructor 6526 * Only initializes the pkt_cmd_dma (which is where the outbound datagram 6527 * is stuffed) since we don't expect response 6528 */ 6529 static fcip_pkt_t * 6530 fcip_pkt_alloc(struct fcip *fptr, mblk_t *bp, int flags, int datalen) 6531 { 6532 fcip_pkt_t *fcip_pkt; 6533 fc_packet_t *fc_pkt; 6534 ddi_dma_cookie_t pkt_cookie; 6535 ddi_dma_cookie_t *cp; 6536 uint32_t cnt; 6537 fcip_port_info_t *fport = fptr->fcip_port_info; 6538 6539 fcip_pkt = kmem_cache_alloc(fptr->fcip_xmit_cache, flags); 6540 if (fcip_pkt == NULL) { 6541 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, (CE_WARN, 6542 "fcip_pkt_alloc: kmem_cache_alloc failed")); 6543 return (NULL); 6544 } 6545 6546 fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 6547 fcip_pkt->fcip_pkt_fcpktp = fc_pkt; 6548 fc_pkt->pkt_tran_flags = 0; 6549 fcip_pkt->fcip_pkt_dma_flags = 0; 6550 6551 /* 6552 * the cache constructor has allocated the dma handle 6553 */ 6554 fc_pkt->pkt_cmd = (caddr_t)bp->b_rptr; 6555 if (ddi_dma_addr_bind_handle(fc_pkt->pkt_cmd_dma, NULL, 6556 (caddr_t)bp->b_rptr, datalen, DDI_DMA_WRITE | DDI_DMA_CONSISTENT, 6557 DDI_DMA_DONTWAIT, NULL, &pkt_cookie, 6558 &fc_pkt->pkt_cmd_cookie_cnt) != DDI_DMA_MAPPED) { 6559 goto fail; 6560 } 6561 6562 fcip_pkt->fcip_pkt_dma_flags |= FCIP_CMD_DMA_BOUND; 6563 6564 if (fc_pkt->pkt_cmd_cookie_cnt > 6565 fport->fcipp_cmd_dma_attr.dma_attr_sgllen) { 6566 goto fail; 6567 } 6568 6569 ASSERT(fc_pkt->pkt_cmd_cookie_cnt != 0); 6570 6571 cp = fc_pkt->pkt_cmd_cookie = (ddi_dma_cookie_t *)kmem_alloc( 6572 fc_pkt->pkt_cmd_cookie_cnt * sizeof (pkt_cookie), 6573 KM_NOSLEEP); 6574 6575 if (cp == NULL) { 6576 goto fail; 6577 } 6578 6579 *cp = pkt_cookie; 6580 cp++; 6581 for (cnt = 1; cnt < fc_pkt->pkt_cmd_cookie_cnt; cnt++, cp++) { 6582 ddi_dma_nextcookie(fc_pkt->pkt_cmd_dma, &pkt_cookie); 6583 *cp = pkt_cookie; 6584 } 6585 6586 fc_pkt->pkt_cmdlen = datalen; 6587 6588 fcip_pkt->fcip_pkt_mp = NULL; 6589 fcip_pkt->fcip_pkt_wq = NULL; 6590 fcip_pkt->fcip_pkt_dest = NULL; 6591 fcip_pkt->fcip_pkt_next = NULL; 6592 fcip_pkt->fcip_pkt_prev = NULL; 6593 fcip_pkt->fcip_pkt_state = 0; 6594 fcip_pkt->fcip_pkt_reason = 0; 6595 fcip_pkt->fcip_pkt_flags = 0; 6596 fcip_pkt->fcip_pkt_frp = NULL; 6597 6598 return (fcip_pkt); 6599 fail: 6600 if (fcip_pkt) { 6601 fcip_pkt_free(fcip_pkt, 0); 6602 } 6603 return ((fcip_pkt_t *)0); 6604 } 6605 6606 /* 6607 * Free a packet and all its associated resources 6608 */ 6609 static void 6610 fcip_pkt_free(struct fcip_pkt *fcip_pkt, int free_mblk) 6611 { 6612 fc_packet_t *fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 6613 struct fcip *fptr = fcip_pkt->fcip_pkt_fptr; 6614 6615 if (fc_pkt->pkt_cmd_cookie != NULL) { 6616 kmem_free(fc_pkt->pkt_cmd_cookie, fc_pkt->pkt_cmd_cookie_cnt * 6617 sizeof (ddi_dma_cookie_t)); 6618 fc_pkt->pkt_cmd_cookie = NULL; 6619 } 6620 6621 fcip_free_pkt_dma(fcip_pkt); 6622 if (free_mblk && fcip_pkt->fcip_pkt_mp) { 6623 freemsg(fcip_pkt->fcip_pkt_mp); 6624 fcip_pkt->fcip_pkt_mp = NULL; 6625 } 6626 6627 (void) fc_ulp_uninit_packet(fptr->fcip_port_info->fcipp_handle, fc_pkt); 6628 6629 kmem_cache_free(fptr->fcip_xmit_cache, (void *)fcip_pkt); 6630 } 6631 6632 /* 6633 * Allocate a Packet for internal driver use. This is for requests 6634 * that originate from within the driver 6635 */ 6636 static fcip_pkt_t * 6637 fcip_ipkt_alloc(struct fcip *fptr, int cmdlen, int resplen, 6638 opaque_t pd, int flags) 6639 { 6640 fcip_pkt_t *fcip_pkt; 6641 fc_packet_t *fc_pkt; 6642 int (*cb)(caddr_t); 6643 fcip_port_info_t *fport = fptr->fcip_port_info; 6644 size_t real_len; 6645 uint_t held_here = 0; 6646 ddi_dma_cookie_t pkt_cookie; 6647 ddi_dma_cookie_t *cp; 6648 uint32_t cnt; 6649 6650 cb = (flags == KM_SLEEP) ? DDI_DMA_SLEEP : DDI_DMA_DONTWAIT; 6651 6652 fcip_pkt = kmem_zalloc((sizeof (fcip_pkt_t) + 6653 fport->fcipp_fca_pkt_size), flags); 6654 6655 if (fcip_pkt == NULL) { 6656 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 6657 (CE_WARN, "pkt alloc of ineternal pkt failed")); 6658 goto fail; 6659 } 6660 6661 fcip_pkt->fcip_pkt_flags = FCIP_PKT_INTERNAL; 6662 fcip_pkt->fcip_pkt_fptr = fptr; 6663 fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 6664 fcip_pkt->fcip_pkt_fcpktp = fc_pkt; 6665 fc_pkt->pkt_tran_flags = 0; 6666 fc_pkt->pkt_cmdlen = 0; 6667 fc_pkt->pkt_rsplen = 0; 6668 fc_pkt->pkt_datalen = 0; 6669 fc_pkt->pkt_fca_private = (opaque_t)((caddr_t)fcip_pkt + 6670 sizeof (fcip_pkt_t)); 6671 fc_pkt->pkt_ulp_private = (opaque_t)fcip_pkt; 6672 6673 if (cmdlen) { 6674 if (ddi_dma_alloc_handle(fptr->fcip_dip, 6675 &fport->fcipp_cmd_dma_attr, cb, NULL, 6676 &fc_pkt->pkt_cmd_dma) != DDI_SUCCESS) { 6677 goto fail; 6678 } 6679 6680 if (ddi_dma_mem_alloc(fc_pkt->pkt_cmd_dma, cmdlen, 6681 &fport->fcipp_fca_acc_attr, DDI_DMA_CONSISTENT, 6682 cb, NULL, (caddr_t *)&fc_pkt->pkt_cmd, 6683 &real_len, &fc_pkt->pkt_cmd_acc) != DDI_SUCCESS) { 6684 goto fail; 6685 } 6686 6687 fcip_pkt->fcip_pkt_dma_flags |= FCIP_CMD_DMA_MEM; 6688 fc_pkt->pkt_cmdlen = cmdlen; 6689 6690 if (real_len < cmdlen) { 6691 goto fail; 6692 } 6693 6694 if (ddi_dma_addr_bind_handle(fc_pkt->pkt_cmd_dma, NULL, 6695 (caddr_t)fc_pkt->pkt_cmd, real_len, 6696 DDI_DMA_WRITE | DDI_DMA_CONSISTENT, cb, NULL, 6697 &pkt_cookie, &fc_pkt->pkt_cmd_cookie_cnt) != 6698 DDI_DMA_MAPPED) { 6699 goto fail; 6700 } 6701 6702 fcip_pkt->fcip_pkt_dma_flags |= FCIP_CMD_DMA_BOUND; 6703 6704 if (fc_pkt->pkt_cmd_cookie_cnt > 6705 fport->fcipp_cmd_dma_attr.dma_attr_sgllen) { 6706 goto fail; 6707 } 6708 6709 ASSERT(fc_pkt->pkt_cmd_cookie_cnt != 0); 6710 6711 cp = fc_pkt->pkt_cmd_cookie = (ddi_dma_cookie_t *)kmem_alloc( 6712 fc_pkt->pkt_cmd_cookie_cnt * sizeof (pkt_cookie), 6713 KM_NOSLEEP); 6714 6715 if (cp == NULL) { 6716 goto fail; 6717 } 6718 6719 *cp = pkt_cookie; 6720 cp++; 6721 for (cnt = 1; cnt < fc_pkt->pkt_cmd_cookie_cnt; cnt++, cp++) { 6722 ddi_dma_nextcookie(fc_pkt->pkt_cmd_dma, &pkt_cookie); 6723 *cp = pkt_cookie; 6724 } 6725 } 6726 6727 if (resplen) { 6728 if (ddi_dma_alloc_handle(fptr->fcip_dip, 6729 &fport->fcipp_resp_dma_attr, cb, NULL, 6730 &fc_pkt->pkt_resp_dma) != DDI_SUCCESS) { 6731 goto fail; 6732 } 6733 6734 if (ddi_dma_mem_alloc(fc_pkt->pkt_resp_dma, resplen, 6735 &fport->fcipp_fca_acc_attr, DDI_DMA_CONSISTENT, 6736 cb, NULL, (caddr_t *)&fc_pkt->pkt_resp, 6737 &real_len, &fc_pkt->pkt_resp_acc) != DDI_SUCCESS) { 6738 goto fail; 6739 } 6740 6741 fcip_pkt->fcip_pkt_dma_flags |= FCIP_RESP_DMA_MEM; 6742 6743 if (real_len < resplen) { 6744 goto fail; 6745 } 6746 6747 if (ddi_dma_addr_bind_handle(fc_pkt->pkt_resp_dma, NULL, 6748 (caddr_t)fc_pkt->pkt_resp, real_len, 6749 DDI_DMA_WRITE | DDI_DMA_CONSISTENT, cb, NULL, 6750 &pkt_cookie, &fc_pkt->pkt_resp_cookie_cnt) != 6751 DDI_DMA_MAPPED) { 6752 goto fail; 6753 } 6754 6755 fcip_pkt->fcip_pkt_dma_flags |= FCIP_RESP_DMA_BOUND; 6756 fc_pkt->pkt_rsplen = resplen; 6757 6758 if (fc_pkt->pkt_resp_cookie_cnt > 6759 fport->fcipp_resp_dma_attr.dma_attr_sgllen) { 6760 goto fail; 6761 } 6762 6763 ASSERT(fc_pkt->pkt_resp_cookie_cnt != 0); 6764 6765 cp = fc_pkt->pkt_resp_cookie = (ddi_dma_cookie_t *)kmem_alloc( 6766 fc_pkt->pkt_resp_cookie_cnt * sizeof (pkt_cookie), 6767 KM_NOSLEEP); 6768 6769 if (cp == NULL) { 6770 goto fail; 6771 } 6772 6773 *cp = pkt_cookie; 6774 cp++; 6775 for (cnt = 1; cnt < fc_pkt->pkt_resp_cookie_cnt; cnt++, cp++) { 6776 ddi_dma_nextcookie(fc_pkt->pkt_resp_dma, &pkt_cookie); 6777 *cp = pkt_cookie; 6778 } 6779 } 6780 6781 /* 6782 * Initialize pkt_pd prior to calling fc_ulp_init_packet 6783 */ 6784 6785 fc_pkt->pkt_pd = pd; 6786 6787 /* 6788 * Ask the FCA to bless the internal packet 6789 */ 6790 if (fc_ulp_init_packet((opaque_t)fport->fcipp_handle, 6791 fc_pkt, flags) != FC_SUCCESS) { 6792 goto fail; 6793 } 6794 6795 /* 6796 * Keep track of # of ipkts alloc-ed 6797 * This function can get called with mutex either held or not. So, we'll 6798 * grab mutex if it is not already held by this thread. 6799 * This has to be cleaned up someday. 6800 */ 6801 if (!MUTEX_HELD(&fptr->fcip_mutex)) { 6802 held_here = 1; 6803 mutex_enter(&fptr->fcip_mutex); 6804 } 6805 6806 fptr->fcip_num_ipkts_pending++; 6807 6808 if (held_here) 6809 mutex_exit(&fptr->fcip_mutex); 6810 6811 return (fcip_pkt); 6812 fail: 6813 if (fcip_pkt) { 6814 fcip_ipkt_free(fcip_pkt); 6815 } 6816 6817 return (NULL); 6818 } 6819 6820 /* 6821 * free up an internal IP packet (like a FARP pkt etc) 6822 */ 6823 static void 6824 fcip_ipkt_free(fcip_pkt_t *fcip_pkt) 6825 { 6826 fc_packet_t *fc_pkt; 6827 struct fcip *fptr = fcip_pkt->fcip_pkt_fptr; 6828 fcip_port_info_t *fport = fptr->fcip_port_info; 6829 6830 ASSERT(fptr != NULL); 6831 ASSERT(!mutex_owned(&fptr->fcip_mutex)); 6832 6833 /* One less ipkt to wait for */ 6834 mutex_enter(&fptr->fcip_mutex); 6835 if (fptr->fcip_num_ipkts_pending) /* Safety check */ 6836 fptr->fcip_num_ipkts_pending--; 6837 mutex_exit(&fptr->fcip_mutex); 6838 6839 fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 6840 6841 if (fc_pkt->pkt_cmd_cookie != NULL) { 6842 kmem_free(fc_pkt->pkt_cmd_cookie, fc_pkt->pkt_cmd_cookie_cnt * 6843 sizeof (ddi_dma_cookie_t)); 6844 fc_pkt->pkt_cmd_cookie = NULL; 6845 } 6846 6847 if (fc_pkt->pkt_resp_cookie != NULL) { 6848 kmem_free(fc_pkt->pkt_resp_cookie, fc_pkt->pkt_resp_cookie_cnt * 6849 sizeof (ddi_dma_cookie_t)); 6850 fc_pkt->pkt_resp_cookie = NULL; 6851 } 6852 6853 if (fc_ulp_uninit_packet(fport->fcipp_handle, fc_pkt) != FC_SUCCESS) { 6854 FCIP_DEBUG(FCIP_DEBUG_ELS, (CE_WARN, 6855 "fc_ulp_uninit_pkt failed for internal fc pkt 0x%p", 6856 (void *)fc_pkt)); 6857 } 6858 fcip_free_pkt_dma(fcip_pkt); 6859 kmem_free(fcip_pkt, (sizeof (fcip_pkt_t) + fport->fcipp_fca_pkt_size)); 6860 } 6861 6862 /* 6863 * initialize a unicast request. This is a misnomer because even the 6864 * broadcast requests are initialized with this routine 6865 */ 6866 static void 6867 fcip_init_unicast_pkt(fcip_pkt_t *fcip_pkt, fc_portid_t sid, fc_portid_t did, 6868 void (*comp) ()) 6869 { 6870 fc_packet_t *fc_pkt; 6871 fc_frame_hdr_t *fr_hdr; 6872 struct fcip *fptr = fcip_pkt->fcip_pkt_fptr; 6873 6874 fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 6875 fr_hdr = &fc_pkt->pkt_cmd_fhdr; 6876 6877 fr_hdr->r_ctl = R_CTL_DEVICE_DATA | R_CTL_UNSOL_DATA; 6878 fr_hdr->s_id = sid.port_id; 6879 fr_hdr->d_id = did.port_id; 6880 fr_hdr->type = FC_TYPE_IS8802_SNAP; 6881 fr_hdr->f_ctl = F_CTL_FIRST_SEQ | F_CTL_LAST_SEQ; 6882 fr_hdr->df_ctl = DF_CTL_NET_HDR; 6883 fr_hdr->seq_cnt = 0; 6884 fr_hdr->ox_id = 0xffff; 6885 fr_hdr->rx_id = 0xffff; 6886 fr_hdr->ro = 0; 6887 /* 6888 * reset all the length fields 6889 */ 6890 fc_pkt->pkt_rsplen = 0; 6891 fc_pkt->pkt_datalen = 0; 6892 fc_pkt->pkt_comp = comp; 6893 if (comp) { 6894 fc_pkt->pkt_tran_flags |= FC_TRAN_INTR; 6895 } else { 6896 fc_pkt->pkt_tran_flags |= FC_TRAN_NO_INTR; 6897 } 6898 fc_pkt->pkt_tran_type = FC_PKT_OUTBOUND | FC_PKT_IP_WRITE; 6899 fc_pkt->pkt_timeout = fcip_pkt_ttl_ticks; 6900 fcip_pkt->fcip_pkt_ttl = fptr->fcip_timeout_ticks + fc_pkt->pkt_timeout; 6901 } 6902 6903 6904 /* 6905 * Initialize a fcip_packet for broadcast data transfers 6906 */ 6907 static void 6908 fcip_init_broadcast_pkt(fcip_pkt_t *fcip_pkt, void (*comp) (), int is_els) 6909 { 6910 fc_packet_t *fc_pkt; 6911 fc_frame_hdr_t *fr_hdr; 6912 struct fcip *fptr = fcip_pkt->fcip_pkt_fptr; 6913 fcip_port_info_t *fport = fptr->fcip_port_info; 6914 uint32_t sid; 6915 uint32_t did; 6916 6917 FCIP_TNF_PROBE_1((fcip_init_broadcast_pkt, "fcip io", /* CSTYLED */, 6918 tnf_string, msg, "enter")); 6919 fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 6920 fr_hdr = &fc_pkt->pkt_cmd_fhdr; 6921 sid = fport->fcipp_sid.port_id; 6922 6923 if (is_els) { 6924 fr_hdr->r_ctl = R_CTL_ELS_REQ; 6925 } else { 6926 fr_hdr->r_ctl = R_CTL_DEVICE_DATA | R_CTL_UNSOL_DATA; 6927 } 6928 fr_hdr->s_id = sid; 6929 /* 6930 * The destination broadcast address depends on the topology 6931 * of the underlying port 6932 */ 6933 did = fptr->fcip_broadcast_did; 6934 /* 6935 * mark pkt a broadcast pkt 6936 */ 6937 fc_pkt->pkt_tran_type = FC_PKT_BROADCAST; 6938 6939 fr_hdr->d_id = did; 6940 fr_hdr->type = FC_TYPE_IS8802_SNAP; 6941 fr_hdr->f_ctl = F_CTL_FIRST_SEQ | F_CTL_LAST_SEQ | F_CTL_END_SEQ; 6942 fr_hdr->f_ctl &= ~(F_CTL_SEQ_INITIATIVE); 6943 fr_hdr->df_ctl = DF_CTL_NET_HDR; 6944 fr_hdr->seq_cnt = 0; 6945 fr_hdr->ox_id = 0xffff; 6946 fr_hdr->rx_id = 0xffff; 6947 fr_hdr->ro = 0; 6948 fc_pkt->pkt_comp = comp; 6949 6950 if (comp) { 6951 fc_pkt->pkt_tran_flags |= FC_TRAN_INTR; 6952 } else { 6953 fc_pkt->pkt_tran_flags |= FC_TRAN_NO_INTR; 6954 } 6955 6956 fc_pkt->pkt_tran_type = FC_PKT_BROADCAST; 6957 fc_pkt->pkt_timeout = fcip_pkt_ttl_ticks; 6958 fcip_pkt->fcip_pkt_ttl = fptr->fcip_timeout_ticks + fc_pkt->pkt_timeout; 6959 } 6960 6961 6962 6963 /* 6964 * Free up all DMA resources associated with an allocated packet 6965 */ 6966 static void 6967 fcip_free_pkt_dma(fcip_pkt_t *fcip_pkt) 6968 { 6969 fc_packet_t *fc_pkt; 6970 6971 fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 6972 6973 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 6974 (CE_NOTE, "in freepktdma : flags 0x%x", 6975 fcip_pkt->fcip_pkt_dma_flags)); 6976 6977 if (fcip_pkt->fcip_pkt_dma_flags & FCIP_CMD_DMA_BOUND) { 6978 (void) ddi_dma_unbind_handle(fc_pkt->pkt_cmd_dma); 6979 } 6980 if (fcip_pkt->fcip_pkt_dma_flags & FCIP_CMD_DMA_MEM) { 6981 ddi_dma_mem_free(&fc_pkt->pkt_cmd_acc); 6982 } 6983 6984 if (fcip_pkt->fcip_pkt_dma_flags & FCIP_RESP_DMA_BOUND) { 6985 (void) ddi_dma_unbind_handle(fc_pkt->pkt_resp_dma); 6986 } 6987 if (fcip_pkt->fcip_pkt_dma_flags & FCIP_RESP_DMA_MEM) { 6988 ddi_dma_mem_free(&fc_pkt->pkt_resp_acc); 6989 } 6990 /* 6991 * for internal commands, we need to free up the dma handles too. 6992 * This is done in the cache destructor for non internal cmds 6993 */ 6994 if (fcip_pkt->fcip_pkt_flags & FCIP_PKT_INTERNAL) { 6995 if (fc_pkt->pkt_cmd_dma) { 6996 ddi_dma_free_handle(&fc_pkt->pkt_cmd_dma); 6997 } 6998 if (fc_pkt->pkt_resp_dma) { 6999 ddi_dma_free_handle(&fc_pkt->pkt_resp_dma); 7000 } 7001 } 7002 } 7003 7004 7005 /* 7006 * helper routine to generate a string, given an ether addr 7007 */ 7008 static void 7009 fcip_ether_to_str(struct ether_addr *e, caddr_t s) 7010 { 7011 int i; 7012 7013 for (i = 0; i < sizeof (struct ether_addr); i++, s += 2) { 7014 FCIP_DEBUG(FCIP_DEBUG_MISC, 7015 (CE_CONT, "0x%02X:", e->ether_addr_octet[i])); 7016 (void) sprintf(s, "%02X", e->ether_addr_octet[i]); 7017 } 7018 7019 *s = '\0'; 7020 } 7021 7022 /* 7023 * When a broadcast request comes from the upper streams modules, it 7024 * is ugly to look into every datagram to figure out if it is a broadcast 7025 * datagram or a unicast packet. Instead just add the broadcast entries 7026 * into our routing and dest tables and the standard hash table look ups 7027 * will find the entries. It is a lot cleaner this way. Also Solaris ifconfig 7028 * seems to be very ethernet specific and it requires broadcasts to the 7029 * ether broadcast addr of 0xffffffffff to succeed even though we specified 7030 * in the dl_info request that our broadcast MAC addr is 0x0000000000 7031 * (can't figure out why RFC2625 did this though). So add broadcast entries 7032 * for both MAC address 7033 */ 7034 static int 7035 fcip_dest_add_broadcast_entry(struct fcip *fptr, int new_flag) 7036 { 7037 fc_portmap_t map; 7038 struct fcip_routing_table *frp; 7039 uint32_t did; 7040 la_wwn_t broadcast_wwn; 7041 7042 /* 7043 * get port_id of destination for broadcast - this is topology 7044 * dependent 7045 */ 7046 did = fptr->fcip_broadcast_did; 7047 7048 ether_to_wwn(&fcip_arpbroadcast_addr, &broadcast_wwn); 7049 bcopy((void *)&broadcast_wwn, (void *)&map.map_pwwn, sizeof (la_wwn_t)); 7050 bcopy((void *)&broadcast_wwn, (void *)&map.map_nwwn, sizeof (la_wwn_t)); 7051 7052 map.map_did.port_id = did; 7053 map.map_hard_addr.hard_addr = did; 7054 map.map_state = PORT_DEVICE_VALID; 7055 if (new_flag) { 7056 map.map_type = PORT_DEVICE_NEW; 7057 } else { 7058 map.map_type = PORT_DEVICE_CHANGED; 7059 } 7060 map.map_flags = 0; 7061 map.map_pd = NULL; 7062 bzero(&map.map_fc4_types, sizeof (map.map_fc4_types)); 7063 fcip_rt_update(fptr, &map, 1); 7064 mutex_enter(&fptr->fcip_rt_mutex); 7065 frp = fcip_lookup_rtable(fptr, &broadcast_wwn, FCIP_COMPARE_NWWN); 7066 mutex_exit(&fptr->fcip_rt_mutex); 7067 if (frp == NULL) { 7068 return (FC_FAILURE); 7069 } 7070 (void) fcip_add_dest(fptr, frp); 7071 /* 7072 * The Upper IP layers expect the traditional broadcast MAC addr 7073 * of 0xff ff ff ff ff ff to work too if we want to plumb the fcip 7074 * stream through the /etc/hostname.fcipXX file. Instead of checking 7075 * each phys addr for a match with fcip's ARP header broadcast 7076 * addr (0x00 00 00 00 00 00), its simply easier to add another 7077 * broadcast entry for 0xff ff ff ff ff ff. 7078 */ 7079 ether_to_wwn(&fcipnhbroadcastaddr, &broadcast_wwn); 7080 bcopy((void *)&broadcast_wwn, (void *)&map.map_pwwn, sizeof (la_wwn_t)); 7081 bcopy((void *)&broadcast_wwn, (void *)&map.map_nwwn, sizeof (la_wwn_t)); 7082 fcip_rt_update(fptr, &map, 1); 7083 mutex_enter(&fptr->fcip_rt_mutex); 7084 frp = fcip_lookup_rtable(fptr, &broadcast_wwn, FCIP_COMPARE_NWWN); 7085 mutex_exit(&fptr->fcip_rt_mutex); 7086 if (frp == NULL) { 7087 return (FC_FAILURE); 7088 } 7089 (void) fcip_add_dest(fptr, frp); 7090 return (FC_SUCCESS); 7091 } 7092 7093 /* 7094 * We need to obtain the D_ID of the broadcast port for transmitting all 7095 * our broadcast (and multicast) requests. The broadcast D_ID as we know 7096 * is dependent on the link topology 7097 */ 7098 static uint32_t 7099 fcip_get_broadcast_did(struct fcip *fptr) 7100 { 7101 fcip_port_info_t *fport = fptr->fcip_port_info; 7102 uint32_t did = 0; 7103 uint32_t sid; 7104 7105 FCIP_TNF_PROBE_2((fcip_get_broadcast_did, "fcip io", /* CSTYLED */, 7106 tnf_string, msg, "enter", 7107 tnf_opaque, fptr, fptr)); 7108 7109 sid = fport->fcipp_sid.port_id; 7110 7111 switch (fport->fcipp_topology) { 7112 7113 case FC_TOP_PT_PT: { 7114 fc_portmap_t *port_map = NULL; 7115 uint32_t listlen = 0; 7116 7117 if (fc_ulp_getportmap(fport->fcipp_handle, &port_map, 7118 &listlen, FC_ULP_PLOGI_DONTCARE) == FC_SUCCESS) { 7119 FCIP_DEBUG(FCIP_DEBUG_INIT, (CE_NOTE, 7120 "fcip_gpmap: listlen : 0x%x", listlen)); 7121 if (listlen == 1) { 7122 did = port_map->map_did.port_id; 7123 } 7124 } 7125 if (port_map) { 7126 kmem_free(port_map, listlen * sizeof (fc_portmap_t)); 7127 } 7128 if (listlen != 1) { 7129 /* Dummy return value */ 7130 return (0x00FFFFFF); 7131 } 7132 break; 7133 } 7134 7135 case FC_TOP_NO_NS: 7136 /* FALLTHROUGH */ 7137 case FC_TOP_FABRIC: 7138 /* 7139 * The broadcast address is the same whether or not 7140 * the switch/fabric contains a Name service. 7141 */ 7142 did = 0x00FFFFFF; 7143 break; 7144 7145 case FC_TOP_PUBLIC_LOOP: 7146 /* 7147 * The open replicate primitive must not be used. The 7148 * broadcast sequence is simply sent to ALPA 0x00. The 7149 * fabric controller then propagates the broadcast to all 7150 * other ports. The fabric propagates the broadcast by 7151 * using the OPNfr primitive. 7152 */ 7153 did = 0x00; 7154 break; 7155 7156 case FC_TOP_PRIVATE_LOOP: 7157 /* 7158 * The source port for broadcast in private loop mode 7159 * must send an OPN(fr) signal forcing all ports in the 7160 * loop to replicate the frames that they receive. 7161 */ 7162 did = 0x00FFFFFF; 7163 break; 7164 7165 case FC_TOP_UNKNOWN: 7166 /* FALLTHROUGH */ 7167 default: 7168 did = sid; 7169 FCIP_DEBUG(FCIP_DEBUG_INIT, (CE_WARN, 7170 "fcip(0x%x):unknown topology in init_broadcast_pkt", 7171 fptr->fcip_instance)); 7172 break; 7173 } 7174 FCIP_TNF_PROBE_2((fcip_get_broadcast_did, "fcip io", /* CSTYLED */, 7175 tnf_string, msg, "return", 7176 tnf_opaque, did, did)); 7177 7178 return (did); 7179 } 7180 7181 7182 /* 7183 * fcip timeout performs 2 operations: 7184 * 1. timeout any packets sent to the FCA for which a callback hasn't 7185 * happened. If you are wondering why we need a callback since all 7186 * traffic in FCIP is unidirectional, hence all exchanges are unidirectional 7187 * but wait, we can only free up the resources after we know the FCA has 7188 * DMA'ed out the data. pretty obvious eh :) 7189 * 7190 * 2. Retire and routing table entries we marked up for retiring. This is 7191 * to give the link a chance to recover instead of marking a port down 7192 * when we have lost all communication with it after a link transition 7193 */ 7194 static void 7195 fcip_timeout(void *arg) 7196 { 7197 struct fcip *fptr = (struct fcip *)arg; 7198 int i; 7199 fcip_pkt_t *fcip_pkt; 7200 struct fcip_dest *fdestp; 7201 int index; 7202 struct fcip_routing_table *frtp; 7203 int dispatch_rte_removal = 0; 7204 7205 mutex_enter(&fptr->fcip_mutex); 7206 7207 fptr->fcip_flags |= FCIP_IN_TIMEOUT; 7208 fptr->fcip_timeout_ticks += fcip_tick_incr; 7209 7210 if (fptr->fcip_flags & (FCIP_DETACHED | FCIP_DETACHING | \ 7211 FCIP_SUSPENDED | FCIP_POWER_DOWN)) { 7212 fptr->fcip_flags &= ~(FCIP_IN_TIMEOUT); 7213 mutex_exit(&fptr->fcip_mutex); 7214 return; 7215 } 7216 7217 if (fptr->fcip_port_state == FCIP_PORT_OFFLINE) { 7218 if (fptr->fcip_timeout_ticks > fptr->fcip_mark_offline) { 7219 fptr->fcip_flags |= FCIP_LINK_DOWN; 7220 } 7221 } 7222 if (!fptr->fcip_flags & FCIP_RTE_REMOVING) { 7223 dispatch_rte_removal = 1; 7224 } 7225 mutex_exit(&fptr->fcip_mutex); 7226 7227 /* 7228 * Check if we have any Invalid routing table entries in our 7229 * hashtable we have marked off for deferred removal. If any, 7230 * we can spawn a taskq thread to do the cleanup for us. We 7231 * need to avoid cleanup in the timeout thread since we may 7232 * have to wait for outstanding commands to complete before 7233 * we retire a routing table entry. Also dispatch the taskq 7234 * thread only if we are already do not have a taskq thread 7235 * dispatched. 7236 */ 7237 if (dispatch_rte_removal) { 7238 mutex_enter(&fptr->fcip_rt_mutex); 7239 for (index = 0; index < FCIP_RT_HASH_ELEMS; index++) { 7240 frtp = fptr->fcip_rtable[index]; 7241 while (frtp) { 7242 if ((frtp->fcipr_state == FCIP_RT_INVALID) && 7243 (fptr->fcip_timeout_ticks > 7244 frtp->fcipr_invalid_timeout)) { 7245 /* 7246 * If we cannot schedule a task thread 7247 * let us attempt again on the next 7248 * tick rather than call 7249 * fcip_rte_remove_deferred() from here 7250 * directly since the routine can sleep. 7251 */ 7252 frtp->fcipr_state = FCIP_RT_RETIRED; 7253 7254 mutex_enter(&fptr->fcip_mutex); 7255 fptr->fcip_flags |= FCIP_RTE_REMOVING; 7256 mutex_exit(&fptr->fcip_mutex); 7257 7258 if (taskq_dispatch(fptr->fcip_tq, 7259 fcip_rte_remove_deferred, fptr, 7260 KM_NOSLEEP) == 0) { 7261 /* 7262 * failed - so mark the entry 7263 * as invalid again. 7264 */ 7265 frtp->fcipr_state = 7266 FCIP_RT_INVALID; 7267 7268 mutex_enter(&fptr->fcip_mutex); 7269 fptr->fcip_flags &= 7270 ~FCIP_RTE_REMOVING; 7271 mutex_exit(&fptr->fcip_mutex); 7272 } 7273 } 7274 frtp = frtp->fcipr_next; 7275 } 7276 } 7277 mutex_exit(&fptr->fcip_rt_mutex); 7278 } 7279 7280 mutex_enter(&fptr->fcip_dest_mutex); 7281 7282 /* 7283 * Now timeout any packets stuck with the transport/FCA for too long 7284 */ 7285 for (i = 0; i < FCIP_DEST_HASH_ELEMS; i++) { 7286 fdestp = fptr->fcip_dest[i]; 7287 while (fdestp != NULL) { 7288 mutex_enter(&fdestp->fcipd_mutex); 7289 for (fcip_pkt = fdestp->fcipd_head; fcip_pkt != NULL; 7290 fcip_pkt = fcip_pkt->fcip_pkt_next) { 7291 if (fcip_pkt->fcip_pkt_flags & 7292 (FCIP_PKT_RETURNED | FCIP_PKT_IN_TIMEOUT | 7293 FCIP_PKT_IN_ABORT)) { 7294 continue; 7295 } 7296 if (fptr->fcip_timeout_ticks > 7297 fcip_pkt->fcip_pkt_ttl) { 7298 fcip_pkt->fcip_pkt_flags |= 7299 FCIP_PKT_IN_TIMEOUT; 7300 7301 mutex_exit(&fdestp->fcipd_mutex); 7302 if (taskq_dispatch(fptr->fcip_tq, 7303 fcip_pkt_timeout, fcip_pkt, 7304 KM_NOSLEEP) == 0) { 7305 /* 7306 * timeout immediately 7307 */ 7308 fcip_pkt_timeout(fcip_pkt); 7309 } 7310 mutex_enter(&fdestp->fcipd_mutex); 7311 /* 7312 * The linked list is altered because 7313 * of one of the following reasons: 7314 * a. Timeout code dequeued a pkt 7315 * b. Pkt completion happened 7316 * 7317 * So restart the spin starting at 7318 * the head again; This is a bit 7319 * excessive, but okay since 7320 * fcip_timeout_ticks isn't incremented 7321 * for this spin, we will skip the 7322 * not-to-be-timedout packets quickly 7323 */ 7324 fcip_pkt = fdestp->fcipd_head; 7325 if (fcip_pkt == NULL) { 7326 break; 7327 } 7328 } 7329 } 7330 mutex_exit(&fdestp->fcipd_mutex); 7331 fdestp = fdestp->fcipd_next; 7332 } 7333 } 7334 mutex_exit(&fptr->fcip_dest_mutex); 7335 7336 /* 7337 * reschedule the timeout thread 7338 */ 7339 mutex_enter(&fptr->fcip_mutex); 7340 7341 fptr->fcip_timeout_id = timeout(fcip_timeout, fptr, 7342 drv_usectohz(1000000)); 7343 fptr->fcip_flags &= ~(FCIP_IN_TIMEOUT); 7344 mutex_exit(&fptr->fcip_mutex); 7345 } 7346 7347 7348 /* 7349 * This routine is either called from taskq or directly from fcip_timeout 7350 * does the actual job of aborting the packet 7351 */ 7352 static void 7353 fcip_pkt_timeout(void *arg) 7354 { 7355 fcip_pkt_t *fcip_pkt = (fcip_pkt_t *)arg; 7356 struct fcip_dest *fdestp; 7357 struct fcip *fptr; 7358 fc_packet_t *fc_pkt; 7359 fcip_port_info_t *fport; 7360 int rval; 7361 7362 fdestp = fcip_pkt->fcip_pkt_dest; 7363 fptr = fcip_pkt->fcip_pkt_fptr; 7364 fport = fptr->fcip_port_info; 7365 fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 7366 7367 /* 7368 * try to abort the pkt 7369 */ 7370 fcip_pkt->fcip_pkt_flags |= FCIP_PKT_IN_ABORT; 7371 rval = fc_ulp_abort(fport->fcipp_handle, fc_pkt, KM_NOSLEEP); 7372 7373 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 7374 (CE_NOTE, "fc_ulp_abort returns: 0x%x", rval)); 7375 7376 if (rval == FC_SUCCESS) { 7377 ASSERT(fdestp != NULL); 7378 7379 /* 7380 * dequeue the pkt from the dest structure pkt list 7381 */ 7382 fcip_pkt->fcip_pkt_flags &= ~FCIP_PKT_IN_ABORT; 7383 mutex_enter(&fdestp->fcipd_mutex); 7384 rval = fcip_fdestp_dequeue_pkt(fdestp, fcip_pkt); 7385 ASSERT(rval == 1); 7386 mutex_exit(&fdestp->fcipd_mutex); 7387 7388 /* 7389 * Now cleanup the pkt and free the mblk 7390 */ 7391 fcip_pkt_free(fcip_pkt, 1); 7392 } else { 7393 /* 7394 * abort failed - just mark the pkt as done and 7395 * wait for it to complete in fcip_pkt_callback since 7396 * the pkt has already been xmitted by the FCA 7397 */ 7398 fcip_pkt->fcip_pkt_flags &= ~FCIP_PKT_IN_TIMEOUT; 7399 if (fcip_pkt->fcip_pkt_flags & FCIP_PKT_RETURNED) { 7400 fcip_pkt->fcip_pkt_flags &= ~FCIP_PKT_IN_ABORT; 7401 mutex_enter(&fdestp->fcipd_mutex); 7402 rval = fcip_fdestp_dequeue_pkt(fdestp, fcip_pkt); 7403 ASSERT(rval == 1); 7404 mutex_exit(&fdestp->fcipd_mutex); 7405 7406 fcip_pkt_free(fcip_pkt, 1); 7407 } 7408 return; 7409 } 7410 } 7411 7412 7413 /* 7414 * Remove a routing table entry marked for deferred removal. This routine 7415 * unlike fcip_pkt_timeout, is always called from a taskq context 7416 */ 7417 static void 7418 fcip_rte_remove_deferred(void *arg) 7419 { 7420 struct fcip *fptr = (struct fcip *)arg; 7421 int hash_bucket; 7422 struct fcip_dest *fdestp; 7423 la_wwn_t *pwwn; 7424 int index; 7425 struct fcip_routing_table *frtp, *frtp_next, *frtp_prev; 7426 7427 7428 mutex_enter(&fptr->fcip_rt_mutex); 7429 for (index = 0; index < FCIP_RT_HASH_ELEMS; index++) { 7430 frtp = fptr->fcip_rtable[index]; 7431 frtp_prev = NULL; 7432 while (frtp) { 7433 frtp_next = frtp->fcipr_next; 7434 7435 if (frtp->fcipr_state == FCIP_RT_RETIRED) { 7436 7437 pwwn = &frtp->fcipr_pwwn; 7438 /* 7439 * Get hold of destination pointer 7440 */ 7441 mutex_enter(&fptr->fcip_dest_mutex); 7442 7443 hash_bucket = FCIP_DEST_HASH(pwwn->raw_wwn); 7444 ASSERT(hash_bucket < FCIP_DEST_HASH_ELEMS); 7445 7446 fdestp = fptr->fcip_dest[hash_bucket]; 7447 while (fdestp != NULL) { 7448 mutex_enter(&fdestp->fcipd_mutex); 7449 if (fdestp->fcipd_rtable) { 7450 if (fcip_wwn_compare(pwwn, 7451 &fdestp->fcipd_pwwn, 7452 FCIP_COMPARE_PWWN) == 0) { 7453 mutex_exit( 7454 &fdestp->fcipd_mutex); 7455 break; 7456 } 7457 } 7458 mutex_exit(&fdestp->fcipd_mutex); 7459 fdestp = fdestp->fcipd_next; 7460 } 7461 7462 mutex_exit(&fptr->fcip_dest_mutex); 7463 if (fdestp == NULL) { 7464 frtp_prev = frtp; 7465 frtp = frtp_next; 7466 continue; 7467 } 7468 7469 mutex_enter(&fdestp->fcipd_mutex); 7470 if (fdestp->fcipd_ncmds) { 7471 /* 7472 * Instead of waiting to drain commands 7473 * let us revisit this RT entry in 7474 * the next pass. 7475 */ 7476 mutex_exit(&fdestp->fcipd_mutex); 7477 frtp_prev = frtp; 7478 frtp = frtp_next; 7479 continue; 7480 } 7481 7482 /* 7483 * We are clean, so remove the RTE 7484 */ 7485 fdestp->fcipd_rtable = NULL; 7486 mutex_exit(&fdestp->fcipd_mutex); 7487 7488 FCIP_TNF_PROBE_2((fcip_rte_remove_deferred, 7489 "fcip io", /* CSTYLED */, 7490 tnf_string, msg, 7491 "remove retired routing entry", 7492 tnf_int, index, index)); 7493 7494 if (frtp_prev == NULL) { 7495 /* first element */ 7496 fptr->fcip_rtable[index] = 7497 frtp->fcipr_next; 7498 } else { 7499 frtp_prev->fcipr_next = 7500 frtp->fcipr_next; 7501 } 7502 kmem_free(frtp, 7503 sizeof (struct fcip_routing_table)); 7504 7505 frtp = frtp_next; 7506 } else { 7507 frtp_prev = frtp; 7508 frtp = frtp_next; 7509 } 7510 } 7511 } 7512 mutex_exit(&fptr->fcip_rt_mutex); 7513 /* 7514 * Clear the RTE_REMOVING flag 7515 */ 7516 mutex_enter(&fptr->fcip_mutex); 7517 fptr->fcip_flags &= ~FCIP_RTE_REMOVING; 7518 mutex_exit(&fptr->fcip_mutex); 7519 } 7520 7521 /* 7522 * Walk through all the dest hash table entries and count up the total 7523 * no. of packets outstanding against a given port 7524 */ 7525 static int 7526 fcip_port_get_num_pkts(struct fcip *fptr) 7527 { 7528 int num_cmds = 0; 7529 int i; 7530 struct fcip_dest *fdestp; 7531 7532 ASSERT(mutex_owned(&fptr->fcip_dest_mutex)); 7533 7534 for (i = 0; i < FCIP_DEST_HASH_ELEMS; i++) { 7535 fdestp = fptr->fcip_dest[i]; 7536 while (fdestp != NULL) { 7537 mutex_enter(&fdestp->fcipd_mutex); 7538 7539 ASSERT(fdestp->fcipd_ncmds >= 0); 7540 7541 if (fdestp->fcipd_ncmds > 0) { 7542 num_cmds += fdestp->fcipd_ncmds; 7543 } 7544 mutex_exit(&fdestp->fcipd_mutex); 7545 fdestp = fdestp->fcipd_next; 7546 } 7547 } 7548 7549 return (num_cmds); 7550 } 7551 7552 7553 /* 7554 * Walk through the routing table for this state instance and see if there is a 7555 * PLOGI in progress for any of the entries. Return success even if we find one. 7556 */ 7557 static int 7558 fcip_plogi_in_progress(struct fcip *fptr) 7559 { 7560 int i; 7561 struct fcip_routing_table *frp; 7562 7563 ASSERT(mutex_owned(&fptr->fcip_rt_mutex)); 7564 7565 for (i = 0; i < FCIP_RT_HASH_ELEMS; i++) { 7566 frp = fptr->fcip_rtable[i]; 7567 while (frp) { 7568 if (frp->fcipr_state == FCIP_RT_LOGIN_PROGRESS) { 7569 /* Found an entry where PLOGI is in progress */ 7570 return (1); 7571 } 7572 frp = frp->fcipr_next; 7573 } 7574 } 7575 7576 return (0); 7577 } 7578 7579 /* 7580 * Walk through the fcip port global list and check if the given port exists in 7581 * the list. Returns "0" if port exists and "1" if otherwise. 7582 */ 7583 static int 7584 fcip_check_port_exists(struct fcip *fptr) 7585 { 7586 fcip_port_info_t *cur_fport; 7587 fcip_port_info_t *fport; 7588 7589 mutex_enter(&fcip_global_mutex); 7590 fport = fptr->fcip_port_info; 7591 cur_fport = fcip_port_head; 7592 while (cur_fport != NULL) { 7593 if (cur_fport == fport) { 7594 /* Found */ 7595 mutex_exit(&fcip_global_mutex); 7596 return (0); 7597 } else { 7598 cur_fport = cur_fport->fcipp_next; 7599 } 7600 } 7601 mutex_exit(&fcip_global_mutex); 7602 7603 return (1); 7604 } 7605 7606 /* 7607 * Constructor to initialize the sendup elements for callback into 7608 * modules upstream 7609 */ 7610 7611 /* ARGSUSED */ 7612 static int 7613 fcip_sendup_constructor(void *buf, void *arg, int flags) 7614 { 7615 struct fcip_sendup_elem *msg_elem = (struct fcip_sendup_elem *)buf; 7616 fcip_port_info_t *fport = (fcip_port_info_t *)arg; 7617 7618 ASSERT(fport != NULL); 7619 7620 msg_elem->fcipsu_mp = NULL; 7621 msg_elem->fcipsu_func = NULL; 7622 msg_elem->fcipsu_next = NULL; 7623 7624 return (FCIP_SUCCESS); 7625 } 7626