1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 /* 22 * Copyright 2006 Sun Microsystems, Inc. All rights reserved. 23 * Use is subject to license terms. 24 */ 25 26 #pragma ident "%Z%%M% %I% %E% SMI" 27 28 /* 29 * Data-Link Driver 30 */ 31 32 #include <sys/stropts.h> 33 #include <sys/strsun.h> 34 #include <sys/strsubr.h> 35 #include <sys/atomic.h> 36 #include <sys/mkdev.h> 37 #include <sys/vlan.h> 38 #include <sys/dld.h> 39 #include <sys/dld_impl.h> 40 #include <sys/dls_impl.h> 41 #include <inet/common.h> 42 43 static int str_constructor(void *, void *, int); 44 static void str_destructor(void *, void *); 45 static mblk_t *str_unitdata_ind(dld_str_t *, mblk_t *, boolean_t); 46 static void str_notify_promisc_on_phys(dld_str_t *); 47 static void str_notify_promisc_off_phys(dld_str_t *); 48 static void str_notify_phys_addr(dld_str_t *, const uint8_t *); 49 static void str_notify_link_up(dld_str_t *); 50 static void str_notify_link_down(dld_str_t *); 51 static void str_notify_capab_reneg(dld_str_t *); 52 static void str_notify_speed(dld_str_t *, uint32_t); 53 static void str_notify(void *, mac_notify_type_t); 54 55 static void ioc_raw(dld_str_t *, mblk_t *); 56 static void ioc_fast(dld_str_t *, mblk_t *); 57 static void ioc(dld_str_t *, mblk_t *); 58 static void dld_ioc(dld_str_t *, mblk_t *); 59 static minor_t dld_minor_hold(boolean_t); 60 static void dld_minor_rele(minor_t); 61 static void str_mdata_raw_put(dld_str_t *, mblk_t *); 62 static mblk_t *i_dld_ether_header_update_tag(mblk_t *, uint_t, uint16_t); 63 static mblk_t *i_dld_ether_header_strip_tag(mblk_t *); 64 65 static uint32_t str_count; 66 static kmem_cache_t *str_cachep; 67 static vmem_t *minor_arenap; 68 static uint32_t minor_count; 69 static mod_hash_t *str_hashp; 70 71 #define MINOR_TO_PTR(minor) ((void *)(uintptr_t)(minor)) 72 #define PTR_TO_MINOR(ptr) ((minor_t)(uintptr_t)(ptr)) 73 74 #define STR_HASHSZ 64 75 #define STR_HASH_KEY(key) ((mod_hash_key_t)(uintptr_t)(key)) 76 77 /* 78 * Some notes on entry points, flow-control, queueing and locking: 79 * 80 * This driver exports the traditional STREAMS put entry point as well as 81 * the non-STREAMS fast-path transmit routine which is provided to IP via 82 * the DL_CAPAB_POLL negotiation. The put procedure handles all control 83 * and data operations, while the fast-path routine deals only with M_DATA 84 * fast-path packets. Regardless of the entry point, all outbound packets 85 * will end up in dld_tx_single(), where they will be delivered to the MAC 86 * driver. 87 * 88 * The transmit logic operates in two modes: a "not busy" mode where the 89 * packets will be delivered to the MAC for a send attempt, or "busy" mode 90 * where they will be enqueued in the internal queue because of flow-control. 91 * Flow-control happens when the MAC driver indicates the packets couldn't 92 * be transmitted due to lack of resources (e.g. running out of descriptors). 93 * In such case, the driver will place a dummy message on its write-side 94 * STREAMS queue so that the queue is marked as "full". Any subsequent 95 * packets arriving at the driver will be enqueued in the internal queue, 96 * which is drained in the context of the service thread that gets scheduled 97 * whenever the driver is in the "busy" mode. When all packets have been 98 * successfully delivered by MAC and the internal queue is empty, it will 99 * transition to the "not busy" mode by removing the dummy message from the 100 * write-side STREAMS queue; in effect this will trigger backenabling. 101 * The sizes of q_hiwat and q_lowat are set to 1 and 0, respectively, due 102 * to the above reasons. 103 * 104 * The driver implements an internal transmit queue independent of STREAMS. 105 * This allows for flexibility and provides a fast enqueue/dequeue mechanism 106 * compared to the putq() and get() STREAMS interfaces. The only putq() and 107 * getq() operations done by the driver are those related to placing and 108 * removing the dummy message to/from the write-side STREAMS queue for flow- 109 * control purposes. 110 * 111 * Locking is done independent of STREAMS due to the driver being fully MT. 112 * Threads entering the driver (either from put or service entry points) 113 * will most likely be readers, with the exception of a few writer cases 114 * such those handling DLPI attach/detach/bind/unbind/etc. or any of the 115 * DLD-related ioctl requests. The DLPI detach case is special, because 116 * it involves freeing resources and therefore must be single-threaded. 117 * Unfortunately the readers/writers lock can't be used to protect against 118 * it, because the lock is dropped prior to the driver calling places where 119 * putnext() may be invoked, and such places may depend on those resources 120 * to exist. Because of this, the driver always completes the DLPI detach 121 * process when there are no other threads running in the driver. This is 122 * done by keeping track of the number of threads, such that the the last 123 * thread leaving the driver will finish the pending DLPI detach operation. 124 */ 125 126 /* 127 * dld_max_q_count is the queue depth threshold used to limit the number of 128 * outstanding packets or bytes allowed in the queue; once this limit is 129 * reached the driver will free any incoming ones until the queue depth 130 * drops below the threshold. 131 * 132 * This buffering is provided to accomodate clients which do not employ 133 * their own buffering scheme, and to handle occasional packet bursts. 134 * Clients which handle their own buffering will receive positive feedback 135 * from this driver as soon as it transitions into the "busy" state, i.e. 136 * when the queue is initially filled up; they will get backenabled once 137 * the queue is empty. 138 * 139 * The value chosen here is rather arbitrary; in future some intelligent 140 * heuristics may be involved which could take into account the hardware's 141 * transmit ring size, etc. 142 */ 143 uint_t dld_max_q_count = (16 * 1024 *1024); 144 145 /* 146 * dld_finddevinfo() returns the dev_info_t * corresponding to a particular 147 * dev_t. It searches str_hashp (a table of dld_str_t's) for streams that 148 * match dev_t. If a stream is found and it is attached, its dev_info_t * 149 * is returned. 150 */ 151 typedef struct i_dld_str_state_s { 152 major_t ds_major; 153 minor_t ds_minor; 154 dev_info_t *ds_dip; 155 } i_dld_str_state_t; 156 157 /* ARGSUSED */ 158 static uint_t 159 i_dld_str_walker(mod_hash_key_t key, mod_hash_val_t *val, void *arg) 160 { 161 i_dld_str_state_t *statep = arg; 162 dld_str_t *dsp = (dld_str_t *)val; 163 164 if (statep->ds_major != dsp->ds_major) 165 return (MH_WALK_CONTINUE); 166 167 ASSERT(statep->ds_minor != 0); 168 169 /* 170 * Access to ds_ppa and ds_mh need to be protected by ds_lock. 171 */ 172 rw_enter(&dsp->ds_lock, RW_READER); 173 if (statep->ds_minor <= DLD_MAX_MINOR) { 174 /* 175 * Style 1: minor can be derived from the ppa. we 176 * continue to walk until we find a matching stream 177 * in attached state. 178 */ 179 if (statep->ds_minor == DLS_PPA2MINOR(dsp->ds_ppa) && 180 dsp->ds_mh != NULL) { 181 statep->ds_dip = mac_devinfo_get(dsp->ds_mh); 182 rw_exit(&dsp->ds_lock); 183 return (MH_WALK_TERMINATE); 184 } 185 } else { 186 /* 187 * Clone: a clone minor is unique. we can terminate the 188 * walk if we find a matching stream -- even if we fail 189 * to obtain the devinfo. 190 */ 191 if (statep->ds_minor == dsp->ds_minor) { 192 if (dsp->ds_mh != NULL) 193 statep->ds_dip = mac_devinfo_get(dsp->ds_mh); 194 rw_exit(&dsp->ds_lock); 195 return (MH_WALK_TERMINATE); 196 } 197 } 198 rw_exit(&dsp->ds_lock); 199 return (MH_WALK_CONTINUE); 200 } 201 202 static dev_info_t * 203 dld_finddevinfo(dev_t dev) 204 { 205 i_dld_str_state_t state; 206 207 state.ds_minor = getminor(dev); 208 state.ds_major = getmajor(dev); 209 state.ds_dip = NULL; 210 211 if (state.ds_minor == 0) 212 return (NULL); 213 214 mod_hash_walk(str_hashp, i_dld_str_walker, &state); 215 return (state.ds_dip); 216 } 217 218 219 /* 220 * devo_getinfo: getinfo(9e) 221 */ 222 /*ARGSUSED*/ 223 int 224 dld_getinfo(dev_info_t *dip, ddi_info_cmd_t cmd, void *arg, void **resp) 225 { 226 dev_info_t *devinfo; 227 minor_t minor = getminor((dev_t)arg); 228 int rc = DDI_FAILURE; 229 230 switch (cmd) { 231 case DDI_INFO_DEVT2DEVINFO: 232 if ((devinfo = dld_finddevinfo((dev_t)arg)) != NULL) { 233 *(dev_info_t **)resp = devinfo; 234 rc = DDI_SUCCESS; 235 } 236 break; 237 case DDI_INFO_DEVT2INSTANCE: 238 if (minor > 0 && minor <= DLD_MAX_MINOR) { 239 *resp = (void *)(uintptr_t)DLS_MINOR2INST(minor); 240 rc = DDI_SUCCESS; 241 } else if (minor > DLD_MAX_MINOR && 242 (devinfo = dld_finddevinfo((dev_t)arg)) != NULL) { 243 *resp = (void *)(uintptr_t)ddi_get_instance(devinfo); 244 rc = DDI_SUCCESS; 245 } 246 break; 247 } 248 return (rc); 249 } 250 251 /* 252 * qi_qopen: open(9e) 253 */ 254 /*ARGSUSED*/ 255 int 256 dld_open(queue_t *rq, dev_t *devp, int flag, int sflag, cred_t *credp) 257 { 258 dld_str_t *dsp; 259 major_t major; 260 minor_t minor; 261 int err; 262 263 if (sflag == MODOPEN) 264 return (ENOTSUP); 265 266 /* 267 * This is a cloning driver and therefore each queue should only 268 * ever get opened once. 269 */ 270 if (rq->q_ptr != NULL) 271 return (EBUSY); 272 273 major = getmajor(*devp); 274 minor = getminor(*devp); 275 if (minor > DLD_MAX_MINOR) 276 return (ENODEV); 277 278 /* 279 * Create a new dld_str_t for the stream. This will grab a new minor 280 * number that will be handed back in the cloned dev_t. Creation may 281 * fail if we can't allocate the dummy mblk used for flow-control. 282 */ 283 dsp = dld_str_create(rq, DLD_DLPI, major, 284 ((minor == 0) ? DL_STYLE2 : DL_STYLE1)); 285 if (dsp == NULL) 286 return (ENOSR); 287 288 ASSERT(dsp->ds_dlstate == DL_UNATTACHED); 289 if (minor != 0) { 290 /* 291 * Style 1 open 292 */ 293 294 if ((err = dld_str_attach(dsp, (t_uscalar_t)minor - 1)) != 0) 295 goto failed; 296 ASSERT(dsp->ds_dlstate == DL_UNBOUND); 297 } else { 298 (void) qassociate(rq, -1); 299 } 300 301 /* 302 * Enable the queue srv(9e) routine. 303 */ 304 qprocson(rq); 305 306 /* 307 * Construct a cloned dev_t to hand back. 308 */ 309 *devp = makedevice(getmajor(*devp), dsp->ds_minor); 310 return (0); 311 312 failed: 313 dld_str_destroy(dsp); 314 return (err); 315 } 316 317 /* 318 * qi_qclose: close(9e) 319 */ 320 int 321 dld_close(queue_t *rq) 322 { 323 dld_str_t *dsp = rq->q_ptr; 324 325 /* 326 * Wait until pending requests are processed. 327 */ 328 mutex_enter(&dsp->ds_thr_lock); 329 while (dsp->ds_pending_cnt > 0) 330 cv_wait(&dsp->ds_pending_cv, &dsp->ds_thr_lock); 331 mutex_exit(&dsp->ds_thr_lock); 332 333 /* 334 * Disable the queue srv(9e) routine. 335 */ 336 qprocsoff(rq); 337 338 /* 339 * At this point we can not be entered by any threads via STREAMS 340 * or the direct call interface, which is available only to IP. 341 * After the interface is unplumbed, IP wouldn't have any reference 342 * to this instance, and therefore we are now effectively single 343 * threaded and don't require any lock protection. Flush all 344 * pending packets which are sitting in the transmit queue. 345 */ 346 ASSERT(dsp->ds_thr == 0); 347 dld_tx_flush(dsp); 348 349 /* 350 * This stream was open to a provider node. Check to see 351 * if it has been cleanly shut down. 352 */ 353 if (dsp->ds_dlstate != DL_UNATTACHED) { 354 /* 355 * The stream is either open to a style 1 provider or 356 * this is not clean shutdown. Detach from the PPA. 357 * (This is still ok even in the style 1 case). 358 */ 359 dld_str_detach(dsp); 360 } 361 362 dld_str_destroy(dsp); 363 return (0); 364 } 365 366 /* 367 * qi_qputp: put(9e) 368 */ 369 void 370 dld_wput(queue_t *wq, mblk_t *mp) 371 { 372 dld_str_t *dsp = (dld_str_t *)wq->q_ptr; 373 374 DLD_ENTER(dsp); 375 376 switch (DB_TYPE(mp)) { 377 case M_DATA: 378 rw_enter(&dsp->ds_lock, RW_READER); 379 if (dsp->ds_dlstate != DL_IDLE || 380 dsp->ds_mode == DLD_UNITDATA) { 381 freemsg(mp); 382 } else if (dsp->ds_mode == DLD_FASTPATH) { 383 str_mdata_fastpath_put(dsp, mp); 384 } else if (dsp->ds_mode == DLD_RAW) { 385 str_mdata_raw_put(dsp, mp); 386 } 387 rw_exit(&dsp->ds_lock); 388 break; 389 case M_PROTO: 390 case M_PCPROTO: 391 dld_proto(dsp, mp); 392 break; 393 case M_IOCTL: 394 dld_ioc(dsp, mp); 395 break; 396 case M_FLUSH: 397 if (*mp->b_rptr & FLUSHW) { 398 dld_tx_flush(dsp); 399 *mp->b_rptr &= ~FLUSHW; 400 } 401 402 if (*mp->b_rptr & FLUSHR) { 403 qreply(wq, mp); 404 } else { 405 freemsg(mp); 406 } 407 break; 408 default: 409 freemsg(mp); 410 break; 411 } 412 413 DLD_EXIT(dsp); 414 } 415 416 /* 417 * qi_srvp: srv(9e) 418 */ 419 void 420 dld_wsrv(queue_t *wq) 421 { 422 mblk_t *mp; 423 dld_str_t *dsp = wq->q_ptr; 424 425 DLD_ENTER(dsp); 426 rw_enter(&dsp->ds_lock, RW_READER); 427 /* 428 * Grab all packets (chained via b_next) off our transmit queue 429 * and try to send them all to the MAC layer. Since the queue 430 * is independent of streams, we are able to dequeue all messages 431 * at once without looping through getq() and manually chaining 432 * them. Note that the queue size parameters (byte and message 433 * counts) are cleared as well, but we postpone the backenabling 434 * until after the MAC transmit since some packets may end up 435 * back at our transmit queue. 436 */ 437 mutex_enter(&dsp->ds_tx_list_lock); 438 if ((mp = dsp->ds_tx_list_head) == NULL) { 439 ASSERT(!dsp->ds_tx_qbusy); 440 ASSERT(dsp->ds_tx_flow_mp != NULL); 441 ASSERT(dsp->ds_tx_list_head == NULL); 442 ASSERT(dsp->ds_tx_list_tail == NULL); 443 ASSERT(dsp->ds_tx_cnt == 0); 444 ASSERT(dsp->ds_tx_msgcnt == 0); 445 mutex_exit(&dsp->ds_tx_list_lock); 446 rw_exit(&dsp->ds_lock); 447 DLD_EXIT(dsp); 448 return; 449 } 450 dsp->ds_tx_list_head = dsp->ds_tx_list_tail = NULL; 451 dsp->ds_tx_cnt = dsp->ds_tx_msgcnt = 0; 452 mutex_exit(&dsp->ds_tx_list_lock); 453 454 /* 455 * Discard packets unless we are attached and bound; note that 456 * the driver mode (fastpath/raw/unitdata) is irrelevant here, 457 * because regardless of the mode all transmit will end up in 458 * dld_tx_single() where the packets may be queued. 459 */ 460 ASSERT(DB_TYPE(mp) == M_DATA); 461 if (dsp->ds_dlstate != DL_IDLE) { 462 freemsgchain(mp); 463 goto done; 464 } 465 466 /* 467 * Attempt to transmit one or more packets. If the MAC can't 468 * send them all, re-queue the packet(s) at the beginning of 469 * the transmit queue to avoid any re-ordering. 470 */ 471 if ((mp = dls_tx(dsp->ds_dc, mp)) != NULL) 472 dld_tx_enqueue(dsp, mp, B_TRUE); 473 474 done: 475 /* 476 * Grab the list lock again and check if the transmit queue is 477 * really empty; if so, lift up flow-control and backenable any 478 * writer queues. If the queue is not empty, schedule service 479 * thread to drain it. 480 */ 481 mutex_enter(&dsp->ds_tx_list_lock); 482 if (dsp->ds_tx_list_head == NULL) { 483 dsp->ds_tx_flow_mp = getq(wq); 484 ASSERT(dsp->ds_tx_flow_mp != NULL); 485 dsp->ds_tx_qbusy = B_FALSE; 486 } 487 mutex_exit(&dsp->ds_tx_list_lock); 488 489 rw_exit(&dsp->ds_lock); 490 DLD_EXIT(dsp); 491 } 492 493 void 494 dld_init_ops(struct dev_ops *ops, const char *name) 495 { 496 struct streamtab *stream; 497 struct qinit *rq, *wq; 498 struct module_info *modinfo; 499 500 modinfo = kmem_zalloc(sizeof (struct module_info), KM_SLEEP); 501 modinfo->mi_idname = kmem_zalloc(FMNAMESZ, KM_SLEEP); 502 (void) snprintf(modinfo->mi_idname, FMNAMESZ, "%s", name); 503 modinfo->mi_minpsz = 0; 504 modinfo->mi_maxpsz = 64*1024; 505 modinfo->mi_hiwat = 1; 506 modinfo->mi_lowat = 0; 507 508 rq = kmem_zalloc(sizeof (struct qinit), KM_SLEEP); 509 rq->qi_qopen = dld_open; 510 rq->qi_qclose = dld_close; 511 rq->qi_minfo = modinfo; 512 513 wq = kmem_zalloc(sizeof (struct qinit), KM_SLEEP); 514 wq->qi_putp = (pfi_t)dld_wput; 515 wq->qi_srvp = (pfi_t)dld_wsrv; 516 wq->qi_minfo = modinfo; 517 518 stream = kmem_zalloc(sizeof (struct streamtab), KM_SLEEP); 519 stream->st_rdinit = rq; 520 stream->st_wrinit = wq; 521 ops->devo_cb_ops->cb_str = stream; 522 523 ops->devo_getinfo = &dld_getinfo; 524 } 525 526 void 527 dld_fini_ops(struct dev_ops *ops) 528 { 529 struct streamtab *stream; 530 struct qinit *rq, *wq; 531 struct module_info *modinfo; 532 533 stream = ops->devo_cb_ops->cb_str; 534 rq = stream->st_rdinit; 535 wq = stream->st_wrinit; 536 modinfo = rq->qi_minfo; 537 ASSERT(wq->qi_minfo == modinfo); 538 539 kmem_free(stream, sizeof (struct streamtab)); 540 kmem_free(wq, sizeof (struct qinit)); 541 kmem_free(rq, sizeof (struct qinit)); 542 kmem_free(modinfo->mi_idname, FMNAMESZ); 543 kmem_free(modinfo, sizeof (struct module_info)); 544 } 545 546 /* 547 * Initialize this module's data structures. 548 */ 549 void 550 dld_str_init(void) 551 { 552 /* 553 * Create dld_str_t object cache. 554 */ 555 str_cachep = kmem_cache_create("dld_str_cache", sizeof (dld_str_t), 556 0, str_constructor, str_destructor, NULL, NULL, NULL, 0); 557 ASSERT(str_cachep != NULL); 558 559 /* 560 * Allocate a vmem arena to manage minor numbers. The range of the 561 * arena will be from DLD_MAX_MINOR + 1 to MAXMIN (maximum legal 562 * minor number). 563 */ 564 minor_arenap = vmem_create("dld_minor_arena", 565 MINOR_TO_PTR(DLD_MAX_MINOR + 1), MAXMIN, 1, NULL, NULL, NULL, 0, 566 VM_SLEEP | VMC_IDENTIFIER); 567 ASSERT(minor_arenap != NULL); 568 569 /* 570 * Create a hash table for maintaining dld_str_t's. 571 * The ds_minor field (the clone minor number) of a dld_str_t 572 * is used as a key for this hash table because this number is 573 * globally unique (allocated from "dld_minor_arena"). 574 */ 575 str_hashp = mod_hash_create_idhash("dld_str_hash", STR_HASHSZ, 576 mod_hash_null_valdtor); 577 } 578 579 /* 580 * Tear down this module's data structures. 581 */ 582 int 583 dld_str_fini(void) 584 { 585 /* 586 * Make sure that there are no objects in use. 587 */ 588 if (str_count != 0) 589 return (EBUSY); 590 591 /* 592 * Check to see if there are any minor numbers still in use. 593 */ 594 if (minor_count != 0) 595 return (EBUSY); 596 597 /* 598 * Destroy object cache. 599 */ 600 kmem_cache_destroy(str_cachep); 601 vmem_destroy(minor_arenap); 602 mod_hash_destroy_idhash(str_hashp); 603 return (0); 604 } 605 606 /* 607 * Create a new dld_str_t object. 608 */ 609 dld_str_t * 610 dld_str_create(queue_t *rq, uint_t type, major_t major, t_uscalar_t style) 611 { 612 dld_str_t *dsp; 613 int err; 614 615 /* 616 * Allocate an object from the cache. 617 */ 618 atomic_add_32(&str_count, 1); 619 dsp = kmem_cache_alloc(str_cachep, KM_SLEEP); 620 621 /* 622 * Allocate the dummy mblk for flow-control. 623 */ 624 dsp->ds_tx_flow_mp = allocb(1, BPRI_HI); 625 if (dsp->ds_tx_flow_mp == NULL) { 626 kmem_cache_free(str_cachep, dsp); 627 atomic_add_32(&str_count, -1); 628 return (NULL); 629 } 630 dsp->ds_type = type; 631 dsp->ds_major = major; 632 dsp->ds_style = style; 633 634 /* 635 * Initialize the queue pointers. 636 */ 637 ASSERT(RD(rq) == rq); 638 dsp->ds_rq = rq; 639 dsp->ds_wq = WR(rq); 640 rq->q_ptr = WR(rq)->q_ptr = (void *)dsp; 641 642 /* 643 * We want explicit control over our write-side STREAMS queue 644 * where the dummy mblk gets added/removed for flow-control. 645 */ 646 noenable(WR(rq)); 647 648 err = mod_hash_insert(str_hashp, STR_HASH_KEY(dsp->ds_minor), 649 (mod_hash_val_t)dsp); 650 ASSERT(err == 0); 651 return (dsp); 652 } 653 654 /* 655 * Destroy a dld_str_t object. 656 */ 657 void 658 dld_str_destroy(dld_str_t *dsp) 659 { 660 queue_t *rq; 661 queue_t *wq; 662 mod_hash_val_t val; 663 /* 664 * Clear the queue pointers. 665 */ 666 rq = dsp->ds_rq; 667 wq = dsp->ds_wq; 668 ASSERT(wq == WR(rq)); 669 670 rq->q_ptr = wq->q_ptr = NULL; 671 dsp->ds_rq = dsp->ds_wq = NULL; 672 673 ASSERT(!RW_LOCK_HELD(&dsp->ds_lock)); 674 ASSERT(MUTEX_NOT_HELD(&dsp->ds_tx_list_lock)); 675 ASSERT(dsp->ds_tx_list_head == NULL); 676 ASSERT(dsp->ds_tx_list_tail == NULL); 677 ASSERT(dsp->ds_tx_cnt == 0); 678 ASSERT(dsp->ds_tx_msgcnt == 0); 679 ASSERT(!dsp->ds_tx_qbusy); 680 681 ASSERT(MUTEX_NOT_HELD(&dsp->ds_thr_lock)); 682 ASSERT(dsp->ds_thr == 0); 683 ASSERT(dsp->ds_pending_req == NULL); 684 685 /* 686 * Reinitialize all the flags. 687 */ 688 dsp->ds_notifications = 0; 689 dsp->ds_passivestate = DLD_UNINITIALIZED; 690 dsp->ds_mode = DLD_UNITDATA; 691 692 /* 693 * Free the dummy mblk if exists. 694 */ 695 if (dsp->ds_tx_flow_mp != NULL) { 696 freeb(dsp->ds_tx_flow_mp); 697 dsp->ds_tx_flow_mp = NULL; 698 } 699 700 (void) mod_hash_remove(str_hashp, STR_HASH_KEY(dsp->ds_minor), &val); 701 ASSERT(dsp == (dld_str_t *)val); 702 703 /* 704 * Free the object back to the cache. 705 */ 706 kmem_cache_free(str_cachep, dsp); 707 atomic_add_32(&str_count, -1); 708 } 709 710 /* 711 * kmem_cache contructor function: see kmem_cache_create(9f). 712 */ 713 /*ARGSUSED*/ 714 static int 715 str_constructor(void *buf, void *cdrarg, int kmflags) 716 { 717 dld_str_t *dsp = buf; 718 719 bzero(buf, sizeof (dld_str_t)); 720 721 /* 722 * Allocate a new minor number. 723 */ 724 if ((dsp->ds_minor = dld_minor_hold(kmflags == KM_SLEEP)) == 0) 725 return (-1); 726 727 /* 728 * Initialize the DLPI state machine. 729 */ 730 dsp->ds_dlstate = DL_UNATTACHED; 731 dsp->ds_ppa = (t_uscalar_t)-1; 732 733 mutex_init(&dsp->ds_thr_lock, NULL, MUTEX_DRIVER, NULL); 734 rw_init(&dsp->ds_lock, NULL, RW_DRIVER, NULL); 735 mutex_init(&dsp->ds_tx_list_lock, NULL, MUTEX_DRIVER, NULL); 736 cv_init(&dsp->ds_pending_cv, NULL, CV_DRIVER, NULL); 737 738 return (0); 739 } 740 741 /* 742 * kmem_cache destructor function. 743 */ 744 /*ARGSUSED*/ 745 static void 746 str_destructor(void *buf, void *cdrarg) 747 { 748 dld_str_t *dsp = buf; 749 750 /* 751 * Make sure the DLPI state machine was reset. 752 */ 753 ASSERT(dsp->ds_dlstate == DL_UNATTACHED); 754 755 /* 756 * Make sure the data-link interface was closed. 757 */ 758 ASSERT(dsp->ds_mh == NULL); 759 ASSERT(dsp->ds_dc == NULL); 760 761 /* 762 * Make sure enabled notifications are cleared. 763 */ 764 ASSERT(dsp->ds_notifications == 0); 765 766 /* 767 * Make sure polling is disabled. 768 */ 769 ASSERT(!dsp->ds_polling); 770 771 /* 772 * Release the minor number. 773 */ 774 dld_minor_rele(dsp->ds_minor); 775 776 ASSERT(!RW_LOCK_HELD(&dsp->ds_lock)); 777 rw_destroy(&dsp->ds_lock); 778 779 ASSERT(MUTEX_NOT_HELD(&dsp->ds_tx_list_lock)); 780 mutex_destroy(&dsp->ds_tx_list_lock); 781 ASSERT(dsp->ds_tx_flow_mp == NULL); 782 783 ASSERT(MUTEX_NOT_HELD(&dsp->ds_thr_lock)); 784 mutex_destroy(&dsp->ds_thr_lock); 785 ASSERT(dsp->ds_pending_req == NULL); 786 ASSERT(dsp->ds_pending_op == NULL); 787 ASSERT(dsp->ds_pending_cnt == 0); 788 cv_destroy(&dsp->ds_pending_cv); 789 } 790 791 /* 792 * M_DATA put. Note that mp is a single message, not a chained message. 793 */ 794 void 795 dld_tx_single(dld_str_t *dsp, mblk_t *mp) 796 { 797 /* 798 * This function can be called from within dld or from an upper 799 * layer protocol (currently only tcp). If we are in the busy 800 * mode enqueue the packet(s) and return. Otherwise hand them 801 * over to the MAC driver for transmission; any remaining one(s) 802 * which didn't get sent will be queued. 803 * 804 * Note here that we don't grab the list lock prior to checking 805 * the busy flag. This is okay, because a missed transition 806 * will not cause any packet reordering for any particular TCP 807 * connection (which is single-threaded). The enqueue routine 808 * will atomically set the busy flag and schedule the service 809 * thread to run; the flag is only cleared by the service thread 810 * when there is no more packet to be transmitted. 811 */ 812 if (dsp->ds_tx_qbusy || (mp = dls_tx(dsp->ds_dc, mp)) != NULL) 813 dld_tx_enqueue(dsp, mp, B_FALSE); 814 } 815 816 /* 817 * Update the priority bits and VID (may need to insert tag if mp points 818 * to an untagged packet. 819 * If vid is VLAN_ID_NONE, use the VID encoded in the packet. 820 */ 821 static mblk_t * 822 i_dld_ether_header_update_tag(mblk_t *mp, uint_t pri, uint16_t vid) 823 { 824 mblk_t *hmp; 825 struct ether_vlan_header *evhp; 826 struct ether_header *ehp; 827 uint16_t old_tci = 0; 828 size_t len; 829 830 ASSERT(pri != 0 || vid != VLAN_ID_NONE); 831 832 evhp = (struct ether_vlan_header *)mp->b_rptr; 833 if (ntohs(evhp->ether_tpid) == ETHERTYPE_VLAN) { 834 /* 835 * Tagged packet, update the priority bits. 836 */ 837 old_tci = ntohs(evhp->ether_tci); 838 len = sizeof (struct ether_vlan_header); 839 840 if ((DB_REF(mp) > 1) || (MBLKL(mp) < len)) { 841 /* 842 * In case some drivers only check the db_ref 843 * count of the first mblk, we pullup the 844 * message into a single mblk. 845 */ 846 hmp = msgpullup(mp, -1); 847 if ((hmp == NULL) || (MBLKL(hmp) < len)) { 848 freemsg(hmp); 849 return (NULL); 850 } else { 851 freemsg(mp); 852 mp = hmp; 853 } 854 } 855 856 evhp = (struct ether_vlan_header *)mp->b_rptr; 857 } else { 858 /* 859 * Untagged packet. Insert the special priority tag. 860 * First allocate a header mblk. 861 */ 862 hmp = allocb(sizeof (struct ether_vlan_header), BPRI_MED); 863 if (hmp == NULL) 864 return (NULL); 865 866 evhp = (struct ether_vlan_header *)hmp->b_rptr; 867 ehp = (struct ether_header *)mp->b_rptr; 868 869 /* 870 * Copy the MAC addresses and typelen 871 */ 872 bcopy(ehp, evhp, (ETHERADDRL * 2)); 873 evhp->ether_type = ehp->ether_type; 874 evhp->ether_tpid = htons(ETHERTYPE_VLAN); 875 876 hmp->b_wptr += sizeof (struct ether_vlan_header); 877 mp->b_rptr += sizeof (struct ether_header); 878 879 /* 880 * Free the original message if it's now empty. Link the 881 * rest of messages to the header message. 882 */ 883 if (MBLKL(mp) == 0) { 884 hmp->b_cont = mp->b_cont; 885 freeb(mp); 886 } else { 887 hmp->b_cont = mp; 888 } 889 mp = hmp; 890 } 891 892 if (pri == 0) 893 pri = VLAN_PRI(old_tci); 894 if (vid == VLAN_ID_NONE) 895 vid = VLAN_ID(old_tci); 896 evhp->ether_tci = htons(VLAN_TCI(pri, VLAN_CFI(old_tci), vid)); 897 return (mp); 898 } 899 900 /* 901 * M_DATA put (IP fast-path mode) 902 */ 903 void 904 str_mdata_fastpath_put(dld_str_t *dsp, mblk_t *mp) 905 { 906 boolean_t is_ethernet = (dsp->ds_mip->mi_media == DL_ETHER); 907 mblk_t *newmp; 908 uint_t pri; 909 910 if (is_ethernet) { 911 /* 912 * Update the priority bits to the assigned priority. 913 */ 914 pri = (VLAN_MBLKPRI(mp) == 0) ? dsp->ds_pri : VLAN_MBLKPRI(mp); 915 916 if (pri != 0) { 917 newmp = i_dld_ether_header_update_tag(mp, pri, 918 VLAN_ID_NONE); 919 if (newmp == NULL) 920 goto discard; 921 mp = newmp; 922 } 923 } 924 925 dld_tx_single(dsp, mp); 926 return; 927 928 discard: 929 /* TODO: bump kstat? */ 930 freemsg(mp); 931 } 932 933 /* 934 * M_DATA put (DLIOCRAW mode) 935 */ 936 static void 937 str_mdata_raw_put(dld_str_t *dsp, mblk_t *mp) 938 { 939 boolean_t is_ethernet = (dsp->ds_mip->mi_media == DL_ETHER); 940 mblk_t *bp, *newmp; 941 size_t size; 942 mac_header_info_t mhi; 943 uint_t pri, vid; 944 945 /* 946 * Certain MAC type plugins provide an illusion for raw DLPI 947 * consumers. They pretend that the MAC layer is something that 948 * it's not for the benefit of observability tools. For example, a 949 * wifi plugin might pretend that it's Ethernet for such consumers. 950 * Here, we call into the MAC layer so that this illusion can be 951 * maintained. The plugin will optionally transform the MAC header 952 * here into something that can be passed down. The header goes 953 * from raw mode to "cooked" mode. 954 */ 955 if ((newmp = mac_header_cook(dsp->ds_mh, mp)) == NULL) 956 goto discard; 957 mp = newmp; 958 959 size = MBLKL(mp); 960 961 /* 962 * Check the packet is not too big and that any remaining 963 * fragment list is composed entirely of M_DATA messages. (We 964 * know the first fragment was M_DATA otherwise we could not 965 * have got here). 966 */ 967 for (bp = mp->b_cont; bp != NULL; bp = bp->b_cont) { 968 if (DB_TYPE(bp) != M_DATA) 969 goto discard; 970 size += MBLKL(bp); 971 } 972 973 if (dls_header_info(dsp->ds_dc, mp, &mhi) != 0) 974 goto discard; 975 976 if (size > dsp->ds_mip->mi_sdu_max + mhi.mhi_hdrsize) 977 goto discard; 978 979 if (is_ethernet) { 980 /* 981 * Discard the packet if this is a VLAN stream but the VID in 982 * the packet is not correct. 983 */ 984 vid = VLAN_ID(mhi.mhi_tci); 985 if ((dsp->ds_vid != VLAN_ID_NONE) && (vid != VLAN_ID_NONE)) 986 goto discard; 987 988 /* 989 * Discard the packet if this packet is a tagged packet 990 * but both pri and VID are 0. 991 */ 992 pri = VLAN_PRI(mhi.mhi_tci); 993 if (mhi.mhi_istagged && (pri == 0) && (vid == VLAN_ID_NONE)) 994 goto discard; 995 996 /* 997 * Update the priority bits to the per-stream priority if 998 * priority is not set in the packet. Update the VID for 999 * packets on a VLAN stream. 1000 */ 1001 pri = (pri == 0) ? dsp->ds_pri : 0; 1002 if ((pri != 0) || (dsp->ds_vid != VLAN_ID_NONE)) { 1003 if ((newmp = i_dld_ether_header_update_tag(mp, 1004 pri, dsp->ds_vid)) == NULL) { 1005 goto discard; 1006 } 1007 mp = newmp; 1008 } 1009 } 1010 1011 dld_tx_single(dsp, mp); 1012 return; 1013 1014 discard: 1015 /* TODO: bump kstat? */ 1016 freemsg(mp); 1017 } 1018 1019 /* 1020 * Process DL_ATTACH_REQ (style 2) or open(2) (style 1). 1021 */ 1022 int 1023 dld_str_attach(dld_str_t *dsp, t_uscalar_t ppa) 1024 { 1025 int err; 1026 const char *drvname; 1027 char name[MAXNAMELEN]; 1028 dls_channel_t dc; 1029 uint_t addr_length; 1030 1031 ASSERT(dsp->ds_dc == NULL); 1032 1033 if ((drvname = ddi_major_to_name(dsp->ds_major)) == NULL) 1034 return (EINVAL); 1035 1036 (void) snprintf(name, MAXNAMELEN, "%s%u", drvname, ppa); 1037 1038 if (strcmp(drvname, "aggr") != 0 && 1039 qassociate(dsp->ds_wq, DLS_PPA2INST(ppa)) != 0) 1040 return (EINVAL); 1041 1042 /* 1043 * Open a channel. 1044 */ 1045 if ((err = dls_open(name, &dc)) != 0) { 1046 (void) qassociate(dsp->ds_wq, -1); 1047 return (err); 1048 } 1049 1050 /* 1051 * Cache the MAC interface handle, a pointer to the immutable MAC 1052 * information and the current and 'factory' MAC address. 1053 */ 1054 dsp->ds_mh = dls_mac(dc); 1055 dsp->ds_mip = mac_info(dsp->ds_mh); 1056 1057 mac_unicst_get(dsp->ds_mh, dsp->ds_curr_addr); 1058 1059 addr_length = dsp->ds_mip->mi_addr_length; 1060 bcopy(dsp->ds_mip->mi_unicst_addr, dsp->ds_fact_addr, addr_length); 1061 1062 /* 1063 * Cache the interface VLAN identifier. (This will be VLAN_ID_NONE for 1064 * a non-VLAN interface). 1065 */ 1066 dsp->ds_vid = dls_vid(dc); 1067 1068 /* 1069 * Set the default packet priority. 1070 */ 1071 dsp->ds_pri = 0; 1072 1073 /* 1074 * Add a notify function so that the we get updates from the MAC. 1075 */ 1076 dsp->ds_mnh = mac_notify_add(dsp->ds_mh, str_notify, (void *)dsp); 1077 1078 dsp->ds_ppa = ppa; 1079 dsp->ds_dc = dc; 1080 dsp->ds_dlstate = DL_UNBOUND; 1081 1082 return (0); 1083 } 1084 1085 /* 1086 * Process DL_DETACH_REQ (style 2) or close(2) (style 1). Can also be called 1087 * from close(2) for style 2. 1088 */ 1089 void 1090 dld_str_detach(dld_str_t *dsp) 1091 { 1092 ASSERT(dsp->ds_thr == 0); 1093 1094 /* 1095 * Remove the notify function. 1096 */ 1097 mac_notify_remove(dsp->ds_mh, dsp->ds_mnh); 1098 1099 /* 1100 * Clear the polling and promisc flags. 1101 */ 1102 dsp->ds_polling = B_FALSE; 1103 dsp->ds_soft_ring = B_FALSE; 1104 dsp->ds_promisc = 0; 1105 1106 /* 1107 * Close the channel. 1108 */ 1109 dls_close(dsp->ds_dc); 1110 dsp->ds_ppa = (t_uscalar_t)-1; 1111 dsp->ds_dc = NULL; 1112 dsp->ds_mh = NULL; 1113 1114 (void) qassociate(dsp->ds_wq, -1); 1115 1116 /* 1117 * Re-initialize the DLPI state machine. 1118 */ 1119 dsp->ds_dlstate = DL_UNATTACHED; 1120 1121 } 1122 1123 /* 1124 * This function is only called for VLAN streams. In raw mode, we strip VLAN 1125 * tags before sending packets up to the DLS clients, with the exception of 1126 * special priority tagged packets, in that case, we set the VID to 0. 1127 * mp must be a VLAN tagged packet. 1128 */ 1129 static mblk_t * 1130 i_dld_ether_header_strip_tag(mblk_t *mp) 1131 { 1132 mblk_t *newmp; 1133 struct ether_vlan_header *evhp; 1134 uint16_t tci, new_tci; 1135 1136 ASSERT(MBLKL(mp) >= sizeof (struct ether_vlan_header)); 1137 if (DB_REF(mp) > 1) { 1138 newmp = copymsg(mp); 1139 if (newmp == NULL) 1140 return (NULL); 1141 freemsg(mp); 1142 mp = newmp; 1143 } 1144 evhp = (struct ether_vlan_header *)mp->b_rptr; 1145 1146 tci = ntohs(evhp->ether_tci); 1147 if (VLAN_PRI(tci) == 0) { 1148 /* 1149 * Priority is 0, strip the tag. 1150 */ 1151 ovbcopy(mp->b_rptr, mp->b_rptr + VLAN_TAGSZ, 2 * ETHERADDRL); 1152 mp->b_rptr += VLAN_TAGSZ; 1153 } else { 1154 /* 1155 * Priority is not 0, update the VID to 0. 1156 */ 1157 new_tci = VLAN_TCI(VLAN_PRI(tci), VLAN_CFI(tci), VLAN_ID_NONE); 1158 evhp->ether_tci = htons(new_tci); 1159 } 1160 return (mp); 1161 } 1162 1163 /* 1164 * Raw mode receive function. 1165 */ 1166 /*ARGSUSED*/ 1167 void 1168 dld_str_rx_raw(void *arg, mac_resource_handle_t mrh, mblk_t *mp, 1169 mac_header_info_t *mhip) 1170 { 1171 dld_str_t *dsp = (dld_str_t *)arg; 1172 boolean_t is_ethernet = (dsp->ds_mip->mi_media == DL_ETHER); 1173 mblk_t *next, *newmp; 1174 1175 ASSERT(mp != NULL); 1176 do { 1177 /* 1178 * Get the pointer to the next packet in the chain and then 1179 * clear b_next before the packet gets passed on. 1180 */ 1181 next = mp->b_next; 1182 mp->b_next = NULL; 1183 1184 /* 1185 * Wind back b_rptr to point at the MAC header. 1186 */ 1187 ASSERT(mp->b_rptr >= DB_BASE(mp) + mhip->mhi_hdrsize); 1188 mp->b_rptr -= mhip->mhi_hdrsize; 1189 1190 /* 1191 * Certain MAC type plugins provide an illusion for raw 1192 * DLPI consumers. They pretend that the MAC layer is 1193 * something that it's not for the benefit of observability 1194 * tools. For example, a wifi plugin might pretend that 1195 * it's Ethernet for such consumers. Here, we call into 1196 * the MAC layer so that this illusion can be maintained. 1197 * The plugin will optionally transform the MAC header here 1198 * into something that can be passed up to raw consumers. 1199 * The header goes from "cooked" mode to raw mode. 1200 */ 1201 if ((newmp = mac_header_uncook(dsp->ds_mh, mp)) == NULL) { 1202 freemsg(mp); 1203 goto next; 1204 } 1205 mp = newmp; 1206 1207 /* 1208 * Strip the VLAN tag for VLAN streams. 1209 */ 1210 if (is_ethernet && dsp->ds_vid != VLAN_ID_NONE) { 1211 newmp = i_dld_ether_header_strip_tag(mp); 1212 if (newmp == NULL) { 1213 freemsg(mp); 1214 goto next; 1215 } 1216 mp = newmp; 1217 } 1218 1219 /* 1220 * Pass the packet on. 1221 */ 1222 if (canputnext(dsp->ds_rq)) 1223 putnext(dsp->ds_rq, mp); 1224 else 1225 freemsg(mp); 1226 1227 next: 1228 /* 1229 * Move on to the next packet in the chain. 1230 */ 1231 mp = next; 1232 } while (mp != NULL); 1233 } 1234 1235 /* 1236 * Fast-path receive function. 1237 */ 1238 /*ARGSUSED*/ 1239 void 1240 dld_str_rx_fastpath(void *arg, mac_resource_handle_t mrh, mblk_t *mp, 1241 mac_header_info_t *mhip) 1242 { 1243 dld_str_t *dsp = (dld_str_t *)arg; 1244 mblk_t *next; 1245 size_t offset = 0; 1246 1247 /* 1248 * MAC header stripping rules: 1249 * - Tagged packets: 1250 * a. VLAN streams. Strip the whole VLAN header including the tag. 1251 * b. Physical streams 1252 * - VLAN packets (non-zero VID). The stream must be either a 1253 * DL_PROMISC_SAP listener or a ETHERTYPE_VLAN listener. 1254 * Strip the Ethernet header but keep the VLAN header. 1255 * - Special tagged packets (zero VID) 1256 * * The stream is either a DL_PROMISC_SAP listener or a 1257 * ETHERTYPE_VLAN listener, strip the Ethernet header but 1258 * keep the VLAN header. 1259 * * Otherwise, strip the whole VLAN header. 1260 * - Untagged packets. Strip the whole MAC header. 1261 */ 1262 if (mhip->mhi_istagged && (dsp->ds_vid == VLAN_ID_NONE) && 1263 ((dsp->ds_sap == ETHERTYPE_VLAN) || 1264 (dsp->ds_promisc & DLS_PROMISC_SAP))) { 1265 offset = VLAN_TAGSZ; 1266 } 1267 1268 ASSERT(mp != NULL); 1269 do { 1270 /* 1271 * Get the pointer to the next packet in the chain and then 1272 * clear b_next before the packet gets passed on. 1273 */ 1274 next = mp->b_next; 1275 mp->b_next = NULL; 1276 1277 /* 1278 * Wind back b_rptr to point at the VLAN header. 1279 */ 1280 ASSERT(mp->b_rptr >= DB_BASE(mp) + offset); 1281 mp->b_rptr -= offset; 1282 1283 /* 1284 * Pass the packet on. 1285 */ 1286 if (canputnext(dsp->ds_rq)) 1287 putnext(dsp->ds_rq, mp); 1288 else 1289 freemsg(mp); 1290 /* 1291 * Move on to the next packet in the chain. 1292 */ 1293 mp = next; 1294 } while (mp != NULL); 1295 } 1296 1297 /* 1298 * Default receive function (send DL_UNITDATA_IND messages). 1299 */ 1300 /*ARGSUSED*/ 1301 void 1302 dld_str_rx_unitdata(void *arg, mac_resource_handle_t mrh, mblk_t *mp, 1303 mac_header_info_t *mhip) 1304 { 1305 dld_str_t *dsp = (dld_str_t *)arg; 1306 mblk_t *ud_mp; 1307 mblk_t *next; 1308 size_t offset = 0; 1309 boolean_t strip_vlan = B_TRUE; 1310 1311 /* 1312 * See MAC header stripping rules in the dld_str_rx_fastpath() function. 1313 */ 1314 if (mhip->mhi_istagged && (dsp->ds_vid == VLAN_ID_NONE) && 1315 ((dsp->ds_sap == ETHERTYPE_VLAN) || 1316 (dsp->ds_promisc & DLS_PROMISC_SAP))) { 1317 offset = VLAN_TAGSZ; 1318 strip_vlan = B_FALSE; 1319 } 1320 1321 ASSERT(mp != NULL); 1322 do { 1323 /* 1324 * Get the pointer to the next packet in the chain and then 1325 * clear b_next before the packet gets passed on. 1326 */ 1327 next = mp->b_next; 1328 mp->b_next = NULL; 1329 1330 /* 1331 * Wind back b_rptr to point at the MAC header. 1332 */ 1333 ASSERT(mp->b_rptr >= DB_BASE(mp) + mhip->mhi_hdrsize); 1334 mp->b_rptr -= mhip->mhi_hdrsize; 1335 1336 /* 1337 * Create the DL_UNITDATA_IND M_PROTO. 1338 */ 1339 if ((ud_mp = str_unitdata_ind(dsp, mp, strip_vlan)) == NULL) { 1340 freemsgchain(mp); 1341 return; 1342 } 1343 1344 /* 1345 * Advance b_rptr to point at the payload (or the VLAN header). 1346 */ 1347 mp->b_rptr += (mhip->mhi_hdrsize - offset); 1348 1349 /* 1350 * Prepend the DL_UNITDATA_IND. 1351 */ 1352 ud_mp->b_cont = mp; 1353 1354 /* 1355 * Send the message. 1356 */ 1357 if (canputnext(dsp->ds_rq)) 1358 putnext(dsp->ds_rq, ud_mp); 1359 else 1360 freemsg(ud_mp); 1361 1362 /* 1363 * Move on to the next packet in the chain. 1364 */ 1365 mp = next; 1366 } while (mp != NULL); 1367 } 1368 1369 /* 1370 * Generate DL_NOTIFY_IND messages to notify the DLPI consumer of the 1371 * current state of the interface. 1372 */ 1373 void 1374 dld_str_notify_ind(dld_str_t *dsp) 1375 { 1376 mac_notify_type_t type; 1377 1378 for (type = 0; type < MAC_NNOTE; type++) 1379 str_notify(dsp, type); 1380 } 1381 1382 typedef struct dl_unitdata_ind_wrapper { 1383 dl_unitdata_ind_t dl_unitdata; 1384 uint8_t dl_dest_addr[MAXMACADDRLEN + sizeof (uint16_t)]; 1385 uint8_t dl_src_addr[MAXMACADDRLEN + sizeof (uint16_t)]; 1386 } dl_unitdata_ind_wrapper_t; 1387 1388 /* 1389 * Create a DL_UNITDATA_IND M_PROTO message. 1390 */ 1391 static mblk_t * 1392 str_unitdata_ind(dld_str_t *dsp, mblk_t *mp, boolean_t strip_vlan) 1393 { 1394 mblk_t *nmp; 1395 dl_unitdata_ind_wrapper_t *dlwp; 1396 dl_unitdata_ind_t *dlp; 1397 mac_header_info_t mhi; 1398 uint_t addr_length; 1399 uint8_t *daddr; 1400 uint8_t *saddr; 1401 1402 /* 1403 * Get the packet header information. 1404 */ 1405 if (dls_header_info(dsp->ds_dc, mp, &mhi) != 0) 1406 return (NULL); 1407 1408 /* 1409 * Allocate a message large enough to contain the wrapper structure 1410 * defined above. 1411 */ 1412 if ((nmp = mexchange(dsp->ds_wq, NULL, 1413 sizeof (dl_unitdata_ind_wrapper_t), M_PROTO, 1414 DL_UNITDATA_IND)) == NULL) 1415 return (NULL); 1416 1417 dlwp = (dl_unitdata_ind_wrapper_t *)nmp->b_rptr; 1418 1419 dlp = &(dlwp->dl_unitdata); 1420 ASSERT(dlp == (dl_unitdata_ind_t *)nmp->b_rptr); 1421 ASSERT(dlp->dl_primitive == DL_UNITDATA_IND); 1422 1423 /* 1424 * Copy in the destination address. 1425 */ 1426 addr_length = dsp->ds_mip->mi_addr_length; 1427 daddr = dlwp->dl_dest_addr; 1428 dlp->dl_dest_addr_offset = (uintptr_t)daddr - (uintptr_t)dlp; 1429 bcopy(mhi.mhi_daddr, daddr, addr_length); 1430 1431 /* 1432 * Set the destination DLSAP to the SAP value encoded in the packet. 1433 */ 1434 if (mhi.mhi_istagged && !strip_vlan) 1435 *(uint16_t *)(daddr + addr_length) = ETHERTYPE_VLAN; 1436 else 1437 *(uint16_t *)(daddr + addr_length) = mhi.mhi_bindsap; 1438 dlp->dl_dest_addr_length = addr_length + sizeof (uint16_t); 1439 1440 /* 1441 * If the destination address was multicast or broadcast then the 1442 * dl_group_address field should be non-zero. 1443 */ 1444 dlp->dl_group_address = (mhi.mhi_dsttype == MAC_ADDRTYPE_MULTICAST) || 1445 (mhi.mhi_dsttype == MAC_ADDRTYPE_BROADCAST); 1446 1447 /* 1448 * Copy in the source address if one exists. Some MAC types (DL_IB 1449 * for example) may not have access to source information. 1450 */ 1451 if (mhi.mhi_saddr == NULL) { 1452 dlp->dl_src_addr_offset = dlp->dl_src_addr_length = 0; 1453 } else { 1454 saddr = dlwp->dl_src_addr; 1455 dlp->dl_src_addr_offset = (uintptr_t)saddr - (uintptr_t)dlp; 1456 bcopy(mhi.mhi_saddr, saddr, addr_length); 1457 1458 /* 1459 * Set the source DLSAP to the packet ethertype. 1460 */ 1461 *(uint16_t *)(saddr + addr_length) = mhi.mhi_origsap; 1462 dlp->dl_src_addr_length = addr_length + sizeof (uint16_t); 1463 } 1464 1465 return (nmp); 1466 } 1467 1468 /* 1469 * DL_NOTIFY_IND: DL_NOTE_PROMISC_ON_PHYS 1470 */ 1471 static void 1472 str_notify_promisc_on_phys(dld_str_t *dsp) 1473 { 1474 mblk_t *mp; 1475 dl_notify_ind_t *dlip; 1476 1477 if (!(dsp->ds_notifications & DL_NOTE_PROMISC_ON_PHYS)) 1478 return; 1479 1480 if ((mp = mexchange(dsp->ds_wq, NULL, sizeof (dl_notify_ind_t), 1481 M_PROTO, 0)) == NULL) 1482 return; 1483 1484 bzero(mp->b_rptr, sizeof (dl_notify_ind_t)); 1485 dlip = (dl_notify_ind_t *)mp->b_rptr; 1486 dlip->dl_primitive = DL_NOTIFY_IND; 1487 dlip->dl_notification = DL_NOTE_PROMISC_ON_PHYS; 1488 1489 qreply(dsp->ds_wq, mp); 1490 } 1491 1492 /* 1493 * DL_NOTIFY_IND: DL_NOTE_PROMISC_OFF_PHYS 1494 */ 1495 static void 1496 str_notify_promisc_off_phys(dld_str_t *dsp) 1497 { 1498 mblk_t *mp; 1499 dl_notify_ind_t *dlip; 1500 1501 if (!(dsp->ds_notifications & DL_NOTE_PROMISC_OFF_PHYS)) 1502 return; 1503 1504 if ((mp = mexchange(dsp->ds_wq, NULL, sizeof (dl_notify_ind_t), 1505 M_PROTO, 0)) == NULL) 1506 return; 1507 1508 bzero(mp->b_rptr, sizeof (dl_notify_ind_t)); 1509 dlip = (dl_notify_ind_t *)mp->b_rptr; 1510 dlip->dl_primitive = DL_NOTIFY_IND; 1511 dlip->dl_notification = DL_NOTE_PROMISC_OFF_PHYS; 1512 1513 qreply(dsp->ds_wq, mp); 1514 } 1515 1516 /* 1517 * DL_NOTIFY_IND: DL_NOTE_PHYS_ADDR 1518 */ 1519 static void 1520 str_notify_phys_addr(dld_str_t *dsp, const uint8_t *addr) 1521 { 1522 mblk_t *mp; 1523 dl_notify_ind_t *dlip; 1524 uint_t addr_length; 1525 uint16_t ethertype; 1526 1527 if (!(dsp->ds_notifications & DL_NOTE_PHYS_ADDR)) 1528 return; 1529 1530 addr_length = dsp->ds_mip->mi_addr_length; 1531 if ((mp = mexchange(dsp->ds_wq, NULL, 1532 sizeof (dl_notify_ind_t) + addr_length + sizeof (uint16_t), 1533 M_PROTO, 0)) == NULL) 1534 return; 1535 1536 bzero(mp->b_rptr, sizeof (dl_notify_ind_t)); 1537 dlip = (dl_notify_ind_t *)mp->b_rptr; 1538 dlip->dl_primitive = DL_NOTIFY_IND; 1539 dlip->dl_notification = DL_NOTE_PHYS_ADDR; 1540 dlip->dl_data = DL_CURR_PHYS_ADDR; 1541 dlip->dl_addr_offset = sizeof (dl_notify_ind_t); 1542 dlip->dl_addr_length = addr_length + sizeof (uint16_t); 1543 1544 bcopy(addr, &dlip[1], addr_length); 1545 1546 ethertype = (dsp->ds_sap < ETHERTYPE_802_MIN) ? 0 : dsp->ds_sap; 1547 *(uint16_t *)((uchar_t *)(dlip + 1) + addr_length) = 1548 ethertype; 1549 1550 qreply(dsp->ds_wq, mp); 1551 } 1552 1553 /* 1554 * DL_NOTIFY_IND: DL_NOTE_LINK_UP 1555 */ 1556 static void 1557 str_notify_link_up(dld_str_t *dsp) 1558 { 1559 mblk_t *mp; 1560 dl_notify_ind_t *dlip; 1561 1562 if (!(dsp->ds_notifications & DL_NOTE_LINK_UP)) 1563 return; 1564 1565 if ((mp = mexchange(dsp->ds_wq, NULL, sizeof (dl_notify_ind_t), 1566 M_PROTO, 0)) == NULL) 1567 return; 1568 1569 bzero(mp->b_rptr, sizeof (dl_notify_ind_t)); 1570 dlip = (dl_notify_ind_t *)mp->b_rptr; 1571 dlip->dl_primitive = DL_NOTIFY_IND; 1572 dlip->dl_notification = DL_NOTE_LINK_UP; 1573 1574 qreply(dsp->ds_wq, mp); 1575 } 1576 1577 /* 1578 * DL_NOTIFY_IND: DL_NOTE_LINK_DOWN 1579 */ 1580 static void 1581 str_notify_link_down(dld_str_t *dsp) 1582 { 1583 mblk_t *mp; 1584 dl_notify_ind_t *dlip; 1585 1586 if (!(dsp->ds_notifications & DL_NOTE_LINK_DOWN)) 1587 return; 1588 1589 if ((mp = mexchange(dsp->ds_wq, NULL, sizeof (dl_notify_ind_t), 1590 M_PROTO, 0)) == NULL) 1591 return; 1592 1593 bzero(mp->b_rptr, sizeof (dl_notify_ind_t)); 1594 dlip = (dl_notify_ind_t *)mp->b_rptr; 1595 dlip->dl_primitive = DL_NOTIFY_IND; 1596 dlip->dl_notification = DL_NOTE_LINK_DOWN; 1597 1598 qreply(dsp->ds_wq, mp); 1599 } 1600 1601 /* 1602 * DL_NOTIFY_IND: DL_NOTE_SPEED 1603 */ 1604 static void 1605 str_notify_speed(dld_str_t *dsp, uint32_t speed) 1606 { 1607 mblk_t *mp; 1608 dl_notify_ind_t *dlip; 1609 1610 if (!(dsp->ds_notifications & DL_NOTE_SPEED)) 1611 return; 1612 1613 if ((mp = mexchange(dsp->ds_wq, NULL, sizeof (dl_notify_ind_t), 1614 M_PROTO, 0)) == NULL) 1615 return; 1616 1617 bzero(mp->b_rptr, sizeof (dl_notify_ind_t)); 1618 dlip = (dl_notify_ind_t *)mp->b_rptr; 1619 dlip->dl_primitive = DL_NOTIFY_IND; 1620 dlip->dl_notification = DL_NOTE_SPEED; 1621 dlip->dl_data = speed; 1622 1623 qreply(dsp->ds_wq, mp); 1624 } 1625 1626 /* 1627 * DL_NOTIFY_IND: DL_NOTE_CAPAB_RENEG 1628 */ 1629 static void 1630 str_notify_capab_reneg(dld_str_t *dsp) 1631 { 1632 mblk_t *mp; 1633 dl_notify_ind_t *dlip; 1634 1635 if (!(dsp->ds_notifications & DL_NOTE_CAPAB_RENEG)) 1636 return; 1637 1638 if ((mp = mexchange(dsp->ds_wq, NULL, sizeof (dl_notify_ind_t), 1639 M_PROTO, 0)) == NULL) 1640 return; 1641 1642 bzero(mp->b_rptr, sizeof (dl_notify_ind_t)); 1643 dlip = (dl_notify_ind_t *)mp->b_rptr; 1644 dlip->dl_primitive = DL_NOTIFY_IND; 1645 dlip->dl_notification = DL_NOTE_CAPAB_RENEG; 1646 1647 qreply(dsp->ds_wq, mp); 1648 } 1649 1650 /* 1651 * DL_NOTIFY_IND: DL_NOTE_FASTPATH_FLUSH 1652 */ 1653 static void 1654 str_notify_fastpath_flush(dld_str_t *dsp) 1655 { 1656 mblk_t *mp; 1657 dl_notify_ind_t *dlip; 1658 1659 if (!(dsp->ds_notifications & DL_NOTE_FASTPATH_FLUSH)) 1660 return; 1661 1662 if ((mp = mexchange(dsp->ds_wq, NULL, sizeof (dl_notify_ind_t), 1663 M_PROTO, 0)) == NULL) 1664 return; 1665 1666 bzero(mp->b_rptr, sizeof (dl_notify_ind_t)); 1667 dlip = (dl_notify_ind_t *)mp->b_rptr; 1668 dlip->dl_primitive = DL_NOTIFY_IND; 1669 dlip->dl_notification = DL_NOTE_FASTPATH_FLUSH; 1670 1671 qreply(dsp->ds_wq, mp); 1672 } 1673 1674 /* 1675 * MAC notification callback. 1676 */ 1677 static void 1678 str_notify(void *arg, mac_notify_type_t type) 1679 { 1680 dld_str_t *dsp = (dld_str_t *)arg; 1681 queue_t *q = dsp->ds_wq; 1682 1683 switch (type) { 1684 case MAC_NOTE_TX: 1685 qenable(q); 1686 break; 1687 1688 case MAC_NOTE_DEVPROMISC: 1689 /* 1690 * Send the appropriate DL_NOTIFY_IND. 1691 */ 1692 if (mac_promisc_get(dsp->ds_mh, MAC_DEVPROMISC)) 1693 str_notify_promisc_on_phys(dsp); 1694 else 1695 str_notify_promisc_off_phys(dsp); 1696 break; 1697 1698 case MAC_NOTE_PROMISC: 1699 break; 1700 1701 case MAC_NOTE_UNICST: 1702 /* 1703 * This notification is sent whenever the MAC unicast address 1704 * changes. We need to re-cache the address. 1705 */ 1706 mac_unicst_get(dsp->ds_mh, dsp->ds_curr_addr); 1707 1708 /* 1709 * Send the appropriate DL_NOTIFY_IND. 1710 */ 1711 str_notify_phys_addr(dsp, dsp->ds_curr_addr); 1712 break; 1713 1714 case MAC_NOTE_LINK: 1715 /* 1716 * This notification is sent every time the MAC driver 1717 * updates the link state. 1718 */ 1719 switch (mac_link_get(dsp->ds_mh)) { 1720 case LINK_STATE_UP: { 1721 uint64_t speed; 1722 /* 1723 * The link is up so send the appropriate 1724 * DL_NOTIFY_IND. 1725 */ 1726 str_notify_link_up(dsp); 1727 1728 speed = mac_stat_get(dsp->ds_mh, MAC_STAT_IFSPEED); 1729 str_notify_speed(dsp, (uint32_t)(speed / 1000ull)); 1730 break; 1731 } 1732 case LINK_STATE_DOWN: 1733 /* 1734 * The link is down so send the appropriate 1735 * DL_NOTIFY_IND. 1736 */ 1737 str_notify_link_down(dsp); 1738 break; 1739 1740 default: 1741 break; 1742 } 1743 break; 1744 1745 case MAC_NOTE_RESOURCE: 1746 /* 1747 * This notification is sent whenever the MAC resources 1748 * change. We need to renegotiate the capabilities. 1749 * Send the appropriate DL_NOTIFY_IND. 1750 */ 1751 str_notify_capab_reneg(dsp); 1752 break; 1753 1754 case MAC_NOTE_FASTPATH_FLUSH: 1755 str_notify_fastpath_flush(dsp); 1756 break; 1757 1758 default: 1759 ASSERT(B_FALSE); 1760 break; 1761 } 1762 } 1763 1764 /* 1765 * Enqueue one or more messages to the transmit queue. 1766 * Caller specifies the insertion position (head/tail). 1767 */ 1768 void 1769 dld_tx_enqueue(dld_str_t *dsp, mblk_t *mp, boolean_t head_insert) 1770 { 1771 mblk_t *tail; 1772 queue_t *q = dsp->ds_wq; 1773 uint_t cnt, msgcnt; 1774 uint_t tot_cnt, tot_msgcnt; 1775 1776 ASSERT(DB_TYPE(mp) == M_DATA); 1777 /* Calculate total size and count of the packet(s) */ 1778 for (tail = mp, cnt = msgdsize(mp), msgcnt = 1; 1779 tail->b_next != NULL; tail = tail->b_next) { 1780 ASSERT(DB_TYPE(tail->b_next) == M_DATA); 1781 cnt += msgdsize(tail->b_next); 1782 msgcnt++; 1783 } 1784 1785 mutex_enter(&dsp->ds_tx_list_lock); 1786 /* 1787 * If the queue depth would exceed the allowed threshold, drop 1788 * new packet(s) and drain those already in the queue. 1789 */ 1790 tot_cnt = dsp->ds_tx_cnt + cnt; 1791 tot_msgcnt = dsp->ds_tx_msgcnt + msgcnt; 1792 1793 if (!head_insert && 1794 (tot_cnt >= dld_max_q_count || tot_msgcnt >= dld_max_q_count)) { 1795 ASSERT(dsp->ds_tx_qbusy); 1796 mutex_exit(&dsp->ds_tx_list_lock); 1797 freemsgchain(mp); 1798 goto done; 1799 } 1800 1801 /* Update the queue size parameters */ 1802 dsp->ds_tx_cnt = tot_cnt; 1803 dsp->ds_tx_msgcnt = tot_msgcnt; 1804 1805 /* 1806 * If the transmit queue is currently empty and we are 1807 * about to deposit the packet(s) there, switch mode to 1808 * "busy" and raise flow-control condition. 1809 */ 1810 if (!dsp->ds_tx_qbusy) { 1811 dsp->ds_tx_qbusy = B_TRUE; 1812 ASSERT(dsp->ds_tx_flow_mp != NULL); 1813 (void) putq(q, dsp->ds_tx_flow_mp); 1814 dsp->ds_tx_flow_mp = NULL; 1815 } 1816 1817 if (!head_insert) { 1818 /* Tail insertion */ 1819 if (dsp->ds_tx_list_head == NULL) 1820 dsp->ds_tx_list_head = mp; 1821 else 1822 dsp->ds_tx_list_tail->b_next = mp; 1823 dsp->ds_tx_list_tail = tail; 1824 } else { 1825 /* Head insertion */ 1826 tail->b_next = dsp->ds_tx_list_head; 1827 if (dsp->ds_tx_list_head == NULL) 1828 dsp->ds_tx_list_tail = tail; 1829 dsp->ds_tx_list_head = mp; 1830 } 1831 mutex_exit(&dsp->ds_tx_list_lock); 1832 done: 1833 /* Schedule service thread to drain the transmit queue */ 1834 qenable(q); 1835 } 1836 1837 void 1838 dld_tx_flush(dld_str_t *dsp) 1839 { 1840 mutex_enter(&dsp->ds_tx_list_lock); 1841 if (dsp->ds_tx_list_head != NULL) { 1842 freemsgchain(dsp->ds_tx_list_head); 1843 dsp->ds_tx_list_head = dsp->ds_tx_list_tail = NULL; 1844 dsp->ds_tx_cnt = dsp->ds_tx_msgcnt = 0; 1845 if (dsp->ds_tx_qbusy) { 1846 dsp->ds_tx_flow_mp = getq(dsp->ds_wq); 1847 ASSERT(dsp->ds_tx_flow_mp != NULL); 1848 dsp->ds_tx_qbusy = B_FALSE; 1849 } 1850 } 1851 mutex_exit(&dsp->ds_tx_list_lock); 1852 } 1853 1854 /* 1855 * Process an M_IOCTL message. 1856 */ 1857 static void 1858 dld_ioc(dld_str_t *dsp, mblk_t *mp) 1859 { 1860 uint_t cmd; 1861 1862 cmd = ((struct iocblk *)mp->b_rptr)->ioc_cmd; 1863 ASSERT(dsp->ds_type == DLD_DLPI); 1864 1865 switch (cmd) { 1866 case DLIOCRAW: 1867 ioc_raw(dsp, mp); 1868 break; 1869 case DLIOCHDRINFO: 1870 ioc_fast(dsp, mp); 1871 break; 1872 default: 1873 ioc(dsp, mp); 1874 } 1875 } 1876 1877 /* 1878 * DLIOCRAW 1879 */ 1880 static void 1881 ioc_raw(dld_str_t *dsp, mblk_t *mp) 1882 { 1883 queue_t *q = dsp->ds_wq; 1884 1885 rw_enter(&dsp->ds_lock, RW_WRITER); 1886 if (dsp->ds_polling || dsp->ds_soft_ring) { 1887 rw_exit(&dsp->ds_lock); 1888 miocnak(q, mp, 0, EPROTO); 1889 return; 1890 } 1891 1892 if (dsp->ds_mode != DLD_RAW && dsp->ds_dlstate == DL_IDLE) { 1893 /* 1894 * Set the receive callback. 1895 */ 1896 dls_rx_set(dsp->ds_dc, dld_str_rx_raw, (void *)dsp); 1897 } 1898 1899 /* 1900 * Note that raw mode is enabled. 1901 */ 1902 dsp->ds_mode = DLD_RAW; 1903 1904 rw_exit(&dsp->ds_lock); 1905 miocack(q, mp, 0, 0); 1906 } 1907 1908 /* 1909 * DLIOCHDRINFO 1910 */ 1911 static void 1912 ioc_fast(dld_str_t *dsp, mblk_t *mp) 1913 { 1914 dl_unitdata_req_t *dlp; 1915 off_t off; 1916 size_t len; 1917 const uint8_t *addr; 1918 uint16_t sap; 1919 mblk_t *nmp; 1920 mblk_t *hmp; 1921 uint_t addr_length; 1922 queue_t *q = dsp->ds_wq; 1923 int err; 1924 dls_channel_t dc; 1925 1926 if (dld_opt & DLD_OPT_NO_FASTPATH) { 1927 err = ENOTSUP; 1928 goto failed; 1929 } 1930 1931 /* 1932 * DLIOCHDRINFO should only come from IP. The one initiated from 1933 * user-land should not be allowed. 1934 */ 1935 if (((struct iocblk *)mp->b_rptr)->ioc_cr != kcred) { 1936 err = EINVAL; 1937 goto failed; 1938 } 1939 1940 nmp = mp->b_cont; 1941 if (nmp == NULL || MBLKL(nmp) < sizeof (dl_unitdata_req_t) || 1942 (dlp = (dl_unitdata_req_t *)nmp->b_rptr, 1943 dlp->dl_primitive != DL_UNITDATA_REQ)) { 1944 err = EINVAL; 1945 goto failed; 1946 } 1947 1948 off = dlp->dl_dest_addr_offset; 1949 len = dlp->dl_dest_addr_length; 1950 1951 if (!MBLKIN(nmp, off, len)) { 1952 err = EINVAL; 1953 goto failed; 1954 } 1955 1956 rw_enter(&dsp->ds_lock, RW_READER); 1957 if (dsp->ds_dlstate != DL_IDLE) { 1958 rw_exit(&dsp->ds_lock); 1959 err = ENOTSUP; 1960 goto failed; 1961 } 1962 1963 addr_length = dsp->ds_mip->mi_addr_length; 1964 if (len != addr_length + sizeof (uint16_t)) { 1965 rw_exit(&dsp->ds_lock); 1966 err = EINVAL; 1967 goto failed; 1968 } 1969 1970 addr = nmp->b_rptr + off; 1971 sap = *(uint16_t *)(nmp->b_rptr + off + addr_length); 1972 dc = dsp->ds_dc; 1973 1974 if ((hmp = dls_header(dc, addr, sap, 0, NULL)) == NULL) { 1975 rw_exit(&dsp->ds_lock); 1976 err = ENOMEM; 1977 goto failed; 1978 } 1979 1980 /* 1981 * This is a performance optimization. We originally entered 1982 * as reader and only become writer upon transitioning into 1983 * the DLD_FASTPATH mode for the first time. Otherwise we 1984 * stay as reader and return the fast-path header to IP. 1985 */ 1986 if (dsp->ds_mode != DLD_FASTPATH) { 1987 if (!rw_tryupgrade(&dsp->ds_lock)) { 1988 rw_exit(&dsp->ds_lock); 1989 rw_enter(&dsp->ds_lock, RW_WRITER); 1990 1991 /* 1992 * State may have changed before we re-acquired 1993 * the writer lock in case the upgrade failed. 1994 */ 1995 if (dsp->ds_dlstate != DL_IDLE) { 1996 rw_exit(&dsp->ds_lock); 1997 err = ENOTSUP; 1998 goto failed; 1999 } 2000 } 2001 2002 /* 2003 * Set the receive callback (unless polling is enabled). 2004 */ 2005 if (!dsp->ds_polling && !dsp->ds_soft_ring) 2006 dls_rx_set(dc, dld_str_rx_fastpath, (void *)dsp); 2007 2008 /* 2009 * Note that fast-path mode is enabled. 2010 */ 2011 dsp->ds_mode = DLD_FASTPATH; 2012 } 2013 rw_exit(&dsp->ds_lock); 2014 2015 freemsg(nmp->b_cont); 2016 nmp->b_cont = hmp; 2017 2018 miocack(q, mp, MBLKL(nmp) + MBLKL(hmp), 0); 2019 return; 2020 failed: 2021 miocnak(q, mp, 0, err); 2022 } 2023 2024 /* 2025 * Catch-all handler. 2026 */ 2027 static void 2028 ioc(dld_str_t *dsp, mblk_t *mp) 2029 { 2030 queue_t *q = dsp->ds_wq; 2031 mac_handle_t mh; 2032 2033 rw_enter(&dsp->ds_lock, RW_READER); 2034 if (dsp->ds_dlstate == DL_UNATTACHED) { 2035 rw_exit(&dsp->ds_lock); 2036 miocnak(q, mp, 0, EINVAL); 2037 return; 2038 } 2039 mh = dsp->ds_mh; 2040 ASSERT(mh != NULL); 2041 rw_exit(&dsp->ds_lock); 2042 mac_ioctl(mh, q, mp); 2043 } 2044 2045 /* 2046 * Allocate a new minor number. 2047 */ 2048 static minor_t 2049 dld_minor_hold(boolean_t sleep) 2050 { 2051 minor_t minor; 2052 2053 /* 2054 * Grab a value from the arena. 2055 */ 2056 atomic_add_32(&minor_count, 1); 2057 if ((minor = PTR_TO_MINOR(vmem_alloc(minor_arenap, 1, 2058 (sleep) ? VM_SLEEP : VM_NOSLEEP))) == 0) { 2059 atomic_add_32(&minor_count, -1); 2060 return (0); 2061 } 2062 2063 return (minor); 2064 } 2065 2066 /* 2067 * Release a previously allocated minor number. 2068 */ 2069 static void 2070 dld_minor_rele(minor_t minor) 2071 { 2072 /* 2073 * Return the value to the arena. 2074 */ 2075 vmem_free(minor_arenap, MINOR_TO_PTR(minor), 1); 2076 2077 atomic_add_32(&minor_count, -1); 2078 } 2079