1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 /* 22 * Copyright 2008 Sun Microsystems, Inc. All rights reserved. 23 * Use is subject to license terms. 24 */ 25 26 /* 27 * utility routines for the /dev fs 28 */ 29 30 #include <sys/types.h> 31 #include <sys/param.h> 32 #include <sys/t_lock.h> 33 #include <sys/systm.h> 34 #include <sys/sysmacros.h> 35 #include <sys/user.h> 36 #include <sys/time.h> 37 #include <sys/vfs.h> 38 #include <sys/vnode.h> 39 #include <sys/file.h> 40 #include <sys/fcntl.h> 41 #include <sys/flock.h> 42 #include <sys/kmem.h> 43 #include <sys/uio.h> 44 #include <sys/errno.h> 45 #include <sys/stat.h> 46 #include <sys/cred.h> 47 #include <sys/dirent.h> 48 #include <sys/pathname.h> 49 #include <sys/cmn_err.h> 50 #include <sys/debug.h> 51 #include <sys/mode.h> 52 #include <sys/policy.h> 53 #include <fs/fs_subr.h> 54 #include <sys/mount.h> 55 #include <sys/fs/snode.h> 56 #include <sys/fs/dv_node.h> 57 #include <sys/fs/sdev_impl.h> 58 #include <sys/fs/sdev_node.h> 59 #include <sys/sunndi.h> 60 #include <sys/sunmdi.h> 61 #include <sys/conf.h> 62 #include <sys/proc.h> 63 #include <sys/user.h> 64 #include <sys/modctl.h> 65 66 #ifdef DEBUG 67 int sdev_debug = 0x00000001; 68 int sdev_debug_cache_flags = 0; 69 #endif 70 71 /* 72 * globals 73 */ 74 /* prototype memory vattrs */ 75 vattr_t sdev_vattr_dir = { 76 AT_TYPE|AT_MODE|AT_UID|AT_GID, /* va_mask */ 77 VDIR, /* va_type */ 78 SDEV_DIRMODE_DEFAULT, /* va_mode */ 79 SDEV_UID_DEFAULT, /* va_uid */ 80 SDEV_GID_DEFAULT, /* va_gid */ 81 0, /* va_fsid */ 82 0, /* va_nodeid */ 83 0, /* va_nlink */ 84 0, /* va_size */ 85 0, /* va_atime */ 86 0, /* va_mtime */ 87 0, /* va_ctime */ 88 0, /* va_rdev */ 89 0, /* va_blksize */ 90 0, /* va_nblocks */ 91 0 /* va_vcode */ 92 }; 93 94 vattr_t sdev_vattr_lnk = { 95 AT_TYPE|AT_MODE, /* va_mask */ 96 VLNK, /* va_type */ 97 SDEV_LNKMODE_DEFAULT, /* va_mode */ 98 SDEV_UID_DEFAULT, /* va_uid */ 99 SDEV_GID_DEFAULT, /* va_gid */ 100 0, /* va_fsid */ 101 0, /* va_nodeid */ 102 0, /* va_nlink */ 103 0, /* va_size */ 104 0, /* va_atime */ 105 0, /* va_mtime */ 106 0, /* va_ctime */ 107 0, /* va_rdev */ 108 0, /* va_blksize */ 109 0, /* va_nblocks */ 110 0 /* va_vcode */ 111 }; 112 113 vattr_t sdev_vattr_blk = { 114 AT_TYPE|AT_MODE|AT_UID|AT_GID, /* va_mask */ 115 VBLK, /* va_type */ 116 S_IFBLK | SDEV_DEVMODE_DEFAULT, /* va_mode */ 117 SDEV_UID_DEFAULT, /* va_uid */ 118 SDEV_GID_DEFAULT, /* va_gid */ 119 0, /* va_fsid */ 120 0, /* va_nodeid */ 121 0, /* va_nlink */ 122 0, /* va_size */ 123 0, /* va_atime */ 124 0, /* va_mtime */ 125 0, /* va_ctime */ 126 0, /* va_rdev */ 127 0, /* va_blksize */ 128 0, /* va_nblocks */ 129 0 /* va_vcode */ 130 }; 131 132 vattr_t sdev_vattr_chr = { 133 AT_TYPE|AT_MODE|AT_UID|AT_GID, /* va_mask */ 134 VCHR, /* va_type */ 135 S_IFCHR | SDEV_DEVMODE_DEFAULT, /* va_mode */ 136 SDEV_UID_DEFAULT, /* va_uid */ 137 SDEV_GID_DEFAULT, /* va_gid */ 138 0, /* va_fsid */ 139 0, /* va_nodeid */ 140 0, /* va_nlink */ 141 0, /* va_size */ 142 0, /* va_atime */ 143 0, /* va_mtime */ 144 0, /* va_ctime */ 145 0, /* va_rdev */ 146 0, /* va_blksize */ 147 0, /* va_nblocks */ 148 0 /* va_vcode */ 149 }; 150 151 kmem_cache_t *sdev_node_cache; /* sdev_node cache */ 152 int devtype; /* fstype */ 153 154 struct devname_ops *devname_ns_ops; /* default name service directory ops */ 155 kmutex_t devname_nsmaps_lock; /* protect devname_nsmaps */ 156 157 /* static */ 158 static struct devname_nsmap *devname_nsmaps = NULL; 159 /* contents from /etc/dev/devname_master */ 160 static int devname_nsmaps_invalidated = 0; /* "devfsadm -m" has run */ 161 162 static struct vnodeops *sdev_get_vop(struct sdev_node *); 163 static void sdev_set_no_nocache(struct sdev_node *); 164 static int sdev_get_moduleops(struct sdev_node *); 165 static fs_operation_def_t *sdev_merge_vtab(const fs_operation_def_t []); 166 static void sdev_free_vtab(fs_operation_def_t *); 167 168 static void 169 sdev_prof_free(struct sdev_node *dv) 170 { 171 ASSERT(!SDEV_IS_GLOBAL(dv)); 172 if (dv->sdev_prof.dev_name) 173 nvlist_free(dv->sdev_prof.dev_name); 174 if (dv->sdev_prof.dev_map) 175 nvlist_free(dv->sdev_prof.dev_map); 176 if (dv->sdev_prof.dev_symlink) 177 nvlist_free(dv->sdev_prof.dev_symlink); 178 if (dv->sdev_prof.dev_glob_incdir) 179 nvlist_free(dv->sdev_prof.dev_glob_incdir); 180 if (dv->sdev_prof.dev_glob_excdir) 181 nvlist_free(dv->sdev_prof.dev_glob_excdir); 182 bzero(&dv->sdev_prof, sizeof (dv->sdev_prof)); 183 } 184 185 /* sdev_node cache constructor */ 186 /*ARGSUSED1*/ 187 static int 188 i_sdev_node_ctor(void *buf, void *cfarg, int flag) 189 { 190 struct sdev_node *dv = (struct sdev_node *)buf; 191 struct vnode *vp; 192 193 bzero(buf, sizeof (struct sdev_node)); 194 vp = dv->sdev_vnode = vn_alloc(flag); 195 if (vp == NULL) { 196 return (-1); 197 } 198 vp->v_data = dv; 199 rw_init(&dv->sdev_contents, NULL, RW_DEFAULT, NULL); 200 return (0); 201 } 202 203 /* sdev_node cache destructor */ 204 /*ARGSUSED1*/ 205 static void 206 i_sdev_node_dtor(void *buf, void *arg) 207 { 208 struct sdev_node *dv = (struct sdev_node *)buf; 209 struct vnode *vp = SDEVTOV(dv); 210 211 rw_destroy(&dv->sdev_contents); 212 vn_free(vp); 213 } 214 215 /* initialize sdev_node cache */ 216 void 217 sdev_node_cache_init() 218 { 219 int flags = 0; 220 221 #ifdef DEBUG 222 flags = sdev_debug_cache_flags; 223 if (flags) 224 sdcmn_err(("cache debug flags 0x%x\n", flags)); 225 #endif /* DEBUG */ 226 227 ASSERT(sdev_node_cache == NULL); 228 sdev_node_cache = kmem_cache_create("sdev_node_cache", 229 sizeof (struct sdev_node), 0, i_sdev_node_ctor, i_sdev_node_dtor, 230 NULL, NULL, NULL, flags); 231 } 232 233 /* destroy sdev_node cache */ 234 void 235 sdev_node_cache_fini() 236 { 237 ASSERT(sdev_node_cache != NULL); 238 kmem_cache_destroy(sdev_node_cache); 239 sdev_node_cache = NULL; 240 } 241 242 /* 243 * Compare two nodes lexographically to balance avl tree 244 */ 245 static int 246 sdev_compare_nodes(const struct sdev_node *dv1, const struct sdev_node *dv2) 247 { 248 int rv; 249 if ((rv = strcmp(dv1->sdev_name, dv2->sdev_name)) == 0) 250 return (0); 251 return ((rv < 0) ? -1 : 1); 252 } 253 254 void 255 sdev_set_nodestate(struct sdev_node *dv, sdev_node_state_t state) 256 { 257 ASSERT(dv); 258 ASSERT(RW_WRITE_HELD(&dv->sdev_contents)); 259 dv->sdev_state = state; 260 } 261 262 static void 263 sdev_attrinit(struct sdev_node *dv, vattr_t *vap) 264 { 265 timestruc_t now; 266 267 ASSERT(vap); 268 269 dv->sdev_attr = kmem_zalloc(sizeof (struct vattr), KM_SLEEP); 270 *dv->sdev_attr = *vap; 271 272 dv->sdev_attr->va_mode = MAKEIMODE(vap->va_type, vap->va_mode); 273 274 gethrestime(&now); 275 dv->sdev_attr->va_atime = now; 276 dv->sdev_attr->va_mtime = now; 277 dv->sdev_attr->va_ctime = now; 278 } 279 280 /* alloc and initialize a sdev_node */ 281 int 282 sdev_nodeinit(struct sdev_node *ddv, char *nm, struct sdev_node **newdv, 283 vattr_t *vap) 284 { 285 struct sdev_node *dv = NULL; 286 struct vnode *vp; 287 size_t nmlen, len; 288 devname_handle_t *dhl; 289 290 nmlen = strlen(nm) + 1; 291 if (nmlen > MAXNAMELEN) { 292 sdcmn_err9(("sdev_nodeinit: node name %s" 293 " too long\n", nm)); 294 *newdv = NULL; 295 return (ENAMETOOLONG); 296 } 297 298 dv = kmem_cache_alloc(sdev_node_cache, KM_SLEEP); 299 300 dv->sdev_name = kmem_alloc(nmlen, KM_SLEEP); 301 bcopy(nm, dv->sdev_name, nmlen); 302 dv->sdev_namelen = nmlen - 1; /* '\0' not included */ 303 len = strlen(ddv->sdev_path) + strlen(nm) + 2; 304 dv->sdev_path = kmem_alloc(len, KM_SLEEP); 305 (void) snprintf(dv->sdev_path, len, "%s/%s", ddv->sdev_path, nm); 306 /* overwritten for VLNK nodes */ 307 dv->sdev_symlink = NULL; 308 309 vp = SDEVTOV(dv); 310 vn_reinit(vp); 311 vp->v_vfsp = SDEVTOV(ddv)->v_vfsp; 312 if (vap) 313 vp->v_type = vap->va_type; 314 315 /* 316 * initialized to the parent's vnodeops. 317 * maybe overwriten for a VDIR 318 */ 319 vn_setops(vp, vn_getops(SDEVTOV(ddv))); 320 vn_exists(vp); 321 322 dv->sdev_dotdot = NULL; 323 dv->sdev_attrvp = NULL; 324 if (vap) { 325 sdev_attrinit(dv, vap); 326 } else { 327 dv->sdev_attr = NULL; 328 } 329 330 dv->sdev_ino = sdev_mkino(dv); 331 dv->sdev_nlink = 0; /* updated on insert */ 332 dv->sdev_flags = ddv->sdev_flags; /* inherit from the parent first */ 333 dv->sdev_flags |= SDEV_BUILD; 334 mutex_init(&dv->sdev_lookup_lock, NULL, MUTEX_DEFAULT, NULL); 335 cv_init(&dv->sdev_lookup_cv, NULL, CV_DEFAULT, NULL); 336 if (SDEV_IS_GLOBAL(ddv)) { 337 dv->sdev_flags |= SDEV_GLOBAL; 338 dv->sdev_mapinfo = NULL; 339 dhl = &(dv->sdev_handle); 340 dhl->dh_data = dv; 341 dhl->dh_spec = DEVNAME_NS_NONE; 342 dhl->dh_args = NULL; 343 sdev_set_no_nocache(dv); 344 dv->sdev_gdir_gen = 0; 345 } else { 346 dv->sdev_flags &= ~SDEV_GLOBAL; 347 dv->sdev_origin = NULL; /* set later */ 348 bzero(&dv->sdev_prof, sizeof (dv->sdev_prof)); 349 dv->sdev_ldir_gen = 0; 350 dv->sdev_devtree_gen = 0; 351 } 352 353 rw_enter(&dv->sdev_contents, RW_WRITER); 354 sdev_set_nodestate(dv, SDEV_INIT); 355 rw_exit(&dv->sdev_contents); 356 *newdv = dv; 357 358 return (0); 359 } 360 361 /* 362 * transition a sdev_node into SDEV_READY state 363 */ 364 int 365 sdev_nodeready(struct sdev_node *dv, struct vattr *vap, struct vnode *avp, 366 void *args, struct cred *cred) 367 { 368 int error = 0; 369 struct vnode *vp = SDEVTOV(dv); 370 vtype_t type; 371 372 ASSERT(dv && (dv->sdev_state != SDEV_READY) && vap); 373 374 type = vap->va_type; 375 vp->v_type = type; 376 vp->v_rdev = vap->va_rdev; 377 rw_enter(&dv->sdev_contents, RW_WRITER); 378 if (type == VDIR) { 379 dv->sdev_nlink = 2; 380 dv->sdev_flags &= ~SDEV_PERSIST; 381 dv->sdev_flags &= ~SDEV_DYNAMIC; 382 vn_setops(vp, sdev_get_vop(dv)); /* from internal vtab */ 383 error = sdev_get_moduleops(dv); /* from plug-in module */ 384 ASSERT(dv->sdev_dotdot); 385 ASSERT(SDEVTOV(dv->sdev_dotdot)->v_type == VDIR); 386 vp->v_rdev = SDEVTOV(dv->sdev_dotdot)->v_rdev; 387 avl_create(&dv->sdev_entries, 388 (int (*)(const void *, const void *))sdev_compare_nodes, 389 sizeof (struct sdev_node), 390 offsetof(struct sdev_node, sdev_avllink)); 391 } else if (type == VLNK) { 392 ASSERT(args); 393 dv->sdev_nlink = 1; 394 dv->sdev_symlink = i_ddi_strdup((char *)args, KM_SLEEP); 395 } else { 396 dv->sdev_nlink = 1; 397 } 398 399 if (!(SDEV_IS_GLOBAL(dv))) { 400 dv->sdev_origin = (struct sdev_node *)args; 401 dv->sdev_flags &= ~SDEV_PERSIST; 402 } 403 404 /* 405 * shadow node is created here OR 406 * if failed (indicated by dv->sdev_attrvp == NULL), 407 * created later in sdev_setattr 408 */ 409 if (avp) { 410 dv->sdev_attrvp = avp; 411 } else { 412 if (dv->sdev_attr == NULL) 413 sdev_attrinit(dv, vap); 414 else 415 *dv->sdev_attr = *vap; 416 417 if ((SDEV_IS_PERSIST(dv) && (dv->sdev_attrvp == NULL)) || 418 ((SDEVTOV(dv)->v_type == VDIR) && 419 (dv->sdev_attrvp == NULL))) { 420 error = sdev_shadow_node(dv, cred); 421 } 422 } 423 424 if (error == 0) { 425 /* transition to READY state */ 426 sdev_set_nodestate(dv, SDEV_READY); 427 sdev_nc_node_exists(dv); 428 } else { 429 sdev_set_nodestate(dv, SDEV_ZOMBIE); 430 } 431 rw_exit(&dv->sdev_contents); 432 return (error); 433 } 434 435 /* 436 * setting ZOMBIE state 437 */ 438 static int 439 sdev_nodezombied(struct sdev_node *dv) 440 { 441 rw_enter(&dv->sdev_contents, RW_WRITER); 442 sdev_set_nodestate(dv, SDEV_ZOMBIE); 443 rw_exit(&dv->sdev_contents); 444 return (0); 445 } 446 447 /* 448 * Build the VROOT sdev_node. 449 */ 450 /*ARGSUSED*/ 451 struct sdev_node * 452 sdev_mkroot(struct vfs *vfsp, dev_t devdev, struct vnode *mvp, 453 struct vnode *avp, struct cred *cred) 454 { 455 struct sdev_node *dv; 456 struct vnode *vp; 457 char devdir[] = "/dev"; 458 459 ASSERT(sdev_node_cache != NULL); 460 ASSERT(avp); 461 dv = kmem_cache_alloc(sdev_node_cache, KM_SLEEP); 462 vp = SDEVTOV(dv); 463 vn_reinit(vp); 464 vp->v_flag |= VROOT; 465 vp->v_vfsp = vfsp; 466 vp->v_type = VDIR; 467 vp->v_rdev = devdev; 468 vn_setops(vp, sdev_vnodeops); /* apply the default vnodeops at /dev */ 469 vn_exists(vp); 470 471 if (vfsp->vfs_mntpt) 472 dv->sdev_name = i_ddi_strdup( 473 (char *)refstr_value(vfsp->vfs_mntpt), KM_SLEEP); 474 else 475 /* vfs_mountdev1 set mount point later */ 476 dv->sdev_name = i_ddi_strdup("/dev", KM_SLEEP); 477 dv->sdev_namelen = strlen(dv->sdev_name); /* '\0' not included */ 478 dv->sdev_path = i_ddi_strdup(devdir, KM_SLEEP); 479 dv->sdev_ino = SDEV_ROOTINO; 480 dv->sdev_nlink = 2; /* name + . (no sdev_insert) */ 481 dv->sdev_dotdot = dv; /* .. == self */ 482 dv->sdev_attrvp = avp; 483 dv->sdev_attr = NULL; 484 mutex_init(&dv->sdev_lookup_lock, NULL, MUTEX_DEFAULT, NULL); 485 cv_init(&dv->sdev_lookup_cv, NULL, CV_DEFAULT, NULL); 486 if (strcmp(dv->sdev_name, "/dev") == 0) { 487 mutex_init(&devname_nsmaps_lock, NULL, MUTEX_DEFAULT, NULL); 488 dv->sdev_mapinfo = NULL; 489 dv->sdev_flags = SDEV_BUILD|SDEV_GLOBAL|SDEV_PERSIST; 490 bzero(&dv->sdev_handle, sizeof (dv->sdev_handle)); 491 dv->sdev_gdir_gen = 0; 492 } else { 493 dv->sdev_flags = SDEV_BUILD; 494 dv->sdev_flags &= ~SDEV_PERSIST; 495 bzero(&dv->sdev_prof, sizeof (dv->sdev_prof)); 496 dv->sdev_ldir_gen = 0; 497 dv->sdev_devtree_gen = 0; 498 } 499 500 avl_create(&dv->sdev_entries, 501 (int (*)(const void *, const void *))sdev_compare_nodes, 502 sizeof (struct sdev_node), 503 offsetof(struct sdev_node, sdev_avllink)); 504 505 rw_enter(&dv->sdev_contents, RW_WRITER); 506 sdev_set_nodestate(dv, SDEV_READY); 507 rw_exit(&dv->sdev_contents); 508 sdev_nc_node_exists(dv); 509 return (dv); 510 } 511 512 /* 513 * 1. load the module 514 * 2. modload invokes sdev_module_register, which in turn sets 515 * the dv->sdev_mapinfo->dir_ops 516 * 517 * note: locking order: 518 * dv->sdev_contents -> map->dir_lock 519 */ 520 static int 521 sdev_get_moduleops(struct sdev_node *dv) 522 { 523 int error = 0; 524 struct devname_nsmap *map = NULL; 525 char *module; 526 char *path; 527 int load = 1; 528 529 ASSERT(SDEVTOV(dv)->v_type == VDIR); 530 531 if (devname_nsmaps == NULL) 532 return (0); 533 534 if (!sdev_nsmaps_loaded() && !sdev_nsmaps_reloaded()) 535 return (0); 536 537 538 path = dv->sdev_path; 539 if ((map = sdev_get_nsmap_by_dir(path, 0))) { 540 rw_enter(&map->dir_lock, RW_READER); 541 if (map->dir_invalid) { 542 if (map->dir_module && map->dir_newmodule && 543 (strcmp(map->dir_module, 544 map->dir_newmodule) == 0)) { 545 load = 0; 546 } 547 sdev_replace_nsmap(map, map->dir_newmodule, 548 map->dir_newmap); 549 } 550 551 module = map->dir_module; 552 if (module && load) { 553 sdcmn_err6(("sdev_get_moduleops: " 554 "load module %s", module)); 555 rw_exit(&map->dir_lock); 556 error = modload("devname", module); 557 sdcmn_err6(("sdev_get_moduleops: error %d\n", error)); 558 if (error < 0) { 559 return (-1); 560 } 561 } else if (module == NULL) { 562 /* 563 * loading the module ops for name services 564 */ 565 if (devname_ns_ops == NULL) { 566 sdcmn_err6(( 567 "sdev_get_moduleops: modload default\n")); 568 error = modload("devname", DEVNAME_NSCONFIG); 569 sdcmn_err6(( 570 "sdev_get_moduleops: error %d\n", error)); 571 if (error < 0) { 572 return (-1); 573 } 574 } 575 576 if (!rw_tryupgrade(&map->dir_lock)) { 577 rw_exit(&map->dir_lock); 578 rw_enter(&map->dir_lock, RW_WRITER); 579 } 580 ASSERT(devname_ns_ops); 581 map->dir_ops = devname_ns_ops; 582 rw_exit(&map->dir_lock); 583 } 584 } 585 586 dv->sdev_mapinfo = map; 587 return (0); 588 } 589 590 /* directory dependent vop table */ 591 struct sdev_vop_table { 592 char *vt_name; /* subdirectory name */ 593 const fs_operation_def_t *vt_service; /* vnodeops table */ 594 struct vnodeops *vt_vops; /* constructed vop */ 595 struct vnodeops **vt_global_vops; /* global container for vop */ 596 int (*vt_vtor)(struct sdev_node *); /* validate sdev_node */ 597 int vt_flags; 598 }; 599 600 /* 601 * A nice improvement would be to provide a plug-in mechanism 602 * for this table instead of a const table. 603 */ 604 static struct sdev_vop_table vtab[] = 605 { 606 { "pts", devpts_vnodeops_tbl, NULL, &devpts_vnodeops, devpts_validate, 607 SDEV_DYNAMIC | SDEV_VTOR }, 608 609 { "vt", devvt_vnodeops_tbl, NULL, &devvt_vnodeops, devvt_validate, 610 SDEV_DYNAMIC | SDEV_VTOR }, 611 612 { "zcons", NULL, NULL, NULL, NULL, SDEV_NO_NCACHE }, 613 614 { "net", devnet_vnodeops_tbl, NULL, &devnet_vnodeops, devnet_validate, 615 SDEV_DYNAMIC | SDEV_VTOR }, 616 617 { NULL, NULL, NULL, NULL, NULL, 0} 618 }; 619 620 621 /* 622 * sets a directory's vnodeops if the directory is in the vtab; 623 */ 624 static struct vnodeops * 625 sdev_get_vop(struct sdev_node *dv) 626 { 627 int i; 628 char *path; 629 630 path = dv->sdev_path; 631 ASSERT(path); 632 633 /* gets the relative path to /dev/ */ 634 path += 5; 635 636 /* gets the vtab entry if matches */ 637 for (i = 0; vtab[i].vt_name; i++) { 638 if (strcmp(vtab[i].vt_name, path) != 0) 639 continue; 640 dv->sdev_flags |= vtab[i].vt_flags; 641 642 if (vtab[i].vt_vops) { 643 if (vtab[i].vt_global_vops) 644 *(vtab[i].vt_global_vops) = vtab[i].vt_vops; 645 return (vtab[i].vt_vops); 646 } 647 648 if (vtab[i].vt_service) { 649 fs_operation_def_t *templ; 650 templ = sdev_merge_vtab(vtab[i].vt_service); 651 if (vn_make_ops(vtab[i].vt_name, 652 (const fs_operation_def_t *)templ, 653 &vtab[i].vt_vops) != 0) { 654 cmn_err(CE_PANIC, "%s: malformed vnode ops\n", 655 vtab[i].vt_name); 656 /*NOTREACHED*/ 657 } 658 if (vtab[i].vt_global_vops) { 659 *(vtab[i].vt_global_vops) = vtab[i].vt_vops; 660 } 661 sdev_free_vtab(templ); 662 return (vtab[i].vt_vops); 663 } 664 return (sdev_vnodeops); 665 } 666 667 /* child inherits the persistence of the parent */ 668 if (SDEV_IS_PERSIST(dv->sdev_dotdot)) 669 dv->sdev_flags |= SDEV_PERSIST; 670 671 return (sdev_vnodeops); 672 } 673 674 static void 675 sdev_set_no_nocache(struct sdev_node *dv) 676 { 677 int i; 678 char *path; 679 680 ASSERT(dv->sdev_path); 681 path = dv->sdev_path + strlen("/dev/"); 682 683 for (i = 0; vtab[i].vt_name; i++) { 684 if (strcmp(vtab[i].vt_name, path) == 0) { 685 if (vtab[i].vt_flags & SDEV_NO_NCACHE) 686 dv->sdev_flags |= SDEV_NO_NCACHE; 687 break; 688 } 689 } 690 } 691 692 void * 693 sdev_get_vtor(struct sdev_node *dv) 694 { 695 int i; 696 697 for (i = 0; vtab[i].vt_name; i++) { 698 if (strcmp(vtab[i].vt_name, dv->sdev_name) != 0) 699 continue; 700 return ((void *)vtab[i].vt_vtor); 701 } 702 return (NULL); 703 } 704 705 /* 706 * Build the base root inode 707 */ 708 ino_t 709 sdev_mkino(struct sdev_node *dv) 710 { 711 ino_t ino; 712 713 /* 714 * for now, follow the lead of tmpfs here 715 * need to someday understand the requirements here 716 */ 717 ino = (ino_t)(uint32_t)((uintptr_t)dv >> 3); 718 ino += SDEV_ROOTINO + 1; 719 720 return (ino); 721 } 722 723 static int 724 sdev_getlink(struct vnode *linkvp, char **link) 725 { 726 int err; 727 char *buf; 728 struct uio uio = {0}; 729 struct iovec iov = {0}; 730 731 if (linkvp == NULL) 732 return (ENOENT); 733 ASSERT(linkvp->v_type == VLNK); 734 735 buf = kmem_zalloc(MAXPATHLEN, KM_SLEEP); 736 iov.iov_base = buf; 737 iov.iov_len = MAXPATHLEN; 738 uio.uio_iov = &iov; 739 uio.uio_iovcnt = 1; 740 uio.uio_resid = MAXPATHLEN; 741 uio.uio_segflg = UIO_SYSSPACE; 742 uio.uio_llimit = MAXOFFSET_T; 743 744 err = VOP_READLINK(linkvp, &uio, kcred, NULL); 745 if (err) { 746 cmn_err(CE_WARN, "readlink %s failed in dev\n", buf); 747 kmem_free(buf, MAXPATHLEN); 748 return (ENOENT); 749 } 750 751 /* mission complete */ 752 *link = i_ddi_strdup(buf, KM_SLEEP); 753 kmem_free(buf, MAXPATHLEN); 754 return (0); 755 } 756 757 /* 758 * A convenient wrapper to get the devfs node vnode for a device 759 * minor functionality: readlink() of a /dev symlink 760 * Place the link into dv->sdev_symlink 761 */ 762 static int 763 sdev_follow_link(struct sdev_node *dv) 764 { 765 int err; 766 struct vnode *linkvp; 767 char *link = NULL; 768 769 linkvp = SDEVTOV(dv); 770 if (linkvp == NULL) 771 return (ENOENT); 772 ASSERT(linkvp->v_type == VLNK); 773 err = sdev_getlink(linkvp, &link); 774 if (err) { 775 (void) sdev_nodezombied(dv); 776 dv->sdev_symlink = NULL; 777 return (ENOENT); 778 } 779 780 ASSERT(link != NULL); 781 dv->sdev_symlink = link; 782 return (0); 783 } 784 785 static int 786 sdev_node_check(struct sdev_node *dv, struct vattr *nvap, void *nargs) 787 { 788 vtype_t otype = SDEVTOV(dv)->v_type; 789 790 /* 791 * existing sdev_node has a different type. 792 */ 793 if (otype != nvap->va_type) { 794 sdcmn_err9(("sdev_node_check: existing node " 795 " %s type %d does not match new node type %d\n", 796 dv->sdev_name, otype, nvap->va_type)); 797 return (EEXIST); 798 } 799 800 /* 801 * For a symlink, the target should be the same. 802 */ 803 if (otype == VLNK) { 804 ASSERT(nargs != NULL); 805 ASSERT(dv->sdev_symlink != NULL); 806 if (strcmp(dv->sdev_symlink, (char *)nargs) != 0) { 807 sdcmn_err9(("sdev_node_check: existing node " 808 " %s has different symlink %s as new node " 809 " %s\n", dv->sdev_name, dv->sdev_symlink, 810 (char *)nargs)); 811 return (EEXIST); 812 } 813 } 814 815 return (0); 816 } 817 818 /* 819 * sdev_mknode - a wrapper for sdev_nodeinit(), sdev_nodeready() 820 * 821 * arguments: 822 * - ddv (parent) 823 * - nm (child name) 824 * - newdv (sdev_node for nm is returned here) 825 * - vap (vattr for the node to be created, va_type should be set. 826 * - avp (attribute vnode) 827 * the defaults should be used if unknown) 828 * - cred 829 * - args 830 * . tnm (for VLNK) 831 * . global sdev_node (for !SDEV_GLOBAL) 832 * - state: SDEV_INIT, SDEV_READY 833 * 834 * only ddv, nm, newddv, vap, cred are required for sdev_mknode(SDEV_INIT) 835 * 836 * NOTE: directory contents writers lock needs to be held before 837 * calling this routine. 838 */ 839 int 840 sdev_mknode(struct sdev_node *ddv, char *nm, struct sdev_node **newdv, 841 struct vattr *vap, struct vnode *avp, void *args, struct cred *cred, 842 sdev_node_state_t state) 843 { 844 int error = 0; 845 sdev_node_state_t node_state; 846 struct sdev_node *dv = NULL; 847 848 ASSERT(state != SDEV_ZOMBIE); 849 ASSERT(RW_WRITE_HELD(&ddv->sdev_contents)); 850 851 if (*newdv) { 852 dv = *newdv; 853 } else { 854 /* allocate and initialize a sdev_node */ 855 if (ddv->sdev_state == SDEV_ZOMBIE) { 856 sdcmn_err9(("sdev_mknode: parent %s ZOMBIEd\n", 857 ddv->sdev_path)); 858 return (ENOENT); 859 } 860 861 error = sdev_nodeinit(ddv, nm, &dv, vap); 862 if (error != 0) { 863 sdcmn_err9(("sdev_mknode: error %d," 864 " name %s can not be initialized\n", 865 error, nm)); 866 return (error); 867 } 868 ASSERT(dv); 869 870 /* insert into the directory cache */ 871 error = sdev_cache_update(ddv, &dv, nm, SDEV_CACHE_ADD); 872 if (error) { 873 sdcmn_err9(("sdev_mknode: node %s can not" 874 " be added into directory cache\n", nm)); 875 return (ENOENT); 876 } 877 } 878 879 ASSERT(dv); 880 node_state = dv->sdev_state; 881 ASSERT(node_state != SDEV_ZOMBIE); 882 883 if (state == SDEV_READY) { 884 switch (node_state) { 885 case SDEV_INIT: 886 error = sdev_nodeready(dv, vap, avp, args, cred); 887 if (error) { 888 sdcmn_err9(("sdev_mknode: node %s can NOT" 889 " be transitioned into READY state, " 890 "error %d\n", nm, error)); 891 } 892 break; 893 case SDEV_READY: 894 /* 895 * Do some sanity checking to make sure 896 * the existing sdev_node is what has been 897 * asked for. 898 */ 899 error = sdev_node_check(dv, vap, args); 900 break; 901 default: 902 break; 903 } 904 } 905 906 if (!error) { 907 *newdv = dv; 908 ASSERT((*newdv)->sdev_state != SDEV_ZOMBIE); 909 } else { 910 SDEV_SIMPLE_RELE(dv); 911 *newdv = NULL; 912 } 913 914 return (error); 915 } 916 917 /* 918 * convenient wrapper to change vp's ATIME, CTIME and MTIME 919 */ 920 void 921 sdev_update_timestamps(struct vnode *vp, cred_t *cred, uint_t mask) 922 { 923 struct vattr attr; 924 timestruc_t now; 925 int err; 926 927 ASSERT(vp); 928 gethrestime(&now); 929 if (mask & AT_CTIME) 930 attr.va_ctime = now; 931 if (mask & AT_MTIME) 932 attr.va_mtime = now; 933 if (mask & AT_ATIME) 934 attr.va_atime = now; 935 936 attr.va_mask = (mask & AT_TIMES); 937 err = VOP_SETATTR(vp, &attr, 0, cred, NULL); 938 if (err && (err != EROFS)) { 939 sdcmn_err(("update timestamps error %d\n", err)); 940 } 941 } 942 943 /* 944 * the backing store vnode is released here 945 */ 946 /*ARGSUSED1*/ 947 void 948 sdev_nodedestroy(struct sdev_node *dv, uint_t flags) 949 { 950 /* no references */ 951 ASSERT(dv->sdev_nlink == 0); 952 953 if (dv->sdev_attrvp != NULLVP) { 954 VN_RELE(dv->sdev_attrvp); 955 /* 956 * reset the attrvp so that no more 957 * references can be made on this already 958 * vn_rele() vnode 959 */ 960 dv->sdev_attrvp = NULLVP; 961 } 962 963 if (dv->sdev_attr != NULL) { 964 kmem_free(dv->sdev_attr, sizeof (struct vattr)); 965 dv->sdev_attr = NULL; 966 } 967 968 if (dv->sdev_name != NULL) { 969 kmem_free(dv->sdev_name, dv->sdev_namelen + 1); 970 dv->sdev_name = NULL; 971 } 972 973 if (dv->sdev_symlink != NULL) { 974 kmem_free(dv->sdev_symlink, strlen(dv->sdev_symlink) + 1); 975 dv->sdev_symlink = NULL; 976 } 977 978 if (dv->sdev_path) { 979 kmem_free(dv->sdev_path, strlen(dv->sdev_path) + 1); 980 dv->sdev_path = NULL; 981 } 982 983 if (!SDEV_IS_GLOBAL(dv)) 984 sdev_prof_free(dv); 985 986 if (SDEVTOV(dv)->v_type == VDIR) { 987 ASSERT(SDEV_FIRST_ENTRY(dv) == NULL); 988 avl_destroy(&dv->sdev_entries); 989 } 990 991 mutex_destroy(&dv->sdev_lookup_lock); 992 cv_destroy(&dv->sdev_lookup_cv); 993 994 /* return node to initial state as per constructor */ 995 (void) memset((void *)&dv->sdev_instance_data, 0, 996 sizeof (dv->sdev_instance_data)); 997 vn_invalid(SDEVTOV(dv)); 998 kmem_cache_free(sdev_node_cache, dv); 999 } 1000 1001 /* 1002 * DIRECTORY CACHE lookup 1003 */ 1004 struct sdev_node * 1005 sdev_findbyname(struct sdev_node *ddv, char *nm) 1006 { 1007 struct sdev_node *dv; 1008 struct sdev_node dvtmp; 1009 avl_index_t where; 1010 1011 ASSERT(RW_LOCK_HELD(&ddv->sdev_contents)); 1012 1013 dvtmp.sdev_name = nm; 1014 dv = avl_find(&ddv->sdev_entries, &dvtmp, &where); 1015 if (dv) { 1016 ASSERT(dv->sdev_dotdot == ddv); 1017 ASSERT(strcmp(dv->sdev_name, nm) == 0); 1018 SDEV_HOLD(dv); 1019 return (dv); 1020 } 1021 return (NULL); 1022 } 1023 1024 /* 1025 * Inserts a new sdev_node in a parent directory 1026 */ 1027 void 1028 sdev_direnter(struct sdev_node *ddv, struct sdev_node *dv) 1029 { 1030 avl_index_t where; 1031 1032 ASSERT(RW_WRITE_HELD(&ddv->sdev_contents)); 1033 ASSERT(SDEVTOV(ddv)->v_type == VDIR); 1034 ASSERT(ddv->sdev_nlink >= 2); 1035 ASSERT(dv->sdev_nlink == 0); 1036 1037 dv->sdev_dotdot = ddv; 1038 VERIFY(avl_find(&ddv->sdev_entries, dv, &where) == NULL); 1039 avl_insert(&ddv->sdev_entries, dv, where); 1040 ddv->sdev_nlink++; 1041 } 1042 1043 /* 1044 * The following check is needed because while sdev_nodes are linked 1045 * in SDEV_INIT state, they have their link counts incremented only 1046 * in SDEV_READY state. 1047 */ 1048 static void 1049 decr_link(struct sdev_node *dv) 1050 { 1051 if (dv->sdev_state != SDEV_INIT) 1052 dv->sdev_nlink--; 1053 else 1054 ASSERT(dv->sdev_nlink == 0); 1055 } 1056 1057 /* 1058 * Delete an existing dv from directory cache 1059 * 1060 * In the case of a node is still held by non-zero reference count, 1061 * the node is put into ZOMBIE state. Once the reference count 1062 * reaches "0", the node is unlinked and destroyed, 1063 * in sdev_inactive(). 1064 */ 1065 static int 1066 sdev_dirdelete(struct sdev_node *ddv, struct sdev_node *dv) 1067 { 1068 struct vnode *vp; 1069 1070 ASSERT(RW_WRITE_HELD(&ddv->sdev_contents)); 1071 1072 vp = SDEVTOV(dv); 1073 mutex_enter(&vp->v_lock); 1074 1075 /* dv is held still */ 1076 if (vp->v_count > 1) { 1077 rw_enter(&dv->sdev_contents, RW_WRITER); 1078 if (dv->sdev_state == SDEV_READY) { 1079 sdcmn_err9(( 1080 "sdev_delete: node %s busy with count %d\n", 1081 dv->sdev_name, vp->v_count)); 1082 dv->sdev_state = SDEV_ZOMBIE; 1083 } 1084 rw_exit(&dv->sdev_contents); 1085 --vp->v_count; 1086 mutex_exit(&vp->v_lock); 1087 return (EBUSY); 1088 } 1089 ASSERT(vp->v_count == 1); 1090 1091 /* unlink from the memory cache */ 1092 ddv->sdev_nlink--; /* .. to above */ 1093 if (vp->v_type == VDIR) { 1094 decr_link(dv); /* . to self */ 1095 } 1096 1097 avl_remove(&ddv->sdev_entries, dv); 1098 decr_link(dv); /* name, back to zero */ 1099 vp->v_count--; 1100 mutex_exit(&vp->v_lock); 1101 1102 /* destroy the node */ 1103 sdev_nodedestroy(dv, 0); 1104 return (0); 1105 } 1106 1107 /* 1108 * check if the source is in the path of the target 1109 * 1110 * source and target are different 1111 */ 1112 /*ARGSUSED2*/ 1113 static int 1114 sdev_checkpath(struct sdev_node *sdv, struct sdev_node *tdv, struct cred *cred) 1115 { 1116 int error = 0; 1117 struct sdev_node *dotdot, *dir; 1118 1119 dotdot = tdv->sdev_dotdot; 1120 ASSERT(dotdot); 1121 1122 /* fs root */ 1123 if (dotdot == tdv) { 1124 return (0); 1125 } 1126 1127 for (;;) { 1128 /* 1129 * avoid error cases like 1130 * mv a a/b 1131 * mv a a/b/c 1132 * etc. 1133 */ 1134 if (dotdot == sdv) { 1135 error = EINVAL; 1136 break; 1137 } 1138 1139 dir = dotdot; 1140 dotdot = dir->sdev_dotdot; 1141 1142 /* done checking because root is reached */ 1143 if (dir == dotdot) { 1144 break; 1145 } 1146 } 1147 return (error); 1148 } 1149 1150 int 1151 sdev_rnmnode(struct sdev_node *oddv, struct sdev_node *odv, 1152 struct sdev_node *nddv, struct sdev_node **ndvp, char *nnm, 1153 struct cred *cred) 1154 { 1155 int error = 0; 1156 struct vnode *ovp = SDEVTOV(odv); 1157 struct vnode *nvp; 1158 struct vattr vattr; 1159 int doingdir = (ovp->v_type == VDIR); 1160 char *link = NULL; 1161 int samedir = (oddv == nddv) ? 1 : 0; 1162 int bkstore = 0; 1163 struct sdev_node *idv = NULL; 1164 struct sdev_node *ndv = NULL; 1165 timestruc_t now; 1166 1167 vattr.va_mask = AT_MODE|AT_UID|AT_GID; 1168 error = VOP_GETATTR(ovp, &vattr, 0, cred, NULL); 1169 if (error) 1170 return (error); 1171 1172 if (!samedir) 1173 rw_enter(&oddv->sdev_contents, RW_WRITER); 1174 rw_enter(&nddv->sdev_contents, RW_WRITER); 1175 1176 /* 1177 * the source may have been deleted by another thread before 1178 * we gets here. 1179 */ 1180 if (odv->sdev_state != SDEV_READY) { 1181 error = ENOENT; 1182 goto err_out; 1183 } 1184 1185 if (doingdir && (odv == nddv)) { 1186 error = EINVAL; 1187 goto err_out; 1188 } 1189 1190 /* 1191 * If renaming a directory, and the parents are different (".." must be 1192 * changed) then the source dir must not be in the dir hierarchy above 1193 * the target since it would orphan everything below the source dir. 1194 */ 1195 if (doingdir && (oddv != nddv)) { 1196 error = sdev_checkpath(odv, nddv, cred); 1197 if (error) 1198 goto err_out; 1199 } 1200 1201 /* destination existing */ 1202 if (*ndvp) { 1203 nvp = SDEVTOV(*ndvp); 1204 ASSERT(nvp); 1205 1206 /* handling renaming to itself */ 1207 if (odv == *ndvp) { 1208 error = 0; 1209 goto err_out; 1210 } 1211 1212 if (nvp->v_type == VDIR) { 1213 if (!doingdir) { 1214 error = EISDIR; 1215 goto err_out; 1216 } 1217 1218 if (vn_vfswlock(nvp)) { 1219 error = EBUSY; 1220 goto err_out; 1221 } 1222 1223 if (vn_mountedvfs(nvp) != NULL) { 1224 vn_vfsunlock(nvp); 1225 error = EBUSY; 1226 goto err_out; 1227 } 1228 1229 /* in case dir1 exists in dir2 and "mv dir1 dir2" */ 1230 if ((*ndvp)->sdev_nlink > 2) { 1231 vn_vfsunlock(nvp); 1232 error = EEXIST; 1233 goto err_out; 1234 } 1235 vn_vfsunlock(nvp); 1236 1237 (void) sdev_dirdelete(nddv, *ndvp); 1238 *ndvp = NULL; 1239 ASSERT(nddv->sdev_attrvp); 1240 error = VOP_RMDIR(nddv->sdev_attrvp, nnm, 1241 nddv->sdev_attrvp, cred, NULL, 0); 1242 if (error) 1243 goto err_out; 1244 } else { 1245 if (doingdir) { 1246 error = ENOTDIR; 1247 goto err_out; 1248 } 1249 1250 if (SDEV_IS_PERSIST((*ndvp))) { 1251 bkstore = 1; 1252 } 1253 1254 /* 1255 * get rid of the node from the directory cache 1256 * note, in case EBUSY is returned, the ZOMBIE 1257 * node is taken care in sdev_mknode. 1258 */ 1259 (void) sdev_dirdelete(nddv, *ndvp); 1260 *ndvp = NULL; 1261 if (bkstore) { 1262 ASSERT(nddv->sdev_attrvp); 1263 error = VOP_REMOVE(nddv->sdev_attrvp, 1264 nnm, cred, NULL, 0); 1265 if (error) 1266 goto err_out; 1267 } 1268 } 1269 } 1270 1271 /* fix the source for a symlink */ 1272 if (vattr.va_type == VLNK) { 1273 if (odv->sdev_symlink == NULL) { 1274 error = sdev_follow_link(odv); 1275 if (error) { 1276 error = ENOENT; 1277 goto err_out; 1278 } 1279 } 1280 ASSERT(odv->sdev_symlink); 1281 link = i_ddi_strdup(odv->sdev_symlink, KM_SLEEP); 1282 } 1283 1284 /* 1285 * make a fresh node from the source attrs 1286 */ 1287 ASSERT(RW_WRITE_HELD(&nddv->sdev_contents)); 1288 error = sdev_mknode(nddv, nnm, ndvp, &vattr, 1289 NULL, (void *)link, cred, SDEV_READY); 1290 1291 if (link) 1292 kmem_free(link, strlen(link) + 1); 1293 1294 if (error) 1295 goto err_out; 1296 ASSERT(*ndvp); 1297 ASSERT((*ndvp)->sdev_state == SDEV_READY); 1298 1299 /* move dir contents */ 1300 if (doingdir) { 1301 for (idv = SDEV_FIRST_ENTRY(odv); idv; 1302 idv = SDEV_NEXT_ENTRY(odv, idv)) { 1303 error = sdev_rnmnode(odv, idv, 1304 (struct sdev_node *)(*ndvp), &ndv, 1305 idv->sdev_name, cred); 1306 if (error) 1307 goto err_out; 1308 ndv = NULL; 1309 } 1310 } 1311 1312 if ((*ndvp)->sdev_attrvp) { 1313 sdev_update_timestamps((*ndvp)->sdev_attrvp, kcred, 1314 AT_CTIME|AT_ATIME); 1315 } else { 1316 ASSERT((*ndvp)->sdev_attr); 1317 gethrestime(&now); 1318 (*ndvp)->sdev_attr->va_ctime = now; 1319 (*ndvp)->sdev_attr->va_atime = now; 1320 } 1321 1322 if (nddv->sdev_attrvp) { 1323 sdev_update_timestamps(nddv->sdev_attrvp, kcred, 1324 AT_MTIME|AT_ATIME); 1325 } else { 1326 ASSERT(nddv->sdev_attr); 1327 gethrestime(&now); 1328 nddv->sdev_attr->va_mtime = now; 1329 nddv->sdev_attr->va_atime = now; 1330 } 1331 rw_exit(&nddv->sdev_contents); 1332 if (!samedir) 1333 rw_exit(&oddv->sdev_contents); 1334 1335 SDEV_RELE(*ndvp); 1336 return (error); 1337 1338 err_out: 1339 rw_exit(&nddv->sdev_contents); 1340 if (!samedir) 1341 rw_exit(&oddv->sdev_contents); 1342 return (error); 1343 } 1344 1345 /* 1346 * Merge sdev_node specific information into an attribute structure. 1347 * 1348 * note: sdev_node is not locked here 1349 */ 1350 void 1351 sdev_vattr_merge(struct sdev_node *dv, struct vattr *vap) 1352 { 1353 struct vnode *vp = SDEVTOV(dv); 1354 1355 vap->va_nlink = dv->sdev_nlink; 1356 vap->va_nodeid = dv->sdev_ino; 1357 vap->va_fsid = SDEVTOV(dv->sdev_dotdot)->v_rdev; 1358 vap->va_type = vp->v_type; 1359 1360 if (vp->v_type == VDIR) { 1361 vap->va_rdev = 0; 1362 vap->va_fsid = vp->v_rdev; 1363 } else if (vp->v_type == VLNK) { 1364 vap->va_rdev = 0; 1365 vap->va_mode &= ~S_IFMT; 1366 vap->va_mode |= S_IFLNK; 1367 } else if ((vp->v_type == VCHR) || (vp->v_type == VBLK)) { 1368 vap->va_rdev = vp->v_rdev; 1369 vap->va_mode &= ~S_IFMT; 1370 if (vap->va_type == VCHR) 1371 vap->va_mode |= S_IFCHR; 1372 else 1373 vap->va_mode |= S_IFBLK; 1374 } else { 1375 vap->va_rdev = 0; 1376 } 1377 } 1378 1379 static struct vattr * 1380 sdev_getdefault_attr(enum vtype type) 1381 { 1382 if (type == VDIR) 1383 return (&sdev_vattr_dir); 1384 else if (type == VCHR) 1385 return (&sdev_vattr_chr); 1386 else if (type == VBLK) 1387 return (&sdev_vattr_blk); 1388 else if (type == VLNK) 1389 return (&sdev_vattr_lnk); 1390 else 1391 return (NULL); 1392 } 1393 int 1394 sdev_to_vp(struct sdev_node *dv, struct vnode **vpp) 1395 { 1396 int rv = 0; 1397 struct vnode *vp = SDEVTOV(dv); 1398 1399 switch (vp->v_type) { 1400 case VCHR: 1401 case VBLK: 1402 /* 1403 * If vnode is a device, return special vnode instead 1404 * (though it knows all about -us- via sp->s_realvp) 1405 */ 1406 *vpp = specvp(vp, vp->v_rdev, vp->v_type, kcred); 1407 VN_RELE(vp); 1408 if (*vpp == NULLVP) 1409 rv = ENOSYS; 1410 break; 1411 default: /* most types are returned as is */ 1412 *vpp = vp; 1413 break; 1414 } 1415 return (rv); 1416 } 1417 1418 /* 1419 * loopback into sdev_lookup() 1420 */ 1421 static struct vnode * 1422 devname_find_by_devpath(char *devpath, struct vattr *vattr) 1423 { 1424 int error = 0; 1425 struct vnode *vp; 1426 1427 error = lookupname(devpath, UIO_SYSSPACE, NO_FOLLOW, NULLVPP, &vp); 1428 if (error) { 1429 return (NULL); 1430 } 1431 1432 if (vattr) 1433 (void) VOP_GETATTR(vp, vattr, 0, kcred, NULL); 1434 return (vp); 1435 } 1436 1437 /* 1438 * the junction between devname and devfs 1439 */ 1440 static struct vnode * 1441 devname_configure_by_path(char *physpath, struct vattr *vattr) 1442 { 1443 int error = 0; 1444 struct vnode *vp; 1445 1446 ASSERT(strncmp(physpath, "/devices/", sizeof ("/devices/") - 1) 1447 == 0); 1448 1449 error = devfs_lookupname(physpath + sizeof ("/devices/") - 1, 1450 NULLVPP, &vp); 1451 if (error != 0) { 1452 if (error == ENODEV) { 1453 cmn_err(CE_CONT, "%s: not found (line %d)\n", 1454 physpath, __LINE__); 1455 } 1456 1457 return (NULL); 1458 } 1459 1460 if (vattr) 1461 (void) VOP_GETATTR(vp, vattr, 0, kcred, NULL); 1462 return (vp); 1463 } 1464 1465 /* 1466 * junction between devname and root file system, e.g. ufs 1467 */ 1468 int 1469 devname_backstore_lookup(struct sdev_node *ddv, char *nm, struct vnode **rvp) 1470 { 1471 struct vnode *rdvp = ddv->sdev_attrvp; 1472 int rval = 0; 1473 1474 ASSERT(rdvp); 1475 1476 rval = VOP_LOOKUP(rdvp, nm, rvp, NULL, 0, NULL, kcred, NULL, NULL, 1477 NULL); 1478 return (rval); 1479 } 1480 1481 static int 1482 sdev_filldir_from_store(struct sdev_node *ddv, int dlen, struct cred *cred) 1483 { 1484 struct sdev_node *dv = NULL; 1485 char *nm; 1486 struct vnode *dirvp; 1487 int error; 1488 vnode_t *vp; 1489 int eof; 1490 struct iovec iov; 1491 struct uio uio; 1492 struct dirent64 *dp; 1493 dirent64_t *dbuf; 1494 size_t dbuflen; 1495 struct vattr vattr; 1496 char *link = NULL; 1497 1498 if (ddv->sdev_attrvp == NULL) 1499 return (0); 1500 if (!(ddv->sdev_flags & SDEV_BUILD)) 1501 return (0); 1502 1503 dirvp = ddv->sdev_attrvp; 1504 VN_HOLD(dirvp); 1505 dbuf = kmem_zalloc(dlen, KM_SLEEP); 1506 1507 uio.uio_iov = &iov; 1508 uio.uio_iovcnt = 1; 1509 uio.uio_segflg = UIO_SYSSPACE; 1510 uio.uio_fmode = 0; 1511 uio.uio_extflg = UIO_COPY_CACHED; 1512 uio.uio_loffset = 0; 1513 uio.uio_llimit = MAXOFFSET_T; 1514 1515 eof = 0; 1516 error = 0; 1517 while (!error && !eof) { 1518 uio.uio_resid = dlen; 1519 iov.iov_base = (char *)dbuf; 1520 iov.iov_len = dlen; 1521 (void) VOP_RWLOCK(dirvp, V_WRITELOCK_FALSE, NULL); 1522 error = VOP_READDIR(dirvp, &uio, kcred, &eof, NULL, 0); 1523 VOP_RWUNLOCK(dirvp, V_WRITELOCK_FALSE, NULL); 1524 1525 dbuflen = dlen - uio.uio_resid; 1526 if (error || dbuflen == 0) 1527 break; 1528 1529 if (!(ddv->sdev_flags & SDEV_BUILD)) { 1530 error = 0; 1531 break; 1532 } 1533 1534 for (dp = dbuf; ((intptr_t)dp < 1535 (intptr_t)dbuf + dbuflen); 1536 dp = (dirent64_t *)((intptr_t)dp + dp->d_reclen)) { 1537 nm = dp->d_name; 1538 1539 if (strcmp(nm, ".") == 0 || 1540 strcmp(nm, "..") == 0) 1541 continue; 1542 1543 vp = NULLVP; 1544 dv = sdev_cache_lookup(ddv, nm); 1545 if (dv) { 1546 if (dv->sdev_state != SDEV_ZOMBIE) { 1547 SDEV_SIMPLE_RELE(dv); 1548 } else { 1549 /* 1550 * A ZOMBIE node may not have been 1551 * cleaned up from the backing store, 1552 * bypass this entry in this case, 1553 * and clean it up from the directory 1554 * cache if this is the last call. 1555 */ 1556 (void) sdev_dirdelete(ddv, dv); 1557 } 1558 continue; 1559 } 1560 1561 /* refill the cache if not already */ 1562 error = devname_backstore_lookup(ddv, nm, &vp); 1563 if (error) 1564 continue; 1565 1566 vattr.va_mask = AT_MODE|AT_UID|AT_GID; 1567 error = VOP_GETATTR(vp, &vattr, 0, cred, NULL); 1568 if (error) 1569 continue; 1570 1571 if (vattr.va_type == VLNK) { 1572 error = sdev_getlink(vp, &link); 1573 if (error) { 1574 continue; 1575 } 1576 ASSERT(link != NULL); 1577 } 1578 1579 if (!rw_tryupgrade(&ddv->sdev_contents)) { 1580 rw_exit(&ddv->sdev_contents); 1581 rw_enter(&ddv->sdev_contents, RW_WRITER); 1582 } 1583 error = sdev_mknode(ddv, nm, &dv, &vattr, vp, link, 1584 cred, SDEV_READY); 1585 rw_downgrade(&ddv->sdev_contents); 1586 1587 if (link != NULL) { 1588 kmem_free(link, strlen(link) + 1); 1589 link = NULL; 1590 } 1591 1592 if (!error) { 1593 ASSERT(dv); 1594 ASSERT(dv->sdev_state != SDEV_ZOMBIE); 1595 SDEV_SIMPLE_RELE(dv); 1596 } 1597 vp = NULL; 1598 dv = NULL; 1599 } 1600 } 1601 1602 done: 1603 VN_RELE(dirvp); 1604 kmem_free(dbuf, dlen); 1605 1606 return (error); 1607 } 1608 1609 void 1610 sdev_filldir_dynamic(struct sdev_node *ddv) 1611 { 1612 int error; 1613 int i; 1614 struct vattr *vap; 1615 char *nm = NULL; 1616 struct sdev_node *dv = NULL; 1617 1618 ASSERT(RW_WRITE_HELD(&ddv->sdev_contents)); 1619 ASSERT((ddv->sdev_flags & SDEV_BUILD)); 1620 1621 vap = sdev_getdefault_attr(VDIR); 1622 for (i = 0; vtab[i].vt_name != NULL; i++) { 1623 nm = vtab[i].vt_name; 1624 ASSERT(RW_WRITE_HELD(&ddv->sdev_contents)); 1625 dv = NULL; 1626 error = sdev_mknode(ddv, nm, &dv, vap, NULL, 1627 NULL, kcred, SDEV_READY); 1628 if (error) { 1629 cmn_err(CE_WARN, "%s/%s: error %d\n", 1630 ddv->sdev_name, nm, error); 1631 } else { 1632 ASSERT(dv); 1633 ASSERT(dv->sdev_state != SDEV_ZOMBIE); 1634 SDEV_SIMPLE_RELE(dv); 1635 } 1636 } 1637 } 1638 1639 /* 1640 * Creating a backing store entry based on sdev_attr. 1641 * This is called either as part of node creation in a persistent directory 1642 * or from setattr/setsecattr to persist access attributes across reboot. 1643 */ 1644 int 1645 sdev_shadow_node(struct sdev_node *dv, struct cred *cred) 1646 { 1647 int error = 0; 1648 struct vnode *dvp = SDEVTOV(dv->sdev_dotdot); 1649 struct vnode *rdvp = VTOSDEV(dvp)->sdev_attrvp; 1650 struct vattr *vap = dv->sdev_attr; 1651 char *nm = dv->sdev_name; 1652 struct vnode *tmpvp, **rvp = &tmpvp, *rrvp = NULL; 1653 1654 ASSERT(dv && dv->sdev_name && rdvp); 1655 ASSERT(RW_WRITE_HELD(&dv->sdev_contents) && dv->sdev_attrvp == NULL); 1656 1657 lookup: 1658 /* try to find it in the backing store */ 1659 error = VOP_LOOKUP(rdvp, nm, rvp, NULL, 0, NULL, cred, NULL, NULL, 1660 NULL); 1661 if (error == 0) { 1662 if (VOP_REALVP(*rvp, &rrvp, NULL) == 0) { 1663 VN_HOLD(rrvp); 1664 VN_RELE(*rvp); 1665 *rvp = rrvp; 1666 } 1667 1668 kmem_free(dv->sdev_attr, sizeof (vattr_t)); 1669 dv->sdev_attr = NULL; 1670 dv->sdev_attrvp = *rvp; 1671 return (0); 1672 } 1673 1674 /* let's try to persist the node */ 1675 gethrestime(&vap->va_atime); 1676 vap->va_mtime = vap->va_atime; 1677 vap->va_ctime = vap->va_atime; 1678 vap->va_mask |= AT_TYPE|AT_MODE; 1679 switch (vap->va_type) { 1680 case VDIR: 1681 error = VOP_MKDIR(rdvp, nm, vap, rvp, cred, NULL, 0, NULL); 1682 sdcmn_err9(("sdev_shadow_node: mkdir vp %p error %d\n", 1683 (void *)(*rvp), error)); 1684 break; 1685 case VCHR: 1686 case VBLK: 1687 case VREG: 1688 case VDOOR: 1689 error = VOP_CREATE(rdvp, nm, vap, NONEXCL, VREAD|VWRITE, 1690 rvp, cred, 0, NULL, NULL); 1691 sdcmn_err9(("sdev_shadow_node: create vp %p, error %d\n", 1692 (void *)(*rvp), error)); 1693 if (!error) 1694 VN_RELE(*rvp); 1695 break; 1696 case VLNK: 1697 ASSERT(dv->sdev_symlink); 1698 error = VOP_SYMLINK(rdvp, nm, vap, dv->sdev_symlink, cred, 1699 NULL, 0); 1700 sdcmn_err9(("sdev_shadow_node: create symlink error %d\n", 1701 error)); 1702 break; 1703 default: 1704 cmn_err(CE_PANIC, "dev: %s: sdev_shadow_node " 1705 "create\n", nm); 1706 /*NOTREACHED*/ 1707 } 1708 1709 /* go back to lookup to factor out spec node and set attrvp */ 1710 if (error == 0) 1711 goto lookup; 1712 1713 sdcmn_err(("cannot persist %s - error %d\n", dv->sdev_path, error)); 1714 return (error); 1715 } 1716 1717 static int 1718 sdev_cache_add(struct sdev_node *ddv, struct sdev_node **dv, char *nm) 1719 { 1720 int error = 0; 1721 struct sdev_node *dup = NULL; 1722 1723 ASSERT(RW_WRITE_HELD(&ddv->sdev_contents)); 1724 if ((dup = sdev_findbyname(ddv, nm)) == NULL) { 1725 sdev_direnter(ddv, *dv); 1726 } else { 1727 if (dup->sdev_state == SDEV_ZOMBIE) { 1728 error = sdev_dirdelete(ddv, dup); 1729 /* 1730 * The ZOMBIE node is still hanging 1731 * around with more than one reference counts. 1732 * Fail the new node creation so that 1733 * the directory cache won't have 1734 * duplicate entries for the same named node 1735 */ 1736 if (error == EBUSY) { 1737 SDEV_SIMPLE_RELE(*dv); 1738 sdev_nodedestroy(*dv, 0); 1739 *dv = NULL; 1740 return (error); 1741 } 1742 sdev_direnter(ddv, *dv); 1743 } else { 1744 ASSERT((*dv)->sdev_state != SDEV_ZOMBIE); 1745 SDEV_SIMPLE_RELE(*dv); 1746 sdev_nodedestroy(*dv, 0); 1747 *dv = dup; 1748 } 1749 } 1750 1751 return (0); 1752 } 1753 1754 static int 1755 sdev_cache_delete(struct sdev_node *ddv, struct sdev_node **dv) 1756 { 1757 ASSERT(RW_WRITE_HELD(&ddv->sdev_contents)); 1758 return (sdev_dirdelete(ddv, *dv)); 1759 } 1760 1761 /* 1762 * update the in-core directory cache 1763 */ 1764 int 1765 sdev_cache_update(struct sdev_node *ddv, struct sdev_node **dv, char *nm, 1766 sdev_cache_ops_t ops) 1767 { 1768 int error = 0; 1769 1770 ASSERT((SDEV_HELD(*dv))); 1771 1772 ASSERT(RW_WRITE_HELD(&ddv->sdev_contents)); 1773 switch (ops) { 1774 case SDEV_CACHE_ADD: 1775 error = sdev_cache_add(ddv, dv, nm); 1776 break; 1777 case SDEV_CACHE_DELETE: 1778 error = sdev_cache_delete(ddv, dv); 1779 break; 1780 default: 1781 break; 1782 } 1783 1784 return (error); 1785 } 1786 1787 /* 1788 * retrieve the named entry from the directory cache 1789 */ 1790 struct sdev_node * 1791 sdev_cache_lookup(struct sdev_node *ddv, char *nm) 1792 { 1793 struct sdev_node *dv = NULL; 1794 1795 ASSERT(RW_LOCK_HELD(&ddv->sdev_contents)); 1796 dv = sdev_findbyname(ddv, nm); 1797 1798 return (dv); 1799 } 1800 1801 /* 1802 * Implicit reconfig for nodes constructed by a link generator 1803 * Start devfsadm if needed, or if devfsadm is in progress, 1804 * prepare to block on devfsadm either completing or 1805 * constructing the desired node. As devfsadmd is global 1806 * in scope, constructing all necessary nodes, we only 1807 * need to initiate it once. 1808 */ 1809 static int 1810 sdev_call_devfsadmd(struct sdev_node *ddv, struct sdev_node *dv, char *nm) 1811 { 1812 int error = 0; 1813 1814 if (DEVNAME_DEVFSADM_IS_RUNNING(devfsadm_state)) { 1815 sdcmn_err6(("lookup: waiting for %s/%s, 0x%x\n", 1816 ddv->sdev_name, nm, devfsadm_state)); 1817 mutex_enter(&dv->sdev_lookup_lock); 1818 SDEV_BLOCK_OTHERS(dv, (SDEV_LOOKUP | SDEV_LGWAITING)); 1819 mutex_exit(&dv->sdev_lookup_lock); 1820 error = 0; 1821 } else if (!DEVNAME_DEVFSADM_HAS_RUN(devfsadm_state)) { 1822 sdcmn_err6(("lookup %s/%s starting devfsadm, 0x%x\n", 1823 ddv->sdev_name, nm, devfsadm_state)); 1824 1825 sdev_devfsadmd_thread(ddv, dv, kcred); 1826 mutex_enter(&dv->sdev_lookup_lock); 1827 SDEV_BLOCK_OTHERS(dv, 1828 (SDEV_LOOKUP | SDEV_LGWAITING)); 1829 mutex_exit(&dv->sdev_lookup_lock); 1830 error = 0; 1831 } else { 1832 error = -1; 1833 } 1834 1835 return (error); 1836 } 1837 1838 static int 1839 sdev_call_modulelookup(struct sdev_node *ddv, struct sdev_node **dvp, char *nm, 1840 int (*fn)(char *, devname_handle_t *, struct cred *), struct cred *cred) 1841 { 1842 struct vnode *rvp = NULL; 1843 int error = 0; 1844 struct vattr *vap; 1845 devname_spec_t spec; 1846 devname_handle_t *hdl; 1847 void *args = NULL; 1848 struct sdev_node *dv = *dvp; 1849 1850 ASSERT(dv && ddv); 1851 hdl = &(dv->sdev_handle); 1852 ASSERT(hdl->dh_data == dv); 1853 mutex_enter(&dv->sdev_lookup_lock); 1854 SDEV_BLOCK_OTHERS(dv, SDEV_LOOKUP); 1855 mutex_exit(&dv->sdev_lookup_lock); 1856 error = (*fn)(nm, hdl, cred); 1857 if (error) { 1858 return (error); 1859 } 1860 1861 spec = hdl->dh_spec; 1862 args = hdl->dh_args; 1863 ASSERT(args); 1864 1865 switch (spec) { 1866 case DEVNAME_NS_PATH: 1867 /* 1868 * symlink of: 1869 * /dev/dir/nm -> /device/... 1870 */ 1871 rvp = devname_configure_by_path((char *)args, NULL); 1872 break; 1873 case DEVNAME_NS_DEV: 1874 /* 1875 * symlink of: 1876 * /dev/dir/nm -> /dev/... 1877 */ 1878 rvp = devname_find_by_devpath((char *)args, NULL); 1879 break; 1880 default: 1881 if (args) 1882 kmem_free((char *)args, strlen(args) + 1); 1883 return (ENOENT); 1884 1885 } 1886 1887 if (rvp == NULL) { 1888 if (args) 1889 kmem_free((char *)args, strlen(args) + 1); 1890 return (ENOENT); 1891 } else { 1892 vap = sdev_getdefault_attr(VLNK); 1893 ASSERT(RW_READ_HELD(&ddv->sdev_contents)); 1894 /* 1895 * Could sdev_mknode return a different dv_node 1896 * once the lock is dropped? 1897 */ 1898 if (!rw_tryupgrade(&ddv->sdev_contents)) { 1899 rw_exit(&ddv->sdev_contents); 1900 rw_enter(&ddv->sdev_contents, RW_WRITER); 1901 } 1902 error = sdev_mknode(ddv, nm, &dv, vap, NULL, args, cred, 1903 SDEV_READY); 1904 rw_downgrade(&ddv->sdev_contents); 1905 if (error) { 1906 if (args) 1907 kmem_free((char *)args, strlen(args) + 1); 1908 return (error); 1909 } else { 1910 mutex_enter(&dv->sdev_lookup_lock); 1911 SDEV_UNBLOCK_OTHERS(dv, SDEV_LOOKUP); 1912 mutex_exit(&dv->sdev_lookup_lock); 1913 error = 0; 1914 } 1915 } 1916 1917 if (args) 1918 kmem_free((char *)args, strlen(args) + 1); 1919 1920 *dvp = dv; 1921 return (0); 1922 } 1923 1924 /* 1925 * Support for specialized device naming construction mechanisms 1926 */ 1927 static int 1928 sdev_call_dircallback(struct sdev_node *ddv, struct sdev_node **dvp, char *nm, 1929 int (*callback)(struct sdev_node *, char *, void **, struct cred *, 1930 void *, char *), int flags, struct cred *cred) 1931 { 1932 int rv = 0; 1933 char *physpath = NULL; 1934 struct vnode *rvp = NULL; 1935 struct vattr vattr; 1936 struct vattr *vap; 1937 struct sdev_node *dv = *dvp; 1938 1939 mutex_enter(&dv->sdev_lookup_lock); 1940 SDEV_BLOCK_OTHERS(dv, SDEV_LOOKUP); 1941 mutex_exit(&dv->sdev_lookup_lock); 1942 1943 /* for non-devfsadm devices */ 1944 if (flags & SDEV_PATH) { 1945 physpath = kmem_zalloc(MAXPATHLEN, KM_SLEEP); 1946 rv = callback(ddv, nm, (void *)&physpath, kcred, NULL, 1947 NULL); 1948 if (rv) { 1949 kmem_free(physpath, MAXPATHLEN); 1950 return (-1); 1951 } 1952 1953 rvp = devname_configure_by_path(physpath, NULL); 1954 if (rvp == NULL) { 1955 sdcmn_err3(("devname_configure_by_path: " 1956 "failed for /dev/%s/%s\n", 1957 ddv->sdev_name, nm)); 1958 kmem_free(physpath, MAXPATHLEN); 1959 rv = -1; 1960 } else { 1961 vap = sdev_getdefault_attr(VLNK); 1962 ASSERT(RW_READ_HELD(&ddv->sdev_contents)); 1963 1964 /* 1965 * Sdev_mknode may return back a different sdev_node 1966 * that was created by another thread that 1967 * raced to the directroy cache before this thread. 1968 * 1969 * With current directory cache mechanism 1970 * (linked list with the sdev_node name as 1971 * the entity key), this is a way to make sure 1972 * only one entry exists for the same name 1973 * in the same directory. The outcome is 1974 * the winner wins. 1975 */ 1976 if (!rw_tryupgrade(&ddv->sdev_contents)) { 1977 rw_exit(&ddv->sdev_contents); 1978 rw_enter(&ddv->sdev_contents, RW_WRITER); 1979 } 1980 rv = sdev_mknode(ddv, nm, &dv, vap, NULL, 1981 (void *)physpath, cred, SDEV_READY); 1982 rw_downgrade(&ddv->sdev_contents); 1983 kmem_free(physpath, MAXPATHLEN); 1984 if (rv) { 1985 return (rv); 1986 } else { 1987 mutex_enter(&dv->sdev_lookup_lock); 1988 SDEV_UNBLOCK_OTHERS(dv, SDEV_LOOKUP); 1989 mutex_exit(&dv->sdev_lookup_lock); 1990 return (0); 1991 } 1992 } 1993 } else if (flags & SDEV_VLINK) { 1994 physpath = kmem_zalloc(MAXPATHLEN, KM_SLEEP); 1995 rv = callback(ddv, nm, (void *)&physpath, kcred, NULL, 1996 NULL); 1997 if (rv) { 1998 kmem_free(physpath, MAXPATHLEN); 1999 return (-1); 2000 } 2001 2002 vap = sdev_getdefault_attr(VLNK); 2003 vap->va_size = strlen(physpath); 2004 ASSERT(RW_READ_HELD(&ddv->sdev_contents)); 2005 2006 if (!rw_tryupgrade(&ddv->sdev_contents)) { 2007 rw_exit(&ddv->sdev_contents); 2008 rw_enter(&ddv->sdev_contents, RW_WRITER); 2009 } 2010 rv = sdev_mknode(ddv, nm, &dv, vap, NULL, 2011 (void *)physpath, cred, SDEV_READY); 2012 rw_downgrade(&ddv->sdev_contents); 2013 kmem_free(physpath, MAXPATHLEN); 2014 if (rv) 2015 return (rv); 2016 2017 mutex_enter(&dv->sdev_lookup_lock); 2018 SDEV_UNBLOCK_OTHERS(dv, SDEV_LOOKUP); 2019 mutex_exit(&dv->sdev_lookup_lock); 2020 return (0); 2021 } else if (flags & SDEV_VNODE) { 2022 /* 2023 * DBNR has its own way to create the device 2024 * and return a backing store vnode in rvp 2025 */ 2026 ASSERT(callback); 2027 rv = callback(ddv, nm, (void *)&rvp, kcred, NULL, NULL); 2028 if (rv || (rvp == NULL)) { 2029 sdcmn_err3(("devname_lookup_func: SDEV_VNODE " 2030 "callback failed \n")); 2031 return (-1); 2032 } 2033 vap = sdev_getdefault_attr(rvp->v_type); 2034 if (vap == NULL) 2035 return (-1); 2036 2037 ASSERT(RW_READ_HELD(&ddv->sdev_contents)); 2038 if (!rw_tryupgrade(&ddv->sdev_contents)) { 2039 rw_exit(&ddv->sdev_contents); 2040 rw_enter(&ddv->sdev_contents, RW_WRITER); 2041 } 2042 rv = sdev_mknode(ddv, nm, &dv, vap, rvp, NULL, 2043 cred, SDEV_READY); 2044 rw_downgrade(&ddv->sdev_contents); 2045 if (rv) 2046 return (rv); 2047 2048 mutex_enter(&dv->sdev_lookup_lock); 2049 SDEV_UNBLOCK_OTHERS(dv, SDEV_LOOKUP); 2050 mutex_exit(&dv->sdev_lookup_lock); 2051 return (0); 2052 } else if (flags & SDEV_VATTR) { 2053 /* 2054 * /dev/pts 2055 * 2056 * DBNR has its own way to create the device 2057 * "0" is returned upon success. 2058 * 2059 * callback is responsible to set the basic attributes, 2060 * e.g. va_type/va_uid/va_gid/ 2061 * dev_t if VCHR or VBLK/ 2062 */ 2063 ASSERT(callback); 2064 rv = callback(ddv, nm, (void *)&vattr, kcred, NULL, NULL); 2065 if (rv) { 2066 sdcmn_err3(("devname_lookup_func: SDEV_NONE " 2067 "callback failed \n")); 2068 return (-1); 2069 } 2070 2071 ASSERT(RW_READ_HELD(&ddv->sdev_contents)); 2072 if (!rw_tryupgrade(&ddv->sdev_contents)) { 2073 rw_exit(&ddv->sdev_contents); 2074 rw_enter(&ddv->sdev_contents, RW_WRITER); 2075 } 2076 rv = sdev_mknode(ddv, nm, &dv, &vattr, NULL, NULL, 2077 cred, SDEV_READY); 2078 rw_downgrade(&ddv->sdev_contents); 2079 2080 if (rv) 2081 return (rv); 2082 2083 mutex_enter(&dv->sdev_lookup_lock); 2084 SDEV_UNBLOCK_OTHERS(dv, SDEV_LOOKUP); 2085 mutex_exit(&dv->sdev_lookup_lock); 2086 return (0); 2087 } else { 2088 impossible(("lookup: %s/%s by %s not supported (%d)\n", 2089 SDEVTOV(ddv)->v_path, nm, curproc->p_user.u_comm, 2090 __LINE__)); 2091 rv = -1; 2092 } 2093 2094 *dvp = dv; 2095 return (rv); 2096 } 2097 2098 static int 2099 is_devfsadm_thread(char *exec_name) 2100 { 2101 /* 2102 * note: because devfsadmd -> /usr/sbin/devfsadm 2103 * it is safe to use "devfsadm" to capture the lookups 2104 * from devfsadm and its daemon version. 2105 */ 2106 if (strcmp(exec_name, "devfsadm") == 0) 2107 return (1); 2108 return (0); 2109 } 2110 2111 2112 /* 2113 * Lookup Order: 2114 * sdev_node cache; 2115 * backing store (SDEV_PERSIST); 2116 * DBNR: a. dir_ops implemented in the loadable modules; 2117 * b. vnode ops in vtab. 2118 */ 2119 int 2120 devname_lookup_func(struct sdev_node *ddv, char *nm, struct vnode **vpp, 2121 struct cred *cred, int (*callback)(struct sdev_node *, char *, void **, 2122 struct cred *, void *, char *), int flags) 2123 { 2124 int rv = 0, nmlen; 2125 struct vnode *rvp = NULL; 2126 struct sdev_node *dv = NULL; 2127 int retried = 0; 2128 int error = 0; 2129 struct devname_nsmap *map = NULL; 2130 struct devname_ops *dirops = NULL; 2131 int (*fn)(char *, devname_handle_t *, struct cred *) = NULL; 2132 struct vattr vattr; 2133 char *lookup_thread = curproc->p_user.u_comm; 2134 int failed_flags = 0; 2135 int (*vtor)(struct sdev_node *) = NULL; 2136 int state; 2137 int parent_state; 2138 char *link = NULL; 2139 2140 if (SDEVTOV(ddv)->v_type != VDIR) 2141 return (ENOTDIR); 2142 2143 /* 2144 * Empty name or ., return node itself. 2145 */ 2146 nmlen = strlen(nm); 2147 if ((nmlen == 0) || ((nmlen == 1) && (nm[0] == '.'))) { 2148 *vpp = SDEVTOV(ddv); 2149 VN_HOLD(*vpp); 2150 return (0); 2151 } 2152 2153 /* 2154 * .., return the parent directory 2155 */ 2156 if ((nmlen == 2) && (strcmp(nm, "..") == 0)) { 2157 *vpp = SDEVTOV(ddv->sdev_dotdot); 2158 VN_HOLD(*vpp); 2159 return (0); 2160 } 2161 2162 rw_enter(&ddv->sdev_contents, RW_READER); 2163 if (ddv->sdev_flags & SDEV_VTOR) { 2164 vtor = (int (*)(struct sdev_node *))sdev_get_vtor(ddv); 2165 ASSERT(vtor); 2166 } 2167 2168 tryagain: 2169 /* 2170 * (a) directory cache lookup: 2171 */ 2172 ASSERT(RW_READ_HELD(&ddv->sdev_contents)); 2173 parent_state = ddv->sdev_state; 2174 dv = sdev_cache_lookup(ddv, nm); 2175 if (dv) { 2176 state = dv->sdev_state; 2177 switch (state) { 2178 case SDEV_INIT: 2179 if (is_devfsadm_thread(lookup_thread)) 2180 break; 2181 2182 /* ZOMBIED parent won't allow node creation */ 2183 if (parent_state == SDEV_ZOMBIE) { 2184 SD_TRACE_FAILED_LOOKUP(ddv, nm, 2185 retried); 2186 goto nolock_notfound; 2187 } 2188 2189 mutex_enter(&dv->sdev_lookup_lock); 2190 /* compensate the threads started after devfsadm */ 2191 if (DEVNAME_DEVFSADM_IS_RUNNING(devfsadm_state) && 2192 !(SDEV_IS_LOOKUP(dv))) 2193 SDEV_BLOCK_OTHERS(dv, 2194 (SDEV_LOOKUP | SDEV_LGWAITING)); 2195 2196 if (SDEV_IS_LOOKUP(dv)) { 2197 failed_flags |= SLF_REBUILT; 2198 rw_exit(&ddv->sdev_contents); 2199 error = sdev_wait4lookup(dv, SDEV_LOOKUP); 2200 mutex_exit(&dv->sdev_lookup_lock); 2201 rw_enter(&ddv->sdev_contents, RW_READER); 2202 2203 if (error != 0) { 2204 SD_TRACE_FAILED_LOOKUP(ddv, nm, 2205 retried); 2206 goto nolock_notfound; 2207 } 2208 2209 state = dv->sdev_state; 2210 if (state == SDEV_INIT) { 2211 SD_TRACE_FAILED_LOOKUP(ddv, nm, 2212 retried); 2213 goto nolock_notfound; 2214 } else if (state == SDEV_READY) { 2215 goto found; 2216 } else if (state == SDEV_ZOMBIE) { 2217 rw_exit(&ddv->sdev_contents); 2218 SD_TRACE_FAILED_LOOKUP(ddv, nm, 2219 retried); 2220 SDEV_RELE(dv); 2221 goto lookup_failed; 2222 } 2223 } else { 2224 mutex_exit(&dv->sdev_lookup_lock); 2225 } 2226 break; 2227 case SDEV_READY: 2228 goto found; 2229 case SDEV_ZOMBIE: 2230 rw_exit(&ddv->sdev_contents); 2231 SD_TRACE_FAILED_LOOKUP(ddv, nm, retried); 2232 SDEV_RELE(dv); 2233 goto lookup_failed; 2234 default: 2235 rw_exit(&ddv->sdev_contents); 2236 SD_TRACE_FAILED_LOOKUP(ddv, nm, retried); 2237 sdev_lookup_failed(ddv, nm, failed_flags); 2238 *vpp = NULLVP; 2239 return (ENOENT); 2240 } 2241 } 2242 ASSERT(RW_READ_HELD(&ddv->sdev_contents)); 2243 2244 /* 2245 * ZOMBIED parent does not allow new node creation. 2246 * bail out early 2247 */ 2248 if (parent_state == SDEV_ZOMBIE) { 2249 rw_exit(&ddv->sdev_contents); 2250 *vpp = NULL; 2251 SD_TRACE_FAILED_LOOKUP(ddv, nm, retried); 2252 return (ENOENT); 2253 } 2254 2255 /* 2256 * (b0): backing store lookup 2257 * SDEV_PERSIST is default except: 2258 * 1) pts nodes 2259 * 2) non-chmod'ed local nodes 2260 */ 2261 if (SDEV_IS_PERSIST(ddv)) { 2262 error = devname_backstore_lookup(ddv, nm, &rvp); 2263 2264 if (!error) { 2265 sdcmn_err3(("devname_backstore_lookup: " 2266 "found attrvp %p for %s\n", (void *)rvp, nm)); 2267 2268 vattr.va_mask = AT_MODE|AT_UID|AT_GID; 2269 error = VOP_GETATTR(rvp, &vattr, 0, cred, NULL); 2270 if (error) { 2271 rw_exit(&ddv->sdev_contents); 2272 if (dv) 2273 SDEV_RELE(dv); 2274 SD_TRACE_FAILED_LOOKUP(ddv, nm, retried); 2275 sdev_lookup_failed(ddv, nm, failed_flags); 2276 *vpp = NULLVP; 2277 return (ENOENT); 2278 } 2279 2280 if (vattr.va_type == VLNK) { 2281 error = sdev_getlink(rvp, &link); 2282 if (error) { 2283 rw_exit(&ddv->sdev_contents); 2284 if (dv) 2285 SDEV_RELE(dv); 2286 SD_TRACE_FAILED_LOOKUP(ddv, nm, 2287 retried); 2288 sdev_lookup_failed(ddv, nm, 2289 failed_flags); 2290 *vpp = NULLVP; 2291 return (ENOENT); 2292 } 2293 ASSERT(link != NULL); 2294 } 2295 2296 if (!rw_tryupgrade(&ddv->sdev_contents)) { 2297 rw_exit(&ddv->sdev_contents); 2298 rw_enter(&ddv->sdev_contents, RW_WRITER); 2299 } 2300 error = sdev_mknode(ddv, nm, &dv, &vattr, 2301 rvp, link, cred, SDEV_READY); 2302 rw_downgrade(&ddv->sdev_contents); 2303 2304 if (link != NULL) { 2305 kmem_free(link, strlen(link) + 1); 2306 link = NULL; 2307 } 2308 2309 if (error) { 2310 SD_TRACE_FAILED_LOOKUP(ddv, nm, retried); 2311 rw_exit(&ddv->sdev_contents); 2312 if (dv) 2313 SDEV_RELE(dv); 2314 goto lookup_failed; 2315 } else { 2316 goto found; 2317 } 2318 } else if (retried) { 2319 rw_exit(&ddv->sdev_contents); 2320 sdcmn_err3(("retry of lookup of %s/%s: failed\n", 2321 ddv->sdev_name, nm)); 2322 if (dv) 2323 SDEV_RELE(dv); 2324 SD_TRACE_FAILED_LOOKUP(ddv, nm, retried); 2325 sdev_lookup_failed(ddv, nm, failed_flags); 2326 *vpp = NULLVP; 2327 return (ENOENT); 2328 } 2329 } 2330 2331 2332 /* first thread that is doing the lookup on this node */ 2333 if (!dv) { 2334 if (!rw_tryupgrade(&ddv->sdev_contents)) { 2335 rw_exit(&ddv->sdev_contents); 2336 rw_enter(&ddv->sdev_contents, RW_WRITER); 2337 } 2338 error = sdev_mknode(ddv, nm, &dv, NULL, NULL, NULL, 2339 cred, SDEV_INIT); 2340 if (!dv) { 2341 rw_exit(&ddv->sdev_contents); 2342 SD_TRACE_FAILED_LOOKUP(ddv, nm, retried); 2343 sdev_lookup_failed(ddv, nm, failed_flags); 2344 *vpp = NULLVP; 2345 return (ENOENT); 2346 } 2347 rw_downgrade(&ddv->sdev_contents); 2348 } 2349 ASSERT(dv); 2350 ASSERT(SDEV_HELD(dv)); 2351 2352 if (SDEV_IS_NO_NCACHE(dv)) { 2353 failed_flags |= SLF_NO_NCACHE; 2354 } 2355 2356 if (SDEV_IS_GLOBAL(ddv)) { 2357 map = sdev_get_map(ddv, 1); 2358 dirops = map ? map->dir_ops : NULL; 2359 fn = dirops ? dirops->devnops_lookup : NULL; 2360 } 2361 2362 /* 2363 * (b1) invoking devfsadm once per life time for devfsadm nodes 2364 */ 2365 if ((fn == NULL) && !callback) { 2366 2367 if (sdev_reconfig_boot || !i_ddi_io_initialized() || 2368 SDEV_IS_DYNAMIC(ddv) || SDEV_IS_NO_NCACHE(dv) || 2369 ((moddebug & MODDEBUG_FINI_EBUSY) != 0)) { 2370 ASSERT(SDEV_HELD(dv)); 2371 SD_TRACE_FAILED_LOOKUP(ddv, nm, retried); 2372 goto nolock_notfound; 2373 } 2374 2375 /* 2376 * filter out known non-existent devices recorded 2377 * during initial reconfiguration boot for which 2378 * reconfig should not be done and lookup may 2379 * be short-circuited now. 2380 */ 2381 if (sdev_lookup_filter(ddv, nm)) { 2382 SD_TRACE_FAILED_LOOKUP(ddv, nm, retried); 2383 goto nolock_notfound; 2384 } 2385 2386 /* bypassing devfsadm internal nodes */ 2387 if (is_devfsadm_thread(lookup_thread)) { 2388 SD_TRACE_FAILED_LOOKUP(ddv, nm, retried); 2389 goto nolock_notfound; 2390 } 2391 2392 if (sdev_reconfig_disable) { 2393 SD_TRACE_FAILED_LOOKUP(ddv, nm, retried); 2394 goto nolock_notfound; 2395 } 2396 2397 error = sdev_call_devfsadmd(ddv, dv, nm); 2398 if (error == 0) { 2399 sdcmn_err8(("lookup of %s/%s by %s: reconfig\n", 2400 ddv->sdev_name, nm, curproc->p_user.u_comm)); 2401 if (sdev_reconfig_verbose) { 2402 cmn_err(CE_CONT, 2403 "?lookup of %s/%s by %s: reconfig\n", 2404 ddv->sdev_name, nm, curproc->p_user.u_comm); 2405 } 2406 retried = 1; 2407 failed_flags |= SLF_REBUILT; 2408 ASSERT(dv->sdev_state != SDEV_ZOMBIE); 2409 SDEV_SIMPLE_RELE(dv); 2410 goto tryagain; 2411 } else { 2412 SD_TRACE_FAILED_LOOKUP(ddv, nm, retried); 2413 goto nolock_notfound; 2414 } 2415 } 2416 2417 /* 2418 * (b2) Directory Based Name Resolution (DBNR): 2419 * ddv - parent 2420 * nm - /dev/(ddv->sdev_name)/nm 2421 * 2422 * note: module vnode ops take precedence than the build-in ones 2423 */ 2424 if (fn) { 2425 error = sdev_call_modulelookup(ddv, &dv, nm, fn, cred); 2426 if (error) { 2427 SD_TRACE_FAILED_LOOKUP(ddv, nm, retried); 2428 goto notfound; 2429 } else { 2430 goto found; 2431 } 2432 } else if (callback) { 2433 error = sdev_call_dircallback(ddv, &dv, nm, callback, 2434 flags, cred); 2435 if (error == 0) { 2436 goto found; 2437 } else { 2438 SD_TRACE_FAILED_LOOKUP(ddv, nm, retried); 2439 goto notfound; 2440 } 2441 } 2442 ASSERT(rvp); 2443 2444 found: 2445 ASSERT(!(dv->sdev_flags & SDEV_STALE)); 2446 ASSERT(dv->sdev_state == SDEV_READY); 2447 if (vtor) { 2448 /* 2449 * Check validity of returned node 2450 */ 2451 switch (vtor(dv)) { 2452 case SDEV_VTOR_VALID: 2453 break; 2454 case SDEV_VTOR_INVALID: 2455 SD_TRACE_FAILED_LOOKUP(ddv, nm, retried); 2456 sdcmn_err7(("lookup: destroy invalid " 2457 "node: %s(%p)\n", dv->sdev_name, (void *)dv)); 2458 goto nolock_notfound; 2459 case SDEV_VTOR_SKIP: 2460 sdcmn_err7(("lookup: node not applicable - " 2461 "skipping: %s(%p)\n", dv->sdev_name, (void *)dv)); 2462 rw_exit(&ddv->sdev_contents); 2463 SD_TRACE_FAILED_LOOKUP(ddv, nm, retried); 2464 SDEV_RELE(dv); 2465 goto lookup_failed; 2466 default: 2467 cmn_err(CE_PANIC, 2468 "dev fs: validator failed: %s(%p)\n", 2469 dv->sdev_name, (void *)dv); 2470 break; 2471 /*NOTREACHED*/ 2472 } 2473 } 2474 2475 if ((SDEVTOV(dv)->v_type == VDIR) && SDEV_IS_GLOBAL(dv)) { 2476 rw_enter(&dv->sdev_contents, RW_READER); 2477 (void) sdev_get_map(dv, 1); 2478 rw_exit(&dv->sdev_contents); 2479 } 2480 rw_exit(&ddv->sdev_contents); 2481 rv = sdev_to_vp(dv, vpp); 2482 sdcmn_err3(("devname_lookup_func: returning vp %p v_count %d state %d " 2483 "for nm %s, error %d\n", (void *)*vpp, (*vpp)->v_count, 2484 dv->sdev_state, nm, rv)); 2485 return (rv); 2486 2487 notfound: 2488 mutex_enter(&dv->sdev_lookup_lock); 2489 SDEV_UNBLOCK_OTHERS(dv, SDEV_LOOKUP); 2490 mutex_exit(&dv->sdev_lookup_lock); 2491 nolock_notfound: 2492 /* 2493 * Destroy the node that is created for synchronization purposes. 2494 */ 2495 sdcmn_err3(("devname_lookup_func: %s with state %d\n", 2496 nm, dv->sdev_state)); 2497 ASSERT(RW_READ_HELD(&ddv->sdev_contents)); 2498 if (dv->sdev_state == SDEV_INIT) { 2499 if (!rw_tryupgrade(&ddv->sdev_contents)) { 2500 rw_exit(&ddv->sdev_contents); 2501 rw_enter(&ddv->sdev_contents, RW_WRITER); 2502 } 2503 2504 /* 2505 * Node state may have changed during the lock 2506 * changes. Re-check. 2507 */ 2508 if (dv->sdev_state == SDEV_INIT) { 2509 (void) sdev_dirdelete(ddv, dv); 2510 rw_exit(&ddv->sdev_contents); 2511 sdev_lookup_failed(ddv, nm, failed_flags); 2512 *vpp = NULL; 2513 return (ENOENT); 2514 } 2515 } 2516 2517 rw_exit(&ddv->sdev_contents); 2518 SDEV_RELE(dv); 2519 2520 lookup_failed: 2521 sdev_lookup_failed(ddv, nm, failed_flags); 2522 *vpp = NULL; 2523 return (ENOENT); 2524 } 2525 2526 /* 2527 * Given a directory node, mark all nodes beneath as 2528 * STALE, i.e. nodes that don't exist as far as new 2529 * consumers are concerned. Remove them from the 2530 * list of directory entries so that no lookup or 2531 * directory traversal will find them. The node 2532 * not deallocated so existing holds are not affected. 2533 */ 2534 void 2535 sdev_stale(struct sdev_node *ddv) 2536 { 2537 struct sdev_node *dv; 2538 struct vnode *vp; 2539 2540 ASSERT(SDEVTOV(ddv)->v_type == VDIR); 2541 2542 rw_enter(&ddv->sdev_contents, RW_WRITER); 2543 for (dv = SDEV_FIRST_ENTRY(ddv); dv; dv = SDEV_NEXT_ENTRY(ddv, dv)) { 2544 vp = SDEVTOV(dv); 2545 if (vp->v_type == VDIR) 2546 sdev_stale(dv); 2547 2548 sdcmn_err9(("sdev_stale: setting stale %s\n", 2549 dv->sdev_path)); 2550 dv->sdev_flags |= SDEV_STALE; 2551 avl_remove(&ddv->sdev_entries, dv); 2552 } 2553 ddv->sdev_flags |= SDEV_BUILD; 2554 rw_exit(&ddv->sdev_contents); 2555 } 2556 2557 /* 2558 * Given a directory node, clean out all the nodes beneath. 2559 * If expr is specified, clean node with names matching expr. 2560 * If SDEV_ENFORCE is specified in flags, busy nodes are made stale, 2561 * so they are excluded from future lookups. 2562 */ 2563 int 2564 sdev_cleandir(struct sdev_node *ddv, char *expr, uint_t flags) 2565 { 2566 int error = 0; 2567 int busy = 0; 2568 struct vnode *vp; 2569 struct sdev_node *dv, *next = NULL; 2570 int bkstore = 0; 2571 int len = 0; 2572 char *bks_name = NULL; 2573 2574 ASSERT(SDEVTOV(ddv)->v_type == VDIR); 2575 2576 /* 2577 * We try our best to destroy all unused sdev_node's 2578 */ 2579 rw_enter(&ddv->sdev_contents, RW_WRITER); 2580 for (dv = SDEV_FIRST_ENTRY(ddv); dv; dv = next) { 2581 next = SDEV_NEXT_ENTRY(ddv, dv); 2582 vp = SDEVTOV(dv); 2583 2584 if (expr && gmatch(dv->sdev_name, expr) == 0) 2585 continue; 2586 2587 if (vp->v_type == VDIR && 2588 sdev_cleandir(dv, NULL, flags) != 0) { 2589 sdcmn_err9(("sdev_cleandir: dir %s busy\n", 2590 dv->sdev_name)); 2591 busy++; 2592 continue; 2593 } 2594 2595 if (vp->v_count > 0 && (flags & SDEV_ENFORCE) == 0) { 2596 sdcmn_err9(("sdev_cleandir: dir %s busy\n", 2597 dv->sdev_name)); 2598 busy++; 2599 continue; 2600 } 2601 2602 /* 2603 * at this point, either dv is not held or SDEV_ENFORCE 2604 * is specified. In either case, dv needs to be deleted 2605 */ 2606 SDEV_HOLD(dv); 2607 2608 bkstore = SDEV_IS_PERSIST(dv) ? 1 : 0; 2609 if (bkstore && (vp->v_type == VDIR)) 2610 bkstore += 1; 2611 2612 if (bkstore) { 2613 len = strlen(dv->sdev_name) + 1; 2614 bks_name = kmem_alloc(len, KM_SLEEP); 2615 bcopy(dv->sdev_name, bks_name, len); 2616 } 2617 2618 error = sdev_dirdelete(ddv, dv); 2619 2620 if (error == EBUSY) { 2621 sdcmn_err9(("sdev_cleandir: dir busy\n")); 2622 busy++; 2623 } 2624 2625 /* take care the backing store clean up */ 2626 if (bkstore && (error == 0)) { 2627 ASSERT(bks_name); 2628 ASSERT(ddv->sdev_attrvp); 2629 2630 if (bkstore == 1) { 2631 error = VOP_REMOVE(ddv->sdev_attrvp, 2632 bks_name, kcred, NULL, 0); 2633 } else if (bkstore == 2) { 2634 error = VOP_RMDIR(ddv->sdev_attrvp, 2635 bks_name, ddv->sdev_attrvp, kcred, NULL, 0); 2636 } 2637 2638 /* do not propagate the backing store errors */ 2639 if (error) { 2640 sdcmn_err9(("sdev_cleandir: backing store" 2641 "not cleaned\n")); 2642 error = 0; 2643 } 2644 2645 bkstore = 0; 2646 kmem_free(bks_name, len); 2647 bks_name = NULL; 2648 len = 0; 2649 } 2650 } 2651 2652 ddv->sdev_flags |= SDEV_BUILD; 2653 rw_exit(&ddv->sdev_contents); 2654 2655 if (busy) { 2656 error = EBUSY; 2657 } 2658 2659 return (error); 2660 } 2661 2662 /* 2663 * a convenient wrapper for readdir() funcs 2664 */ 2665 size_t 2666 add_dir_entry(dirent64_t *de, char *nm, size_t size, ino_t ino, offset_t off) 2667 { 2668 size_t reclen = DIRENT64_RECLEN(strlen(nm)); 2669 if (reclen > size) 2670 return (0); 2671 2672 de->d_ino = (ino64_t)ino; 2673 de->d_off = (off64_t)off + 1; 2674 de->d_reclen = (ushort_t)reclen; 2675 (void) strncpy(de->d_name, nm, DIRENT64_NAMELEN(reclen)); 2676 return (reclen); 2677 } 2678 2679 /* 2680 * sdev_mount service routines 2681 */ 2682 int 2683 sdev_copyin_mountargs(struct mounta *uap, struct sdev_mountargs *args) 2684 { 2685 int error; 2686 2687 if (uap->datalen != sizeof (*args)) 2688 return (EINVAL); 2689 2690 if (error = copyin(uap->dataptr, args, sizeof (*args))) { 2691 cmn_err(CE_WARN, "sdev_copyin_mountargs: can not" 2692 "get user data. error %d\n", error); 2693 return (EFAULT); 2694 } 2695 2696 return (0); 2697 } 2698 2699 #ifdef nextdp 2700 #undef nextdp 2701 #endif 2702 #define nextdp(dp) ((struct dirent64 *) \ 2703 (intptr_t)((char *)(dp) + (dp)->d_reclen)) 2704 2705 /* 2706 * readdir helper func 2707 */ 2708 int 2709 devname_readdir_func(vnode_t *vp, uio_t *uiop, cred_t *cred, int *eofp, 2710 int flags) 2711 { 2712 struct sdev_node *ddv = VTOSDEV(vp); 2713 struct sdev_node *dv; 2714 dirent64_t *dp; 2715 ulong_t outcount = 0; 2716 size_t namelen; 2717 ulong_t alloc_count; 2718 void *outbuf; 2719 struct iovec *iovp; 2720 int error = 0; 2721 size_t reclen; 2722 offset_t diroff; 2723 offset_t soff; 2724 int this_reclen; 2725 struct devname_nsmap *map = NULL; 2726 struct devname_ops *dirops = NULL; 2727 int (*fn)(devname_handle_t *, struct cred *) = NULL; 2728 int (*vtor)(struct sdev_node *) = NULL; 2729 struct vattr attr; 2730 timestruc_t now; 2731 2732 ASSERT(ddv->sdev_attr || ddv->sdev_attrvp); 2733 ASSERT(RW_READ_HELD(&ddv->sdev_contents)); 2734 2735 if (uiop->uio_loffset >= MAXOFF_T) { 2736 if (eofp) 2737 *eofp = 1; 2738 return (0); 2739 } 2740 2741 if (uiop->uio_iovcnt != 1) 2742 return (EINVAL); 2743 2744 if (vp->v_type != VDIR) 2745 return (ENOTDIR); 2746 2747 if (ddv->sdev_flags & SDEV_VTOR) { 2748 vtor = (int (*)(struct sdev_node *))sdev_get_vtor(ddv); 2749 ASSERT(vtor); 2750 } 2751 2752 if (eofp != NULL) 2753 *eofp = 0; 2754 2755 soff = uiop->uio_loffset; 2756 iovp = uiop->uio_iov; 2757 alloc_count = iovp->iov_len; 2758 dp = outbuf = kmem_alloc(alloc_count, KM_SLEEP); 2759 outcount = 0; 2760 2761 if (ddv->sdev_state == SDEV_ZOMBIE) 2762 goto get_cache; 2763 2764 if (SDEV_IS_GLOBAL(ddv)) { 2765 map = sdev_get_map(ddv, 0); 2766 dirops = map ? map->dir_ops : NULL; 2767 fn = dirops ? dirops->devnops_readdir : NULL; 2768 2769 if (map && map->dir_map) { 2770 /* 2771 * load the name mapping rule database 2772 * through invoking devfsadm and symlink 2773 * all the entries in the map 2774 */ 2775 devname_rdr_result_t rdr_result; 2776 int do_thread = 0; 2777 2778 rw_enter(&map->dir_lock, RW_READER); 2779 do_thread = map->dir_maploaded ? 0 : 1; 2780 rw_exit(&map->dir_lock); 2781 2782 if (do_thread) { 2783 mutex_enter(&ddv->sdev_lookup_lock); 2784 SDEV_BLOCK_OTHERS(ddv, SDEV_READDIR); 2785 mutex_exit(&ddv->sdev_lookup_lock); 2786 2787 sdev_dispatch_to_nsrdr_thread(ddv, 2788 map->dir_map, &rdr_result); 2789 } 2790 } else if ((sdev_boot_state == SDEV_BOOT_STATE_COMPLETE) && 2791 !sdev_reconfig_boot && (flags & SDEV_BROWSE) && 2792 !SDEV_IS_DYNAMIC(ddv) && !SDEV_IS_NO_NCACHE(ddv) && 2793 ((moddebug & MODDEBUG_FINI_EBUSY) == 0) && 2794 !DEVNAME_DEVFSADM_HAS_RUN(devfsadm_state) && 2795 !DEVNAME_DEVFSADM_IS_RUNNING(devfsadm_state) && 2796 !sdev_reconfig_disable) { 2797 /* 2798 * invoking "devfsadm" to do system device reconfig 2799 */ 2800 mutex_enter(&ddv->sdev_lookup_lock); 2801 SDEV_BLOCK_OTHERS(ddv, 2802 (SDEV_READDIR|SDEV_LGWAITING)); 2803 mutex_exit(&ddv->sdev_lookup_lock); 2804 2805 sdcmn_err8(("readdir of %s by %s: reconfig\n", 2806 ddv->sdev_path, curproc->p_user.u_comm)); 2807 if (sdev_reconfig_verbose) { 2808 cmn_err(CE_CONT, 2809 "?readdir of %s by %s: reconfig\n", 2810 ddv->sdev_path, curproc->p_user.u_comm); 2811 } 2812 2813 sdev_devfsadmd_thread(ddv, NULL, kcred); 2814 } else if (DEVNAME_DEVFSADM_IS_RUNNING(devfsadm_state)) { 2815 /* 2816 * compensate the "ls" started later than "devfsadm" 2817 */ 2818 mutex_enter(&ddv->sdev_lookup_lock); 2819 SDEV_BLOCK_OTHERS(ddv, (SDEV_READDIR|SDEV_LGWAITING)); 2820 mutex_exit(&ddv->sdev_lookup_lock); 2821 } 2822 2823 /* 2824 * release the contents lock so that 2825 * the cache may be updated by devfsadmd 2826 */ 2827 rw_exit(&ddv->sdev_contents); 2828 mutex_enter(&ddv->sdev_lookup_lock); 2829 if (SDEV_IS_READDIR(ddv)) 2830 (void) sdev_wait4lookup(ddv, SDEV_READDIR); 2831 mutex_exit(&ddv->sdev_lookup_lock); 2832 rw_enter(&ddv->sdev_contents, RW_READER); 2833 2834 sdcmn_err4(("readdir of directory %s by %s\n", 2835 ddv->sdev_name, curproc->p_user.u_comm)); 2836 if (ddv->sdev_flags & SDEV_BUILD) { 2837 if (SDEV_IS_PERSIST(ddv)) { 2838 error = sdev_filldir_from_store(ddv, 2839 alloc_count, cred); 2840 } 2841 ddv->sdev_flags &= ~SDEV_BUILD; 2842 } 2843 } 2844 2845 get_cache: 2846 /* handle "." and ".." */ 2847 diroff = 0; 2848 if (soff == 0) { 2849 /* first time */ 2850 this_reclen = DIRENT64_RECLEN(1); 2851 if (alloc_count < this_reclen) { 2852 error = EINVAL; 2853 goto done; 2854 } 2855 2856 dp->d_ino = (ino64_t)ddv->sdev_ino; 2857 dp->d_off = (off64_t)1; 2858 dp->d_reclen = (ushort_t)this_reclen; 2859 2860 (void) strncpy(dp->d_name, ".", 2861 DIRENT64_NAMELEN(this_reclen)); 2862 outcount += dp->d_reclen; 2863 dp = nextdp(dp); 2864 } 2865 2866 diroff++; 2867 if (soff <= 1) { 2868 this_reclen = DIRENT64_RECLEN(2); 2869 if (alloc_count < outcount + this_reclen) { 2870 error = EINVAL; 2871 goto done; 2872 } 2873 2874 dp->d_reclen = (ushort_t)this_reclen; 2875 dp->d_ino = (ino64_t)ddv->sdev_dotdot->sdev_ino; 2876 dp->d_off = (off64_t)2; 2877 2878 (void) strncpy(dp->d_name, "..", 2879 DIRENT64_NAMELEN(this_reclen)); 2880 outcount += dp->d_reclen; 2881 2882 dp = nextdp(dp); 2883 } 2884 2885 2886 /* gets the cache */ 2887 diroff++; 2888 for (dv = SDEV_FIRST_ENTRY(ddv); dv; 2889 dv = SDEV_NEXT_ENTRY(ddv, dv), diroff++) { 2890 sdcmn_err3(("sdev_readdir: diroff %lld soff %lld for '%s' \n", 2891 diroff, soff, dv->sdev_name)); 2892 2893 /* bypassing pre-matured nodes */ 2894 if (diroff < soff || (dv->sdev_state != SDEV_READY)) { 2895 sdcmn_err3(("sdev_readdir: pre-mature node " 2896 "%s\n", dv->sdev_name)); 2897 continue; 2898 } 2899 2900 /* 2901 * Check validity of node 2902 */ 2903 if (vtor) { 2904 switch (vtor(dv)) { 2905 case SDEV_VTOR_VALID: 2906 break; 2907 case SDEV_VTOR_INVALID: 2908 case SDEV_VTOR_SKIP: 2909 continue; 2910 default: 2911 cmn_err(CE_PANIC, 2912 "dev fs: validator failed: %s(%p)\n", 2913 dv->sdev_name, (void *)dv); 2914 break; 2915 /*NOTREACHED*/ 2916 } 2917 } 2918 2919 /* 2920 * call back into the module for the validity/bookkeeping 2921 * of this entry 2922 */ 2923 if (fn) { 2924 error = (*fn)(&(dv->sdev_handle), cred); 2925 if (error) { 2926 sdcmn_err4(("sdev_readdir: module did not " 2927 "validate %s\n", dv->sdev_name)); 2928 continue; 2929 } 2930 } 2931 2932 namelen = strlen(dv->sdev_name); 2933 reclen = DIRENT64_RECLEN(namelen); 2934 if (outcount + reclen > alloc_count) { 2935 goto full; 2936 } 2937 dp->d_reclen = (ushort_t)reclen; 2938 dp->d_ino = (ino64_t)dv->sdev_ino; 2939 dp->d_off = (off64_t)diroff + 1; 2940 (void) strncpy(dp->d_name, dv->sdev_name, 2941 DIRENT64_NAMELEN(reclen)); 2942 outcount += reclen; 2943 dp = nextdp(dp); 2944 } 2945 2946 full: 2947 sdcmn_err4(("sdev_readdir: moving %lu bytes: " 2948 "diroff %lld, soff %lld, dv %p\n", outcount, diroff, soff, 2949 (void *)dv)); 2950 2951 if (outcount) 2952 error = uiomove(outbuf, outcount, UIO_READ, uiop); 2953 2954 if (!error) { 2955 uiop->uio_loffset = diroff; 2956 if (eofp) 2957 *eofp = dv ? 0 : 1; 2958 } 2959 2960 2961 if (ddv->sdev_attrvp) { 2962 gethrestime(&now); 2963 attr.va_ctime = now; 2964 attr.va_atime = now; 2965 attr.va_mask = AT_CTIME|AT_ATIME; 2966 2967 (void) VOP_SETATTR(ddv->sdev_attrvp, &attr, 0, kcred, NULL); 2968 } 2969 done: 2970 kmem_free(outbuf, alloc_count); 2971 return (error); 2972 } 2973 2974 static int 2975 sdev_modctl_lookup(const char *path, vnode_t **r_vp) 2976 { 2977 vnode_t *vp; 2978 vnode_t *cvp; 2979 struct sdev_node *svp; 2980 char *nm; 2981 struct pathname pn; 2982 int error; 2983 int persisted = 0; 2984 2985 ASSERT(INGLOBALZONE(curproc)); 2986 2987 if (error = pn_get((char *)path, UIO_SYSSPACE, &pn)) 2988 return (error); 2989 nm = kmem_alloc(MAXNAMELEN, KM_SLEEP); 2990 2991 vp = rootdir; 2992 VN_HOLD(vp); 2993 2994 while (pn_pathleft(&pn)) { 2995 ASSERT(vp->v_type == VDIR || vp->v_type == VLNK); 2996 (void) pn_getcomponent(&pn, nm); 2997 2998 /* 2999 * Deal with the .. special case where we may be 3000 * traversing up across a mount point, to the 3001 * root of this filesystem or global root. 3002 */ 3003 if (nm[0] == '.' && nm[1] == '.' && nm[2] == 0) { 3004 checkforroot: 3005 if (VN_CMP(vp, rootdir)) { 3006 nm[1] = 0; 3007 } else if (vp->v_flag & VROOT) { 3008 vfs_t *vfsp; 3009 cvp = vp; 3010 vfsp = cvp->v_vfsp; 3011 vfs_rlock_wait(vfsp); 3012 vp = cvp->v_vfsp->vfs_vnodecovered; 3013 if (vp == NULL || 3014 (cvp->v_vfsp->vfs_flag & VFS_UNMOUNTED)) { 3015 vfs_unlock(vfsp); 3016 VN_RELE(cvp); 3017 error = EIO; 3018 break; 3019 } 3020 VN_HOLD(vp); 3021 vfs_unlock(vfsp); 3022 VN_RELE(cvp); 3023 cvp = NULL; 3024 goto checkforroot; 3025 } 3026 } 3027 3028 error = VOP_LOOKUP(vp, nm, &cvp, NULL, 0, NULL, kcred, NULL, 3029 NULL, NULL); 3030 if (error) { 3031 VN_RELE(vp); 3032 break; 3033 } 3034 3035 /* traverse mount points encountered on our journey */ 3036 if (vn_ismntpt(cvp) && (error = traverse(&cvp)) != 0) { 3037 VN_RELE(vp); 3038 VN_RELE(cvp); 3039 break; 3040 } 3041 3042 /* 3043 * symbolic link, can be either relative and absolute 3044 */ 3045 if ((cvp->v_type == VLNK) && pn_pathleft(&pn)) { 3046 struct pathname linkpath; 3047 pn_alloc(&linkpath); 3048 if (error = pn_getsymlink(cvp, &linkpath, kcred)) { 3049 pn_free(&linkpath); 3050 break; 3051 } 3052 if (pn_pathleft(&linkpath) == 0) 3053 (void) pn_set(&linkpath, "."); 3054 error = pn_insert(&pn, &linkpath, strlen(nm)); 3055 pn_free(&linkpath); 3056 if (pn.pn_pathlen == 0) { 3057 VN_RELE(vp); 3058 return (ENOENT); 3059 } 3060 if (pn.pn_path[0] == '/') { 3061 pn_skipslash(&pn); 3062 VN_RELE(vp); 3063 VN_RELE(cvp); 3064 vp = rootdir; 3065 VN_HOLD(vp); 3066 } else { 3067 VN_RELE(cvp); 3068 } 3069 continue; 3070 } 3071 3072 VN_RELE(vp); 3073 3074 /* 3075 * Direct the operation to the persisting filesystem 3076 * underlying /dev. Bail if we encounter a 3077 * non-persistent dev entity here. 3078 */ 3079 if (cvp->v_vfsp->vfs_fstype == devtype) { 3080 3081 if ((VTOSDEV(cvp)->sdev_flags & SDEV_PERSIST) == 0) { 3082 error = ENOENT; 3083 VN_RELE(cvp); 3084 break; 3085 } 3086 3087 if (VTOSDEV(cvp) == NULL) { 3088 error = ENOENT; 3089 VN_RELE(cvp); 3090 break; 3091 } 3092 svp = VTOSDEV(cvp); 3093 if ((vp = svp->sdev_attrvp) == NULL) { 3094 error = ENOENT; 3095 VN_RELE(cvp); 3096 break; 3097 } 3098 persisted = 1; 3099 VN_HOLD(vp); 3100 VN_RELE(cvp); 3101 cvp = vp; 3102 } 3103 3104 vp = cvp; 3105 pn_skipslash(&pn); 3106 } 3107 3108 kmem_free(nm, MAXNAMELEN); 3109 pn_free(&pn); 3110 3111 if (error) 3112 return (error); 3113 3114 /* 3115 * Only return persisted nodes in the filesystem underlying /dev. 3116 */ 3117 if (!persisted) { 3118 VN_RELE(vp); 3119 return (ENOENT); 3120 } 3121 3122 *r_vp = vp; 3123 return (0); 3124 } 3125 3126 int 3127 sdev_modctl_readdir(const char *dir, char ***dirlistp, 3128 int *npathsp, int *npathsp_alloc, int checking_empty) 3129 { 3130 char **pathlist = NULL; 3131 char **newlist = NULL; 3132 int npaths = 0; 3133 int npaths_alloc = 0; 3134 dirent64_t *dbuf = NULL; 3135 int n; 3136 char *s; 3137 int error; 3138 vnode_t *vp; 3139 int eof; 3140 struct iovec iov; 3141 struct uio uio; 3142 struct dirent64 *dp; 3143 size_t dlen; 3144 size_t dbuflen; 3145 int ndirents = 64; 3146 char *nm; 3147 3148 error = sdev_modctl_lookup(dir, &vp); 3149 sdcmn_err11(("modctl readdir: %s by %s: %s\n", 3150 dir, curproc->p_user.u_comm, 3151 (error == 0) ? "ok" : "failed")); 3152 if (error) 3153 return (error); 3154 3155 dlen = ndirents * (sizeof (*dbuf)); 3156 dbuf = kmem_alloc(dlen, KM_SLEEP); 3157 3158 uio.uio_iov = &iov; 3159 uio.uio_iovcnt = 1; 3160 uio.uio_segflg = UIO_SYSSPACE; 3161 uio.uio_fmode = 0; 3162 uio.uio_extflg = UIO_COPY_CACHED; 3163 uio.uio_loffset = 0; 3164 uio.uio_llimit = MAXOFFSET_T; 3165 3166 eof = 0; 3167 error = 0; 3168 while (!error && !eof) { 3169 uio.uio_resid = dlen; 3170 iov.iov_base = (char *)dbuf; 3171 iov.iov_len = dlen; 3172 3173 (void) VOP_RWLOCK(vp, V_WRITELOCK_FALSE, NULL); 3174 error = VOP_READDIR(vp, &uio, kcred, &eof, NULL, 0); 3175 VOP_RWUNLOCK(vp, V_WRITELOCK_FALSE, NULL); 3176 3177 dbuflen = dlen - uio.uio_resid; 3178 3179 if (error || dbuflen == 0) 3180 break; 3181 3182 for (dp = dbuf; ((intptr_t)dp < (intptr_t)dbuf + dbuflen); 3183 dp = (dirent64_t *)((intptr_t)dp + dp->d_reclen)) { 3184 3185 nm = dp->d_name; 3186 3187 if (strcmp(nm, ".") == 0 || strcmp(nm, "..") == 0) 3188 continue; 3189 if (npaths == npaths_alloc) { 3190 npaths_alloc += 64; 3191 newlist = (char **) 3192 kmem_zalloc((npaths_alloc + 1) * 3193 sizeof (char *), KM_SLEEP); 3194 if (pathlist) { 3195 bcopy(pathlist, newlist, 3196 npaths * sizeof (char *)); 3197 kmem_free(pathlist, 3198 (npaths + 1) * sizeof (char *)); 3199 } 3200 pathlist = newlist; 3201 } 3202 n = strlen(nm) + 1; 3203 s = kmem_alloc(n, KM_SLEEP); 3204 bcopy(nm, s, n); 3205 pathlist[npaths++] = s; 3206 sdcmn_err11((" %s/%s\n", dir, s)); 3207 3208 /* if checking empty, one entry is as good as many */ 3209 if (checking_empty) { 3210 eof = 1; 3211 break; 3212 } 3213 } 3214 } 3215 3216 exit: 3217 VN_RELE(vp); 3218 3219 if (dbuf) 3220 kmem_free(dbuf, dlen); 3221 3222 if (error) 3223 return (error); 3224 3225 *dirlistp = pathlist; 3226 *npathsp = npaths; 3227 *npathsp_alloc = npaths_alloc; 3228 3229 return (0); 3230 } 3231 3232 void 3233 sdev_modctl_readdir_free(char **pathlist, int npaths, int npaths_alloc) 3234 { 3235 int i, n; 3236 3237 for (i = 0; i < npaths; i++) { 3238 n = strlen(pathlist[i]) + 1; 3239 kmem_free(pathlist[i], n); 3240 } 3241 3242 kmem_free(pathlist, (npaths_alloc + 1) * sizeof (char *)); 3243 } 3244 3245 int 3246 sdev_modctl_devexists(const char *path) 3247 { 3248 vnode_t *vp; 3249 int error; 3250 3251 error = sdev_modctl_lookup(path, &vp); 3252 sdcmn_err11(("modctl dev exists: %s by %s: %s\n", 3253 path, curproc->p_user.u_comm, 3254 (error == 0) ? "ok" : "failed")); 3255 if (error == 0) 3256 VN_RELE(vp); 3257 3258 return (error); 3259 } 3260 3261 void 3262 sdev_update_newnsmap(struct devname_nsmap *map, char *module, char *mapname) 3263 { 3264 rw_enter(&map->dir_lock, RW_WRITER); 3265 if (module) { 3266 ASSERT(map->dir_newmodule == NULL); 3267 map->dir_newmodule = i_ddi_strdup(module, KM_SLEEP); 3268 } 3269 if (mapname) { 3270 ASSERT(map->dir_newmap == NULL); 3271 map->dir_newmap = i_ddi_strdup(mapname, KM_SLEEP); 3272 } 3273 3274 map->dir_invalid = 1; 3275 rw_exit(&map->dir_lock); 3276 } 3277 3278 void 3279 sdev_replace_nsmap(struct devname_nsmap *map, char *module, char *mapname) 3280 { 3281 char *old_module = NULL; 3282 char *old_map = NULL; 3283 3284 ASSERT(RW_LOCK_HELD(&map->dir_lock)); 3285 if (!rw_tryupgrade(&map->dir_lock)) { 3286 rw_exit(&map->dir_lock); 3287 rw_enter(&map->dir_lock, RW_WRITER); 3288 } 3289 3290 old_module = map->dir_module; 3291 if (module) { 3292 if (old_module && strcmp(old_module, module) != 0) { 3293 kmem_free(old_module, strlen(old_module) + 1); 3294 } 3295 map->dir_module = module; 3296 map->dir_newmodule = NULL; 3297 } 3298 3299 old_map = map->dir_map; 3300 if (mapname) { 3301 if (old_map && strcmp(old_map, mapname) != 0) { 3302 kmem_free(old_map, strlen(old_map) + 1); 3303 } 3304 3305 map->dir_map = mapname; 3306 map->dir_newmap = NULL; 3307 } 3308 map->dir_maploaded = 0; 3309 map->dir_invalid = 0; 3310 rw_downgrade(&map->dir_lock); 3311 } 3312 3313 /* 3314 * dir_name should have at least one attribute, 3315 * dir_module 3316 * or dir_map 3317 * or both 3318 * caller holds the devname_nsmaps_lock 3319 */ 3320 void 3321 sdev_insert_nsmap(char *dir_name, char *dir_module, char *dir_map) 3322 { 3323 struct devname_nsmap *map; 3324 int len = 0; 3325 3326 ASSERT(dir_name); 3327 ASSERT(dir_module || dir_map); 3328 ASSERT(MUTEX_HELD(&devname_nsmaps_lock)); 3329 3330 if (map = sdev_get_nsmap_by_dir(dir_name, 1)) { 3331 sdev_update_newnsmap(map, dir_module, dir_map); 3332 return; 3333 } 3334 3335 map = (struct devname_nsmap *)kmem_zalloc(sizeof (*map), KM_SLEEP); 3336 map->dir_name = i_ddi_strdup(dir_name, KM_SLEEP); 3337 if (dir_module) { 3338 map->dir_module = i_ddi_strdup(dir_module, KM_SLEEP); 3339 } 3340 3341 if (dir_map) { 3342 if (dir_map[0] != '/') { 3343 len = strlen(ETC_DEV_DIR) + strlen(dir_map) + 2; 3344 map->dir_map = kmem_zalloc(len, KM_SLEEP); 3345 (void) snprintf(map->dir_map, len, "%s/%s", ETC_DEV_DIR, 3346 dir_map); 3347 } else { 3348 map->dir_map = i_ddi_strdup(dir_map, KM_SLEEP); 3349 } 3350 } 3351 3352 map->dir_ops = NULL; 3353 map->dir_maploaded = 0; 3354 map->dir_invalid = 0; 3355 rw_init(&map->dir_lock, NULL, RW_DEFAULT, NULL); 3356 3357 map->next = devname_nsmaps; 3358 map->prev = NULL; 3359 if (devname_nsmaps) { 3360 devname_nsmaps->prev = map; 3361 } 3362 devname_nsmaps = map; 3363 } 3364 3365 struct devname_nsmap * 3366 sdev_get_nsmap_by_dir(char *dir_path, int locked) 3367 { 3368 struct devname_nsmap *map = NULL; 3369 3370 if (!locked) 3371 mutex_enter(&devname_nsmaps_lock); 3372 for (map = devname_nsmaps; map; map = map->next) { 3373 sdcmn_err6(("sdev_get_nsmap_by_dir: dir %s\n", map->dir_name)); 3374 if (strcmp(map->dir_name, dir_path) == 0) { 3375 if (!locked) 3376 mutex_exit(&devname_nsmaps_lock); 3377 return (map); 3378 } 3379 } 3380 if (!locked) 3381 mutex_exit(&devname_nsmaps_lock); 3382 return (NULL); 3383 } 3384 3385 struct devname_nsmap * 3386 sdev_get_nsmap_by_module(char *mod_name) 3387 { 3388 struct devname_nsmap *map = NULL; 3389 3390 mutex_enter(&devname_nsmaps_lock); 3391 for (map = devname_nsmaps; map; map = map->next) { 3392 sdcmn_err7(("sdev_get_nsmap_by_module: module %s\n", 3393 map->dir_module)); 3394 if (map->dir_module && strcmp(map->dir_module, mod_name) == 0) { 3395 mutex_exit(&devname_nsmaps_lock); 3396 return (map); 3397 } 3398 } 3399 mutex_exit(&devname_nsmaps_lock); 3400 return (NULL); 3401 } 3402 3403 void 3404 sdev_invalidate_nsmaps() 3405 { 3406 struct devname_nsmap *map = NULL; 3407 3408 ASSERT(MUTEX_HELD(&devname_nsmaps_lock)); 3409 3410 if (devname_nsmaps == NULL) 3411 return; 3412 3413 for (map = devname_nsmaps; map; map = map->next) { 3414 rw_enter(&map->dir_lock, RW_WRITER); 3415 map->dir_invalid = 1; 3416 rw_exit(&map->dir_lock); 3417 } 3418 devname_nsmaps_invalidated = 1; 3419 } 3420 3421 3422 int 3423 sdev_nsmaps_loaded() 3424 { 3425 int ret = 0; 3426 3427 mutex_enter(&devname_nsmaps_lock); 3428 if (devname_nsmaps_loaded) 3429 ret = 1; 3430 3431 mutex_exit(&devname_nsmaps_lock); 3432 return (ret); 3433 } 3434 3435 int 3436 sdev_nsmaps_reloaded() 3437 { 3438 int ret = 0; 3439 3440 mutex_enter(&devname_nsmaps_lock); 3441 if (devname_nsmaps_invalidated) 3442 ret = 1; 3443 3444 mutex_exit(&devname_nsmaps_lock); 3445 return (ret); 3446 } 3447 3448 static void 3449 sdev_free_nsmap(struct devname_nsmap *map) 3450 { 3451 ASSERT(map); 3452 if (map->dir_name) 3453 kmem_free(map->dir_name, strlen(map->dir_name) + 1); 3454 if (map->dir_module) 3455 kmem_free(map->dir_module, strlen(map->dir_module) + 1); 3456 if (map->dir_map) 3457 kmem_free(map->dir_map, strlen(map->dir_map) + 1); 3458 rw_destroy(&map->dir_lock); 3459 kmem_free(map, sizeof (*map)); 3460 } 3461 3462 void 3463 sdev_validate_nsmaps() 3464 { 3465 struct devname_nsmap *map = NULL; 3466 struct devname_nsmap *oldmap = NULL; 3467 3468 ASSERT(MUTEX_HELD(&devname_nsmaps_lock)); 3469 map = devname_nsmaps; 3470 while (map) { 3471 rw_enter(&map->dir_lock, RW_READER); 3472 if ((map->dir_invalid == 1) && (map->dir_newmodule == NULL) && 3473 (map->dir_newmap == NULL)) { 3474 oldmap = map; 3475 rw_exit(&map->dir_lock); 3476 if (map->prev) 3477 map->prev->next = oldmap->next; 3478 if (map == devname_nsmaps) 3479 devname_nsmaps = oldmap->next; 3480 3481 map = oldmap->next; 3482 if (map) 3483 map->prev = oldmap->prev; 3484 sdev_free_nsmap(oldmap); 3485 oldmap = NULL; 3486 } else { 3487 rw_exit(&map->dir_lock); 3488 map = map->next; 3489 } 3490 } 3491 devname_nsmaps_invalidated = 0; 3492 } 3493 3494 static int 3495 sdev_map_is_invalid(struct devname_nsmap *map) 3496 { 3497 int ret = 0; 3498 3499 ASSERT(map); 3500 rw_enter(&map->dir_lock, RW_READER); 3501 if (map->dir_invalid) 3502 ret = 1; 3503 rw_exit(&map->dir_lock); 3504 return (ret); 3505 } 3506 3507 static int 3508 sdev_check_map(struct devname_nsmap *map) 3509 { 3510 struct devname_nsmap *mapp; 3511 3512 mutex_enter(&devname_nsmaps_lock); 3513 if (devname_nsmaps == NULL) { 3514 mutex_exit(&devname_nsmaps_lock); 3515 return (1); 3516 } 3517 3518 for (mapp = devname_nsmaps; mapp; mapp = mapp->next) { 3519 if (mapp == map) { 3520 mutex_exit(&devname_nsmaps_lock); 3521 return (0); 3522 } 3523 } 3524 3525 mutex_exit(&devname_nsmaps_lock); 3526 return (1); 3527 3528 } 3529 3530 struct devname_nsmap * 3531 sdev_get_map(struct sdev_node *dv, int validate) 3532 { 3533 struct devname_nsmap *map; 3534 int error; 3535 3536 ASSERT(RW_READ_HELD(&dv->sdev_contents)); 3537 map = dv->sdev_mapinfo; 3538 if (map && sdev_check_map(map)) { 3539 if (!rw_tryupgrade(&dv->sdev_contents)) { 3540 rw_exit(&dv->sdev_contents); 3541 rw_enter(&dv->sdev_contents, RW_WRITER); 3542 } 3543 dv->sdev_mapinfo = NULL; 3544 rw_downgrade(&dv->sdev_contents); 3545 return (NULL); 3546 } 3547 3548 if (validate && (!map || (map && sdev_map_is_invalid(map)))) { 3549 if (!rw_tryupgrade(&dv->sdev_contents)) { 3550 rw_exit(&dv->sdev_contents); 3551 rw_enter(&dv->sdev_contents, RW_WRITER); 3552 } 3553 error = sdev_get_moduleops(dv); 3554 if (!error) 3555 map = dv->sdev_mapinfo; 3556 rw_downgrade(&dv->sdev_contents); 3557 } 3558 return (map); 3559 } 3560 3561 extern int sdev_vnodeops_tbl_size; 3562 3563 /* 3564 * construct a new template with overrides from vtab 3565 */ 3566 static fs_operation_def_t * 3567 sdev_merge_vtab(const fs_operation_def_t tab[]) 3568 { 3569 fs_operation_def_t *new; 3570 const fs_operation_def_t *tab_entry; 3571 3572 /* make a copy of standard vnode ops table */ 3573 new = kmem_alloc(sdev_vnodeops_tbl_size, KM_SLEEP); 3574 bcopy((void *)sdev_vnodeops_tbl, new, sdev_vnodeops_tbl_size); 3575 3576 /* replace the overrides from tab */ 3577 for (tab_entry = tab; tab_entry->name != NULL; tab_entry++) { 3578 fs_operation_def_t *std_entry = new; 3579 while (std_entry->name) { 3580 if (strcmp(tab_entry->name, std_entry->name) == 0) { 3581 std_entry->func = tab_entry->func; 3582 break; 3583 } 3584 std_entry++; 3585 } 3586 if (std_entry->name == NULL) 3587 cmn_err(CE_NOTE, "sdev_merge_vtab: entry %s unused.", 3588 tab_entry->name); 3589 } 3590 3591 return (new); 3592 } 3593 3594 /* free memory allocated by sdev_merge_vtab */ 3595 static void 3596 sdev_free_vtab(fs_operation_def_t *new) 3597 { 3598 kmem_free(new, sdev_vnodeops_tbl_size); 3599 } 3600 3601 void 3602 devname_get_vnode(devname_handle_t *hdl, vnode_t **vpp) 3603 { 3604 struct sdev_node *dv = hdl->dh_data; 3605 3606 ASSERT(dv); 3607 3608 rw_enter(&dv->sdev_contents, RW_READER); 3609 *vpp = SDEVTOV(dv); 3610 rw_exit(&dv->sdev_contents); 3611 } 3612 3613 int 3614 devname_get_path(devname_handle_t *hdl, char **path) 3615 { 3616 struct sdev_node *dv = hdl->dh_data; 3617 3618 ASSERT(dv); 3619 3620 rw_enter(&dv->sdev_contents, RW_READER); 3621 *path = dv->sdev_path; 3622 rw_exit(&dv->sdev_contents); 3623 return (0); 3624 } 3625 3626 int 3627 devname_get_name(devname_handle_t *hdl, char **entry) 3628 { 3629 struct sdev_node *dv = hdl->dh_data; 3630 3631 ASSERT(dv); 3632 rw_enter(&dv->sdev_contents, RW_READER); 3633 *entry = dv->sdev_name; 3634 rw_exit(&dv->sdev_contents); 3635 return (0); 3636 } 3637 3638 void 3639 devname_get_dir_vnode(devname_handle_t *hdl, vnode_t **vpp) 3640 { 3641 struct sdev_node *dv = hdl->dh_data->sdev_dotdot; 3642 3643 ASSERT(dv); 3644 3645 rw_enter(&dv->sdev_contents, RW_READER); 3646 *vpp = SDEVTOV(dv); 3647 rw_exit(&dv->sdev_contents); 3648 } 3649 3650 int 3651 devname_get_dir_path(devname_handle_t *hdl, char **path) 3652 { 3653 struct sdev_node *dv = hdl->dh_data->sdev_dotdot; 3654 3655 ASSERT(dv); 3656 rw_enter(&dv->sdev_contents, RW_READER); 3657 *path = dv->sdev_path; 3658 rw_exit(&dv->sdev_contents); 3659 return (0); 3660 } 3661 3662 int 3663 devname_get_dir_name(devname_handle_t *hdl, char **entry) 3664 { 3665 struct sdev_node *dv = hdl->dh_data->sdev_dotdot; 3666 3667 ASSERT(dv); 3668 rw_enter(&dv->sdev_contents, RW_READER); 3669 *entry = dv->sdev_name; 3670 rw_exit(&dv->sdev_contents); 3671 return (0); 3672 } 3673 3674 int 3675 devname_get_dir_nsmap(devname_handle_t *hdl, struct devname_nsmap **map) 3676 { 3677 struct sdev_node *dv = hdl->dh_data->sdev_dotdot; 3678 3679 ASSERT(dv); 3680 rw_enter(&dv->sdev_contents, RW_READER); 3681 *map = dv->sdev_mapinfo; 3682 rw_exit(&dv->sdev_contents); 3683 return (0); 3684 } 3685 3686 int 3687 devname_get_dir_handle(devname_handle_t *hdl, devname_handle_t **dir_hdl) 3688 { 3689 struct sdev_node *dv = hdl->dh_data->sdev_dotdot; 3690 3691 ASSERT(dv); 3692 rw_enter(&dv->sdev_contents, RW_READER); 3693 *dir_hdl = &(dv->sdev_handle); 3694 rw_exit(&dv->sdev_contents); 3695 return (0); 3696 } 3697 3698 void 3699 devname_set_nodetype(devname_handle_t *hdl, void *args, int spec) 3700 { 3701 struct sdev_node *dv = hdl->dh_data; 3702 3703 ASSERT(dv); 3704 rw_enter(&dv->sdev_contents, RW_WRITER); 3705 hdl->dh_spec = (devname_spec_t)spec; 3706 hdl->dh_args = (void *)i_ddi_strdup((char *)args, KM_SLEEP); 3707 rw_exit(&dv->sdev_contents); 3708 } 3709 3710 /* 3711 * a generic setattr() function 3712 * 3713 * note: flags only supports AT_UID and AT_GID. 3714 * Future enhancements can be done for other types, e.g. AT_MODE 3715 */ 3716 int 3717 devname_setattr_func(struct vnode *vp, struct vattr *vap, int flags, 3718 struct cred *cred, int (*callback)(struct sdev_node *, struct vattr *, 3719 int), int protocol) 3720 { 3721 struct sdev_node *dv = VTOSDEV(vp); 3722 struct sdev_node *parent = dv->sdev_dotdot; 3723 struct vattr *get; 3724 uint_t mask = vap->va_mask; 3725 int error; 3726 3727 /* some sanity checks */ 3728 if (vap->va_mask & AT_NOSET) 3729 return (EINVAL); 3730 3731 if (vap->va_mask & AT_SIZE) { 3732 if (vp->v_type == VDIR) { 3733 return (EISDIR); 3734 } 3735 } 3736 3737 /* no need to set attribute, but do not fail either */ 3738 ASSERT(parent); 3739 rw_enter(&parent->sdev_contents, RW_READER); 3740 if (dv->sdev_state == SDEV_ZOMBIE) { 3741 rw_exit(&parent->sdev_contents); 3742 return (0); 3743 } 3744 3745 /* If backing store exists, just set it. */ 3746 if (dv->sdev_attrvp) { 3747 rw_exit(&parent->sdev_contents); 3748 return (VOP_SETATTR(dv->sdev_attrvp, vap, flags, cred, NULL)); 3749 } 3750 3751 /* 3752 * Otherwise, for nodes with the persistence attribute, create it. 3753 */ 3754 ASSERT(dv->sdev_attr); 3755 if (SDEV_IS_PERSIST(dv) || 3756 ((vap->va_mask & ~AT_TIMES) != 0 && !SDEV_IS_DYNAMIC(dv))) { 3757 sdev_vattr_merge(dv, vap); 3758 rw_enter(&dv->sdev_contents, RW_WRITER); 3759 error = sdev_shadow_node(dv, cred); 3760 rw_exit(&dv->sdev_contents); 3761 rw_exit(&parent->sdev_contents); 3762 3763 if (error) 3764 return (error); 3765 return (VOP_SETATTR(dv->sdev_attrvp, vap, flags, cred, NULL)); 3766 } 3767 3768 3769 /* 3770 * sdev_attr was allocated in sdev_mknode 3771 */ 3772 rw_enter(&dv->sdev_contents, RW_WRITER); 3773 error = secpolicy_vnode_setattr(cred, vp, vap, 3774 dv->sdev_attr, flags, sdev_unlocked_access, dv); 3775 if (error) { 3776 rw_exit(&dv->sdev_contents); 3777 rw_exit(&parent->sdev_contents); 3778 return (error); 3779 } 3780 3781 get = dv->sdev_attr; 3782 if (mask & AT_MODE) { 3783 get->va_mode &= S_IFMT; 3784 get->va_mode |= vap->va_mode & ~S_IFMT; 3785 } 3786 3787 if ((mask & AT_UID) || (mask & AT_GID)) { 3788 if (mask & AT_UID) 3789 get->va_uid = vap->va_uid; 3790 if (mask & AT_GID) 3791 get->va_gid = vap->va_gid; 3792 /* 3793 * a callback must be provided if the protocol is set 3794 */ 3795 if ((protocol & AT_UID) || (protocol & AT_GID)) { 3796 ASSERT(callback); 3797 error = callback(dv, get, protocol); 3798 if (error) { 3799 rw_exit(&dv->sdev_contents); 3800 rw_exit(&parent->sdev_contents); 3801 return (error); 3802 } 3803 } 3804 } 3805 3806 if (mask & AT_ATIME) 3807 get->va_atime = vap->va_atime; 3808 if (mask & AT_MTIME) 3809 get->va_mtime = vap->va_mtime; 3810 if (mask & (AT_MODE | AT_UID | AT_GID | AT_CTIME)) { 3811 gethrestime(&get->va_ctime); 3812 } 3813 3814 sdev_vattr_merge(dv, get); 3815 rw_exit(&dv->sdev_contents); 3816 rw_exit(&parent->sdev_contents); 3817 return (0); 3818 } 3819 3820 /* 3821 * a generic inactive() function 3822 */ 3823 void 3824 devname_inactive_func(struct vnode *vp, struct cred *cred, 3825 void (*callback)(struct vnode *)) 3826 { 3827 int clean; 3828 struct sdev_node *dv = VTOSDEV(vp); 3829 struct sdev_node *ddv = dv->sdev_dotdot; 3830 int state; 3831 struct devname_nsmap *map = NULL; 3832 struct devname_ops *dirops = NULL; 3833 void (*fn)(devname_handle_t *, struct cred *) = NULL; 3834 3835 rw_enter(&ddv->sdev_contents, RW_WRITER); 3836 state = dv->sdev_state; 3837 3838 mutex_enter(&vp->v_lock); 3839 ASSERT(vp->v_count >= 1); 3840 3841 if (vp->v_count == 1 && callback != NULL) 3842 callback(vp); 3843 3844 clean = (vp->v_count == 1) && (state == SDEV_ZOMBIE); 3845 3846 /* 3847 * last ref count on the ZOMBIE node is released. 3848 * clean up the sdev_node, and 3849 * release the hold on the backing store node so that 3850 * the ZOMBIE backing stores also cleaned out. 3851 */ 3852 if (clean) { 3853 ASSERT(ddv); 3854 if (SDEV_IS_GLOBAL(dv)) { 3855 map = ddv->sdev_mapinfo; 3856 dirops = map ? map->dir_ops : NULL; 3857 if (dirops && (fn = dirops->devnops_inactive)) 3858 (*fn)(&(dv->sdev_handle), cred); 3859 } 3860 3861 ddv->sdev_nlink--; 3862 if (vp->v_type == VDIR) { 3863 dv->sdev_nlink--; 3864 } 3865 if ((dv->sdev_flags & SDEV_STALE) == 0) 3866 avl_remove(&ddv->sdev_entries, dv); 3867 dv->sdev_nlink--; 3868 --vp->v_count; 3869 mutex_exit(&vp->v_lock); 3870 sdev_nodedestroy(dv, 0); 3871 } else { 3872 --vp->v_count; 3873 mutex_exit(&vp->v_lock); 3874 } 3875 rw_exit(&ddv->sdev_contents); 3876 } 3877