1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License, Version 1.0 only 6 * (the "License"). You may not use this file except in compliance 7 * with the License. 8 * 9 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 10 * or http://www.opensolaris.org/os/licensing. 11 * See the License for the specific language governing permissions 12 * and limitations under the License. 13 * 14 * When distributing Covered Code, include this CDDL HEADER in each 15 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 16 * If applicable, add the following below this CDDL HEADER, with the 17 * fields enclosed by brackets "[]" replaced with your own identifying 18 * information: Portions Copyright [yyyy] [name of copyright owner] 19 * 20 * CDDL HEADER END 21 */ 22 /* 23 * Copyright 2005 Sun Microsystems, Inc. All rights reserved. 24 * Use is subject to license terms. 25 */ 26 27 #pragma ident "%Z%%M% %I% %E% SMI" 28 29 #include <sys/param.h> 30 #include <sys/types.h> 31 #include <sys/systm.h> 32 #include <sys/cred.h> 33 #include <sys/proc.h> 34 #include <sys/user.h> 35 #include <sys/time.h> 36 #include <sys/vnode.h> 37 #include <sys/vfs.h> 38 #include <sys/file.h> 39 #include <sys/filio.h> 40 #include <sys/uio.h> 41 #include <sys/buf.h> 42 #include <sys/mman.h> 43 #include <sys/tiuser.h> 44 #include <sys/pathname.h> 45 #include <sys/dirent.h> 46 #include <sys/conf.h> 47 #include <sys/debug.h> 48 #include <sys/vmsystm.h> 49 #include <sys/fcntl.h> 50 #include <sys/flock.h> 51 #include <sys/swap.h> 52 #include <sys/errno.h> 53 #include <sys/sysmacros.h> 54 #include <sys/disp.h> 55 #include <sys/kmem.h> 56 #include <sys/cmn_err.h> 57 #include <sys/vtrace.h> 58 #include <sys/mount.h> 59 #include <sys/bootconf.h> 60 #include <sys/dnlc.h> 61 #include <sys/stat.h> 62 #include <sys/acl.h> 63 #include <sys/policy.h> 64 #include <rpc/types.h> 65 66 #include <vm/hat.h> 67 #include <vm/as.h> 68 #include <vm/page.h> 69 #include <vm/pvn.h> 70 #include <vm/seg.h> 71 #include <vm/seg_map.h> 72 #include <vm/seg_vn.h> 73 #include <vm/rm.h> 74 #include <sys/fs/cachefs_fs.h> 75 #include <sys/fs/cachefs_dir.h> 76 #include <sys/fs/cachefs_dlog.h> 77 #include <sys/fs/cachefs_ioctl.h> 78 #include <sys/fs/cachefs_log.h> 79 #include <fs/fs_subr.h> 80 81 int cachefs_dnlc; /* use dnlc, debugging */ 82 83 static void cachefs_attr_setup(vattr_t *srcp, vattr_t *targp, cnode_t *cp, 84 cred_t *cr); 85 static void cachefs_creategid(cnode_t *dcp, cnode_t *newcp, vattr_t *vap, 86 cred_t *cr); 87 static void cachefs_createacl(cnode_t *dcp, cnode_t *newcp); 88 static int cachefs_getaclfromcache(cnode_t *cp, vsecattr_t *vsec); 89 static int cachefs_getacldirvp(cnode_t *cp); 90 static void cachefs_acl2perm(cnode_t *cp, vsecattr_t *vsec); 91 static int cachefs_access_local(void *cp, int mode, cred_t *cr); 92 static int cachefs_acl_access(struct cnode *cp, int mode, cred_t *cr); 93 static int cachefs_push_connected(vnode_t *vp, struct buf *bp, size_t iolen, 94 u_offset_t iooff, cred_t *cr); 95 static int cachefs_push_front(vnode_t *vp, struct buf *bp, size_t iolen, 96 u_offset_t iooff, cred_t *cr); 97 static int cachefs_setattr_connected(vnode_t *vp, vattr_t *vap, int flags, 98 cred_t *cr, caller_context_t *ct); 99 static int cachefs_setattr_disconnected(vnode_t *vp, vattr_t *vap, 100 int flags, cred_t *cr, caller_context_t *ct); 101 static int cachefs_access_connected(struct vnode *vp, int mode, 102 int flags, cred_t *cr); 103 static int cachefs_lookup_back(vnode_t *dvp, char *nm, vnode_t **vpp, 104 cred_t *cr); 105 static int cachefs_symlink_connected(vnode_t *dvp, char *lnm, vattr_t *tva, 106 char *tnm, cred_t *cr); 107 static int cachefs_symlink_disconnected(vnode_t *dvp, char *lnm, 108 vattr_t *tva, char *tnm, cred_t *cr); 109 static int cachefs_link_connected(vnode_t *tdvp, vnode_t *fvp, char *tnm, 110 cred_t *cr); 111 static int cachefs_link_disconnected(vnode_t *tdvp, vnode_t *fvp, 112 char *tnm, cred_t *cr); 113 static int cachefs_mkdir_connected(vnode_t *dvp, char *nm, vattr_t *vap, 114 vnode_t **vpp, cred_t *cr); 115 static int cachefs_mkdir_disconnected(vnode_t *dvp, char *nm, vattr_t *vap, 116 vnode_t **vpp, cred_t *cr); 117 static int cachefs_stickyrmchk(struct cnode *dcp, struct cnode *cp, cred_t *cr); 118 static int cachefs_rmdir_connected(vnode_t *dvp, char *nm, 119 vnode_t *cdir, cred_t *cr, vnode_t *vp); 120 static int cachefs_rmdir_disconnected(vnode_t *dvp, char *nm, 121 vnode_t *cdir, cred_t *cr, vnode_t *vp); 122 static char *cachefs_newname(void); 123 static int cachefs_remove_dolink(vnode_t *dvp, vnode_t *vp, char *nm, 124 cred_t *cr); 125 static int cachefs_rename_connected(vnode_t *odvp, char *onm, 126 vnode_t *ndvp, char *nnm, cred_t *cr, vnode_t *delvp); 127 static int cachefs_rename_disconnected(vnode_t *odvp, char *onm, 128 vnode_t *ndvp, char *nnm, cred_t *cr, vnode_t *delvp); 129 static int cachefs_readdir_connected(vnode_t *vp, uio_t *uiop, cred_t *cr, 130 int *eofp); 131 static int cachefs_readdir_disconnected(vnode_t *vp, uio_t *uiop, 132 cred_t *cr, int *eofp); 133 static int cachefs_readback_translate(cnode_t *cp, uio_t *uiop, 134 cred_t *cr, int *eofp); 135 136 static int cachefs_setattr_common(vnode_t *vp, vattr_t *vap, int flags, 137 cred_t *cr, caller_context_t *ct); 138 139 static int cachefs_open(struct vnode **, int, cred_t *); 140 static int cachefs_close(struct vnode *, int, int, offset_t, 141 cred_t *); 142 static int cachefs_read(struct vnode *, struct uio *, int, cred_t *, 143 caller_context_t *); 144 static int cachefs_write(struct vnode *, struct uio *, int, cred_t *, 145 caller_context_t *); 146 static int cachefs_ioctl(struct vnode *, int, intptr_t, int, cred_t *, 147 int *); 148 static int cachefs_getattr(struct vnode *, struct vattr *, int, 149 cred_t *); 150 static int cachefs_setattr(struct vnode *, struct vattr *, 151 int, cred_t *, caller_context_t *); 152 static int cachefs_access(struct vnode *, int, int, cred_t *); 153 static int cachefs_lookup(struct vnode *, char *, struct vnode **, 154 struct pathname *, int, struct vnode *, cred_t *); 155 static int cachefs_create(struct vnode *, char *, struct vattr *, 156 enum vcexcl, int, struct vnode **, cred_t *, int); 157 static int cachefs_create_connected(vnode_t *dvp, char *nm, 158 vattr_t *vap, enum vcexcl exclusive, int mode, 159 vnode_t **vpp, cred_t *cr); 160 static int cachefs_create_disconnected(vnode_t *dvp, char *nm, 161 vattr_t *vap, enum vcexcl exclusive, int mode, 162 vnode_t **vpp, cred_t *cr); 163 static int cachefs_remove(struct vnode *, char *, cred_t *); 164 static int cachefs_link(struct vnode *, struct vnode *, char *, 165 cred_t *); 166 static int cachefs_rename(struct vnode *, char *, struct vnode *, 167 char *, cred_t *); 168 static int cachefs_mkdir(struct vnode *, char *, struct 169 vattr *, struct vnode **, cred_t *); 170 static int cachefs_rmdir(struct vnode *, char *, struct vnode *, 171 cred_t *); 172 static int cachefs_readdir(struct vnode *, struct uio *, 173 cred_t *, int *); 174 static int cachefs_symlink(struct vnode *, char *, struct vattr *, 175 char *, cred_t *); 176 static int cachefs_readlink(struct vnode *, struct uio *, cred_t *); 177 static int cachefs_readlink_connected(vnode_t *vp, uio_t *uiop, cred_t *cr); 178 static int cachefs_readlink_disconnected(vnode_t *vp, uio_t *uiop); 179 static int cachefs_fsync(struct vnode *, int, cred_t *); 180 static void cachefs_inactive(struct vnode *, cred_t *); 181 static int cachefs_fid(struct vnode *, struct fid *); 182 static int cachefs_rwlock(struct vnode *, int, caller_context_t *); 183 static void cachefs_rwunlock(struct vnode *, int, caller_context_t *); 184 static int cachefs_seek(struct vnode *, offset_t, offset_t *); 185 static int cachefs_frlock(struct vnode *, int, struct flock64 *, 186 int, offset_t, struct flk_callback *, cred_t *); 187 static int cachefs_space(struct vnode *, int, struct flock64 *, int, 188 offset_t, cred_t *, caller_context_t *); 189 static int cachefs_realvp(struct vnode *, struct vnode **); 190 static int cachefs_getpage(struct vnode *, offset_t, size_t, uint_t *, 191 struct page *[], size_t, struct seg *, caddr_t, 192 enum seg_rw, cred_t *); 193 static int cachefs_getapage(struct vnode *, u_offset_t, size_t, uint_t *, 194 struct page *[], size_t, struct seg *, caddr_t, 195 enum seg_rw, cred_t *); 196 static int cachefs_getapage_back(struct vnode *, u_offset_t, size_t, 197 uint_t *, struct page *[], size_t, struct seg *, caddr_t, 198 enum seg_rw, cred_t *); 199 static int cachefs_putpage(struct vnode *, offset_t, size_t, int, 200 cred_t *); 201 static int cachefs_map(struct vnode *, offset_t, struct as *, 202 caddr_t *, size_t, uchar_t, uchar_t, uint_t, cred_t *); 203 static int cachefs_addmap(struct vnode *, offset_t, struct as *, 204 caddr_t, size_t, uchar_t, uchar_t, uint_t, cred_t *); 205 static int cachefs_delmap(struct vnode *, offset_t, struct as *, 206 caddr_t, size_t, uint_t, uint_t, uint_t, cred_t *); 207 static int cachefs_setsecattr(vnode_t *vp, vsecattr_t *vsec, 208 int flag, cred_t *cr); 209 static int cachefs_getsecattr(vnode_t *vp, vsecattr_t *vsec, 210 int flag, cred_t *cr); 211 static int cachefs_shrlock(vnode_t *, int, struct shrlock *, int, 212 cred_t *); 213 static int cachefs_getsecattr_connected(vnode_t *vp, vsecattr_t *vsec, int flag, 214 cred_t *cr); 215 static int cachefs_getsecattr_disconnected(vnode_t *vp, vsecattr_t *vsec, 216 int flag, cred_t *cr); 217 218 static int cachefs_dump(struct vnode *, caddr_t, int, int); 219 static int cachefs_pageio(struct vnode *, page_t *, 220 u_offset_t, size_t, int, cred_t *); 221 static int cachefs_writepage(struct vnode *vp, caddr_t base, 222 int tcount, struct uio *uiop); 223 static int cachefs_pathconf(vnode_t *, int, ulong_t *, cred_t *); 224 225 static int cachefs_read_backfs_nfsv4(vnode_t *vp, uio_t *uiop, int ioflag, 226 cred_t *cr, caller_context_t *ct); 227 static int cachefs_write_backfs_nfsv4(vnode_t *vp, uio_t *uiop, int ioflag, 228 cred_t *cr, caller_context_t *ct); 229 static int cachefs_getattr_backfs_nfsv4(vnode_t *vp, vattr_t *vap, 230 int flags, cred_t *cr); 231 static int cachefs_remove_backfs_nfsv4(vnode_t *dvp, char *nm, cred_t *cr, 232 vnode_t *vp); 233 static int cachefs_getpage_backfs_nfsv4(struct vnode *vp, offset_t off, 234 size_t len, uint_t *protp, struct page *pl[], 235 size_t plsz, struct seg *seg, caddr_t addr, 236 enum seg_rw rw, cred_t *cr); 237 static int cachefs_putpage_backfs_nfsv4(vnode_t *vp, offset_t off, 238 size_t len, int flags, cred_t *cr); 239 static int cachefs_map_backfs_nfsv4(struct vnode *vp, offset_t off, 240 struct as *as, caddr_t *addrp, size_t len, uchar_t prot, 241 uchar_t maxprot, uint_t flags, cred_t *cr); 242 static int cachefs_space_backfs_nfsv4(struct vnode *vp, int cmd, 243 struct flock64 *bfp, int flag, offset_t offset, 244 cred_t *cr, caller_context_t *ct); 245 246 struct vnodeops *cachefs_vnodeops; 247 248 static const fs_operation_def_t cachefs_vnodeops_template[] = { 249 VOPNAME_OPEN, cachefs_open, 250 VOPNAME_CLOSE, cachefs_close, 251 VOPNAME_READ, cachefs_read, 252 VOPNAME_WRITE, cachefs_write, 253 VOPNAME_IOCTL, cachefs_ioctl, 254 VOPNAME_GETATTR, cachefs_getattr, 255 VOPNAME_SETATTR, cachefs_setattr, 256 VOPNAME_ACCESS, cachefs_access, 257 VOPNAME_LOOKUP, cachefs_lookup, 258 VOPNAME_CREATE, cachefs_create, 259 VOPNAME_REMOVE, cachefs_remove, 260 VOPNAME_LINK, cachefs_link, 261 VOPNAME_RENAME, cachefs_rename, 262 VOPNAME_MKDIR, cachefs_mkdir, 263 VOPNAME_RMDIR, cachefs_rmdir, 264 VOPNAME_READDIR, cachefs_readdir, 265 VOPNAME_SYMLINK, cachefs_symlink, 266 VOPNAME_READLINK, cachefs_readlink, 267 VOPNAME_FSYNC, cachefs_fsync, 268 VOPNAME_INACTIVE, (fs_generic_func_p) cachefs_inactive, 269 VOPNAME_FID, cachefs_fid, 270 VOPNAME_RWLOCK, cachefs_rwlock, 271 VOPNAME_RWUNLOCK, (fs_generic_func_p) cachefs_rwunlock, 272 VOPNAME_SEEK, cachefs_seek, 273 VOPNAME_FRLOCK, cachefs_frlock, 274 VOPNAME_SPACE, cachefs_space, 275 VOPNAME_REALVP, cachefs_realvp, 276 VOPNAME_GETPAGE, cachefs_getpage, 277 VOPNAME_PUTPAGE, cachefs_putpage, 278 VOPNAME_MAP, (fs_generic_func_p) cachefs_map, 279 VOPNAME_ADDMAP, (fs_generic_func_p) cachefs_addmap, 280 VOPNAME_DELMAP, cachefs_delmap, 281 VOPNAME_DUMP, cachefs_dump, 282 VOPNAME_PATHCONF, cachefs_pathconf, 283 VOPNAME_PAGEIO, cachefs_pageio, 284 VOPNAME_SETSECATTR, cachefs_setsecattr, 285 VOPNAME_GETSECATTR, cachefs_getsecattr, 286 VOPNAME_SHRLOCK, cachefs_shrlock, 287 NULL, NULL 288 }; 289 290 /* forward declarations of statics */ 291 static void cachefs_modified(cnode_t *cp); 292 static int cachefs_modified_alloc(cnode_t *cp); 293 294 int 295 cachefs_init_vnops(char *name) 296 { 297 return (vn_make_ops(name, 298 cachefs_vnodeops_template, &cachefs_vnodeops)); 299 } 300 301 struct vnodeops * 302 cachefs_getvnodeops(void) 303 { 304 return (cachefs_vnodeops); 305 } 306 307 static int 308 cachefs_open(vnode_t **vpp, int flag, cred_t *cr) 309 { 310 int error = 0; 311 cnode_t *cp = VTOC(*vpp); 312 fscache_t *fscp = C_TO_FSCACHE(cp); 313 int held = 0; 314 int type; 315 int connected = 0; 316 317 #ifdef CFSDEBUG 318 CFS_DEBUG(CFSDEBUG_VOPS) 319 printf("cachefs_open: ENTER vpp %p flag %x\n", 320 (void *)vpp, flag); 321 #endif 322 if (getzoneid() != GLOBAL_ZONEID) { 323 error = EPERM; 324 goto out; 325 } 326 if ((flag & FWRITE) && 327 ((*vpp)->v_type == VDIR || (*vpp)->v_type == VLNK)) { 328 error = EISDIR; 329 goto out; 330 } 331 332 /* 333 * Cachefs only provides pass-through support for NFSv4, 334 * and all vnode operations are passed through to the 335 * back file system. For NFSv4 pass-through to work, only 336 * connected operation is supported, the cnode backvp must 337 * exist, and cachefs optional (eg., disconnectable) flags 338 * are turned off. Assert these conditions to ensure that 339 * the backfilesystem is called for the open operation. 340 */ 341 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 342 CFS_BACKFS_NFSV4_ASSERT_CNODE(cp); 343 344 for (;;) { 345 /* get (or renew) access to the file system */ 346 if (held) { 347 /* Won't loop with NFSv4 connected behavior */ 348 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 349 cachefs_cd_release(fscp); 350 held = 0; 351 } 352 error = cachefs_cd_access(fscp, connected, 0); 353 if (error) 354 goto out; 355 held = 1; 356 357 mutex_enter(&cp->c_statelock); 358 359 /* grab creds if we do not have any yet */ 360 if (cp->c_cred == NULL) { 361 crhold(cr); 362 cp->c_cred = cr; 363 } 364 cp->c_flags |= CN_NEEDOPEN; 365 366 /* if we are disconnected */ 367 if (fscp->fs_cdconnected != CFS_CD_CONNECTED) { 368 /* if we cannot write to the file system */ 369 if ((flag & FWRITE) && CFS_ISFS_WRITE_AROUND(fscp)) { 370 mutex_exit(&cp->c_statelock); 371 connected = 1; 372 continue; 373 } 374 /* 375 * Allow read only requests to continue 376 */ 377 if ((flag & (FWRITE|FREAD)) == FREAD) { 378 /* track the flag for opening the backvp */ 379 cp->c_rdcnt++; 380 mutex_exit(&cp->c_statelock); 381 error = 0; 382 break; 383 } 384 385 /* 386 * check credentials - if this procs 387 * credentials don't match the creds in the 388 * cnode disallow writing while disconnected. 389 */ 390 if (crcmp(cp->c_cred, CRED()) != 0 && 391 secpolicy_vnode_access(CRED(), *vpp, 392 cp->c_attr.va_uid, VWRITE) != 0) { 393 mutex_exit(&cp->c_statelock); 394 connected = 1; 395 continue; 396 } 397 /* to get here, we know that the WRITE flag is on */ 398 cp->c_wrcnt++; 399 if (flag & FREAD) 400 cp->c_rdcnt++; 401 } 402 403 /* else if we are connected */ 404 else { 405 /* if cannot use the cached copy of the file */ 406 if ((flag & FWRITE) && CFS_ISFS_WRITE_AROUND(fscp) && 407 ((cp->c_flags & CN_NOCACHE) == 0)) 408 cachefs_nocache(cp); 409 410 /* pass open to the back file */ 411 if (cp->c_backvp) { 412 cp->c_flags &= ~CN_NEEDOPEN; 413 CFS_DPRINT_BACKFS_NFSV4(fscp, 414 ("cachefs_open (nfsv4): cnode %p, " 415 "backvp %p\n", cp, cp->c_backvp)); 416 error = VOP_OPEN(&cp->c_backvp, flag, cr); 417 if (CFS_TIMEOUT(fscp, error)) { 418 mutex_exit(&cp->c_statelock); 419 cachefs_cd_release(fscp); 420 held = 0; 421 cachefs_cd_timedout(fscp); 422 continue; 423 } else if (error) { 424 mutex_exit(&cp->c_statelock); 425 break; 426 } 427 } else { 428 /* backvp will be VOP_OPEN'd later */ 429 if (flag & FREAD) 430 cp->c_rdcnt++; 431 if (flag & FWRITE) 432 cp->c_wrcnt++; 433 } 434 435 /* 436 * Now perform a consistency check on the file. 437 * If strict consistency then force a check to 438 * the backfs even if the timeout has not expired 439 * for close-to-open consistency. 440 */ 441 type = 0; 442 if (fscp->fs_consttype == CFS_FS_CONST_STRICT) 443 type = C_BACK_CHECK; 444 error = CFSOP_CHECK_COBJECT(fscp, cp, type, cr); 445 if (CFS_TIMEOUT(fscp, error)) { 446 mutex_exit(&cp->c_statelock); 447 cachefs_cd_release(fscp); 448 held = 0; 449 cachefs_cd_timedout(fscp); 450 continue; 451 } 452 } 453 mutex_exit(&cp->c_statelock); 454 break; 455 } 456 if (held) 457 cachefs_cd_release(fscp); 458 out: 459 #ifdef CFS_CD_DEBUG 460 ASSERT((curthread->t_flag & T_CD_HELD) == 0); 461 #endif 462 #ifdef CFSDEBUG 463 CFS_DEBUG(CFSDEBUG_VOPS) 464 printf("cachefs_open: EXIT vpp %p error %d\n", 465 (void *)vpp, error); 466 #endif 467 return (error); 468 } 469 470 /* ARGSUSED */ 471 static int 472 cachefs_close(vnode_t *vp, int flag, int count, offset_t offset, cred_t *cr) 473 { 474 int error = 0; 475 cnode_t *cp = VTOC(vp); 476 fscache_t *fscp = C_TO_FSCACHE(cp); 477 int held = 0; 478 int connected = 0; 479 int close_cnt = 1; 480 cachefscache_t *cachep; 481 482 #ifdef CFSDEBUG 483 CFS_DEBUG(CFSDEBUG_VOPS) 484 printf("cachefs_close: ENTER vp %p\n", (void *)vp); 485 #endif 486 /* 487 * Cachefs only provides pass-through support for NFSv4, 488 * and all vnode operations are passed through to the 489 * back file system. For NFSv4 pass-through to work, only 490 * connected operation is supported, the cnode backvp must 491 * exist, and cachefs optional (eg., disconnectable) flags 492 * are turned off. Assert these conditions to ensure that 493 * the backfilesystem is called for the close operation. 494 */ 495 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 496 CFS_BACKFS_NFSV4_ASSERT_CNODE(cp); 497 498 /* 499 * File could have been passed in or inherited from the global zone, so 500 * we don't want to flat out reject the request; we'll just leave things 501 * the way they are and let the backfs (NFS) deal with it. 502 */ 503 /* get rid of any local locks */ 504 if (CFS_ISFS_LLOCK(fscp)) { 505 (void) cleanlocks(vp, ttoproc(curthread)->p_pid, 0); 506 } 507 508 /* clean up if this is the daemon closing down */ 509 if ((fscp->fs_cddaemonid == ttoproc(curthread)->p_pid) && 510 ((ttoproc(curthread)->p_pid) != 0) && 511 (vp == fscp->fs_rootvp) && 512 (count == 1)) { 513 mutex_enter(&fscp->fs_cdlock); 514 fscp->fs_cddaemonid = 0; 515 if (fscp->fs_dlogfile) 516 fscp->fs_cdconnected = CFS_CD_DISCONNECTED; 517 else 518 fscp->fs_cdconnected = CFS_CD_CONNECTED; 519 cv_broadcast(&fscp->fs_cdwaitcv); 520 mutex_exit(&fscp->fs_cdlock); 521 if (fscp->fs_flags & CFS_FS_ROOTFS) { 522 cachep = fscp->fs_cache; 523 mutex_enter(&cachep->c_contentslock); 524 ASSERT(cachep->c_rootdaemonid != 0); 525 cachep->c_rootdaemonid = 0; 526 mutex_exit(&cachep->c_contentslock); 527 } 528 return (0); 529 } 530 531 for (;;) { 532 /* get (or renew) access to the file system */ 533 if (held) { 534 /* Won't loop with NFSv4 connected behavior */ 535 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 536 cachefs_cd_release(fscp); 537 held = 0; 538 } 539 error = cachefs_cd_access(fscp, connected, 0); 540 if (error) 541 goto out; 542 held = 1; 543 connected = 0; 544 545 /* if not the last close */ 546 if (count > 1) { 547 if (fscp->fs_cdconnected != CFS_CD_CONNECTED) 548 goto out; 549 mutex_enter(&cp->c_statelock); 550 if (cp->c_backvp) { 551 CFS_DPRINT_BACKFS_NFSV4(fscp, 552 ("cachefs_close (nfsv4): cnode %p, " 553 "backvp %p\n", cp, cp->c_backvp)); 554 error = VOP_CLOSE(cp->c_backvp, flag, count, 555 offset, cr); 556 if (CFS_TIMEOUT(fscp, error)) { 557 mutex_exit(&cp->c_statelock); 558 cachefs_cd_release(fscp); 559 held = 0; 560 cachefs_cd_timedout(fscp); 561 continue; 562 } 563 } 564 mutex_exit(&cp->c_statelock); 565 goto out; 566 } 567 568 /* 569 * If the file is an unlinked file, then flush the lookup 570 * cache so that inactive will be called if this is 571 * the last reference. It will invalidate all of the 572 * cached pages, without writing them out. Writing them 573 * out is not required because they will be written to a 574 * file which will be immediately removed. 575 */ 576 if (cp->c_unldvp != NULL) { 577 dnlc_purge_vp(vp); 578 mutex_enter(&cp->c_statelock); 579 error = cp->c_error; 580 cp->c_error = 0; 581 mutex_exit(&cp->c_statelock); 582 /* always call VOP_CLOSE() for back fs vnode */ 583 } 584 585 /* force dirty data to stable storage */ 586 else if ((vp->v_type == VREG) && (flag & FWRITE) && 587 !CFS_ISFS_BACKFS_NFSV4(fscp)) { 588 /* clean the cachefs pages synchronously */ 589 error = cachefs_putpage_common(vp, (offset_t)0, 590 0, 0, cr); 591 if (CFS_TIMEOUT(fscp, error)) { 592 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 593 cachefs_cd_release(fscp); 594 held = 0; 595 cachefs_cd_timedout(fscp); 596 continue; 597 } else { 598 connected = 1; 599 continue; 600 } 601 } 602 603 /* if no space left in cache, wait until connected */ 604 if ((error == ENOSPC) && 605 (fscp->fs_cdconnected != CFS_CD_CONNECTED)) { 606 connected = 1; 607 continue; 608 } 609 610 /* clear the cnode error if putpage worked */ 611 if ((error == 0) && cp->c_error) { 612 mutex_enter(&cp->c_statelock); 613 cp->c_error = 0; 614 mutex_exit(&cp->c_statelock); 615 } 616 617 /* if any other important error */ 618 if (cp->c_error) { 619 /* get rid of the pages */ 620 (void) cachefs_putpage_common(vp, 621 (offset_t)0, 0, B_INVAL | B_FORCE, cr); 622 dnlc_purge_vp(vp); 623 } 624 } 625 626 mutex_enter(&cp->c_statelock); 627 if (cp->c_backvp && 628 (fscp->fs_cdconnected == CFS_CD_CONNECTED)) { 629 error = VOP_CLOSE(cp->c_backvp, flag, close_cnt, 630 offset, cr); 631 if (CFS_TIMEOUT(fscp, error)) { 632 mutex_exit(&cp->c_statelock); 633 cachefs_cd_release(fscp); 634 held = 0; 635 cachefs_cd_timedout(fscp); 636 /* don't decrement the vnode counts again */ 637 close_cnt = 0; 638 continue; 639 } 640 } 641 mutex_exit(&cp->c_statelock); 642 break; 643 } 644 645 mutex_enter(&cp->c_statelock); 646 if (!error) 647 error = cp->c_error; 648 cp->c_error = 0; 649 mutex_exit(&cp->c_statelock); 650 651 out: 652 if (held) 653 cachefs_cd_release(fscp); 654 #ifdef CFS_CD_DEBUG 655 ASSERT((curthread->t_flag & T_CD_HELD) == 0); 656 #endif 657 658 #ifdef CFSDEBUG 659 CFS_DEBUG(CFSDEBUG_VOPS) 660 printf("cachefs_close: EXIT vp %p\n", (void *)vp); 661 #endif 662 return (error); 663 } 664 665 /*ARGSUSED*/ 666 static int 667 cachefs_read(vnode_t *vp, uio_t *uiop, int ioflag, cred_t *cr, 668 caller_context_t *ct) 669 { 670 struct cnode *cp = VTOC(vp); 671 fscache_t *fscp = C_TO_FSCACHE(cp); 672 register u_offset_t off; 673 register int mapoff; 674 register caddr_t base; 675 int n; 676 offset_t diff; 677 uint_t flags = 0; 678 int error = 0; 679 680 #if 0 681 if (vp->v_flag & VNOCACHE) 682 flags = SM_INVAL; 683 #endif 684 if (getzoneid() != GLOBAL_ZONEID) 685 return (EPERM); 686 if (vp->v_type != VREG) 687 return (EISDIR); 688 689 ASSERT(RW_READ_HELD(&cp->c_rwlock)); 690 691 if (uiop->uio_resid == 0) 692 return (0); 693 694 695 if (uiop->uio_loffset < (offset_t)0) 696 return (EINVAL); 697 698 /* 699 * Call backfilesystem to read if NFSv4, the cachefs code 700 * does the read from the back filesystem asynchronously 701 * which is not supported by pass-through functionality. 702 */ 703 if (CFS_ISFS_BACKFS_NFSV4(fscp)) { 704 error = cachefs_read_backfs_nfsv4(vp, uiop, ioflag, cr, ct); 705 goto out; 706 } 707 708 if (MANDLOCK(vp, cp->c_attr.va_mode)) { 709 error = chklock(vp, FREAD, (offset_t)uiop->uio_loffset, 710 uiop->uio_resid, uiop->uio_fmode, ct); 711 if (error) 712 return (error); 713 } 714 715 /* 716 * Sit in a loop and transfer (uiomove) the data in up to 717 * MAXBSIZE chunks. Each chunk is mapped into the kernel's 718 * address space as needed and then released. 719 */ 720 do { 721 /* 722 * off Offset of current MAXBSIZE chunk 723 * mapoff Offset within the current chunk 724 * n Number of bytes to move from this chunk 725 * base kernel address of mapped in chunk 726 */ 727 off = uiop->uio_loffset & (offset_t)MAXBMASK; 728 mapoff = uiop->uio_loffset & MAXBOFFSET; 729 n = MAXBSIZE - mapoff; 730 if (n > uiop->uio_resid) 731 n = (uint_t)uiop->uio_resid; 732 733 /* perform consistency check */ 734 error = cachefs_cd_access(fscp, 0, 0); 735 if (error) 736 break; 737 mutex_enter(&cp->c_statelock); 738 error = CFSOP_CHECK_COBJECT(fscp, cp, 0, cr); 739 diff = cp->c_size - uiop->uio_loffset; 740 mutex_exit(&cp->c_statelock); 741 if (CFS_TIMEOUT(fscp, error)) { 742 cachefs_cd_release(fscp); 743 cachefs_cd_timedout(fscp); 744 error = 0; 745 continue; 746 } 747 cachefs_cd_release(fscp); 748 749 if (error) 750 break; 751 752 if (diff <= (offset_t)0) 753 break; 754 if (diff < (offset_t)n) 755 n = diff; 756 757 base = segmap_getmapflt(segkmap, vp, off, (uint_t)n, 1, S_READ); 758 759 error = segmap_fault(kas.a_hat, segkmap, base, n, 760 F_SOFTLOCK, S_READ); 761 if (error) { 762 (void) segmap_release(segkmap, base, 0); 763 if (FC_CODE(error) == FC_OBJERR) 764 error = FC_ERRNO(error); 765 else 766 error = EIO; 767 break; 768 } 769 error = uiomove(base+mapoff, n, UIO_READ, uiop); 770 (void) segmap_fault(kas.a_hat, segkmap, base, n, 771 F_SOFTUNLOCK, S_READ); 772 if (error == 0) { 773 /* 774 * if we read a whole page(s), or to eof, 775 * we won't need this page(s) again soon. 776 */ 777 if (n + mapoff == MAXBSIZE || 778 uiop->uio_loffset == cp->c_size) 779 flags |= SM_DONTNEED; 780 } 781 (void) segmap_release(segkmap, base, flags); 782 } while (error == 0 && uiop->uio_resid > 0); 783 784 out: 785 #ifdef CFSDEBUG 786 CFS_DEBUG(CFSDEBUG_VOPS) 787 printf("cachefs_read: EXIT error %d resid %ld\n", error, 788 uiop->uio_resid); 789 #endif 790 return (error); 791 } 792 793 /* 794 * cachefs_read_backfs_nfsv4 795 * 796 * Call NFSv4 back filesystem to handle the read (cachefs 797 * pass-through support for NFSv4). 798 */ 799 static int 800 cachefs_read_backfs_nfsv4(vnode_t *vp, uio_t *uiop, int ioflag, cred_t *cr, 801 caller_context_t *ct) 802 { 803 cnode_t *cp = VTOC(vp); 804 fscache_t *fscp = C_TO_FSCACHE(cp); 805 vnode_t *backvp; 806 int error; 807 808 /* 809 * For NFSv4 pass-through to work, only connected operation 810 * is supported, the cnode backvp must exist, and cachefs 811 * optional (eg., disconnectable) flags are turned off. Assert 812 * these conditions for the read operation. 813 */ 814 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 815 CFS_BACKFS_NFSV4_ASSERT_CNODE(cp); 816 817 /* Call backfs vnode op after extracting backvp */ 818 mutex_enter(&cp->c_statelock); 819 backvp = cp->c_backvp; 820 mutex_exit(&cp->c_statelock); 821 822 CFS_DPRINT_BACKFS_NFSV4(fscp, ("cachefs_read_backfs_nfsv4: cnode %p, " 823 "backvp %p\n", cp, backvp)); 824 825 (void) VOP_RWLOCK(backvp, V_WRITELOCK_FALSE, ct); 826 error = VOP_READ(backvp, uiop, ioflag, cr, ct); 827 VOP_RWUNLOCK(backvp, V_WRITELOCK_FALSE, ct); 828 829 /* Increment cache miss counter */ 830 fscp->fs_stats.st_misses++; 831 832 return (error); 833 } 834 835 /*ARGSUSED*/ 836 static int 837 cachefs_write(vnode_t *vp, uio_t *uiop, int ioflag, cred_t *cr, 838 caller_context_t *ct) 839 { 840 struct cnode *cp = VTOC(vp); 841 fscache_t *fscp = C_TO_FSCACHE(cp); 842 int error = 0; 843 u_offset_t off; 844 caddr_t base; 845 uint_t bsize; 846 uint_t flags; 847 int n, on; 848 rlim64_t limit = uiop->uio_llimit; 849 ssize_t resid; 850 offset_t offset; 851 offset_t remainder; 852 853 #ifdef CFSDEBUG 854 CFS_DEBUG(CFSDEBUG_VOPS) 855 printf( 856 "cachefs_write: ENTER vp %p offset %llu count %ld cflags %x\n", 857 (void *)vp, uiop->uio_loffset, uiop->uio_resid, 858 cp->c_flags); 859 #endif 860 if (getzoneid() != GLOBAL_ZONEID) { 861 error = EPERM; 862 goto out; 863 } 864 if (vp->v_type != VREG) { 865 error = EISDIR; 866 goto out; 867 } 868 869 ASSERT(RW_WRITE_HELD(&cp->c_rwlock)); 870 871 if (uiop->uio_resid == 0) { 872 goto out; 873 } 874 875 /* Call backfilesystem to write if NFSv4 */ 876 if (CFS_ISFS_BACKFS_NFSV4(fscp)) { 877 error = cachefs_write_backfs_nfsv4(vp, uiop, ioflag, cr, ct); 878 goto out2; 879 } 880 881 if (MANDLOCK(vp, cp->c_attr.va_mode)) { 882 error = chklock(vp, FWRITE, (offset_t)uiop->uio_loffset, 883 uiop->uio_resid, uiop->uio_fmode, ct); 884 if (error) 885 goto out; 886 } 887 888 if (ioflag & FAPPEND) { 889 for (;;) { 890 /* do consistency check to get correct file size */ 891 error = cachefs_cd_access(fscp, 0, 1); 892 if (error) 893 goto out; 894 mutex_enter(&cp->c_statelock); 895 error = CFSOP_CHECK_COBJECT(fscp, cp, 0, cr); 896 uiop->uio_loffset = cp->c_size; 897 mutex_exit(&cp->c_statelock); 898 if (CFS_TIMEOUT(fscp, error)) { 899 cachefs_cd_release(fscp); 900 cachefs_cd_timedout(fscp); 901 continue; 902 } 903 cachefs_cd_release(fscp); 904 if (error) 905 goto out; 906 break; 907 } 908 } 909 910 if (limit == RLIM64_INFINITY || limit > MAXOFFSET_T) 911 limit = MAXOFFSET_T; 912 913 if (uiop->uio_loffset >= limit) { 914 proc_t *p = ttoproc(curthread); 915 916 mutex_enter(&p->p_lock); 917 (void) rctl_action(rctlproc_legacy[RLIMIT_FSIZE], p->p_rctls, 918 p, RCA_UNSAFE_SIGINFO); 919 mutex_exit(&p->p_lock); 920 error = EFBIG; 921 goto out; 922 } 923 if (uiop->uio_loffset > fscp->fs_offmax) { 924 error = EFBIG; 925 goto out; 926 } 927 928 if (limit > fscp->fs_offmax) 929 limit = fscp->fs_offmax; 930 931 if (uiop->uio_loffset < (offset_t)0) { 932 error = EINVAL; 933 goto out; 934 } 935 936 offset = uiop->uio_loffset + uiop->uio_resid; 937 /* 938 * Check to make sure that the process will not exceed 939 * its limit on file size. It is okay to write up to 940 * the limit, but not beyond. Thus, the write which 941 * reaches the limit will be short and the next write 942 * will return an error. 943 */ 944 remainder = 0; 945 if (offset > limit) { 946 remainder = (int)(offset - (u_offset_t)limit); 947 uiop->uio_resid = limit - uiop->uio_loffset; 948 if (uiop->uio_resid <= 0) { 949 proc_t *p = ttoproc(curthread); 950 951 uiop->uio_resid += remainder; 952 mutex_enter(&p->p_lock); 953 (void) rctl_action(rctlproc_legacy[RLIMIT_FSIZE], 954 p->p_rctls, p, RCA_UNSAFE_SIGINFO); 955 mutex_exit(&p->p_lock); 956 error = EFBIG; 957 goto out; 958 } 959 } 960 961 resid = uiop->uio_resid; 962 offset = uiop->uio_loffset; 963 bsize = vp->v_vfsp->vfs_bsize; 964 965 /* loop around and do the write in MAXBSIZE chunks */ 966 do { 967 /* mapping offset */ 968 off = uiop->uio_loffset & (offset_t)MAXBMASK; 969 on = uiop->uio_loffset & MAXBOFFSET; /* Rel. offset */ 970 n = MAXBSIZE - on; 971 if (n > uiop->uio_resid) 972 n = (int)uiop->uio_resid; 973 base = segmap_getmap(segkmap, vp, off); 974 error = cachefs_writepage(vp, (base + on), n, uiop); 975 if (error == 0) { 976 flags = 0; 977 /* 978 * Have written a whole block.Start an 979 * asynchronous write and mark the buffer to 980 * indicate that it won't be needed again 981 * soon. 982 */ 983 if (n + on == bsize) { 984 flags = SM_WRITE |SM_ASYNC |SM_DONTNEED; 985 } 986 #if 0 987 /* XXX need to understand this */ 988 if ((ioflag & (FSYNC|FDSYNC)) || 989 (cp->c_backvp && vn_has_flocks(cp->c_backvp))) { 990 flags &= ~SM_ASYNC; 991 flags |= SM_WRITE; 992 } 993 #else 994 if (ioflag & (FSYNC|FDSYNC)) { 995 flags &= ~SM_ASYNC; 996 flags |= SM_WRITE; 997 } 998 #endif 999 error = segmap_release(segkmap, base, flags); 1000 } else { 1001 (void) segmap_release(segkmap, base, 0); 1002 } 1003 } while (error == 0 && uiop->uio_resid > 0); 1004 1005 out: 1006 if (error == EINTR && (ioflag & (FSYNC|FDSYNC))) { 1007 uiop->uio_resid = resid; 1008 uiop->uio_loffset = offset; 1009 } else 1010 uiop->uio_resid += remainder; 1011 1012 out2: 1013 #ifdef CFSDEBUG 1014 CFS_DEBUG(CFSDEBUG_VOPS) 1015 printf("cachefs_write: EXIT error %d\n", error); 1016 #endif 1017 return (error); 1018 } 1019 1020 /* 1021 * cachefs_write_backfs_nfsv4 1022 * 1023 * Call NFSv4 back filesystem to handle the write (cachefs 1024 * pass-through support for NFSv4). 1025 */ 1026 static int 1027 cachefs_write_backfs_nfsv4(vnode_t *vp, uio_t *uiop, int ioflag, cred_t *cr, 1028 caller_context_t *ct) 1029 { 1030 cnode_t *cp = VTOC(vp); 1031 fscache_t *fscp = C_TO_FSCACHE(cp); 1032 vnode_t *backvp; 1033 int error; 1034 1035 /* 1036 * For NFSv4 pass-through to work, only connected operation 1037 * is supported, the cnode backvp must exist, and cachefs 1038 * optional (eg., disconnectable) flags are turned off. Assert 1039 * these conditions for the read operation. 1040 */ 1041 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 1042 CFS_BACKFS_NFSV4_ASSERT_CNODE(cp); 1043 1044 /* Call backfs vnode op after extracting the backvp */ 1045 mutex_enter(&cp->c_statelock); 1046 backvp = cp->c_backvp; 1047 mutex_exit(&cp->c_statelock); 1048 1049 CFS_DPRINT_BACKFS_NFSV4(fscp, ("cachefs_write_backfs_nfsv4: cnode %p, " 1050 "backvp %p\n", cp, backvp)); 1051 (void) VOP_RWLOCK(backvp, V_WRITELOCK_TRUE, ct); 1052 error = VOP_WRITE(backvp, uiop, ioflag, cr, ct); 1053 VOP_RWUNLOCK(backvp, V_WRITELOCK_TRUE, ct); 1054 1055 return (error); 1056 } 1057 1058 /* 1059 * see if we've charged ourselves for frontfile data at 1060 * the given offset. If not, allocate a block for it now. 1061 */ 1062 static int 1063 cachefs_charge_page(struct cnode *cp, u_offset_t offset) 1064 { 1065 u_offset_t blockoff; 1066 int error; 1067 int inc; 1068 1069 ASSERT(MUTEX_HELD(&cp->c_statelock)); 1070 ASSERT(PAGESIZE <= MAXBSIZE); 1071 1072 error = 0; 1073 blockoff = offset & (offset_t)MAXBMASK; 1074 1075 /* get the front file if necessary so allocblocks works */ 1076 if ((cp->c_frontvp == NULL) && 1077 ((cp->c_flags & CN_NOCACHE) == 0)) { 1078 (void) cachefs_getfrontfile(cp); 1079 } 1080 if (cp->c_flags & CN_NOCACHE) 1081 return (1); 1082 1083 if (cachefs_check_allocmap(cp, blockoff)) 1084 return (0); 1085 1086 for (inc = PAGESIZE; inc < MAXBSIZE; inc += PAGESIZE) 1087 if (cachefs_check_allocmap(cp, blockoff+inc)) 1088 return (0); 1089 1090 error = cachefs_allocblocks(C_TO_FSCACHE(cp)->fs_cache, 1, 1091 cp->c_metadata.md_rltype); 1092 if (error == 0) { 1093 cp->c_metadata.md_frontblks++; 1094 cp->c_flags |= CN_UPDATED; 1095 } 1096 return (error); 1097 } 1098 1099 /* 1100 * Called only by cachefs_write to write 1 page or less of data. 1101 * base - base address kernel addr space 1102 * tcount - Total bytes to move - < MAXBSIZE 1103 */ 1104 static int 1105 cachefs_writepage(vnode_t *vp, caddr_t base, int tcount, uio_t *uiop) 1106 { 1107 struct cnode *cp = VTOC(vp); 1108 fscache_t *fscp = C_TO_FSCACHE(cp); 1109 register int n; 1110 register u_offset_t offset; 1111 int error = 0, terror; 1112 extern struct as kas; 1113 u_offset_t lastpage_off; 1114 int pagecreate = 0; 1115 int newpage; 1116 1117 #ifdef CFSDEBUG 1118 CFS_DEBUG(CFSDEBUG_VOPS) 1119 printf( 1120 "cachefs_writepage: ENTER vp %p offset %llu len %ld\\\n", 1121 (void *)vp, uiop->uio_loffset, uiop->uio_resid); 1122 #endif 1123 1124 /* 1125 * Move bytes in PAGESIZE chunks. We must avoid spanning pages in 1126 * uiomove() because page faults may cause the cache to be invalidated 1127 * out from under us. 1128 */ 1129 do { 1130 offset = uiop->uio_loffset; 1131 lastpage_off = (cp->c_size - 1) & (offset_t)PAGEMASK; 1132 1133 /* 1134 * If not connected then need to make sure we have space 1135 * to perform the write. We could make this check 1136 * a little tighter by only doing it if we are growing the file. 1137 */ 1138 if (fscp->fs_cdconnected != CFS_CD_CONNECTED) { 1139 error = cachefs_allocblocks(fscp->fs_cache, 1, 1140 cp->c_metadata.md_rltype); 1141 if (error) 1142 break; 1143 cachefs_freeblocks(fscp->fs_cache, 1, 1144 cp->c_metadata.md_rltype); 1145 } 1146 1147 /* 1148 * n is the number of bytes required to satisfy the request 1149 * or the number of bytes to fill out the page. 1150 */ 1151 n = (int)(PAGESIZE - ((uintptr_t)base & PAGEOFFSET)); 1152 if (n > tcount) 1153 n = tcount; 1154 1155 /* 1156 * The number of bytes of data in the last page can not 1157 * be accurately be determined while page is being 1158 * uiomove'd to and the size of the file being updated. 1159 * Thus, inform threads which need to know accurately 1160 * how much data is in the last page of the file. They 1161 * will not do the i/o immediately, but will arrange for 1162 * the i/o to happen later when this modify operation 1163 * will have finished. 1164 * 1165 * in similar NFS code, this is done right before the 1166 * uiomove(), which is best. but here in cachefs, we 1167 * have two uiomove()s, so we must do it here. 1168 */ 1169 ASSERT(!(cp->c_flags & CN_CMODINPROG)); 1170 mutex_enter(&cp->c_statelock); 1171 cp->c_flags |= CN_CMODINPROG; 1172 cp->c_modaddr = (offset & (offset_t)MAXBMASK); 1173 mutex_exit(&cp->c_statelock); 1174 1175 /* 1176 * Check to see if we can skip reading in the page 1177 * and just allocate the memory. We can do this 1178 * if we are going to rewrite the entire mapping 1179 * or if we are going to write to or beyond the current 1180 * end of file from the beginning of the mapping. 1181 */ 1182 if ((offset > (lastpage_off + PAGEOFFSET)) || 1183 ((cp->c_size == 0) && (offset < PAGESIZE)) || 1184 ((uintptr_t)base & PAGEOFFSET) == 0 && (n == PAGESIZE || 1185 ((offset + n) >= cp->c_size))) { 1186 pagecreate = 1; 1187 1188 /* 1189 * segmap_pagecreate() returns 1 if it calls 1190 * page_create_va() to allocate any pages. 1191 */ 1192 newpage = segmap_pagecreate(segkmap, 1193 (caddr_t)((uintptr_t)base & (uintptr_t)PAGEMASK), 1194 PAGESIZE, 0); 1195 /* do not zero page if we are overwriting all of it */ 1196 if (!((((uintptr_t)base & PAGEOFFSET) == 0) && 1197 (n == PAGESIZE))) { 1198 (void) kzero((void *) 1199 ((uintptr_t)base & (uintptr_t)PAGEMASK), 1200 PAGESIZE); 1201 } 1202 error = uiomove(base, n, UIO_WRITE, uiop); 1203 1204 /* 1205 * Unlock the page allocated by page_create_va() 1206 * in segmap_pagecreate() 1207 */ 1208 if (newpage) 1209 segmap_pageunlock(segkmap, 1210 (caddr_t)((uintptr_t)base & 1211 (uintptr_t)PAGEMASK), 1212 PAGESIZE, S_WRITE); 1213 } else { 1214 /* 1215 * KLUDGE ! Use segmap_fault instead of faulting and 1216 * using as_fault() to avoid a recursive readers lock 1217 * on kas. 1218 */ 1219 error = segmap_fault(kas.a_hat, segkmap, (caddr_t) 1220 ((uintptr_t)base & (uintptr_t)PAGEMASK), 1221 PAGESIZE, F_SOFTLOCK, S_WRITE); 1222 if (error) { 1223 if (FC_CODE(error) == FC_OBJERR) 1224 error = FC_ERRNO(error); 1225 else 1226 error = EIO; 1227 break; 1228 } 1229 error = uiomove(base, n, UIO_WRITE, uiop); 1230 (void) segmap_fault(kas.a_hat, segkmap, (caddr_t) 1231 ((uintptr_t)base & (uintptr_t)PAGEMASK), 1232 PAGESIZE, F_SOFTUNLOCK, S_WRITE); 1233 } 1234 n = (int)(uiop->uio_loffset - offset); /* n = # bytes written */ 1235 base += n; 1236 tcount -= n; 1237 1238 /* get access to the file system */ 1239 if ((terror = cachefs_cd_access(fscp, 0, 1)) != 0) { 1240 error = terror; 1241 break; 1242 } 1243 1244 /* 1245 * cp->c_attr.va_size is the maximum number of 1246 * bytes known to be in the file. 1247 * Make sure it is at least as high as the 1248 * last byte we just wrote into the buffer. 1249 */ 1250 mutex_enter(&cp->c_statelock); 1251 if (cp->c_size < uiop->uio_loffset) { 1252 cp->c_size = uiop->uio_loffset; 1253 } 1254 if (cp->c_size != cp->c_attr.va_size) { 1255 cp->c_attr.va_size = cp->c_size; 1256 cp->c_flags |= CN_UPDATED; 1257 } 1258 /* c_size is now correct, so we can clear modinprog */ 1259 cp->c_flags &= ~CN_CMODINPROG; 1260 if (error == 0) { 1261 cp->c_flags |= CDIRTY; 1262 if (pagecreate && (cp->c_flags & CN_NOCACHE) == 0) { 1263 /* 1264 * if we're not in NOCACHE mode 1265 * (i.e., single-writer), we update the 1266 * allocmap here rather than waiting until 1267 * cachefspush is called. This prevents 1268 * getpage from clustering up pages from 1269 * the backfile and stomping over the changes 1270 * we make here. 1271 */ 1272 if (cachefs_charge_page(cp, offset) == 0) { 1273 cachefs_update_allocmap(cp, 1274 offset & (offset_t)PAGEMASK, 1275 (size_t)PAGESIZE); 1276 } 1277 1278 /* else we ran out of space */ 1279 else { 1280 /* nocache file if connected */ 1281 if (fscp->fs_cdconnected == 1282 CFS_CD_CONNECTED) 1283 cachefs_nocache(cp); 1284 /* 1285 * If disconnected then cannot 1286 * nocache the file. Let it have 1287 * the space. 1288 */ 1289 else { 1290 cp->c_metadata.md_frontblks++; 1291 cp->c_flags |= CN_UPDATED; 1292 cachefs_update_allocmap(cp, 1293 offset & (offset_t)PAGEMASK, 1294 (size_t)PAGESIZE); 1295 } 1296 } 1297 } 1298 } 1299 mutex_exit(&cp->c_statelock); 1300 cachefs_cd_release(fscp); 1301 } while (tcount > 0 && error == 0); 1302 1303 if (cp->c_flags & CN_CMODINPROG) { 1304 /* XXX assert error != 0? FC_ERRNO() makes this more risky. */ 1305 mutex_enter(&cp->c_statelock); 1306 cp->c_flags &= ~CN_CMODINPROG; 1307 mutex_exit(&cp->c_statelock); 1308 } 1309 1310 #ifdef CFS_CD_DEBUG 1311 ASSERT((curthread->t_flag & T_CD_HELD) == 0); 1312 #endif 1313 1314 #ifdef CFSDEBUG 1315 CFS_DEBUG(CFSDEBUG_VOPS) 1316 printf("cachefs_writepage: EXIT error %d\n", error); 1317 #endif 1318 1319 return (error); 1320 } 1321 1322 /* 1323 * Pushes out pages to the back and/or front file system. 1324 */ 1325 static int 1326 cachefs_push(vnode_t *vp, page_t *pp, u_offset_t *offp, size_t *lenp, 1327 int flags, cred_t *cr) 1328 { 1329 struct cnode *cp = VTOC(vp); 1330 struct buf *bp; 1331 int error; 1332 fscache_t *fscp = C_TO_FSCACHE(cp); 1333 u_offset_t iooff; 1334 size_t iolen; 1335 u_offset_t lbn; 1336 u_offset_t lbn_off; 1337 uint_t bsize; 1338 1339 ASSERT((flags & B_ASYNC) == 0); 1340 ASSERT(!vn_is_readonly(vp)); 1341 ASSERT(pp != NULL); 1342 ASSERT(cr != NULL); 1343 1344 bsize = MAX(vp->v_vfsp->vfs_bsize, PAGESIZE); 1345 lbn = pp->p_offset / bsize; 1346 lbn_off = lbn * bsize; 1347 1348 /* 1349 * Find a kluster that fits in one block, or in 1350 * one page if pages are bigger than blocks. If 1351 * there is less file space allocated than a whole 1352 * page, we'll shorten the i/o request below. 1353 */ 1354 1355 pp = pvn_write_kluster(vp, pp, &iooff, &iolen, lbn_off, 1356 roundup(bsize, PAGESIZE), flags); 1357 1358 /* 1359 * The CN_CMODINPROG flag makes sure that we use a correct 1360 * value of c_size, below. CN_CMODINPROG is set in 1361 * cachefs_writepage(). When CN_CMODINPROG is set it 1362 * indicates that a uiomove() is in progress and the c_size 1363 * has not been made consistent with the new size of the 1364 * file. When the uiomove() completes the c_size is updated 1365 * and the CN_CMODINPROG flag is cleared. 1366 * 1367 * The CN_CMODINPROG flag makes sure that cachefs_push_front 1368 * and cachefs_push_connected see a consistent value of 1369 * c_size. Without this handshaking, it is possible that 1370 * these routines will pick up the old value of c_size before 1371 * the uiomove() in cachefs_writepage() completes. This will 1372 * result in the vn_rdwr() being too small, and data loss. 1373 * 1374 * More precisely, there is a window between the time the 1375 * uiomove() completes and the time the c_size is updated. If 1376 * a VOP_PUTPAGE() operation intervenes in this window, the 1377 * page will be picked up, because it is dirty; it will be 1378 * unlocked, unless it was pagecreate'd. When the page is 1379 * picked up as dirty, the dirty bit is reset 1380 * (pvn_getdirty()). In cachefs_push_connected(), c_size is 1381 * checked. This will still be the old size. Therefore, the 1382 * page will not be written out to the correct length, and the 1383 * page will be clean, so the data may disappear. 1384 */ 1385 if (cp->c_flags & CN_CMODINPROG) { 1386 mutex_enter(&cp->c_statelock); 1387 if ((cp->c_flags & CN_CMODINPROG) && 1388 cp->c_modaddr + MAXBSIZE > iooff && 1389 cp->c_modaddr < iooff + iolen) { 1390 page_t *plist; 1391 1392 /* 1393 * A write is in progress for this region of 1394 * the file. If we did not detect 1395 * CN_CMODINPROG here then this path through 1396 * cachefs_push_connected() would eventually 1397 * do the vn_rdwr() and may not write out all 1398 * of the data in the pages. We end up losing 1399 * data. So we decide to set the modified bit 1400 * on each page in the page list and mark the 1401 * cnode with CDIRTY. This push will be 1402 * restarted at some later time. 1403 */ 1404 1405 plist = pp; 1406 while (plist != NULL) { 1407 pp = plist; 1408 page_sub(&plist, pp); 1409 hat_setmod(pp); 1410 page_io_unlock(pp); 1411 page_unlock(pp); 1412 } 1413 cp->c_flags |= CDIRTY; 1414 mutex_exit(&cp->c_statelock); 1415 if (offp) 1416 *offp = iooff; 1417 if (lenp) 1418 *lenp = iolen; 1419 return (0); 1420 } 1421 mutex_exit(&cp->c_statelock); 1422 } 1423 1424 /* 1425 * Set the pages up for pageout. 1426 */ 1427 bp = pageio_setup(pp, iolen, CTOV(cp), B_WRITE | flags); 1428 if (bp == NULL) { 1429 1430 /* 1431 * currently, there is no way for pageio_setup() to 1432 * return NULL, since it uses its own scheme for 1433 * kmem_alloc()ing that shouldn't return NULL, and 1434 * since pageio_setup() itself dereferences the thing 1435 * it's about to return. still, we need to be ready 1436 * in case this ever does start happening. 1437 */ 1438 1439 error = ENOMEM; 1440 goto writedone; 1441 } 1442 /* 1443 * pageio_setup should have set b_addr to 0. This 1444 * is correct since we want to do I/O on a page 1445 * boundary. bp_mapin will use this addr to calculate 1446 * an offset, and then set b_addr to the kernel virtual 1447 * address it allocated for us. 1448 */ 1449 bp->b_edev = 0; 1450 bp->b_dev = 0; 1451 bp->b_lblkno = (diskaddr_t)lbtodb(iooff); 1452 bp_mapin(bp); 1453 1454 iolen = cp->c_size - ldbtob(bp->b_blkno); 1455 if (iolen > bp->b_bcount) 1456 iolen = bp->b_bcount; 1457 1458 /* if connected */ 1459 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 1460 /* write to the back file first */ 1461 error = cachefs_push_connected(vp, bp, iolen, iooff, cr); 1462 1463 /* write to the front file if allowed */ 1464 if ((error == 0) && CFS_ISFS_NONSHARED(fscp) && 1465 ((cp->c_flags & CN_NOCACHE) == 0)) { 1466 /* try to write to the front file */ 1467 (void) cachefs_push_front(vp, bp, iolen, iooff, cr); 1468 } 1469 } 1470 1471 /* else if disconnected */ 1472 else { 1473 /* try to write to the front file */ 1474 error = cachefs_push_front(vp, bp, iolen, iooff, cr); 1475 } 1476 1477 bp_mapout(bp); 1478 pageio_done(bp); 1479 1480 writedone: 1481 1482 pvn_write_done(pp, ((error) ? B_ERROR : 0) | B_WRITE | flags); 1483 if (offp) 1484 *offp = iooff; 1485 if (lenp) 1486 *lenp = iolen; 1487 1488 /* XXX ask bob mastors how to fix this someday */ 1489 mutex_enter(&cp->c_statelock); 1490 if (error) { 1491 if (error == ENOSPC) { 1492 if ((fscp->fs_cdconnected == CFS_CD_CONNECTED) || 1493 CFS_ISFS_SOFT(fscp)) { 1494 CFSOP_INVALIDATE_COBJECT(fscp, cp, cr); 1495 cp->c_error = error; 1496 } 1497 } else if ((CFS_TIMEOUT(fscp, error) == 0) && 1498 (error != EINTR)) { 1499 CFSOP_INVALIDATE_COBJECT(fscp, cp, cr); 1500 cp->c_error = error; 1501 } 1502 } else if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 1503 CFSOP_MODIFY_COBJECT(fscp, cp, cr); 1504 } 1505 mutex_exit(&cp->c_statelock); 1506 1507 return (error); 1508 } 1509 1510 /* 1511 * Pushes out pages to the back file system. 1512 */ 1513 static int 1514 cachefs_push_connected(vnode_t *vp, struct buf *bp, size_t iolen, 1515 u_offset_t iooff, cred_t *cr) 1516 { 1517 struct cnode *cp = VTOC(vp); 1518 int error = 0; 1519 int mode = 0; 1520 fscache_t *fscp = C_TO_FSCACHE(cp); 1521 ssize_t resid; 1522 vnode_t *backvp; 1523 1524 /* get the back file if necessary */ 1525 mutex_enter(&cp->c_statelock); 1526 if (cp->c_backvp == NULL) { 1527 error = cachefs_getbackvp(fscp, cp); 1528 if (error) { 1529 mutex_exit(&cp->c_statelock); 1530 goto out; 1531 } 1532 } 1533 backvp = cp->c_backvp; 1534 VN_HOLD(backvp); 1535 mutex_exit(&cp->c_statelock); 1536 1537 if (CFS_ISFS_NONSHARED(fscp) && CFS_ISFS_SNR(fscp)) 1538 mode = FSYNC; 1539 1540 /* write to the back file */ 1541 error = bp->b_error = vn_rdwr(UIO_WRITE, backvp, bp->b_un.b_addr, 1542 iolen, iooff, UIO_SYSSPACE, mode, 1543 RLIM64_INFINITY, cr, &resid); 1544 if (error) { 1545 #ifdef CFSDEBUG 1546 CFS_DEBUG(CFSDEBUG_VOPS | CFSDEBUG_BACK) 1547 printf("cachefspush: error %d cr %p\n", 1548 error, (void *)cr); 1549 #endif 1550 bp->b_flags |= B_ERROR; 1551 } 1552 VN_RELE(backvp); 1553 out: 1554 return (error); 1555 } 1556 1557 /* 1558 * Pushes out pages to the front file system. 1559 * Called for both connected and disconnected states. 1560 */ 1561 static int 1562 cachefs_push_front(vnode_t *vp, struct buf *bp, size_t iolen, 1563 u_offset_t iooff, cred_t *cr) 1564 { 1565 struct cnode *cp = VTOC(vp); 1566 fscache_t *fscp = C_TO_FSCACHE(cp); 1567 int error = 0; 1568 ssize_t resid; 1569 u_offset_t popoff; 1570 off_t commit = 0; 1571 uint_t seq; 1572 enum cachefs_rl_type type; 1573 vnode_t *frontvp = NULL; 1574 1575 mutex_enter(&cp->c_statelock); 1576 1577 if (!CFS_ISFS_NONSHARED(fscp)) { 1578 error = ETIMEDOUT; 1579 goto out; 1580 } 1581 1582 /* get the front file if necessary */ 1583 if ((cp->c_frontvp == NULL) && 1584 ((cp->c_flags & CN_NOCACHE) == 0)) { 1585 (void) cachefs_getfrontfile(cp); 1586 } 1587 if (cp->c_flags & CN_NOCACHE) { 1588 error = ETIMEDOUT; 1589 goto out; 1590 } 1591 1592 /* if disconnected, needs to be populated and have good attributes */ 1593 if ((fscp->fs_cdconnected != CFS_CD_CONNECTED) && 1594 (((cp->c_metadata.md_flags & MD_POPULATED) == 0) || 1595 (cp->c_metadata.md_flags & MD_NEEDATTRS))) { 1596 error = ETIMEDOUT; 1597 goto out; 1598 } 1599 1600 for (popoff = iooff; popoff < (iooff + iolen); popoff += MAXBSIZE) { 1601 if (cachefs_charge_page(cp, popoff)) { 1602 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 1603 cachefs_nocache(cp); 1604 goto out; 1605 } else { 1606 error = ENOSPC; 1607 goto out; 1608 } 1609 } 1610 } 1611 1612 if (fscp->fs_cdconnected != CFS_CD_CONNECTED) { 1613 /* log the first putpage to a file */ 1614 if ((cp->c_metadata.md_flags & MD_PUTPAGE) == 0) { 1615 /* uses open's creds if we have them */ 1616 if (cp->c_cred) 1617 cr = cp->c_cred; 1618 1619 if ((cp->c_metadata.md_flags & MD_MAPPING) == 0) { 1620 error = cachefs_dlog_cidmap(fscp); 1621 if (error) { 1622 error = ENOSPC; 1623 goto out; 1624 } 1625 cp->c_metadata.md_flags |= MD_MAPPING; 1626 } 1627 1628 commit = cachefs_dlog_modify(fscp, cp, cr, &seq); 1629 if (commit == 0) { 1630 /* out of space */ 1631 error = ENOSPC; 1632 goto out; 1633 } 1634 1635 cp->c_metadata.md_seq = seq; 1636 type = cp->c_metadata.md_rltype; 1637 cachefs_modified(cp); 1638 cp->c_metadata.md_flags |= MD_PUTPAGE; 1639 cp->c_metadata.md_flags &= ~MD_PUSHDONE; 1640 cp->c_flags |= CN_UPDATED; 1641 } 1642 1643 /* subsequent putpages just get a new sequence number */ 1644 else { 1645 /* but only if it matters */ 1646 if (cp->c_metadata.md_seq != fscp->fs_dlogseq) { 1647 seq = cachefs_dlog_seqnext(fscp); 1648 if (seq == 0) { 1649 error = ENOSPC; 1650 goto out; 1651 } 1652 cp->c_metadata.md_seq = seq; 1653 cp->c_flags |= CN_UPDATED; 1654 /* XXX maybe should do write_metadata here */ 1655 } 1656 } 1657 } 1658 1659 frontvp = cp->c_frontvp; 1660 VN_HOLD(frontvp); 1661 mutex_exit(&cp->c_statelock); 1662 error = bp->b_error = vn_rdwr(UIO_WRITE, frontvp, 1663 bp->b_un.b_addr, iolen, iooff, UIO_SYSSPACE, 0, 1664 RLIM64_INFINITY, kcred, &resid); 1665 mutex_enter(&cp->c_statelock); 1666 VN_RELE(frontvp); 1667 frontvp = NULL; 1668 if (error) { 1669 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 1670 cachefs_nocache(cp); 1671 error = 0; 1672 goto out; 1673 } else { 1674 goto out; 1675 } 1676 } 1677 1678 (void) cachefs_update_allocmap(cp, iooff, iolen); 1679 cp->c_flags |= (CN_UPDATED | CN_NEED_FRONT_SYNC | 1680 CN_POPULATION_PENDING); 1681 if (fscp->fs_cdconnected != CFS_CD_CONNECTED) { 1682 gethrestime(&cp->c_metadata.md_localmtime); 1683 cp->c_metadata.md_flags |= MD_LOCALMTIME; 1684 } 1685 1686 out: 1687 if (commit) { 1688 /* commit the log record */ 1689 ASSERT(fscp->fs_cdconnected == CFS_CD_DISCONNECTED); 1690 if (cachefs_dlog_commit(fscp, commit, error)) { 1691 /*EMPTY*/ 1692 /* XXX fix on panic */ 1693 } 1694 } 1695 1696 if (error && commit) { 1697 cp->c_metadata.md_flags &= ~MD_PUTPAGE; 1698 cachefs_rlent_moveto(fscp->fs_cache, type, 1699 cp->c_metadata.md_rlno, cp->c_metadata.md_frontblks); 1700 cp->c_metadata.md_rltype = type; 1701 cp->c_flags |= CN_UPDATED; 1702 } 1703 mutex_exit(&cp->c_statelock); 1704 return (error); 1705 } 1706 1707 /*ARGSUSED*/ 1708 static int 1709 cachefs_dump(struct vnode *vp, caddr_t foo1, int foo2, int foo3) 1710 { 1711 return (ENOSYS); /* should we panic if we get here? */ 1712 } 1713 1714 /*ARGSUSED*/ 1715 static int 1716 cachefs_ioctl(struct vnode *vp, int cmd, intptr_t arg, int flag, cred_t *cred, 1717 int *rvalp) 1718 { 1719 int error; 1720 struct cnode *cp = VTOC(vp); 1721 struct fscache *fscp = C_TO_FSCACHE(cp); 1722 struct cachefscache *cachep; 1723 extern kmutex_t cachefs_cachelock; 1724 extern cachefscache_t *cachefs_cachelist; 1725 cachefsio_pack_t *packp; 1726 STRUCT_DECL(cachefsio_dcmd, dcmd); 1727 int inlen, outlen; /* LP64: generic int for struct in/out len */ 1728 void *dinp, *doutp; 1729 int (*dcmd_routine)(vnode_t *, void *, void *); 1730 1731 if (getzoneid() != GLOBAL_ZONEID) 1732 return (EPERM); 1733 1734 /* 1735 * Cachefs only provides pass-through support for NFSv4, 1736 * and all vnode operations are passed through to the 1737 * back file system. For NFSv4 pass-through to work, only 1738 * connected operation is supported, the cnode backvp must 1739 * exist, and cachefs optional (eg., disconnectable) flags 1740 * are turned off. Assert these conditions which ensure 1741 * that only a subset of the ioctls are "truly supported" 1742 * for NFSv4 (these are CFSDCMD_DAEMONID and CFSDCMD_GETSTATS. 1743 * The packing operations are meaningless since there is 1744 * no caching for NFSv4, and the called functions silently 1745 * return if the backfilesystem is NFSv4. The daemon 1746 * commands except for those above are essentially used 1747 * for disconnectable operation support (including log 1748 * rolling), so in each called function, we assert that 1749 * NFSv4 is not in use. The _FIO* calls (except _FIOCOD) 1750 * are from "cfsfstype" which is not a documented 1751 * command. However, the command is visible in 1752 * /usr/lib/fs/cachefs so the commands are simply let 1753 * through (don't seem to impact pass-through functionality). 1754 */ 1755 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 1756 CFS_BACKFS_NFSV4_ASSERT_CNODE(cp); 1757 1758 switch (cmd) { 1759 case CACHEFSIO_PACK: 1760 packp = cachefs_kmem_alloc(sizeof (cachefsio_pack_t), KM_SLEEP); 1761 error = xcopyin((void *)arg, packp, sizeof (cachefsio_pack_t)); 1762 if (!error) 1763 error = cachefs_pack(vp, packp->p_name, cred); 1764 cachefs_kmem_free(packp, sizeof (cachefsio_pack_t)); 1765 break; 1766 1767 case CACHEFSIO_UNPACK: 1768 packp = cachefs_kmem_alloc(sizeof (cachefsio_pack_t), KM_SLEEP); 1769 error = xcopyin((void *)arg, packp, sizeof (cachefsio_pack_t)); 1770 if (!error) 1771 error = cachefs_unpack(vp, packp->p_name, cred); 1772 cachefs_kmem_free(packp, sizeof (cachefsio_pack_t)); 1773 break; 1774 1775 case CACHEFSIO_PACKINFO: 1776 packp = cachefs_kmem_alloc(sizeof (cachefsio_pack_t), KM_SLEEP); 1777 error = xcopyin((void *)arg, packp, sizeof (cachefsio_pack_t)); 1778 if (!error) 1779 error = cachefs_packinfo(vp, packp->p_name, 1780 &packp->p_status, cred); 1781 if (!error) 1782 error = xcopyout(packp, (void *)arg, 1783 sizeof (cachefsio_pack_t)); 1784 cachefs_kmem_free(packp, sizeof (cachefsio_pack_t)); 1785 break; 1786 1787 case CACHEFSIO_UNPACKALL: 1788 error = cachefs_unpackall(vp); 1789 break; 1790 1791 case CACHEFSIO_DCMD: 1792 /* 1793 * This is a private interface between the cachefsd and 1794 * this file system. 1795 */ 1796 1797 /* must be root to use these commands */ 1798 if (secpolicy_fs_config(cred, vp->v_vfsp) != 0) 1799 return (EPERM); 1800 1801 /* get the command packet */ 1802 STRUCT_INIT(dcmd, flag & DATAMODEL_MASK); 1803 error = xcopyin((void *)arg, STRUCT_BUF(dcmd), 1804 SIZEOF_STRUCT(cachefsio_dcmd, DATAMODEL_NATIVE)); 1805 if (error) 1806 return (error); 1807 1808 /* copy in the data for the operation */ 1809 dinp = NULL; 1810 if ((inlen = STRUCT_FGET(dcmd, d_slen)) > 0) { 1811 dinp = cachefs_kmem_alloc(inlen, KM_SLEEP); 1812 error = xcopyin(STRUCT_FGETP(dcmd, d_sdata), dinp, 1813 inlen); 1814 if (error) 1815 return (error); 1816 } 1817 1818 /* allocate space for the result */ 1819 doutp = NULL; 1820 if ((outlen = STRUCT_FGET(dcmd, d_rlen)) > 0) 1821 doutp = cachefs_kmem_alloc(outlen, KM_SLEEP); 1822 1823 /* 1824 * Assert NFSv4 only allows the daemonid and getstats 1825 * daemon requests 1826 */ 1827 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0 || 1828 STRUCT_FGET(dcmd, d_cmd) == CFSDCMD_DAEMONID || 1829 STRUCT_FGET(dcmd, d_cmd) == CFSDCMD_GETSTATS); 1830 1831 /* get the routine to execute */ 1832 dcmd_routine = NULL; 1833 switch (STRUCT_FGET(dcmd, d_cmd)) { 1834 case CFSDCMD_DAEMONID: 1835 dcmd_routine = cachefs_io_daemonid; 1836 break; 1837 case CFSDCMD_STATEGET: 1838 dcmd_routine = cachefs_io_stateget; 1839 break; 1840 case CFSDCMD_STATESET: 1841 dcmd_routine = cachefs_io_stateset; 1842 break; 1843 case CFSDCMD_XWAIT: 1844 dcmd_routine = cachefs_io_xwait; 1845 break; 1846 case CFSDCMD_EXISTS: 1847 dcmd_routine = cachefs_io_exists; 1848 break; 1849 case CFSDCMD_LOSTFOUND: 1850 dcmd_routine = cachefs_io_lostfound; 1851 break; 1852 case CFSDCMD_GETINFO: 1853 dcmd_routine = cachefs_io_getinfo; 1854 break; 1855 case CFSDCMD_CIDTOFID: 1856 dcmd_routine = cachefs_io_cidtofid; 1857 break; 1858 case CFSDCMD_GETATTRFID: 1859 dcmd_routine = cachefs_io_getattrfid; 1860 break; 1861 case CFSDCMD_GETATTRNAME: 1862 dcmd_routine = cachefs_io_getattrname; 1863 break; 1864 case CFSDCMD_GETSTATS: 1865 dcmd_routine = cachefs_io_getstats; 1866 break; 1867 case CFSDCMD_ROOTFID: 1868 dcmd_routine = cachefs_io_rootfid; 1869 break; 1870 case CFSDCMD_CREATE: 1871 dcmd_routine = cachefs_io_create; 1872 break; 1873 case CFSDCMD_REMOVE: 1874 dcmd_routine = cachefs_io_remove; 1875 break; 1876 case CFSDCMD_LINK: 1877 dcmd_routine = cachefs_io_link; 1878 break; 1879 case CFSDCMD_RENAME: 1880 dcmd_routine = cachefs_io_rename; 1881 break; 1882 case CFSDCMD_MKDIR: 1883 dcmd_routine = cachefs_io_mkdir; 1884 break; 1885 case CFSDCMD_RMDIR: 1886 dcmd_routine = cachefs_io_rmdir; 1887 break; 1888 case CFSDCMD_SYMLINK: 1889 dcmd_routine = cachefs_io_symlink; 1890 break; 1891 case CFSDCMD_SETATTR: 1892 dcmd_routine = cachefs_io_setattr; 1893 break; 1894 case CFSDCMD_SETSECATTR: 1895 dcmd_routine = cachefs_io_setsecattr; 1896 break; 1897 case CFSDCMD_PUSHBACK: 1898 dcmd_routine = cachefs_io_pushback; 1899 break; 1900 default: 1901 error = ENOTTY; 1902 break; 1903 } 1904 1905 /* execute the routine */ 1906 if (dcmd_routine) 1907 error = (*dcmd_routine)(vp, dinp, doutp); 1908 1909 /* copy out the result */ 1910 if ((error == 0) && doutp) 1911 error = xcopyout(doutp, STRUCT_FGETP(dcmd, d_rdata), 1912 outlen); 1913 1914 /* free allocated memory */ 1915 if (dinp) 1916 cachefs_kmem_free(dinp, inlen); 1917 if (doutp) 1918 cachefs_kmem_free(doutp, outlen); 1919 1920 break; 1921 1922 case _FIOCOD: 1923 if (secpolicy_fs_config(cred, vp->v_vfsp) != 0) { 1924 error = EPERM; 1925 break; 1926 } 1927 1928 error = EBUSY; 1929 if (arg) { 1930 /* non-zero arg means do all filesystems */ 1931 mutex_enter(&cachefs_cachelock); 1932 for (cachep = cachefs_cachelist; cachep != NULL; 1933 cachep = cachep->c_next) { 1934 mutex_enter(&cachep->c_fslistlock); 1935 for (fscp = cachep->c_fslist; 1936 fscp != NULL; 1937 fscp = fscp->fs_next) { 1938 if (CFS_ISFS_CODCONST(fscp)) { 1939 gethrestime(&fscp->fs_cod_time); 1940 error = 0; 1941 } 1942 } 1943 mutex_exit(&cachep->c_fslistlock); 1944 } 1945 mutex_exit(&cachefs_cachelock); 1946 } else { 1947 if (CFS_ISFS_CODCONST(fscp)) { 1948 gethrestime(&fscp->fs_cod_time); 1949 error = 0; 1950 } 1951 } 1952 break; 1953 1954 case _FIOSTOPCACHE: 1955 error = cachefs_stop_cache(cp); 1956 break; 1957 1958 default: 1959 error = ENOTTY; 1960 break; 1961 } 1962 1963 /* return the result */ 1964 return (error); 1965 } 1966 1967 ino64_t 1968 cachefs_fileno_conflict(fscache_t *fscp, ino64_t old) 1969 { 1970 ino64_t new; 1971 1972 ASSERT(MUTEX_HELD(&fscp->fs_fslock)); 1973 1974 for (;;) { 1975 fscp->fs_info.fi_localfileno++; 1976 if (fscp->fs_info.fi_localfileno == 0) 1977 fscp->fs_info.fi_localfileno = 3; 1978 fscp->fs_flags |= CFS_FS_DIRTYINFO; 1979 1980 new = fscp->fs_info.fi_localfileno; 1981 if (! cachefs_fileno_inuse(fscp, new)) 1982 break; 1983 } 1984 1985 cachefs_inum_register(fscp, old, new); 1986 cachefs_inum_register(fscp, new, 0); 1987 return (new); 1988 } 1989 1990 /*ARGSUSED*/ 1991 static int 1992 cachefs_getattr(vnode_t *vp, vattr_t *vap, int flags, cred_t *cr) 1993 { 1994 struct cnode *cp = VTOC(vp); 1995 fscache_t *fscp = C_TO_FSCACHE(cp); 1996 int error = 0; 1997 int held = 0; 1998 int connected = 0; 1999 2000 #ifdef CFSDEBUG 2001 CFS_DEBUG(CFSDEBUG_VOPS) 2002 printf("cachefs_getattr: ENTER vp %p\n", (void *)vp); 2003 #endif 2004 2005 if (getzoneid() != GLOBAL_ZONEID) 2006 return (EPERM); 2007 2008 /* Call backfilesystem getattr if NFSv4 */ 2009 if (CFS_ISFS_BACKFS_NFSV4(fscp)) { 2010 error = cachefs_getattr_backfs_nfsv4(vp, vap, flags, cr); 2011 goto out; 2012 } 2013 2014 /* 2015 * If it has been specified that the return value will 2016 * just be used as a hint, and we are only being asked 2017 * for size, fsid or rdevid, then return the client's 2018 * notion of these values without checking to make sure 2019 * that the attribute cache is up to date. 2020 * The whole point is to avoid an over the wire GETATTR 2021 * call. 2022 */ 2023 if (flags & ATTR_HINT) { 2024 if (vap->va_mask == 2025 (vap->va_mask & (AT_SIZE | AT_FSID | AT_RDEV))) { 2026 if (vap->va_mask | AT_SIZE) 2027 vap->va_size = cp->c_size; 2028 /* 2029 * Return the FSID of the cachefs filesystem, 2030 * not the back filesystem 2031 */ 2032 if (vap->va_mask | AT_FSID) 2033 vap->va_fsid = vp->v_vfsp->vfs_dev; 2034 if (vap->va_mask | AT_RDEV) 2035 vap->va_rdev = cp->c_attr.va_rdev; 2036 return (0); 2037 } 2038 } 2039 2040 /* 2041 * Only need to flush pages if asking for the mtime 2042 * and if there any dirty pages. 2043 */ 2044 if (vap->va_mask & AT_MTIME) { 2045 /*EMPTY*/ 2046 #if 0 2047 /* 2048 * XXX bob: stolen from nfs code, need to do something similar 2049 */ 2050 rp = VTOR(vp); 2051 if ((rp->r_flags & RDIRTY) || rp->r_iocnt > 0) 2052 (void) nfs3_putpage(vp, (offset_t)0, 0, 0, cr); 2053 #endif 2054 } 2055 2056 for (;;) { 2057 /* get (or renew) access to the file system */ 2058 if (held) { 2059 cachefs_cd_release(fscp); 2060 held = 0; 2061 } 2062 error = cachefs_cd_access(fscp, connected, 0); 2063 if (error) 2064 goto out; 2065 held = 1; 2066 2067 /* 2068 * If it has been specified that the return value will 2069 * just be used as a hint, and we are only being asked 2070 * for size, fsid or rdevid, then return the client's 2071 * notion of these values without checking to make sure 2072 * that the attribute cache is up to date. 2073 * The whole point is to avoid an over the wire GETATTR 2074 * call. 2075 */ 2076 if (flags & ATTR_HINT) { 2077 if (vap->va_mask == 2078 (vap->va_mask & (AT_SIZE | AT_FSID | AT_RDEV))) { 2079 if (vap->va_mask | AT_SIZE) 2080 vap->va_size = cp->c_size; 2081 /* 2082 * Return the FSID of the cachefs filesystem, 2083 * not the back filesystem 2084 */ 2085 if (vap->va_mask | AT_FSID) 2086 vap->va_fsid = vp->v_vfsp->vfs_dev; 2087 if (vap->va_mask | AT_RDEV) 2088 vap->va_rdev = cp->c_attr.va_rdev; 2089 goto out; 2090 } 2091 } 2092 2093 mutex_enter(&cp->c_statelock); 2094 if ((cp->c_metadata.md_flags & MD_NEEDATTRS) && 2095 (fscp->fs_cdconnected != CFS_CD_CONNECTED)) { 2096 mutex_exit(&cp->c_statelock); 2097 connected = 1; 2098 continue; 2099 } 2100 2101 error = CFSOP_CHECK_COBJECT(fscp, cp, 0, cr); 2102 if (CFS_TIMEOUT(fscp, error)) { 2103 mutex_exit(&cp->c_statelock); 2104 cachefs_cd_release(fscp); 2105 held = 0; 2106 cachefs_cd_timedout(fscp); 2107 continue; 2108 } 2109 if (error) { 2110 mutex_exit(&cp->c_statelock); 2111 break; 2112 } 2113 2114 /* check for fileno conflict */ 2115 if ((fscp->fs_inum_size > 0) && 2116 ((cp->c_metadata.md_flags & MD_LOCALFILENO) == 0)) { 2117 ino64_t fakenum; 2118 2119 mutex_exit(&cp->c_statelock); 2120 mutex_enter(&fscp->fs_fslock); 2121 fakenum = cachefs_inum_real2fake(fscp, 2122 cp->c_attr.va_nodeid); 2123 if (fakenum == 0) { 2124 fakenum = cachefs_fileno_conflict(fscp, 2125 cp->c_attr.va_nodeid); 2126 } 2127 mutex_exit(&fscp->fs_fslock); 2128 2129 mutex_enter(&cp->c_statelock); 2130 cp->c_metadata.md_flags |= MD_LOCALFILENO; 2131 cp->c_metadata.md_localfileno = fakenum; 2132 cp->c_flags |= CN_UPDATED; 2133 } 2134 2135 /* copy out the attributes */ 2136 *vap = cp->c_attr; 2137 2138 /* 2139 * return the FSID of the cachefs filesystem, 2140 * not the back filesystem 2141 */ 2142 vap->va_fsid = vp->v_vfsp->vfs_dev; 2143 2144 /* return our idea of the size */ 2145 if (cp->c_size > vap->va_size) 2146 vap->va_size = cp->c_size; 2147 2148 /* overwrite with our version of fileno and timestamps */ 2149 vap->va_nodeid = cp->c_metadata.md_localfileno; 2150 vap->va_mtime = cp->c_metadata.md_localmtime; 2151 vap->va_ctime = cp->c_metadata.md_localctime; 2152 2153 mutex_exit(&cp->c_statelock); 2154 break; 2155 } 2156 out: 2157 if (held) 2158 cachefs_cd_release(fscp); 2159 #ifdef CFS_CD_DEBUG 2160 ASSERT((curthread->t_flag & T_CD_HELD) == 0); 2161 #endif 2162 2163 #ifdef CFSDEBUG 2164 CFS_DEBUG(CFSDEBUG_VOPS) 2165 printf("cachefs_getattr: EXIT error = %d\n", error); 2166 #endif 2167 return (error); 2168 } 2169 2170 /* 2171 * cachefs_getattr_backfs_nfsv4 2172 * 2173 * Call NFSv4 back filesystem to handle the getattr (cachefs 2174 * pass-through support for NFSv4). 2175 */ 2176 static int 2177 cachefs_getattr_backfs_nfsv4(vnode_t *vp, vattr_t *vap, 2178 int flags, cred_t *cr) 2179 { 2180 cnode_t *cp = VTOC(vp); 2181 fscache_t *fscp = C_TO_FSCACHE(cp); 2182 vnode_t *backvp; 2183 int error; 2184 2185 /* 2186 * For NFSv4 pass-through to work, only connected operation 2187 * is supported, the cnode backvp must exist, and cachefs 2188 * optional (eg., disconnectable) flags are turned off. Assert 2189 * these conditions for the getattr operation. 2190 */ 2191 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 2192 CFS_BACKFS_NFSV4_ASSERT_CNODE(cp); 2193 2194 /* Call backfs vnode op after extracting backvp */ 2195 mutex_enter(&cp->c_statelock); 2196 backvp = cp->c_backvp; 2197 mutex_exit(&cp->c_statelock); 2198 2199 CFS_DPRINT_BACKFS_NFSV4(fscp, ("cachefs_getattr_backfs_nfsv4: cnode %p," 2200 " backvp %p\n", cp, backvp)); 2201 error = VOP_GETATTR(backvp, vap, flags, cr); 2202 2203 /* Update attributes */ 2204 cp->c_attr = *vap; 2205 2206 /* 2207 * return the FSID of the cachefs filesystem, 2208 * not the back filesystem 2209 */ 2210 vap->va_fsid = vp->v_vfsp->vfs_dev; 2211 2212 return (error); 2213 } 2214 2215 /*ARGSUSED4*/ 2216 static int 2217 cachefs_setattr( 2218 vnode_t *vp, 2219 vattr_t *vap, 2220 int flags, 2221 cred_t *cr, 2222 caller_context_t *ct) 2223 { 2224 cnode_t *cp = VTOC(vp); 2225 fscache_t *fscp = C_TO_FSCACHE(cp); 2226 int error; 2227 int connected; 2228 int held = 0; 2229 2230 if (getzoneid() != GLOBAL_ZONEID) 2231 return (EPERM); 2232 2233 /* 2234 * Cachefs only provides pass-through support for NFSv4, 2235 * and all vnode operations are passed through to the 2236 * back file system. For NFSv4 pass-through to work, only 2237 * connected operation is supported, the cnode backvp must 2238 * exist, and cachefs optional (eg., disconnectable) flags 2239 * are turned off. Assert these conditions to ensure that 2240 * the backfilesystem is called for the setattr operation. 2241 */ 2242 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 2243 CFS_BACKFS_NFSV4_ASSERT_CNODE(cp); 2244 2245 connected = 0; 2246 for (;;) { 2247 /* drop hold on file system */ 2248 if (held) { 2249 /* Won't loop with NFSv4 connected behavior */ 2250 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 2251 cachefs_cd_release(fscp); 2252 held = 0; 2253 } 2254 2255 /* aquire access to the file system */ 2256 error = cachefs_cd_access(fscp, connected, 1); 2257 if (error) 2258 break; 2259 held = 1; 2260 2261 /* perform the setattr */ 2262 error = cachefs_setattr_common(vp, vap, flags, cr, ct); 2263 if (error) { 2264 /* if connected */ 2265 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 2266 if (CFS_TIMEOUT(fscp, error)) { 2267 cachefs_cd_release(fscp); 2268 held = 0; 2269 cachefs_cd_timedout(fscp); 2270 connected = 0; 2271 continue; 2272 } 2273 } 2274 2275 /* else must be disconnected */ 2276 else { 2277 if (CFS_TIMEOUT(fscp, error)) { 2278 connected = 1; 2279 continue; 2280 } 2281 } 2282 } 2283 break; 2284 } 2285 2286 if (held) { 2287 cachefs_cd_release(fscp); 2288 } 2289 #ifdef CFS_CD_DEBUG 2290 ASSERT((curthread->t_flag & T_CD_HELD) == 0); 2291 #endif 2292 return (error); 2293 } 2294 2295 static int 2296 cachefs_setattr_common( 2297 vnode_t *vp, 2298 vattr_t *vap, 2299 int flags, 2300 cred_t *cr, 2301 caller_context_t *ct) 2302 { 2303 cnode_t *cp = VTOC(vp); 2304 fscache_t *fscp = C_TO_FSCACHE(cp); 2305 cachefscache_t *cachep = fscp->fs_cache; 2306 uint_t mask = vap->va_mask; 2307 int error = 0; 2308 uint_t bcnt; 2309 2310 /* Cannot set these attributes. */ 2311 if (mask & AT_NOSET) 2312 return (EINVAL); 2313 2314 /* 2315 * Truncate file. Must have write permission and not be a directory. 2316 */ 2317 if (mask & AT_SIZE) { 2318 if (vp->v_type == VDIR) { 2319 if (CACHEFS_LOG_LOGGING(cachep, CACHEFS_LOG_TRUNCATE)) 2320 cachefs_log_truncate(cachep, EISDIR, 2321 fscp->fs_cfsvfsp, 2322 &cp->c_metadata.md_cookie, 2323 cp->c_id.cid_fileno, 2324 crgetuid(cr), vap->va_size); 2325 return (EISDIR); 2326 } 2327 } 2328 2329 /* 2330 * Gotta deal with one special case here, where we're setting the 2331 * size of the file. First, we zero out part of the page after the 2332 * new size of the file. Then we toss (not write) all pages after 2333 * page in which the new offset occurs. Note that the NULL passed 2334 * in instead of a putapage() fn parameter is correct, since 2335 * no dirty pages will be found (B_TRUNC | B_INVAL). 2336 */ 2337 2338 rw_enter(&cp->c_rwlock, RW_WRITER); 2339 2340 /* sync dirty pages */ 2341 if (!CFS_ISFS_BACKFS_NFSV4(fscp)) { 2342 error = cachefs_putpage_common(vp, (offset_t)0, 0, 0, cr); 2343 if (error == EINTR) 2344 goto out; 2345 } 2346 error = 0; 2347 2348 /* if connected */ 2349 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 2350 error = cachefs_setattr_connected(vp, vap, flags, cr, ct); 2351 } 2352 /* else must be disconnected */ 2353 else { 2354 error = cachefs_setattr_disconnected(vp, vap, flags, cr, ct); 2355 } 2356 if (error) 2357 goto out; 2358 2359 /* 2360 * If the file size has been changed then 2361 * toss whole pages beyond the end of the file and zero 2362 * the portion of the last page that is beyond the end of the file. 2363 */ 2364 if (mask & AT_SIZE && !CFS_ISFS_BACKFS_NFSV4(fscp)) { 2365 bcnt = (uint_t)(cp->c_size & PAGEOFFSET); 2366 if (bcnt) 2367 pvn_vpzero(vp, cp->c_size, PAGESIZE - bcnt); 2368 (void) pvn_vplist_dirty(vp, cp->c_size, cachefs_push, 2369 B_TRUNC | B_INVAL, cr); 2370 } 2371 2372 out: 2373 rw_exit(&cp->c_rwlock); 2374 2375 if ((mask & AT_SIZE) && 2376 (CACHEFS_LOG_LOGGING(cachep, CACHEFS_LOG_TRUNCATE))) 2377 cachefs_log_truncate(cachep, error, fscp->fs_cfsvfsp, 2378 &cp->c_metadata.md_cookie, cp->c_id.cid_fileno, 2379 crgetuid(cr), vap->va_size); 2380 2381 return (error); 2382 } 2383 2384 static int 2385 cachefs_setattr_connected( 2386 vnode_t *vp, 2387 vattr_t *vap, 2388 int flags, 2389 cred_t *cr, 2390 caller_context_t *ct) 2391 { 2392 cnode_t *cp = VTOC(vp); 2393 fscache_t *fscp = C_TO_FSCACHE(cp); 2394 uint_t mask = vap->va_mask; 2395 int error = 0; 2396 int setsize; 2397 2398 mutex_enter(&cp->c_statelock); 2399 2400 if (cp->c_backvp == NULL) { 2401 error = cachefs_getbackvp(fscp, cp); 2402 if (error) 2403 goto out; 2404 } 2405 2406 error = CFSOP_CHECK_COBJECT(fscp, cp, 0, cr); 2407 if (error) 2408 goto out; 2409 2410 CFS_DPRINT_BACKFS_NFSV4(fscp, ("cachefs_setattr (nfsv4): cnode %p, " 2411 "backvp %p\n", cp, cp->c_backvp)); 2412 error = VOP_SETATTR(cp->c_backvp, vap, flags, cr, ct); 2413 if (error) { 2414 goto out; 2415 } 2416 2417 /* if the size of the file is being changed */ 2418 if (mask & AT_SIZE) { 2419 cp->c_size = vap->va_size; 2420 error = 0; 2421 setsize = 0; 2422 2423 /* see if okay to try to set the file size */ 2424 if (((cp->c_flags & CN_NOCACHE) == 0) && 2425 CFS_ISFS_NONSHARED(fscp)) { 2426 /* okay to set size if file is populated */ 2427 if (cp->c_metadata.md_flags & MD_POPULATED) 2428 setsize = 1; 2429 2430 /* 2431 * Okay to set size if front file exists and setting 2432 * file size to zero. 2433 */ 2434 if ((cp->c_metadata.md_flags & MD_FILE) && 2435 (vap->va_size == 0)) 2436 setsize = 1; 2437 } 2438 2439 /* if okay to try to set the file size */ 2440 if (setsize) { 2441 error = 0; 2442 if (cp->c_frontvp == NULL) 2443 error = cachefs_getfrontfile(cp); 2444 if (error == 0) 2445 error = cachefs_frontfile_size(cp, cp->c_size); 2446 } else if (cp->c_metadata.md_flags & MD_FILE) { 2447 /* make sure file gets nocached */ 2448 error = EEXIST; 2449 } 2450 2451 /* if we have to nocache the file */ 2452 if (error) { 2453 if ((cp->c_flags & CN_NOCACHE) == 0 && 2454 !CFS_ISFS_BACKFS_NFSV4(fscp)) 2455 cachefs_nocache(cp); 2456 error = 0; 2457 } 2458 } 2459 2460 cp->c_flags |= CN_UPDATED; 2461 2462 /* XXX bob: given what modify_cobject does this seems unnecessary */ 2463 cp->c_attr.va_mask = AT_ALL; 2464 error = VOP_GETATTR(cp->c_backvp, &cp->c_attr, 0, cr); 2465 if (error) 2466 goto out; 2467 2468 cp->c_attr.va_size = MAX(cp->c_attr.va_size, cp->c_size); 2469 cp->c_size = cp->c_attr.va_size; 2470 2471 CFSOP_MODIFY_COBJECT(fscp, cp, cr); 2472 out: 2473 mutex_exit(&cp->c_statelock); 2474 return (error); 2475 } 2476 2477 /* 2478 * perform the setattr on the local file system 2479 */ 2480 /*ARGSUSED4*/ 2481 static int 2482 cachefs_setattr_disconnected( 2483 vnode_t *vp, 2484 vattr_t *vap, 2485 int flags, 2486 cred_t *cr, 2487 caller_context_t *ct) 2488 { 2489 cnode_t *cp = VTOC(vp); 2490 fscache_t *fscp = C_TO_FSCACHE(cp); 2491 int mask; 2492 int error; 2493 int newfile; 2494 off_t commit = 0; 2495 2496 if (CFS_ISFS_WRITE_AROUND(fscp)) 2497 return (ETIMEDOUT); 2498 2499 /* if we do not have good attributes */ 2500 if (cp->c_metadata.md_flags & MD_NEEDATTRS) 2501 return (ETIMEDOUT); 2502 2503 /* primary concern is to keep this routine as much like ufs_setattr */ 2504 2505 mutex_enter(&cp->c_statelock); 2506 2507 error = secpolicy_vnode_setattr(cr, vp, vap, &cp->c_attr, flags, 2508 cachefs_access_local, cp); 2509 2510 if (error) 2511 goto out; 2512 2513 mask = vap->va_mask; 2514 2515 /* if changing the size of the file */ 2516 if (mask & AT_SIZE) { 2517 if (vp->v_type == VDIR) { 2518 error = EISDIR; 2519 goto out; 2520 } 2521 2522 if (vp->v_type == VFIFO) { 2523 error = 0; 2524 goto out; 2525 } 2526 2527 if ((vp->v_type != VREG) && 2528 !((vp->v_type == VLNK) && (vap->va_size == 0))) { 2529 error = EINVAL; 2530 goto out; 2531 } 2532 2533 if (vap->va_size > fscp->fs_offmax) { 2534 error = EFBIG; 2535 goto out; 2536 } 2537 2538 /* if the file is not populated and we are not truncating it */ 2539 if (((cp->c_metadata.md_flags & MD_POPULATED) == 0) && 2540 (vap->va_size != 0)) { 2541 error = ETIMEDOUT; 2542 goto out; 2543 } 2544 2545 if ((cp->c_metadata.md_flags & MD_MAPPING) == 0) { 2546 error = cachefs_dlog_cidmap(fscp); 2547 if (error) { 2548 error = ENOSPC; 2549 goto out; 2550 } 2551 cp->c_metadata.md_flags |= MD_MAPPING; 2552 } 2553 2554 /* log the operation */ 2555 commit = cachefs_dlog_setattr(fscp, vap, flags, cp, cr); 2556 if (commit == 0) { 2557 error = ENOSPC; 2558 goto out; 2559 } 2560 cp->c_flags &= ~CN_NOCACHE; 2561 2562 /* special case truncating fast sym links */ 2563 if ((vp->v_type == VLNK) && 2564 (cp->c_metadata.md_flags & MD_FASTSYMLNK)) { 2565 /* XXX how can we get here */ 2566 /* XXX should update mtime */ 2567 cp->c_size = 0; 2568 error = 0; 2569 goto out; 2570 } 2571 2572 /* get the front file, this may create one */ 2573 newfile = (cp->c_metadata.md_flags & MD_FILE) ? 0 : 1; 2574 if (cp->c_frontvp == NULL) { 2575 error = cachefs_getfrontfile(cp); 2576 if (error) 2577 goto out; 2578 } 2579 ASSERT(cp->c_frontvp); 2580 if (newfile && (cp->c_flags & CN_UPDATED)) { 2581 /* allocate space for the metadata */ 2582 ASSERT((cp->c_flags & CN_ALLOC_PENDING) == 0); 2583 ASSERT((cp->c_filegrp->fg_flags & CFS_FG_ALLOC_ATTR) 2584 == 0); 2585 error = filegrp_write_metadata(cp->c_filegrp, 2586 &cp->c_id, &cp->c_metadata); 2587 if (error) 2588 goto out; 2589 } 2590 2591 /* change the size of the front file */ 2592 error = cachefs_frontfile_size(cp, vap->va_size); 2593 if (error) 2594 goto out; 2595 cp->c_attr.va_size = cp->c_size = vap->va_size; 2596 gethrestime(&cp->c_metadata.md_localmtime); 2597 cp->c_metadata.md_flags |= MD_POPULATED | MD_LOCALMTIME; 2598 cachefs_modified(cp); 2599 cp->c_flags |= CN_UPDATED; 2600 } 2601 2602 if (mask & AT_MODE) { 2603 /* mark as modified */ 2604 if (cachefs_modified_alloc(cp)) { 2605 error = ENOSPC; 2606 goto out; 2607 } 2608 2609 if ((cp->c_metadata.md_flags & MD_MAPPING) == 0) { 2610 error = cachefs_dlog_cidmap(fscp); 2611 if (error) { 2612 error = ENOSPC; 2613 goto out; 2614 } 2615 cp->c_metadata.md_flags |= MD_MAPPING; 2616 } 2617 2618 /* log the operation if not already logged */ 2619 if (commit == 0) { 2620 commit = cachefs_dlog_setattr(fscp, vap, flags, cp, cr); 2621 if (commit == 0) { 2622 error = ENOSPC; 2623 goto out; 2624 } 2625 } 2626 2627 cp->c_attr.va_mode &= S_IFMT; 2628 cp->c_attr.va_mode |= vap->va_mode & ~S_IFMT; 2629 gethrestime(&cp->c_metadata.md_localctime); 2630 cp->c_metadata.md_flags |= MD_LOCALCTIME; 2631 cp->c_flags |= CN_UPDATED; 2632 } 2633 2634 if (mask & (AT_UID|AT_GID)) { 2635 2636 /* mark as modified */ 2637 if (cachefs_modified_alloc(cp)) { 2638 error = ENOSPC; 2639 goto out; 2640 } 2641 2642 if ((cp->c_metadata.md_flags & MD_MAPPING) == 0) { 2643 error = cachefs_dlog_cidmap(fscp); 2644 if (error) { 2645 error = ENOSPC; 2646 goto out; 2647 } 2648 cp->c_metadata.md_flags |= MD_MAPPING; 2649 } 2650 2651 /* log the operation if not already logged */ 2652 if (commit == 0) { 2653 commit = cachefs_dlog_setattr(fscp, vap, flags, cp, cr); 2654 if (commit == 0) { 2655 error = ENOSPC; 2656 goto out; 2657 } 2658 } 2659 2660 if (mask & AT_UID) 2661 cp->c_attr.va_uid = vap->va_uid; 2662 2663 if (mask & AT_GID) 2664 cp->c_attr.va_gid = vap->va_gid; 2665 gethrestime(&cp->c_metadata.md_localctime); 2666 cp->c_metadata.md_flags |= MD_LOCALCTIME; 2667 cp->c_flags |= CN_UPDATED; 2668 } 2669 2670 2671 if (mask & (AT_MTIME|AT_ATIME)) { 2672 /* mark as modified */ 2673 if (cachefs_modified_alloc(cp)) { 2674 error = ENOSPC; 2675 goto out; 2676 } 2677 2678 if ((cp->c_metadata.md_flags & MD_MAPPING) == 0) { 2679 error = cachefs_dlog_cidmap(fscp); 2680 if (error) { 2681 error = ENOSPC; 2682 goto out; 2683 } 2684 cp->c_metadata.md_flags |= MD_MAPPING; 2685 } 2686 2687 /* log the operation if not already logged */ 2688 if (commit == 0) { 2689 commit = cachefs_dlog_setattr(fscp, vap, flags, cp, cr); 2690 if (commit == 0) { 2691 error = ENOSPC; 2692 goto out; 2693 } 2694 } 2695 2696 if (mask & AT_MTIME) { 2697 cp->c_metadata.md_localmtime = vap->va_mtime; 2698 cp->c_metadata.md_flags |= MD_LOCALMTIME; 2699 } 2700 if (mask & AT_ATIME) 2701 cp->c_attr.va_atime = vap->va_atime; 2702 gethrestime(&cp->c_metadata.md_localctime); 2703 cp->c_metadata.md_flags |= MD_LOCALCTIME; 2704 cp->c_flags |= CN_UPDATED; 2705 } 2706 2707 out: 2708 mutex_exit(&cp->c_statelock); 2709 2710 /* commit the log entry */ 2711 if (commit) { 2712 if (cachefs_dlog_commit(fscp, commit, error)) { 2713 /*EMPTY*/ 2714 /* XXX bob: fix on panic */ 2715 } 2716 } 2717 return (error); 2718 } 2719 2720 /* ARGSUSED */ 2721 static int 2722 cachefs_access(vnode_t *vp, int mode, int flags, cred_t *cr) 2723 { 2724 cnode_t *cp = VTOC(vp); 2725 fscache_t *fscp = C_TO_FSCACHE(cp); 2726 int error; 2727 int held = 0; 2728 int connected = 0; 2729 2730 #ifdef CFSDEBUG 2731 CFS_DEBUG(CFSDEBUG_VOPS) 2732 printf("cachefs_access: ENTER vp %p\n", (void *)vp); 2733 #endif 2734 if (getzoneid() != GLOBAL_ZONEID) { 2735 error = EPERM; 2736 goto out; 2737 } 2738 2739 /* 2740 * Cachefs only provides pass-through support for NFSv4, 2741 * and all vnode operations are passed through to the 2742 * back file system. For NFSv4 pass-through to work, only 2743 * connected operation is supported, the cnode backvp must 2744 * exist, and cachefs optional (eg., disconnectable) flags 2745 * are turned off. Assert these conditions to ensure that 2746 * the backfilesystem is called for the access operation. 2747 */ 2748 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 2749 CFS_BACKFS_NFSV4_ASSERT_CNODE(cp); 2750 2751 for (;;) { 2752 /* get (or renew) access to the file system */ 2753 if (held) { 2754 /* Won't loop with NFSv4 connected behavior */ 2755 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 2756 cachefs_cd_release(fscp); 2757 held = 0; 2758 } 2759 error = cachefs_cd_access(fscp, connected, 0); 2760 if (error) 2761 break; 2762 held = 1; 2763 2764 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 2765 error = cachefs_access_connected(vp, mode, flags, 2766 cr); 2767 if (CFS_TIMEOUT(fscp, error)) { 2768 cachefs_cd_release(fscp); 2769 held = 0; 2770 cachefs_cd_timedout(fscp); 2771 connected = 0; 2772 continue; 2773 } 2774 } else { 2775 mutex_enter(&cp->c_statelock); 2776 error = cachefs_access_local(cp, mode, cr); 2777 mutex_exit(&cp->c_statelock); 2778 if (CFS_TIMEOUT(fscp, error)) { 2779 if (cachefs_cd_access_miss(fscp)) { 2780 mutex_enter(&cp->c_statelock); 2781 if (cp->c_backvp == NULL) { 2782 (void) cachefs_getbackvp(fscp, 2783 cp); 2784 } 2785 mutex_exit(&cp->c_statelock); 2786 error = cachefs_access_connected(vp, 2787 mode, flags, cr); 2788 if (!CFS_TIMEOUT(fscp, error)) 2789 break; 2790 delay(5*hz); 2791 connected = 0; 2792 continue; 2793 } 2794 connected = 1; 2795 continue; 2796 } 2797 } 2798 break; 2799 } 2800 if (held) 2801 cachefs_cd_release(fscp); 2802 #ifdef CFS_CD_DEBUG 2803 ASSERT((curthread->t_flag & T_CD_HELD) == 0); 2804 #endif 2805 out: 2806 #ifdef CFSDEBUG 2807 CFS_DEBUG(CFSDEBUG_VOPS) 2808 printf("cachefs_access: EXIT error = %d\n", error); 2809 #endif 2810 return (error); 2811 } 2812 2813 static int 2814 cachefs_access_connected(struct vnode *vp, int mode, int flags, cred_t *cr) 2815 { 2816 cnode_t *cp = VTOC(vp); 2817 fscache_t *fscp = C_TO_FSCACHE(cp); 2818 int error = 0; 2819 2820 mutex_enter(&cp->c_statelock); 2821 2822 /* Make sure the cnode attrs are valid first. */ 2823 error = CFSOP_CHECK_COBJECT(fscp, cp, 0, cr); 2824 if (error) 2825 goto out; 2826 2827 /* see if can do a local file system check */ 2828 if ((fscp->fs_info.fi_mntflags & CFS_ACCESS_BACKFS) == 0 && 2829 !CFS_ISFS_BACKFS_NFSV4(fscp)) { 2830 error = cachefs_access_local(cp, mode, cr); 2831 goto out; 2832 } 2833 2834 /* else do a remote file system check */ 2835 else { 2836 if (cp->c_backvp == NULL) { 2837 error = cachefs_getbackvp(fscp, cp); 2838 if (error) 2839 goto out; 2840 } 2841 2842 CFS_DPRINT_BACKFS_NFSV4(fscp, 2843 ("cachefs_access (nfsv4): cnode %p, backvp %p\n", 2844 cp, cp->c_backvp)); 2845 error = VOP_ACCESS(cp->c_backvp, mode, flags, cr); 2846 2847 /* 2848 * even though we don't `need' the ACL to do access 2849 * via the backvp, we should cache it here to make our 2850 * behavior more reasonable if we go disconnected. 2851 */ 2852 2853 if (((fscp->fs_info.fi_mntflags & CFS_NOACL) == 0) && 2854 (cachefs_vtype_aclok(vp)) && 2855 ((cp->c_flags & CN_NOCACHE) == 0) && 2856 (!CFS_ISFS_BACKFS_NFSV4(fscp)) && 2857 ((cp->c_metadata.md_flags & MD_ACL) == 0)) 2858 (void) cachefs_cacheacl(cp, NULL); 2859 } 2860 out: 2861 /* 2862 * If NFS returned ESTALE, mark this cnode as stale, so that 2863 * the vn_open retry will read the file anew from backfs 2864 */ 2865 if (error == ESTALE) 2866 cachefs_cnode_stale(cp); 2867 2868 mutex_exit(&cp->c_statelock); 2869 return (error); 2870 } 2871 2872 /* 2873 * CFS has a fastsymlink scheme. If the size of the link is < C_FSL_SIZE, then 2874 * the link is placed in the metadata itself (no front file is allocated). 2875 */ 2876 static int 2877 cachefs_readlink(vnode_t *vp, uio_t *uiop, cred_t *cr) 2878 { 2879 int error = 0; 2880 cnode_t *cp = VTOC(vp); 2881 fscache_t *fscp = C_TO_FSCACHE(cp); 2882 cachefscache_t *cachep = fscp->fs_cache; 2883 int held = 0; 2884 int connected = 0; 2885 2886 if (getzoneid() != GLOBAL_ZONEID) 2887 return (EPERM); 2888 2889 if (vp->v_type != VLNK) 2890 return (EINVAL); 2891 2892 /* 2893 * Cachefs only provides pass-through support for NFSv4, 2894 * and all vnode operations are passed through to the 2895 * back file system. For NFSv4 pass-through to work, only 2896 * connected operation is supported, the cnode backvp must 2897 * exist, and cachefs optional (eg., disconnectable) flags 2898 * are turned off. Assert these conditions to ensure that 2899 * the backfilesystem is called for the readlink operation. 2900 */ 2901 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 2902 CFS_BACKFS_NFSV4_ASSERT_CNODE(cp); 2903 2904 for (;;) { 2905 /* get (or renew) access to the file system */ 2906 if (held) { 2907 /* Won't loop with NFSv4 connected behavior */ 2908 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 2909 cachefs_cd_release(fscp); 2910 held = 0; 2911 } 2912 error = cachefs_cd_access(fscp, connected, 0); 2913 if (error) 2914 break; 2915 held = 1; 2916 2917 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 2918 /* 2919 * since readlink_connected will call stuffsymlink 2920 * on success, have to serialize access 2921 */ 2922 if (!rw_tryenter(&cp->c_rwlock, RW_WRITER)) { 2923 cachefs_cd_release(fscp); 2924 rw_enter(&cp->c_rwlock, RW_WRITER); 2925 error = cachefs_cd_access(fscp, connected, 0); 2926 if (error) { 2927 held = 0; 2928 rw_exit(&cp->c_rwlock); 2929 break; 2930 } 2931 } 2932 error = cachefs_readlink_connected(vp, uiop, cr); 2933 rw_exit(&cp->c_rwlock); 2934 if (CFS_TIMEOUT(fscp, error)) { 2935 cachefs_cd_release(fscp); 2936 held = 0; 2937 cachefs_cd_timedout(fscp); 2938 connected = 0; 2939 continue; 2940 } 2941 } else { 2942 error = cachefs_readlink_disconnected(vp, uiop); 2943 if (CFS_TIMEOUT(fscp, error)) { 2944 if (cachefs_cd_access_miss(fscp)) { 2945 /* as above */ 2946 if (!rw_tryenter(&cp->c_rwlock, 2947 RW_WRITER)) { 2948 cachefs_cd_release(fscp); 2949 rw_enter(&cp->c_rwlock, 2950 RW_WRITER); 2951 error = cachefs_cd_access(fscp, 2952 connected, 0); 2953 if (error) { 2954 held = 0; 2955 rw_exit(&cp->c_rwlock); 2956 break; 2957 } 2958 } 2959 error = cachefs_readlink_connected(vp, 2960 uiop, cr); 2961 rw_exit(&cp->c_rwlock); 2962 if (!CFS_TIMEOUT(fscp, error)) 2963 break; 2964 delay(5*hz); 2965 connected = 0; 2966 continue; 2967 } 2968 connected = 1; 2969 continue; 2970 } 2971 } 2972 break; 2973 } 2974 if (CACHEFS_LOG_LOGGING(cachep, CACHEFS_LOG_READLINK)) 2975 cachefs_log_readlink(cachep, error, fscp->fs_cfsvfsp, 2976 &cp->c_metadata.md_cookie, cp->c_id.cid_fileno, 2977 crgetuid(cr), cp->c_size); 2978 2979 if (held) 2980 cachefs_cd_release(fscp); 2981 #ifdef CFS_CD_DEBUG 2982 ASSERT((curthread->t_flag & T_CD_HELD) == 0); 2983 #endif 2984 2985 /* 2986 * The over the wire error for attempting to readlink something 2987 * other than a symbolic link is ENXIO. However, we need to 2988 * return EINVAL instead of ENXIO, so we map it here. 2989 */ 2990 return (error == ENXIO ? EINVAL : error); 2991 } 2992 2993 static int 2994 cachefs_readlink_connected(vnode_t *vp, uio_t *uiop, cred_t *cr) 2995 { 2996 int error; 2997 cnode_t *cp = VTOC(vp); 2998 fscache_t *fscp = C_TO_FSCACHE(cp); 2999 caddr_t buf; 3000 int buflen; 3001 int readcache = 0; 3002 3003 mutex_enter(&cp->c_statelock); 3004 3005 error = CFSOP_CHECK_COBJECT(fscp, cp, 0, cr); 3006 if (error) 3007 goto out; 3008 3009 /* if the sym link is cached as a fast sym link */ 3010 if (cp->c_metadata.md_flags & MD_FASTSYMLNK) { 3011 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 3012 error = uiomove(cp->c_metadata.md_allocinfo, 3013 MIN(cp->c_size, uiop->uio_resid), UIO_READ, uiop); 3014 #ifdef CFSDEBUG 3015 readcache = 1; 3016 goto out; 3017 #else /* CFSDEBUG */ 3018 /* XXX KLUDGE! correct for insidious 0-len symlink */ 3019 if (cp->c_size != 0) { 3020 readcache = 1; 3021 goto out; 3022 } 3023 #endif /* CFSDEBUG */ 3024 } 3025 3026 /* if the sym link is cached in a front file */ 3027 if (cp->c_metadata.md_flags & MD_POPULATED) { 3028 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 3029 ASSERT(cp->c_metadata.md_flags & MD_FILE); 3030 if (cp->c_frontvp == NULL) { 3031 (void) cachefs_getfrontfile(cp); 3032 } 3033 if (cp->c_metadata.md_flags & MD_POPULATED) { 3034 /* read symlink data from frontfile */ 3035 uiop->uio_offset = 0; 3036 (void) VOP_RWLOCK(cp->c_frontvp, 3037 V_WRITELOCK_FALSE, NULL); 3038 error = VOP_READ(cp->c_frontvp, uiop, 0, kcred, NULL); 3039 VOP_RWUNLOCK(cp->c_frontvp, V_WRITELOCK_FALSE, NULL); 3040 3041 /* XXX KLUDGE! correct for insidious 0-len symlink */ 3042 if (cp->c_size != 0) { 3043 readcache = 1; 3044 goto out; 3045 } 3046 } 3047 } 3048 3049 /* get the sym link contents from the back fs */ 3050 error = cachefs_readlink_back(cp, cr, &buf, &buflen); 3051 if (error) 3052 goto out; 3053 3054 /* copy the contents out to the user */ 3055 error = uiomove(buf, MIN(buflen, uiop->uio_resid), UIO_READ, uiop); 3056 3057 /* 3058 * try to cache the sym link, note that its a noop if NOCACHE is set 3059 * or if NFSv4 pass-through is enabled. 3060 */ 3061 if (cachefs_stuffsymlink(cp, buf, buflen)) { 3062 cachefs_nocache(cp); 3063 } 3064 3065 cachefs_kmem_free(buf, MAXPATHLEN); 3066 3067 out: 3068 mutex_exit(&cp->c_statelock); 3069 if (error == 0) { 3070 if (readcache) 3071 fscp->fs_stats.st_hits++; 3072 else 3073 fscp->fs_stats.st_misses++; 3074 } 3075 return (error); 3076 } 3077 3078 static int 3079 cachefs_readlink_disconnected(vnode_t *vp, uio_t *uiop) 3080 { 3081 int error; 3082 cnode_t *cp = VTOC(vp); 3083 fscache_t *fscp = C_TO_FSCACHE(cp); 3084 int readcache = 0; 3085 3086 mutex_enter(&cp->c_statelock); 3087 3088 /* if the sym link is cached as a fast sym link */ 3089 if (cp->c_metadata.md_flags & MD_FASTSYMLNK) { 3090 error = uiomove(cp->c_metadata.md_allocinfo, 3091 MIN(cp->c_size, uiop->uio_resid), UIO_READ, uiop); 3092 readcache = 1; 3093 goto out; 3094 } 3095 3096 /* if the sym link is cached in a front file */ 3097 if (cp->c_metadata.md_flags & MD_POPULATED) { 3098 ASSERT(cp->c_metadata.md_flags & MD_FILE); 3099 if (cp->c_frontvp == NULL) { 3100 (void) cachefs_getfrontfile(cp); 3101 } 3102 if (cp->c_metadata.md_flags & MD_POPULATED) { 3103 /* read symlink data from frontfile */ 3104 uiop->uio_offset = 0; 3105 (void) VOP_RWLOCK(cp->c_frontvp, 3106 V_WRITELOCK_FALSE, NULL); 3107 error = VOP_READ(cp->c_frontvp, uiop, 0, kcred, NULL); 3108 VOP_RWUNLOCK(cp->c_frontvp, V_WRITELOCK_FALSE, NULL); 3109 readcache = 1; 3110 goto out; 3111 } 3112 } 3113 error = ETIMEDOUT; 3114 3115 out: 3116 mutex_exit(&cp->c_statelock); 3117 if (error == 0) { 3118 if (readcache) 3119 fscp->fs_stats.st_hits++; 3120 else 3121 fscp->fs_stats.st_misses++; 3122 } 3123 return (error); 3124 } 3125 3126 /*ARGSUSED*/ 3127 static int 3128 cachefs_fsync(vnode_t *vp, int syncflag, cred_t *cr) 3129 { 3130 cnode_t *cp = VTOC(vp); 3131 int error = 0; 3132 fscache_t *fscp = C_TO_FSCACHE(cp); 3133 int held = 0; 3134 int connected = 0; 3135 3136 #ifdef CFSDEBUG 3137 CFS_DEBUG(CFSDEBUG_VOPS) 3138 printf("cachefs_fsync: ENTER vp %p\n", (void *)vp); 3139 #endif 3140 3141 if (getzoneid() != GLOBAL_ZONEID) { 3142 error = EPERM; 3143 goto out; 3144 } 3145 3146 if (fscp->fs_backvfsp && fscp->fs_backvfsp->vfs_flag & VFS_RDONLY) 3147 goto out; 3148 3149 /* 3150 * Cachefs only provides pass-through support for NFSv4, 3151 * and all vnode operations are passed through to the 3152 * back file system. For NFSv4 pass-through to work, only 3153 * connected operation is supported, the cnode backvp must 3154 * exist, and cachefs optional (eg., disconnectable) flags 3155 * are turned off. Assert these conditions to ensure that 3156 * the backfilesystem is called for the fsync operation. 3157 */ 3158 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 3159 CFS_BACKFS_NFSV4_ASSERT_CNODE(cp); 3160 3161 for (;;) { 3162 /* get (or renew) access to the file system */ 3163 if (held) { 3164 /* Won't loop with NFSv4 connected behavior */ 3165 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 3166 cachefs_cd_release(fscp); 3167 held = 0; 3168 } 3169 error = cachefs_cd_access(fscp, connected, 1); 3170 if (error) 3171 break; 3172 held = 1; 3173 connected = 0; 3174 3175 /* if a regular file, write out the pages */ 3176 if ((vp->v_type == VREG) && vn_has_cached_data(vp) && 3177 !CFS_ISFS_BACKFS_NFSV4(fscp)) { 3178 error = cachefs_putpage_common(vp, (offset_t)0, 3179 0, 0, cr); 3180 if (CFS_TIMEOUT(fscp, error)) { 3181 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 3182 cachefs_cd_release(fscp); 3183 held = 0; 3184 cachefs_cd_timedout(fscp); 3185 continue; 3186 } else { 3187 connected = 1; 3188 continue; 3189 } 3190 } 3191 3192 /* if no space left in cache, wait until connected */ 3193 if ((error == ENOSPC) && 3194 (fscp->fs_cdconnected != CFS_CD_CONNECTED)) { 3195 connected = 1; 3196 continue; 3197 } 3198 3199 /* clear the cnode error if putpage worked */ 3200 if ((error == 0) && cp->c_error) { 3201 mutex_enter(&cp->c_statelock); 3202 cp->c_error = 0; 3203 mutex_exit(&cp->c_statelock); 3204 } 3205 3206 if (error) 3207 break; 3208 } 3209 3210 /* if connected, sync the backvp */ 3211 if ((fscp->fs_cdconnected == CFS_CD_CONNECTED) && 3212 cp->c_backvp) { 3213 mutex_enter(&cp->c_statelock); 3214 if (cp->c_backvp) { 3215 CFS_DPRINT_BACKFS_NFSV4(fscp, 3216 ("cachefs_fsync (nfsv4): cnode %p, " 3217 "backvp %p\n", cp, cp->c_backvp)); 3218 error = VOP_FSYNC(cp->c_backvp, syncflag, cr); 3219 if (CFS_TIMEOUT(fscp, error)) { 3220 mutex_exit(&cp->c_statelock); 3221 cachefs_cd_release(fscp); 3222 held = 0; 3223 cachefs_cd_timedout(fscp); 3224 continue; 3225 } else if (error && (error != EINTR)) 3226 cp->c_error = error; 3227 } 3228 mutex_exit(&cp->c_statelock); 3229 } 3230 3231 /* sync the metadata and the front file to the front fs */ 3232 if (!CFS_ISFS_BACKFS_NFSV4(fscp)) { 3233 error = cachefs_sync_metadata(cp); 3234 if (error && 3235 (fscp->fs_cdconnected == CFS_CD_CONNECTED)) 3236 error = 0; 3237 } 3238 break; 3239 } 3240 3241 if (error == 0) 3242 error = cp->c_error; 3243 3244 if (held) 3245 cachefs_cd_release(fscp); 3246 3247 out: 3248 #ifdef CFS_CD_DEBUG 3249 ASSERT((curthread->t_flag & T_CD_HELD) == 0); 3250 #endif 3251 3252 #ifdef CFSDEBUG 3253 CFS_DEBUG(CFSDEBUG_VOPS) 3254 printf("cachefs_fsync: EXIT vp %p\n", (void *)vp); 3255 #endif 3256 return (error); 3257 } 3258 3259 /* 3260 * Called from cachefs_inactive(), to make sure all the data goes out to disk. 3261 */ 3262 int 3263 cachefs_sync_metadata(cnode_t *cp) 3264 { 3265 int error = 0; 3266 struct filegrp *fgp; 3267 struct vattr va; 3268 fscache_t *fscp = C_TO_FSCACHE(cp); 3269 3270 #ifdef CFSDEBUG 3271 CFS_DEBUG(CFSDEBUG_VOPS) 3272 printf("c_sync_metadata: ENTER cp %p cflag %x\n", 3273 (void *)cp, cp->c_flags); 3274 #endif 3275 3276 mutex_enter(&cp->c_statelock); 3277 if ((cp->c_flags & CN_UPDATED) == 0) 3278 goto out; 3279 if (cp->c_flags & (CN_STALE | CN_DESTROY)) 3280 goto out; 3281 fgp = cp->c_filegrp; 3282 if ((fgp->fg_flags & CFS_FG_WRITE) == 0) 3283 goto out; 3284 if (CFS_ISFS_BACKFS_NFSV4(fscp)) 3285 goto out; 3286 3287 if (fgp->fg_flags & CFS_FG_ALLOC_ATTR) { 3288 mutex_exit(&cp->c_statelock); 3289 error = filegrp_allocattr(fgp); 3290 mutex_enter(&cp->c_statelock); 3291 if (error) { 3292 error = 0; 3293 goto out; 3294 } 3295 } 3296 3297 if (cp->c_flags & CN_ALLOC_PENDING) { 3298 error = filegrp_create_metadata(fgp, &cp->c_metadata, 3299 &cp->c_id); 3300 if (error) 3301 goto out; 3302 cp->c_flags &= ~CN_ALLOC_PENDING; 3303 } 3304 3305 if (cp->c_flags & CN_NEED_FRONT_SYNC) { 3306 if (cp->c_frontvp != NULL) { 3307 error = VOP_FSYNC(cp->c_frontvp, FSYNC, kcred); 3308 if (error) { 3309 cp->c_metadata.md_timestamp.tv_sec = 0; 3310 } else { 3311 va.va_mask = AT_MTIME; 3312 error = VOP_GETATTR(cp->c_frontvp, &va, 0, 3313 kcred); 3314 if (error) 3315 goto out; 3316 cp->c_metadata.md_timestamp = va.va_mtime; 3317 cp->c_flags &= 3318 ~(CN_NEED_FRONT_SYNC | CN_POPULATION_PENDING); 3319 } 3320 } else { 3321 cp->c_flags &= 3322 ~(CN_NEED_FRONT_SYNC | CN_POPULATION_PENDING); 3323 } 3324 } 3325 3326 /* 3327 * XXX tony: How can CN_ALLOC_PENDING still be set?? 3328 * XXX tony: How can CN_UPDATED not be set????? 3329 */ 3330 if ((cp->c_flags & CN_ALLOC_PENDING) == 0 && 3331 (cp->c_flags & CN_UPDATED)) { 3332 error = filegrp_write_metadata(fgp, &cp->c_id, 3333 &cp->c_metadata); 3334 if (error) 3335 goto out; 3336 } 3337 out: 3338 if (error) { 3339 /* XXX modified files? */ 3340 if (cp->c_metadata.md_rlno) { 3341 cachefs_removefrontfile(&cp->c_metadata, 3342 &cp->c_id, fgp); 3343 cachefs_rlent_moveto(C_TO_FSCACHE(cp)->fs_cache, 3344 CACHEFS_RL_FREE, cp->c_metadata.md_rlno, 0); 3345 cp->c_metadata.md_rlno = 0; 3346 cp->c_metadata.md_rltype = CACHEFS_RL_NONE; 3347 if (cp->c_frontvp) { 3348 VN_RELE(cp->c_frontvp); 3349 cp->c_frontvp = NULL; 3350 } 3351 } 3352 if ((cp->c_flags & CN_ALLOC_PENDING) == 0) 3353 (void) filegrp_destroy_metadata(fgp, &cp->c_id); 3354 cp->c_flags |= CN_ALLOC_PENDING; 3355 cachefs_nocache(cp); 3356 } 3357 /* 3358 * we clear the updated bit even on errors because a retry 3359 * will probably fail also. 3360 */ 3361 cp->c_flags &= ~CN_UPDATED; 3362 mutex_exit(&cp->c_statelock); 3363 3364 #ifdef CFSDEBUG 3365 CFS_DEBUG(CFSDEBUG_VOPS) 3366 printf("c_sync_metadata: EXIT cp %p cflag %x\n", 3367 (void *)cp, cp->c_flags); 3368 #endif 3369 3370 return (error); 3371 } 3372 3373 /* 3374 * This is the vop entry point for inactivating a vnode. 3375 * It just queues the request for the async thread which 3376 * calls cachefs_inactive. 3377 * Because of the dnlc, it is not safe to grab most locks here. 3378 */ 3379 static void 3380 cachefs_inactive(struct vnode *vp, cred_t *cr) 3381 { 3382 cnode_t *cp; 3383 struct cachefs_req *rp; 3384 fscache_t *fscp; 3385 3386 #ifdef CFSDEBUG 3387 CFS_DEBUG(CFSDEBUG_VOPS) 3388 printf("cachefs_inactive: ENTER vp %p\n", (void *)vp); 3389 #endif 3390 3391 cp = VTOC(vp); 3392 fscp = C_TO_FSCACHE(cp); 3393 3394 ASSERT((cp->c_flags & CN_IDLE) == 0); 3395 3396 /* 3397 * Cachefs only provides pass-through support for NFSv4, 3398 * and all vnode operations are passed through to the 3399 * back file system. For NFSv4 pass-through to work, only 3400 * connected operation is supported, the cnode backvp must 3401 * exist, and cachefs optional (eg., disconnectable) flags 3402 * are turned off. Assert these conditions to ensure that 3403 * the backfilesystem is called for the inactive operation. 3404 */ 3405 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 3406 CFS_BACKFS_NFSV4_ASSERT_CNODE(cp); 3407 3408 /* vn_rele() set the v_count == 1 */ 3409 3410 cp->c_ipending = 1; 3411 3412 rp = kmem_cache_alloc(cachefs_req_cache, KM_SLEEP); 3413 rp->cfs_cmd = CFS_IDLE; 3414 rp->cfs_cr = cr; 3415 crhold(rp->cfs_cr); 3416 rp->cfs_req_u.cu_idle.ci_vp = vp; 3417 cachefs_addqueue(rp, &(C_TO_FSCACHE(cp)->fs_workq)); 3418 3419 #ifdef CFSDEBUG 3420 CFS_DEBUG(CFSDEBUG_VOPS) 3421 printf("cachefs_inactive: EXIT vp %p\n", (void *)vp); 3422 #endif 3423 } 3424 3425 /* ARGSUSED */ 3426 static int 3427 cachefs_lookup(vnode_t *dvp, char *nm, vnode_t **vpp, 3428 struct pathname *pnp, int flags, vnode_t *rdir, cred_t *cr) 3429 { 3430 int error = 0; 3431 cnode_t *dcp = VTOC(dvp); 3432 fscache_t *fscp = C_TO_FSCACHE(dcp); 3433 int held = 0; 3434 int connected = 0; 3435 3436 #ifdef CFSDEBUG 3437 CFS_DEBUG(CFSDEBUG_VOPS) 3438 printf("cachefs_lookup: ENTER dvp %p nm %s\n", (void *)dvp, nm); 3439 #endif 3440 3441 if (getzoneid() != GLOBAL_ZONEID) { 3442 error = EPERM; 3443 goto out; 3444 } 3445 3446 /* 3447 * Cachefs only provides pass-through support for NFSv4, 3448 * and all vnode operations are passed through to the 3449 * back file system. For NFSv4 pass-through to work, only 3450 * connected operation is supported, the cnode backvp must 3451 * exist, and cachefs optional (eg., disconnectable) flags 3452 * are turned off. Assert these conditions to ensure that 3453 * the backfilesystem is called for the lookup operation. 3454 */ 3455 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 3456 CFS_BACKFS_NFSV4_ASSERT_CNODE(dcp); 3457 3458 for (;;) { 3459 /* get (or renew) access to the file system */ 3460 if (held) { 3461 /* Won't loop with NFSv4 connected behavior */ 3462 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 3463 cachefs_cd_release(fscp); 3464 held = 0; 3465 } 3466 error = cachefs_cd_access(fscp, connected, 0); 3467 if (error) 3468 break; 3469 held = 1; 3470 3471 error = cachefs_lookup_common(dvp, nm, vpp, pnp, 3472 flags, rdir, cr); 3473 if (CFS_TIMEOUT(fscp, error)) { 3474 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 3475 cachefs_cd_release(fscp); 3476 held = 0; 3477 cachefs_cd_timedout(fscp); 3478 connected = 0; 3479 continue; 3480 } else { 3481 if (cachefs_cd_access_miss(fscp)) { 3482 rw_enter(&dcp->c_rwlock, RW_READER); 3483 error = cachefs_lookup_back(dvp, nm, 3484 vpp, cr); 3485 rw_exit(&dcp->c_rwlock); 3486 if (!CFS_TIMEOUT(fscp, error)) 3487 break; 3488 delay(5*hz); 3489 connected = 0; 3490 continue; 3491 } 3492 connected = 1; 3493 continue; 3494 } 3495 } 3496 break; 3497 } 3498 if (held) 3499 cachefs_cd_release(fscp); 3500 3501 if (error == 0 && IS_DEVVP(*vpp)) { 3502 struct vnode *newvp; 3503 newvp = specvp(*vpp, (*vpp)->v_rdev, (*vpp)->v_type, cr); 3504 VN_RELE(*vpp); 3505 if (newvp == NULL) { 3506 error = ENOSYS; 3507 } else { 3508 *vpp = newvp; 3509 } 3510 } 3511 3512 #ifdef CFS_CD_DEBUG 3513 ASSERT((curthread->t_flag & T_CD_HELD) == 0); 3514 #endif 3515 out: 3516 #ifdef CFSDEBUG 3517 CFS_DEBUG(CFSDEBUG_VOPS) 3518 printf("cachefs_lookup: EXIT error = %d\n", error); 3519 #endif 3520 3521 return (error); 3522 } 3523 3524 /* ARGSUSED */ 3525 int 3526 cachefs_lookup_common(vnode_t *dvp, char *nm, vnode_t **vpp, 3527 struct pathname *pnp, int flags, vnode_t *rdir, cred_t *cr) 3528 { 3529 int error = 0; 3530 cnode_t *cp, *dcp = VTOC(dvp); 3531 fscache_t *fscp = C_TO_FSCACHE(dcp); 3532 struct fid cookie; 3533 u_offset_t d_offset; 3534 struct cachefs_req *rp; 3535 cfs_cid_t cid, dircid; 3536 uint_t flag; 3537 uint_t uncached = 0; 3538 3539 *vpp = NULL; 3540 3541 /* 3542 * If lookup is for "", just return dvp. Don't need 3543 * to send it over the wire, look it up in the dnlc, 3544 * or perform any access checks. 3545 */ 3546 if (*nm == '\0') { 3547 VN_HOLD(dvp); 3548 *vpp = dvp; 3549 return (0); 3550 } 3551 3552 /* can't do lookups in non-directories */ 3553 if (dvp->v_type != VDIR) 3554 return (ENOTDIR); 3555 3556 /* perform access check, also does consistency check if connected */ 3557 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 3558 error = cachefs_access_connected(dvp, VEXEC, 0, cr); 3559 } else { 3560 mutex_enter(&dcp->c_statelock); 3561 error = cachefs_access_local(dcp, VEXEC, cr); 3562 mutex_exit(&dcp->c_statelock); 3563 } 3564 if (error) 3565 return (error); 3566 3567 /* 3568 * If lookup is for ".", just return dvp. Don't need 3569 * to send it over the wire or look it up in the dnlc, 3570 * just need to check access. 3571 */ 3572 if (strcmp(nm, ".") == 0) { 3573 VN_HOLD(dvp); 3574 *vpp = dvp; 3575 return (0); 3576 } 3577 3578 /* check the dnlc */ 3579 *vpp = (vnode_t *)dnlc_lookup(dvp, nm); 3580 if (*vpp) 3581 return (0); 3582 3583 /* read lock the dir before starting the search */ 3584 rw_enter(&dcp->c_rwlock, RW_READER); 3585 3586 mutex_enter(&dcp->c_statelock); 3587 dircid = dcp->c_id; 3588 3589 dcp->c_usage++; 3590 3591 /* if front file is not usable, lookup on the back fs */ 3592 if ((dcp->c_flags & (CN_NOCACHE | CN_ASYNC_POPULATE)) || 3593 CFS_ISFS_BACKFS_NFSV4(fscp) || 3594 ((dcp->c_filegrp->fg_flags & CFS_FG_READ) == 0)) { 3595 mutex_exit(&dcp->c_statelock); 3596 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) 3597 error = cachefs_lookup_back(dvp, nm, vpp, cr); 3598 else 3599 error = ETIMEDOUT; 3600 goto out; 3601 } 3602 3603 /* if the front file is not populated, try to populate it */ 3604 if ((dcp->c_metadata.md_flags & MD_POPULATED) == 0) { 3605 if (fscp->fs_cdconnected != CFS_CD_CONNECTED) { 3606 error = ETIMEDOUT; 3607 mutex_exit(&dcp->c_statelock); 3608 goto out; 3609 } 3610 3611 if (cachefs_async_okay()) { 3612 /* cannot populate if cache is not writable */ 3613 ASSERT((dcp->c_flags & 3614 (CN_ASYNC_POPULATE | CN_NOCACHE)) == 0); 3615 dcp->c_flags |= CN_ASYNC_POPULATE; 3616 3617 rp = kmem_cache_alloc(cachefs_req_cache, KM_SLEEP); 3618 rp->cfs_cmd = CFS_POPULATE; 3619 rp->cfs_req_u.cu_populate.cpop_vp = dvp; 3620 rp->cfs_cr = cr; 3621 3622 crhold(cr); 3623 VN_HOLD(dvp); 3624 3625 cachefs_addqueue(rp, &fscp->fs_workq); 3626 } else if (fscp->fs_info.fi_mntflags & CFS_NOACL) { 3627 error = cachefs_dir_fill(dcp, cr); 3628 if (error != 0) { 3629 mutex_exit(&dcp->c_statelock); 3630 goto out; 3631 } 3632 } 3633 /* no populate if too many asyncs and we have to cache ACLs */ 3634 3635 mutex_exit(&dcp->c_statelock); 3636 3637 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) 3638 error = cachefs_lookup_back(dvp, nm, vpp, cr); 3639 else 3640 error = ETIMEDOUT; 3641 goto out; 3642 } 3643 3644 /* by now we have a valid cached front file that we can search */ 3645 3646 ASSERT((dcp->c_flags & CN_ASYNC_POPULATE) == 0); 3647 error = cachefs_dir_look(dcp, nm, &cookie, &flag, 3648 &d_offset, &cid); 3649 mutex_exit(&dcp->c_statelock); 3650 3651 if (error) { 3652 /* if the entry does not have the fid, go get it */ 3653 if (error == EINVAL) { 3654 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) 3655 error = cachefs_lookup_back(dvp, nm, vpp, cr); 3656 else 3657 error = ETIMEDOUT; 3658 } 3659 3660 /* errors other than does not exist */ 3661 else if (error != ENOENT) { 3662 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) 3663 error = cachefs_lookup_back(dvp, nm, vpp, cr); 3664 else 3665 error = ETIMEDOUT; 3666 } 3667 goto out; 3668 } 3669 3670 /* 3671 * Else we found the entry in the cached directory. 3672 * Make a cnode for it. 3673 */ 3674 error = cachefs_cnode_make(&cid, fscp, &cookie, NULL, NULL, 3675 cr, 0, &cp); 3676 if (error == ESTALE) { 3677 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 3678 mutex_enter(&dcp->c_statelock); 3679 cachefs_nocache(dcp); 3680 mutex_exit(&dcp->c_statelock); 3681 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 3682 error = cachefs_lookup_back(dvp, nm, vpp, cr); 3683 uncached = 1; 3684 } else 3685 error = ETIMEDOUT; 3686 } else if (error == 0) { 3687 *vpp = CTOV(cp); 3688 } 3689 3690 out: 3691 if (error == 0) { 3692 /* put the entry in the dnlc */ 3693 if (cachefs_dnlc) 3694 dnlc_enter(dvp, nm, *vpp); 3695 3696 /* save the cid of the parent so can find the name */ 3697 cp = VTOC(*vpp); 3698 if (bcmp(&cp->c_metadata.md_parent, &dircid, 3699 sizeof (cfs_cid_t)) != 0) { 3700 mutex_enter(&cp->c_statelock); 3701 cp->c_metadata.md_parent = dircid; 3702 cp->c_flags |= CN_UPDATED; 3703 mutex_exit(&cp->c_statelock); 3704 } 3705 } 3706 3707 rw_exit(&dcp->c_rwlock); 3708 if (uncached && dcp->c_metadata.md_flags & MD_PACKED) 3709 (void) cachefs_pack_common(dvp, cr); 3710 return (error); 3711 } 3712 3713 /* 3714 * Called from cachefs_lookup_common when the back file system needs to be 3715 * examined to perform the lookup. 3716 */ 3717 static int 3718 cachefs_lookup_back(vnode_t *dvp, char *nm, vnode_t **vpp, 3719 cred_t *cr) 3720 { 3721 int error = 0; 3722 cnode_t *cp, *dcp = VTOC(dvp); 3723 fscache_t *fscp = C_TO_FSCACHE(dcp); 3724 vnode_t *backvp = NULL; 3725 struct vattr va; 3726 struct fid cookie; 3727 cfs_cid_t cid; 3728 uint32_t valid_fid; 3729 3730 mutex_enter(&dcp->c_statelock); 3731 3732 /* do a lookup on the back FS to get the back vnode */ 3733 if (dcp->c_backvp == NULL) { 3734 error = cachefs_getbackvp(fscp, dcp); 3735 if (error) 3736 goto out; 3737 } 3738 3739 CFS_DPRINT_BACKFS_NFSV4(fscp, 3740 ("cachefs_lookup (nfsv4): dcp %p, dbackvp %p, name %s\n", 3741 dcp, dcp->c_backvp, nm)); 3742 error = VOP_LOOKUP(dcp->c_backvp, nm, &backvp, (struct pathname *)NULL, 3743 0, (vnode_t *)NULL, cr); 3744 if (error) 3745 goto out; 3746 if (IS_DEVVP(backvp)) { 3747 struct vnode *devvp = backvp; 3748 3749 if (VOP_REALVP(devvp, &backvp) == 0) { 3750 VN_HOLD(backvp); 3751 VN_RELE(devvp); 3752 } 3753 } 3754 3755 /* get the fid and attrs from the back fs */ 3756 valid_fid = (CFS_ISFS_BACKFS_NFSV4(fscp) ? FALSE : TRUE); 3757 error = cachefs_getcookie(backvp, &cookie, &va, cr, valid_fid); 3758 if (error) 3759 goto out; 3760 3761 cid.cid_fileno = va.va_nodeid; 3762 cid.cid_flags = 0; 3763 3764 #if 0 3765 /* XXX bob: this is probably no longer necessary */ 3766 /* if the directory entry was incomplete, we can complete it now */ 3767 if ((dcp->c_metadata.md_flags & MD_POPULATED) && 3768 ((dcp->c_flags & CN_ASYNC_POPULATE) == 0) && 3769 (dcp->c_filegrp->fg_flags & CFS_FG_WRITE)) { 3770 cachefs_dir_modentry(dcp, d_offset, &cookie, &cid); 3771 } 3772 #endif 3773 3774 out: 3775 mutex_exit(&dcp->c_statelock); 3776 3777 /* create the cnode */ 3778 if (error == 0) { 3779 error = cachefs_cnode_make(&cid, fscp, 3780 (valid_fid ? &cookie : NULL), 3781 &va, backvp, cr, 0, &cp); 3782 if (error == 0) { 3783 *vpp = CTOV(cp); 3784 } 3785 } 3786 3787 if (backvp) 3788 VN_RELE(backvp); 3789 3790 return (error); 3791 } 3792 3793 /*ARGSUSED7*/ 3794 static int 3795 cachefs_create(vnode_t *dvp, char *nm, vattr_t *vap, 3796 vcexcl_t exclusive, int mode, vnode_t **vpp, cred_t *cr, int flag) 3797 { 3798 cnode_t *dcp = VTOC(dvp); 3799 fscache_t *fscp = C_TO_FSCACHE(dcp); 3800 cachefscache_t *cachep = fscp->fs_cache; 3801 int error; 3802 int connected = 0; 3803 int held = 0; 3804 3805 #ifdef CFSDEBUG 3806 CFS_DEBUG(CFSDEBUG_VOPS) 3807 printf("cachefs_create: ENTER dvp %p, nm %s\n", 3808 (void *)dvp, nm); 3809 #endif 3810 if (getzoneid() != GLOBAL_ZONEID) { 3811 error = EPERM; 3812 goto out; 3813 } 3814 3815 /* 3816 * Cachefs only provides pass-through support for NFSv4, 3817 * and all vnode operations are passed through to the 3818 * back file system. For NFSv4 pass-through to work, only 3819 * connected operation is supported, the cnode backvp must 3820 * exist, and cachefs optional (eg., disconnectable) flags 3821 * are turned off. Assert these conditions to ensure that 3822 * the backfilesystem is called for the create operation. 3823 */ 3824 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 3825 CFS_BACKFS_NFSV4_ASSERT_CNODE(dcp); 3826 3827 for (;;) { 3828 /* get (or renew) access to the file system */ 3829 if (held) { 3830 /* Won't loop with NFSv4 connected behavior */ 3831 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 3832 cachefs_cd_release(fscp); 3833 held = 0; 3834 } 3835 error = cachefs_cd_access(fscp, connected, 1); 3836 if (error) 3837 break; 3838 held = 1; 3839 3840 /* 3841 * if we are connected, perform the remote portion of the 3842 * create. 3843 */ 3844 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 3845 error = cachefs_create_connected(dvp, nm, vap, 3846 exclusive, mode, vpp, cr); 3847 if (CFS_TIMEOUT(fscp, error)) { 3848 cachefs_cd_release(fscp); 3849 held = 0; 3850 cachefs_cd_timedout(fscp); 3851 connected = 0; 3852 continue; 3853 } else if (error) { 3854 break; 3855 } 3856 } 3857 3858 /* else we must be disconnected */ 3859 else { 3860 error = cachefs_create_disconnected(dvp, nm, vap, 3861 exclusive, mode, vpp, cr); 3862 if (CFS_TIMEOUT(fscp, error)) { 3863 connected = 1; 3864 continue; 3865 } else if (error) { 3866 break; 3867 } 3868 } 3869 break; 3870 } 3871 3872 if (CACHEFS_LOG_LOGGING(cachep, CACHEFS_LOG_CREATE)) { 3873 fid_t *fidp = NULL; 3874 ino64_t fileno = 0; 3875 cnode_t *cp = NULL; 3876 if (error == 0) 3877 cp = VTOC(*vpp); 3878 3879 if (cp != NULL) { 3880 fidp = &cp->c_metadata.md_cookie; 3881 fileno = cp->c_id.cid_fileno; 3882 } 3883 cachefs_log_create(cachep, error, fscp->fs_cfsvfsp, 3884 fidp, fileno, crgetuid(cr)); 3885 } 3886 3887 if (held) 3888 cachefs_cd_release(fscp); 3889 3890 if (error == 0 && CFS_ISFS_NONSHARED(fscp)) 3891 (void) cachefs_pack(dvp, nm, cr); 3892 if (error == 0 && IS_DEVVP(*vpp)) { 3893 struct vnode *spcvp; 3894 3895 spcvp = specvp(*vpp, (*vpp)->v_rdev, (*vpp)->v_type, cr); 3896 VN_RELE(*vpp); 3897 if (spcvp == NULL) { 3898 error = ENOSYS; 3899 } else { 3900 *vpp = spcvp; 3901 } 3902 } 3903 3904 #ifdef CFS_CD_DEBUG 3905 ASSERT((curthread->t_flag & T_CD_HELD) == 0); 3906 #endif 3907 out: 3908 #ifdef CFSDEBUG 3909 CFS_DEBUG(CFSDEBUG_VOPS) 3910 printf("cachefs_create: EXIT error %d\n", error); 3911 #endif 3912 return (error); 3913 } 3914 3915 3916 static int 3917 cachefs_create_connected(vnode_t *dvp, char *nm, vattr_t *vap, 3918 enum vcexcl exclusive, int mode, vnode_t **vpp, cred_t *cr) 3919 { 3920 cnode_t *dcp = VTOC(dvp); 3921 fscache_t *fscp = C_TO_FSCACHE(dcp); 3922 int error; 3923 vnode_t *tvp = NULL; 3924 vnode_t *devvp; 3925 fid_t cookie; 3926 vattr_t va; 3927 cnode_t *ncp; 3928 cfs_cid_t cid; 3929 vnode_t *vp; 3930 uint32_t valid_fid; 3931 3932 /* special case if file already exists */ 3933 error = cachefs_lookup_common(dvp, nm, &vp, NULL, 0, NULL, cr); 3934 if (CFS_TIMEOUT(fscp, error)) 3935 return (error); 3936 if (error == 0) { 3937 if (exclusive == EXCL) 3938 error = EEXIST; 3939 else if (vp->v_type == VDIR && (mode & VWRITE)) 3940 error = EISDIR; 3941 else if ((error = 3942 cachefs_access_connected(vp, mode, 0, cr)) == 0) { 3943 if ((vap->va_mask & AT_SIZE) && (vp->v_type == VREG)) { 3944 vap->va_mask = AT_SIZE; 3945 error = 3946 cachefs_setattr_common(vp, vap, 3947 0, cr, NULL); 3948 } 3949 } 3950 if (error) { 3951 VN_RELE(vp); 3952 } else 3953 *vpp = vp; 3954 return (error); 3955 } 3956 3957 rw_enter(&dcp->c_rwlock, RW_WRITER); 3958 mutex_enter(&dcp->c_statelock); 3959 3960 /* consistency check the directory */ 3961 error = CFSOP_CHECK_COBJECT(fscp, dcp, 0, cr); 3962 if (error) { 3963 mutex_exit(&dcp->c_statelock); 3964 goto out; 3965 } 3966 3967 /* get the backvp if necessary */ 3968 if (dcp->c_backvp == NULL) { 3969 error = cachefs_getbackvp(fscp, dcp); 3970 if (error) { 3971 mutex_exit(&dcp->c_statelock); 3972 goto out; 3973 } 3974 } 3975 3976 /* create the file on the back fs */ 3977 CFS_DPRINT_BACKFS_NFSV4(fscp, 3978 ("cachefs_create (nfsv4): dcp %p, dbackvp %p," 3979 "name %s\n", dcp, dcp->c_backvp, nm)); 3980 error = VOP_CREATE(dcp->c_backvp, nm, vap, exclusive, mode, 3981 &devvp, cr, 0); 3982 mutex_exit(&dcp->c_statelock); 3983 if (error) 3984 goto out; 3985 if (VOP_REALVP(devvp, &tvp) == 0) { 3986 VN_HOLD(tvp); 3987 VN_RELE(devvp); 3988 } else { 3989 tvp = devvp; 3990 } 3991 3992 /* get the fid and attrs from the back fs */ 3993 valid_fid = (CFS_ISFS_BACKFS_NFSV4(fscp) ? FALSE : TRUE); 3994 error = cachefs_getcookie(tvp, &cookie, &va, cr, valid_fid); 3995 if (error) 3996 goto out; 3997 3998 /* make the cnode */ 3999 cid.cid_fileno = va.va_nodeid; 4000 cid.cid_flags = 0; 4001 error = cachefs_cnode_make(&cid, fscp, (valid_fid ? &cookie : NULL), 4002 &va, tvp, cr, 0, &ncp); 4003 if (error) 4004 goto out; 4005 4006 *vpp = CTOV(ncp); 4007 4008 /* enter it in the parent directory */ 4009 mutex_enter(&dcp->c_statelock); 4010 if (CFS_ISFS_NONSHARED(fscp) && 4011 (dcp->c_metadata.md_flags & MD_POPULATED)) { 4012 /* see if entry already exists */ 4013 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 4014 error = cachefs_dir_look(dcp, nm, NULL, NULL, NULL, NULL); 4015 if (error == ENOENT) { 4016 /* entry, does not exist, add the new file */ 4017 error = cachefs_dir_enter(dcp, nm, &ncp->c_cookie, 4018 &ncp->c_id, SM_ASYNC); 4019 if (error) { 4020 cachefs_nocache(dcp); 4021 error = 0; 4022 } 4023 /* XXX should this be done elsewhere, too? */ 4024 dnlc_enter(dvp, nm, *vpp); 4025 } else { 4026 /* entry exists or some other problem */ 4027 cachefs_nocache(dcp); 4028 error = 0; 4029 } 4030 } 4031 CFSOP_MODIFY_COBJECT(fscp, dcp, cr); 4032 mutex_exit(&dcp->c_statelock); 4033 4034 out: 4035 rw_exit(&dcp->c_rwlock); 4036 if (tvp) 4037 VN_RELE(tvp); 4038 4039 return (error); 4040 } 4041 4042 static int 4043 cachefs_create_disconnected(vnode_t *dvp, char *nm, vattr_t *vap, 4044 enum vcexcl exclusive, int mode, vnode_t **vpp, cred_t *cr) 4045 { 4046 cnode_t *dcp = VTOC(dvp); 4047 cnode_t *cp; 4048 cnode_t *ncp = NULL; 4049 vnode_t *vp; 4050 fscache_t *fscp = C_TO_FSCACHE(dcp); 4051 int error = 0; 4052 struct vattr va; 4053 timestruc_t current_time; 4054 off_t commit = 0; 4055 fid_t cookie; 4056 cfs_cid_t cid; 4057 4058 rw_enter(&dcp->c_rwlock, RW_WRITER); 4059 mutex_enter(&dcp->c_statelock); 4060 4061 /* give up if the directory is not populated */ 4062 if ((dcp->c_metadata.md_flags & MD_POPULATED) == 0) { 4063 mutex_exit(&dcp->c_statelock); 4064 rw_exit(&dcp->c_rwlock); 4065 return (ETIMEDOUT); 4066 } 4067 4068 /* special case if file already exists */ 4069 error = cachefs_dir_look(dcp, nm, &cookie, NULL, NULL, &cid); 4070 if (error == EINVAL) { 4071 mutex_exit(&dcp->c_statelock); 4072 rw_exit(&dcp->c_rwlock); 4073 return (ETIMEDOUT); 4074 } 4075 if (error == 0) { 4076 mutex_exit(&dcp->c_statelock); 4077 rw_exit(&dcp->c_rwlock); 4078 error = cachefs_cnode_make(&cid, fscp, &cookie, NULL, NULL, 4079 cr, 0, &cp); 4080 if (error) { 4081 return (error); 4082 } 4083 vp = CTOV(cp); 4084 4085 if (cp->c_metadata.md_flags & MD_NEEDATTRS) 4086 error = ETIMEDOUT; 4087 else if (exclusive == EXCL) 4088 error = EEXIST; 4089 else if (vp->v_type == VDIR && (mode & VWRITE)) 4090 error = EISDIR; 4091 else { 4092 mutex_enter(&cp->c_statelock); 4093 error = cachefs_access_local(cp, mode, cr); 4094 mutex_exit(&cp->c_statelock); 4095 if (!error) { 4096 if ((vap->va_mask & AT_SIZE) && 4097 (vp->v_type == VREG)) { 4098 vap->va_mask = AT_SIZE; 4099 error = 4100 cachefs_setattr_common(vp, vap, 4101 0, cr, NULL); 4102 } 4103 } 4104 } 4105 if (error) { 4106 VN_RELE(vp); 4107 } else 4108 *vpp = vp; 4109 return (error); 4110 } 4111 4112 /* give up if cannot modify the cache */ 4113 if (CFS_ISFS_WRITE_AROUND(fscp)) { 4114 mutex_exit(&dcp->c_statelock); 4115 error = ETIMEDOUT; 4116 goto out; 4117 } 4118 4119 /* check access */ 4120 if (error = cachefs_access_local(dcp, VWRITE, cr)) { 4121 mutex_exit(&dcp->c_statelock); 4122 goto out; 4123 } 4124 4125 /* mark dir as modified */ 4126 cachefs_modified(dcp); 4127 mutex_exit(&dcp->c_statelock); 4128 4129 /* must be privileged to set sticky bit */ 4130 if ((vap->va_mode & VSVTX) && secpolicy_vnode_stky_modify(cr) != 0) 4131 vap->va_mode &= ~VSVTX; 4132 4133 /* make up a reasonable set of attributes */ 4134 cachefs_attr_setup(vap, &va, dcp, cr); 4135 4136 /* create the cnode */ 4137 error = cachefs_cnode_create(fscp, &va, 0, &ncp); 4138 if (error) 4139 goto out; 4140 4141 mutex_enter(&ncp->c_statelock); 4142 4143 /* get the front file now instead of later */ 4144 if (vap->va_type == VREG) { 4145 error = cachefs_getfrontfile(ncp); 4146 if (error) { 4147 mutex_exit(&ncp->c_statelock); 4148 goto out; 4149 } 4150 ASSERT(ncp->c_frontvp != NULL); 4151 ASSERT((ncp->c_flags & CN_ALLOC_PENDING) == 0); 4152 ncp->c_metadata.md_flags |= MD_POPULATED; 4153 } else { 4154 ASSERT(ncp->c_flags & CN_ALLOC_PENDING); 4155 if (ncp->c_filegrp->fg_flags & CFS_FG_ALLOC_ATTR) { 4156 (void) filegrp_allocattr(ncp->c_filegrp); 4157 } 4158 error = filegrp_create_metadata(ncp->c_filegrp, 4159 &ncp->c_metadata, &ncp->c_id); 4160 if (error) { 4161 mutex_exit(&ncp->c_statelock); 4162 goto out; 4163 } 4164 ncp->c_flags &= ~CN_ALLOC_PENDING; 4165 } 4166 mutex_enter(&dcp->c_statelock); 4167 cachefs_creategid(dcp, ncp, vap, cr); 4168 cachefs_createacl(dcp, ncp); 4169 mutex_exit(&dcp->c_statelock); 4170 4171 /* set times on the file */ 4172 gethrestime(¤t_time); 4173 ncp->c_metadata.md_vattr.va_atime = current_time; 4174 ncp->c_metadata.md_localctime = current_time; 4175 ncp->c_metadata.md_localmtime = current_time; 4176 ncp->c_metadata.md_flags |= MD_LOCALMTIME | MD_LOCALCTIME; 4177 4178 /* reserve space for the daemon cid mapping */ 4179 error = cachefs_dlog_cidmap(fscp); 4180 if (error) { 4181 mutex_exit(&ncp->c_statelock); 4182 goto out; 4183 } 4184 ncp->c_metadata.md_flags |= MD_MAPPING; 4185 4186 /* mark the new file as modified */ 4187 if (cachefs_modified_alloc(ncp)) { 4188 mutex_exit(&ncp->c_statelock); 4189 error = ENOSPC; 4190 goto out; 4191 } 4192 ncp->c_flags |= CN_UPDATED; 4193 4194 /* 4195 * write the metadata now rather than waiting until 4196 * inactive so that if there's no space we can let 4197 * the caller know. 4198 */ 4199 ASSERT((ncp->c_flags & CN_ALLOC_PENDING) == 0); 4200 ASSERT((ncp->c_filegrp->fg_flags & CFS_FG_ALLOC_ATTR) == 0); 4201 error = filegrp_write_metadata(ncp->c_filegrp, 4202 &ncp->c_id, &ncp->c_metadata); 4203 if (error) { 4204 mutex_exit(&ncp->c_statelock); 4205 goto out; 4206 } 4207 4208 /* log the operation */ 4209 commit = cachefs_dlog_create(fscp, dcp, nm, vap, exclusive, 4210 mode, ncp, 0, cr); 4211 if (commit == 0) { 4212 mutex_exit(&ncp->c_statelock); 4213 error = ENOSPC; 4214 goto out; 4215 } 4216 4217 mutex_exit(&ncp->c_statelock); 4218 4219 mutex_enter(&dcp->c_statelock); 4220 4221 /* update parent dir times */ 4222 dcp->c_metadata.md_localmtime = current_time; 4223 dcp->c_metadata.md_flags |= MD_LOCALMTIME; 4224 dcp->c_flags |= CN_UPDATED; 4225 4226 /* enter new file name in the parent directory */ 4227 if (dcp->c_metadata.md_flags & MD_POPULATED) { 4228 error = cachefs_dir_enter(dcp, nm, &ncp->c_cookie, 4229 &ncp->c_id, 0); 4230 if (error) { 4231 cachefs_nocache(dcp); 4232 mutex_exit(&dcp->c_statelock); 4233 error = ETIMEDOUT; 4234 goto out; 4235 } 4236 dnlc_enter(dvp, nm, CTOV(ncp)); 4237 } else { 4238 mutex_exit(&dcp->c_statelock); 4239 error = ETIMEDOUT; 4240 goto out; 4241 } 4242 mutex_exit(&dcp->c_statelock); 4243 4244 out: 4245 rw_exit(&dcp->c_rwlock); 4246 4247 if (commit) { 4248 if (cachefs_dlog_commit(fscp, commit, error)) { 4249 /*EMPTY*/ 4250 /* XXX bob: fix on panic */ 4251 } 4252 } 4253 if (error) { 4254 /* destroy the cnode we created */ 4255 if (ncp) { 4256 mutex_enter(&ncp->c_statelock); 4257 ncp->c_flags |= CN_DESTROY; 4258 mutex_exit(&ncp->c_statelock); 4259 VN_RELE(CTOV(ncp)); 4260 } 4261 } else { 4262 *vpp = CTOV(ncp); 4263 } 4264 return (error); 4265 } 4266 4267 static int 4268 cachefs_remove(vnode_t *dvp, char *nm, cred_t *cr) 4269 { 4270 cnode_t *dcp = VTOC(dvp); 4271 fscache_t *fscp = C_TO_FSCACHE(dcp); 4272 cachefscache_t *cachep = fscp->fs_cache; 4273 int error = 0; 4274 int held = 0; 4275 int connected = 0; 4276 size_t namlen; 4277 vnode_t *vp = NULL; 4278 int vfslock = 0; 4279 4280 #ifdef CFSDEBUG 4281 CFS_DEBUG(CFSDEBUG_VOPS) 4282 printf("cachefs_remove: ENTER dvp %p name %s\n", 4283 (void *)dvp, nm); 4284 #endif 4285 if (getzoneid() != GLOBAL_ZONEID) { 4286 error = EPERM; 4287 goto out; 4288 } 4289 4290 if (fscp->fs_cache->c_flags & (CACHE_NOFILL | CACHE_NOCACHE)) 4291 ASSERT(dcp->c_flags & CN_NOCACHE); 4292 4293 /* 4294 * Cachefs only provides pass-through support for NFSv4, 4295 * and all vnode operations are passed through to the 4296 * back file system. For NFSv4 pass-through to work, only 4297 * connected operation is supported, the cnode backvp must 4298 * exist, and cachefs optional (eg., disconnectable) flags 4299 * are turned off. Assert these conditions to ensure that 4300 * the backfilesystem is called for the remove operation. 4301 */ 4302 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 4303 CFS_BACKFS_NFSV4_ASSERT_CNODE(dcp); 4304 4305 for (;;) { 4306 if (vfslock) { 4307 vn_vfsunlock(vp); 4308 vfslock = 0; 4309 } 4310 if (vp) { 4311 VN_RELE(vp); 4312 vp = NULL; 4313 } 4314 4315 /* get (or renew) access to the file system */ 4316 if (held) { 4317 /* Won't loop with NFSv4 connected behavior */ 4318 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 4319 cachefs_cd_release(fscp); 4320 held = 0; 4321 } 4322 error = cachefs_cd_access(fscp, connected, 1); 4323 if (error) 4324 break; 4325 held = 1; 4326 4327 /* if disconnected, do some extra error checking */ 4328 if (fscp->fs_cdconnected != CFS_CD_CONNECTED) { 4329 /* check permissions */ 4330 mutex_enter(&dcp->c_statelock); 4331 error = cachefs_access_local(dcp, (VEXEC|VWRITE), cr); 4332 mutex_exit(&dcp->c_statelock); 4333 if (CFS_TIMEOUT(fscp, error)) { 4334 connected = 1; 4335 continue; 4336 } 4337 if (error) 4338 break; 4339 4340 namlen = strlen(nm); 4341 if (namlen == 0) { 4342 error = EINVAL; 4343 break; 4344 } 4345 4346 /* cannot remove . and .. */ 4347 if (nm[0] == '.') { 4348 if (namlen == 1) { 4349 error = EINVAL; 4350 break; 4351 } else if (namlen == 2 && nm[1] == '.') { 4352 error = EEXIST; 4353 break; 4354 } 4355 } 4356 4357 } 4358 4359 /* get the cnode of the file to delete */ 4360 error = cachefs_lookup_common(dvp, nm, &vp, NULL, 0, NULL, cr); 4361 if (error) { 4362 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 4363 if (CFS_TIMEOUT(fscp, error)) { 4364 cachefs_cd_release(fscp); 4365 held = 0; 4366 cachefs_cd_timedout(fscp); 4367 connected = 0; 4368 continue; 4369 } 4370 } else { 4371 if (CFS_TIMEOUT(fscp, error)) { 4372 connected = 1; 4373 continue; 4374 } 4375 } 4376 if (CACHEFS_LOG_LOGGING(cachep, CACHEFS_LOG_REMOVE)) { 4377 struct fid foo; 4378 4379 bzero(&foo, sizeof (foo)); 4380 cachefs_log_remove(cachep, error, 4381 fscp->fs_cfsvfsp, &foo, 0, crgetuid(cr)); 4382 } 4383 break; 4384 } 4385 4386 if (vp->v_type == VDIR) { 4387 /* must be privileged to remove dirs with unlink() */ 4388 if ((error = secpolicy_fs_linkdir(cr, vp->v_vfsp)) != 0) 4389 break; 4390 4391 /* see ufs_dirremove for why this is done, mount race */ 4392 if (vn_vfslock(vp)) { 4393 error = EBUSY; 4394 break; 4395 } 4396 vfslock = 1; 4397 if (vn_mountedvfs(vp) != NULL) { 4398 error = EBUSY; 4399 break; 4400 } 4401 } 4402 4403 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 4404 error = cachefs_remove_connected(dvp, nm, cr, vp); 4405 if (CFS_TIMEOUT(fscp, error)) { 4406 cachefs_cd_release(fscp); 4407 held = 0; 4408 cachefs_cd_timedout(fscp); 4409 connected = 0; 4410 continue; 4411 } 4412 } else { 4413 error = cachefs_remove_disconnected(dvp, nm, cr, 4414 vp); 4415 if (CFS_TIMEOUT(fscp, error)) { 4416 connected = 1; 4417 continue; 4418 } 4419 } 4420 break; 4421 } 4422 4423 #if 0 4424 if (CACHEFS_LOG_LOGGING(cachep, CACHEFS_LOG_REMOVE)) 4425 cachefs_log_remove(cachep, error, fscp->fs_cfsvfsp, 4426 &cp->c_metadata.md_cookie, cp->c_id.cid_fileno, 4427 crgetuid(cr)); 4428 #endif 4429 4430 if (held) 4431 cachefs_cd_release(fscp); 4432 4433 if (vfslock) 4434 vn_vfsunlock(vp); 4435 4436 if (vp) 4437 VN_RELE(vp); 4438 4439 #ifdef CFS_CD_DEBUG 4440 ASSERT((curthread->t_flag & T_CD_HELD) == 0); 4441 #endif 4442 out: 4443 #ifdef CFSDEBUG 4444 CFS_DEBUG(CFSDEBUG_VOPS) 4445 printf("cachefs_remove: EXIT dvp %p\n", (void *)dvp); 4446 #endif 4447 4448 return (error); 4449 } 4450 4451 int 4452 cachefs_remove_connected(vnode_t *dvp, char *nm, cred_t *cr, vnode_t *vp) 4453 { 4454 cnode_t *dcp = VTOC(dvp); 4455 cnode_t *cp = VTOC(vp); 4456 fscache_t *fscp = C_TO_FSCACHE(dcp); 4457 int error = 0; 4458 4459 /* 4460 * Acquire the rwlock (WRITER) on the directory to prevent other 4461 * activity on the directory. 4462 */ 4463 rw_enter(&dcp->c_rwlock, RW_WRITER); 4464 4465 /* purge dnlc of this entry so can get accurate vnode count */ 4466 dnlc_purge_vp(vp); 4467 4468 /* 4469 * If the cnode is active, make a link to the file 4470 * so operations on the file will continue. 4471 */ 4472 if ((vp->v_type != VDIR) && 4473 !((vp->v_count == 1) || ((vp->v_count == 2) && cp->c_ipending))) { 4474 error = cachefs_remove_dolink(dvp, vp, nm, cr); 4475 if (error) 4476 goto out; 4477 } 4478 4479 /* else call backfs NFSv4 handler if NFSv4 */ 4480 else if (CFS_ISFS_BACKFS_NFSV4(fscp)) { 4481 error = cachefs_remove_backfs_nfsv4(dvp, nm, cr, vp); 4482 goto out; 4483 } 4484 4485 /* else drop the backvp so nfs does not do rename */ 4486 else if (cp->c_backvp) { 4487 mutex_enter(&cp->c_statelock); 4488 if (cp->c_backvp) { 4489 VN_RELE(cp->c_backvp); 4490 cp->c_backvp = NULL; 4491 } 4492 mutex_exit(&cp->c_statelock); 4493 } 4494 4495 mutex_enter(&dcp->c_statelock); 4496 4497 /* get the backvp */ 4498 if (dcp->c_backvp == NULL) { 4499 error = cachefs_getbackvp(fscp, dcp); 4500 if (error) { 4501 mutex_exit(&dcp->c_statelock); 4502 goto out; 4503 } 4504 } 4505 4506 /* check directory consistency */ 4507 error = CFSOP_CHECK_COBJECT(fscp, dcp, 0, cr); 4508 if (error) { 4509 mutex_exit(&dcp->c_statelock); 4510 goto out; 4511 } 4512 4513 /* perform the remove on the back fs */ 4514 error = VOP_REMOVE(dcp->c_backvp, nm, cr); 4515 if (error) { 4516 mutex_exit(&dcp->c_statelock); 4517 goto out; 4518 } 4519 4520 /* the dir has been modified */ 4521 CFSOP_MODIFY_COBJECT(fscp, dcp, cr); 4522 4523 /* remove the entry from the populated directory */ 4524 if (CFS_ISFS_NONSHARED(fscp) && 4525 (dcp->c_metadata.md_flags & MD_POPULATED)) { 4526 error = cachefs_dir_rmentry(dcp, nm); 4527 if (error) { 4528 cachefs_nocache(dcp); 4529 error = 0; 4530 } 4531 } 4532 mutex_exit(&dcp->c_statelock); 4533 4534 /* fix up the file we deleted */ 4535 mutex_enter(&cp->c_statelock); 4536 if (cp->c_attr.va_nlink == 1) 4537 cp->c_flags |= CN_DESTROY; 4538 else 4539 cp->c_flags |= CN_UPDATED; 4540 4541 cp->c_attr.va_nlink--; 4542 CFSOP_MODIFY_COBJECT(fscp, cp, cr); 4543 mutex_exit(&cp->c_statelock); 4544 4545 out: 4546 rw_exit(&dcp->c_rwlock); 4547 return (error); 4548 } 4549 4550 /* 4551 * cachefs_remove_backfs_nfsv4 4552 * 4553 * Call NFSv4 back filesystem to handle the remove (cachefs 4554 * pass-through support for NFSv4). 4555 */ 4556 int 4557 cachefs_remove_backfs_nfsv4(vnode_t *dvp, char *nm, cred_t *cr, vnode_t *vp) 4558 { 4559 cnode_t *dcp = VTOC(dvp); 4560 cnode_t *cp = VTOC(vp); 4561 vnode_t *dbackvp; 4562 fscache_t *fscp = C_TO_FSCACHE(dcp); 4563 int error = 0; 4564 4565 /* 4566 * For NFSv4 pass-through to work, only connected operation 4567 * is supported, the cnode backvp must exist, and cachefs 4568 * optional (eg., disconnectable) flags are turned off. Assert 4569 * these conditions for the getattr operation. 4570 */ 4571 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 4572 CFS_BACKFS_NFSV4_ASSERT_CNODE(cp); 4573 4574 /* Should hold the directory readwrite lock to update directory */ 4575 ASSERT(RW_WRITE_HELD(&dcp->c_rwlock)); 4576 4577 /* 4578 * Update attributes for directory. Note that 4579 * CFSOP_CHECK_COBJECT asserts for c_statelock being 4580 * held, so grab it before calling the routine. 4581 */ 4582 mutex_enter(&dcp->c_statelock); 4583 error = CFSOP_CHECK_COBJECT(fscp, dcp, 0, cr); 4584 mutex_exit(&dcp->c_statelock); 4585 if (error) 4586 goto out; 4587 4588 /* 4589 * Update attributes for cp. Note that CFSOP_CHECK_COBJECT 4590 * asserts for c_statelock being held, so grab it before 4591 * calling the routine. 4592 */ 4593 mutex_enter(&cp->c_statelock); 4594 error = CFSOP_CHECK_COBJECT(fscp, cp, 0, cr); 4595 if (error) { 4596 mutex_exit(&cp->c_statelock); 4597 goto out; 4598 } 4599 4600 /* 4601 * Drop the backvp so nfs if the link count is 1 so that 4602 * nfs does not do rename. Ensure that we will destroy the cnode 4603 * since this cnode no longer contains the backvp. Note that we 4604 * maintain lock on this cnode to prevent change till the remove 4605 * completes, otherwise other operations will encounter an ESTALE 4606 * if they try to use the cnode with CN_DESTROY set (see 4607 * cachefs_get_backvp()), or change the state of the cnode 4608 * while we're removing it. 4609 */ 4610 if (cp->c_attr.va_nlink == 1) { 4611 /* 4612 * The unldvp information is created for the case 4613 * when there is more than one reference on the 4614 * vnode when a remove operation is called. If the 4615 * remove itself was holding a reference to the 4616 * vnode, then a subsequent remove will remove the 4617 * backvp, so we need to get rid of the unldvp 4618 * before removing the backvp. An alternate would 4619 * be to simply ignore the remove and let the 4620 * inactivation routine do the deletion of the 4621 * unldvp. 4622 */ 4623 if (cp->c_unldvp) { 4624 VN_RELE(cp->c_unldvp); 4625 cachefs_kmem_free(cp->c_unlname, MAXNAMELEN); 4626 crfree(cp->c_unlcred); 4627 cp->c_unldvp = NULL; 4628 cp->c_unlcred = NULL; 4629 } 4630 cp->c_flags |= CN_DESTROY; 4631 cp->c_attr.va_nlink = 0; 4632 VN_RELE(cp->c_backvp); 4633 cp->c_backvp = NULL; 4634 } 4635 4636 /* perform the remove on back fs after extracting directory backvp */ 4637 mutex_enter(&dcp->c_statelock); 4638 dbackvp = dcp->c_backvp; 4639 mutex_exit(&dcp->c_statelock); 4640 4641 CFS_DPRINT_BACKFS_NFSV4(fscp, 4642 ("cachefs_remove (nfsv4): dcp %p, dbackvp %p, name %s\n", 4643 dcp, dbackvp, nm)); 4644 error = VOP_REMOVE(dbackvp, nm, cr); 4645 if (error) { 4646 mutex_exit(&cp->c_statelock); 4647 goto out; 4648 } 4649 4650 /* fix up the file we deleted, if not destroying the cnode */ 4651 if ((cp->c_flags & CN_DESTROY) == 0) { 4652 cp->c_attr.va_nlink--; 4653 cp->c_flags |= CN_UPDATED; 4654 } 4655 4656 mutex_exit(&cp->c_statelock); 4657 4658 out: 4659 return (error); 4660 } 4661 4662 int 4663 cachefs_remove_disconnected(vnode_t *dvp, char *nm, cred_t *cr, 4664 vnode_t *vp) 4665 { 4666 cnode_t *dcp = VTOC(dvp); 4667 cnode_t *cp = VTOC(vp); 4668 fscache_t *fscp = C_TO_FSCACHE(dcp); 4669 int error = 0; 4670 off_t commit = 0; 4671 timestruc_t current_time; 4672 4673 if (CFS_ISFS_WRITE_AROUND(fscp)) 4674 return (ETIMEDOUT); 4675 4676 if (cp->c_metadata.md_flags & MD_NEEDATTRS) 4677 return (ETIMEDOUT); 4678 4679 /* 4680 * Acquire the rwlock (WRITER) on the directory to prevent other 4681 * activity on the directory. 4682 */ 4683 rw_enter(&dcp->c_rwlock, RW_WRITER); 4684 4685 /* dir must be populated */ 4686 if ((dcp->c_metadata.md_flags & MD_POPULATED) == 0) { 4687 error = ETIMEDOUT; 4688 goto out; 4689 } 4690 4691 mutex_enter(&dcp->c_statelock); 4692 mutex_enter(&cp->c_statelock); 4693 4694 error = cachefs_stickyrmchk(dcp, cp, cr); 4695 4696 mutex_exit(&cp->c_statelock); 4697 mutex_exit(&dcp->c_statelock); 4698 if (error) 4699 goto out; 4700 4701 /* purge dnlc of this entry so can get accurate vnode count */ 4702 dnlc_purge_vp(vp); 4703 4704 /* 4705 * If the cnode is active, make a link to the file 4706 * so operations on the file will continue. 4707 */ 4708 if ((vp->v_type != VDIR) && 4709 !((vp->v_count == 1) || ((vp->v_count == 2) && cp->c_ipending))) { 4710 error = cachefs_remove_dolink(dvp, vp, nm, cr); 4711 if (error) 4712 goto out; 4713 } 4714 4715 if (cp->c_attr.va_nlink > 1) { 4716 mutex_enter(&cp->c_statelock); 4717 if (cachefs_modified_alloc(cp)) { 4718 mutex_exit(&cp->c_statelock); 4719 error = ENOSPC; 4720 goto out; 4721 } 4722 if ((cp->c_metadata.md_flags & MD_MAPPING) == 0) { 4723 error = cachefs_dlog_cidmap(fscp); 4724 if (error) { 4725 mutex_exit(&cp->c_statelock); 4726 error = ENOSPC; 4727 goto out; 4728 } 4729 cp->c_metadata.md_flags |= MD_MAPPING; 4730 cp->c_flags |= CN_UPDATED; 4731 } 4732 mutex_exit(&cp->c_statelock); 4733 } 4734 4735 /* log the remove */ 4736 commit = cachefs_dlog_remove(fscp, dcp, nm, cp, cr); 4737 if (commit == 0) { 4738 error = ENOSPC; 4739 goto out; 4740 } 4741 4742 /* remove the file from the dir */ 4743 mutex_enter(&dcp->c_statelock); 4744 if ((dcp->c_metadata.md_flags & MD_POPULATED) == 0) { 4745 mutex_exit(&dcp->c_statelock); 4746 error = ETIMEDOUT; 4747 goto out; 4748 4749 } 4750 cachefs_modified(dcp); 4751 error = cachefs_dir_rmentry(dcp, nm); 4752 if (error) { 4753 mutex_exit(&dcp->c_statelock); 4754 if (error == ENOTDIR) 4755 error = ETIMEDOUT; 4756 goto out; 4757 } 4758 4759 /* update parent dir times */ 4760 gethrestime(¤t_time); 4761 dcp->c_metadata.md_localctime = current_time; 4762 dcp->c_metadata.md_localmtime = current_time; 4763 dcp->c_metadata.md_flags |= MD_LOCALCTIME | MD_LOCALMTIME; 4764 dcp->c_flags |= CN_UPDATED; 4765 mutex_exit(&dcp->c_statelock); 4766 4767 /* adjust file we are deleting */ 4768 mutex_enter(&cp->c_statelock); 4769 cp->c_attr.va_nlink--; 4770 cp->c_metadata.md_localctime = current_time; 4771 cp->c_metadata.md_flags |= MD_LOCALCTIME; 4772 if (cp->c_attr.va_nlink == 0) { 4773 cp->c_flags |= CN_DESTROY; 4774 } else { 4775 cp->c_flags |= CN_UPDATED; 4776 } 4777 mutex_exit(&cp->c_statelock); 4778 4779 out: 4780 if (commit) { 4781 /* commit the log entry */ 4782 if (cachefs_dlog_commit(fscp, commit, error)) { 4783 /*EMPTY*/ 4784 /* XXX bob: fix on panic */ 4785 } 4786 } 4787 4788 rw_exit(&dcp->c_rwlock); 4789 return (error); 4790 } 4791 4792 static int 4793 cachefs_link(vnode_t *tdvp, vnode_t *fvp, char *tnm, cred_t *cr) 4794 { 4795 fscache_t *fscp = VFS_TO_FSCACHE(tdvp->v_vfsp); 4796 cnode_t *tdcp = VTOC(tdvp); 4797 struct vnode *realvp; 4798 int error = 0; 4799 int held = 0; 4800 int connected = 0; 4801 4802 #ifdef CFSDEBUG 4803 CFS_DEBUG(CFSDEBUG_VOPS) 4804 printf("cachefs_link: ENTER fvp %p tdvp %p tnm %s\n", 4805 (void *)fvp, (void *)tdvp, tnm); 4806 #endif 4807 4808 if (getzoneid() != GLOBAL_ZONEID) { 4809 error = EPERM; 4810 goto out; 4811 } 4812 4813 if (fscp->fs_cache->c_flags & (CACHE_NOFILL | CACHE_NOCACHE)) 4814 ASSERT(tdcp->c_flags & CN_NOCACHE); 4815 4816 if (VOP_REALVP(fvp, &realvp) == 0) { 4817 fvp = realvp; 4818 } 4819 4820 /* 4821 * Cachefs only provides pass-through support for NFSv4, 4822 * and all vnode operations are passed through to the 4823 * back file system. For NFSv4 pass-through to work, only 4824 * connected operation is supported, the cnode backvp must 4825 * exist, and cachefs optional (eg., disconnectable) flags 4826 * are turned off. Assert these conditions to ensure that 4827 * the backfilesystem is called for the link operation. 4828 */ 4829 4830 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 4831 CFS_BACKFS_NFSV4_ASSERT_CNODE(tdcp); 4832 4833 for (;;) { 4834 /* get (or renew) access to the file system */ 4835 if (held) { 4836 /* Won't loop with NFSv4 connected behavior */ 4837 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 4838 rw_exit(&tdcp->c_rwlock); 4839 cachefs_cd_release(fscp); 4840 held = 0; 4841 } 4842 error = cachefs_cd_access(fscp, connected, 1); 4843 if (error) 4844 break; 4845 rw_enter(&tdcp->c_rwlock, RW_WRITER); 4846 held = 1; 4847 4848 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 4849 error = cachefs_link_connected(tdvp, fvp, tnm, cr); 4850 if (CFS_TIMEOUT(fscp, error)) { 4851 rw_exit(&tdcp->c_rwlock); 4852 cachefs_cd_release(fscp); 4853 held = 0; 4854 cachefs_cd_timedout(fscp); 4855 connected = 0; 4856 continue; 4857 } 4858 } else { 4859 error = cachefs_link_disconnected(tdvp, fvp, tnm, 4860 cr); 4861 if (CFS_TIMEOUT(fscp, error)) { 4862 connected = 1; 4863 continue; 4864 } 4865 } 4866 break; 4867 } 4868 4869 if (held) { 4870 rw_exit(&tdcp->c_rwlock); 4871 cachefs_cd_release(fscp); 4872 } 4873 4874 #ifdef CFS_CD_DEBUG 4875 ASSERT((curthread->t_flag & T_CD_HELD) == 0); 4876 #endif 4877 out: 4878 #ifdef CFSDEBUG 4879 CFS_DEBUG(CFSDEBUG_VOPS) 4880 printf("cachefs_link: EXIT fvp %p tdvp %p tnm %s\n", 4881 (void *)fvp, (void *)tdvp, tnm); 4882 #endif 4883 return (error); 4884 } 4885 4886 static int 4887 cachefs_link_connected(vnode_t *tdvp, vnode_t *fvp, char *tnm, cred_t *cr) 4888 { 4889 cnode_t *tdcp = VTOC(tdvp); 4890 cnode_t *fcp = VTOC(fvp); 4891 fscache_t *fscp = VFS_TO_FSCACHE(tdvp->v_vfsp); 4892 int error = 0; 4893 vnode_t *backvp = NULL; 4894 4895 if (tdcp != fcp) { 4896 mutex_enter(&fcp->c_statelock); 4897 4898 if (fcp->c_backvp == NULL) { 4899 error = cachefs_getbackvp(fscp, fcp); 4900 if (error) { 4901 mutex_exit(&fcp->c_statelock); 4902 goto out; 4903 } 4904 } 4905 4906 error = CFSOP_CHECK_COBJECT(fscp, fcp, 0, cr); 4907 if (error) { 4908 mutex_exit(&fcp->c_statelock); 4909 goto out; 4910 } 4911 backvp = fcp->c_backvp; 4912 VN_HOLD(backvp); 4913 mutex_exit(&fcp->c_statelock); 4914 } 4915 4916 mutex_enter(&tdcp->c_statelock); 4917 4918 /* get backvp of target directory */ 4919 if (tdcp->c_backvp == NULL) { 4920 error = cachefs_getbackvp(fscp, tdcp); 4921 if (error) { 4922 mutex_exit(&tdcp->c_statelock); 4923 goto out; 4924 } 4925 } 4926 4927 /* consistency check target directory */ 4928 error = CFSOP_CHECK_COBJECT(fscp, tdcp, 0, cr); 4929 if (error) { 4930 mutex_exit(&tdcp->c_statelock); 4931 goto out; 4932 } 4933 if (backvp == NULL) { 4934 backvp = tdcp->c_backvp; 4935 VN_HOLD(backvp); 4936 } 4937 4938 /* perform the link on the back fs */ 4939 CFS_DPRINT_BACKFS_NFSV4(fscp, 4940 ("cachefs_link (nfsv4): tdcp %p, tdbackvp %p, " 4941 "name %s\n", tdcp, tdcp->c_backvp, tnm)); 4942 error = VOP_LINK(tdcp->c_backvp, backvp, tnm, cr); 4943 if (error) { 4944 mutex_exit(&tdcp->c_statelock); 4945 goto out; 4946 } 4947 4948 CFSOP_MODIFY_COBJECT(fscp, tdcp, cr); 4949 4950 /* if the dir is populated, add the new link */ 4951 if (CFS_ISFS_NONSHARED(fscp) && 4952 (tdcp->c_metadata.md_flags & MD_POPULATED)) { 4953 error = cachefs_dir_enter(tdcp, tnm, &fcp->c_cookie, 4954 &fcp->c_id, SM_ASYNC); 4955 if (error) { 4956 cachefs_nocache(tdcp); 4957 error = 0; 4958 } 4959 } 4960 mutex_exit(&tdcp->c_statelock); 4961 4962 /* get the new link count on the file */ 4963 mutex_enter(&fcp->c_statelock); 4964 fcp->c_flags |= CN_UPDATED; 4965 CFSOP_MODIFY_COBJECT(fscp, fcp, cr); 4966 if (fcp->c_backvp == NULL) { 4967 error = cachefs_getbackvp(fscp, fcp); 4968 if (error) { 4969 mutex_exit(&fcp->c_statelock); 4970 goto out; 4971 } 4972 } 4973 4974 /* XXX bob: given what modify_cobject does this seems unnecessary */ 4975 fcp->c_attr.va_mask = AT_ALL; 4976 error = VOP_GETATTR(fcp->c_backvp, &fcp->c_attr, 0, cr); 4977 mutex_exit(&fcp->c_statelock); 4978 out: 4979 if (backvp) 4980 VN_RELE(backvp); 4981 4982 return (error); 4983 } 4984 4985 static int 4986 cachefs_link_disconnected(vnode_t *tdvp, vnode_t *fvp, char *tnm, 4987 cred_t *cr) 4988 { 4989 cnode_t *tdcp = VTOC(tdvp); 4990 cnode_t *fcp = VTOC(fvp); 4991 fscache_t *fscp = VFS_TO_FSCACHE(tdvp->v_vfsp); 4992 int error = 0; 4993 timestruc_t current_time; 4994 off_t commit = 0; 4995 4996 if (fvp->v_type == VDIR && secpolicy_fs_linkdir(cr, fvp->v_vfsp) != 0 || 4997 fcp->c_attr.va_uid != crgetuid(cr) && secpolicy_basic_link(cr) != 0) 4998 return (EPERM); 4999 5000 if (CFS_ISFS_WRITE_AROUND(fscp)) 5001 return (ETIMEDOUT); 5002 5003 if (fcp->c_metadata.md_flags & MD_NEEDATTRS) 5004 return (ETIMEDOUT); 5005 5006 mutex_enter(&tdcp->c_statelock); 5007 5008 /* check permissions */ 5009 if (error = cachefs_access_local(tdcp, (VEXEC|VWRITE), cr)) { 5010 mutex_exit(&tdcp->c_statelock); 5011 goto out; 5012 } 5013 5014 /* the directory front file must be populated */ 5015 if ((tdcp->c_metadata.md_flags & MD_POPULATED) == 0) { 5016 error = ETIMEDOUT; 5017 mutex_exit(&tdcp->c_statelock); 5018 goto out; 5019 } 5020 5021 /* make sure tnm does not already exist in the directory */ 5022 error = cachefs_dir_look(tdcp, tnm, NULL, NULL, NULL, NULL); 5023 if (error == ENOTDIR) { 5024 error = ETIMEDOUT; 5025 mutex_exit(&tdcp->c_statelock); 5026 goto out; 5027 } 5028 if (error != ENOENT) { 5029 error = EEXIST; 5030 mutex_exit(&tdcp->c_statelock); 5031 goto out; 5032 } 5033 5034 mutex_enter(&fcp->c_statelock); 5035 5036 /* create a mapping for the file if necessary */ 5037 if ((fcp->c_metadata.md_flags & MD_MAPPING) == 0) { 5038 error = cachefs_dlog_cidmap(fscp); 5039 if (error) { 5040 mutex_exit(&fcp->c_statelock); 5041 mutex_exit(&tdcp->c_statelock); 5042 error = ENOSPC; 5043 goto out; 5044 } 5045 fcp->c_metadata.md_flags |= MD_MAPPING; 5046 fcp->c_flags |= CN_UPDATED; 5047 } 5048 5049 /* mark file as modified */ 5050 if (cachefs_modified_alloc(fcp)) { 5051 mutex_exit(&fcp->c_statelock); 5052 mutex_exit(&tdcp->c_statelock); 5053 error = ENOSPC; 5054 goto out; 5055 } 5056 mutex_exit(&fcp->c_statelock); 5057 5058 /* log the operation */ 5059 commit = cachefs_dlog_link(fscp, tdcp, tnm, fcp, cr); 5060 if (commit == 0) { 5061 mutex_exit(&tdcp->c_statelock); 5062 error = ENOSPC; 5063 goto out; 5064 } 5065 5066 gethrestime(¤t_time); 5067 5068 /* make the new link */ 5069 cachefs_modified(tdcp); 5070 error = cachefs_dir_enter(tdcp, tnm, &fcp->c_cookie, 5071 &fcp->c_id, SM_ASYNC); 5072 if (error) { 5073 error = 0; 5074 mutex_exit(&tdcp->c_statelock); 5075 goto out; 5076 } 5077 5078 /* Update mtime/ctime of parent dir */ 5079 tdcp->c_metadata.md_localmtime = current_time; 5080 tdcp->c_metadata.md_localctime = current_time; 5081 tdcp->c_metadata.md_flags |= MD_LOCALCTIME | MD_LOCALMTIME; 5082 tdcp->c_flags |= CN_UPDATED; 5083 mutex_exit(&tdcp->c_statelock); 5084 5085 /* update the file we linked to */ 5086 mutex_enter(&fcp->c_statelock); 5087 fcp->c_attr.va_nlink++; 5088 fcp->c_metadata.md_localctime = current_time; 5089 fcp->c_metadata.md_flags |= MD_LOCALCTIME; 5090 fcp->c_flags |= CN_UPDATED; 5091 mutex_exit(&fcp->c_statelock); 5092 5093 out: 5094 if (commit) { 5095 /* commit the log entry */ 5096 if (cachefs_dlog_commit(fscp, commit, error)) { 5097 /*EMPTY*/ 5098 /* XXX bob: fix on panic */ 5099 } 5100 } 5101 5102 return (error); 5103 } 5104 5105 /* 5106 * Serialize all renames in CFS, to avoid deadlocks - We have to hold two 5107 * cnodes atomically. 5108 */ 5109 kmutex_t cachefs_rename_lock; 5110 5111 static int 5112 cachefs_rename(vnode_t *odvp, char *onm, vnode_t *ndvp, 5113 char *nnm, cred_t *cr) 5114 { 5115 fscache_t *fscp = C_TO_FSCACHE(VTOC(odvp)); 5116 cachefscache_t *cachep = fscp->fs_cache; 5117 int error = 0; 5118 int held = 0; 5119 int connected = 0; 5120 vnode_t *delvp = NULL; 5121 vnode_t *tvp = NULL; 5122 int vfslock = 0; 5123 struct vnode *realvp; 5124 5125 if (getzoneid() != GLOBAL_ZONEID) 5126 return (EPERM); 5127 5128 if (VOP_REALVP(ndvp, &realvp) == 0) 5129 ndvp = realvp; 5130 5131 /* 5132 * if the fs NOFILL or NOCACHE flags are on, then the old and new 5133 * directory cnodes better indicate NOCACHE mode as well. 5134 */ 5135 ASSERT 5136 ((fscp->fs_cache->c_flags & (CACHE_NOFILL | CACHE_NOCACHE)) == 0 || 5137 ((VTOC(odvp)->c_flags & CN_NOCACHE) && 5138 (VTOC(ndvp)->c_flags & CN_NOCACHE))); 5139 5140 /* 5141 * Cachefs only provides pass-through support for NFSv4, 5142 * and all vnode operations are passed through to the 5143 * back file system. For NFSv4 pass-through to work, only 5144 * connected operation is supported, the cnode backvp must 5145 * exist, and cachefs optional (eg., disconnectable) flags 5146 * are turned off. Assert these conditions to ensure that 5147 * the backfilesystem is called for the rename operation. 5148 */ 5149 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 5150 CFS_BACKFS_NFSV4_ASSERT_CNODE(VTOC(odvp)); 5151 CFS_BACKFS_NFSV4_ASSERT_CNODE(VTOC(ndvp)); 5152 5153 for (;;) { 5154 if (vfslock) { 5155 vn_vfsunlock(delvp); 5156 vfslock = 0; 5157 } 5158 if (delvp) { 5159 VN_RELE(delvp); 5160 delvp = NULL; 5161 } 5162 5163 /* get (or renew) access to the file system */ 5164 if (held) { 5165 /* Won't loop for NFSv4 connected support */ 5166 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 5167 cachefs_cd_release(fscp); 5168 held = 0; 5169 } 5170 error = cachefs_cd_access(fscp, connected, 1); 5171 if (error) 5172 break; 5173 held = 1; 5174 5175 /* sanity check */ 5176 if ((odvp->v_type != VDIR) || (ndvp->v_type != VDIR)) { 5177 error = EINVAL; 5178 break; 5179 } 5180 5181 /* cannot rename from or to . or .. */ 5182 if (strcmp(onm, ".") == 0 || strcmp(onm, "..") == 0 || 5183 strcmp(nnm, ".") == 0 || strcmp(nnm, "..") == 0) { 5184 error = EINVAL; 5185 break; 5186 } 5187 5188 if (odvp != ndvp) { 5189 /* 5190 * if moving a directory, its notion 5191 * of ".." will change 5192 */ 5193 error = cachefs_lookup_common(odvp, onm, &tvp, 5194 NULL, 0, NULL, cr); 5195 if (error == 0) { 5196 ASSERT(tvp != NULL); 5197 if (tvp->v_type == VDIR) { 5198 cnode_t *cp = VTOC(tvp); 5199 5200 dnlc_remove(tvp, ".."); 5201 5202 mutex_enter(&cp->c_statelock); 5203 CFSOP_MODIFY_COBJECT(fscp, cp, cr); 5204 mutex_exit(&cp->c_statelock); 5205 } 5206 } else { 5207 tvp = NULL; 5208 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 5209 if (CFS_TIMEOUT(fscp, error)) { 5210 cachefs_cd_release(fscp); 5211 held = 0; 5212 cachefs_cd_timedout(fscp); 5213 connected = 0; 5214 continue; 5215 } 5216 } else { 5217 if (CFS_TIMEOUT(fscp, error)) { 5218 connected = 1; 5219 continue; 5220 } 5221 } 5222 break; 5223 } 5224 } 5225 5226 /* get the cnode if file being deleted */ 5227 error = cachefs_lookup_common(ndvp, nnm, &delvp, NULL, 0, 5228 NULL, cr); 5229 if (error) { 5230 delvp = NULL; 5231 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 5232 if (CFS_TIMEOUT(fscp, error)) { 5233 cachefs_cd_release(fscp); 5234 held = 0; 5235 cachefs_cd_timedout(fscp); 5236 connected = 0; 5237 continue; 5238 } 5239 } else { 5240 if (CFS_TIMEOUT(fscp, error)) { 5241 connected = 1; 5242 continue; 5243 } 5244 } 5245 if (error != ENOENT) 5246 break; 5247 } 5248 5249 if (delvp && delvp->v_type == VDIR) { 5250 /* see ufs_dirremove for why this is done, mount race */ 5251 if (vn_vfslock(delvp)) { 5252 error = EBUSY; 5253 break; 5254 } 5255 vfslock = 1; 5256 if (vn_mountedvfs(delvp) != NULL) { 5257 error = EBUSY; 5258 break; 5259 } 5260 } 5261 5262 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 5263 error = cachefs_rename_connected(odvp, onm, 5264 ndvp, nnm, cr, delvp); 5265 if (CFS_TIMEOUT(fscp, error)) { 5266 cachefs_cd_release(fscp); 5267 held = 0; 5268 cachefs_cd_timedout(fscp); 5269 connected = 0; 5270 continue; 5271 } 5272 } else { 5273 error = cachefs_rename_disconnected(odvp, onm, 5274 ndvp, nnm, cr, delvp); 5275 if (CFS_TIMEOUT(fscp, error)) { 5276 connected = 1; 5277 continue; 5278 } 5279 } 5280 break; 5281 } 5282 5283 if (CACHEFS_LOG_LOGGING(cachep, CACHEFS_LOG_RENAME)) { 5284 struct fid gone; 5285 5286 bzero(&gone, sizeof (gone)); 5287 gone.fid_len = MAXFIDSZ; 5288 if (delvp != NULL) 5289 (void) VOP_FID(delvp, &gone); 5290 5291 cachefs_log_rename(cachep, error, fscp->fs_cfsvfsp, 5292 &gone, 0, (delvp != NULL), crgetuid(cr)); 5293 } 5294 5295 if (held) 5296 cachefs_cd_release(fscp); 5297 5298 if (vfslock) 5299 vn_vfsunlock(delvp); 5300 5301 if (delvp) 5302 VN_RELE(delvp); 5303 if (tvp) 5304 VN_RELE(tvp); 5305 5306 #ifdef CFS_CD_DEBUG 5307 ASSERT((curthread->t_flag & T_CD_HELD) == 0); 5308 #endif 5309 return (error); 5310 } 5311 5312 static int 5313 cachefs_rename_connected(vnode_t *odvp, char *onm, vnode_t *ndvp, 5314 char *nnm, cred_t *cr, vnode_t *delvp) 5315 { 5316 cnode_t *odcp = VTOC(odvp); 5317 cnode_t *ndcp = VTOC(ndvp); 5318 vnode_t *revp = NULL; 5319 cnode_t *recp; 5320 cnode_t *delcp; 5321 fscache_t *fscp = C_TO_FSCACHE(odcp); 5322 int error = 0; 5323 struct fid cookie; 5324 struct fid *cookiep; 5325 cfs_cid_t cid; 5326 int gotdirent; 5327 5328 /* find the file we are renaming */ 5329 error = cachefs_lookup_common(odvp, onm, &revp, NULL, 0, NULL, cr); 5330 if (error) 5331 return (error); 5332 recp = VTOC(revp); 5333 5334 /* 5335 * To avoid deadlock, we acquire this global rename lock before 5336 * we try to get the locks for the source and target directories. 5337 */ 5338 mutex_enter(&cachefs_rename_lock); 5339 rw_enter(&odcp->c_rwlock, RW_WRITER); 5340 if (odcp != ndcp) { 5341 rw_enter(&ndcp->c_rwlock, RW_WRITER); 5342 } 5343 mutex_exit(&cachefs_rename_lock); 5344 5345 ASSERT((odcp->c_flags & CN_ASYNC_POP_WORKING) == 0); 5346 ASSERT((ndcp->c_flags & CN_ASYNC_POP_WORKING) == 0); 5347 5348 mutex_enter(&odcp->c_statelock); 5349 if (odcp->c_backvp == NULL) { 5350 error = cachefs_getbackvp(fscp, odcp); 5351 if (error) { 5352 mutex_exit(&odcp->c_statelock); 5353 goto out; 5354 } 5355 } 5356 5357 error = CFSOP_CHECK_COBJECT(fscp, odcp, 0, cr); 5358 if (error) { 5359 mutex_exit(&odcp->c_statelock); 5360 goto out; 5361 } 5362 mutex_exit(&odcp->c_statelock); 5363 5364 if (odcp != ndcp) { 5365 mutex_enter(&ndcp->c_statelock); 5366 if (ndcp->c_backvp == NULL) { 5367 error = cachefs_getbackvp(fscp, ndcp); 5368 if (error) { 5369 mutex_exit(&ndcp->c_statelock); 5370 goto out; 5371 } 5372 } 5373 5374 error = CFSOP_CHECK_COBJECT(fscp, ndcp, 0, cr); 5375 if (error) { 5376 mutex_exit(&ndcp->c_statelock); 5377 goto out; 5378 } 5379 mutex_exit(&ndcp->c_statelock); 5380 } 5381 5382 /* if a file is being deleted because of this rename */ 5383 if (delvp) { 5384 /* if src and dest file are same */ 5385 if (delvp == revp) { 5386 error = 0; 5387 goto out; 5388 } 5389 5390 /* 5391 * If the cnode is active, make a link to the file 5392 * so operations on the file will continue. 5393 */ 5394 dnlc_purge_vp(delvp); 5395 delcp = VTOC(delvp); 5396 if ((delvp->v_type != VDIR) && 5397 !((delvp->v_count == 1) || 5398 ((delvp->v_count == 2) && delcp->c_ipending))) { 5399 error = cachefs_remove_dolink(ndvp, delvp, nnm, cr); 5400 if (error) 5401 goto out; 5402 } 5403 } 5404 5405 /* do the rename on the back fs */ 5406 CFS_DPRINT_BACKFS_NFSV4(fscp, 5407 ("cachefs_rename (nfsv4): odcp %p, odbackvp %p, " 5408 " ndcp %p, ndbackvp %p, onm %s, nnm %s\n", 5409 odcp, odcp->c_backvp, ndcp, ndcp->c_backvp, onm, nnm)); 5410 error = VOP_RENAME(odcp->c_backvp, onm, ndcp->c_backvp, nnm, cr); 5411 if (error) 5412 goto out; 5413 5414 /* purge mappings to file in the old directory */ 5415 dnlc_purge_vp(odvp); 5416 5417 /* purge mappings in the new dir if we deleted a file */ 5418 if (delvp && (odvp != ndvp)) 5419 dnlc_purge_vp(ndvp); 5420 5421 /* update the file we just deleted */ 5422 if (delvp) { 5423 mutex_enter(&delcp->c_statelock); 5424 if (delcp->c_attr.va_nlink == 1) { 5425 delcp->c_flags |= CN_DESTROY; 5426 } else { 5427 delcp->c_flags |= CN_UPDATED; 5428 } 5429 delcp->c_attr.va_nlink--; 5430 CFSOP_MODIFY_COBJECT(fscp, delcp, cr); 5431 mutex_exit(&delcp->c_statelock); 5432 } 5433 5434 /* find the entry in the old directory */ 5435 mutex_enter(&odcp->c_statelock); 5436 gotdirent = 0; 5437 cookiep = NULL; 5438 if (CFS_ISFS_NONSHARED(fscp) && 5439 (odcp->c_metadata.md_flags & MD_POPULATED)) { 5440 error = cachefs_dir_look(odcp, onm, &cookie, 5441 NULL, NULL, &cid); 5442 if (error == 0 || error == EINVAL) { 5443 gotdirent = 1; 5444 if (error == 0) 5445 cookiep = &cookie; 5446 } else { 5447 cachefs_inval_object(odcp); 5448 } 5449 } 5450 error = 0; 5451 5452 /* remove the directory entry from the old directory */ 5453 if (gotdirent) { 5454 error = cachefs_dir_rmentry(odcp, onm); 5455 if (error) { 5456 cachefs_nocache(odcp); 5457 error = 0; 5458 } 5459 } 5460 CFSOP_MODIFY_COBJECT(fscp, odcp, cr); 5461 mutex_exit(&odcp->c_statelock); 5462 5463 /* install the directory entry in the new directory */ 5464 mutex_enter(&ndcp->c_statelock); 5465 if (CFS_ISFS_NONSHARED(fscp) && 5466 (ndcp->c_metadata.md_flags & MD_POPULATED)) { 5467 error = 1; 5468 if (gotdirent) { 5469 ASSERT(cid.cid_fileno != 0); 5470 error = 0; 5471 if (delvp) { 5472 error = cachefs_dir_rmentry(ndcp, nnm); 5473 } 5474 if (error == 0) { 5475 error = cachefs_dir_enter(ndcp, nnm, cookiep, 5476 &cid, SM_ASYNC); 5477 } 5478 } 5479 if (error) { 5480 cachefs_nocache(ndcp); 5481 error = 0; 5482 } 5483 } 5484 if (odcp != ndcp) 5485 CFSOP_MODIFY_COBJECT(fscp, ndcp, cr); 5486 mutex_exit(&ndcp->c_statelock); 5487 5488 /* ctime of renamed file has changed */ 5489 mutex_enter(&recp->c_statelock); 5490 CFSOP_MODIFY_COBJECT(fscp, recp, cr); 5491 mutex_exit(&recp->c_statelock); 5492 5493 out: 5494 if (odcp != ndcp) 5495 rw_exit(&ndcp->c_rwlock); 5496 rw_exit(&odcp->c_rwlock); 5497 5498 VN_RELE(revp); 5499 5500 return (error); 5501 } 5502 5503 static int 5504 cachefs_rename_disconnected(vnode_t *odvp, char *onm, vnode_t *ndvp, 5505 char *nnm, cred_t *cr, vnode_t *delvp) 5506 { 5507 cnode_t *odcp = VTOC(odvp); 5508 cnode_t *ndcp = VTOC(ndvp); 5509 cnode_t *delcp = NULL; 5510 vnode_t *revp = NULL; 5511 cnode_t *recp; 5512 fscache_t *fscp = C_TO_FSCACHE(odcp); 5513 int error = 0; 5514 struct fid cookie; 5515 struct fid *cookiep; 5516 cfs_cid_t cid; 5517 off_t commit = 0; 5518 timestruc_t current_time; 5519 5520 if (CFS_ISFS_WRITE_AROUND(fscp)) 5521 return (ETIMEDOUT); 5522 5523 /* find the file we are renaming */ 5524 error = cachefs_lookup_common(odvp, onm, &revp, NULL, 0, NULL, cr); 5525 if (error) 5526 return (error); 5527 recp = VTOC(revp); 5528 5529 /* 5530 * To avoid deadlock, we acquire this global rename lock before 5531 * we try to get the locks for the source and target directories. 5532 */ 5533 mutex_enter(&cachefs_rename_lock); 5534 rw_enter(&odcp->c_rwlock, RW_WRITER); 5535 if (odcp != ndcp) { 5536 rw_enter(&ndcp->c_rwlock, RW_WRITER); 5537 } 5538 mutex_exit(&cachefs_rename_lock); 5539 5540 if (recp->c_metadata.md_flags & MD_NEEDATTRS) { 5541 error = ETIMEDOUT; 5542 goto out; 5543 } 5544 5545 if ((recp->c_metadata.md_flags & MD_MAPPING) == 0) { 5546 mutex_enter(&recp->c_statelock); 5547 if ((recp->c_metadata.md_flags & MD_MAPPING) == 0) { 5548 error = cachefs_dlog_cidmap(fscp); 5549 if (error) { 5550 mutex_exit(&recp->c_statelock); 5551 error = ENOSPC; 5552 goto out; 5553 } 5554 recp->c_metadata.md_flags |= MD_MAPPING; 5555 recp->c_flags |= CN_UPDATED; 5556 } 5557 mutex_exit(&recp->c_statelock); 5558 } 5559 5560 /* check permissions */ 5561 /* XXX clean up this mutex junk sometime */ 5562 mutex_enter(&odcp->c_statelock); 5563 error = cachefs_access_local(odcp, (VEXEC|VWRITE), cr); 5564 mutex_exit(&odcp->c_statelock); 5565 if (error != 0) 5566 goto out; 5567 mutex_enter(&ndcp->c_statelock); 5568 error = cachefs_access_local(ndcp, (VEXEC|VWRITE), cr); 5569 mutex_exit(&ndcp->c_statelock); 5570 if (error != 0) 5571 goto out; 5572 mutex_enter(&odcp->c_statelock); 5573 error = cachefs_stickyrmchk(odcp, recp, cr); 5574 mutex_exit(&odcp->c_statelock); 5575 if (error != 0) 5576 goto out; 5577 5578 /* dirs must be populated */ 5579 if (((odcp->c_metadata.md_flags & MD_POPULATED) == 0) || 5580 ((ndcp->c_metadata.md_flags & MD_POPULATED) == 0)) { 5581 error = ETIMEDOUT; 5582 goto out; 5583 } 5584 5585 /* for now do not allow moving dirs because could cause cycles */ 5586 if ((((revp->v_type == VDIR) && (odvp != ndvp))) || 5587 (revp == odvp)) { 5588 error = ETIMEDOUT; 5589 goto out; 5590 } 5591 5592 /* if a file is being deleted because of this rename */ 5593 if (delvp) { 5594 delcp = VTOC(delvp); 5595 5596 /* if src and dest file are the same */ 5597 if (delvp == revp) { 5598 error = 0; 5599 goto out; 5600 } 5601 5602 if (delcp->c_metadata.md_flags & MD_NEEDATTRS) { 5603 error = ETIMEDOUT; 5604 goto out; 5605 } 5606 5607 /* if there are hard links to this file */ 5608 if (delcp->c_attr.va_nlink > 1) { 5609 mutex_enter(&delcp->c_statelock); 5610 if (cachefs_modified_alloc(delcp)) { 5611 mutex_exit(&delcp->c_statelock); 5612 error = ENOSPC; 5613 goto out; 5614 } 5615 5616 if ((delcp->c_metadata.md_flags & MD_MAPPING) == 0) { 5617 error = cachefs_dlog_cidmap(fscp); 5618 if (error) { 5619 mutex_exit(&delcp->c_statelock); 5620 error = ENOSPC; 5621 goto out; 5622 } 5623 delcp->c_metadata.md_flags |= MD_MAPPING; 5624 delcp->c_flags |= CN_UPDATED; 5625 } 5626 mutex_exit(&delcp->c_statelock); 5627 } 5628 5629 /* make sure we can delete file */ 5630 mutex_enter(&ndcp->c_statelock); 5631 error = cachefs_stickyrmchk(ndcp, delcp, cr); 5632 mutex_exit(&ndcp->c_statelock); 5633 if (error != 0) 5634 goto out; 5635 5636 /* 5637 * If the cnode is active, make a link to the file 5638 * so operations on the file will continue. 5639 */ 5640 dnlc_purge_vp(delvp); 5641 if ((delvp->v_type != VDIR) && 5642 !((delvp->v_count == 1) || 5643 ((delvp->v_count == 2) && delcp->c_ipending))) { 5644 error = cachefs_remove_dolink(ndvp, delvp, nnm, cr); 5645 if (error) 5646 goto out; 5647 } 5648 } 5649 5650 /* purge mappings to file in the old directory */ 5651 dnlc_purge_vp(odvp); 5652 5653 /* purge mappings in the new dir if we deleted a file */ 5654 if (delvp && (odvp != ndvp)) 5655 dnlc_purge_vp(ndvp); 5656 5657 /* find the entry in the old directory */ 5658 mutex_enter(&odcp->c_statelock); 5659 if ((odcp->c_metadata.md_flags & MD_POPULATED) == 0) { 5660 mutex_exit(&odcp->c_statelock); 5661 error = ETIMEDOUT; 5662 goto out; 5663 } 5664 cookiep = NULL; 5665 error = cachefs_dir_look(odcp, onm, &cookie, NULL, NULL, &cid); 5666 if (error == 0 || error == EINVAL) { 5667 if (error == 0) 5668 cookiep = &cookie; 5669 } else { 5670 mutex_exit(&odcp->c_statelock); 5671 if (error == ENOTDIR) 5672 error = ETIMEDOUT; 5673 goto out; 5674 } 5675 error = 0; 5676 5677 /* write the log entry */ 5678 commit = cachefs_dlog_rename(fscp, odcp, onm, ndcp, nnm, cr, 5679 recp, delcp); 5680 if (commit == 0) { 5681 mutex_exit(&odcp->c_statelock); 5682 error = ENOSPC; 5683 goto out; 5684 } 5685 5686 /* remove the directory entry from the old directory */ 5687 cachefs_modified(odcp); 5688 error = cachefs_dir_rmentry(odcp, onm); 5689 if (error) { 5690 mutex_exit(&odcp->c_statelock); 5691 if (error == ENOTDIR) 5692 error = ETIMEDOUT; 5693 goto out; 5694 } 5695 mutex_exit(&odcp->c_statelock); 5696 5697 /* install the directory entry in the new directory */ 5698 mutex_enter(&ndcp->c_statelock); 5699 error = ENOTDIR; 5700 if (ndcp->c_metadata.md_flags & MD_POPULATED) { 5701 ASSERT(cid.cid_fileno != 0); 5702 cachefs_modified(ndcp); 5703 error = 0; 5704 if (delvp) { 5705 error = cachefs_dir_rmentry(ndcp, nnm); 5706 } 5707 if (error == 0) { 5708 error = cachefs_dir_enter(ndcp, nnm, cookiep, 5709 &cid, SM_ASYNC); 5710 } 5711 } 5712 if (error) { 5713 cachefs_nocache(ndcp); 5714 mutex_exit(&ndcp->c_statelock); 5715 mutex_enter(&odcp->c_statelock); 5716 cachefs_nocache(odcp); 5717 mutex_exit(&odcp->c_statelock); 5718 if (error == ENOTDIR) 5719 error = ETIMEDOUT; 5720 goto out; 5721 } 5722 mutex_exit(&ndcp->c_statelock); 5723 5724 gethrestime(¤t_time); 5725 5726 /* update the file we just deleted */ 5727 if (delvp) { 5728 mutex_enter(&delcp->c_statelock); 5729 delcp->c_attr.va_nlink--; 5730 delcp->c_metadata.md_localctime = current_time; 5731 delcp->c_metadata.md_flags |= MD_LOCALCTIME; 5732 if (delcp->c_attr.va_nlink == 0) { 5733 delcp->c_flags |= CN_DESTROY; 5734 } else { 5735 delcp->c_flags |= CN_UPDATED; 5736 } 5737 mutex_exit(&delcp->c_statelock); 5738 } 5739 5740 /* update the file we renamed */ 5741 mutex_enter(&recp->c_statelock); 5742 recp->c_metadata.md_localctime = current_time; 5743 recp->c_metadata.md_flags |= MD_LOCALCTIME; 5744 recp->c_flags |= CN_UPDATED; 5745 mutex_exit(&recp->c_statelock); 5746 5747 /* update the source directory */ 5748 mutex_enter(&odcp->c_statelock); 5749 odcp->c_metadata.md_localctime = current_time; 5750 odcp->c_metadata.md_localmtime = current_time; 5751 odcp->c_metadata.md_flags |= MD_LOCALCTIME | MD_LOCALMTIME; 5752 odcp->c_flags |= CN_UPDATED; 5753 mutex_exit(&odcp->c_statelock); 5754 5755 /* update the destination directory */ 5756 if (odcp != ndcp) { 5757 mutex_enter(&ndcp->c_statelock); 5758 ndcp->c_metadata.md_localctime = current_time; 5759 ndcp->c_metadata.md_localmtime = current_time; 5760 ndcp->c_metadata.md_flags |= MD_LOCALCTIME | MD_LOCALMTIME; 5761 ndcp->c_flags |= CN_UPDATED; 5762 mutex_exit(&ndcp->c_statelock); 5763 } 5764 5765 out: 5766 if (commit) { 5767 /* commit the log entry */ 5768 if (cachefs_dlog_commit(fscp, commit, error)) { 5769 /*EMPTY*/ 5770 /* XXX bob: fix on panic */ 5771 } 5772 } 5773 5774 if (odcp != ndcp) 5775 rw_exit(&ndcp->c_rwlock); 5776 rw_exit(&odcp->c_rwlock); 5777 5778 VN_RELE(revp); 5779 5780 return (error); 5781 } 5782 5783 static int 5784 cachefs_mkdir(vnode_t *dvp, char *nm, vattr_t *vap, vnode_t **vpp, 5785 cred_t *cr) 5786 { 5787 cnode_t *dcp = VTOC(dvp); 5788 fscache_t *fscp = C_TO_FSCACHE(dcp); 5789 cachefscache_t *cachep = fscp->fs_cache; 5790 int error = 0; 5791 int held = 0; 5792 int connected = 0; 5793 5794 #ifdef CFSDEBUG 5795 CFS_DEBUG(CFSDEBUG_VOPS) 5796 printf("cachefs_mkdir: ENTER dvp %p\n", (void *)dvp); 5797 #endif 5798 5799 if (getzoneid() != GLOBAL_ZONEID) { 5800 error = EPERM; 5801 goto out; 5802 } 5803 5804 if (fscp->fs_cache->c_flags & (CACHE_NOFILL | CACHE_NOCACHE)) 5805 ASSERT(dcp->c_flags & CN_NOCACHE); 5806 5807 /* 5808 * Cachefs only provides pass-through support for NFSv4, 5809 * and all vnode operations are passed through to the 5810 * back file system. For NFSv4 pass-through to work, only 5811 * connected operation is supported, the cnode backvp must 5812 * exist, and cachefs optional (eg., disconnectable) flags 5813 * are turned off. Assert these conditions to ensure that 5814 * the backfilesystem is called for the mkdir operation. 5815 */ 5816 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 5817 CFS_BACKFS_NFSV4_ASSERT_CNODE(dcp); 5818 5819 for (;;) { 5820 /* get (or renew) access to the file system */ 5821 if (held) { 5822 /* Won't loop with NFSv4 connected behavior */ 5823 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 5824 rw_exit(&dcp->c_rwlock); 5825 cachefs_cd_release(fscp); 5826 held = 0; 5827 } 5828 error = cachefs_cd_access(fscp, connected, 1); 5829 if (error) 5830 break; 5831 rw_enter(&dcp->c_rwlock, RW_WRITER); 5832 held = 1; 5833 5834 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 5835 error = cachefs_mkdir_connected(dvp, nm, vap, 5836 vpp, cr); 5837 if (CFS_TIMEOUT(fscp, error)) { 5838 rw_exit(&dcp->c_rwlock); 5839 cachefs_cd_release(fscp); 5840 held = 0; 5841 cachefs_cd_timedout(fscp); 5842 connected = 0; 5843 continue; 5844 } 5845 } else { 5846 error = cachefs_mkdir_disconnected(dvp, nm, vap, 5847 vpp, cr); 5848 if (CFS_TIMEOUT(fscp, error)) { 5849 connected = 1; 5850 continue; 5851 } 5852 } 5853 break; 5854 } 5855 5856 if (CACHEFS_LOG_LOGGING(cachep, CACHEFS_LOG_MKDIR)) { 5857 fid_t *fidp = NULL; 5858 ino64_t fileno = 0; 5859 cnode_t *cp = NULL; 5860 if (error == 0) 5861 cp = VTOC(*vpp); 5862 5863 if (cp != NULL) { 5864 fidp = &cp->c_metadata.md_cookie; 5865 fileno = cp->c_id.cid_fileno; 5866 } 5867 5868 cachefs_log_mkdir(cachep, error, fscp->fs_cfsvfsp, 5869 fidp, fileno, crgetuid(cr)); 5870 } 5871 5872 if (held) { 5873 rw_exit(&dcp->c_rwlock); 5874 cachefs_cd_release(fscp); 5875 } 5876 if (error == 0 && CFS_ISFS_NONSHARED(fscp)) 5877 (void) cachefs_pack(dvp, nm, cr); 5878 5879 #ifdef CFS_CD_DEBUG 5880 ASSERT((curthread->t_flag & T_CD_HELD) == 0); 5881 #endif 5882 out: 5883 #ifdef CFSDEBUG 5884 CFS_DEBUG(CFSDEBUG_VOPS) 5885 printf("cachefs_mkdir: EXIT error = %d\n", error); 5886 #endif 5887 return (error); 5888 } 5889 5890 static int 5891 cachefs_mkdir_connected(vnode_t *dvp, char *nm, vattr_t *vap, 5892 vnode_t **vpp, cred_t *cr) 5893 { 5894 cnode_t *newcp = NULL, *dcp = VTOC(dvp); 5895 struct vnode *vp = NULL; 5896 int error = 0; 5897 fscache_t *fscp = C_TO_FSCACHE(dcp); 5898 struct fid cookie; 5899 struct vattr attr; 5900 cfs_cid_t cid, dircid; 5901 uint32_t valid_fid; 5902 5903 if (fscp->fs_cache->c_flags & (CACHE_NOFILL | CACHE_NOCACHE)) 5904 ASSERT(dcp->c_flags & CN_NOCACHE); 5905 5906 mutex_enter(&dcp->c_statelock); 5907 5908 /* get backvp of dir */ 5909 if (dcp->c_backvp == NULL) { 5910 error = cachefs_getbackvp(fscp, dcp); 5911 if (error) { 5912 mutex_exit(&dcp->c_statelock); 5913 goto out; 5914 } 5915 } 5916 5917 /* consistency check the directory */ 5918 error = CFSOP_CHECK_COBJECT(fscp, dcp, 0, cr); 5919 if (error) { 5920 mutex_exit(&dcp->c_statelock); 5921 goto out; 5922 } 5923 dircid = dcp->c_id; 5924 5925 /* make the dir on the back fs */ 5926 CFS_DPRINT_BACKFS_NFSV4(fscp, 5927 ("cachefs_mkdir (nfsv4): dcp %p, dbackvp %p, " 5928 "name %s\n", dcp, dcp->c_backvp, nm)); 5929 error = VOP_MKDIR(dcp->c_backvp, nm, vap, &vp, cr); 5930 mutex_exit(&dcp->c_statelock); 5931 if (error) { 5932 goto out; 5933 } 5934 5935 /* get the cookie and make the cnode */ 5936 attr.va_mask = AT_ALL; 5937 valid_fid = (CFS_ISFS_BACKFS_NFSV4(fscp) ? FALSE : TRUE); 5938 error = cachefs_getcookie(vp, &cookie, &attr, cr, valid_fid); 5939 if (error) { 5940 goto out; 5941 } 5942 cid.cid_flags = 0; 5943 cid.cid_fileno = attr.va_nodeid; 5944 error = cachefs_cnode_make(&cid, fscp, (valid_fid ? &cookie : NULL), 5945 &attr, vp, cr, 0, &newcp); 5946 if (error) { 5947 goto out; 5948 } 5949 ASSERT(CTOV(newcp)->v_type == VDIR); 5950 *vpp = CTOV(newcp); 5951 5952 /* if the dir is populated, add the new entry */ 5953 mutex_enter(&dcp->c_statelock); 5954 if (CFS_ISFS_NONSHARED(fscp) && 5955 (dcp->c_metadata.md_flags & MD_POPULATED)) { 5956 error = cachefs_dir_enter(dcp, nm, &cookie, &newcp->c_id, 5957 SM_ASYNC); 5958 if (error) { 5959 cachefs_nocache(dcp); 5960 error = 0; 5961 } 5962 } 5963 dcp->c_attr.va_nlink++; 5964 dcp->c_flags |= CN_UPDATED; 5965 CFSOP_MODIFY_COBJECT(fscp, dcp, cr); 5966 mutex_exit(&dcp->c_statelock); 5967 5968 /* XXX bob: should we do a filldir here? or just add . and .. */ 5969 /* maybe should kick off an async filldir so caller does not wait */ 5970 5971 /* put the entry in the dnlc */ 5972 if (cachefs_dnlc) 5973 dnlc_enter(dvp, nm, *vpp); 5974 5975 /* save the fileno of the parent so can find the name */ 5976 if (bcmp(&newcp->c_metadata.md_parent, &dircid, 5977 sizeof (cfs_cid_t)) != 0) { 5978 mutex_enter(&newcp->c_statelock); 5979 newcp->c_metadata.md_parent = dircid; 5980 newcp->c_flags |= CN_UPDATED; 5981 mutex_exit(&newcp->c_statelock); 5982 } 5983 out: 5984 if (vp) 5985 VN_RELE(vp); 5986 5987 return (error); 5988 } 5989 5990 static int 5991 cachefs_mkdir_disconnected(vnode_t *dvp, char *nm, vattr_t *vap, 5992 vnode_t **vpp, cred_t *cr) 5993 { 5994 cnode_t *dcp = VTOC(dvp); 5995 fscache_t *fscp = C_TO_FSCACHE(dcp); 5996 int error; 5997 cnode_t *newcp = NULL; 5998 struct vattr va; 5999 timestruc_t current_time; 6000 off_t commit = 0; 6001 char *s; 6002 int namlen; 6003 6004 /* don't allow '/' characters in pathname component */ 6005 for (s = nm, namlen = 0; *s; s++, namlen++) 6006 if (*s == '/') 6007 return (EACCES); 6008 if (namlen == 0) 6009 return (EINVAL); 6010 6011 if (CFS_ISFS_WRITE_AROUND(fscp)) 6012 return (ETIMEDOUT); 6013 6014 mutex_enter(&dcp->c_statelock); 6015 6016 /* check permissions */ 6017 if (error = cachefs_access_local(dcp, (VEXEC|VWRITE), cr)) { 6018 mutex_exit(&dcp->c_statelock); 6019 goto out; 6020 } 6021 6022 /* the directory front file must be populated */ 6023 if ((dcp->c_metadata.md_flags & MD_POPULATED) == 0) { 6024 error = ETIMEDOUT; 6025 mutex_exit(&dcp->c_statelock); 6026 goto out; 6027 } 6028 6029 /* make sure nm does not already exist in the directory */ 6030 error = cachefs_dir_look(dcp, nm, NULL, NULL, NULL, NULL); 6031 if (error == ENOTDIR) { 6032 error = ETIMEDOUT; 6033 mutex_exit(&dcp->c_statelock); 6034 goto out; 6035 } 6036 if (error != ENOENT) { 6037 error = EEXIST; 6038 mutex_exit(&dcp->c_statelock); 6039 goto out; 6040 } 6041 6042 /* make up a reasonable set of attributes */ 6043 cachefs_attr_setup(vap, &va, dcp, cr); 6044 va.va_type = VDIR; 6045 va.va_mode |= S_IFDIR; 6046 va.va_nlink = 2; 6047 6048 mutex_exit(&dcp->c_statelock); 6049 6050 /* create the cnode */ 6051 error = cachefs_cnode_create(fscp, &va, 0, &newcp); 6052 if (error) 6053 goto out; 6054 6055 mutex_enter(&newcp->c_statelock); 6056 6057 error = cachefs_dlog_cidmap(fscp); 6058 if (error) { 6059 mutex_exit(&newcp->c_statelock); 6060 goto out; 6061 } 6062 6063 cachefs_creategid(dcp, newcp, vap, cr); 6064 mutex_enter(&dcp->c_statelock); 6065 cachefs_createacl(dcp, newcp); 6066 mutex_exit(&dcp->c_statelock); 6067 gethrestime(¤t_time); 6068 newcp->c_metadata.md_vattr.va_atime = current_time; 6069 newcp->c_metadata.md_localctime = current_time; 6070 newcp->c_metadata.md_localmtime = current_time; 6071 newcp->c_metadata.md_flags |= MD_MAPPING | MD_LOCALMTIME | 6072 MD_LOCALCTIME; 6073 newcp->c_flags |= CN_UPDATED; 6074 6075 /* make a front file for the new directory, add . and .. */ 6076 error = cachefs_dir_new(dcp, newcp); 6077 if (error) { 6078 mutex_exit(&newcp->c_statelock); 6079 goto out; 6080 } 6081 cachefs_modified(newcp); 6082 6083 /* 6084 * write the metadata now rather than waiting until 6085 * inactive so that if there's no space we can let 6086 * the caller know. 6087 */ 6088 ASSERT(newcp->c_frontvp); 6089 ASSERT((newcp->c_filegrp->fg_flags & CFS_FG_ALLOC_ATTR) == 0); 6090 ASSERT((newcp->c_flags & CN_ALLOC_PENDING) == 0); 6091 error = filegrp_write_metadata(newcp->c_filegrp, 6092 &newcp->c_id, &newcp->c_metadata); 6093 if (error) { 6094 mutex_exit(&newcp->c_statelock); 6095 goto out; 6096 } 6097 mutex_exit(&newcp->c_statelock); 6098 6099 /* log the operation */ 6100 commit = cachefs_dlog_mkdir(fscp, dcp, newcp, nm, &va, cr); 6101 if (commit == 0) { 6102 error = ENOSPC; 6103 goto out; 6104 } 6105 6106 mutex_enter(&dcp->c_statelock); 6107 6108 /* make sure directory is still populated */ 6109 if ((dcp->c_metadata.md_flags & MD_POPULATED) == 0) { 6110 mutex_exit(&dcp->c_statelock); 6111 error = ETIMEDOUT; 6112 goto out; 6113 } 6114 cachefs_modified(dcp); 6115 6116 /* enter the new file in the directory */ 6117 error = cachefs_dir_enter(dcp, nm, &newcp->c_metadata.md_cookie, 6118 &newcp->c_id, SM_ASYNC); 6119 if (error) { 6120 mutex_exit(&dcp->c_statelock); 6121 goto out; 6122 } 6123 6124 /* update parent dir times */ 6125 dcp->c_metadata.md_localctime = current_time; 6126 dcp->c_metadata.md_localmtime = current_time; 6127 dcp->c_metadata.md_flags |= MD_LOCALCTIME | MD_LOCALMTIME; 6128 dcp->c_attr.va_nlink++; 6129 dcp->c_flags |= CN_UPDATED; 6130 mutex_exit(&dcp->c_statelock); 6131 6132 out: 6133 if (commit) { 6134 /* commit the log entry */ 6135 if (cachefs_dlog_commit(fscp, commit, error)) { 6136 /*EMPTY*/ 6137 /* XXX bob: fix on panic */ 6138 } 6139 } 6140 if (error) { 6141 if (newcp) { 6142 mutex_enter(&newcp->c_statelock); 6143 newcp->c_flags |= CN_DESTROY; 6144 mutex_exit(&newcp->c_statelock); 6145 VN_RELE(CTOV(newcp)); 6146 } 6147 } else { 6148 *vpp = CTOV(newcp); 6149 } 6150 return (error); 6151 } 6152 6153 static int 6154 cachefs_rmdir(vnode_t *dvp, char *nm, vnode_t *cdir, cred_t *cr) 6155 { 6156 cnode_t *dcp = VTOC(dvp); 6157 fscache_t *fscp = C_TO_FSCACHE(dcp); 6158 cachefscache_t *cachep = fscp->fs_cache; 6159 int error = 0; 6160 int held = 0; 6161 int connected = 0; 6162 size_t namlen; 6163 vnode_t *vp = NULL; 6164 int vfslock = 0; 6165 6166 #ifdef CFSDEBUG 6167 CFS_DEBUG(CFSDEBUG_VOPS) 6168 printf("cachefs_rmdir: ENTER vp %p\n", (void *)dvp); 6169 #endif 6170 6171 if (getzoneid() != GLOBAL_ZONEID) { 6172 error = EPERM; 6173 goto out; 6174 } 6175 6176 if (fscp->fs_cache->c_flags & (CACHE_NOFILL | CACHE_NOCACHE)) 6177 ASSERT(dcp->c_flags & CN_NOCACHE); 6178 6179 /* 6180 * Cachefs only provides pass-through support for NFSv4, 6181 * and all vnode operations are passed through to the 6182 * back file system. For NFSv4 pass-through to work, only 6183 * connected operation is supported, the cnode backvp must 6184 * exist, and cachefs optional (eg., disconnectable) flags 6185 * are turned off. Assert these conditions to ensure that 6186 * the backfilesystem is called for the rmdir operation. 6187 */ 6188 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 6189 CFS_BACKFS_NFSV4_ASSERT_CNODE(dcp); 6190 6191 for (;;) { 6192 if (vfslock) { 6193 vn_vfsunlock(vp); 6194 vfslock = 0; 6195 } 6196 if (vp) { 6197 VN_RELE(vp); 6198 vp = NULL; 6199 } 6200 6201 /* get (or renew) access to the file system */ 6202 if (held) { 6203 /* Won't loop with NFSv4 connected behavior */ 6204 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 6205 cachefs_cd_release(fscp); 6206 held = 0; 6207 } 6208 error = cachefs_cd_access(fscp, connected, 1); 6209 if (error) 6210 break; 6211 held = 1; 6212 6213 /* if disconnected, do some extra error checking */ 6214 if (fscp->fs_cdconnected != CFS_CD_CONNECTED) { 6215 /* check permissions */ 6216 mutex_enter(&dcp->c_statelock); 6217 error = cachefs_access_local(dcp, (VEXEC|VWRITE), cr); 6218 mutex_exit(&dcp->c_statelock); 6219 if (CFS_TIMEOUT(fscp, error)) { 6220 connected = 1; 6221 continue; 6222 } 6223 if (error) 6224 break; 6225 6226 namlen = strlen(nm); 6227 if (namlen == 0) { 6228 error = EINVAL; 6229 break; 6230 } 6231 6232 /* cannot remove . and .. */ 6233 if (nm[0] == '.') { 6234 if (namlen == 1) { 6235 error = EINVAL; 6236 break; 6237 } else if (namlen == 2 && nm[1] == '.') { 6238 error = EEXIST; 6239 break; 6240 } 6241 } 6242 6243 } 6244 6245 /* get the cnode of the dir to remove */ 6246 error = cachefs_lookup_common(dvp, nm, &vp, NULL, 0, NULL, cr); 6247 if (error) { 6248 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 6249 if (CFS_TIMEOUT(fscp, error)) { 6250 cachefs_cd_release(fscp); 6251 held = 0; 6252 cachefs_cd_timedout(fscp); 6253 connected = 0; 6254 continue; 6255 } 6256 } else { 6257 if (CFS_TIMEOUT(fscp, error)) { 6258 connected = 1; 6259 continue; 6260 } 6261 } 6262 break; 6263 } 6264 6265 /* must be a dir */ 6266 if (vp->v_type != VDIR) { 6267 error = ENOTDIR; 6268 break; 6269 } 6270 6271 /* must not be current dir */ 6272 if (VOP_CMP(vp, cdir)) { 6273 error = EINVAL; 6274 break; 6275 } 6276 6277 /* see ufs_dirremove for why this is done, mount race */ 6278 if (vn_vfslock(vp)) { 6279 error = EBUSY; 6280 break; 6281 } 6282 vfslock = 1; 6283 if (vn_mountedvfs(vp) != NULL) { 6284 error = EBUSY; 6285 break; 6286 } 6287 6288 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 6289 error = cachefs_rmdir_connected(dvp, nm, cdir, 6290 cr, vp); 6291 if (CFS_TIMEOUT(fscp, error)) { 6292 cachefs_cd_release(fscp); 6293 held = 0; 6294 cachefs_cd_timedout(fscp); 6295 connected = 0; 6296 continue; 6297 } 6298 } else { 6299 error = cachefs_rmdir_disconnected(dvp, nm, cdir, 6300 cr, vp); 6301 if (CFS_TIMEOUT(fscp, error)) { 6302 connected = 1; 6303 continue; 6304 } 6305 } 6306 break; 6307 } 6308 6309 if (CACHEFS_LOG_LOGGING(cachep, CACHEFS_LOG_RMDIR)) { 6310 ino64_t fileno = 0; 6311 fid_t *fidp = NULL; 6312 cnode_t *cp = NULL; 6313 if (vp) 6314 cp = VTOC(vp); 6315 6316 if (cp != NULL) { 6317 fidp = &cp->c_metadata.md_cookie; 6318 fileno = cp->c_id.cid_fileno; 6319 } 6320 6321 cachefs_log_rmdir(cachep, error, fscp->fs_cfsvfsp, 6322 fidp, fileno, crgetuid(cr)); 6323 } 6324 6325 if (held) { 6326 cachefs_cd_release(fscp); 6327 } 6328 6329 if (vfslock) 6330 vn_vfsunlock(vp); 6331 6332 if (vp) 6333 VN_RELE(vp); 6334 6335 #ifdef CFS_CD_DEBUG 6336 ASSERT((curthread->t_flag & T_CD_HELD) == 0); 6337 #endif 6338 out: 6339 #ifdef CFSDEBUG 6340 CFS_DEBUG(CFSDEBUG_VOPS) 6341 printf("cachefs_rmdir: EXIT error = %d\n", error); 6342 #endif 6343 6344 return (error); 6345 } 6346 6347 static int 6348 cachefs_rmdir_connected(vnode_t *dvp, char *nm, vnode_t *cdir, cred_t *cr, 6349 vnode_t *vp) 6350 { 6351 cnode_t *dcp = VTOC(dvp); 6352 cnode_t *cp = VTOC(vp); 6353 int error = 0; 6354 fscache_t *fscp = C_TO_FSCACHE(dcp); 6355 6356 rw_enter(&dcp->c_rwlock, RW_WRITER); 6357 mutex_enter(&dcp->c_statelock); 6358 mutex_enter(&cp->c_statelock); 6359 6360 if (dcp->c_backvp == NULL) { 6361 error = cachefs_getbackvp(fscp, dcp); 6362 if (error) { 6363 goto out; 6364 } 6365 } 6366 6367 error = CFSOP_CHECK_COBJECT(fscp, dcp, 0, cr); 6368 if (error) 6369 goto out; 6370 6371 /* rmdir on the back fs */ 6372 CFS_DPRINT_BACKFS_NFSV4(fscp, 6373 ("cachefs_rmdir (nfsv4): dcp %p, dbackvp %p, " 6374 "name %s\n", dcp, dcp->c_backvp, nm)); 6375 error = VOP_RMDIR(dcp->c_backvp, nm, cdir, cr); 6376 if (error) 6377 goto out; 6378 6379 /* if the dir is populated, remove the entry from it */ 6380 if (CFS_ISFS_NONSHARED(fscp) && 6381 (dcp->c_metadata.md_flags & MD_POPULATED)) { 6382 error = cachefs_dir_rmentry(dcp, nm); 6383 if (error) { 6384 cachefs_nocache(dcp); 6385 error = 0; 6386 } 6387 } 6388 6389 /* 6390 * *if* the (hard) link count goes to 0, then we set the CDESTROY 6391 * flag on the cnode. The cached object will then be destroyed 6392 * at inactive time where the chickens come home to roost :-) 6393 * The link cnt for directories is bumped down by 2 'cause the "." 6394 * entry has to be elided too ! The link cnt for the parent goes down 6395 * by 1 (because of ".."). 6396 */ 6397 cp->c_attr.va_nlink -= 2; 6398 dcp->c_attr.va_nlink--; 6399 if (cp->c_attr.va_nlink == 0) { 6400 cp->c_flags |= CN_DESTROY; 6401 } else { 6402 cp->c_flags |= CN_UPDATED; 6403 } 6404 dcp->c_flags |= CN_UPDATED; 6405 6406 dnlc_purge_vp(vp); 6407 CFSOP_MODIFY_COBJECT(fscp, dcp, cr); 6408 6409 out: 6410 mutex_exit(&cp->c_statelock); 6411 mutex_exit(&dcp->c_statelock); 6412 rw_exit(&dcp->c_rwlock); 6413 6414 return (error); 6415 } 6416 6417 static int 6418 /*ARGSUSED*/ 6419 cachefs_rmdir_disconnected(vnode_t *dvp, char *nm, vnode_t *cdir, 6420 cred_t *cr, vnode_t *vp) 6421 { 6422 cnode_t *dcp = VTOC(dvp); 6423 cnode_t *cp = VTOC(vp); 6424 fscache_t *fscp = C_TO_FSCACHE(dcp); 6425 int error = 0; 6426 off_t commit = 0; 6427 timestruc_t current_time; 6428 6429 if (CFS_ISFS_WRITE_AROUND(fscp)) 6430 return (ETIMEDOUT); 6431 6432 rw_enter(&dcp->c_rwlock, RW_WRITER); 6433 mutex_enter(&dcp->c_statelock); 6434 mutex_enter(&cp->c_statelock); 6435 6436 /* both directories must be populated */ 6437 if (((dcp->c_metadata.md_flags & MD_POPULATED) == 0) || 6438 ((cp->c_metadata.md_flags & MD_POPULATED) == 0)) { 6439 error = ETIMEDOUT; 6440 goto out; 6441 } 6442 6443 /* if sticky bit set on the dir, more access checks to perform */ 6444 if (error = cachefs_stickyrmchk(dcp, cp, cr)) { 6445 goto out; 6446 } 6447 6448 /* make sure dir is empty */ 6449 if (cp->c_attr.va_nlink > 2) { 6450 error = cachefs_dir_empty(cp); 6451 if (error) { 6452 if (error == ENOTDIR) 6453 error = ETIMEDOUT; 6454 goto out; 6455 } 6456 cachefs_modified(cp); 6457 } 6458 cachefs_modified(dcp); 6459 6460 /* log the operation */ 6461 commit = cachefs_dlog_rmdir(fscp, dcp, nm, cp, cr); 6462 if (commit == 0) { 6463 error = ENOSPC; 6464 goto out; 6465 } 6466 6467 /* remove name from parent dir */ 6468 error = cachefs_dir_rmentry(dcp, nm); 6469 if (error == ENOTDIR) { 6470 error = ETIMEDOUT; 6471 goto out; 6472 } 6473 if (error) 6474 goto out; 6475 6476 gethrestime(¤t_time); 6477 6478 /* update deleted dir values */ 6479 cp->c_attr.va_nlink -= 2; 6480 if (cp->c_attr.va_nlink == 0) 6481 cp->c_flags |= CN_DESTROY; 6482 else { 6483 cp->c_metadata.md_localctime = current_time; 6484 cp->c_metadata.md_flags |= MD_LOCALCTIME; 6485 cp->c_flags |= CN_UPDATED; 6486 } 6487 6488 /* update parent values */ 6489 dcp->c_metadata.md_localctime = current_time; 6490 dcp->c_metadata.md_localmtime = current_time; 6491 dcp->c_metadata.md_flags |= MD_LOCALCTIME | MD_LOCALMTIME; 6492 dcp->c_attr.va_nlink--; 6493 dcp->c_flags |= CN_UPDATED; 6494 6495 out: 6496 mutex_exit(&cp->c_statelock); 6497 mutex_exit(&dcp->c_statelock); 6498 rw_exit(&dcp->c_rwlock); 6499 if (commit) { 6500 /* commit the log entry */ 6501 if (cachefs_dlog_commit(fscp, commit, error)) { 6502 /*EMPTY*/ 6503 /* XXX bob: fix on panic */ 6504 } 6505 dnlc_purge_vp(vp); 6506 } 6507 return (error); 6508 } 6509 6510 static int 6511 cachefs_symlink(vnode_t *dvp, char *lnm, vattr_t *tva, 6512 char *tnm, cred_t *cr) 6513 { 6514 cnode_t *dcp = VTOC(dvp); 6515 fscache_t *fscp = C_TO_FSCACHE(dcp); 6516 cachefscache_t *cachep = fscp->fs_cache; 6517 int error = 0; 6518 int held = 0; 6519 int connected = 0; 6520 6521 #ifdef CFSDEBUG 6522 CFS_DEBUG(CFSDEBUG_VOPS) 6523 printf("cachefs_symlink: ENTER dvp %p lnm %s tnm %s\n", 6524 (void *)dvp, lnm, tnm); 6525 #endif 6526 6527 if (getzoneid() != GLOBAL_ZONEID) { 6528 error = EPERM; 6529 goto out; 6530 } 6531 6532 if (fscp->fs_cache->c_flags & CACHE_NOCACHE) 6533 ASSERT(dcp->c_flags & CN_NOCACHE); 6534 6535 /* 6536 * Cachefs only provides pass-through support for NFSv4, 6537 * and all vnode operations are passed through to the 6538 * back file system. For NFSv4 pass-through to work, only 6539 * connected operation is supported, the cnode backvp must 6540 * exist, and cachefs optional (eg., disconnectable) flags 6541 * are turned off. Assert these conditions to ensure that 6542 * the backfilesystem is called for the symlink operation. 6543 */ 6544 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 6545 CFS_BACKFS_NFSV4_ASSERT_CNODE(dcp); 6546 6547 for (;;) { 6548 /* get (or renew) access to the file system */ 6549 if (held) { 6550 /* Won't loop with NFSv4 connected behavior */ 6551 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 6552 rw_exit(&dcp->c_rwlock); 6553 cachefs_cd_release(fscp); 6554 held = 0; 6555 } 6556 error = cachefs_cd_access(fscp, connected, 1); 6557 if (error) 6558 break; 6559 rw_enter(&dcp->c_rwlock, RW_WRITER); 6560 held = 1; 6561 6562 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 6563 error = cachefs_symlink_connected(dvp, lnm, tva, 6564 tnm, cr); 6565 if (CFS_TIMEOUT(fscp, error)) { 6566 rw_exit(&dcp->c_rwlock); 6567 cachefs_cd_release(fscp); 6568 held = 0; 6569 cachefs_cd_timedout(fscp); 6570 connected = 0; 6571 continue; 6572 } 6573 } else { 6574 error = cachefs_symlink_disconnected(dvp, lnm, tva, 6575 tnm, cr); 6576 if (CFS_TIMEOUT(fscp, error)) { 6577 connected = 1; 6578 continue; 6579 } 6580 } 6581 break; 6582 } 6583 6584 if (CACHEFS_LOG_LOGGING(cachep, CACHEFS_LOG_SYMLINK)) 6585 cachefs_log_symlink(cachep, error, fscp->fs_cfsvfsp, 6586 &dcp->c_metadata.md_cookie, dcp->c_id.cid_fileno, 6587 crgetuid(cr), (uint_t)strlen(tnm)); 6588 6589 if (held) { 6590 rw_exit(&dcp->c_rwlock); 6591 cachefs_cd_release(fscp); 6592 } 6593 6594 #ifdef CFS_CD_DEBUG 6595 ASSERT((curthread->t_flag & T_CD_HELD) == 0); 6596 #endif 6597 out: 6598 #ifdef CFSDEBUG 6599 CFS_DEBUG(CFSDEBUG_VOPS) 6600 printf("cachefs_symlink: EXIT error = %d\n", error); 6601 #endif 6602 return (error); 6603 } 6604 6605 static int 6606 cachefs_symlink_connected(vnode_t *dvp, char *lnm, vattr_t *tva, 6607 char *tnm, cred_t *cr) 6608 { 6609 cnode_t *dcp = VTOC(dvp); 6610 fscache_t *fscp = C_TO_FSCACHE(dcp); 6611 int error = 0; 6612 vnode_t *backvp = NULL; 6613 cnode_t *newcp = NULL; 6614 struct vattr va; 6615 struct fid cookie; 6616 cfs_cid_t cid; 6617 uint32_t valid_fid; 6618 6619 mutex_enter(&dcp->c_statelock); 6620 6621 if (dcp->c_backvp == NULL) { 6622 error = cachefs_getbackvp(fscp, dcp); 6623 if (error) { 6624 cachefs_nocache(dcp); 6625 mutex_exit(&dcp->c_statelock); 6626 goto out; 6627 } 6628 } 6629 6630 error = CFSOP_CHECK_COBJECT(fscp, dcp, 0, cr); 6631 if (error) { 6632 mutex_exit(&dcp->c_statelock); 6633 goto out; 6634 } 6635 CFS_DPRINT_BACKFS_NFSV4(fscp, 6636 ("cachefs_symlink (nfsv4): dcp %p, dbackvp %p, " 6637 "lnm %s, tnm %s\n", dcp, dcp->c_backvp, lnm, tnm)); 6638 error = VOP_SYMLINK(dcp->c_backvp, lnm, tva, tnm, cr); 6639 if (error) { 6640 mutex_exit(&dcp->c_statelock); 6641 goto out; 6642 } 6643 if ((dcp->c_filegrp->fg_flags & CFS_FG_WRITE) == 0 && 6644 !CFS_ISFS_BACKFS_NFSV4(fscp)) { 6645 cachefs_nocache(dcp); 6646 mutex_exit(&dcp->c_statelock); 6647 goto out; 6648 } 6649 6650 CFSOP_MODIFY_COBJECT(fscp, dcp, cr); 6651 6652 /* lookup the symlink we just created and get its fid and attrs */ 6653 (void) VOP_LOOKUP(dcp->c_backvp, lnm, &backvp, NULL, 0, NULL, cr); 6654 if (backvp == NULL) { 6655 if (CFS_ISFS_BACKFS_NFSV4(fscp) == 0) 6656 cachefs_nocache(dcp); 6657 mutex_exit(&dcp->c_statelock); 6658 goto out; 6659 } 6660 6661 valid_fid = (CFS_ISFS_BACKFS_NFSV4(fscp) ? FALSE : TRUE); 6662 error = cachefs_getcookie(backvp, &cookie, &va, cr, valid_fid); 6663 if (error) { 6664 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 6665 error = 0; 6666 cachefs_nocache(dcp); 6667 mutex_exit(&dcp->c_statelock); 6668 goto out; 6669 } 6670 cid.cid_fileno = va.va_nodeid; 6671 cid.cid_flags = 0; 6672 6673 /* if the dir is cached, add the symlink to it */ 6674 if (CFS_ISFS_NONSHARED(fscp) && 6675 (dcp->c_metadata.md_flags & MD_POPULATED)) { 6676 error = cachefs_dir_enter(dcp, lnm, &cookie, &cid, SM_ASYNC); 6677 if (error) { 6678 cachefs_nocache(dcp); 6679 error = 0; 6680 } 6681 } 6682 mutex_exit(&dcp->c_statelock); 6683 6684 /* make the cnode for the sym link */ 6685 error = cachefs_cnode_make(&cid, fscp, (valid_fid ? &cookie : NULL), 6686 &va, backvp, cr, 0, &newcp); 6687 if (error) { 6688 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 6689 cachefs_nocache(dcp); 6690 error = 0; 6691 goto out; 6692 } 6693 6694 /* try to cache the symlink contents */ 6695 rw_enter(&newcp->c_rwlock, RW_WRITER); 6696 mutex_enter(&newcp->c_statelock); 6697 6698 /* 6699 * try to cache the sym link, note that its a noop if NOCACHE 6700 * or NFSv4 is set 6701 */ 6702 error = cachefs_stuffsymlink(newcp, tnm, (int)newcp->c_size); 6703 if (error) { 6704 cachefs_nocache(newcp); 6705 error = 0; 6706 } 6707 mutex_exit(&newcp->c_statelock); 6708 rw_exit(&newcp->c_rwlock); 6709 6710 out: 6711 if (backvp) 6712 VN_RELE(backvp); 6713 if (newcp) 6714 VN_RELE(CTOV(newcp)); 6715 return (error); 6716 } 6717 6718 static int 6719 cachefs_symlink_disconnected(vnode_t *dvp, char *lnm, vattr_t *tva, 6720 char *tnm, cred_t *cr) 6721 { 6722 cnode_t *dcp = VTOC(dvp); 6723 fscache_t *fscp = C_TO_FSCACHE(dcp); 6724 int error; 6725 cnode_t *newcp = NULL; 6726 struct vattr va; 6727 timestruc_t current_time; 6728 off_t commit = 0; 6729 6730 if (CFS_ISFS_WRITE_AROUND(fscp)) 6731 return (ETIMEDOUT); 6732 6733 mutex_enter(&dcp->c_statelock); 6734 6735 /* check permissions */ 6736 if (error = cachefs_access_local(dcp, (VEXEC|VWRITE), cr)) { 6737 mutex_exit(&dcp->c_statelock); 6738 goto out; 6739 } 6740 6741 /* the directory front file must be populated */ 6742 if ((dcp->c_metadata.md_flags & MD_POPULATED) == 0) { 6743 error = ETIMEDOUT; 6744 mutex_exit(&dcp->c_statelock); 6745 goto out; 6746 } 6747 6748 /* make sure lnm does not already exist in the directory */ 6749 error = cachefs_dir_look(dcp, lnm, NULL, NULL, NULL, NULL); 6750 if (error == ENOTDIR) { 6751 error = ETIMEDOUT; 6752 mutex_exit(&dcp->c_statelock); 6753 goto out; 6754 } 6755 if (error != ENOENT) { 6756 error = EEXIST; 6757 mutex_exit(&dcp->c_statelock); 6758 goto out; 6759 } 6760 6761 /* make up a reasonable set of attributes */ 6762 cachefs_attr_setup(tva, &va, dcp, cr); 6763 va.va_type = VLNK; 6764 va.va_mode |= S_IFLNK; 6765 va.va_size = strlen(tnm); 6766 6767 mutex_exit(&dcp->c_statelock); 6768 6769 /* create the cnode */ 6770 error = cachefs_cnode_create(fscp, &va, 0, &newcp); 6771 if (error) 6772 goto out; 6773 6774 rw_enter(&newcp->c_rwlock, RW_WRITER); 6775 mutex_enter(&newcp->c_statelock); 6776 6777 error = cachefs_dlog_cidmap(fscp); 6778 if (error) { 6779 mutex_exit(&newcp->c_statelock); 6780 rw_exit(&newcp->c_rwlock); 6781 error = ENOSPC; 6782 goto out; 6783 } 6784 6785 cachefs_creategid(dcp, newcp, tva, cr); 6786 mutex_enter(&dcp->c_statelock); 6787 cachefs_createacl(dcp, newcp); 6788 mutex_exit(&dcp->c_statelock); 6789 gethrestime(¤t_time); 6790 newcp->c_metadata.md_vattr.va_atime = current_time; 6791 newcp->c_metadata.md_localctime = current_time; 6792 newcp->c_metadata.md_localmtime = current_time; 6793 newcp->c_metadata.md_flags |= MD_MAPPING | MD_LOCALMTIME | 6794 MD_LOCALCTIME; 6795 newcp->c_flags |= CN_UPDATED; 6796 6797 /* log the operation */ 6798 commit = cachefs_dlog_symlink(fscp, dcp, newcp, lnm, tva, tnm, cr); 6799 if (commit == 0) { 6800 mutex_exit(&newcp->c_statelock); 6801 rw_exit(&newcp->c_rwlock); 6802 error = ENOSPC; 6803 goto out; 6804 } 6805 6806 /* store the symlink contents */ 6807 error = cachefs_stuffsymlink(newcp, tnm, (int)newcp->c_size); 6808 if (error) { 6809 mutex_exit(&newcp->c_statelock); 6810 rw_exit(&newcp->c_rwlock); 6811 goto out; 6812 } 6813 if (cachefs_modified_alloc(newcp)) { 6814 mutex_exit(&newcp->c_statelock); 6815 rw_exit(&newcp->c_rwlock); 6816 error = ENOSPC; 6817 goto out; 6818 } 6819 6820 /* 6821 * write the metadata now rather than waiting until 6822 * inactive so that if there's no space we can let 6823 * the caller know. 6824 */ 6825 if (newcp->c_flags & CN_ALLOC_PENDING) { 6826 if (newcp->c_filegrp->fg_flags & CFS_FG_ALLOC_ATTR) { 6827 (void) filegrp_allocattr(newcp->c_filegrp); 6828 } 6829 error = filegrp_create_metadata(newcp->c_filegrp, 6830 &newcp->c_metadata, &newcp->c_id); 6831 if (error) { 6832 mutex_exit(&newcp->c_statelock); 6833 rw_exit(&newcp->c_rwlock); 6834 goto out; 6835 } 6836 newcp->c_flags &= ~CN_ALLOC_PENDING; 6837 } 6838 error = filegrp_write_metadata(newcp->c_filegrp, 6839 &newcp->c_id, &newcp->c_metadata); 6840 if (error) { 6841 mutex_exit(&newcp->c_statelock); 6842 rw_exit(&newcp->c_rwlock); 6843 goto out; 6844 } 6845 mutex_exit(&newcp->c_statelock); 6846 rw_exit(&newcp->c_rwlock); 6847 6848 mutex_enter(&dcp->c_statelock); 6849 6850 /* enter the new file in the directory */ 6851 if ((dcp->c_metadata.md_flags & MD_POPULATED) == 0) { 6852 error = ETIMEDOUT; 6853 mutex_exit(&dcp->c_statelock); 6854 goto out; 6855 } 6856 cachefs_modified(dcp); 6857 error = cachefs_dir_enter(dcp, lnm, &newcp->c_metadata.md_cookie, 6858 &newcp->c_id, SM_ASYNC); 6859 if (error) { 6860 mutex_exit(&dcp->c_statelock); 6861 goto out; 6862 } 6863 6864 /* update parent dir times */ 6865 dcp->c_metadata.md_localctime = current_time; 6866 dcp->c_metadata.md_localmtime = current_time; 6867 dcp->c_metadata.md_flags |= MD_LOCALMTIME | MD_LOCALCTIME; 6868 dcp->c_flags |= CN_UPDATED; 6869 mutex_exit(&dcp->c_statelock); 6870 6871 out: 6872 if (commit) { 6873 /* commit the log entry */ 6874 if (cachefs_dlog_commit(fscp, commit, error)) { 6875 /*EMPTY*/ 6876 /* XXX bob: fix on panic */ 6877 } 6878 } 6879 6880 if (error) { 6881 if (newcp) { 6882 mutex_enter(&newcp->c_statelock); 6883 newcp->c_flags |= CN_DESTROY; 6884 mutex_exit(&newcp->c_statelock); 6885 } 6886 } 6887 if (newcp) { 6888 VN_RELE(CTOV(newcp)); 6889 } 6890 6891 return (error); 6892 } 6893 6894 static int 6895 cachefs_readdir(vnode_t *vp, uio_t *uiop, cred_t *cr, int *eofp) 6896 { 6897 cnode_t *dcp = VTOC(vp); 6898 fscache_t *fscp = C_TO_FSCACHE(dcp); 6899 cachefscache_t *cachep = fscp->fs_cache; 6900 int error = 0; 6901 int held = 0; 6902 int connected = 0; 6903 6904 #ifdef CFSDEBUG 6905 CFS_DEBUG(CFSDEBUG_VOPS) 6906 printf("cachefs_readdir: ENTER vp %p\n", (void *)vp); 6907 #endif 6908 if (getzoneid() != GLOBAL_ZONEID) { 6909 error = EPERM; 6910 goto out; 6911 } 6912 6913 /* 6914 * Cachefs only provides pass-through support for NFSv4, 6915 * and all vnode operations are passed through to the 6916 * back file system. For NFSv4 pass-through to work, only 6917 * connected operation is supported, the cnode backvp must 6918 * exist, and cachefs optional (eg., disconnectable) flags 6919 * are turned off. Assert these conditions to ensure that 6920 * the backfilesystem is called for the readdir operation. 6921 */ 6922 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 6923 CFS_BACKFS_NFSV4_ASSERT_CNODE(dcp); 6924 6925 for (;;) { 6926 /* get (or renew) access to the file system */ 6927 if (held) { 6928 /* Won't loop with NFSv4 connected behavior */ 6929 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 6930 rw_exit(&dcp->c_rwlock); 6931 cachefs_cd_release(fscp); 6932 held = 0; 6933 } 6934 error = cachefs_cd_access(fscp, connected, 0); 6935 if (error) 6936 break; 6937 rw_enter(&dcp->c_rwlock, RW_READER); 6938 held = 1; 6939 6940 /* quit if link count of zero (posix) */ 6941 if (dcp->c_attr.va_nlink == 0) { 6942 if (eofp) 6943 *eofp = 1; 6944 error = 0; 6945 break; 6946 } 6947 6948 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 6949 error = cachefs_readdir_connected(vp, uiop, cr, 6950 eofp); 6951 if (CFS_TIMEOUT(fscp, error)) { 6952 rw_exit(&dcp->c_rwlock); 6953 cachefs_cd_release(fscp); 6954 held = 0; 6955 cachefs_cd_timedout(fscp); 6956 connected = 0; 6957 continue; 6958 } 6959 } else { 6960 error = cachefs_readdir_disconnected(vp, uiop, cr, 6961 eofp); 6962 if (CFS_TIMEOUT(fscp, error)) { 6963 if (cachefs_cd_access_miss(fscp)) { 6964 error = cachefs_readdir_connected(vp, 6965 uiop, cr, eofp); 6966 if (!CFS_TIMEOUT(fscp, error)) 6967 break; 6968 delay(5*hz); 6969 connected = 0; 6970 continue; 6971 } 6972 connected = 1; 6973 continue; 6974 } 6975 } 6976 break; 6977 } 6978 6979 if (CACHEFS_LOG_LOGGING(cachep, CACHEFS_LOG_READDIR)) 6980 cachefs_log_readdir(cachep, error, fscp->fs_cfsvfsp, 6981 &dcp->c_metadata.md_cookie, dcp->c_id.cid_fileno, 6982 crgetuid(cr), uiop->uio_loffset, *eofp); 6983 6984 if (held) { 6985 rw_exit(&dcp->c_rwlock); 6986 cachefs_cd_release(fscp); 6987 } 6988 6989 #ifdef CFS_CD_DEBUG 6990 ASSERT((curthread->t_flag & T_CD_HELD) == 0); 6991 #endif 6992 out: 6993 #ifdef CFSDEBUG 6994 CFS_DEBUG(CFSDEBUG_VOPS) 6995 printf("cachefs_readdir: EXIT error = %d\n", error); 6996 #endif 6997 6998 return (error); 6999 } 7000 7001 static int 7002 cachefs_readdir_connected(vnode_t *vp, uio_t *uiop, cred_t *cr, int *eofp) 7003 { 7004 cnode_t *dcp = VTOC(vp); 7005 int error; 7006 fscache_t *fscp = C_TO_FSCACHE(dcp); 7007 struct cachefs_req *rp; 7008 7009 mutex_enter(&dcp->c_statelock); 7010 7011 /* check directory consistency */ 7012 error = CFSOP_CHECK_COBJECT(fscp, dcp, 0, cr); 7013 if (error) 7014 goto out; 7015 dcp->c_usage++; 7016 7017 /* if dir was modified, toss old contents */ 7018 if (dcp->c_metadata.md_flags & MD_INVALREADDIR) { 7019 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 7020 cachefs_inval_object(dcp); 7021 } 7022 7023 error = 0; 7024 if (((dcp->c_metadata.md_flags & MD_POPULATED) == 0) && 7025 ((dcp->c_flags & (CN_ASYNC_POPULATE | CN_NOCACHE)) == 0) && 7026 !CFS_ISFS_BACKFS_NFSV4(fscp) && 7027 (fscp->fs_cdconnected == CFS_CD_CONNECTED)) { 7028 7029 if (cachefs_async_okay()) { 7030 7031 /* 7032 * Set up asynchronous request to fill this 7033 * directory. 7034 */ 7035 7036 dcp->c_flags |= CN_ASYNC_POPULATE; 7037 7038 rp = kmem_cache_alloc(cachefs_req_cache, KM_SLEEP); 7039 rp->cfs_cmd = CFS_POPULATE; 7040 rp->cfs_req_u.cu_populate.cpop_vp = vp; 7041 rp->cfs_cr = cr; 7042 7043 crhold(cr); 7044 VN_HOLD(vp); 7045 7046 cachefs_addqueue(rp, &fscp->fs_workq); 7047 } else { 7048 error = cachefs_dir_fill(dcp, cr); 7049 if (error != 0) 7050 cachefs_nocache(dcp); 7051 } 7052 } 7053 7054 /* if front file is populated */ 7055 if (((dcp->c_flags & (CN_NOCACHE | CN_ASYNC_POPULATE)) == 0) && 7056 !CFS_ISFS_BACKFS_NFSV4(fscp) && 7057 (dcp->c_metadata.md_flags & MD_POPULATED)) { 7058 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 7059 error = cachefs_dir_read(dcp, uiop, eofp); 7060 if (error == 0) 7061 fscp->fs_stats.st_hits++; 7062 } 7063 7064 /* if front file could not be used */ 7065 if ((error != 0) || 7066 CFS_ISFS_BACKFS_NFSV4(fscp) || 7067 (dcp->c_flags & (CN_NOCACHE | CN_ASYNC_POPULATE)) || 7068 ((dcp->c_metadata.md_flags & MD_POPULATED) == 0)) { 7069 7070 if (error && !(dcp->c_flags & CN_NOCACHE) && 7071 !CFS_ISFS_BACKFS_NFSV4(fscp)) 7072 cachefs_nocache(dcp); 7073 7074 /* get the back vp */ 7075 if (dcp->c_backvp == NULL) { 7076 error = cachefs_getbackvp(fscp, dcp); 7077 if (error) 7078 goto out; 7079 } 7080 7081 if (fscp->fs_inum_size > 0) { 7082 error = cachefs_readback_translate(dcp, uiop, cr, eofp); 7083 } else { 7084 /* do the dir read from the back fs */ 7085 (void) VOP_RWLOCK(dcp->c_backvp, 7086 V_WRITELOCK_FALSE, NULL); 7087 CFS_DPRINT_BACKFS_NFSV4(fscp, 7088 ("cachefs_readdir (nfsv4): " 7089 "dcp %p, dbackvp %p\n", dcp, dcp->c_backvp)); 7090 error = VOP_READDIR(dcp->c_backvp, uiop, cr, eofp); 7091 VOP_RWUNLOCK(dcp->c_backvp, V_WRITELOCK_FALSE, NULL); 7092 } 7093 7094 if (error == 0) 7095 fscp->fs_stats.st_misses++; 7096 } 7097 7098 out: 7099 mutex_exit(&dcp->c_statelock); 7100 7101 return (error); 7102 } 7103 7104 static int 7105 cachefs_readback_translate(cnode_t *cp, uio_t *uiop, cred_t *cr, int *eofp) 7106 { 7107 int error = 0; 7108 fscache_t *fscp = C_TO_FSCACHE(cp); 7109 caddr_t buffy = NULL; 7110 int buffysize = MAXBSIZE; 7111 caddr_t chrp, end; 7112 ino64_t newinum; 7113 struct dirent64 *de; 7114 uio_t uioin; 7115 iovec_t iov; 7116 7117 ASSERT(cp->c_backvp != NULL); 7118 ASSERT(fscp->fs_inum_size > 0); 7119 7120 if (uiop->uio_resid < buffysize) 7121 buffysize = (int)uiop->uio_resid; 7122 buffy = cachefs_kmem_alloc(buffysize, KM_SLEEP); 7123 7124 iov.iov_base = buffy; 7125 iov.iov_len = buffysize; 7126 uioin.uio_iov = &iov; 7127 uioin.uio_iovcnt = 1; 7128 uioin.uio_segflg = UIO_SYSSPACE; 7129 uioin.uio_fmode = 0; 7130 uioin.uio_extflg = UIO_COPY_CACHED; 7131 uioin.uio_loffset = uiop->uio_loffset; 7132 uioin.uio_resid = buffysize; 7133 7134 (void) VOP_RWLOCK(cp->c_backvp, V_WRITELOCK_FALSE, NULL); 7135 error = VOP_READDIR(cp->c_backvp, &uioin, cr, eofp); 7136 VOP_RWUNLOCK(cp->c_backvp, V_WRITELOCK_FALSE, NULL); 7137 7138 if (error != 0) 7139 goto out; 7140 7141 end = buffy + buffysize - uioin.uio_resid; 7142 7143 mutex_exit(&cp->c_statelock); 7144 mutex_enter(&fscp->fs_fslock); 7145 7146 7147 for (chrp = buffy; chrp < end; chrp += de->d_reclen) { 7148 de = (dirent64_t *)chrp; 7149 newinum = cachefs_inum_real2fake(fscp, de->d_ino); 7150 if (newinum == 0) 7151 newinum = cachefs_fileno_conflict(fscp, de->d_ino); 7152 de->d_ino = newinum; 7153 } 7154 mutex_exit(&fscp->fs_fslock); 7155 mutex_enter(&cp->c_statelock); 7156 7157 error = uiomove(buffy, end - buffy, UIO_READ, uiop); 7158 uiop->uio_loffset = uioin.uio_loffset; 7159 7160 out: 7161 7162 if (buffy != NULL) 7163 cachefs_kmem_free(buffy, buffysize); 7164 7165 return (error); 7166 } 7167 7168 static int 7169 /*ARGSUSED*/ 7170 cachefs_readdir_disconnected(vnode_t *vp, uio_t *uiop, cred_t *cr, 7171 int *eofp) 7172 { 7173 cnode_t *dcp = VTOC(vp); 7174 int error; 7175 7176 mutex_enter(&dcp->c_statelock); 7177 if ((dcp->c_metadata.md_flags & MD_POPULATED) == 0) { 7178 error = ETIMEDOUT; 7179 } else { 7180 error = cachefs_dir_read(dcp, uiop, eofp); 7181 if (error == ENOTDIR) 7182 error = ETIMEDOUT; 7183 } 7184 mutex_exit(&dcp->c_statelock); 7185 7186 return (error); 7187 } 7188 7189 static int 7190 cachefs_fid(struct vnode *vp, struct fid *fidp) 7191 { 7192 int error = 0; 7193 struct cnode *cp = VTOC(vp); 7194 fscache_t *fscp = C_TO_FSCACHE(cp); 7195 7196 /* 7197 * Cachefs only provides pass-through support for NFSv4, 7198 * and all vnode operations are passed through to the 7199 * back file system. For NFSv4 pass-through to work, only 7200 * connected operation is supported, the cnode backvp must 7201 * exist, and cachefs optional (eg., disconnectable) flags 7202 * are turned off. Assert these conditions, then bail 7203 * as NFSv4 doesn't support VOP_FID. 7204 */ 7205 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 7206 CFS_BACKFS_NFSV4_ASSERT_CNODE(cp); 7207 if (CFS_ISFS_BACKFS_NFSV4(fscp)) { 7208 return (ENOTSUP); 7209 } 7210 7211 mutex_enter(&cp->c_statelock); 7212 if (fidp->fid_len < cp->c_metadata.md_cookie.fid_len) { 7213 fidp->fid_len = cp->c_metadata.md_cookie.fid_len; 7214 error = ENOSPC; 7215 } else { 7216 bcopy(cp->c_metadata.md_cookie.fid_data, fidp->fid_data, 7217 cp->c_metadata.md_cookie.fid_len); 7218 fidp->fid_len = cp->c_metadata.md_cookie.fid_len; 7219 } 7220 mutex_exit(&cp->c_statelock); 7221 return (error); 7222 } 7223 7224 /* ARGSUSED2 */ 7225 static int 7226 cachefs_rwlock(struct vnode *vp, int write_lock, caller_context_t *ctp) 7227 { 7228 cnode_t *cp = VTOC(vp); 7229 7230 /* 7231 * XXX - This is ifdef'ed out for now. The problem - 7232 * getdents() acquires the read version of rwlock, then we come 7233 * into cachefs_readdir() and that wants to acquire the write version 7234 * of this lock (if its going to populate the directory). This is 7235 * a problem, this can be solved by introducing another lock in the 7236 * cnode. 7237 */ 7238 /* XXX */ 7239 if (vp->v_type != VREG) 7240 return (-1); 7241 if (write_lock) 7242 rw_enter(&cp->c_rwlock, RW_WRITER); 7243 else 7244 rw_enter(&cp->c_rwlock, RW_READER); 7245 return (write_lock); 7246 } 7247 7248 /* ARGSUSED */ 7249 static void 7250 cachefs_rwunlock(struct vnode *vp, int write_lock, caller_context_t *ctp) 7251 { 7252 cnode_t *cp = VTOC(vp); 7253 if (vp->v_type != VREG) 7254 return; 7255 rw_exit(&cp->c_rwlock); 7256 } 7257 7258 /* ARGSUSED */ 7259 static int 7260 cachefs_seek(struct vnode *vp, offset_t ooff, offset_t *noffp) 7261 { 7262 return (0); 7263 } 7264 7265 static int cachefs_lostpage = 0; 7266 /* 7267 * Return all the pages from [off..off+len] in file 7268 */ 7269 static int 7270 cachefs_getpage(struct vnode *vp, offset_t off, size_t len, 7271 uint_t *protp, struct page *pl[], size_t plsz, struct seg *seg, 7272 caddr_t addr, enum seg_rw rw, cred_t *cr) 7273 { 7274 cnode_t *cp = VTOC(vp); 7275 int error; 7276 fscache_t *fscp = C_TO_FSCACHE(cp); 7277 cachefscache_t *cachep = fscp->fs_cache; 7278 int held = 0; 7279 int connected = 0; 7280 7281 #ifdef CFSDEBUG 7282 u_offset_t offx = (u_offset_t)off; 7283 7284 CFS_DEBUG(CFSDEBUG_VOPS) 7285 printf("cachefs_getpage: ENTER vp %p off %lld len %lu rw %d\n", 7286 (void *)vp, offx, len, rw); 7287 #endif 7288 if (getzoneid() != GLOBAL_ZONEID) { 7289 error = EPERM; 7290 goto out; 7291 } 7292 7293 if (vp->v_flag & VNOMAP) { 7294 error = ENOSYS; 7295 goto out; 7296 } 7297 7298 /* Call backfilesystem if NFSv4 */ 7299 if (CFS_ISFS_BACKFS_NFSV4(fscp)) { 7300 error = cachefs_getpage_backfs_nfsv4(vp, off, len, protp, pl, 7301 plsz, seg, addr, rw, cr); 7302 goto out; 7303 } 7304 7305 /* XXX sam: make this do an async populate? */ 7306 if (pl == NULL) { 7307 error = 0; 7308 goto out; 7309 } 7310 if (protp != NULL) 7311 *protp = PROT_ALL; 7312 7313 for (;;) { 7314 /* get (or renew) access to the file system */ 7315 if (held) { 7316 cachefs_cd_release(fscp); 7317 held = 0; 7318 } 7319 error = cachefs_cd_access(fscp, connected, 0); 7320 if (error) 7321 break; 7322 held = 1; 7323 7324 /* 7325 * If we are getting called as a side effect of a 7326 * cachefs_write() 7327 * operation the local file size might not be extended yet. 7328 * In this case we want to be able to return pages of zeroes. 7329 */ 7330 if ((u_offset_t)off + len > 7331 ((cp->c_size + PAGEOFFSET) & (offset_t)PAGEMASK)) { 7332 if (seg != segkmap) { 7333 error = EFAULT; 7334 break; 7335 } 7336 } 7337 if (len <= PAGESIZE) 7338 error = cachefs_getapage(vp, (u_offset_t)off, len, 7339 protp, pl, plsz, seg, addr, rw, cr); 7340 else 7341 error = pvn_getpages(cachefs_getapage, vp, 7342 (u_offset_t)off, len, protp, pl, plsz, seg, addr, 7343 rw, cr); 7344 if (error == 0) 7345 break; 7346 7347 if (((cp->c_flags & CN_NOCACHE) && (error == ENOSPC)) || 7348 error == EAGAIN) { 7349 connected = 0; 7350 continue; 7351 } 7352 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 7353 if (CFS_TIMEOUT(fscp, error)) { 7354 cachefs_cd_release(fscp); 7355 held = 0; 7356 cachefs_cd_timedout(fscp); 7357 connected = 0; 7358 continue; 7359 } 7360 } else { 7361 if (CFS_TIMEOUT(fscp, error)) { 7362 if (cachefs_cd_access_miss(fscp)) { 7363 if (len <= PAGESIZE) 7364 error = cachefs_getapage_back( 7365 vp, (u_offset_t)off, 7366 len, protp, pl, 7367 plsz, seg, addr, rw, cr); 7368 else 7369 error = pvn_getpages( 7370 cachefs_getapage_back, vp, 7371 (u_offset_t)off, len, 7372 protp, pl, 7373 plsz, seg, addr, rw, cr); 7374 if (!CFS_TIMEOUT(fscp, error) && 7375 (error != EAGAIN)) 7376 break; 7377 delay(5*hz); 7378 connected = 0; 7379 continue; 7380 } 7381 connected = 1; 7382 continue; 7383 } 7384 } 7385 break; 7386 } 7387 7388 if (CACHEFS_LOG_LOGGING(cachep, CACHEFS_LOG_GETPAGE)) 7389 cachefs_log_getpage(cachep, error, vp->v_vfsp, 7390 &cp->c_metadata.md_cookie, cp->c_id.cid_fileno, 7391 crgetuid(cr), off, len); 7392 7393 if (held) { 7394 cachefs_cd_release(fscp); 7395 } 7396 7397 out: 7398 #ifdef CFS_CD_DEBUG 7399 ASSERT((curthread->t_flag & T_CD_HELD) == 0); 7400 #endif 7401 #ifdef CFSDEBUG 7402 CFS_DEBUG(CFSDEBUG_VOPS) 7403 printf("cachefs_getpage: EXIT vp %p error %d\n", 7404 (void *)vp, error); 7405 #endif 7406 return (error); 7407 } 7408 7409 /* 7410 * cachefs_getpage_backfs_nfsv4 7411 * 7412 * Call NFSv4 back filesystem to handle the getpage (cachefs 7413 * pass-through support for NFSv4). 7414 */ 7415 static int 7416 cachefs_getpage_backfs_nfsv4(struct vnode *vp, offset_t off, size_t len, 7417 uint_t *protp, struct page *pl[], size_t plsz, 7418 struct seg *seg, caddr_t addr, enum seg_rw rw, 7419 cred_t *cr) 7420 { 7421 cnode_t *cp = VTOC(vp); 7422 fscache_t *fscp = C_TO_FSCACHE(cp); 7423 vnode_t *backvp; 7424 int error; 7425 7426 /* 7427 * For NFSv4 pass-through to work, only connected operation is 7428 * supported, the cnode backvp must exist, and cachefs optional 7429 * (eg., disconnectable) flags are turned off. Assert these 7430 * conditions for the getpage operation. 7431 */ 7432 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 7433 CFS_BACKFS_NFSV4_ASSERT_CNODE(cp); 7434 7435 /* Call backfs vnode op after extracting backvp */ 7436 mutex_enter(&cp->c_statelock); 7437 backvp = cp->c_backvp; 7438 mutex_exit(&cp->c_statelock); 7439 7440 CFS_DPRINT_BACKFS_NFSV4(fscp, 7441 ("cachefs_getpage_backfs_nfsv4: cnode %p, backvp %p\n", 7442 cp, backvp)); 7443 error = VOP_GETPAGE(backvp, off, len, protp, pl, plsz, seg, 7444 addr, rw, cr); 7445 7446 return (error); 7447 } 7448 7449 /* 7450 * Called from pvn_getpages or cachefs_getpage to get a particular page. 7451 */ 7452 /*ARGSUSED*/ 7453 static int 7454 cachefs_getapage(struct vnode *vp, u_offset_t off, size_t len, uint_t *protp, 7455 struct page *pl[], size_t plsz, struct seg *seg, caddr_t addr, 7456 enum seg_rw rw, cred_t *cr) 7457 { 7458 cnode_t *cp = VTOC(vp); 7459 page_t **ppp, *pp = NULL; 7460 fscache_t *fscp = C_TO_FSCACHE(cp); 7461 cachefscache_t *cachep = fscp->fs_cache; 7462 int error = 0; 7463 struct page **ourpl; 7464 struct page *ourstackpl[17]; /* see ASSERT() below for 17 */ 7465 int index = 0; 7466 int downgrade; 7467 int have_statelock = 0; 7468 u_offset_t popoff; 7469 size_t popsize = 0; 7470 7471 ASSERT(((DEF_POP_SIZE / PAGESIZE) + 1) <= 17); 7472 7473 if (fscp->fs_info.fi_popsize > DEF_POP_SIZE) 7474 ourpl = cachefs_kmem_alloc(sizeof (struct page *) * 7475 ((fscp->fs_info.fi_popsize / PAGESIZE) + 1), KM_SLEEP); 7476 else 7477 ourpl = ourstackpl; 7478 7479 ourpl[0] = NULL; 7480 off = off & (offset_t)PAGEMASK; 7481 again: 7482 /* 7483 * Look for the page 7484 */ 7485 if (page_exists(vp, off) == 0) { 7486 /* 7487 * Need to do work to get the page. 7488 * Grab our lock because we are going to 7489 * modify the state of the cnode. 7490 */ 7491 if (! have_statelock) { 7492 mutex_enter(&cp->c_statelock); 7493 have_statelock = 1; 7494 } 7495 /* 7496 * If we're in NOCACHE mode, we will need a backvp 7497 */ 7498 if (cp->c_flags & CN_NOCACHE) { 7499 if (fscp->fs_cdconnected != CFS_CD_CONNECTED) { 7500 error = ETIMEDOUT; 7501 goto out; 7502 } 7503 if (cp->c_backvp == NULL) { 7504 error = cachefs_getbackvp(fscp, cp); 7505 if (error) 7506 goto out; 7507 } 7508 error = VOP_GETPAGE(cp->c_backvp, off, 7509 PAGESIZE, protp, ourpl, PAGESIZE, seg, 7510 addr, S_READ, cr); 7511 /* 7512 * backfs returns EFAULT when we are trying for a 7513 * page beyond EOF but cachefs has the knowledge that 7514 * it is not beyond EOF be cause cp->c_size is 7515 * greater then the offset requested. 7516 */ 7517 if (error == EFAULT) { 7518 error = 0; 7519 pp = page_create_va(vp, off, PAGESIZE, 7520 PG_EXCL | PG_WAIT, seg, addr); 7521 if (pp == NULL) 7522 goto again; 7523 pagezero(pp, 0, PAGESIZE); 7524 pvn_plist_init(pp, pl, plsz, off, PAGESIZE, rw); 7525 goto out; 7526 } 7527 if (error) 7528 goto out; 7529 goto getpages; 7530 } 7531 /* 7532 * We need a front file. If we can't get it, 7533 * put the cnode in NOCACHE mode and try again. 7534 */ 7535 if (cp->c_frontvp == NULL) { 7536 error = cachefs_getfrontfile(cp); 7537 if (error) { 7538 cachefs_nocache(cp); 7539 error = EAGAIN; 7540 goto out; 7541 } 7542 } 7543 /* 7544 * Check if the front file needs population. 7545 * If population is necessary, make sure we have a 7546 * backvp as well. We will get the page from the backvp. 7547 * bug 4152459- 7548 * But if the file system is in disconnected mode 7549 * and the file is a local file then do not check the 7550 * allocmap. 7551 */ 7552 if (((fscp->fs_cdconnected == CFS_CD_CONNECTED) || 7553 ((cp->c_metadata.md_flags & MD_LOCALFILENO) == 0)) && 7554 (cachefs_check_allocmap(cp, off) == 0)) { 7555 if (fscp->fs_cdconnected != CFS_CD_CONNECTED) { 7556 error = ETIMEDOUT; 7557 goto out; 7558 } 7559 if (cp->c_backvp == NULL) { 7560 error = cachefs_getbackvp(fscp, cp); 7561 if (error) 7562 goto out; 7563 } 7564 if (cp->c_filegrp->fg_flags & CFS_FG_WRITE) { 7565 cachefs_cluster_allocmap(off, &popoff, 7566 &popsize, 7567 fscp->fs_info.fi_popsize, cp); 7568 if (popsize != 0) { 7569 error = cachefs_populate(cp, 7570 popoff, popsize, 7571 cp->c_frontvp, cp->c_backvp, 7572 cp->c_size, cr); 7573 if (error) { 7574 cachefs_nocache(cp); 7575 error = EAGAIN; 7576 goto out; 7577 } else { 7578 cp->c_flags |= 7579 CN_UPDATED | 7580 CN_NEED_FRONT_SYNC | 7581 CN_POPULATION_PENDING; 7582 } 7583 popsize = popsize - (off - popoff); 7584 } else { 7585 popsize = PAGESIZE; 7586 } 7587 } 7588 /* else XXX assert CN_NOCACHE? */ 7589 error = VOP_GETPAGE(cp->c_backvp, (offset_t)off, 7590 PAGESIZE, protp, ourpl, popsize, 7591 seg, addr, S_READ, cr); 7592 if (error) 7593 goto out; 7594 fscp->fs_stats.st_misses++; 7595 } else { 7596 if (cp->c_flags & CN_POPULATION_PENDING) { 7597 error = VOP_FSYNC(cp->c_frontvp, FSYNC, cr); 7598 cp->c_flags &= ~CN_POPULATION_PENDING; 7599 if (error) { 7600 cachefs_nocache(cp); 7601 error = EAGAIN; 7602 goto out; 7603 } 7604 } 7605 /* 7606 * File was populated so we get the page from the 7607 * frontvp 7608 */ 7609 error = VOP_GETPAGE(cp->c_frontvp, (offset_t)off, 7610 PAGESIZE, protp, ourpl, PAGESIZE, seg, addr, 7611 rw, cr); 7612 if (CACHEFS_LOG_LOGGING(cachep, CACHEFS_LOG_GPFRONT)) 7613 cachefs_log_gpfront(cachep, error, 7614 fscp->fs_cfsvfsp, 7615 &cp->c_metadata.md_cookie, cp->c_fileno, 7616 crgetuid(cr), off, PAGESIZE); 7617 if (error) { 7618 cachefs_nocache(cp); 7619 error = EAGAIN; 7620 goto out; 7621 } 7622 fscp->fs_stats.st_hits++; 7623 } 7624 getpages: 7625 ASSERT(have_statelock); 7626 if (have_statelock) { 7627 mutex_exit(&cp->c_statelock); 7628 have_statelock = 0; 7629 } 7630 downgrade = 0; 7631 for (ppp = ourpl; *ppp; ppp++) { 7632 if ((*ppp)->p_offset < off) { 7633 index++; 7634 page_unlock(*ppp); 7635 continue; 7636 } 7637 if (PAGE_SHARED(*ppp)) { 7638 if (page_tryupgrade(*ppp) == 0) { 7639 for (ppp = &ourpl[index]; *ppp; ppp++) 7640 page_unlock(*ppp); 7641 error = EAGAIN; 7642 goto out; 7643 } 7644 downgrade = 1; 7645 } 7646 ASSERT(PAGE_EXCL(*ppp)); 7647 (void) hat_pageunload((*ppp), HAT_FORCE_PGUNLOAD); 7648 page_rename(*ppp, vp, (*ppp)->p_offset); 7649 } 7650 pl[0] = ourpl[index]; 7651 pl[1] = NULL; 7652 if (downgrade) { 7653 page_downgrade(ourpl[index]); 7654 } 7655 /* Unlock the rest of the pages from the cluster */ 7656 for (ppp = &ourpl[index+1]; *ppp; ppp++) 7657 page_unlock(*ppp); 7658 } else { 7659 ASSERT(! have_statelock); 7660 if (have_statelock) { 7661 mutex_exit(&cp->c_statelock); 7662 have_statelock = 0; 7663 } 7664 /* XXX SE_SHARED probably isn't what we *always* want */ 7665 if ((pp = page_lookup(vp, off, SE_SHARED)) == NULL) { 7666 cachefs_lostpage++; 7667 goto again; 7668 } 7669 pl[0] = pp; 7670 pl[1] = NULL; 7671 /* XXX increment st_hits? i don't think so, but... */ 7672 } 7673 7674 out: 7675 if (have_statelock) { 7676 mutex_exit(&cp->c_statelock); 7677 have_statelock = 0; 7678 } 7679 if (fscp->fs_info.fi_popsize > DEF_POP_SIZE) 7680 cachefs_kmem_free(ourpl, sizeof (struct page *) * 7681 ((fscp->fs_info.fi_popsize / PAGESIZE) + 1)); 7682 return (error); 7683 } 7684 7685 /* gets a page but only from the back fs */ 7686 /*ARGSUSED*/ 7687 static int 7688 cachefs_getapage_back(struct vnode *vp, u_offset_t off, size_t len, 7689 uint_t *protp, struct page *pl[], size_t plsz, struct seg *seg, 7690 caddr_t addr, enum seg_rw rw, cred_t *cr) 7691 { 7692 cnode_t *cp = VTOC(vp); 7693 page_t **ppp, *pp = NULL; 7694 fscache_t *fscp = C_TO_FSCACHE(cp); 7695 int error = 0; 7696 struct page *ourpl[17]; 7697 int index = 0; 7698 int have_statelock = 0; 7699 int downgrade; 7700 7701 /* 7702 * Grab the cnode statelock so the cnode state won't change 7703 * while we're in here. 7704 */ 7705 ourpl[0] = NULL; 7706 off = off & (offset_t)PAGEMASK; 7707 again: 7708 if (page_exists(vp, off) == 0) { 7709 if (! have_statelock) { 7710 mutex_enter(&cp->c_statelock); 7711 have_statelock = 1; 7712 } 7713 7714 if (cp->c_backvp == NULL) { 7715 error = cachefs_getbackvp(fscp, cp); 7716 if (error) 7717 goto out; 7718 } 7719 error = VOP_GETPAGE(cp->c_backvp, (offset_t)off, 7720 PAGESIZE, protp, ourpl, PAGESIZE, seg, 7721 addr, S_READ, cr); 7722 if (error) 7723 goto out; 7724 7725 if (have_statelock) { 7726 mutex_exit(&cp->c_statelock); 7727 have_statelock = 0; 7728 } 7729 downgrade = 0; 7730 for (ppp = ourpl; *ppp; ppp++) { 7731 if ((*ppp)->p_offset < off) { 7732 index++; 7733 page_unlock(*ppp); 7734 continue; 7735 } 7736 if (PAGE_SHARED(*ppp)) { 7737 if (page_tryupgrade(*ppp) == 0) { 7738 for (ppp = &ourpl[index]; *ppp; ppp++) 7739 page_unlock(*ppp); 7740 error = EAGAIN; 7741 goto out; 7742 } 7743 downgrade = 1; 7744 } 7745 ASSERT(PAGE_EXCL(*ppp)); 7746 (void) hat_pageunload((*ppp), HAT_FORCE_PGUNLOAD); 7747 page_rename(*ppp, vp, (*ppp)->p_offset); 7748 } 7749 pl[0] = ourpl[index]; 7750 pl[1] = NULL; 7751 if (downgrade) { 7752 page_downgrade(ourpl[index]); 7753 } 7754 /* Unlock the rest of the pages from the cluster */ 7755 for (ppp = &ourpl[index+1]; *ppp; ppp++) 7756 page_unlock(*ppp); 7757 } else { 7758 ASSERT(! have_statelock); 7759 if (have_statelock) { 7760 mutex_exit(&cp->c_statelock); 7761 have_statelock = 0; 7762 } 7763 if ((pp = page_lookup(vp, off, SE_SHARED)) == NULL) { 7764 cachefs_lostpage++; 7765 goto again; 7766 } 7767 pl[0] = pp; 7768 pl[1] = NULL; 7769 } 7770 7771 out: 7772 if (have_statelock) { 7773 mutex_exit(&cp->c_statelock); 7774 have_statelock = 0; 7775 } 7776 return (error); 7777 } 7778 7779 static int 7780 cachefs_putpage(vnode_t *vp, offset_t off, size_t len, int flags, cred_t *cr) 7781 { 7782 cnode_t *cp = VTOC(vp); 7783 int error = 0; 7784 fscache_t *fscp = C_TO_FSCACHE(cp); 7785 int held = 0; 7786 int connected = 0; 7787 7788 if (getzoneid() != GLOBAL_ZONEID) 7789 return (EPERM); 7790 7791 /* Call backfilesytem if NFSv4 */ 7792 if (CFS_ISFS_BACKFS_NFSV4(fscp)) { 7793 error = cachefs_putpage_backfs_nfsv4(vp, off, len, flags, cr); 7794 goto out; 7795 } 7796 7797 for (;;) { 7798 /* get (or renew) access to the file system */ 7799 if (held) { 7800 cachefs_cd_release(fscp); 7801 held = 0; 7802 } 7803 error = cachefs_cd_access(fscp, connected, 1); 7804 if (error) 7805 break; 7806 held = 1; 7807 7808 error = cachefs_putpage_common(vp, off, len, flags, cr); 7809 if (error == 0) 7810 break; 7811 7812 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 7813 if (CFS_TIMEOUT(fscp, error)) { 7814 cachefs_cd_release(fscp); 7815 held = 0; 7816 cachefs_cd_timedout(fscp); 7817 connected = 0; 7818 continue; 7819 } 7820 } else { 7821 if (NOMEMWAIT()) { 7822 error = 0; 7823 goto out; 7824 } 7825 if (CFS_TIMEOUT(fscp, error)) { 7826 connected = 1; 7827 continue; 7828 } 7829 } 7830 break; 7831 } 7832 7833 out: 7834 7835 if (held) { 7836 cachefs_cd_release(fscp); 7837 } 7838 7839 #ifdef CFS_CD_DEBUG 7840 ASSERT((curthread->t_flag & T_CD_HELD) == 0); 7841 #endif 7842 return (error); 7843 } 7844 7845 /* 7846 * cachefs_putpage_backfs_nfsv4 7847 * 7848 * Call NFSv4 back filesystem to handle the putpage (cachefs 7849 * pass-through support for NFSv4). 7850 */ 7851 static int 7852 cachefs_putpage_backfs_nfsv4(vnode_t *vp, offset_t off, size_t len, int flags, 7853 cred_t *cr) 7854 { 7855 cnode_t *cp = VTOC(vp); 7856 fscache_t *fscp = C_TO_FSCACHE(cp); 7857 vnode_t *backvp; 7858 int error; 7859 7860 /* 7861 * For NFSv4 pass-through to work, only connected operation is 7862 * supported, the cnode backvp must exist, and cachefs optional 7863 * (eg., disconnectable) flags are turned off. Assert these 7864 * conditions for the putpage operation. 7865 */ 7866 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 7867 CFS_BACKFS_NFSV4_ASSERT_CNODE(cp); 7868 7869 /* Call backfs vnode op after extracting backvp */ 7870 mutex_enter(&cp->c_statelock); 7871 backvp = cp->c_backvp; 7872 mutex_exit(&cp->c_statelock); 7873 7874 CFS_DPRINT_BACKFS_NFSV4(fscp, 7875 ("cachefs_putpage_backfs_nfsv4: cnode %p, backvp %p\n", 7876 cp, backvp)); 7877 error = VOP_PUTPAGE(backvp, off, len, flags, cr); 7878 7879 return (error); 7880 } 7881 7882 /* 7883 * Flags are composed of {B_INVAL, B_FREE, B_DONTNEED, B_FORCE} 7884 * If len == 0, do from off to EOF. 7885 * 7886 * The normal cases should be len == 0 & off == 0 (entire vp list), 7887 * len == MAXBSIZE (from segmap_release actions), and len == PAGESIZE 7888 * (from pageout). 7889 */ 7890 7891 /*ARGSUSED*/ 7892 int 7893 cachefs_putpage_common(struct vnode *vp, offset_t off, size_t len, 7894 int flags, cred_t *cr) 7895 { 7896 struct cnode *cp = VTOC(vp); 7897 struct page *pp; 7898 size_t io_len; 7899 u_offset_t eoff, io_off; 7900 int error = 0; 7901 fscache_t *fscp = C_TO_FSCACHE(cp); 7902 cachefscache_t *cachep = fscp->fs_cache; 7903 7904 if (len == 0 && (flags & B_INVAL) == 0 && vn_is_readonly(vp)) { 7905 return (0); 7906 } 7907 if (!vn_has_cached_data(vp) || (off >= cp->c_size && 7908 (flags & B_INVAL) == 0)) 7909 return (0); 7910 7911 /* 7912 * Should never have cached data for the cachefs vnode 7913 * if NFSv4 is in use. 7914 */ 7915 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 7916 7917 /* 7918 * If this is an async putpage let a thread handle it. 7919 */ 7920 if (flags & B_ASYNC) { 7921 struct cachefs_req *rp; 7922 int tflags = (flags & ~(B_ASYNC|B_DONTNEED)); 7923 7924 if (ttoproc(curthread) == proc_pageout) { 7925 /* 7926 * If this is the page daemon we 7927 * do the push synchronously (Dangerous!) and hope 7928 * we can free enough to keep running... 7929 */ 7930 flags &= ~B_ASYNC; 7931 goto again; 7932 } 7933 7934 if (! cachefs_async_okay()) { 7935 7936 /* 7937 * this is somewhat like NFS's behavior. keep 7938 * the system from thrashing. we've seen 7939 * cases where async queues get out of 7940 * control, especially if 7941 * madvise(MADV_SEQUENTIAL) is done on a large 7942 * mmap()ed file that is read sequentially. 7943 */ 7944 7945 flags &= ~B_ASYNC; 7946 goto again; 7947 } 7948 7949 /* 7950 * if no flags other than B_ASYNC were set, 7951 * we coalesce putpage requests into a single one for the 7952 * whole file (len = off = 0). If such a request is 7953 * already queued, we're done. 7954 * 7955 * If there are other flags set (e.g., B_INVAL), we don't 7956 * attempt to coalesce and we use the specified length and 7957 * offset. 7958 */ 7959 rp = kmem_cache_alloc(cachefs_req_cache, KM_SLEEP); 7960 mutex_enter(&cp->c_iomutex); 7961 if ((cp->c_ioflags & CIO_PUTPAGES) == 0 || tflags != 0) { 7962 rp->cfs_cmd = CFS_PUTPAGE; 7963 rp->cfs_req_u.cu_putpage.cp_vp = vp; 7964 if (tflags == 0) { 7965 off = len = 0; 7966 cp->c_ioflags |= CIO_PUTPAGES; 7967 } 7968 rp->cfs_req_u.cu_putpage.cp_off = off; 7969 rp->cfs_req_u.cu_putpage.cp_len = (uint_t)len; 7970 rp->cfs_req_u.cu_putpage.cp_flags = flags & ~B_ASYNC; 7971 rp->cfs_cr = cr; 7972 crhold(rp->cfs_cr); 7973 VN_HOLD(vp); 7974 cp->c_nio++; 7975 cachefs_addqueue(rp, &(C_TO_FSCACHE(cp)->fs_workq)); 7976 } else { 7977 kmem_cache_free(cachefs_req_cache, rp); 7978 } 7979 7980 mutex_exit(&cp->c_iomutex); 7981 return (0); 7982 } 7983 7984 7985 again: 7986 if (len == 0) { 7987 /* 7988 * Search the entire vp list for pages >= off 7989 */ 7990 error = pvn_vplist_dirty(vp, off, cachefs_push, flags, cr); 7991 } else { 7992 /* 7993 * Do a range from [off...off + len] looking for pages 7994 * to deal with. 7995 */ 7996 eoff = (u_offset_t)off + len; 7997 for (io_off = off; io_off < eoff && io_off < cp->c_size; 7998 io_off += io_len) { 7999 /* 8000 * If we are not invalidating, synchronously 8001 * freeing or writing pages use the routine 8002 * page_lookup_nowait() to prevent reclaiming 8003 * them from the free list. 8004 */ 8005 if ((flags & B_INVAL) || ((flags & B_ASYNC) == 0)) { 8006 pp = page_lookup(vp, io_off, 8007 (flags & (B_INVAL | B_FREE)) ? 8008 SE_EXCL : SE_SHARED); 8009 } else { 8010 /* XXX this looks like dead code */ 8011 pp = page_lookup_nowait(vp, io_off, 8012 (flags & B_FREE) ? SE_EXCL : SE_SHARED); 8013 } 8014 8015 if (pp == NULL || pvn_getdirty(pp, flags) == 0) 8016 io_len = PAGESIZE; 8017 else { 8018 error = cachefs_push(vp, pp, &io_off, 8019 &io_len, flags, cr); 8020 if (error != 0) 8021 break; 8022 /* 8023 * "io_off" and "io_len" are returned as 8024 * the range of pages we actually wrote. 8025 * This allows us to skip ahead more quickly 8026 * since several pages may've been dealt 8027 * with by this iteration of the loop. 8028 */ 8029 } 8030 } 8031 } 8032 8033 if (error == 0 && off == 0 && (len == 0 || len >= cp->c_size)) { 8034 cp->c_flags &= ~CDIRTY; 8035 } 8036 8037 if (CACHEFS_LOG_LOGGING(cachep, CACHEFS_LOG_PUTPAGE)) 8038 cachefs_log_putpage(cachep, error, fscp->fs_cfsvfsp, 8039 &cp->c_metadata.md_cookie, cp->c_id.cid_fileno, 8040 crgetuid(cr), off, len); 8041 8042 return (error); 8043 8044 } 8045 8046 /*ARGSUSED*/ 8047 static int 8048 cachefs_map(struct vnode *vp, offset_t off, struct as *as, caddr_t *addrp, 8049 size_t len, uchar_t prot, uchar_t maxprot, uint_t flags, cred_t *cr) 8050 { 8051 cnode_t *cp = VTOC(vp); 8052 fscache_t *fscp = C_TO_FSCACHE(cp); 8053 struct segvn_crargs vn_a; 8054 int error; 8055 int held = 0; 8056 int writing; 8057 int connected = 0; 8058 8059 #ifdef CFSDEBUG 8060 u_offset_t offx = (u_offset_t)off; 8061 8062 CFS_DEBUG(CFSDEBUG_VOPS) 8063 printf("cachefs_map: ENTER vp %p off %lld len %lu flags %d\n", 8064 (void *)vp, offx, len, flags); 8065 #endif 8066 if (getzoneid() != GLOBAL_ZONEID) { 8067 error = EPERM; 8068 goto out; 8069 } 8070 8071 if (vp->v_flag & VNOMAP) { 8072 error = ENOSYS; 8073 goto out; 8074 } 8075 if (off < 0 || (offset_t)(off + len) < 0) { 8076 error = ENXIO; 8077 goto out; 8078 } 8079 if (vp->v_type != VREG) { 8080 error = ENODEV; 8081 goto out; 8082 } 8083 8084 /* 8085 * Check to see if the vnode is currently marked as not cachable. 8086 * If so, we have to refuse the map request as this violates the 8087 * don't cache attribute. 8088 */ 8089 if (vp->v_flag & VNOCACHE) 8090 return (EAGAIN); 8091 8092 #ifdef OBSOLETE 8093 /* 8094 * If file is being locked, disallow mapping. 8095 */ 8096 if (vn_has_flocks(vp)) { 8097 error = EAGAIN; 8098 goto out; 8099 } 8100 #endif 8101 8102 /* call backfilesystem if NFSv4 */ 8103 if (CFS_ISFS_BACKFS_NFSV4(fscp)) { 8104 error = cachefs_map_backfs_nfsv4(vp, off, as, addrp, len, prot, 8105 maxprot, flags, cr); 8106 goto out; 8107 } 8108 8109 writing = (prot & PROT_WRITE && ((flags & MAP_PRIVATE) == 0)); 8110 8111 for (;;) { 8112 /* get (or renew) access to the file system */ 8113 if (held) { 8114 cachefs_cd_release(fscp); 8115 held = 0; 8116 } 8117 error = cachefs_cd_access(fscp, connected, writing); 8118 if (error) 8119 break; 8120 held = 1; 8121 8122 if (writing) { 8123 mutex_enter(&cp->c_statelock); 8124 if (CFS_ISFS_WRITE_AROUND(fscp)) { 8125 if (fscp->fs_cdconnected != CFS_CD_CONNECTED) { 8126 connected = 1; 8127 continue; 8128 } else { 8129 cachefs_nocache(cp); 8130 } 8131 } 8132 8133 /* 8134 * CN_MAPWRITE is for an optimization in cachefs_delmap. 8135 * If CN_MAPWRITE is not set then cachefs_delmap does 8136 * not need to try to push out any pages. 8137 * This bit gets cleared when the cnode goes inactive. 8138 */ 8139 cp->c_flags |= CN_MAPWRITE; 8140 8141 mutex_exit(&cp->c_statelock); 8142 } 8143 break; 8144 } 8145 8146 if (held) { 8147 cachefs_cd_release(fscp); 8148 } 8149 8150 as_rangelock(as); 8151 if ((flags & MAP_FIXED) == 0) { 8152 map_addr(addrp, len, off, 1, flags); 8153 if (*addrp == NULL) { 8154 as_rangeunlock(as); 8155 error = ENOMEM; 8156 goto out; 8157 } 8158 } else { 8159 /* 8160 * User specified address - blow away any previous mappings 8161 */ 8162 (void) as_unmap(as, *addrp, len); 8163 } 8164 8165 /* 8166 * package up all the data passed in into a segvn_args struct and 8167 * call as_map with segvn_create function to create a new segment 8168 * in the address space. 8169 */ 8170 vn_a.vp = vp; 8171 vn_a.offset = off; 8172 vn_a.type = flags & MAP_TYPE; 8173 vn_a.prot = (uchar_t)prot; 8174 vn_a.maxprot = (uchar_t)maxprot; 8175 vn_a.cred = cr; 8176 vn_a.amp = NULL; 8177 vn_a.flags = flags & ~MAP_TYPE; 8178 vn_a.szc = 0; 8179 vn_a.lgrp_mem_policy_flags = 0; 8180 error = as_map(as, *addrp, len, segvn_create, &vn_a); 8181 as_rangeunlock(as); 8182 out: 8183 8184 #ifdef CFS_CD_DEBUG 8185 ASSERT((curthread->t_flag & T_CD_HELD) == 0); 8186 #endif 8187 #ifdef CFSDEBUG 8188 CFS_DEBUG(CFSDEBUG_VOPS) 8189 printf("cachefs_map: EXIT vp %p error %d\n", (void *)vp, error); 8190 #endif 8191 return (error); 8192 } 8193 8194 /* 8195 * cachefs_map_backfs_nfsv4 8196 * 8197 * Call NFSv4 back filesystem to handle the map (cachefs 8198 * pass-through support for NFSv4). 8199 */ 8200 static int 8201 cachefs_map_backfs_nfsv4(struct vnode *vp, offset_t off, struct as *as, 8202 caddr_t *addrp, size_t len, uchar_t prot, 8203 uchar_t maxprot, uint_t flags, cred_t *cr) 8204 { 8205 cnode_t *cp = VTOC(vp); 8206 fscache_t *fscp = C_TO_FSCACHE(cp); 8207 vnode_t *backvp; 8208 int error; 8209 8210 /* 8211 * For NFSv4 pass-through to work, only connected operation is 8212 * supported, the cnode backvp must exist, and cachefs optional 8213 * (eg., disconnectable) flags are turned off. Assert these 8214 * conditions for the map operation. 8215 */ 8216 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 8217 CFS_BACKFS_NFSV4_ASSERT_CNODE(cp); 8218 8219 /* Call backfs vnode op after extracting backvp */ 8220 mutex_enter(&cp->c_statelock); 8221 backvp = cp->c_backvp; 8222 mutex_exit(&cp->c_statelock); 8223 8224 CFS_DPRINT_BACKFS_NFSV4(fscp, 8225 ("cachefs_map_backfs_nfsv4: cnode %p, backvp %p\n", 8226 cp, backvp)); 8227 error = VOP_MAP(backvp, off, as, addrp, len, prot, maxprot, flags, cr); 8228 8229 return (error); 8230 } 8231 8232 /*ARGSUSED*/ 8233 static int 8234 cachefs_addmap(struct vnode *vp, offset_t off, struct as *as, 8235 caddr_t addr, size_t len, uchar_t prot, uchar_t maxprot, uint_t flags, 8236 cred_t *cr) 8237 { 8238 cnode_t *cp = VTOC(vp); 8239 fscache_t *fscp = C_TO_FSCACHE(cp); 8240 8241 if (getzoneid() != GLOBAL_ZONEID) 8242 return (EPERM); 8243 8244 if (vp->v_flag & VNOMAP) 8245 return (ENOSYS); 8246 8247 /* 8248 * Check this is not an NFSv4 filesystem, as the mapping 8249 * is not done on the cachefs filesystem if NFSv4 is in 8250 * use. 8251 */ 8252 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 8253 8254 mutex_enter(&cp->c_statelock); 8255 cp->c_mapcnt += btopr(len); 8256 mutex_exit(&cp->c_statelock); 8257 return (0); 8258 } 8259 8260 /*ARGSUSED*/ 8261 static int 8262 cachefs_delmap(struct vnode *vp, offset_t off, struct as *as, 8263 caddr_t addr, size_t len, uint_t prot, uint_t maxprot, uint_t flags, 8264 cred_t *cr) 8265 { 8266 cnode_t *cp = VTOC(vp); 8267 fscache_t *fscp = C_TO_FSCACHE(cp); 8268 int error; 8269 int connected = 0; 8270 int held = 0; 8271 8272 /* 8273 * The file may be passed in to (or inherited into) the zone, so we 8274 * need to let this operation go through since it happens as part of 8275 * exiting. 8276 */ 8277 if (vp->v_flag & VNOMAP) 8278 return (ENOSYS); 8279 8280 /* 8281 * Check this is not an NFSv4 filesystem, as the mapping 8282 * is not done on the cachefs filesystem if NFSv4 is in 8283 * use. 8284 */ 8285 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 8286 8287 mutex_enter(&cp->c_statelock); 8288 cp->c_mapcnt -= btopr(len); 8289 ASSERT(cp->c_mapcnt >= 0); 8290 mutex_exit(&cp->c_statelock); 8291 8292 if (cp->c_mapcnt || !vn_has_cached_data(vp) || 8293 ((cp->c_flags & CN_MAPWRITE) == 0)) 8294 return (0); 8295 8296 for (;;) { 8297 /* get (or renew) access to the file system */ 8298 if (held) { 8299 cachefs_cd_release(fscp); 8300 held = 0; 8301 } 8302 error = cachefs_cd_access(fscp, connected, 1); 8303 if (error) 8304 break; 8305 held = 1; 8306 connected = 0; 8307 8308 error = cachefs_putpage_common(vp, (offset_t)0, 8309 (uint_t)0, 0, cr); 8310 if (CFS_TIMEOUT(fscp, error)) { 8311 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 8312 cachefs_cd_release(fscp); 8313 held = 0; 8314 cachefs_cd_timedout(fscp); 8315 continue; 8316 } else { 8317 connected = 1; 8318 continue; 8319 } 8320 } 8321 8322 /* if no space left in cache, wait until connected */ 8323 if ((error == ENOSPC) && 8324 (fscp->fs_cdconnected != CFS_CD_CONNECTED)) { 8325 connected = 1; 8326 continue; 8327 } 8328 8329 mutex_enter(&cp->c_statelock); 8330 if (!error) 8331 error = cp->c_error; 8332 cp->c_error = 0; 8333 mutex_exit(&cp->c_statelock); 8334 break; 8335 } 8336 8337 if (held) 8338 cachefs_cd_release(fscp); 8339 8340 #ifdef CFS_CD_DEBUG 8341 ASSERT((curthread->t_flag & T_CD_HELD) == 0); 8342 #endif 8343 return (error); 8344 } 8345 8346 /* ARGSUSED */ 8347 static int 8348 cachefs_frlock(struct vnode *vp, int cmd, struct flock64 *bfp, int flag, 8349 offset_t offset, struct flk_callback *flk_cbp, cred_t *cr) 8350 { 8351 struct cnode *cp = VTOC(vp); 8352 int error; 8353 struct fscache *fscp = C_TO_FSCACHE(cp); 8354 vnode_t *backvp; 8355 int held = 0; 8356 int connected = 0; 8357 8358 if (getzoneid() != GLOBAL_ZONEID) 8359 return (EPERM); 8360 8361 if ((cmd != F_GETLK) && (cmd != F_SETLK) && (cmd != F_SETLKW)) 8362 return (EINVAL); 8363 8364 /* Disallow locking of files that are currently mapped */ 8365 if (((cmd == F_SETLK) || (cmd == F_SETLKW)) && (cp->c_mapcnt > 0)) { 8366 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 8367 return (EAGAIN); 8368 } 8369 8370 /* 8371 * Cachefs only provides pass-through support for NFSv4, 8372 * and all vnode operations are passed through to the 8373 * back file system. For NFSv4 pass-through to work, only 8374 * connected operation is supported, the cnode backvp must 8375 * exist, and cachefs optional (eg., disconnectable) flags 8376 * are turned off. Assert these conditions to ensure that 8377 * the backfilesystem is called for the frlock operation. 8378 */ 8379 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 8380 CFS_BACKFS_NFSV4_ASSERT_CNODE(cp); 8381 8382 /* XXX bob: nfs does a bunch more checks than we do */ 8383 if (CFS_ISFS_LLOCK(fscp)) { 8384 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 8385 return (fs_frlock(vp, cmd, bfp, flag, offset, flk_cbp, cr)); 8386 } 8387 8388 for (;;) { 8389 /* get (or renew) access to the file system */ 8390 if (held) { 8391 /* Won't loop with NFSv4 connected behavior */ 8392 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 8393 cachefs_cd_release(fscp); 8394 held = 0; 8395 } 8396 error = cachefs_cd_access(fscp, connected, 0); 8397 if (error) 8398 break; 8399 held = 1; 8400 8401 /* if not connected, quit or wait */ 8402 if (fscp->fs_cdconnected != CFS_CD_CONNECTED) { 8403 connected = 1; 8404 continue; 8405 } 8406 8407 /* nocache the file */ 8408 if ((cp->c_flags & CN_NOCACHE) == 0 && 8409 !CFS_ISFS_BACKFS_NFSV4(fscp)) { 8410 mutex_enter(&cp->c_statelock); 8411 cachefs_nocache(cp); 8412 mutex_exit(&cp->c_statelock); 8413 } 8414 8415 /* 8416 * XXX bob: probably should do a consistency check 8417 * Pass arguments unchanged if NFSv4 is the backfs. 8418 */ 8419 if (bfp->l_whence == 2 && CFS_ISFS_BACKFS_NFSV4(fscp) == 0) { 8420 bfp->l_start += cp->c_size; 8421 bfp->l_whence = 0; 8422 } 8423 8424 /* get the back vp */ 8425 mutex_enter(&cp->c_statelock); 8426 if (cp->c_backvp == NULL) { 8427 error = cachefs_getbackvp(fscp, cp); 8428 if (error) { 8429 mutex_exit(&cp->c_statelock); 8430 break; 8431 } 8432 } 8433 backvp = cp->c_backvp; 8434 VN_HOLD(backvp); 8435 mutex_exit(&cp->c_statelock); 8436 8437 /* 8438 * make sure we can flush currently dirty pages before 8439 * allowing the lock 8440 */ 8441 if (bfp->l_type != F_UNLCK && cmd != F_GETLK && 8442 !CFS_ISFS_BACKFS_NFSV4(fscp)) { 8443 error = cachefs_putpage( 8444 vp, (offset_t)0, 0, B_INVAL, cr); 8445 if (error) { 8446 error = ENOLCK; 8447 VN_RELE(backvp); 8448 break; 8449 } 8450 } 8451 8452 /* do lock on the back file */ 8453 CFS_DPRINT_BACKFS_NFSV4(fscp, 8454 ("cachefs_frlock (nfsv4): cp %p, backvp %p\n", 8455 cp, backvp)); 8456 error = VOP_FRLOCK(backvp, cmd, bfp, flag, offset, NULL, cr); 8457 VN_RELE(backvp); 8458 if (CFS_TIMEOUT(fscp, error)) { 8459 connected = 1; 8460 continue; 8461 } 8462 break; 8463 } 8464 8465 if (held) { 8466 cachefs_cd_release(fscp); 8467 } 8468 8469 /* 8470 * If we are setting a lock mark the vnode VNOCACHE so the page 8471 * cache does not give inconsistent results on locked files shared 8472 * between clients. The VNOCACHE flag is never turned off as long 8473 * as the vnode is active because it is hard to figure out when the 8474 * last lock is gone. 8475 * XXX - what if some already has the vnode mapped in? 8476 * XXX bob: see nfs3_frlock, do not allow locking if vnode mapped in. 8477 */ 8478 if ((error == 0) && (bfp->l_type != F_UNLCK) && (cmd != F_GETLK) && 8479 !CFS_ISFS_BACKFS_NFSV4(fscp)) 8480 vp->v_flag |= VNOCACHE; 8481 8482 #ifdef CFS_CD_DEBUG 8483 ASSERT((curthread->t_flag & T_CD_HELD) == 0); 8484 #endif 8485 return (error); 8486 } 8487 8488 /* 8489 * Free storage space associated with the specified vnode. The portion 8490 * to be freed is specified by bfp->l_start and bfp->l_len (already 8491 * normalized to a "whence" of 0). 8492 * 8493 * This is an experimental facility whose continued existence is not 8494 * guaranteed. Currently, we only support the special case 8495 * of l_len == 0, meaning free to end of file. 8496 */ 8497 /* ARGSUSED */ 8498 static int 8499 cachefs_space(struct vnode *vp, int cmd, struct flock64 *bfp, int flag, 8500 offset_t offset, cred_t *cr, caller_context_t *ct) 8501 { 8502 cnode_t *cp = VTOC(vp); 8503 fscache_t *fscp = C_TO_FSCACHE(cp); 8504 int error; 8505 8506 ASSERT(vp->v_type == VREG); 8507 if (getzoneid() != GLOBAL_ZONEID) 8508 return (EPERM); 8509 if (cmd != F_FREESP) 8510 return (EINVAL); 8511 8512 /* call backfilesystem if NFSv4 */ 8513 if (CFS_ISFS_BACKFS_NFSV4(fscp)) { 8514 error = cachefs_space_backfs_nfsv4(vp, cmd, bfp, flag, 8515 offset, cr, ct); 8516 goto out; 8517 } 8518 8519 if ((error = convoff(vp, bfp, 0, offset)) == 0) { 8520 ASSERT(bfp->l_start >= 0); 8521 if (bfp->l_len == 0) { 8522 struct vattr va; 8523 8524 va.va_size = bfp->l_start; 8525 va.va_mask = AT_SIZE; 8526 error = cachefs_setattr(vp, &va, 0, cr, ct); 8527 } else 8528 error = EINVAL; 8529 } 8530 8531 out: 8532 return (error); 8533 } 8534 8535 /* 8536 * cachefs_space_backfs_nfsv4 8537 * 8538 * Call NFSv4 back filesystem to handle the space (cachefs 8539 * pass-through support for NFSv4). 8540 */ 8541 static int 8542 cachefs_space_backfs_nfsv4(struct vnode *vp, int cmd, struct flock64 *bfp, 8543 int flag, offset_t offset, cred_t *cr, caller_context_t *ct) 8544 { 8545 cnode_t *cp = VTOC(vp); 8546 fscache_t *fscp = C_TO_FSCACHE(cp); 8547 vnode_t *backvp; 8548 int error; 8549 8550 /* 8551 * For NFSv4 pass-through to work, only connected operation is 8552 * supported, the cnode backvp must exist, and cachefs optional 8553 * (eg., disconnectable) flags are turned off. Assert these 8554 * conditions for the space operation. 8555 */ 8556 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 8557 CFS_BACKFS_NFSV4_ASSERT_CNODE(cp); 8558 8559 /* Call backfs vnode op after extracting backvp */ 8560 mutex_enter(&cp->c_statelock); 8561 backvp = cp->c_backvp; 8562 mutex_exit(&cp->c_statelock); 8563 8564 CFS_DPRINT_BACKFS_NFSV4(fscp, 8565 ("cachefs_space_backfs_nfsv4: cnode %p, backvp %p\n", 8566 cp, backvp)); 8567 error = VOP_SPACE(backvp, cmd, bfp, flag, offset, cr, ct); 8568 8569 return (error); 8570 } 8571 8572 /*ARGSUSED*/ 8573 static int 8574 cachefs_realvp(struct vnode *vp, struct vnode **vpp) 8575 { 8576 return (EINVAL); 8577 } 8578 8579 /*ARGSUSED*/ 8580 static int 8581 cachefs_pageio(struct vnode *vp, page_t *pp, u_offset_t io_off, size_t io_len, 8582 int flags, cred_t *cr) 8583 { 8584 return (ENOSYS); 8585 } 8586 8587 static int 8588 cachefs_setsecattr_connected(cnode_t *cp, 8589 vsecattr_t *vsec, int flag, cred_t *cr) 8590 { 8591 fscache_t *fscp = C_TO_FSCACHE(cp); 8592 int error = 0; 8593 8594 ASSERT(RW_WRITE_HELD(&cp->c_rwlock)); 8595 ASSERT((fscp->fs_info.fi_mntflags & CFS_NOACL) == 0); 8596 8597 mutex_enter(&cp->c_statelock); 8598 8599 if (cp->c_backvp == NULL) { 8600 error = cachefs_getbackvp(fscp, cp); 8601 if (error) { 8602 cachefs_nocache(cp); 8603 goto out; 8604 } 8605 } 8606 8607 error = CFSOP_CHECK_COBJECT(fscp, cp, 0, cr); 8608 if (error) 8609 goto out; 8610 8611 /* only owner can set acl */ 8612 if (cp->c_metadata.md_vattr.va_uid != crgetuid(cr)) { 8613 error = EINVAL; 8614 goto out; 8615 } 8616 8617 8618 CFS_DPRINT_BACKFS_NFSV4(fscp, 8619 ("cachefs_setsecattr (nfsv4): cp %p, backvp %p", 8620 cp, cp->c_backvp)); 8621 error = VOP_SETSECATTR(cp->c_backvp, vsec, flag, cr); 8622 if (error) { 8623 goto out; 8624 } 8625 8626 if ((cp->c_filegrp->fg_flags & CFS_FG_WRITE) == 0 && 8627 !CFS_ISFS_BACKFS_NFSV4(fscp)) { 8628 cachefs_nocache(cp); 8629 goto out; 8630 } 8631 8632 CFSOP_MODIFY_COBJECT(fscp, cp, cr); 8633 8634 /* acl may have changed permissions -- handle this. */ 8635 if (!CFS_ISFS_BACKFS_NFSV4(fscp)) 8636 cachefs_acl2perm(cp, vsec); 8637 8638 if ((cp->c_flags & CN_NOCACHE) == 0 && 8639 !CFS_ISFS_BACKFS_NFSV4(fscp)) { 8640 error = cachefs_cacheacl(cp, vsec); 8641 if (error != 0) { 8642 #ifdef CFSDEBUG 8643 CFS_DEBUG(CFSDEBUG_VOPS) 8644 printf("cachefs_setacl: cacheacl: error %d\n", 8645 error); 8646 #endif /* CFSDEBUG */ 8647 error = 0; 8648 cachefs_nocache(cp); 8649 } 8650 } 8651 8652 out: 8653 mutex_exit(&cp->c_statelock); 8654 8655 return (error); 8656 } 8657 8658 static int 8659 cachefs_setsecattr_disconnected(cnode_t *cp, 8660 vsecattr_t *vsec, int flag, cred_t *cr) 8661 { 8662 fscache_t *fscp = C_TO_FSCACHE(cp); 8663 mode_t failmode = cp->c_metadata.md_vattr.va_mode; 8664 off_t commit = 0; 8665 int error = 0; 8666 8667 ASSERT((fscp->fs_info.fi_mntflags & CFS_NOACL) == 0); 8668 8669 if (CFS_ISFS_WRITE_AROUND(fscp)) 8670 return (ETIMEDOUT); 8671 8672 mutex_enter(&cp->c_statelock); 8673 8674 /* only owner can set acl */ 8675 if (cp->c_metadata.md_vattr.va_uid != crgetuid(cr)) { 8676 error = EINVAL; 8677 goto out; 8678 } 8679 8680 if (cp->c_metadata.md_flags & MD_NEEDATTRS) { 8681 error = ETIMEDOUT; 8682 goto out; 8683 } 8684 8685 /* XXX do i need this? is this right? */ 8686 if (cp->c_flags & CN_ALLOC_PENDING) { 8687 if (cp->c_filegrp->fg_flags & CFS_FG_ALLOC_ATTR) { 8688 (void) filegrp_allocattr(cp->c_filegrp); 8689 } 8690 error = filegrp_create_metadata(cp->c_filegrp, 8691 &cp->c_metadata, &cp->c_id); 8692 if (error) { 8693 goto out; 8694 } 8695 cp->c_flags &= ~CN_ALLOC_PENDING; 8696 } 8697 8698 /* XXX is this right? */ 8699 if ((cp->c_metadata.md_flags & MD_MAPPING) == 0) { 8700 error = cachefs_dlog_cidmap(fscp); 8701 if (error) { 8702 error = ENOSPC; 8703 goto out; 8704 } 8705 cp->c_metadata.md_flags |= MD_MAPPING; 8706 cp->c_flags |= CN_UPDATED; 8707 } 8708 8709 commit = cachefs_dlog_setsecattr(fscp, vsec, flag, cp, cr); 8710 if (commit == 0) 8711 goto out; 8712 8713 /* fix modes in metadata */ 8714 cachefs_acl2perm(cp, vsec); 8715 8716 if ((cp->c_flags & CN_NOCACHE) == 0) { 8717 error = cachefs_cacheacl(cp, vsec); 8718 if (error != 0) { 8719 goto out; 8720 } 8721 } 8722 8723 /* XXX is this right? */ 8724 if (cachefs_modified_alloc(cp)) { 8725 error = ENOSPC; 8726 goto out; 8727 } 8728 8729 out: 8730 if (error != 0) 8731 cp->c_metadata.md_vattr.va_mode = failmode; 8732 8733 mutex_exit(&cp->c_statelock); 8734 8735 if (commit) { 8736 if (cachefs_dlog_commit(fscp, commit, error)) { 8737 /*EMPTY*/ 8738 /* XXX fix on panic? */ 8739 } 8740 } 8741 8742 return (error); 8743 } 8744 8745 static int 8746 cachefs_setsecattr(vnode_t *vp, vsecattr_t *vsec, int flag, cred_t *cr) 8747 { 8748 cnode_t *cp = VTOC(vp); 8749 fscache_t *fscp = C_TO_FSCACHE(cp); 8750 int connected = 0; 8751 int held = 0; 8752 int error = 0; 8753 8754 #ifdef CFSDEBUG 8755 CFS_DEBUG(CFSDEBUG_VOPS) 8756 printf("cachefs_setsecattr: ENTER vp %p\n", (void *)vp); 8757 #endif 8758 if (getzoneid() != GLOBAL_ZONEID) { 8759 error = EPERM; 8760 goto out; 8761 } 8762 8763 if (fscp->fs_info.fi_mntflags & CFS_NOACL) { 8764 error = ENOSYS; 8765 goto out; 8766 } 8767 8768 if (! cachefs_vtype_aclok(vp)) { 8769 error = EINVAL; 8770 goto out; 8771 } 8772 8773 /* 8774 * Cachefs only provides pass-through support for NFSv4, 8775 * and all vnode operations are passed through to the 8776 * back file system. For NFSv4 pass-through to work, only 8777 * connected operation is supported, the cnode backvp must 8778 * exist, and cachefs optional (eg., disconnectable) flags 8779 * are turned off. Assert these conditions to ensure that 8780 * the backfilesystem is called for the setsecattr operation. 8781 */ 8782 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 8783 CFS_BACKFS_NFSV4_ASSERT_CNODE(cp); 8784 8785 for (;;) { 8786 /* drop hold on file system */ 8787 if (held) { 8788 /* Won't loop with NFSv4 connected operation */ 8789 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 8790 cachefs_cd_release(fscp); 8791 held = 0; 8792 } 8793 8794 /* aquire access to the file system */ 8795 error = cachefs_cd_access(fscp, connected, 1); 8796 if (error) 8797 break; 8798 held = 1; 8799 8800 /* perform the setattr */ 8801 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) 8802 error = cachefs_setsecattr_connected(cp, 8803 vsec, flag, cr); 8804 else 8805 error = cachefs_setsecattr_disconnected(cp, 8806 vsec, flag, cr); 8807 if (error) { 8808 /* if connected */ 8809 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 8810 if (CFS_TIMEOUT(fscp, error)) { 8811 cachefs_cd_release(fscp); 8812 held = 0; 8813 cachefs_cd_timedout(fscp); 8814 connected = 0; 8815 continue; 8816 } 8817 } 8818 8819 /* else must be disconnected */ 8820 else { 8821 if (CFS_TIMEOUT(fscp, error)) { 8822 connected = 1; 8823 continue; 8824 } 8825 } 8826 } 8827 break; 8828 } 8829 8830 if (held) { 8831 cachefs_cd_release(fscp); 8832 } 8833 return (error); 8834 8835 out: 8836 #ifdef CFS_CD_DEBUG 8837 ASSERT((curthread->t_flag & T_CD_HELD) == 0); 8838 #endif 8839 8840 #ifdef CFSDEBUG 8841 CFS_DEBUG(CFSDEBUG_VOPS) 8842 printf("cachefs_setsecattr: EXIT error = %d\n", error); 8843 #endif 8844 return (error); 8845 } 8846 8847 /* 8848 * call this BEFORE calling cachefs_cacheacl(), as the latter will 8849 * sanitize the acl. 8850 */ 8851 8852 static void 8853 cachefs_acl2perm(cnode_t *cp, vsecattr_t *vsec) 8854 { 8855 aclent_t *aclp; 8856 int i; 8857 8858 for (i = 0; i < vsec->vsa_aclcnt; i++) { 8859 aclp = ((aclent_t *)vsec->vsa_aclentp) + i; 8860 switch (aclp->a_type) { 8861 case USER_OBJ: 8862 cp->c_metadata.md_vattr.va_mode &= (~0700); 8863 cp->c_metadata.md_vattr.va_mode |= (aclp->a_perm << 6); 8864 break; 8865 8866 case GROUP_OBJ: 8867 cp->c_metadata.md_vattr.va_mode &= (~070); 8868 cp->c_metadata.md_vattr.va_mode |= (aclp->a_perm << 3); 8869 break; 8870 8871 case OTHER_OBJ: 8872 cp->c_metadata.md_vattr.va_mode &= (~07); 8873 cp->c_metadata.md_vattr.va_mode |= (aclp->a_perm); 8874 break; 8875 8876 case CLASS_OBJ: 8877 cp->c_metadata.md_aclclass = aclp->a_perm; 8878 break; 8879 } 8880 } 8881 8882 cp->c_flags |= CN_UPDATED; 8883 } 8884 8885 static int 8886 cachefs_getsecattr(vnode_t *vp, vsecattr_t *vsec, int flag, cred_t *cr) 8887 { 8888 cnode_t *cp = VTOC(vp); 8889 fscache_t *fscp = C_TO_FSCACHE(cp); 8890 int held = 0, connected = 0; 8891 int error = 0; 8892 8893 #ifdef CFSDEBUG 8894 CFS_DEBUG(CFSDEBUG_VOPS) 8895 printf("cachefs_getsecattr: ENTER vp %p\n", (void *)vp); 8896 #endif 8897 8898 if (getzoneid() != GLOBAL_ZONEID) { 8899 error = EPERM; 8900 goto out; 8901 } 8902 8903 /* 8904 * Cachefs only provides pass-through support for NFSv4, 8905 * and all vnode operations are passed through to the 8906 * back file system. For NFSv4 pass-through to work, only 8907 * connected operation is supported, the cnode backvp must 8908 * exist, and cachefs optional (eg., disconnectable) flags 8909 * are turned off. Assert these conditions to ensure that 8910 * the backfilesystem is called for the getsecattr operation. 8911 */ 8912 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 8913 CFS_BACKFS_NFSV4_ASSERT_CNODE(cp); 8914 8915 if (fscp->fs_info.fi_mntflags & CFS_NOACL) { 8916 error = fs_fab_acl(vp, vsec, flag, cr); 8917 goto out; 8918 } 8919 8920 for (;;) { 8921 if (held) { 8922 /* Won't loop with NFSv4 connected behavior */ 8923 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 8924 cachefs_cd_release(fscp); 8925 held = 0; 8926 } 8927 error = cachefs_cd_access(fscp, connected, 0); 8928 if (error) 8929 break; 8930 held = 1; 8931 8932 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) { 8933 error = cachefs_getsecattr_connected(vp, vsec, flag, 8934 cr); 8935 if (CFS_TIMEOUT(fscp, error)) { 8936 cachefs_cd_release(fscp); 8937 held = 0; 8938 cachefs_cd_timedout(fscp); 8939 connected = 0; 8940 continue; 8941 } 8942 } else { 8943 error = cachefs_getsecattr_disconnected(vp, vsec, flag, 8944 cr); 8945 if (CFS_TIMEOUT(fscp, error)) { 8946 if (cachefs_cd_access_miss(fscp)) { 8947 error = cachefs_getsecattr_connected(vp, 8948 vsec, flag, cr); 8949 if (!CFS_TIMEOUT(fscp, error)) 8950 break; 8951 delay(5*hz); 8952 connected = 0; 8953 continue; 8954 } 8955 connected = 1; 8956 continue; 8957 } 8958 } 8959 break; 8960 } 8961 8962 out: 8963 if (held) 8964 cachefs_cd_release(fscp); 8965 8966 #ifdef CFS_CD_DEBUG 8967 ASSERT((curthread->t_flag & T_CD_HELD) == 0); 8968 #endif 8969 #ifdef CFSDEBUG 8970 CFS_DEBUG(CFSDEBUG_VOPS) 8971 printf("cachefs_getsecattr: EXIT error = %d\n", error); 8972 #endif 8973 return (error); 8974 } 8975 8976 static int 8977 cachefs_shrlock(vnode_t *vp, int cmd, struct shrlock *shr, int flag, cred_t *cr) 8978 { 8979 cnode_t *cp = VTOC(vp); 8980 fscache_t *fscp = C_TO_FSCACHE(cp); 8981 int error = 0; 8982 vnode_t *backvp; 8983 8984 #ifdef CFSDEBUG 8985 CFS_DEBUG(CFSDEBUG_VOPS) 8986 printf("cachefs_shrlock: ENTER vp %p\n", (void *)vp); 8987 #endif 8988 8989 if (getzoneid() != GLOBAL_ZONEID) { 8990 error = EPERM; 8991 goto out; 8992 } 8993 8994 /* 8995 * Cachefs only provides pass-through support for NFSv4, 8996 * and all vnode operations are passed through to the 8997 * back file system. For NFSv4 pass-through to work, only 8998 * connected operation is supported, the cnode backvp must 8999 * exist, and cachefs optional (eg., disconnectable) flags 9000 * are turned off. Assert these conditions to ensure that 9001 * the backfilesystem is called for the shrlock operation. 9002 */ 9003 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 9004 CFS_BACKFS_NFSV4_ASSERT_CNODE(cp); 9005 9006 mutex_enter(&cp->c_statelock); 9007 if (cp->c_backvp == NULL) 9008 error = cachefs_getbackvp(fscp, cp); 9009 backvp = cp->c_backvp; 9010 mutex_exit(&cp->c_statelock); 9011 ASSERT((error != 0) || (backvp != NULL)); 9012 9013 if (error == 0) { 9014 CFS_DPRINT_BACKFS_NFSV4(fscp, 9015 ("cachefs_shrlock (nfsv4): cp %p, backvp %p", 9016 cp, backvp)); 9017 error = VOP_SHRLOCK(backvp, cmd, shr, flag, cr); 9018 } 9019 9020 out: 9021 #ifdef CFSDEBUG 9022 CFS_DEBUG(CFSDEBUG_VOPS) 9023 printf("cachefs_shrlock: EXIT error = %d\n", error); 9024 #endif 9025 return (error); 9026 } 9027 9028 static int 9029 cachefs_getsecattr_connected(vnode_t *vp, vsecattr_t *vsec, int flag, 9030 cred_t *cr) 9031 { 9032 cnode_t *cp = VTOC(vp); 9033 fscache_t *fscp = C_TO_FSCACHE(cp); 9034 int hit = 0; 9035 int error = 0; 9036 9037 9038 mutex_enter(&cp->c_statelock); 9039 error = CFSOP_CHECK_COBJECT(fscp, cp, 0, cr); 9040 if (error) 9041 goto out; 9042 9043 /* read from the cache if we can */ 9044 if ((cp->c_metadata.md_flags & MD_ACL) && 9045 ((cp->c_flags & CN_NOCACHE) == 0) && 9046 !CFS_ISFS_BACKFS_NFSV4(fscp)) { 9047 ASSERT((cp->c_flags & CN_NOCACHE) == 0); 9048 error = cachefs_getaclfromcache(cp, vsec); 9049 if (error) { 9050 cachefs_nocache(cp); 9051 ASSERT((cp->c_metadata.md_flags & MD_ACL) == 0); 9052 error = 0; 9053 } else { 9054 hit = 1; 9055 goto out; 9056 } 9057 } 9058 9059 ASSERT(error == 0); 9060 if (cp->c_backvp == NULL) 9061 error = cachefs_getbackvp(fscp, cp); 9062 if (error) 9063 goto out; 9064 9065 CFS_DPRINT_BACKFS_NFSV4(fscp, 9066 ("cachefs_getsecattr (nfsv4): cp %p, backvp %p", 9067 cp, cp->c_backvp)); 9068 error = VOP_GETSECATTR(cp->c_backvp, vsec, flag, cr); 9069 if (error) 9070 goto out; 9071 9072 if (((fscp->fs_info.fi_mntflags & CFS_NOACL) == 0) && 9073 (cachefs_vtype_aclok(vp)) && 9074 ((cp->c_flags & CN_NOCACHE) == 0) && 9075 !CFS_ISFS_BACKFS_NFSV4(fscp)) { 9076 error = cachefs_cacheacl(cp, vsec); 9077 if (error) { 9078 error = 0; 9079 cachefs_nocache(cp); 9080 } 9081 } 9082 9083 out: 9084 if (error == 0) { 9085 if (hit) 9086 fscp->fs_stats.st_hits++; 9087 else 9088 fscp->fs_stats.st_misses++; 9089 } 9090 mutex_exit(&cp->c_statelock); 9091 9092 return (error); 9093 } 9094 9095 static int 9096 /*ARGSUSED*/ 9097 cachefs_getsecattr_disconnected(vnode_t *vp, vsecattr_t *vsec, int flag, 9098 cred_t *cr) 9099 { 9100 cnode_t *cp = VTOC(vp); 9101 fscache_t *fscp = C_TO_FSCACHE(cp); 9102 int hit = 0; 9103 int error = 0; 9104 9105 9106 mutex_enter(&cp->c_statelock); 9107 9108 /* read from the cache if we can */ 9109 if (((cp->c_flags & CN_NOCACHE) == 0) && 9110 (cp->c_metadata.md_flags & MD_ACL)) { 9111 error = cachefs_getaclfromcache(cp, vsec); 9112 if (error) { 9113 cachefs_nocache(cp); 9114 ASSERT((cp->c_metadata.md_flags & MD_ACL) == 0); 9115 error = 0; 9116 } else { 9117 hit = 1; 9118 goto out; 9119 } 9120 } 9121 error = ETIMEDOUT; 9122 9123 out: 9124 if (error == 0) { 9125 if (hit) 9126 fscp->fs_stats.st_hits++; 9127 else 9128 fscp->fs_stats.st_misses++; 9129 } 9130 mutex_exit(&cp->c_statelock); 9131 9132 return (error); 9133 } 9134 9135 /* 9136 * cachefs_cacheacl() -- cache an ACL, which we do by applying it to 9137 * the frontfile if possible; otherwise, the adjunct directory. 9138 * 9139 * inputs: 9140 * cp - the cnode, with its statelock already held 9141 * vsecp - a pointer to a vsecattr_t you'd like us to cache as-is, 9142 * or NULL if you want us to do the VOP_GETSECATTR(backvp). 9143 * 9144 * returns: 9145 * 0 - all is well 9146 * nonzero - errno 9147 */ 9148 9149 int 9150 cachefs_cacheacl(cnode_t *cp, vsecattr_t *vsecp) 9151 { 9152 fscache_t *fscp = C_TO_FSCACHE(cp); 9153 vsecattr_t vsec; 9154 aclent_t *aclp; 9155 int gotvsec = 0; 9156 int error = 0; 9157 vnode_t *vp = NULL; 9158 void *aclkeep = NULL; 9159 int i; 9160 9161 ASSERT(MUTEX_HELD(&cp->c_statelock)); 9162 ASSERT((cp->c_flags & CN_NOCACHE) == 0); 9163 ASSERT(CFS_ISFS_BACKFS_NFSV4(fscp) == 0); 9164 ASSERT((fscp->fs_info.fi_mntflags & CFS_NOACL) == 0); 9165 ASSERT(cachefs_vtype_aclok(CTOV(cp))); 9166 9167 if (fscp->fs_info.fi_mntflags & CFS_NOACL) { 9168 error = ENOSYS; 9169 goto out; 9170 } 9171 9172 if (vsecp == NULL) { 9173 if (cp->c_backvp == NULL) 9174 error = cachefs_getbackvp(fscp, cp); 9175 if (error != 0) 9176 goto out; 9177 vsecp = &vsec; 9178 bzero(&vsec, sizeof (vsec)); 9179 vsecp->vsa_mask = 9180 VSA_ACL | VSA_ACLCNT | VSA_DFACL | VSA_DFACLCNT; 9181 error = VOP_GETSECATTR(cp->c_backvp, vsecp, 0, kcred); 9182 if (error != 0) { 9183 goto out; 9184 } 9185 gotvsec = 1; 9186 } else if (vsecp->vsa_mask & VSA_ACL) { 9187 aclkeep = vsecp->vsa_aclentp; 9188 vsecp->vsa_aclentp = cachefs_kmem_alloc(vsecp->vsa_aclcnt * 9189 sizeof (aclent_t), KM_SLEEP); 9190 bcopy(aclkeep, vsecp->vsa_aclentp, vsecp->vsa_aclcnt * 9191 sizeof (aclent_t)); 9192 } else if ((vsecp->vsa_mask & (VSA_ACL | VSA_DFACL)) == 0) { 9193 /* unless there's real data, we can cache nothing. */ 9194 return (0); 9195 } 9196 9197 /* 9198 * prevent the ACL from chmoding our frontfile, and 9199 * snarf the class info 9200 */ 9201 9202 if ((vsecp->vsa_mask & (VSA_ACL | VSA_ACLCNT)) == 9203 (VSA_ACL | VSA_ACLCNT)) { 9204 for (i = 0; i < vsecp->vsa_aclcnt; i++) { 9205 aclp = ((aclent_t *)vsecp->vsa_aclentp) + i; 9206 switch (aclp->a_type) { 9207 case CLASS_OBJ: 9208 cp->c_metadata.md_aclclass = 9209 aclp->a_perm; 9210 /*FALLTHROUGH*/ 9211 case USER_OBJ: 9212 case GROUP_OBJ: 9213 case OTHER_OBJ: 9214 aclp->a_perm = 06; 9215 } 9216 } 9217 } 9218 9219 /* 9220 * if the frontfile exists, then we always do the work. but, 9221 * if there's no frontfile, and the ACL isn't a `real' ACL, 9222 * then we don't want to do the work. otherwise, an `ls -l' 9223 * will create tons of emtpy frontfiles. 9224 */ 9225 9226 if (((cp->c_metadata.md_flags & MD_FILE) == 0) && 9227 ((vsecp->vsa_aclcnt + vsecp->vsa_dfaclcnt) 9228 <= MIN_ACL_ENTRIES)) { 9229 cp->c_metadata.md_flags |= MD_ACL; 9230 cp->c_flags |= CN_UPDATED; 9231 goto out; 9232 } 9233 9234 /* 9235 * if we have a default ACL, then we need a 9236 * real live directory in the frontfs that we 9237 * can apply the ACL to. if not, then we just 9238 * use the frontfile. we get the frontfile 9239 * regardless -- that way, we know the 9240 * directory for the frontfile exists. 9241 */ 9242 9243 if (vsecp->vsa_dfaclcnt > 0) { 9244 if (cp->c_acldirvp == NULL) 9245 error = cachefs_getacldirvp(cp); 9246 if (error != 0) 9247 goto out; 9248 vp = cp->c_acldirvp; 9249 } else { 9250 if (cp->c_frontvp == NULL) 9251 error = cachefs_getfrontfile(cp); 9252 if (error != 0) 9253 goto out; 9254 vp = cp->c_frontvp; 9255 } 9256 ASSERT(vp != NULL); 9257 9258 (void) VOP_RWLOCK(vp, V_WRITELOCK_TRUE, NULL); 9259 error = VOP_SETSECATTR(vp, vsecp, 0, kcred); 9260 VOP_RWUNLOCK(vp, V_WRITELOCK_TRUE, NULL); 9261 if (error != 0) { 9262 #ifdef CFSDEBUG 9263 CFS_DEBUG(CFSDEBUG_VOPS) 9264 printf("cachefs_cacheacl: setsecattr: error %d\n", 9265 error); 9266 #endif /* CFSDEBUG */ 9267 goto out; 9268 } 9269 if (vp == cp->c_frontvp) 9270 cp->c_flags |= CN_NEED_FRONT_SYNC; 9271 9272 cp->c_metadata.md_flags |= MD_ACL; 9273 cp->c_flags |= CN_UPDATED; 9274 9275 out: 9276 if ((error) && (fscp->fs_cdconnected == CFS_CD_CONNECTED)) 9277 cachefs_nocache(cp); 9278 9279 if (gotvsec) { 9280 if (vsec.vsa_aclcnt) 9281 kmem_free(vsec.vsa_aclentp, 9282 vsec.vsa_aclcnt * sizeof (aclent_t)); 9283 if (vsec.vsa_dfaclcnt) 9284 kmem_free(vsec.vsa_dfaclentp, 9285 vsec.vsa_dfaclcnt * sizeof (aclent_t)); 9286 } else if (aclkeep != NULL) { 9287 cachefs_kmem_free(vsecp->vsa_aclentp, 9288 vsecp->vsa_aclcnt * sizeof (aclent_t)); 9289 vsecp->vsa_aclentp = aclkeep; 9290 } 9291 9292 return (error); 9293 } 9294 9295 void 9296 cachefs_purgeacl(cnode_t *cp) 9297 { 9298 ASSERT(MUTEX_HELD(&cp->c_statelock)); 9299 9300 ASSERT(!CFS_ISFS_BACKFS_NFSV4(C_TO_FSCACHE(cp))); 9301 9302 if (cp->c_acldirvp != NULL) { 9303 VN_RELE(cp->c_acldirvp); 9304 cp->c_acldirvp = NULL; 9305 } 9306 9307 if (cp->c_metadata.md_flags & MD_ACLDIR) { 9308 char name[CFS_FRONTFILE_NAME_SIZE + 2]; 9309 9310 ASSERT(cp->c_filegrp->fg_dirvp != NULL); 9311 make_ascii_name(&cp->c_id, name); 9312 (void) strcat(name, ".d"); 9313 9314 (void) VOP_RMDIR(cp->c_filegrp->fg_dirvp, name, 9315 cp->c_filegrp->fg_dirvp, kcred); 9316 } 9317 9318 cp->c_metadata.md_flags &= ~(MD_ACL | MD_ACLDIR); 9319 cp->c_flags |= CN_UPDATED; 9320 } 9321 9322 static int 9323 cachefs_getacldirvp(cnode_t *cp) 9324 { 9325 char name[CFS_FRONTFILE_NAME_SIZE + 2]; 9326 int error = 0; 9327 9328 ASSERT(MUTEX_HELD(&cp->c_statelock)); 9329 ASSERT(cp->c_acldirvp == NULL); 9330 9331 if (cp->c_frontvp == NULL) 9332 error = cachefs_getfrontfile(cp); 9333 if (error != 0) 9334 goto out; 9335 9336 ASSERT(cp->c_filegrp->fg_dirvp != NULL); 9337 make_ascii_name(&cp->c_id, name); 9338 (void) strcat(name, ".d"); 9339 error = VOP_LOOKUP(cp->c_filegrp->fg_dirvp, 9340 name, &cp->c_acldirvp, NULL, 0, NULL, kcred); 9341 if ((error != 0) && (error != ENOENT)) 9342 goto out; 9343 9344 if (error != 0) { 9345 vattr_t va; 9346 9347 va.va_mode = S_IFDIR | 0777; 9348 va.va_uid = 0; 9349 va.va_gid = 0; 9350 va.va_type = VDIR; 9351 va.va_mask = AT_TYPE | AT_MODE | 9352 AT_UID | AT_GID; 9353 error = 9354 VOP_MKDIR(cp->c_filegrp->fg_dirvp, 9355 name, &va, &cp->c_acldirvp, kcred); 9356 if (error != 0) 9357 goto out; 9358 } 9359 9360 ASSERT(cp->c_acldirvp != NULL); 9361 cp->c_metadata.md_flags |= MD_ACLDIR; 9362 cp->c_flags |= CN_UPDATED; 9363 9364 out: 9365 if (error != 0) 9366 cp->c_acldirvp = NULL; 9367 return (error); 9368 } 9369 9370 static int 9371 cachefs_getaclfromcache(cnode_t *cp, vsecattr_t *vsec) 9372 { 9373 aclent_t *aclp; 9374 int error = 0; 9375 vnode_t *vp = NULL; 9376 int i; 9377 9378 ASSERT(cp->c_metadata.md_flags & MD_ACL); 9379 ASSERT(MUTEX_HELD(&cp->c_statelock)); 9380 ASSERT(vsec->vsa_aclentp == NULL); 9381 9382 if (cp->c_metadata.md_flags & MD_ACLDIR) { 9383 if (cp->c_acldirvp == NULL) 9384 error = cachefs_getacldirvp(cp); 9385 if (error != 0) 9386 goto out; 9387 vp = cp->c_acldirvp; 9388 } else if (cp->c_metadata.md_flags & MD_FILE) { 9389 if (cp->c_frontvp == NULL) 9390 error = cachefs_getfrontfile(cp); 9391 if (error != 0) 9392 goto out; 9393 vp = cp->c_frontvp; 9394 } else { 9395 9396 /* 9397 * if we get here, then we know that MD_ACL is on, 9398 * meaning an ACL was successfully cached. we also 9399 * know that neither MD_ACLDIR nor MD_FILE are on, so 9400 * this has to be an entry without a `real' ACL. 9401 * thus, we forge whatever is necessary. 9402 */ 9403 9404 if (vsec->vsa_mask & VSA_ACLCNT) 9405 vsec->vsa_aclcnt = MIN_ACL_ENTRIES; 9406 9407 if (vsec->vsa_mask & VSA_ACL) { 9408 vsec->vsa_aclentp = 9409 kmem_zalloc(MIN_ACL_ENTRIES * 9410 sizeof (aclent_t), KM_SLEEP); 9411 aclp = (aclent_t *)vsec->vsa_aclentp; 9412 aclp->a_type = USER_OBJ; 9413 ++aclp; 9414 aclp->a_type = GROUP_OBJ; 9415 ++aclp; 9416 aclp->a_type = OTHER_OBJ; 9417 ++aclp; 9418 aclp->a_type = CLASS_OBJ; 9419 ksort((caddr_t)vsec->vsa_aclentp, MIN_ACL_ENTRIES, 9420 sizeof (aclent_t), cmp2acls); 9421 } 9422 9423 ASSERT(vp == NULL); 9424 } 9425 9426 if (vp != NULL) { 9427 if ((error = VOP_GETSECATTR(vp, vsec, 0, kcred)) != 0) { 9428 #ifdef CFSDEBUG 9429 CFS_DEBUG(CFSDEBUG_VOPS) 9430 printf("cachefs_getaclfromcache: error %d\n", 9431 error); 9432 #endif /* CFSDEBUG */ 9433 goto out; 9434 } 9435 } 9436 9437 if (vsec->vsa_aclentp != NULL) { 9438 for (i = 0; i < vsec->vsa_aclcnt; i++) { 9439 aclp = ((aclent_t *)vsec->vsa_aclentp) + i; 9440 switch (aclp->a_type) { 9441 case USER_OBJ: 9442 aclp->a_id = cp->c_metadata.md_vattr.va_uid; 9443 aclp->a_perm = 9444 cp->c_metadata.md_vattr.va_mode & 0700; 9445 aclp->a_perm >>= 6; 9446 break; 9447 9448 case GROUP_OBJ: 9449 aclp->a_id = cp->c_metadata.md_vattr.va_gid; 9450 aclp->a_perm = 9451 cp->c_metadata.md_vattr.va_mode & 070; 9452 aclp->a_perm >>= 3; 9453 break; 9454 9455 case OTHER_OBJ: 9456 aclp->a_perm = 9457 cp->c_metadata.md_vattr.va_mode & 07; 9458 break; 9459 9460 case CLASS_OBJ: 9461 aclp->a_perm = 9462 cp->c_metadata.md_aclclass; 9463 break; 9464 } 9465 } 9466 } 9467 9468 out: 9469 9470 if (error != 0) 9471 cachefs_nocache(cp); 9472 9473 return (error); 9474 } 9475 9476 /* 9477 * Fills in targp with attribute information from srcp, cp 9478 * and if necessary the system. 9479 */ 9480 static void 9481 cachefs_attr_setup(vattr_t *srcp, vattr_t *targp, cnode_t *cp, cred_t *cr) 9482 { 9483 time_t now; 9484 9485 ASSERT((srcp->va_mask & (AT_TYPE | AT_MODE)) == (AT_TYPE | AT_MODE)); 9486 9487 /* 9488 * Add code to fill in the va struct. We use the fields from 9489 * the srcp struct if they are populated, otherwise we guess 9490 */ 9491 9492 targp->va_mask = 0; /* initialize all fields */ 9493 targp->va_mode = srcp->va_mode; 9494 targp->va_type = srcp->va_type; 9495 targp->va_nlink = 1; 9496 targp->va_nodeid = 0; 9497 9498 if (srcp->va_mask & AT_UID) 9499 targp->va_uid = srcp->va_uid; 9500 else 9501 targp->va_uid = crgetuid(cr); 9502 9503 if (srcp->va_mask & AT_GID) 9504 targp->va_gid = srcp->va_gid; 9505 else 9506 targp->va_gid = crgetgid(cr); 9507 9508 if (srcp->va_mask & AT_FSID) 9509 targp->va_fsid = srcp->va_fsid; 9510 else 9511 targp->va_fsid = 0; /* initialize all fields */ 9512 9513 now = gethrestime_sec(); 9514 if (srcp->va_mask & AT_ATIME) 9515 targp->va_atime = srcp->va_atime; 9516 else 9517 targp->va_atime.tv_sec = now; 9518 9519 if (srcp->va_mask & AT_MTIME) 9520 targp->va_mtime = srcp->va_mtime; 9521 else 9522 targp->va_mtime.tv_sec = now; 9523 9524 if (srcp->va_mask & AT_CTIME) 9525 targp->va_ctime = srcp->va_ctime; 9526 else 9527 targp->va_ctime.tv_sec = now; 9528 9529 9530 if (srcp->va_mask & AT_SIZE) 9531 targp->va_size = srcp->va_size; 9532 else 9533 targp->va_size = 0; 9534 9535 /* 9536 * the remaing fields are set by the fs and not changable. 9537 * we populate these entries useing the parent directory 9538 * values. It's a small hack, but should work. 9539 */ 9540 targp->va_blksize = cp->c_metadata.md_vattr.va_blksize; 9541 targp->va_rdev = cp->c_metadata.md_vattr.va_rdev; 9542 targp->va_nblocks = cp->c_metadata.md_vattr.va_nblocks; 9543 targp->va_seq = 0; /* Never keep the sequence number */ 9544 } 9545 9546 /* 9547 * set the gid for a newly created file. The algorithm is as follows: 9548 * 9549 * 1) If the gid is set in the attribute list, then use it if 9550 * the caller is privileged, belongs to the target group, or 9551 * the group is the same as the parent directory. 9552 * 9553 * 2) If the parent directory's set-gid bit is clear, then use 9554 * the process gid 9555 * 9556 * 3) Otherwise, use the gid of the parent directory. 9557 * 9558 * Note: newcp->c_attr.va_{mode,type} must already be set before calling 9559 * this routine. 9560 */ 9561 static void 9562 cachefs_creategid(cnode_t *dcp, cnode_t *newcp, vattr_t *vap, cred_t *cr) 9563 { 9564 if ((vap->va_mask & AT_GID) && 9565 ((vap->va_gid == dcp->c_attr.va_gid) || 9566 groupmember(vap->va_gid, cr) || 9567 secpolicy_vnode_create_gid(cr) != 0)) { 9568 newcp->c_attr.va_gid = vap->va_gid; 9569 } else { 9570 if (dcp->c_attr.va_mode & S_ISGID) 9571 newcp->c_attr.va_gid = dcp->c_attr.va_gid; 9572 else 9573 newcp->c_attr.va_gid = crgetgid(cr); 9574 } 9575 9576 /* 9577 * if we're creating a directory, and the parent directory has the 9578 * set-GID bit set, set it on the new directory. 9579 * Otherwise, if the user is neither privileged nor a member of the 9580 * file's new group, clear the file's set-GID bit. 9581 */ 9582 if (dcp->c_attr.va_mode & S_ISGID && newcp->c_attr.va_type == VDIR) { 9583 newcp->c_attr.va_mode |= S_ISGID; 9584 } else if ((newcp->c_attr.va_mode & S_ISGID) && 9585 secpolicy_vnode_setids_setgids(cr, newcp->c_attr.va_gid) != 0) 9586 newcp->c_attr.va_mode &= ~S_ISGID; 9587 } 9588 9589 /* 9590 * create an acl for the newly created file. should be called right 9591 * after cachefs_creategid. 9592 */ 9593 9594 static void 9595 cachefs_createacl(cnode_t *dcp, cnode_t *newcp) 9596 { 9597 fscache_t *fscp = C_TO_FSCACHE(dcp); 9598 vsecattr_t vsec; 9599 int gotvsec = 0; 9600 int error = 0; /* placeholder */ 9601 aclent_t *aclp; 9602 o_mode_t *classp = NULL; 9603 o_mode_t gunion = 0; 9604 int i; 9605 9606 if ((fscp->fs_info.fi_mntflags & CFS_NOACL) || 9607 (! cachefs_vtype_aclok(CTOV(newcp)))) 9608 return; 9609 9610 ASSERT(dcp->c_metadata.md_flags & MD_ACL); 9611 ASSERT(MUTEX_HELD(&dcp->c_statelock)); 9612 ASSERT(MUTEX_HELD(&newcp->c_statelock)); 9613 9614 /* 9615 * XXX should probably not do VSA_ACL and VSA_ACLCNT, but that 9616 * would hit code paths that isn't hit anywhere else. 9617 */ 9618 9619 bzero(&vsec, sizeof (vsec)); 9620 vsec.vsa_mask = VSA_ACL | VSA_ACLCNT | VSA_DFACL | VSA_DFACLCNT; 9621 error = cachefs_getaclfromcache(dcp, &vsec); 9622 if (error != 0) 9623 goto out; 9624 gotvsec = 1; 9625 9626 if ((vsec.vsa_dfaclcnt > 0) && (vsec.vsa_dfaclentp != NULL)) { 9627 if ((vsec.vsa_aclcnt > 0) && (vsec.vsa_aclentp != NULL)) 9628 kmem_free(vsec.vsa_aclentp, 9629 vsec.vsa_aclcnt * sizeof (aclent_t)); 9630 9631 vsec.vsa_aclcnt = vsec.vsa_dfaclcnt; 9632 vsec.vsa_aclentp = vsec.vsa_dfaclentp; 9633 vsec.vsa_dfaclcnt = 0; 9634 vsec.vsa_dfaclentp = NULL; 9635 9636 if (newcp->c_attr.va_type == VDIR) { 9637 vsec.vsa_dfaclentp = kmem_alloc(vsec.vsa_aclcnt * 9638 sizeof (aclent_t), KM_SLEEP); 9639 vsec.vsa_dfaclcnt = vsec.vsa_aclcnt; 9640 bcopy(vsec.vsa_aclentp, vsec.vsa_dfaclentp, 9641 vsec.vsa_aclcnt * sizeof (aclent_t)); 9642 } 9643 9644 /* 9645 * this function should be called pretty much after 9646 * the rest of the file creation stuff is done. so, 9647 * uid, gid, etc. should be `right'. we'll go with 9648 * that, rather than trying to determine whether to 9649 * get stuff from cr or va. 9650 */ 9651 9652 for (i = 0; i < vsec.vsa_aclcnt; i++) { 9653 aclp = ((aclent_t *)vsec.vsa_aclentp) + i; 9654 switch (aclp->a_type) { 9655 case DEF_USER_OBJ: 9656 aclp->a_type = USER_OBJ; 9657 aclp->a_id = newcp->c_metadata.md_vattr.va_uid; 9658 aclp->a_perm = 9659 newcp->c_metadata.md_vattr.va_mode; 9660 aclp->a_perm &= 0700; 9661 aclp->a_perm >>= 6; 9662 break; 9663 9664 case DEF_GROUP_OBJ: 9665 aclp->a_type = GROUP_OBJ; 9666 aclp->a_id = newcp->c_metadata.md_vattr.va_gid; 9667 aclp->a_perm = 9668 newcp->c_metadata.md_vattr.va_mode; 9669 aclp->a_perm &= 070; 9670 aclp->a_perm >>= 3; 9671 gunion |= aclp->a_perm; 9672 break; 9673 9674 case DEF_OTHER_OBJ: 9675 aclp->a_type = OTHER_OBJ; 9676 aclp->a_perm = 9677 newcp->c_metadata.md_vattr.va_mode & 07; 9678 break; 9679 9680 case DEF_CLASS_OBJ: 9681 aclp->a_type = CLASS_OBJ; 9682 classp = &(aclp->a_perm); 9683 break; 9684 9685 case DEF_USER: 9686 aclp->a_type = USER; 9687 gunion |= aclp->a_perm; 9688 break; 9689 9690 case DEF_GROUP: 9691 aclp->a_type = GROUP; 9692 gunion |= aclp->a_perm; 9693 break; 9694 } 9695 } 9696 9697 /* XXX is this the POSIX thing to do? */ 9698 if (classp != NULL) 9699 *classp &= gunion; 9700 9701 /* 9702 * we don't need to log this; rather, we clear the 9703 * MD_ACL bit when we reconnect. 9704 */ 9705 9706 error = cachefs_cacheacl(newcp, &vsec); 9707 if (error != 0) 9708 goto out; 9709 } 9710 9711 newcp->c_metadata.md_aclclass = 07; /* XXX check posix */ 9712 newcp->c_metadata.md_flags |= MD_ACL; 9713 newcp->c_flags |= CN_UPDATED; 9714 9715 out: 9716 9717 if (gotvsec) { 9718 if ((vsec.vsa_aclcnt > 0) && (vsec.vsa_aclentp != NULL)) 9719 kmem_free(vsec.vsa_aclentp, 9720 vsec.vsa_aclcnt * sizeof (aclent_t)); 9721 if ((vsec.vsa_dfaclcnt > 0) && (vsec.vsa_dfaclentp != NULL)) 9722 kmem_free(vsec.vsa_dfaclentp, 9723 vsec.vsa_dfaclcnt * sizeof (aclent_t)); 9724 } 9725 } 9726 9727 /* 9728 * this is translated from the UFS code for access checking. 9729 */ 9730 9731 static int 9732 cachefs_access_local(void *vcp, int mode, cred_t *cr) 9733 { 9734 cnode_t *cp = vcp; 9735 fscache_t *fscp = C_TO_FSCACHE(cp); 9736 int shift = 0; 9737 9738 ASSERT(MUTEX_HELD(&cp->c_statelock)); 9739 9740 if (mode & VWRITE) { 9741 /* 9742 * Disallow write attempts on read-only 9743 * file systems, unless the file is special. 9744 */ 9745 struct vnode *vp = CTOV(cp); 9746 if (vn_is_readonly(vp)) { 9747 if (!IS_DEVVP(vp)) { 9748 return (EROFS); 9749 } 9750 } 9751 } 9752 9753 /* 9754 * if we need to do ACLs, do it. this works whether anyone 9755 * has explicitly made an ACL or not. 9756 */ 9757 9758 if (((fscp->fs_info.fi_mntflags & CFS_NOACL) == 0) && 9759 (cachefs_vtype_aclok(CTOV(cp)))) 9760 return (cachefs_acl_access(cp, mode, cr)); 9761 9762 if (crgetuid(cr) != cp->c_attr.va_uid) { 9763 shift += 3; 9764 if (!groupmember(cp->c_attr.va_gid, cr)) 9765 shift += 3; 9766 } 9767 9768 /* compute missing mode bits */ 9769 mode &= ~(cp->c_attr.va_mode << shift); 9770 9771 if (mode == 0) 9772 return (0); 9773 9774 return (secpolicy_vnode_access(cr, CTOV(cp), cp->c_attr.va_uid, mode)); 9775 } 9776 9777 /* 9778 * This is transcribed from ufs_acl_access(). If that changes, then 9779 * this should, too. 9780 * 9781 * Check the cnode's ACL's to see if this mode of access is 9782 * allowed; return 0 if allowed, EACCES if not. 9783 * 9784 * We follow the procedure defined in Sec. 3.3.5, ACL Access 9785 * Check Algorithm, of the POSIX 1003.6 Draft Standard. 9786 */ 9787 9788 #define ACL_MODE_CHECK(M, PERM, C, I) ((((M) & (PERM)) == (M)) ? 0 : \ 9789 secpolicy_vnode_access(C, CTOV(I), owner, (M) & ~(PERM))) 9790 9791 static int 9792 cachefs_acl_access(struct cnode *cp, int mode, cred_t *cr) 9793 { 9794 int error = 0; 9795 9796 fscache_t *fscp = C_TO_FSCACHE(cp); 9797 9798 int mask = ~0; 9799 int ismask = 0; 9800 9801 int gperm = 0; 9802 int ngroup = 0; 9803 9804 vsecattr_t vsec; 9805 int gotvsec = 0; 9806 aclent_t *aclp; 9807 9808 uid_t owner = cp->c_attr.va_uid; 9809 9810 int i; 9811 9812 ASSERT(MUTEX_HELD(&cp->c_statelock)); 9813 ASSERT((fscp->fs_info.fi_mntflags & CFS_NOACL) == 0); 9814 9815 /* 9816 * strictly speaking, we shouldn't set VSA_DFACL and DFACLCNT, 9817 * but then i believe we'd be the only thing exercising those 9818 * code paths -- probably a bad thing. 9819 */ 9820 9821 bzero(&vsec, sizeof (vsec)); 9822 vsec.vsa_mask = VSA_ACL | VSA_ACLCNT | VSA_DFACL | VSA_DFACLCNT; 9823 9824 /* XXX KLUDGE! correct insidious 0-class problem */ 9825 if (cp->c_metadata.md_aclclass == 0 && 9826 fscp->fs_cdconnected == CFS_CD_CONNECTED) 9827 cachefs_purgeacl(cp); 9828 again: 9829 if (cp->c_metadata.md_flags & MD_ACL) { 9830 error = cachefs_getaclfromcache(cp, &vsec); 9831 if (error != 0) { 9832 #ifdef CFSDEBUG 9833 if (error != ETIMEDOUT) 9834 CFS_DEBUG(CFSDEBUG_VOPS) 9835 printf("cachefs_acl_access():" 9836 "error %d from getaclfromcache()\n", 9837 error); 9838 #endif /* CFSDEBUG */ 9839 if ((cp->c_metadata.md_flags & MD_ACL) == 0) { 9840 goto again; 9841 } else { 9842 goto out; 9843 } 9844 } 9845 } else { 9846 if (cp->c_backvp == NULL) { 9847 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) 9848 error = cachefs_getbackvp(fscp, cp); 9849 else 9850 error = ETIMEDOUT; 9851 } 9852 if (error == 0) 9853 error = VOP_GETSECATTR(cp->c_backvp, &vsec, 0, cr); 9854 if (error != 0) { 9855 #ifdef CFSDEBUG 9856 CFS_DEBUG(CFSDEBUG_VOPS) 9857 printf("cachefs_acl_access():" 9858 "error %d from getsecattr(backvp)\n", 9859 error); 9860 #endif /* CFSDEBUG */ 9861 goto out; 9862 } 9863 if ((cp->c_flags & CN_NOCACHE) == 0 && 9864 !CFS_ISFS_BACKFS_NFSV4(fscp)) 9865 (void) cachefs_cacheacl(cp, &vsec); 9866 } 9867 gotvsec = 1; 9868 9869 ASSERT(error == 0); 9870 for (i = 0; i < vsec.vsa_aclcnt; i++) { 9871 aclp = ((aclent_t *)vsec.vsa_aclentp) + i; 9872 switch (aclp->a_type) { 9873 case USER_OBJ: 9874 /* 9875 * this might look cleaner in the 2nd loop 9876 * below, but we do it here as an 9877 * optimization. 9878 */ 9879 9880 owner = aclp->a_id; 9881 if (crgetuid(cr) == owner) { 9882 error = ACL_MODE_CHECK(mode, aclp->a_perm << 6, 9883 cr, cp); 9884 goto out; 9885 } 9886 break; 9887 9888 case CLASS_OBJ: 9889 mask = aclp->a_perm; 9890 ismask = 1; 9891 break; 9892 } 9893 } 9894 9895 ASSERT(error == 0); 9896 for (i = 0; i < vsec.vsa_aclcnt; i++) { 9897 aclp = ((aclent_t *)vsec.vsa_aclentp) + i; 9898 switch (aclp->a_type) { 9899 case USER: 9900 if (crgetuid(cr) == aclp->a_id) { 9901 error = ACL_MODE_CHECK(mode, 9902 (aclp->a_perm & mask) << 6, cr, cp); 9903 goto out; 9904 } 9905 break; 9906 9907 case GROUP_OBJ: 9908 if (groupmember(aclp->a_id, cr)) { 9909 ++ngroup; 9910 gperm |= aclp->a_perm; 9911 if (! ismask) { 9912 error = ACL_MODE_CHECK(mode, 9913 aclp->a_perm << 6, 9914 cr, cp); 9915 goto out; 9916 } 9917 } 9918 break; 9919 9920 case GROUP: 9921 if (groupmember(aclp->a_id, cr)) { 9922 ++ngroup; 9923 gperm |= aclp->a_perm; 9924 } 9925 break; 9926 9927 case OTHER_OBJ: 9928 if (ngroup == 0) { 9929 error = ACL_MODE_CHECK(mode, aclp->a_perm << 6, 9930 cr, cp); 9931 goto out; 9932 } 9933 break; 9934 9935 default: 9936 break; 9937 } 9938 } 9939 9940 ASSERT(ngroup > 0); 9941 error = ACL_MODE_CHECK(mode, (gperm & mask) << 6, cr, cp); 9942 9943 out: 9944 if (gotvsec) { 9945 if (vsec.vsa_aclcnt && vsec.vsa_aclentp) 9946 kmem_free(vsec.vsa_aclentp, 9947 vsec.vsa_aclcnt * sizeof (aclent_t)); 9948 if (vsec.vsa_dfaclcnt && vsec.vsa_dfaclentp) 9949 kmem_free(vsec.vsa_dfaclentp, 9950 vsec.vsa_dfaclcnt * sizeof (aclent_t)); 9951 } 9952 9953 return (error); 9954 } 9955 9956 /* 9957 * see if permissions allow for removal of the given file from 9958 * the given directory. 9959 */ 9960 static int 9961 cachefs_stickyrmchk(struct cnode *dcp, struct cnode *cp, cred_t *cr) 9962 { 9963 uid_t uid; 9964 /* 9965 * If the containing directory is sticky, the user must: 9966 * - own the directory, or 9967 * - own the file, or 9968 * - be able to write the file (if it's a plain file), or 9969 * - be sufficiently privileged. 9970 */ 9971 if ((dcp->c_attr.va_mode & S_ISVTX) && 9972 ((uid = crgetuid(cr)) != dcp->c_attr.va_uid) && 9973 (uid != cp->c_attr.va_uid) && 9974 (cp->c_attr.va_type != VREG || 9975 cachefs_access_local(cp, VWRITE, cr) != 0)) 9976 return (secpolicy_vnode_remove(cr)); 9977 9978 return (0); 9979 } 9980 9981 /* 9982 * Returns a new name, may even be unique. 9983 * Stolen from nfs code. 9984 * Since now we will use renaming to .cfs* in place of .nfs* 9985 * for CacheFS. Both NFS and CacheFS will rename opened files. 9986 */ 9987 static char cachefs_prefix[] = ".cfs"; 9988 kmutex_t cachefs_newnum_lock; 9989 9990 static char * 9991 cachefs_newname(void) 9992 { 9993 static uint_t newnum = 0; 9994 char *news; 9995 char *s, *p; 9996 uint_t id; 9997 9998 mutex_enter(&cachefs_newnum_lock); 9999 if (newnum == 0) { 10000 newnum = gethrestime_sec() & 0xfffff; 10001 newnum |= 0x10000; 10002 } 10003 id = newnum++; 10004 mutex_exit(&cachefs_newnum_lock); 10005 10006 news = cachefs_kmem_alloc(MAXNAMELEN, KM_SLEEP); 10007 s = news; 10008 p = cachefs_prefix; 10009 while (*p != '\0') 10010 *s++ = *p++; 10011 while (id != 0) { 10012 *s++ = "0123456789ABCDEF"[id & 0x0f]; 10013 id >>= 4; 10014 } 10015 *s = '\0'; 10016 return (news); 10017 } 10018 10019 /* 10020 * Called to rename the specified file to a temporary file so 10021 * operations to the file after remove work. 10022 * Must call this routine with the dir c_rwlock held as a writer. 10023 */ 10024 static int 10025 /*ARGSUSED*/ 10026 cachefs_remove_dolink(vnode_t *dvp, vnode_t *vp, char *nm, cred_t *cr) 10027 { 10028 cnode_t *cp = VTOC(vp); 10029 char *tmpname; 10030 fscache_t *fscp = C_TO_FSCACHE(cp); 10031 int error; 10032 10033 ASSERT(RW_WRITE_HELD(&(VTOC(dvp)->c_rwlock))); 10034 10035 /* get the new name for the file */ 10036 tmpname = cachefs_newname(); 10037 10038 /* do the link */ 10039 if (fscp->fs_cdconnected == CFS_CD_CONNECTED) 10040 error = cachefs_link_connected(dvp, vp, tmpname, cr); 10041 else 10042 error = cachefs_link_disconnected(dvp, vp, tmpname, cr); 10043 if (error) { 10044 cachefs_kmem_free(tmpname, MAXNAMELEN); 10045 return (error); 10046 } 10047 10048 mutex_enter(&cp->c_statelock); 10049 if (cp->c_unldvp) { 10050 VN_RELE(cp->c_unldvp); 10051 cachefs_kmem_free(cp->c_unlname, MAXNAMELEN); 10052 crfree(cp->c_unlcred); 10053 } 10054 10055 VN_HOLD(dvp); 10056 cp->c_unldvp = dvp; 10057 crhold(cr); 10058 cp->c_unlcred = cr; 10059 cp->c_unlname = tmpname; 10060 10061 /* drop the backvp so NFS does not also do a rename */ 10062 mutex_exit(&cp->c_statelock); 10063 10064 return (0); 10065 } 10066 10067 /* 10068 * Marks the cnode as modified. 10069 */ 10070 static void 10071 cachefs_modified(cnode_t *cp) 10072 { 10073 fscache_t *fscp = C_TO_FSCACHE(cp); 10074 struct vattr va; 10075 int error; 10076 10077 ASSERT(MUTEX_HELD(&cp->c_statelock)); 10078 ASSERT(cp->c_metadata.md_rlno); 10079 10080 /* if not on the modify list */ 10081 if (cp->c_metadata.md_rltype != CACHEFS_RL_MODIFIED) { 10082 /* put on modified list, also marks the file as modified */ 10083 cachefs_rlent_moveto(fscp->fs_cache, CACHEFS_RL_MODIFIED, 10084 cp->c_metadata.md_rlno, cp->c_metadata.md_frontblks); 10085 cp->c_metadata.md_rltype = CACHEFS_RL_MODIFIED; 10086 cp->c_flags |= CN_UPDATED; 10087 10088 /* if a modified regular file that is not local */ 10089 if (((cp->c_id.cid_flags & CFS_CID_LOCAL) == 0) && 10090 (cp->c_metadata.md_flags & MD_FILE) && 10091 (cp->c_attr.va_type == VREG)) { 10092 10093 if (cp->c_frontvp == NULL) 10094 (void) cachefs_getfrontfile(cp); 10095 if (cp->c_frontvp) { 10096 /* identify file so fsck knows it is modified */ 10097 va.va_mode = 0766; 10098 va.va_mask = AT_MODE; 10099 error = VOP_SETATTR(cp->c_frontvp, &va, 10100 0, kcred, NULL); 10101 if (error) { 10102 cmn_err(CE_WARN, 10103 "Cannot change ff mode.\n"); 10104 } 10105 } 10106 } 10107 } 10108 } 10109 10110 /* 10111 * Marks the cnode as modified. 10112 * Allocates a rl slot for the cnode if necessary. 10113 * Returns 0 for success, !0 if cannot get an rl slot. 10114 */ 10115 static int 10116 cachefs_modified_alloc(cnode_t *cp) 10117 { 10118 fscache_t *fscp = C_TO_FSCACHE(cp); 10119 filegrp_t *fgp = cp->c_filegrp; 10120 int error; 10121 rl_entry_t rl_ent; 10122 10123 ASSERT(MUTEX_HELD(&cp->c_statelock)); 10124 10125 /* get the rl slot if needed */ 10126 if (cp->c_metadata.md_rlno == 0) { 10127 /* get a metadata slot if we do not have one yet */ 10128 if (cp->c_flags & CN_ALLOC_PENDING) { 10129 if (cp->c_filegrp->fg_flags & CFS_FG_ALLOC_ATTR) { 10130 (void) filegrp_allocattr(cp->c_filegrp); 10131 } 10132 error = filegrp_create_metadata(cp->c_filegrp, 10133 &cp->c_metadata, &cp->c_id); 10134 if (error) 10135 return (error); 10136 cp->c_flags &= ~CN_ALLOC_PENDING; 10137 } 10138 10139 /* get a free rl entry */ 10140 rl_ent.rl_fileno = cp->c_id.cid_fileno; 10141 rl_ent.rl_local = (cp->c_id.cid_flags & CFS_CID_LOCAL) ? 1 : 0; 10142 rl_ent.rl_fsid = fscp->fs_cfsid; 10143 rl_ent.rl_attrc = 0; 10144 error = cachefs_rl_alloc(fscp->fs_cache, &rl_ent, 10145 &cp->c_metadata.md_rlno); 10146 if (error) 10147 return (error); 10148 cp->c_metadata.md_rltype = CACHEFS_RL_NONE; 10149 10150 /* hold the filegrp so the attrcache file is not gc */ 10151 error = filegrp_ffhold(fgp); 10152 if (error) { 10153 cachefs_rlent_moveto(fscp->fs_cache, 10154 CACHEFS_RL_FREE, cp->c_metadata.md_rlno, 0); 10155 cp->c_metadata.md_rlno = 0; 10156 return (error); 10157 } 10158 } 10159 cachefs_modified(cp); 10160 return (0); 10161 } 10162 10163 int 10164 cachefs_vtype_aclok(vnode_t *vp) 10165 { 10166 vtype_t *vtp, oktypes[] = {VREG, VDIR, VFIFO, VNON}; 10167 10168 if (vp->v_type == VNON) 10169 return (0); 10170 10171 for (vtp = oktypes; *vtp != VNON; vtp++) 10172 if (vp->v_type == *vtp) 10173 break; 10174 10175 return (*vtp != VNON); 10176 } 10177 10178 static int 10179 cachefs_pathconf(vnode_t *vp, int cmd, ulong_t *valp, cred_t *cr) 10180 { 10181 int error = 0; 10182 fscache_t *fscp = C_TO_FSCACHE(VTOC(vp)); 10183 10184 /* Assert cachefs compatibility if NFSv4 is in use */ 10185 CFS_BACKFS_NFSV4_ASSERT_FSCACHE(fscp); 10186 CFS_BACKFS_NFSV4_ASSERT_CNODE(VTOC(vp)); 10187 10188 if (cmd == _PC_FILESIZEBITS) { 10189 u_offset_t maxsize = fscp->fs_offmax; 10190 (*valp) = 0; 10191 while (maxsize != 0) { 10192 maxsize >>= 1; 10193 (*valp)++; 10194 } 10195 (*valp)++; 10196 } else 10197 error = fs_pathconf(vp, cmd, valp, cr); 10198 10199 return (error); 10200 } 10201