1d1fa59e9SXin LI /* $NetBSD: tmpfs_vfsops.c,v 1.10 2005/12/11 12:24:29 christos Exp $ */ 2d1fa59e9SXin LI 3e08d5567SXin LI /*- 4d63027b6SPedro F. Giffuni * SPDX-License-Identifier: BSD-2-Clause-NetBSD 5d63027b6SPedro F. Giffuni * 6d1fa59e9SXin LI * Copyright (c) 2005 The NetBSD Foundation, Inc. 7d1fa59e9SXin LI * All rights reserved. 8d1fa59e9SXin LI * 9d1fa59e9SXin LI * This code is derived from software contributed to The NetBSD Foundation 10d1fa59e9SXin LI * by Julio M. Merino Vidal, developed as part of Google's Summer of Code 11d1fa59e9SXin LI * 2005 program. 12d1fa59e9SXin LI * 13d1fa59e9SXin LI * Redistribution and use in source and binary forms, with or without 14d1fa59e9SXin LI * modification, are permitted provided that the following conditions 15d1fa59e9SXin LI * are met: 16d1fa59e9SXin LI * 1. Redistributions of source code must retain the above copyright 17d1fa59e9SXin LI * notice, this list of conditions and the following disclaimer. 18d1fa59e9SXin LI * 2. Redistributions in binary form must reproduce the above copyright 19d1fa59e9SXin LI * notice, this list of conditions and the following disclaimer in the 20d1fa59e9SXin LI * documentation and/or other materials provided with the distribution. 21d1fa59e9SXin LI * 22d1fa59e9SXin LI * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS 23d1fa59e9SXin LI * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED 24d1fa59e9SXin LI * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR 25d1fa59e9SXin LI * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS 26d1fa59e9SXin LI * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 27d1fa59e9SXin LI * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 28d1fa59e9SXin LI * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 29d1fa59e9SXin LI * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 30d1fa59e9SXin LI * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 31d1fa59e9SXin LI * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE 32d1fa59e9SXin LI * POSSIBILITY OF SUCH DAMAGE. 33d1fa59e9SXin LI */ 34d1fa59e9SXin LI 35d1fa59e9SXin LI /* 36d1fa59e9SXin LI * Efficient memory file system. 37d1fa59e9SXin LI * 383544b0f6SKonstantin Belousov * tmpfs is a file system that uses FreeBSD's virtual memory 393544b0f6SKonstantin Belousov * sub-system to store file data and metadata in an efficient way. 403544b0f6SKonstantin Belousov * This means that it does not follow the structure of an on-disk file 413544b0f6SKonstantin Belousov * system because it simply does not need to. Instead, it uses 42d1fa59e9SXin LI * memory-specific data structures and algorithms to automatically 43d1fa59e9SXin LI * allocate and release resources. 44d1fa59e9SXin LI */ 45b4b3e349SAllan Jude 46b4b3e349SAllan Jude #include "opt_tmpfs.h" 47b4b3e349SAllan Jude 48d1fa59e9SXin LI #include <sys/cdefs.h> 49d1fa59e9SXin LI __FBSDID("$FreeBSD$"); 50d1fa59e9SXin LI 51d1fa59e9SXin LI #include <sys/param.h> 526d2e2df7SMark Johnston #include <sys/systm.h> 53135beaf6SGleb Smirnoff #include <sys/dirent.h> 541df86a32SXin LI #include <sys/limits.h> 55d1fa59e9SXin LI #include <sys/lock.h> 56135beaf6SGleb Smirnoff #include <sys/mount.h> 57d1fa59e9SXin LI #include <sys/mutex.h> 582454886eSXin LI #include <sys/proc.h> 592454886eSXin LI #include <sys/jail.h> 60d1fa59e9SXin LI #include <sys/kernel.h> 61f40cb1c6SKonstantin Belousov #include <sys/rwlock.h> 62d1fa59e9SXin LI #include <sys/stat.h> 63*5c4ce6faSKonstantin Belousov #include <sys/sx.h> 64d1fa59e9SXin LI #include <sys/sysctl.h> 65135beaf6SGleb Smirnoff #include <sys/vnode.h> 66d1fa59e9SXin LI 67d1fa59e9SXin LI #include <vm/vm.h> 68*5c4ce6faSKonstantin Belousov #include <vm/vm_param.h> 69*5c4ce6faSKonstantin Belousov #include <vm/pmap.h> 70*5c4ce6faSKonstantin Belousov #include <vm/vm_extern.h> 71*5c4ce6faSKonstantin Belousov #include <vm/vm_map.h> 72d1fa59e9SXin LI #include <vm/vm_object.h> 73d1fa59e9SXin LI #include <vm/vm_param.h> 74d1fa59e9SXin LI 75d1fa59e9SXin LI #include <fs/tmpfs/tmpfs.h> 76d1fa59e9SXin LI 77d1fa59e9SXin LI /* 78d1fa59e9SXin LI * Default permission for root node 79d1fa59e9SXin LI */ 80d1fa59e9SXin LI #define TMPFS_DEFAULT_ROOT_MODE (S_IRWXU|S_IRGRP|S_IXGRP|S_IROTH|S_IXOTH) 81d1fa59e9SXin LI 82d1fa59e9SXin LI MALLOC_DEFINE(M_TMPFSMNT, "tmpfs mount", "tmpfs mount structures"); 839b258fcaSXin LI MALLOC_DEFINE(M_TMPFSNAME, "tmpfs name", "tmpfs file names"); 84d1fa59e9SXin LI 85dfd233edSAttilio Rao static int tmpfs_mount(struct mount *); 86dfd233edSAttilio Rao static int tmpfs_unmount(struct mount *, int); 87dfd233edSAttilio Rao static int tmpfs_root(struct mount *, int flags, struct vnode **); 88694a586aSRick Macklem static int tmpfs_fhtovp(struct mount *, struct fid *, int, 89694a586aSRick Macklem struct vnode **); 90dfd233edSAttilio Rao static int tmpfs_statfs(struct mount *, struct statfs *); 915f34e93cSMark Johnston static void tmpfs_susp_clean(struct mount *); 92d1fa59e9SXin LI 93d1fa59e9SXin LI static const char *tmpfs_opts[] = { 94dec3772eSJaakko Heinonen "from", "size", "maxfilesize", "inodes", "uid", "gid", "mode", "export", 9500ac6a98SKonstantin Belousov "union", "nonc", NULL 96d1fa59e9SXin LI }; 97d1fa59e9SXin LI 98c5ab5ce3SJaakko Heinonen static const char *tmpfs_updateopts[] = { 99ac1a10efSMaxim Sobolev "from", "export", "size", NULL 100c5ab5ce3SJaakko Heinonen }; 101c5ab5ce3SJaakko Heinonen 1027adb1776SXin LI static int 1037adb1776SXin LI tmpfs_node_ctor(void *mem, int size, void *arg, int flags) 1047adb1776SXin LI { 1057adb1776SXin LI struct tmpfs_node *node = (struct tmpfs_node *)mem; 1067adb1776SXin LI 1077adb1776SXin LI node->tn_gen++; 1087adb1776SXin LI node->tn_size = 0; 1097adb1776SXin LI node->tn_status = 0; 1107adb1776SXin LI node->tn_flags = 0; 1117adb1776SXin LI node->tn_links = 0; 1127adb1776SXin LI node->tn_vnode = NULL; 1137adb1776SXin LI node->tn_vpstate = 0; 1147adb1776SXin LI 1157adb1776SXin LI return (0); 1167adb1776SXin LI } 1177adb1776SXin LI 1187adb1776SXin LI static void 1197adb1776SXin LI tmpfs_node_dtor(void *mem, int size, void *arg) 1207adb1776SXin LI { 1217adb1776SXin LI struct tmpfs_node *node = (struct tmpfs_node *)mem; 1227adb1776SXin LI node->tn_type = VNON; 1237adb1776SXin LI } 1247adb1776SXin LI 1257adb1776SXin LI static int 1267adb1776SXin LI tmpfs_node_init(void *mem, int size, int flags) 1277adb1776SXin LI { 1287adb1776SXin LI struct tmpfs_node *node = (struct tmpfs_node *)mem; 1297adb1776SXin LI node->tn_id = 0; 1307adb1776SXin LI 1317adb1776SXin LI mtx_init(&node->tn_interlock, "tmpfs node interlock", NULL, MTX_DEF); 1328d9a89a3SXin LI node->tn_gen = arc4random(); 1337adb1776SXin LI 1347adb1776SXin LI return (0); 1357adb1776SXin LI } 1367adb1776SXin LI 1377adb1776SXin LI static void 1387adb1776SXin LI tmpfs_node_fini(void *mem, int size) 1397adb1776SXin LI { 1407adb1776SXin LI struct tmpfs_node *node = (struct tmpfs_node *)mem; 1417adb1776SXin LI 1427adb1776SXin LI mtx_destroy(&node->tn_interlock); 1437adb1776SXin LI } 144d1fa59e9SXin LI 145*5c4ce6faSKonstantin Belousov /* 146*5c4ce6faSKonstantin Belousov * Handle updates of time from writes to mmaped regions. Use 147*5c4ce6faSKonstantin Belousov * MNT_VNODE_FOREACH_ALL instead of MNT_VNODE_FOREACH_ACTIVE, since 148*5c4ce6faSKonstantin Belousov * unmap of the tmpfs-backed vnode does not call vinactive(), due to 149*5c4ce6faSKonstantin Belousov * vm object type is OBJT_SWAP. 150*5c4ce6faSKonstantin Belousov * If lazy, only handle delayed update of mtime due to the writes to 151*5c4ce6faSKonstantin Belousov * mapped files. 152*5c4ce6faSKonstantin Belousov */ 153*5c4ce6faSKonstantin Belousov static void 154*5c4ce6faSKonstantin Belousov tmpfs_update_mtime(struct mount *mp, bool lazy) 155*5c4ce6faSKonstantin Belousov { 156*5c4ce6faSKonstantin Belousov struct vnode *vp, *mvp; 157*5c4ce6faSKonstantin Belousov struct vm_object *obj; 158*5c4ce6faSKonstantin Belousov 159*5c4ce6faSKonstantin Belousov MNT_VNODE_FOREACH_ALL(vp, mp, mvp) { 160*5c4ce6faSKonstantin Belousov if (vp->v_type != VREG) { 161*5c4ce6faSKonstantin Belousov VI_UNLOCK(vp); 162*5c4ce6faSKonstantin Belousov continue; 163*5c4ce6faSKonstantin Belousov } 164*5c4ce6faSKonstantin Belousov obj = vp->v_object; 165*5c4ce6faSKonstantin Belousov KASSERT((obj->flags & (OBJ_TMPFS_NODE | OBJ_TMPFS)) == 166*5c4ce6faSKonstantin Belousov (OBJ_TMPFS_NODE | OBJ_TMPFS), ("non-tmpfs obj")); 167*5c4ce6faSKonstantin Belousov 168*5c4ce6faSKonstantin Belousov /* 169*5c4ce6faSKonstantin Belousov * In lazy case, do unlocked read, avoid taking vnode 170*5c4ce6faSKonstantin Belousov * lock if not needed. Lost update will be handled on 171*5c4ce6faSKonstantin Belousov * the next call. 172*5c4ce6faSKonstantin Belousov * For non-lazy case, we must flush all pending 173*5c4ce6faSKonstantin Belousov * metadata changes now. 174*5c4ce6faSKonstantin Belousov */ 175*5c4ce6faSKonstantin Belousov if (!lazy || (obj->flags & OBJ_TMPFS_DIRTY) != 0) { 176*5c4ce6faSKonstantin Belousov if (vget(vp, LK_EXCLUSIVE | LK_RETRY | LK_INTERLOCK, 177*5c4ce6faSKonstantin Belousov curthread) != 0) 178*5c4ce6faSKonstantin Belousov continue; 179*5c4ce6faSKonstantin Belousov tmpfs_check_mtime(vp); 180*5c4ce6faSKonstantin Belousov if (!lazy) 181*5c4ce6faSKonstantin Belousov tmpfs_update(vp); 182*5c4ce6faSKonstantin Belousov vput(vp); 183*5c4ce6faSKonstantin Belousov } else { 184*5c4ce6faSKonstantin Belousov VI_UNLOCK(vp); 185*5c4ce6faSKonstantin Belousov continue; 186*5c4ce6faSKonstantin Belousov } 187*5c4ce6faSKonstantin Belousov } 188*5c4ce6faSKonstantin Belousov } 189*5c4ce6faSKonstantin Belousov 190*5c4ce6faSKonstantin Belousov struct tmpfs_check_rw_maps_arg { 191*5c4ce6faSKonstantin Belousov bool found; 192*5c4ce6faSKonstantin Belousov }; 193*5c4ce6faSKonstantin Belousov 194*5c4ce6faSKonstantin Belousov static bool 195*5c4ce6faSKonstantin Belousov tmpfs_check_rw_maps_cb(struct mount *mp __unused, vm_map_t map __unused, 196*5c4ce6faSKonstantin Belousov vm_map_entry_t entry __unused, void *arg) 197*5c4ce6faSKonstantin Belousov { 198*5c4ce6faSKonstantin Belousov struct tmpfs_check_rw_maps_arg *a; 199*5c4ce6faSKonstantin Belousov 200*5c4ce6faSKonstantin Belousov a = arg; 201*5c4ce6faSKonstantin Belousov a->found = true; 202*5c4ce6faSKonstantin Belousov return (true); 203*5c4ce6faSKonstantin Belousov } 204*5c4ce6faSKonstantin Belousov 205*5c4ce6faSKonstantin Belousov /* 206*5c4ce6faSKonstantin Belousov * Revoke write permissions from all mappings of regular files 207*5c4ce6faSKonstantin Belousov * belonging to the specified tmpfs mount. 208*5c4ce6faSKonstantin Belousov */ 209*5c4ce6faSKonstantin Belousov static bool 210*5c4ce6faSKonstantin Belousov tmpfs_revoke_rw_maps_cb(struct mount *mp __unused, vm_map_t map, 211*5c4ce6faSKonstantin Belousov vm_map_entry_t entry, void *arg __unused) 212*5c4ce6faSKonstantin Belousov { 213*5c4ce6faSKonstantin Belousov 214*5c4ce6faSKonstantin Belousov /* 215*5c4ce6faSKonstantin Belousov * XXXKIB: might be invalidate the mapping 216*5c4ce6faSKonstantin Belousov * instead ? The process is not going to be 217*5c4ce6faSKonstantin Belousov * happy in any case. 218*5c4ce6faSKonstantin Belousov */ 219*5c4ce6faSKonstantin Belousov entry->max_protection &= ~VM_PROT_WRITE; 220*5c4ce6faSKonstantin Belousov if ((entry->protection & VM_PROT_WRITE) != 0) { 221*5c4ce6faSKonstantin Belousov entry->protection &= ~VM_PROT_WRITE; 222*5c4ce6faSKonstantin Belousov pmap_protect(map->pmap, entry->start, entry->end, 223*5c4ce6faSKonstantin Belousov entry->protection); 224*5c4ce6faSKonstantin Belousov } 225*5c4ce6faSKonstantin Belousov return (false); 226*5c4ce6faSKonstantin Belousov } 227*5c4ce6faSKonstantin Belousov 228*5c4ce6faSKonstantin Belousov static void 229*5c4ce6faSKonstantin Belousov tmpfs_all_rw_maps(struct mount *mp, bool (*cb)(struct mount *mp, vm_map_t, 230*5c4ce6faSKonstantin Belousov vm_map_entry_t, void *), void *cb_arg) 231*5c4ce6faSKonstantin Belousov { 232*5c4ce6faSKonstantin Belousov struct proc *p; 233*5c4ce6faSKonstantin Belousov struct vmspace *vm; 234*5c4ce6faSKonstantin Belousov vm_map_t map; 235*5c4ce6faSKonstantin Belousov vm_map_entry_t entry; 236*5c4ce6faSKonstantin Belousov vm_object_t object; 237*5c4ce6faSKonstantin Belousov struct vnode *vp; 238*5c4ce6faSKonstantin Belousov int gen; 239*5c4ce6faSKonstantin Belousov bool terminate; 240*5c4ce6faSKonstantin Belousov 241*5c4ce6faSKonstantin Belousov terminate = false; 242*5c4ce6faSKonstantin Belousov sx_slock(&allproc_lock); 243*5c4ce6faSKonstantin Belousov again: 244*5c4ce6faSKonstantin Belousov gen = allproc_gen; 245*5c4ce6faSKonstantin Belousov FOREACH_PROC_IN_SYSTEM(p) { 246*5c4ce6faSKonstantin Belousov PROC_LOCK(p); 247*5c4ce6faSKonstantin Belousov if (p->p_state != PRS_NORMAL || (p->p_flag & (P_INEXEC | 248*5c4ce6faSKonstantin Belousov P_SYSTEM | P_WEXIT)) != 0) { 249*5c4ce6faSKonstantin Belousov PROC_UNLOCK(p); 250*5c4ce6faSKonstantin Belousov continue; 251*5c4ce6faSKonstantin Belousov } 252*5c4ce6faSKonstantin Belousov vm = vmspace_acquire_ref(p); 253*5c4ce6faSKonstantin Belousov _PHOLD_LITE(p); 254*5c4ce6faSKonstantin Belousov PROC_UNLOCK(p); 255*5c4ce6faSKonstantin Belousov if (vm == NULL) { 256*5c4ce6faSKonstantin Belousov PRELE(p); 257*5c4ce6faSKonstantin Belousov continue; 258*5c4ce6faSKonstantin Belousov } 259*5c4ce6faSKonstantin Belousov sx_sunlock(&allproc_lock); 260*5c4ce6faSKonstantin Belousov map = &vm->vm_map; 261*5c4ce6faSKonstantin Belousov 262*5c4ce6faSKonstantin Belousov vm_map_lock(map); 263*5c4ce6faSKonstantin Belousov if (map->busy) 264*5c4ce6faSKonstantin Belousov vm_map_wait_busy(map); 265*5c4ce6faSKonstantin Belousov for (entry = map->header.next; entry != &map->header; 266*5c4ce6faSKonstantin Belousov entry = entry->next) { 267*5c4ce6faSKonstantin Belousov if ((entry->eflags & (MAP_ENTRY_GUARD | 268*5c4ce6faSKonstantin Belousov MAP_ENTRY_IS_SUB_MAP | MAP_ENTRY_COW)) != 0 || 269*5c4ce6faSKonstantin Belousov (entry->max_protection & VM_PROT_WRITE) == 0) 270*5c4ce6faSKonstantin Belousov continue; 271*5c4ce6faSKonstantin Belousov object = entry->object.vm_object; 272*5c4ce6faSKonstantin Belousov if (object == NULL || object->type != OBJT_SWAP || 273*5c4ce6faSKonstantin Belousov (object->flags & OBJ_TMPFS_NODE) == 0) 274*5c4ce6faSKonstantin Belousov continue; 275*5c4ce6faSKonstantin Belousov /* 276*5c4ce6faSKonstantin Belousov * No need to dig into shadow chain, mapping 277*5c4ce6faSKonstantin Belousov * of the object not at top is readonly. 278*5c4ce6faSKonstantin Belousov */ 279*5c4ce6faSKonstantin Belousov 280*5c4ce6faSKonstantin Belousov VM_OBJECT_RLOCK(object); 281*5c4ce6faSKonstantin Belousov if (object->type == OBJT_DEAD) { 282*5c4ce6faSKonstantin Belousov VM_OBJECT_RUNLOCK(object); 283*5c4ce6faSKonstantin Belousov continue; 284*5c4ce6faSKonstantin Belousov } 285*5c4ce6faSKonstantin Belousov MPASS(object->ref_count > 1); 286*5c4ce6faSKonstantin Belousov if ((object->flags & (OBJ_TMPFS_NODE | OBJ_TMPFS)) != 287*5c4ce6faSKonstantin Belousov (OBJ_TMPFS_NODE | OBJ_TMPFS)) { 288*5c4ce6faSKonstantin Belousov VM_OBJECT_RUNLOCK(object); 289*5c4ce6faSKonstantin Belousov continue; 290*5c4ce6faSKonstantin Belousov } 291*5c4ce6faSKonstantin Belousov vp = object->un_pager.swp.swp_tmpfs; 292*5c4ce6faSKonstantin Belousov if (vp->v_mount != mp) { 293*5c4ce6faSKonstantin Belousov VM_OBJECT_RUNLOCK(object); 294*5c4ce6faSKonstantin Belousov continue; 295*5c4ce6faSKonstantin Belousov } 296*5c4ce6faSKonstantin Belousov 297*5c4ce6faSKonstantin Belousov terminate = cb(mp, map, entry, cb_arg); 298*5c4ce6faSKonstantin Belousov VM_OBJECT_RUNLOCK(object); 299*5c4ce6faSKonstantin Belousov if (terminate) 300*5c4ce6faSKonstantin Belousov break; 301*5c4ce6faSKonstantin Belousov } 302*5c4ce6faSKonstantin Belousov vm_map_unlock(map); 303*5c4ce6faSKonstantin Belousov 304*5c4ce6faSKonstantin Belousov vmspace_free(vm); 305*5c4ce6faSKonstantin Belousov sx_slock(&allproc_lock); 306*5c4ce6faSKonstantin Belousov PRELE(p); 307*5c4ce6faSKonstantin Belousov if (terminate) 308*5c4ce6faSKonstantin Belousov break; 309*5c4ce6faSKonstantin Belousov } 310*5c4ce6faSKonstantin Belousov if (!terminate && gen != allproc_gen) 311*5c4ce6faSKonstantin Belousov goto again; 312*5c4ce6faSKonstantin Belousov sx_sunlock(&allproc_lock); 313*5c4ce6faSKonstantin Belousov } 314*5c4ce6faSKonstantin Belousov 315*5c4ce6faSKonstantin Belousov static bool 316*5c4ce6faSKonstantin Belousov tmpfs_check_rw_maps(struct mount *mp) 317*5c4ce6faSKonstantin Belousov { 318*5c4ce6faSKonstantin Belousov struct tmpfs_check_rw_maps_arg ca; 319*5c4ce6faSKonstantin Belousov 320*5c4ce6faSKonstantin Belousov ca.found = false; 321*5c4ce6faSKonstantin Belousov tmpfs_all_rw_maps(mp, tmpfs_check_rw_maps_cb, &ca); 322*5c4ce6faSKonstantin Belousov return (ca.found); 323*5c4ce6faSKonstantin Belousov } 324*5c4ce6faSKonstantin Belousov 325*5c4ce6faSKonstantin Belousov static int 326*5c4ce6faSKonstantin Belousov tmpfs_rw_to_ro(struct mount *mp) 327*5c4ce6faSKonstantin Belousov { 328*5c4ce6faSKonstantin Belousov int error, flags; 329*5c4ce6faSKonstantin Belousov bool forced; 330*5c4ce6faSKonstantin Belousov 331*5c4ce6faSKonstantin Belousov forced = (mp->mnt_flag & MNT_FORCE) != 0; 332*5c4ce6faSKonstantin Belousov flags = WRITECLOSE | (forced ? FORCECLOSE : 0); 333*5c4ce6faSKonstantin Belousov 334*5c4ce6faSKonstantin Belousov if ((error = vn_start_write(NULL, &mp, V_WAIT)) != 0) 335*5c4ce6faSKonstantin Belousov return (error); 336*5c4ce6faSKonstantin Belousov error = vfs_write_suspend_umnt(mp); 337*5c4ce6faSKonstantin Belousov if (error != 0) 338*5c4ce6faSKonstantin Belousov return (error); 339*5c4ce6faSKonstantin Belousov if (!forced && tmpfs_check_rw_maps(mp)) { 340*5c4ce6faSKonstantin Belousov error = EBUSY; 341*5c4ce6faSKonstantin Belousov goto out; 342*5c4ce6faSKonstantin Belousov } 343*5c4ce6faSKonstantin Belousov VFS_TO_TMPFS(mp)->tm_ronly = 1; 344*5c4ce6faSKonstantin Belousov MNT_ILOCK(mp); 345*5c4ce6faSKonstantin Belousov mp->mnt_flag |= MNT_RDONLY; 346*5c4ce6faSKonstantin Belousov MNT_IUNLOCK(mp); 347*5c4ce6faSKonstantin Belousov for (;;) { 348*5c4ce6faSKonstantin Belousov tmpfs_all_rw_maps(mp, tmpfs_revoke_rw_maps_cb, NULL); 349*5c4ce6faSKonstantin Belousov tmpfs_update_mtime(mp, false); 350*5c4ce6faSKonstantin Belousov error = vflush(mp, 0, flags, curthread); 351*5c4ce6faSKonstantin Belousov if (error != 0) { 352*5c4ce6faSKonstantin Belousov VFS_TO_TMPFS(mp)->tm_ronly = 0; 353*5c4ce6faSKonstantin Belousov MNT_ILOCK(mp); 354*5c4ce6faSKonstantin Belousov mp->mnt_flag &= ~MNT_RDONLY; 355*5c4ce6faSKonstantin Belousov MNT_IUNLOCK(mp); 356*5c4ce6faSKonstantin Belousov goto out; 357*5c4ce6faSKonstantin Belousov } 358*5c4ce6faSKonstantin Belousov if (!tmpfs_check_rw_maps(mp)) 359*5c4ce6faSKonstantin Belousov break; 360*5c4ce6faSKonstantin Belousov } 361*5c4ce6faSKonstantin Belousov out: 362*5c4ce6faSKonstantin Belousov vfs_write_resume(mp, 0); 363*5c4ce6faSKonstantin Belousov return (error); 364*5c4ce6faSKonstantin Belousov } 365*5c4ce6faSKonstantin Belousov 366d1fa59e9SXin LI static int 367dfd233edSAttilio Rao tmpfs_mount(struct mount *mp) 368d1fa59e9SXin LI { 3699295c628SGleb Kurtsou const size_t nodes_per_page = howmany(PAGE_SIZE, 3709295c628SGleb Kurtsou sizeof(struct tmpfs_dirent) + sizeof(struct tmpfs_node)); 371d1fa59e9SXin LI struct tmpfs_mount *tmp; 372d1fa59e9SXin LI struct tmpfs_node *root; 373*5c4ce6faSKonstantin Belousov int error; 37400ac6a98SKonstantin Belousov bool nonc; 3751df86a32SXin LI /* Size counters. */ 3760ff93c48SGleb Kurtsou u_quad_t pages; 3770ff93c48SGleb Kurtsou off_t nodes_max, size_max, maxfilesize; 3781df86a32SXin LI 3791df86a32SXin LI /* Root node attributes. */ 3801df86a32SXin LI uid_t root_uid; 3811df86a32SXin LI gid_t root_gid; 3821df86a32SXin LI mode_t root_mode; 3831df86a32SXin LI 3841df86a32SXin LI struct vattr va; 385d1fa59e9SXin LI 386d1fa59e9SXin LI if (vfs_filteropt(mp->mnt_optnew, tmpfs_opts)) 387d1fa59e9SXin LI return (EINVAL); 388d1fa59e9SXin LI 389d1fa59e9SXin LI if (mp->mnt_flag & MNT_UPDATE) { 390c5ab5ce3SJaakko Heinonen /* Only support update mounts for certain options. */ 391c5ab5ce3SJaakko Heinonen if (vfs_filteropt(mp->mnt_optnew, tmpfs_updateopts) != 0) 392e0d3195bSKevin Lo return (EOPNOTSUPP); 393ac1a10efSMaxim Sobolev if (vfs_getopt_size(mp->mnt_optnew, "size", &size_max) == 0) { 394ac1a10efSMaxim Sobolev /* 395ac1a10efSMaxim Sobolev * On-the-fly resizing is not supported (yet). We still 396ac1a10efSMaxim Sobolev * need to have "size" listed as "supported", otherwise 397ac1a10efSMaxim Sobolev * trying to update fs that is listed in fstab with size 398ac1a10efSMaxim Sobolev * parameter, say trying to change rw to ro or vice 399ac1a10efSMaxim Sobolev * versa, would cause vfs_filteropt() to bail. 400ac1a10efSMaxim Sobolev */ 4014f207061SMaxim Sobolev if (size_max != VFS_TO_TMPFS(mp)->tm_size_max) 402c5ab5ce3SJaakko Heinonen return (EOPNOTSUPP); 403ac1a10efSMaxim Sobolev } 404ac1a10efSMaxim Sobolev if (vfs_flagopt(mp->mnt_optnew, "ro", NULL, 0) && 4054f207061SMaxim Sobolev !(VFS_TO_TMPFS(mp)->tm_ronly)) { 406ac1a10efSMaxim Sobolev /* RW -> RO */ 407*5c4ce6faSKonstantin Belousov return (tmpfs_rw_to_ro(mp)); 408ac1a10efSMaxim Sobolev } else if (!vfs_flagopt(mp->mnt_optnew, "ro", NULL, 0) && 4094f207061SMaxim Sobolev VFS_TO_TMPFS(mp)->tm_ronly) { 410ac1a10efSMaxim Sobolev /* RO -> RW */ 4114f207061SMaxim Sobolev VFS_TO_TMPFS(mp)->tm_ronly = 0; 412ac1a10efSMaxim Sobolev MNT_ILOCK(mp); 413ac1a10efSMaxim Sobolev mp->mnt_flag &= ~MNT_RDONLY; 414ac1a10efSMaxim Sobolev MNT_IUNLOCK(mp); 415ac1a10efSMaxim Sobolev } 416c5ab5ce3SJaakko Heinonen return (0); 417d1fa59e9SXin LI } 418d1fa59e9SXin LI 419cb05b60aSAttilio Rao vn_lock(mp->mnt_vnodecovered, LK_SHARED | LK_RETRY); 4200359a12eSAttilio Rao error = VOP_GETATTR(mp->mnt_vnodecovered, &va, mp->mnt_cred); 42122db15c0SAttilio Rao VOP_UNLOCK(mp->mnt_vnodecovered, 0); 4221df86a32SXin LI if (error) 4231df86a32SXin LI return (error); 4241df86a32SXin LI 4251df86a32SXin LI if (mp->mnt_cred->cr_ruid != 0 || 4261df86a32SXin LI vfs_scanopt(mp->mnt_optnew, "gid", "%d", &root_gid) != 1) 4271df86a32SXin LI root_gid = va.va_gid; 4281df86a32SXin LI if (mp->mnt_cred->cr_ruid != 0 || 4291df86a32SXin LI vfs_scanopt(mp->mnt_optnew, "uid", "%d", &root_uid) != 1) 4301df86a32SXin LI root_uid = va.va_uid; 4311df86a32SXin LI if (mp->mnt_cred->cr_ruid != 0 || 432eed4ee29SXin LI vfs_scanopt(mp->mnt_optnew, "mode", "%ho", &root_mode) != 1) 4331df86a32SXin LI root_mode = va.va_mode; 4340ff93c48SGleb Kurtsou if (vfs_getopt_size(mp->mnt_optnew, "inodes", &nodes_max) != 0) 4351df86a32SXin LI nodes_max = 0; 4360ff93c48SGleb Kurtsou if (vfs_getopt_size(mp->mnt_optnew, "size", &size_max) != 0) 4371df86a32SXin LI size_max = 0; 4380ff93c48SGleb Kurtsou if (vfs_getopt_size(mp->mnt_optnew, "maxfilesize", &maxfilesize) != 0) 439dec3772eSJaakko Heinonen maxfilesize = 0; 44000ac6a98SKonstantin Belousov nonc = vfs_getopt(mp->mnt_optnew, "nonc", NULL, NULL) == 0; 441d1fa59e9SXin LI 442d1fa59e9SXin LI /* Do not allow mounts if we do not have enough memory to preserve 443d1fa59e9SXin LI * the minimum reserved pages. */ 444da7aa277SGleb Kurtsou if (tmpfs_mem_avail() < TMPFS_PAGES_MINRESERVED) 445bba7ed20SKonstantin Belousov return (ENOSPC); 446d1fa59e9SXin LI 447d1fa59e9SXin LI /* Get the maximum number of memory pages this file system is 448d1fa59e9SXin LI * allowed to use, based on the maximum size the user passed in 449d1fa59e9SXin LI * the mount structure. A value of zero is treated as if the 450d1fa59e9SXin LI * maximum available space was requested. */ 4510742ebc9SBryan Drewery if (size_max == 0 || size_max > OFF_MAX - PAGE_SIZE || 4520ff93c48SGleb Kurtsou (SIZE_MAX < OFF_MAX && size_max / PAGE_SIZE >= SIZE_MAX)) 453d1fa59e9SXin LI pages = SIZE_MAX; 4540742ebc9SBryan Drewery else { 4550742ebc9SBryan Drewery size_max = roundup(size_max, PAGE_SIZE); 4561df86a32SXin LI pages = howmany(size_max, PAGE_SIZE); 4570742ebc9SBryan Drewery } 458d1fa59e9SXin LI MPASS(pages > 0); 459d1fa59e9SXin LI 460189ee6beSJaakko Heinonen if (nodes_max <= 3) { 4619295c628SGleb Kurtsou if (pages < INT_MAX / nodes_per_page) 4629295c628SGleb Kurtsou nodes_max = pages * nodes_per_page; 463d1fa59e9SXin LI else 4649295c628SGleb Kurtsou nodes_max = INT_MAX; 4650ff93c48SGleb Kurtsou } 4669295c628SGleb Kurtsou if (nodes_max > INT_MAX) 4679295c628SGleb Kurtsou nodes_max = INT_MAX; 4680ff93c48SGleb Kurtsou MPASS(nodes_max >= 3); 469d1fa59e9SXin LI 470d1fa59e9SXin LI /* Allocate the tmpfs mount structure and fill it. */ 471d1fa59e9SXin LI tmp = (struct tmpfs_mount *)malloc(sizeof(struct tmpfs_mount), 472d1fa59e9SXin LI M_TMPFSMNT, M_WAITOK | M_ZERO); 473d1fa59e9SXin LI 474280ffa5eSKonstantin Belousov mtx_init(&tmp->tm_allnode_lock, "tmpfs allnode lock", NULL, MTX_DEF); 4750ff93c48SGleb Kurtsou tmp->tm_nodes_max = nodes_max; 476d1fa59e9SXin LI tmp->tm_nodes_inuse = 0; 47764c25043SKonstantin Belousov tmp->tm_refcount = 1; 4780ff93c48SGleb Kurtsou tmp->tm_maxfilesize = maxfilesize > 0 ? maxfilesize : OFF_MAX; 479d1fa59e9SXin LI LIST_INIT(&tmp->tm_nodes_used); 480d1fa59e9SXin LI 481ac1a10efSMaxim Sobolev tmp->tm_size_max = size_max; 482d1fa59e9SXin LI tmp->tm_pages_max = pages; 483d1fa59e9SXin LI tmp->tm_pages_used = 0; 48430e0cf49SMateusz Guzik new_unrhdr64(&tmp->tm_ino_unr, 2); 485e0f51ae7SXin LI tmp->tm_dirent_pool = uma_zcreate("TMPFS dirent", 486bba7ed20SKonstantin Belousov sizeof(struct tmpfs_dirent), NULL, NULL, NULL, NULL, 487e0f51ae7SXin LI UMA_ALIGN_PTR, 0); 488e0f51ae7SXin LI tmp->tm_node_pool = uma_zcreate("TMPFS node", 489bba7ed20SKonstantin Belousov sizeof(struct tmpfs_node), tmpfs_node_ctor, tmpfs_node_dtor, 490bba7ed20SKonstantin Belousov tmpfs_node_init, tmpfs_node_fini, UMA_ALIGN_PTR, 0); 491c5ab5ce3SJaakko Heinonen tmp->tm_ronly = (mp->mnt_flag & MNT_RDONLY) != 0; 49200ac6a98SKonstantin Belousov tmp->tm_nonc = nonc; 493d1fa59e9SXin LI 494d1fa59e9SXin LI /* Allocate the root node. */ 495bba7ed20SKonstantin Belousov error = tmpfs_alloc_node(mp, tmp, VDIR, root_uid, root_gid, 496bba7ed20SKonstantin Belousov root_mode & ALLPERMS, NULL, NULL, VNOVAL, &root); 497d1fa59e9SXin LI 498d1fa59e9SXin LI if (error != 0 || root == NULL) { 4997adb1776SXin LI uma_zdestroy(tmp->tm_node_pool); 5007adb1776SXin LI uma_zdestroy(tmp->tm_dirent_pool); 501d1fa59e9SXin LI free(tmp, M_TMPFSMNT); 502bba7ed20SKonstantin Belousov return (error); 503d1fa59e9SXin LI } 504fc8fdae0SMatthew D Fleming KASSERT(root->tn_id == 2, 505fc8fdae0SMatthew D Fleming ("tmpfs root with invalid ino: %ju", (uintmax_t)root->tn_id)); 506d1fa59e9SXin LI tmp->tm_root = root; 507d1fa59e9SXin LI 508d1fa59e9SXin LI MNT_ILOCK(mp); 509d1fa59e9SXin LI mp->mnt_flag |= MNT_LOCAL; 51031e73fd4SMateusz Guzik mp->mnt_kern_flag |= MNTK_LOOKUP_SHARED | MNTK_EXTENDED_SHARED; 511d1fa59e9SXin LI MNT_IUNLOCK(mp); 512d1fa59e9SXin LI 513d1fa59e9SXin LI mp->mnt_data = tmp; 514d1fa59e9SXin LI mp->mnt_stat.f_namemax = MAXNAMLEN; 515d1fa59e9SXin LI vfs_getnewfsid(mp); 516d1fa59e9SXin LI vfs_mountedfrom(mp, "tmpfs"); 517d1fa59e9SXin LI 518d1fa59e9SXin LI return 0; 519d1fa59e9SXin LI } 520d1fa59e9SXin LI 521d1fa59e9SXin LI /* ARGSUSED2 */ 522d1fa59e9SXin LI static int 523dfd233edSAttilio Rao tmpfs_unmount(struct mount *mp, int mntflags) 524d1fa59e9SXin LI { 525d1fa59e9SXin LI struct tmpfs_mount *tmp; 526d1fa59e9SXin LI struct tmpfs_node *node; 5274cda7f7eSKonstantin Belousov int error, flags; 528d1fa59e9SXin LI 5294cda7f7eSKonstantin Belousov flags = (mntflags & MNT_FORCE) != 0 ? FORCECLOSE : 0; 530d1fa59e9SXin LI tmp = VFS_TO_TMPFS(mp); 531d1fa59e9SXin LI 5324cda7f7eSKonstantin Belousov /* Stop writers */ 5334cda7f7eSKonstantin Belousov error = vfs_write_suspend_umnt(mp); 5344cda7f7eSKonstantin Belousov if (error != 0) 5354cda7f7eSKonstantin Belousov return (error); 5364cda7f7eSKonstantin Belousov /* 5374cda7f7eSKonstantin Belousov * At this point, nodes cannot be destroyed by any other 5384cda7f7eSKonstantin Belousov * thread because write suspension is started. 5394cda7f7eSKonstantin Belousov */ 540d1fa59e9SXin LI 5414cda7f7eSKonstantin Belousov for (;;) { 5424cda7f7eSKonstantin Belousov error = vflush(mp, 0, flags, curthread); 5434cda7f7eSKonstantin Belousov if (error != 0) { 5444cda7f7eSKonstantin Belousov vfs_write_resume(mp, VR_START_WRITE); 5454cda7f7eSKonstantin Belousov return (error); 5464cda7f7eSKonstantin Belousov } 5474cda7f7eSKonstantin Belousov MNT_ILOCK(mp); 5484cda7f7eSKonstantin Belousov if (mp->mnt_nvnodelistsize == 0) { 5494cda7f7eSKonstantin Belousov MNT_IUNLOCK(mp); 5504cda7f7eSKonstantin Belousov break; 5514cda7f7eSKonstantin Belousov } 5524cda7f7eSKonstantin Belousov MNT_IUNLOCK(mp); 5534cda7f7eSKonstantin Belousov if ((mntflags & MNT_FORCE) == 0) { 5544cda7f7eSKonstantin Belousov vfs_write_resume(mp, VR_START_WRITE); 5554cda7f7eSKonstantin Belousov return (EBUSY); 5564cda7f7eSKonstantin Belousov } 5574cda7f7eSKonstantin Belousov } 5584cda7f7eSKonstantin Belousov 5594cda7f7eSKonstantin Belousov TMPFS_LOCK(tmp); 5604cda7f7eSKonstantin Belousov while ((node = LIST_FIRST(&tmp->tm_nodes_used)) != NULL) { 56164c25043SKonstantin Belousov TMPFS_NODE_LOCK(node); 5624fd5efe7SGleb Kurtsou if (node->tn_type == VDIR) 5634fd5efe7SGleb Kurtsou tmpfs_dir_destroy(tmp, node); 56464c25043SKonstantin Belousov if (tmpfs_free_node_locked(tmp, node, true)) 5654cda7f7eSKonstantin Belousov TMPFS_LOCK(tmp); 56664c25043SKonstantin Belousov else 56764c25043SKonstantin Belousov TMPFS_NODE_UNLOCK(node); 56864c25043SKonstantin Belousov } 56964c25043SKonstantin Belousov 57064c25043SKonstantin Belousov mp->mnt_data = NULL; 57164c25043SKonstantin Belousov tmpfs_free_tmp(tmp); 57264c25043SKonstantin Belousov vfs_write_resume(mp, VR_START_WRITE); 57364c25043SKonstantin Belousov 57464c25043SKonstantin Belousov MNT_ILOCK(mp); 57564c25043SKonstantin Belousov mp->mnt_flag &= ~MNT_LOCAL; 57664c25043SKonstantin Belousov MNT_IUNLOCK(mp); 57764c25043SKonstantin Belousov 57864c25043SKonstantin Belousov return (0); 57964c25043SKonstantin Belousov } 58064c25043SKonstantin Belousov 58164c25043SKonstantin Belousov void 58264c25043SKonstantin Belousov tmpfs_free_tmp(struct tmpfs_mount *tmp) 58364c25043SKonstantin Belousov { 58464c25043SKonstantin Belousov 58564c25043SKonstantin Belousov MPASS(tmp->tm_refcount > 0); 58664c25043SKonstantin Belousov tmp->tm_refcount--; 58764c25043SKonstantin Belousov if (tmp->tm_refcount > 0) { 58864c25043SKonstantin Belousov TMPFS_UNLOCK(tmp); 58964c25043SKonstantin Belousov return; 590d1fa59e9SXin LI } 5914cda7f7eSKonstantin Belousov TMPFS_UNLOCK(tmp); 592d1fa59e9SXin LI 5937adb1776SXin LI uma_zdestroy(tmp->tm_dirent_pool); 5947adb1776SXin LI uma_zdestroy(tmp->tm_node_pool); 595d1fa59e9SXin LI 596280ffa5eSKonstantin Belousov mtx_destroy(&tmp->tm_allnode_lock); 597d1fa59e9SXin LI MPASS(tmp->tm_pages_used == 0); 5981df86a32SXin LI MPASS(tmp->tm_nodes_inuse == 0); 599d1fa59e9SXin LI 60064c25043SKonstantin Belousov free(tmp, M_TMPFSMNT); 601d1fa59e9SXin LI } 602d1fa59e9SXin LI 603d1fa59e9SXin LI static int 604dfd233edSAttilio Rao tmpfs_root(struct mount *mp, int flags, struct vnode **vpp) 605d1fa59e9SXin LI { 606d1fa59e9SXin LI int error; 607bba7ed20SKonstantin Belousov 608dfd233edSAttilio Rao error = tmpfs_alloc_vp(mp, VFS_TO_TMPFS(mp)->tm_root, flags, vpp); 609bba7ed20SKonstantin Belousov if (error == 0) 6107adb1776SXin LI (*vpp)->v_vflag |= VV_ROOT; 611bba7ed20SKonstantin Belousov return (error); 612d1fa59e9SXin LI } 613d1fa59e9SXin LI 614d1fa59e9SXin LI static int 615694a586aSRick Macklem tmpfs_fhtovp(struct mount *mp, struct fid *fhp, int flags, 616694a586aSRick Macklem struct vnode **vpp) 617d1fa59e9SXin LI { 618d1fa59e9SXin LI struct tmpfs_fid *tfhp; 619d1fa59e9SXin LI struct tmpfs_mount *tmp; 620d1fa59e9SXin LI struct tmpfs_node *node; 62164c25043SKonstantin Belousov int error; 622d1fa59e9SXin LI 623d1fa59e9SXin LI tmp = VFS_TO_TMPFS(mp); 624d1fa59e9SXin LI 625d1fa59e9SXin LI tfhp = (struct tmpfs_fid *)fhp; 626d1fa59e9SXin LI if (tfhp->tf_len != sizeof(struct tmpfs_fid)) 62764c25043SKonstantin Belousov return (EINVAL); 628d1fa59e9SXin LI 629d1fa59e9SXin LI if (tfhp->tf_id >= tmp->tm_nodes_max) 63064c25043SKonstantin Belousov return (EINVAL); 631d1fa59e9SXin LI 632d1fa59e9SXin LI TMPFS_LOCK(tmp); 633d1fa59e9SXin LI LIST_FOREACH(node, &tmp->tm_nodes_used, tn_entries) { 634d1fa59e9SXin LI if (node->tn_id == tfhp->tf_id && 635d1fa59e9SXin LI node->tn_gen == tfhp->tf_gen) { 63664c25043SKonstantin Belousov tmpfs_ref_node(node); 637d1fa59e9SXin LI break; 638d1fa59e9SXin LI } 639d1fa59e9SXin LI } 640d1fa59e9SXin LI TMPFS_UNLOCK(tmp); 641d1fa59e9SXin LI 64264c25043SKonstantin Belousov if (node != NULL) { 64364c25043SKonstantin Belousov error = tmpfs_alloc_vp(mp, node, LK_EXCLUSIVE, vpp); 64464c25043SKonstantin Belousov tmpfs_free_node(tmp, node); 64564c25043SKonstantin Belousov } else 64664c25043SKonstantin Belousov error = EINVAL; 64764c25043SKonstantin Belousov return (error); 648d1fa59e9SXin LI } 649d1fa59e9SXin LI 650d1fa59e9SXin LI /* ARGSUSED2 */ 651d1fa59e9SXin LI static int 652dfd233edSAttilio Rao tmpfs_statfs(struct mount *mp, struct statfs *sbp) 653d1fa59e9SXin LI { 654d1fa59e9SXin LI struct tmpfs_mount *tmp; 655da7aa277SGleb Kurtsou size_t used; 656d1fa59e9SXin LI 657d1fa59e9SXin LI tmp = VFS_TO_TMPFS(mp); 658d1fa59e9SXin LI 659d1fa59e9SXin LI sbp->f_iosize = PAGE_SIZE; 660d1fa59e9SXin LI sbp->f_bsize = PAGE_SIZE; 661d1fa59e9SXin LI 662da7aa277SGleb Kurtsou used = tmpfs_pages_used(tmp); 663ed2159c9SMateusz Guzik if (tmp->tm_pages_max != ULONG_MAX) 664da7aa277SGleb Kurtsou sbp->f_blocks = tmp->tm_pages_max; 665da7aa277SGleb Kurtsou else 666da7aa277SGleb Kurtsou sbp->f_blocks = used + tmpfs_mem_avail(); 667da7aa277SGleb Kurtsou if (sbp->f_blocks <= used) 668da7aa277SGleb Kurtsou sbp->f_bavail = 0; 669da7aa277SGleb Kurtsou else 670da7aa277SGleb Kurtsou sbp->f_bavail = sbp->f_blocks - used; 671da7aa277SGleb Kurtsou sbp->f_bfree = sbp->f_bavail; 672da7aa277SGleb Kurtsou used = tmp->tm_nodes_inuse; 673da7aa277SGleb Kurtsou sbp->f_files = tmp->tm_nodes_max; 674da7aa277SGleb Kurtsou if (sbp->f_files <= used) 675da7aa277SGleb Kurtsou sbp->f_ffree = 0; 676da7aa277SGleb Kurtsou else 677da7aa277SGleb Kurtsou sbp->f_ffree = sbp->f_files - used; 678d1fa59e9SXin LI /* sbp->f_owner = tmp->tn_uid; */ 679d1fa59e9SXin LI 680d1fa59e9SXin LI return 0; 681d1fa59e9SXin LI } 682d1fa59e9SXin LI 6834cda7f7eSKonstantin Belousov static int 6844cda7f7eSKonstantin Belousov tmpfs_sync(struct mount *mp, int waitfor) 6854cda7f7eSKonstantin Belousov { 6864cda7f7eSKonstantin Belousov 6874cda7f7eSKonstantin Belousov if (waitfor == MNT_SUSPEND) { 6884cda7f7eSKonstantin Belousov MNT_ILOCK(mp); 6894cda7f7eSKonstantin Belousov mp->mnt_kern_flag |= MNTK_SUSPEND2 | MNTK_SUSPENDED; 6904cda7f7eSKonstantin Belousov MNT_IUNLOCK(mp); 691f40cb1c6SKonstantin Belousov } else if (waitfor == MNT_LAZY) { 692*5c4ce6faSKonstantin Belousov tmpfs_update_mtime(mp, true); 6934cda7f7eSKonstantin Belousov } 6944cda7f7eSKonstantin Belousov return (0); 6954cda7f7eSKonstantin Belousov } 6964cda7f7eSKonstantin Belousov 697d1fa59e9SXin LI /* 6985f34e93cSMark Johnston * The presence of a susp_clean method tells the VFS to track writes. 6995f34e93cSMark Johnston */ 7005f34e93cSMark Johnston static void 7015f34e93cSMark Johnston tmpfs_susp_clean(struct mount *mp __unused) 7025f34e93cSMark Johnston { 7035f34e93cSMark Johnston } 7045f34e93cSMark Johnston 7055f34e93cSMark Johnston /* 706d1fa59e9SXin LI * tmpfs vfs operations. 707d1fa59e9SXin LI */ 708d1fa59e9SXin LI 709d1fa59e9SXin LI struct vfsops tmpfs_vfsops = { 710d1fa59e9SXin LI .vfs_mount = tmpfs_mount, 711d1fa59e9SXin LI .vfs_unmount = tmpfs_unmount, 712d1fa59e9SXin LI .vfs_root = tmpfs_root, 713d1fa59e9SXin LI .vfs_statfs = tmpfs_statfs, 714d1fa59e9SXin LI .vfs_fhtovp = tmpfs_fhtovp, 7154cda7f7eSKonstantin Belousov .vfs_sync = tmpfs_sync, 7165f34e93cSMark Johnston .vfs_susp_clean = tmpfs_susp_clean, 717d1fa59e9SXin LI }; 7182454886eSXin LI VFS_SET(tmpfs_vfsops, tmpfs, VFCF_JAIL); 719