xref: /freebsd/sys/kern/vfs_cache.c (revision 952d112864d8008aa87278a30a539d888a8493cd)
1 /*
2  * Copyright (c) 1989, 1993, 1995
3  *	The Regents of the University of California.  All rights reserved.
4  *
5  * This code is derived from software contributed to Berkeley by
6  * Poul-Henning Kamp of the FreeBSD Project.
7  *
8  * Redistribution and use in source and binary forms, with or without
9  * modification, are permitted provided that the following conditions
10  * are met:
11  * 1. Redistributions of source code must retain the above copyright
12  *    notice, this list of conditions and the following disclaimer.
13  * 2. Redistributions in binary form must reproduce the above copyright
14  *    notice, this list of conditions and the following disclaimer in the
15  *    documentation and/or other materials provided with the distribution.
16  * 3. All advertising materials mentioning features or use of this software
17  *    must display the following acknowledgement:
18  *	This product includes software developed by the University of
19  *	California, Berkeley and its contributors.
20  * 4. Neither the name of the University nor the names of its contributors
21  *    may be used to endorse or promote products derived from this software
22  *    without specific prior written permission.
23  *
24  * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
25  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
26  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
27  * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
28  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
29  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
30  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
31  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
32  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
33  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
34  * SUCH DAMAGE.
35  *
36  *	@(#)vfs_cache.c	8.5 (Berkeley) 3/22/95
37  * $Id: vfs_cache.c,v 1.23 1997/02/22 09:39:31 peter Exp $
38  */
39 
40 #include <sys/param.h>
41 #include <sys/systm.h>
42 #include <sys/kernel.h>
43 #include <sys/sysctl.h>
44 #include <sys/time.h>
45 #include <sys/mount.h>
46 #include <sys/vnode.h>
47 #include <sys/namei.h>
48 #include <sys/errno.h>
49 #include <sys/malloc.h>
50 
51 #define MAXVNODEUSE 32
52 
53 /*
54  * Name caching works as follows:
55  *
56  * Names found by directory scans are retained in a cache
57  * for future reference.  It is managed LRU, so frequently
58  * used names will hang around.  Cache is indexed by hash value
59  * obtained from (vp, name) where vp refers to the directory
60  * containing name.
61  *
62  * If it is a "negative" entry, (i.e. for a name that is known NOT to
63  * exist) the vnode pointer will be NULL.
64  *
65  * For simplicity (and economy of storage), names longer than
66  * a maximum length of NCHNAMLEN are not cached; they occur
67  * infrequently in any case, and are almost never of interest.
68  *
69  * Upon reaching the last segment of a path, if the reference
70  * is for DELETE, or NOCACHE is set (rewrite), and the
71  * name is located in the cache, it will be dropped.
72  */
73 
74 /*
75  * Structures associated with name cacheing.
76  */
77 #define NCHHASH(dvp, cnp) \
78 	(&nchashtbl[((dvp)->v_id + (cnp)->cn_hash) % nchash])
79 static LIST_HEAD(nchashhead, namecache) *nchashtbl;	/* Hash Table */
80 static u_long	nchash;			/* size of hash table */
81 static u_long	numcache;		/* number of cache entries allocated */
82 static TAILQ_HEAD(, namecache) nclruhead;	/* LRU chain */
83 struct	nchstats nchstats;		/* cache effectiveness statistics */
84 
85 static int	doingcache = 1;		/* 1 => enable the cache */
86 SYSCTL_INT(_debug, OID_AUTO, vfscache, CTLFLAG_RW, &doingcache, 0, "");
87 
88 #ifdef NCH_STATISTICS
89 u_long	nchnbr;
90 #define NCHNBR(ncp) (ncp)->nc_nbr = ++nchnbr;
91 #define NCHHIT(ncp) (ncp)->nc_hits++
92 #else
93 #define NCHNBR(ncp)
94 #define NCHHIT(ncp)
95 #endif
96 
97 /*
98  * Delete an entry from its hash list and move it to the front
99  * of the LRU list for immediate reuse.
100  */
101 #define PURGE(ncp)  {						\
102 	LIST_REMOVE(ncp, nc_hash);				\
103 	ncp->nc_hash.le_prev = 0;				\
104 	TAILQ_REMOVE(&nclruhead, ncp, nc_lru);			\
105 	TAILQ_INSERT_HEAD(&nclruhead, ncp, nc_lru);		\
106 }
107 
108 /*
109  * Move an entry that has been used to the tail of the LRU list
110  * so that it will be preserved for future use.
111  */
112 #define TOUCH(ncp)  {						\
113 	if (ncp->nc_lru.tqe_next != 0) {			\
114 		TAILQ_REMOVE(&nclruhead, ncp, nc_lru);		\
115 		TAILQ_INSERT_TAIL(&nclruhead, ncp, nc_lru);	\
116 		NCHNBR(ncp);					\
117 	}							\
118 }
119 
120 /*
121  * Lookup an entry in the cache
122  *
123  * We don't do this if the segment name is long, simply so the cache
124  * can avoid holding long names (which would either waste space, or
125  * add greatly to the complexity).
126  *
127  * Lookup is called with dvp pointing to the directory to search,
128  * cnp pointing to the name of the entry being sought. If the lookup
129  * succeeds, the vnode is returned in *vpp, and a status of -1 is
130  * returned. If the lookup determines that the name does not exist
131  * (negative cacheing), a status of ENOENT is returned. If the lookup
132  * fails, a status of zero is returned.
133  */
134 
135 int
136 cache_lookup(dvp, vpp, cnp)
137 	struct vnode *dvp;
138 	struct vnode **vpp;
139 	struct componentname *cnp;
140 {
141 	register struct namecache *ncp, *nnp;
142 	register struct nchashhead *ncpp;
143 
144 	if (!doingcache) {
145 		cnp->cn_flags &= ~MAKEENTRY;
146 		return (0);
147 	}
148 	if (cnp->cn_namelen > NCHNAMLEN) {
149 		nchstats.ncs_long++;
150 		cnp->cn_flags &= ~MAKEENTRY;
151 		return (0);
152 	}
153 
154 	ncpp = NCHHASH(dvp, cnp);
155 	for (ncp = ncpp->lh_first; ncp != 0; ncp = nnp) {
156 		nnp = ncp->nc_hash.le_next;
157 		/* If one of the vp's went stale, don't bother anymore. */
158 		if ((ncp->nc_dvpid != ncp->nc_dvp->v_id) ||
159 		    (ncp->nc_vp && ncp->nc_vpid != ncp->nc_vp->v_id)) {
160 			nchstats.ncs_falsehits++;
161 			PURGE(ncp);
162 			continue;
163 		}
164 		/* Now that we know the vp's to be valid, is it ours ? */
165 		if (ncp->nc_dvp == dvp &&
166 		    ncp->nc_nlen == cnp->cn_namelen &&
167 		    !bcmp(ncp->nc_name, cnp->cn_nameptr, (u_int)ncp->nc_nlen))
168 			break;
169 	}
170 
171 	/* We failed to find an entry */
172 	if (ncp == 0) {
173 		nchstats.ncs_miss++;
174 		return (0);
175 	}
176 
177 	NCHHIT(ncp);
178 
179 	/* We don't want to have an entry, so dump it */
180 	if ((cnp->cn_flags & MAKEENTRY) == 0) {
181 		nchstats.ncs_badhits++;
182 		PURGE(ncp);
183 		return (0);
184 	}
185 
186 	/* We found a "positive" match, return the vnode */
187         if (ncp->nc_vp) {
188 		nchstats.ncs_goodhits++;
189 		TOUCH(ncp);
190 		*vpp = ncp->nc_vp;
191 		if ((*vpp)->v_usage < MAXVNODEUSE)
192 			(*vpp)->v_usage++;
193 		return (-1);
194 	}
195 
196 	/* We found a negative match, and want to create it, so purge */
197 	if (cnp->cn_nameiop == CREATE) {
198 		nchstats.ncs_badhits++;
199 		PURGE(ncp);
200 		return (0);
201 	}
202 
203 	/*
204 	 * We found a "negative" match, ENOENT notifies client of this match.
205 	 * The nc_vpid field records whether this is a whiteout.
206 	 */
207 	nchstats.ncs_neghits++;
208 	TOUCH(ncp);
209 	cnp->cn_flags |= ncp->nc_vpid;
210 	return (ENOENT);
211 }
212 
213 /*
214  * Add an entry to the cache.
215  */
216 void
217 cache_enter(dvp, vp, cnp)
218 	struct vnode *dvp;
219 	struct vnode *vp;
220 	struct componentname *cnp;
221 {
222 	register struct namecache *ncp;
223 	register struct nchashhead *ncpp;
224 
225 	if (!doingcache)
226 		return;
227 
228 	if (cnp->cn_namelen > NCHNAMLEN) {
229 		printf("cache_enter: name too long");
230 		return;
231 	}
232 
233 	/*
234 	 * We allocate a new entry if we are less than the maximum
235 	 * allowed and the one at the front of the LRU list is in use.
236 	 * Otherwise we use the one at the front of the LRU list.
237 	 */
238 	if (numcache < desiredvnodes &&
239 	    ((ncp = nclruhead.tqh_first) == NULL ||
240 	    ncp->nc_hash.le_prev != 0)) {
241 		/* Add one more entry */
242 		ncp = (struct namecache *)
243 			malloc((u_long)sizeof *ncp, M_CACHE, M_WAITOK);
244 		bzero((char *)ncp, sizeof *ncp);
245 		numcache++;
246 	} else if (ncp = nclruhead.tqh_first) {
247 		/* reuse an old entry */
248 		TAILQ_REMOVE(&nclruhead, ncp, nc_lru);
249 		if (ncp->nc_hash.le_prev != 0) {
250 			LIST_REMOVE(ncp, nc_hash);
251 			ncp->nc_hash.le_prev = 0;
252 		}
253 	} else {
254 		/* give up */
255 		return;
256 	}
257 
258 	/*
259 	 * Fill in cache info, if vp is NULL this is a "negative" cache entry.
260 	 * For negative entries, we have to record whether it is a whiteout.
261 	 * the whiteout flag is stored in the nc_vpid field which is
262 	 * otherwise unused.
263 	 */
264 	ncp->nc_vp = vp;
265 	if (vp) {
266 		ncp->nc_vpid = vp->v_id;
267 		if (vp->v_usage < MAXVNODEUSE)
268 			++vp->v_usage;
269 	} else
270 		ncp->nc_vpid = cnp->cn_flags & ISWHITEOUT;
271 	ncp->nc_dvp = dvp;
272 	ncp->nc_dvpid = dvp->v_id;
273 	ncp->nc_nlen = cnp->cn_namelen;
274 	bcopy(cnp->cn_nameptr, ncp->nc_name, (unsigned)ncp->nc_nlen);
275 	TAILQ_INSERT_TAIL(&nclruhead, ncp, nc_lru);
276 	ncpp = NCHHASH(dvp, cnp);
277 	LIST_INSERT_HEAD(ncpp, ncp, nc_hash);
278 }
279 
280 /*
281  * Name cache initialization, from vfs_init() when we are booting
282  */
283 void
284 nchinit()
285 {
286 
287 	TAILQ_INIT(&nclruhead);
288 	nchashtbl = phashinit(desiredvnodes, M_CACHE, &nchash);
289 }
290 
291 /*
292  * Invalidate all entries to particular vnode.
293  *
294  * We actually just increment the v_id, that will do it. The stale entries
295  * will be purged by lookup as they get found. If the v_id wraps around, we
296  * need to ditch the entire cache, to avoid confusion. No valid vnode will
297  * ever have (v_id == 0).
298  */
299 void
300 cache_purge(vp)
301 	struct vnode *vp;
302 {
303 	struct namecache *ncp;
304 	struct nchashhead *ncpp;
305 	static u_long nextvnodeid;
306 
307 	vp->v_id = ++nextvnodeid;
308 	if (nextvnodeid != 0)
309 		return;
310 	for (ncpp = &nchashtbl[nchash - 1]; ncpp >= nchashtbl; ncpp--) {
311 		while (ncp = ncpp->lh_first)
312 			PURGE(ncp);
313 	}
314 	vp->v_id = ++nextvnodeid;
315 }
316 
317 /*
318  * Flush all entries referencing a particular filesystem.
319  *
320  * Since we need to check it anyway, we will flush all the invalid
321  * entries at the same time.
322  */
323 void
324 cache_purgevfs(mp)
325 	struct mount *mp;
326 {
327 	struct nchashhead *ncpp;
328 	struct namecache *ncp, *nnp;
329 
330 	/* Scan hash tables for applicable entries */
331 	for (ncpp = &nchashtbl[nchash - 1]; ncpp >= nchashtbl; ncpp--) {
332 		for (ncp = ncpp->lh_first; ncp != 0; ncp = nnp) {
333 			nnp = ncp->nc_hash.le_next;
334 			if (ncp->nc_dvpid != ncp->nc_dvp->v_id ||
335 			    (ncp->nc_vp && ncp->nc_vpid != ncp->nc_vp->v_id) ||
336 			    ncp->nc_dvp->v_mount == mp) {
337 				PURGE(ncp);
338 			}
339 		}
340 	}
341 }
342