xref: /freebsd/stand/common/bcache.c (revision 0d13f5343fafbf3067ffc33a507ffca0375c4417)
1ca987d46SWarner Losh /*-
2ca987d46SWarner Losh  * Copyright (c) 1998 Michael Smith <msmith@freebsd.org>
3ca987d46SWarner Losh  * Copyright 2015 Toomas Soome <tsoome@me.com>
4ca987d46SWarner Losh  * All rights reserved.
5ca987d46SWarner Losh  *
6ca987d46SWarner Losh  * Redistribution and use in source and binary forms, with or without
7ca987d46SWarner Losh  * modification, are permitted provided that the following conditions
8ca987d46SWarner Losh  * are met:
9ca987d46SWarner Losh  * 1. Redistributions of source code must retain the above copyright
10ca987d46SWarner Losh  *    notice, this list of conditions and the following disclaimer.
11ca987d46SWarner Losh  * 2. Redistributions in binary form must reproduce the above copyright
12ca987d46SWarner Losh  *    notice, this list of conditions and the following disclaimer in the
13ca987d46SWarner Losh  *    documentation and/or other materials provided with the distribution.
14ca987d46SWarner Losh  *
15ca987d46SWarner Losh  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
16ca987d46SWarner Losh  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
17ca987d46SWarner Losh  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
18ca987d46SWarner Losh  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
19ca987d46SWarner Losh  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
20ca987d46SWarner Losh  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
21ca987d46SWarner Losh  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
22ca987d46SWarner Losh  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
23ca987d46SWarner Losh  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
24ca987d46SWarner Losh  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
25ca987d46SWarner Losh  * SUCH DAMAGE.
26ca987d46SWarner Losh  */
27ca987d46SWarner Losh 
28ca987d46SWarner Losh #include <sys/cdefs.h>
29ca987d46SWarner Losh #include <sys/param.h>
30ca987d46SWarner Losh __FBSDID("$FreeBSD$");
31ca987d46SWarner Losh 
32ca987d46SWarner Losh /*
33ca987d46SWarner Losh  * Simple hashed block cache
34ca987d46SWarner Losh  */
35ca987d46SWarner Losh 
36ca987d46SWarner Losh #include <sys/stdint.h>
37ca987d46SWarner Losh 
38ca987d46SWarner Losh #include <stand.h>
39ca987d46SWarner Losh #include <string.h>
40ca987d46SWarner Losh #include <strings.h>
41ca987d46SWarner Losh 
42ca987d46SWarner Losh #include "bootstrap.h"
43ca987d46SWarner Losh 
44ca987d46SWarner Losh /* #define BCACHE_DEBUG */
45ca987d46SWarner Losh 
46ca987d46SWarner Losh #ifdef BCACHE_DEBUG
477325df02SKyle Evans # define DPRINTF(fmt, args...)	printf("%s: " fmt "\n" , __func__ , ## args)
48ca987d46SWarner Losh #else
4974772ef1SToomas Soome # define DPRINTF(fmt, args...)	((void)0)
50ca987d46SWarner Losh #endif
51ca987d46SWarner Losh 
52ca987d46SWarner Losh struct bcachectl
53ca987d46SWarner Losh {
54ca987d46SWarner Losh     daddr_t	bc_blkno;
55ca987d46SWarner Losh     int		bc_count;
56ca987d46SWarner Losh };
57ca987d46SWarner Losh 
58ca987d46SWarner Losh /*
59ca987d46SWarner Losh  * bcache per device node. cache is allocated on device first open and freed
60ca987d46SWarner Losh  * on last close, to save memory. The issue there is the size; biosdisk
61ca987d46SWarner Losh  * supports up to 31 (0x1f) devices. Classic setup would use single disk
62ca987d46SWarner Losh  * to boot from, but this has changed with zfs.
63ca987d46SWarner Losh  */
64ca987d46SWarner Losh struct bcache {
65ca987d46SWarner Losh     struct bcachectl	*bcache_ctl;
66ca987d46SWarner Losh     caddr_t		bcache_data;
67ca987d46SWarner Losh     size_t		bcache_nblks;
68ca987d46SWarner Losh     size_t		ra;
69*0d13f534SMaxim Sobolev     daddr_t		bcache_nextblkno;
70ca987d46SWarner Losh };
71ca987d46SWarner Losh 
72ca987d46SWarner Losh static u_int bcache_total_nblks;	/* set by bcache_init */
73ca987d46SWarner Losh static u_int bcache_blksize;		/* set by bcache_init */
74ca987d46SWarner Losh static u_int bcache_numdev;		/* set by bcache_add_dev */
75ca987d46SWarner Losh /* statistics */
76ca987d46SWarner Losh static u_int bcache_units;	/* number of devices with cache */
77ca987d46SWarner Losh static u_int bcache_unit_nblks;	/* nblocks per unit */
78ca987d46SWarner Losh static u_int bcache_hits;
79ca987d46SWarner Losh static u_int bcache_misses;
80ca987d46SWarner Losh static u_int bcache_ops;
81ca987d46SWarner Losh static u_int bcache_bypasses;
82ca987d46SWarner Losh static u_int bcache_bcount;
83ca987d46SWarner Losh static u_int bcache_rablks;
84ca987d46SWarner Losh 
85ca987d46SWarner Losh #define	BHASH(bc, blkno)	((blkno) & ((bc)->bcache_nblks - 1))
86ca987d46SWarner Losh #define	BCACHE_LOOKUP(bc, blkno)	\
87ca987d46SWarner Losh 	((bc)->bcache_ctl[BHASH((bc), (blkno))].bc_blkno != (blkno))
88ca987d46SWarner Losh #define	BCACHE_READAHEAD	256
89ca987d46SWarner Losh #define	BCACHE_MINREADAHEAD	32
90ca987d46SWarner Losh 
91ca987d46SWarner Losh static void	bcache_invalidate(struct bcache *bc, daddr_t blkno);
92ca987d46SWarner Losh static void	bcache_insert(struct bcache *bc, daddr_t blkno);
93ca987d46SWarner Losh static void	bcache_free_instance(struct bcache *bc);
94ca987d46SWarner Losh 
95ca987d46SWarner Losh /*
96ca987d46SWarner Losh  * Initialise the cache for (nblks) of (bsize).
97ca987d46SWarner Losh  */
98ca987d46SWarner Losh void
99ca987d46SWarner Losh bcache_init(size_t nblks, size_t bsize)
100ca987d46SWarner Losh {
101ca987d46SWarner Losh     /* set up control data */
102ca987d46SWarner Losh     bcache_total_nblks = nblks;
103ca987d46SWarner Losh     bcache_blksize = bsize;
104ca987d46SWarner Losh }
105ca987d46SWarner Losh 
106ca987d46SWarner Losh /*
107ca987d46SWarner Losh  * add number of devices to bcache. we have to divide cache space
108ca987d46SWarner Losh  * between the devices, so bcache_add_dev() can be used to set up the
109ca987d46SWarner Losh  * number. The issue is, we need to get the number before actual allocations.
110ca987d46SWarner Losh  * bcache_add_dev() is supposed to be called from device init() call, so the
111ca987d46SWarner Losh  * assumption is, devsw dv_init is called for plain devices first, and
112ca987d46SWarner Losh  * for zfs, last.
113ca987d46SWarner Losh  */
114ca987d46SWarner Losh void
115ca987d46SWarner Losh bcache_add_dev(int devices)
116ca987d46SWarner Losh {
117ca987d46SWarner Losh     bcache_numdev += devices;
118ca987d46SWarner Losh }
119ca987d46SWarner Losh 
120ca987d46SWarner Losh void *
121ca987d46SWarner Losh bcache_allocate(void)
122ca987d46SWarner Losh {
123ca987d46SWarner Losh     u_int i;
124ca987d46SWarner Losh     struct bcache *bc = malloc(sizeof (struct bcache));
125ca987d46SWarner Losh     int disks = bcache_numdev;
126ca987d46SWarner Losh 
127ca987d46SWarner Losh     if (disks == 0)
128ca987d46SWarner Losh 	disks = 1;	/* safe guard */
129ca987d46SWarner Losh 
130ca987d46SWarner Losh     if (bc == NULL) {
131ca987d46SWarner Losh 	errno = ENOMEM;
132ca987d46SWarner Losh 	return (bc);
133ca987d46SWarner Losh     }
134ca987d46SWarner Losh 
135ca987d46SWarner Losh     /*
136ca987d46SWarner Losh      * the bcache block count must be power of 2 for hash function
137ca987d46SWarner Losh      */
138ca987d46SWarner Losh     i = fls(disks) - 1;		/* highbit - 1 */
139ca987d46SWarner Losh     if (disks > (1 << i))	/* next power of 2 */
140ca987d46SWarner Losh 	i++;
141ca987d46SWarner Losh 
142ca987d46SWarner Losh     bc->bcache_nblks = bcache_total_nblks >> i;
143ca987d46SWarner Losh     bcache_unit_nblks = bc->bcache_nblks;
1446c81fe16SMaxim Sobolev     bc->bcache_data = malloc(bc->bcache_nblks * bcache_blksize);
145ca987d46SWarner Losh     if (bc->bcache_data == NULL) {
146ca987d46SWarner Losh 	/* dont error out yet. fall back to 32 blocks and try again */
147ca987d46SWarner Losh 	bc->bcache_nblks = 32;
148ca987d46SWarner Losh 	bc->bcache_data = malloc(bc->bcache_nblks * bcache_blksize +
149ca987d46SWarner Losh 	sizeof(uint32_t));
150ca987d46SWarner Losh     }
151ca987d46SWarner Losh 
152ca987d46SWarner Losh     bc->bcache_ctl = malloc(bc->bcache_nblks * sizeof(struct bcachectl));
153ca987d46SWarner Losh 
154ca987d46SWarner Losh     if ((bc->bcache_data == NULL) || (bc->bcache_ctl == NULL)) {
155ca987d46SWarner Losh 	bcache_free_instance(bc);
156ca987d46SWarner Losh 	errno = ENOMEM;
157ca987d46SWarner Losh 	return (NULL);
158ca987d46SWarner Losh     }
159ca987d46SWarner Losh 
160ca987d46SWarner Losh     /* Flush the cache */
161ca987d46SWarner Losh     for (i = 0; i < bc->bcache_nblks; i++) {
162ca987d46SWarner Losh 	bc->bcache_ctl[i].bc_count = -1;
163ca987d46SWarner Losh 	bc->bcache_ctl[i].bc_blkno = -1;
164ca987d46SWarner Losh     }
165ca987d46SWarner Losh     bcache_units++;
166ca987d46SWarner Losh     bc->ra = BCACHE_READAHEAD;	/* optimistic read ahead */
167*0d13f534SMaxim Sobolev     bc->bcache_nextblkno = -1;
168ca987d46SWarner Losh     return (bc);
169ca987d46SWarner Losh }
170ca987d46SWarner Losh 
171ca987d46SWarner Losh void
172ca987d46SWarner Losh bcache_free(void *cache)
173ca987d46SWarner Losh {
174ca987d46SWarner Losh     struct bcache *bc = cache;
175ca987d46SWarner Losh 
176ca987d46SWarner Losh     if (bc == NULL)
177ca987d46SWarner Losh 	return;
178ca987d46SWarner Losh 
179ca987d46SWarner Losh     bcache_free_instance(bc);
180ca987d46SWarner Losh     bcache_units--;
181ca987d46SWarner Losh }
182ca987d46SWarner Losh 
183ca987d46SWarner Losh /*
184ca987d46SWarner Losh  * Handle a write request; write directly to the disk, and populate the
185ca987d46SWarner Losh  * cache with the new values.
186ca987d46SWarner Losh  */
187ca987d46SWarner Losh static int
188ca987d46SWarner Losh write_strategy(void *devdata, int rw, daddr_t blk, size_t size,
189ca987d46SWarner Losh     char *buf, size_t *rsize)
190ca987d46SWarner Losh {
191ca987d46SWarner Losh     struct bcache_devdata	*dd = (struct bcache_devdata *)devdata;
192ca987d46SWarner Losh     struct bcache		*bc = dd->dv_cache;
193ca987d46SWarner Losh     daddr_t			i, nblk;
194ca987d46SWarner Losh 
195ca987d46SWarner Losh     nblk = size / bcache_blksize;
196ca987d46SWarner Losh 
197ca987d46SWarner Losh     /* Invalidate the blocks being written */
198ca987d46SWarner Losh     for (i = 0; i < nblk; i++) {
199ca987d46SWarner Losh 	bcache_invalidate(bc, blk + i);
200ca987d46SWarner Losh     }
201ca987d46SWarner Losh 
202ca987d46SWarner Losh     /* Write the blocks */
203ca987d46SWarner Losh     return (dd->dv_strategy(dd->dv_devdata, rw, blk, size, buf, rsize));
204ca987d46SWarner Losh }
205ca987d46SWarner Losh 
206ca987d46SWarner Losh /*
207ca987d46SWarner Losh  * Handle a read request; fill in parts of the request that can
208ca987d46SWarner Losh  * be satisfied by the cache, use the supplied strategy routine to do
209ca987d46SWarner Losh  * device I/O and then use the I/O results to populate the cache.
210ca987d46SWarner Losh  */
211ca987d46SWarner Losh static int
212ca987d46SWarner Losh read_strategy(void *devdata, int rw, daddr_t blk, size_t size,
213ca987d46SWarner Losh     char *buf, size_t *rsize)
214ca987d46SWarner Losh {
215ca987d46SWarner Losh     struct bcache_devdata	*dd = (struct bcache_devdata *)devdata;
216ca987d46SWarner Losh     struct bcache		*bc = dd->dv_cache;
217ca987d46SWarner Losh     size_t			i, nblk, p_size, r_size, complete, ra;
218ca987d46SWarner Losh     int				result;
219ca987d46SWarner Losh     daddr_t			p_blk;
220ca987d46SWarner Losh     caddr_t			p_buf;
221ca987d46SWarner Losh 
222ca987d46SWarner Losh     if (bc == NULL) {
223ca987d46SWarner Losh 	errno = ENODEV;
224ca987d46SWarner Losh 	return (-1);
225ca987d46SWarner Losh     }
226ca987d46SWarner Losh 
227ca987d46SWarner Losh     if (rsize != NULL)
228ca987d46SWarner Losh 	*rsize = 0;
229ca987d46SWarner Losh 
230ca987d46SWarner Losh     nblk = size / bcache_blksize;
231ca987d46SWarner Losh     if (nblk == 0 && size != 0)
232ca987d46SWarner Losh 	nblk++;
233ca987d46SWarner Losh     result = 0;
234ca987d46SWarner Losh     complete = 1;
235ca987d46SWarner Losh 
236ca987d46SWarner Losh     /* Satisfy any cache hits up front, break on first miss */
237ca987d46SWarner Losh     for (i = 0; i < nblk; i++) {
238ca987d46SWarner Losh 	if (BCACHE_LOOKUP(bc, (daddr_t)(blk + i))) {
239ca987d46SWarner Losh 	    bcache_misses += (nblk - i);
240ca987d46SWarner Losh 	    complete = 0;
241ca987d46SWarner Losh 	    if (nblk - i > BCACHE_MINREADAHEAD && bc->ra > BCACHE_MINREADAHEAD)
242ca987d46SWarner Losh 		bc->ra >>= 1;	/* reduce read ahead */
243ca987d46SWarner Losh 	    break;
244ca987d46SWarner Losh 	} else {
245ca987d46SWarner Losh 	    bcache_hits++;
246ca987d46SWarner Losh 	}
247ca987d46SWarner Losh     }
248ca987d46SWarner Losh 
249ca987d46SWarner Losh    if (complete) {	/* whole set was in cache, return it */
250ca987d46SWarner Losh 	if (bc->ra < BCACHE_READAHEAD)
251ca987d46SWarner Losh 		bc->ra <<= 1;	/* increase read ahead */
252ca987d46SWarner Losh 	bcopy(bc->bcache_data + (bcache_blksize * BHASH(bc, blk)), buf, size);
253ca987d46SWarner Losh 	goto done;
254ca987d46SWarner Losh    }
255ca987d46SWarner Losh 
256ca987d46SWarner Losh     /*
257ca987d46SWarner Losh      * Fill in any misses. From check we have i pointing to first missing
258ca987d46SWarner Losh      * block, read in all remaining blocks + readahead.
259ca987d46SWarner Losh      * We have space at least for nblk - i before bcache wraps.
260ca987d46SWarner Losh      */
261ca987d46SWarner Losh     p_blk = blk + i;
262ca987d46SWarner Losh     p_buf = bc->bcache_data + (bcache_blksize * BHASH(bc, p_blk));
263ca987d46SWarner Losh     r_size = bc->bcache_nblks - BHASH(bc, p_blk); /* remaining blocks */
264ca987d46SWarner Losh 
265ca987d46SWarner Losh     p_size = MIN(r_size, nblk - i);	/* read at least those blocks */
266ca987d46SWarner Losh 
267ca987d46SWarner Losh     /*
268ca987d46SWarner Losh      * The read ahead size setup.
269ca987d46SWarner Losh      * While the read ahead can save us IO, it also can complicate things:
270ca987d46SWarner Losh      * 1. We do not want to read ahead by wrapping around the
271ca987d46SWarner Losh      * bcache end - this would complicate the cache management.
272ca987d46SWarner Losh      * 2. We are using bc->ra as dynamic hint for read ahead size,
273ca987d46SWarner Losh      * detected cache hits will increase the read-ahead block count, and
274ca987d46SWarner Losh      * misses will decrease, see the code above.
275ca987d46SWarner Losh      * 3. The bcache is sized by 512B blocks, however, the underlying device
276ca987d46SWarner Losh      * may have a larger sector size, and we should perform the IO by
277ca987d46SWarner Losh      * taking into account these larger sector sizes. We could solve this by
278ca987d46SWarner Losh      * passing the sector size to bcache_allocate(), or by using ioctl(), but
279ca987d46SWarner Losh      * in this version we are using the constant, 16 blocks, and are rounding
280ca987d46SWarner Losh      * read ahead block count down to multiple of 16.
281ca987d46SWarner Losh      * Using the constant has two reasons, we are not entirely sure if the
282ca987d46SWarner Losh      * BIOS disk interface is providing the correct value for sector size.
283ca987d46SWarner Losh      * And secondly, this way we get the most conservative setup for the ra.
284ca987d46SWarner Losh      *
285ca987d46SWarner Losh      * The selection of multiple of 16 blocks (8KB) is quite arbitrary, however,
286ca987d46SWarner Losh      * we want to cover CDs (2K) and 4K disks.
287ca987d46SWarner Losh      * bcache_allocate() will always fall back to a minimum of 32 blocks.
288ca987d46SWarner Losh      * Our choice of 16 read ahead blocks will always fit inside the bcache.
289ca987d46SWarner Losh      */
290ca987d46SWarner Losh 
291ca987d46SWarner Losh     if ((rw & F_NORA) == F_NORA)
292ca987d46SWarner Losh 	ra = 0;
293ca987d46SWarner Losh     else
294ca987d46SWarner Losh 	ra = bc->bcache_nblks - BHASH(bc, p_blk + p_size);
295ca987d46SWarner Losh 
296*0d13f534SMaxim Sobolev     /*
297*0d13f534SMaxim Sobolev      * Only trigger read-ahead if we detect two blocks being read
298*0d13f534SMaxim Sobolev      * sequentially.
299*0d13f534SMaxim Sobolev      */
300*0d13f534SMaxim Sobolev     if ((bc->bcache_nextblkno != blk) && ra != 0) {
301*0d13f534SMaxim Sobolev         ra = 0;
302*0d13f534SMaxim Sobolev     }
303*0d13f534SMaxim Sobolev 
304ca987d46SWarner Losh     if (ra != 0 && ra != bc->bcache_nblks) { /* do we have RA space? */
305ca987d46SWarner Losh 	ra = MIN(bc->ra, ra - 1);
306ca987d46SWarner Losh 	ra = rounddown(ra, 16);		/* multiple of 16 blocks */
307ca987d46SWarner Losh 	p_size += ra;
308ca987d46SWarner Losh     }
309ca987d46SWarner Losh 
310ca987d46SWarner Losh     /* invalidate bcache */
311ca987d46SWarner Losh     for (i = 0; i < p_size; i++) {
312ca987d46SWarner Losh 	bcache_invalidate(bc, p_blk + i);
313ca987d46SWarner Losh     }
314ca987d46SWarner Losh 
315ca987d46SWarner Losh     r_size = 0;
316ca987d46SWarner Losh     /*
317ca987d46SWarner Losh      * with read-ahead, it may happen we are attempting to read past
318ca987d46SWarner Losh      * disk end, as bcache has no information about disk size.
319ca987d46SWarner Losh      * in such case we should get partial read if some blocks can be
320ca987d46SWarner Losh      * read or error, if no blocks can be read.
321ca987d46SWarner Losh      * in either case we should return the data in bcache and only
322ca987d46SWarner Losh      * return error if there is no data.
323ca987d46SWarner Losh      */
324ca987d46SWarner Losh     rw &= F_MASK;
325ca987d46SWarner Losh     result = dd->dv_strategy(dd->dv_devdata, rw, p_blk,
326ca987d46SWarner Losh 	p_size * bcache_blksize, p_buf, &r_size);
327ca987d46SWarner Losh 
328ca987d46SWarner Losh     r_size /= bcache_blksize;
329ca987d46SWarner Losh     for (i = 0; i < r_size; i++)
330ca987d46SWarner Losh 	bcache_insert(bc, p_blk + i);
331ca987d46SWarner Losh 
332ca987d46SWarner Losh     /* update ra statistics */
333ca987d46SWarner Losh     if (r_size != 0) {
334ca987d46SWarner Losh 	if (r_size < p_size)
335ca987d46SWarner Losh 	    bcache_rablks += (p_size - r_size);
336ca987d46SWarner Losh 	else
337ca987d46SWarner Losh 	    bcache_rablks += ra;
338ca987d46SWarner Losh     }
339ca987d46SWarner Losh 
340ca987d46SWarner Losh     /* check how much data can we copy */
341ca987d46SWarner Losh     for (i = 0; i < nblk; i++) {
342ca987d46SWarner Losh 	if (BCACHE_LOOKUP(bc, (daddr_t)(blk + i)))
343ca987d46SWarner Losh 	    break;
344ca987d46SWarner Losh     }
345ca987d46SWarner Losh 
346ca987d46SWarner Losh     if (size > i * bcache_blksize)
347ca987d46SWarner Losh 	size = i * bcache_blksize;
348ca987d46SWarner Losh 
349ca987d46SWarner Losh     if (size != 0) {
350ca987d46SWarner Losh 	bcopy(bc->bcache_data + (bcache_blksize * BHASH(bc, blk)), buf, size);
351ca987d46SWarner Losh 	result = 0;
352ca987d46SWarner Losh     }
353ca987d46SWarner Losh 
354ca987d46SWarner Losh  done:
355*0d13f534SMaxim Sobolev     if (result == 0) {
356*0d13f534SMaxim Sobolev         if (rsize != NULL)
357ca987d46SWarner Losh 	    *rsize = size;
358*0d13f534SMaxim Sobolev         bc->bcache_nextblkno = blk + (size / DEV_BSIZE);
359*0d13f534SMaxim Sobolev     }
360ca987d46SWarner Losh     return(result);
361ca987d46SWarner Losh }
362ca987d46SWarner Losh 
363ca987d46SWarner Losh /*
364ca987d46SWarner Losh  * Requests larger than 1/2 cache size will be bypassed and go
365ca987d46SWarner Losh  * directly to the disk.  XXX tune this.
366ca987d46SWarner Losh  */
367ca987d46SWarner Losh int
368ca987d46SWarner Losh bcache_strategy(void *devdata, int rw, daddr_t blk, size_t size,
369ca987d46SWarner Losh     char *buf, size_t *rsize)
370ca987d46SWarner Losh {
371ca987d46SWarner Losh     struct bcache_devdata	*dd = (struct bcache_devdata *)devdata;
372ca987d46SWarner Losh     struct bcache		*bc = dd->dv_cache;
373ca987d46SWarner Losh     u_int bcache_nblks = 0;
374ca987d46SWarner Losh     int nblk, cblk, ret;
375ca987d46SWarner Losh     size_t csize, isize, total;
376ca987d46SWarner Losh 
377ca987d46SWarner Losh     bcache_ops++;
378ca987d46SWarner Losh 
379ca987d46SWarner Losh     if (bc != NULL)
380ca987d46SWarner Losh 	bcache_nblks = bc->bcache_nblks;
381ca987d46SWarner Losh 
382ca987d46SWarner Losh     /* bypass large requests, or when the cache is inactive */
383ca987d46SWarner Losh     if (bc == NULL ||
384ca987d46SWarner Losh 	((size * 2 / bcache_blksize) > bcache_nblks)) {
3857325df02SKyle Evans 	DPRINTF("bypass %zu from %qu", size / bcache_blksize, blk);
386ca987d46SWarner Losh 	bcache_bypasses++;
387ca987d46SWarner Losh 	rw &= F_MASK;
388ca987d46SWarner Losh 	return (dd->dv_strategy(dd->dv_devdata, rw, blk, size, buf, rsize));
389ca987d46SWarner Losh     }
390ca987d46SWarner Losh 
391ca987d46SWarner Losh     switch (rw & F_MASK) {
392ca987d46SWarner Losh     case F_READ:
393ca987d46SWarner Losh 	nblk = size / bcache_blksize;
394ca987d46SWarner Losh 	if (size != 0 && nblk == 0)
395ca987d46SWarner Losh 	    nblk++;	/* read at least one block */
396ca987d46SWarner Losh 
397ca987d46SWarner Losh 	ret = 0;
398ca987d46SWarner Losh 	total = 0;
399ca987d46SWarner Losh 	while(size) {
400ca987d46SWarner Losh 	    cblk = bcache_nblks - BHASH(bc, blk); /* # of blocks left */
401ca987d46SWarner Losh 	    cblk = MIN(cblk, nblk);
402ca987d46SWarner Losh 
403ca987d46SWarner Losh 	    if (size <= bcache_blksize)
404ca987d46SWarner Losh 		csize = size;
405ca987d46SWarner Losh 	    else
406ca987d46SWarner Losh 		csize = cblk * bcache_blksize;
407ca987d46SWarner Losh 
408ca987d46SWarner Losh 	    ret = read_strategy(devdata, rw, blk, csize, buf+total, &isize);
409ca987d46SWarner Losh 
410ca987d46SWarner Losh 	    /*
411ca987d46SWarner Losh 	     * we may have error from read ahead, if we have read some data
412ca987d46SWarner Losh 	     * return partial read.
413ca987d46SWarner Losh 	     */
414ca987d46SWarner Losh 	    if (ret != 0 || isize == 0) {
415ca987d46SWarner Losh 		if (total != 0)
416ca987d46SWarner Losh 		    ret = 0;
417ca987d46SWarner Losh 		break;
418ca987d46SWarner Losh 	    }
419ca987d46SWarner Losh 	    blk += isize / bcache_blksize;
420ca987d46SWarner Losh 	    total += isize;
421ca987d46SWarner Losh 	    size -= isize;
422ca987d46SWarner Losh 	    nblk = size / bcache_blksize;
423ca987d46SWarner Losh 	}
424ca987d46SWarner Losh 
425ca987d46SWarner Losh 	if (rsize)
426ca987d46SWarner Losh 	    *rsize = total;
427ca987d46SWarner Losh 
428ca987d46SWarner Losh 	return (ret);
429ca987d46SWarner Losh     case F_WRITE:
430ca987d46SWarner Losh 	return write_strategy(devdata, F_WRITE, blk, size, buf, rsize);
431ca987d46SWarner Losh     }
432ca987d46SWarner Losh     return -1;
433ca987d46SWarner Losh }
434ca987d46SWarner Losh 
435ca987d46SWarner Losh /*
436ca987d46SWarner Losh  * Free allocated bcache instance
437ca987d46SWarner Losh  */
438ca987d46SWarner Losh static void
439ca987d46SWarner Losh bcache_free_instance(struct bcache *bc)
440ca987d46SWarner Losh {
441ca987d46SWarner Losh     if (bc != NULL) {
442ca987d46SWarner Losh 	free(bc->bcache_ctl);
443ca987d46SWarner Losh 	free(bc->bcache_data);
444ca987d46SWarner Losh 	free(bc);
445ca987d46SWarner Losh     }
446ca987d46SWarner Losh }
447ca987d46SWarner Losh 
448ca987d46SWarner Losh /*
449ca987d46SWarner Losh  * Insert a block into the cache.
450ca987d46SWarner Losh  */
451ca987d46SWarner Losh static void
452ca987d46SWarner Losh bcache_insert(struct bcache *bc, daddr_t blkno)
453ca987d46SWarner Losh {
454ca987d46SWarner Losh     u_int	cand;
455ca987d46SWarner Losh 
456ca987d46SWarner Losh     cand = BHASH(bc, blkno);
457ca987d46SWarner Losh 
4587325df02SKyle Evans     DPRINTF("insert blk %llu -> %u # %d", blkno, cand, bcache_bcount);
459ca987d46SWarner Losh     bc->bcache_ctl[cand].bc_blkno = blkno;
460ca987d46SWarner Losh     bc->bcache_ctl[cand].bc_count = bcache_bcount++;
461ca987d46SWarner Losh }
462ca987d46SWarner Losh 
463ca987d46SWarner Losh /*
464ca987d46SWarner Losh  * Invalidate a block from the cache.
465ca987d46SWarner Losh  */
466ca987d46SWarner Losh static void
467ca987d46SWarner Losh bcache_invalidate(struct bcache *bc, daddr_t blkno)
468ca987d46SWarner Losh {
469ca987d46SWarner Losh     u_int	i;
470ca987d46SWarner Losh 
471ca987d46SWarner Losh     i = BHASH(bc, blkno);
472ca987d46SWarner Losh     if (bc->bcache_ctl[i].bc_blkno == blkno) {
473ca987d46SWarner Losh 	bc->bcache_ctl[i].bc_count = -1;
474ca987d46SWarner Losh 	bc->bcache_ctl[i].bc_blkno = -1;
4757325df02SKyle Evans 	DPRINTF("invalidate blk %llu", blkno);
476ca987d46SWarner Losh     }
477ca987d46SWarner Losh }
478ca987d46SWarner Losh 
479ca987d46SWarner Losh #ifndef BOOT2
480ca987d46SWarner Losh COMMAND_SET(bcachestat, "bcachestat", "get disk block cache stats", command_bcache);
481ca987d46SWarner Losh 
482ca987d46SWarner Losh static int
483726f09bcSToomas Soome command_bcache(int argc, char *argv[] __unused)
484ca987d46SWarner Losh {
485ca987d46SWarner Losh     if (argc != 1) {
486ca987d46SWarner Losh 	command_errmsg = "wrong number of arguments";
487ca987d46SWarner Losh 	return(CMD_ERROR);
488ca987d46SWarner Losh     }
489ca987d46SWarner Losh 
49013dcff6bSToomas Soome     printf("\ncache blocks: %u\n", bcache_total_nblks);
49113dcff6bSToomas Soome     printf("cache blocksz: %u\n", bcache_blksize);
49213dcff6bSToomas Soome     printf("cache readahead: %u\n", bcache_rablks);
49313dcff6bSToomas Soome     printf("unit cache blocks: %u\n", bcache_unit_nblks);
49413dcff6bSToomas Soome     printf("cached units: %u\n", bcache_units);
49513dcff6bSToomas Soome     printf("%u ops %d bypasses %u hits %u misses\n", bcache_ops,
496ca987d46SWarner Losh 	bcache_bypasses, bcache_hits, bcache_misses);
497ca987d46SWarner Losh     return(CMD_OK);
498ca987d46SWarner Losh }
499ca987d46SWarner Losh #endif
500