1ca987d46SWarner Losh /*-
2ca987d46SWarner Losh * Copyright (c) 1998 Michael Smith <msmith@freebsd.org>
3ca987d46SWarner Losh * Copyright 2015 Toomas Soome <tsoome@me.com>
4ca987d46SWarner Losh * All rights reserved.
5ca987d46SWarner Losh *
6ca987d46SWarner Losh * Redistribution and use in source and binary forms, with or without
7ca987d46SWarner Losh * modification, are permitted provided that the following conditions
8ca987d46SWarner Losh * are met:
9ca987d46SWarner Losh * 1. Redistributions of source code must retain the above copyright
10ca987d46SWarner Losh * notice, this list of conditions and the following disclaimer.
11ca987d46SWarner Losh * 2. Redistributions in binary form must reproduce the above copyright
12ca987d46SWarner Losh * notice, this list of conditions and the following disclaimer in the
13ca987d46SWarner Losh * documentation and/or other materials provided with the distribution.
14ca987d46SWarner Losh *
15ca987d46SWarner Losh * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
16ca987d46SWarner Losh * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
17ca987d46SWarner Losh * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
18ca987d46SWarner Losh * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
19ca987d46SWarner Losh * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
20ca987d46SWarner Losh * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
21ca987d46SWarner Losh * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
22ca987d46SWarner Losh * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
23ca987d46SWarner Losh * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
24ca987d46SWarner Losh * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
25ca987d46SWarner Losh * SUCH DAMAGE.
26ca987d46SWarner Losh */
27ca987d46SWarner Losh
28ca987d46SWarner Losh #include <sys/param.h>
29ca987d46SWarner Losh /*
30ca987d46SWarner Losh * Simple hashed block cache
31ca987d46SWarner Losh */
32ca987d46SWarner Losh
33ca987d46SWarner Losh #include <sys/stdint.h>
34ca987d46SWarner Losh
35ca987d46SWarner Losh #include <stand.h>
36ca987d46SWarner Losh #include <string.h>
37ca987d46SWarner Losh #include <strings.h>
38ca987d46SWarner Losh
39ca987d46SWarner Losh #include "bootstrap.h"
40ca987d46SWarner Losh
41ca987d46SWarner Losh /* #define BCACHE_DEBUG */
42ca987d46SWarner Losh
43ca987d46SWarner Losh #ifdef BCACHE_DEBUG
447325df02SKyle Evans # define DPRINTF(fmt, args...) printf("%s: " fmt "\n" , __func__ , ## args)
45ca987d46SWarner Losh #else
4674772ef1SToomas Soome # define DPRINTF(fmt, args...) ((void)0)
47ca987d46SWarner Losh #endif
48ca987d46SWarner Losh
49ca987d46SWarner Losh struct bcachectl
50ca987d46SWarner Losh {
51ca987d46SWarner Losh daddr_t bc_blkno;
52ca987d46SWarner Losh int bc_count;
53ca987d46SWarner Losh };
54ca987d46SWarner Losh
55ca987d46SWarner Losh /*
56ca987d46SWarner Losh * bcache per device node. cache is allocated on device first open and freed
57ca987d46SWarner Losh * on last close, to save memory. The issue there is the size; biosdisk
58ca987d46SWarner Losh * supports up to 31 (0x1f) devices. Classic setup would use single disk
59ca987d46SWarner Losh * to boot from, but this has changed with zfs.
60ca987d46SWarner Losh */
61ca987d46SWarner Losh struct bcache {
62ca987d46SWarner Losh struct bcachectl *bcache_ctl;
63ca987d46SWarner Losh caddr_t bcache_data;
64ca987d46SWarner Losh size_t bcache_nblks;
65ca987d46SWarner Losh size_t ra;
660d13f534SMaxim Sobolev daddr_t bcache_nextblkno;
6704b9b7c5SColin Percival size_t ralen;
68ca987d46SWarner Losh };
69ca987d46SWarner Losh
70ca987d46SWarner Losh static u_int bcache_total_nblks; /* set by bcache_init */
71ca987d46SWarner Losh static u_int bcache_blksize; /* set by bcache_init */
72ca987d46SWarner Losh static u_int bcache_numdev; /* set by bcache_add_dev */
73ca987d46SWarner Losh /* statistics */
74ca987d46SWarner Losh static u_int bcache_units; /* number of devices with cache */
75ca987d46SWarner Losh static u_int bcache_unit_nblks; /* nblocks per unit */
76ca987d46SWarner Losh static u_int bcache_hits;
77ca987d46SWarner Losh static u_int bcache_misses;
78ca987d46SWarner Losh static u_int bcache_ops;
79ca987d46SWarner Losh static u_int bcache_bypasses;
80ca987d46SWarner Losh static u_int bcache_bcount;
81ca987d46SWarner Losh static u_int bcache_rablks;
82ca987d46SWarner Losh
83ca987d46SWarner Losh #define BHASH(bc, blkno) ((blkno) & ((bc)->bcache_nblks - 1))
84ca987d46SWarner Losh #define BCACHE_LOOKUP(bc, blkno) \
85ca987d46SWarner Losh ((bc)->bcache_ctl[BHASH((bc), (blkno))].bc_blkno != (blkno))
86248682a5SColin Percival #define BCACHE_READAHEAD 512
87ca987d46SWarner Losh #define BCACHE_MINREADAHEAD 32
88248682a5SColin Percival #define BCACHE_MAXIOWRA 512
89ca987d46SWarner Losh
90ca987d46SWarner Losh static void bcache_invalidate(struct bcache *bc, daddr_t blkno);
91ca987d46SWarner Losh static void bcache_insert(struct bcache *bc, daddr_t blkno);
92ca987d46SWarner Losh static void bcache_free_instance(struct bcache *bc);
93ca987d46SWarner Losh
94ca987d46SWarner Losh /*
95ca987d46SWarner Losh * Initialise the cache for (nblks) of (bsize).
96ca987d46SWarner Losh */
97ca987d46SWarner Losh void
bcache_init(size_t nblks,size_t bsize)98ca987d46SWarner Losh bcache_init(size_t nblks, size_t bsize)
99ca987d46SWarner Losh {
100ca987d46SWarner Losh /* set up control data */
101ca987d46SWarner Losh bcache_total_nblks = nblks;
102ca987d46SWarner Losh bcache_blksize = bsize;
103ca987d46SWarner Losh }
104ca987d46SWarner Losh
105ca987d46SWarner Losh /*
106ca987d46SWarner Losh * add number of devices to bcache. we have to divide cache space
107ca987d46SWarner Losh * between the devices, so bcache_add_dev() can be used to set up the
108ca987d46SWarner Losh * number. The issue is, we need to get the number before actual allocations.
109ca987d46SWarner Losh * bcache_add_dev() is supposed to be called from device init() call, so the
110ca987d46SWarner Losh * assumption is, devsw dv_init is called for plain devices first, and
111ca987d46SWarner Losh * for zfs, last.
112ca987d46SWarner Losh */
113ca987d46SWarner Losh void
bcache_add_dev(int devices)114ca987d46SWarner Losh bcache_add_dev(int devices)
115ca987d46SWarner Losh {
116ca987d46SWarner Losh bcache_numdev += devices;
117ca987d46SWarner Losh }
118ca987d46SWarner Losh
119ca987d46SWarner Losh void *
bcache_allocate(void)120ca987d46SWarner Losh bcache_allocate(void)
121ca987d46SWarner Losh {
122ca987d46SWarner Losh u_int i;
123ca987d46SWarner Losh struct bcache *bc = malloc(sizeof (struct bcache));
124ca987d46SWarner Losh int disks = bcache_numdev;
125ca987d46SWarner Losh
126ca987d46SWarner Losh if (disks == 0)
127ca987d46SWarner Losh disks = 1; /* safe guard */
128ca987d46SWarner Losh
129ca987d46SWarner Losh if (bc == NULL) {
130ca987d46SWarner Losh errno = ENOMEM;
131ca987d46SWarner Losh return (bc);
132ca987d46SWarner Losh }
133ca987d46SWarner Losh
134ca987d46SWarner Losh /*
135ca987d46SWarner Losh * the bcache block count must be power of 2 for hash function
136ca987d46SWarner Losh */
137ca987d46SWarner Losh i = fls(disks) - 1; /* highbit - 1 */
138ca987d46SWarner Losh if (disks > (1 << i)) /* next power of 2 */
139ca987d46SWarner Losh i++;
140ca987d46SWarner Losh
141ca987d46SWarner Losh bc->bcache_nblks = bcache_total_nblks >> i;
142ca987d46SWarner Losh bcache_unit_nblks = bc->bcache_nblks;
1436c81fe16SMaxim Sobolev bc->bcache_data = malloc(bc->bcache_nblks * bcache_blksize);
144ca987d46SWarner Losh if (bc->bcache_data == NULL) {
145ca987d46SWarner Losh /* dont error out yet. fall back to 32 blocks and try again */
146ca987d46SWarner Losh bc->bcache_nblks = 32;
147ca987d46SWarner Losh bc->bcache_data = malloc(bc->bcache_nblks * bcache_blksize +
148ca987d46SWarner Losh sizeof(uint32_t));
149ca987d46SWarner Losh }
150ca987d46SWarner Losh
151ca987d46SWarner Losh bc->bcache_ctl = malloc(bc->bcache_nblks * sizeof(struct bcachectl));
152ca987d46SWarner Losh
153ca987d46SWarner Losh if ((bc->bcache_data == NULL) || (bc->bcache_ctl == NULL)) {
154ca987d46SWarner Losh bcache_free_instance(bc);
155ca987d46SWarner Losh errno = ENOMEM;
156ca987d46SWarner Losh return (NULL);
157ca987d46SWarner Losh }
158ca987d46SWarner Losh
159ca987d46SWarner Losh /* Flush the cache */
160ca987d46SWarner Losh for (i = 0; i < bc->bcache_nblks; i++) {
161ca987d46SWarner Losh bc->bcache_ctl[i].bc_count = -1;
162ca987d46SWarner Losh bc->bcache_ctl[i].bc_blkno = -1;
163ca987d46SWarner Losh }
164ca987d46SWarner Losh bcache_units++;
165ca987d46SWarner Losh bc->ra = BCACHE_READAHEAD; /* optimistic read ahead */
1660d13f534SMaxim Sobolev bc->bcache_nextblkno = -1;
167ca987d46SWarner Losh return (bc);
168ca987d46SWarner Losh }
169ca987d46SWarner Losh
170ca987d46SWarner Losh void
bcache_free(void * cache)171ca987d46SWarner Losh bcache_free(void *cache)
172ca987d46SWarner Losh {
173ca987d46SWarner Losh struct bcache *bc = cache;
174ca987d46SWarner Losh
175ca987d46SWarner Losh if (bc == NULL)
176ca987d46SWarner Losh return;
177ca987d46SWarner Losh
178ca987d46SWarner Losh bcache_free_instance(bc);
179ca987d46SWarner Losh bcache_units--;
180ca987d46SWarner Losh }
181ca987d46SWarner Losh
182ca987d46SWarner Losh /*
183ca987d46SWarner Losh * Handle a write request; write directly to the disk, and populate the
184ca987d46SWarner Losh * cache with the new values.
185ca987d46SWarner Losh */
186ca987d46SWarner Losh static int
write_strategy(void * devdata,int rw,daddr_t blk,size_t size,char * buf,size_t * rsize)187ca987d46SWarner Losh write_strategy(void *devdata, int rw, daddr_t blk, size_t size,
188ca987d46SWarner Losh char *buf, size_t *rsize)
189ca987d46SWarner Losh {
190ca987d46SWarner Losh struct bcache_devdata *dd = (struct bcache_devdata *)devdata;
191ca987d46SWarner Losh struct bcache *bc = dd->dv_cache;
192ca987d46SWarner Losh daddr_t i, nblk;
193ca987d46SWarner Losh
194ca987d46SWarner Losh nblk = size / bcache_blksize;
195ca987d46SWarner Losh
196ca987d46SWarner Losh /* Invalidate the blocks being written */
197ca987d46SWarner Losh for (i = 0; i < nblk; i++) {
198ca987d46SWarner Losh bcache_invalidate(bc, blk + i);
199ca987d46SWarner Losh }
200ca987d46SWarner Losh
201ca987d46SWarner Losh /* Write the blocks */
202ca987d46SWarner Losh return (dd->dv_strategy(dd->dv_devdata, rw, blk, size, buf, rsize));
203ca987d46SWarner Losh }
204ca987d46SWarner Losh
205ca987d46SWarner Losh /*
206ca987d46SWarner Losh * Handle a read request; fill in parts of the request that can
207ca987d46SWarner Losh * be satisfied by the cache, use the supplied strategy routine to do
208ca987d46SWarner Losh * device I/O and then use the I/O results to populate the cache.
209ca987d46SWarner Losh */
210ca987d46SWarner Losh static int
read_strategy(void * devdata,int rw,daddr_t blk,size_t size,char * buf,size_t * rsize)211ca987d46SWarner Losh read_strategy(void *devdata, int rw, daddr_t blk, size_t size,
212ca987d46SWarner Losh char *buf, size_t *rsize)
213ca987d46SWarner Losh {
214ca987d46SWarner Losh struct bcache_devdata *dd = (struct bcache_devdata *)devdata;
215ca987d46SWarner Losh struct bcache *bc = dd->dv_cache;
216ca987d46SWarner Losh size_t i, nblk, p_size, r_size, complete, ra;
217ca987d46SWarner Losh int result;
218ca987d46SWarner Losh daddr_t p_blk;
219ca987d46SWarner Losh caddr_t p_buf;
220ca987d46SWarner Losh
221ca987d46SWarner Losh if (bc == NULL) {
222ca987d46SWarner Losh errno = ENODEV;
223ca987d46SWarner Losh return (-1);
224ca987d46SWarner Losh }
225ca987d46SWarner Losh
226ca987d46SWarner Losh if (rsize != NULL)
227ca987d46SWarner Losh *rsize = 0;
228ca987d46SWarner Losh
229ca987d46SWarner Losh nblk = size / bcache_blksize;
230ca987d46SWarner Losh if (nblk == 0 && size != 0)
231ca987d46SWarner Losh nblk++;
232ca987d46SWarner Losh result = 0;
233ca987d46SWarner Losh complete = 1;
234ca987d46SWarner Losh
235ca987d46SWarner Losh /* Satisfy any cache hits up front, break on first miss */
236ca987d46SWarner Losh for (i = 0; i < nblk; i++) {
237ca987d46SWarner Losh if (BCACHE_LOOKUP(bc, (daddr_t)(blk + i))) {
238ca987d46SWarner Losh bcache_misses += (nblk - i);
239ca987d46SWarner Losh complete = 0;
240ca987d46SWarner Losh break;
241ca987d46SWarner Losh } else {
242ca987d46SWarner Losh bcache_hits++;
243ca987d46SWarner Losh }
244ca987d46SWarner Losh }
245ca987d46SWarner Losh
246b841148bSColin Percival /*
247b841148bSColin Percival * Adjust read-ahead size if appropriate. Subject to the requirement
248b841148bSColin Percival * that bc->ra must stay in between MINREADAHEAD and READAHEAD, we
249b841148bSColin Percival * increase it when we notice that readahead was useful and decrease
250b841148bSColin Percival * it when we notice that readahead was not useful.
251b841148bSColin Percival */
25204b9b7c5SColin Percival if (complete || (i == bc->ralen && bc->ralen > 0)) {
253ca987d46SWarner Losh if (bc->ra < BCACHE_READAHEAD)
254ca987d46SWarner Losh bc->ra <<= 1; /* increase read ahead */
255b841148bSColin Percival } else {
25604b9b7c5SColin Percival if (nblk - i > BCACHE_MINREADAHEAD && bc->ralen > 0 &&
25704b9b7c5SColin Percival bc->ra > BCACHE_MINREADAHEAD)
258b841148bSColin Percival bc->ra >>= 1; /* reduce read ahead */
259b841148bSColin Percival }
260b841148bSColin Percival
26104b9b7c5SColin Percival /* Adjust our "unconsumed readahead" value. */
26204b9b7c5SColin Percival if (blk == bc->bcache_nextblkno) {
26304b9b7c5SColin Percival if (nblk > bc->ralen)
26404b9b7c5SColin Percival bc->ralen = 0;
26504b9b7c5SColin Percival else
26604b9b7c5SColin Percival bc->ralen -= nblk;
26704b9b7c5SColin Percival }
26804b9b7c5SColin Percival
269b841148bSColin Percival if (complete) { /* whole set was in cache, return it */
270ca987d46SWarner Losh bcopy(bc->bcache_data + (bcache_blksize * BHASH(bc, blk)), buf, size);
271ca987d46SWarner Losh goto done;
272ca987d46SWarner Losh }
273ca987d46SWarner Losh
274ca987d46SWarner Losh /*
275ca987d46SWarner Losh * Fill in any misses. From check we have i pointing to first missing
276ca987d46SWarner Losh * block, read in all remaining blocks + readahead.
277ca987d46SWarner Losh * We have space at least for nblk - i before bcache wraps.
278ca987d46SWarner Losh */
279ca987d46SWarner Losh p_blk = blk + i;
280ca987d46SWarner Losh p_buf = bc->bcache_data + (bcache_blksize * BHASH(bc, p_blk));
281ca987d46SWarner Losh r_size = bc->bcache_nblks - BHASH(bc, p_blk); /* remaining blocks */
282ca987d46SWarner Losh
283ca987d46SWarner Losh p_size = MIN(r_size, nblk - i); /* read at least those blocks */
284ca987d46SWarner Losh
285ca987d46SWarner Losh /*
286ca987d46SWarner Losh * The read ahead size setup.
287ca987d46SWarner Losh * While the read ahead can save us IO, it also can complicate things:
288ca987d46SWarner Losh * 1. We do not want to read ahead by wrapping around the
289ca987d46SWarner Losh * bcache end - this would complicate the cache management.
290ca987d46SWarner Losh * 2. We are using bc->ra as dynamic hint for read ahead size,
291ca987d46SWarner Losh * detected cache hits will increase the read-ahead block count, and
292ca987d46SWarner Losh * misses will decrease, see the code above.
293ca987d46SWarner Losh * 3. The bcache is sized by 512B blocks, however, the underlying device
294ca987d46SWarner Losh * may have a larger sector size, and we should perform the IO by
295ca987d46SWarner Losh * taking into account these larger sector sizes. We could solve this by
296ca987d46SWarner Losh * passing the sector size to bcache_allocate(), or by using ioctl(), but
297ca987d46SWarner Losh * in this version we are using the constant, 16 blocks, and are rounding
298ca987d46SWarner Losh * read ahead block count down to multiple of 16.
299ca987d46SWarner Losh * Using the constant has two reasons, we are not entirely sure if the
300ca987d46SWarner Losh * BIOS disk interface is providing the correct value for sector size.
301ca987d46SWarner Losh * And secondly, this way we get the most conservative setup for the ra.
302ca987d46SWarner Losh *
303ca987d46SWarner Losh * The selection of multiple of 16 blocks (8KB) is quite arbitrary, however,
304ca987d46SWarner Losh * we want to cover CDs (2K) and 4K disks.
305ca987d46SWarner Losh * bcache_allocate() will always fall back to a minimum of 32 blocks.
306ca987d46SWarner Losh * Our choice of 16 read ahead blocks will always fit inside the bcache.
307ca987d46SWarner Losh */
308ca987d46SWarner Losh
309ca987d46SWarner Losh if ((rw & F_NORA) == F_NORA)
310ca987d46SWarner Losh ra = 0;
311ca987d46SWarner Losh else
312ca987d46SWarner Losh ra = bc->bcache_nblks - BHASH(bc, p_blk + p_size);
313ca987d46SWarner Losh
3140d13f534SMaxim Sobolev /*
3150d13f534SMaxim Sobolev * Only trigger read-ahead if we detect two blocks being read
3160d13f534SMaxim Sobolev * sequentially.
3170d13f534SMaxim Sobolev */
3180d13f534SMaxim Sobolev if ((bc->bcache_nextblkno != blk) && ra != 0) {
3190d13f534SMaxim Sobolev ra = 0;
3200d13f534SMaxim Sobolev }
3210d13f534SMaxim Sobolev
322ca987d46SWarner Losh if (ra != 0 && ra != bc->bcache_nblks) { /* do we have RA space? */
323ca987d46SWarner Losh ra = MIN(bc->ra, ra - 1);
324ca987d46SWarner Losh ra = rounddown(ra, 16); /* multiple of 16 blocks */
325248682a5SColin Percival if (ra + p_size > BCACHE_MAXIOWRA)
326248682a5SColin Percival ra = BCACHE_MAXIOWRA - p_size;
32704b9b7c5SColin Percival bc->ralen = ra;
328ca987d46SWarner Losh p_size += ra;
32904b9b7c5SColin Percival } else {
33004b9b7c5SColin Percival bc->ralen = 0;
331ca987d46SWarner Losh }
332ca987d46SWarner Losh
333ca987d46SWarner Losh /* invalidate bcache */
334ca987d46SWarner Losh for (i = 0; i < p_size; i++) {
335ca987d46SWarner Losh bcache_invalidate(bc, p_blk + i);
336ca987d46SWarner Losh }
337ca987d46SWarner Losh
338ca987d46SWarner Losh r_size = 0;
339ca987d46SWarner Losh /*
340ca987d46SWarner Losh * with read-ahead, it may happen we are attempting to read past
341ca987d46SWarner Losh * disk end, as bcache has no information about disk size.
342ca987d46SWarner Losh * in such case we should get partial read if some blocks can be
343ca987d46SWarner Losh * read or error, if no blocks can be read.
344ca987d46SWarner Losh * in either case we should return the data in bcache and only
345ca987d46SWarner Losh * return error if there is no data.
346ca987d46SWarner Losh */
347ca987d46SWarner Losh rw &= F_MASK;
348ca987d46SWarner Losh result = dd->dv_strategy(dd->dv_devdata, rw, p_blk,
349ca987d46SWarner Losh p_size * bcache_blksize, p_buf, &r_size);
350ca987d46SWarner Losh
351ca987d46SWarner Losh r_size /= bcache_blksize;
352ca987d46SWarner Losh for (i = 0; i < r_size; i++)
353ca987d46SWarner Losh bcache_insert(bc, p_blk + i);
354ca987d46SWarner Losh
355ca987d46SWarner Losh /* update ra statistics */
356ca987d46SWarner Losh if (r_size != 0) {
357ca987d46SWarner Losh if (r_size < p_size)
358ca987d46SWarner Losh bcache_rablks += (p_size - r_size);
359ca987d46SWarner Losh else
360ca987d46SWarner Losh bcache_rablks += ra;
361ca987d46SWarner Losh }
362ca987d46SWarner Losh
363ca987d46SWarner Losh /* check how much data can we copy */
364ca987d46SWarner Losh for (i = 0; i < nblk; i++) {
365ca987d46SWarner Losh if (BCACHE_LOOKUP(bc, (daddr_t)(blk + i)))
366ca987d46SWarner Losh break;
367ca987d46SWarner Losh }
368ca987d46SWarner Losh
369ca987d46SWarner Losh if (size > i * bcache_blksize)
370ca987d46SWarner Losh size = i * bcache_blksize;
371ca987d46SWarner Losh
372ca987d46SWarner Losh if (size != 0) {
373ca987d46SWarner Losh bcopy(bc->bcache_data + (bcache_blksize * BHASH(bc, blk)), buf, size);
374ca987d46SWarner Losh result = 0;
375ca987d46SWarner Losh }
376ca987d46SWarner Losh
377ca987d46SWarner Losh done:
3780d13f534SMaxim Sobolev if (result == 0) {
3790d13f534SMaxim Sobolev if (rsize != NULL)
380ca987d46SWarner Losh *rsize = size;
3810d13f534SMaxim Sobolev bc->bcache_nextblkno = blk + (size / DEV_BSIZE);
3820d13f534SMaxim Sobolev }
383ca987d46SWarner Losh return(result);
384ca987d46SWarner Losh }
385ca987d46SWarner Losh
386ca987d46SWarner Losh /*
387ca987d46SWarner Losh * Requests larger than 1/2 cache size will be bypassed and go
388ca987d46SWarner Losh * directly to the disk. XXX tune this.
389ca987d46SWarner Losh */
390ca987d46SWarner Losh int
bcache_strategy(void * devdata,int rw,daddr_t blk,size_t size,char * buf,size_t * rsize)391ca987d46SWarner Losh bcache_strategy(void *devdata, int rw, daddr_t blk, size_t size,
392ca987d46SWarner Losh char *buf, size_t *rsize)
393ca987d46SWarner Losh {
394ca987d46SWarner Losh struct bcache_devdata *dd = (struct bcache_devdata *)devdata;
395ca987d46SWarner Losh struct bcache *bc = dd->dv_cache;
396ca987d46SWarner Losh u_int bcache_nblks = 0;
397ca987d46SWarner Losh int nblk, cblk, ret;
398ca987d46SWarner Losh size_t csize, isize, total;
399ca987d46SWarner Losh
400ca987d46SWarner Losh bcache_ops++;
401ca987d46SWarner Losh
402ca987d46SWarner Losh if (bc != NULL)
403ca987d46SWarner Losh bcache_nblks = bc->bcache_nblks;
404ca987d46SWarner Losh
405ca987d46SWarner Losh /* bypass large requests, or when the cache is inactive */
406ca987d46SWarner Losh if (bc == NULL ||
407ca987d46SWarner Losh ((size * 2 / bcache_blksize) > bcache_nblks)) {
408*8ed8b420SEmmanuel Vadot DPRINTF("bypass %zu from %jd", size / bcache_blksize, blk);
409ca987d46SWarner Losh bcache_bypasses++;
410ca987d46SWarner Losh rw &= F_MASK;
411ca987d46SWarner Losh return (dd->dv_strategy(dd->dv_devdata, rw, blk, size, buf, rsize));
412ca987d46SWarner Losh }
413ca987d46SWarner Losh
414ca987d46SWarner Losh switch (rw & F_MASK) {
415ca987d46SWarner Losh case F_READ:
416ca987d46SWarner Losh nblk = size / bcache_blksize;
417ca987d46SWarner Losh if (size != 0 && nblk == 0)
418ca987d46SWarner Losh nblk++; /* read at least one block */
419ca987d46SWarner Losh
420ca987d46SWarner Losh ret = 0;
421ca987d46SWarner Losh total = 0;
422ca987d46SWarner Losh while(size) {
423ca987d46SWarner Losh cblk = bcache_nblks - BHASH(bc, blk); /* # of blocks left */
424ca987d46SWarner Losh cblk = MIN(cblk, nblk);
425ca987d46SWarner Losh
426ca987d46SWarner Losh if (size <= bcache_blksize)
427ca987d46SWarner Losh csize = size;
428ca987d46SWarner Losh else
429ca987d46SWarner Losh csize = cblk * bcache_blksize;
430ca987d46SWarner Losh
431ca987d46SWarner Losh ret = read_strategy(devdata, rw, blk, csize, buf+total, &isize);
432ca987d46SWarner Losh
433ca987d46SWarner Losh /*
434ca987d46SWarner Losh * we may have error from read ahead, if we have read some data
435ca987d46SWarner Losh * return partial read.
436ca987d46SWarner Losh */
437ca987d46SWarner Losh if (ret != 0 || isize == 0) {
438ca987d46SWarner Losh if (total != 0)
439ca987d46SWarner Losh ret = 0;
440ca987d46SWarner Losh break;
441ca987d46SWarner Losh }
442ca987d46SWarner Losh blk += isize / bcache_blksize;
443ca987d46SWarner Losh total += isize;
444ca987d46SWarner Losh size -= isize;
445ca987d46SWarner Losh nblk = size / bcache_blksize;
446ca987d46SWarner Losh }
447ca987d46SWarner Losh
448ca987d46SWarner Losh if (rsize)
449ca987d46SWarner Losh *rsize = total;
450ca987d46SWarner Losh
451ca987d46SWarner Losh return (ret);
452ca987d46SWarner Losh case F_WRITE:
453ca987d46SWarner Losh return write_strategy(devdata, F_WRITE, blk, size, buf, rsize);
454ca987d46SWarner Losh }
455ca987d46SWarner Losh return -1;
456ca987d46SWarner Losh }
457ca987d46SWarner Losh
458ca987d46SWarner Losh /*
459ca987d46SWarner Losh * Free allocated bcache instance
460ca987d46SWarner Losh */
461ca987d46SWarner Losh static void
bcache_free_instance(struct bcache * bc)462ca987d46SWarner Losh bcache_free_instance(struct bcache *bc)
463ca987d46SWarner Losh {
464ca987d46SWarner Losh if (bc != NULL) {
465ca987d46SWarner Losh free(bc->bcache_ctl);
466ca987d46SWarner Losh free(bc->bcache_data);
467ca987d46SWarner Losh free(bc);
468ca987d46SWarner Losh }
469ca987d46SWarner Losh }
470ca987d46SWarner Losh
471ca987d46SWarner Losh /*
472ca987d46SWarner Losh * Insert a block into the cache.
473ca987d46SWarner Losh */
474ca987d46SWarner Losh static void
bcache_insert(struct bcache * bc,daddr_t blkno)475ca987d46SWarner Losh bcache_insert(struct bcache *bc, daddr_t blkno)
476ca987d46SWarner Losh {
477ca987d46SWarner Losh u_int cand;
478ca987d46SWarner Losh
479ca987d46SWarner Losh cand = BHASH(bc, blkno);
480ca987d46SWarner Losh
481*8ed8b420SEmmanuel Vadot DPRINTF("insert blk %jd -> %u # %d", blkno, cand, bcache_bcount);
482ca987d46SWarner Losh bc->bcache_ctl[cand].bc_blkno = blkno;
483ca987d46SWarner Losh bc->bcache_ctl[cand].bc_count = bcache_bcount++;
484ca987d46SWarner Losh }
485ca987d46SWarner Losh
486ca987d46SWarner Losh /*
487ca987d46SWarner Losh * Invalidate a block from the cache.
488ca987d46SWarner Losh */
489ca987d46SWarner Losh static void
bcache_invalidate(struct bcache * bc,daddr_t blkno)490ca987d46SWarner Losh bcache_invalidate(struct bcache *bc, daddr_t blkno)
491ca987d46SWarner Losh {
492ca987d46SWarner Losh u_int i;
493ca987d46SWarner Losh
494ca987d46SWarner Losh i = BHASH(bc, blkno);
495ca987d46SWarner Losh if (bc->bcache_ctl[i].bc_blkno == blkno) {
496ca987d46SWarner Losh bc->bcache_ctl[i].bc_count = -1;
497ca987d46SWarner Losh bc->bcache_ctl[i].bc_blkno = -1;
498*8ed8b420SEmmanuel Vadot DPRINTF("invalidate blk %ju", blkno);
499ca987d46SWarner Losh }
500ca987d46SWarner Losh }
501ca987d46SWarner Losh
502ca987d46SWarner Losh #ifndef BOOT2
503ca987d46SWarner Losh COMMAND_SET(bcachestat, "bcachestat", "get disk block cache stats", command_bcache);
504ca987d46SWarner Losh
505ca987d46SWarner Losh static int
command_bcache(int argc,char * argv[]__unused)506726f09bcSToomas Soome command_bcache(int argc, char *argv[] __unused)
507ca987d46SWarner Losh {
508ca987d46SWarner Losh if (argc != 1) {
509ca987d46SWarner Losh command_errmsg = "wrong number of arguments";
510ca987d46SWarner Losh return(CMD_ERROR);
511ca987d46SWarner Losh }
512ca987d46SWarner Losh
51313dcff6bSToomas Soome printf("\ncache blocks: %u\n", bcache_total_nblks);
51413dcff6bSToomas Soome printf("cache blocksz: %u\n", bcache_blksize);
51513dcff6bSToomas Soome printf("cache readahead: %u\n", bcache_rablks);
51613dcff6bSToomas Soome printf("unit cache blocks: %u\n", bcache_unit_nblks);
51713dcff6bSToomas Soome printf("cached units: %u\n", bcache_units);
51813dcff6bSToomas Soome printf("%u ops %d bypasses %u hits %u misses\n", bcache_ops,
519ca987d46SWarner Losh bcache_bypasses, bcache_hits, bcache_misses);
520ca987d46SWarner Losh return(CMD_OK);
521ca987d46SWarner Losh }
522ca987d46SWarner Losh #endif
523