1ca987d46SWarner Losh /*- 2ca987d46SWarner Losh * Copyright (c) 1998 Michael Smith <msmith@freebsd.org> 3ca987d46SWarner Losh * Copyright 2015 Toomas Soome <tsoome@me.com> 4ca987d46SWarner Losh * All rights reserved. 5ca987d46SWarner Losh * 6ca987d46SWarner Losh * Redistribution and use in source and binary forms, with or without 7ca987d46SWarner Losh * modification, are permitted provided that the following conditions 8ca987d46SWarner Losh * are met: 9ca987d46SWarner Losh * 1. Redistributions of source code must retain the above copyright 10ca987d46SWarner Losh * notice, this list of conditions and the following disclaimer. 11ca987d46SWarner Losh * 2. Redistributions in binary form must reproduce the above copyright 12ca987d46SWarner Losh * notice, this list of conditions and the following disclaimer in the 13ca987d46SWarner Losh * documentation and/or other materials provided with the distribution. 14ca987d46SWarner Losh * 15ca987d46SWarner Losh * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 16ca987d46SWarner Losh * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 17ca987d46SWarner Losh * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 18ca987d46SWarner Losh * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE 19ca987d46SWarner Losh * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 20ca987d46SWarner Losh * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 21ca987d46SWarner Losh * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 22ca987d46SWarner Losh * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 23ca987d46SWarner Losh * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 24ca987d46SWarner Losh * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 25ca987d46SWarner Losh * SUCH DAMAGE. 26ca987d46SWarner Losh */ 27ca987d46SWarner Losh 28ca987d46SWarner Losh #include <sys/cdefs.h> 29ca987d46SWarner Losh #include <sys/param.h> 30ca987d46SWarner Losh __FBSDID("$FreeBSD$"); 31ca987d46SWarner Losh 32ca987d46SWarner Losh /* 33ca987d46SWarner Losh * Simple hashed block cache 34ca987d46SWarner Losh */ 35ca987d46SWarner Losh 36ca987d46SWarner Losh #include <sys/stdint.h> 37ca987d46SWarner Losh 38ca987d46SWarner Losh #include <stand.h> 39ca987d46SWarner Losh #include <string.h> 40ca987d46SWarner Losh #include <strings.h> 41ca987d46SWarner Losh 42ca987d46SWarner Losh #include "bootstrap.h" 43ca987d46SWarner Losh 44ca987d46SWarner Losh /* #define BCACHE_DEBUG */ 45ca987d46SWarner Losh 46ca987d46SWarner Losh #ifdef BCACHE_DEBUG 477325df02SKyle Evans # define DPRINTF(fmt, args...) printf("%s: " fmt "\n" , __func__ , ## args) 48ca987d46SWarner Losh #else 4974772ef1SToomas Soome # define DPRINTF(fmt, args...) ((void)0) 50ca987d46SWarner Losh #endif 51ca987d46SWarner Losh 52ca987d46SWarner Losh struct bcachectl 53ca987d46SWarner Losh { 54ca987d46SWarner Losh daddr_t bc_blkno; 55ca987d46SWarner Losh int bc_count; 56ca987d46SWarner Losh }; 57ca987d46SWarner Losh 58ca987d46SWarner Losh /* 59ca987d46SWarner Losh * bcache per device node. cache is allocated on device first open and freed 60ca987d46SWarner Losh * on last close, to save memory. The issue there is the size; biosdisk 61ca987d46SWarner Losh * supports up to 31 (0x1f) devices. Classic setup would use single disk 62ca987d46SWarner Losh * to boot from, but this has changed with zfs. 63ca987d46SWarner Losh */ 64ca987d46SWarner Losh struct bcache { 65ca987d46SWarner Losh struct bcachectl *bcache_ctl; 66ca987d46SWarner Losh caddr_t bcache_data; 67ca987d46SWarner Losh size_t bcache_nblks; 68ca987d46SWarner Losh size_t ra; 690d13f534SMaxim Sobolev daddr_t bcache_nextblkno; 70*04b9b7c5SColin Percival size_t ralen; 71ca987d46SWarner Losh }; 72ca987d46SWarner Losh 73ca987d46SWarner Losh static u_int bcache_total_nblks; /* set by bcache_init */ 74ca987d46SWarner Losh static u_int bcache_blksize; /* set by bcache_init */ 75ca987d46SWarner Losh static u_int bcache_numdev; /* set by bcache_add_dev */ 76ca987d46SWarner Losh /* statistics */ 77ca987d46SWarner Losh static u_int bcache_units; /* number of devices with cache */ 78ca987d46SWarner Losh static u_int bcache_unit_nblks; /* nblocks per unit */ 79ca987d46SWarner Losh static u_int bcache_hits; 80ca987d46SWarner Losh static u_int bcache_misses; 81ca987d46SWarner Losh static u_int bcache_ops; 82ca987d46SWarner Losh static u_int bcache_bypasses; 83ca987d46SWarner Losh static u_int bcache_bcount; 84ca987d46SWarner Losh static u_int bcache_rablks; 85ca987d46SWarner Losh 86ca987d46SWarner Losh #define BHASH(bc, blkno) ((blkno) & ((bc)->bcache_nblks - 1)) 87ca987d46SWarner Losh #define BCACHE_LOOKUP(bc, blkno) \ 88ca987d46SWarner Losh ((bc)->bcache_ctl[BHASH((bc), (blkno))].bc_blkno != (blkno)) 89ca987d46SWarner Losh #define BCACHE_READAHEAD 256 90ca987d46SWarner Losh #define BCACHE_MINREADAHEAD 32 91ca987d46SWarner Losh 92ca987d46SWarner Losh static void bcache_invalidate(struct bcache *bc, daddr_t blkno); 93ca987d46SWarner Losh static void bcache_insert(struct bcache *bc, daddr_t blkno); 94ca987d46SWarner Losh static void bcache_free_instance(struct bcache *bc); 95ca987d46SWarner Losh 96ca987d46SWarner Losh /* 97ca987d46SWarner Losh * Initialise the cache for (nblks) of (bsize). 98ca987d46SWarner Losh */ 99ca987d46SWarner Losh void 100ca987d46SWarner Losh bcache_init(size_t nblks, size_t bsize) 101ca987d46SWarner Losh { 102ca987d46SWarner Losh /* set up control data */ 103ca987d46SWarner Losh bcache_total_nblks = nblks; 104ca987d46SWarner Losh bcache_blksize = bsize; 105ca987d46SWarner Losh } 106ca987d46SWarner Losh 107ca987d46SWarner Losh /* 108ca987d46SWarner Losh * add number of devices to bcache. we have to divide cache space 109ca987d46SWarner Losh * between the devices, so bcache_add_dev() can be used to set up the 110ca987d46SWarner Losh * number. The issue is, we need to get the number before actual allocations. 111ca987d46SWarner Losh * bcache_add_dev() is supposed to be called from device init() call, so the 112ca987d46SWarner Losh * assumption is, devsw dv_init is called for plain devices first, and 113ca987d46SWarner Losh * for zfs, last. 114ca987d46SWarner Losh */ 115ca987d46SWarner Losh void 116ca987d46SWarner Losh bcache_add_dev(int devices) 117ca987d46SWarner Losh { 118ca987d46SWarner Losh bcache_numdev += devices; 119ca987d46SWarner Losh } 120ca987d46SWarner Losh 121ca987d46SWarner Losh void * 122ca987d46SWarner Losh bcache_allocate(void) 123ca987d46SWarner Losh { 124ca987d46SWarner Losh u_int i; 125ca987d46SWarner Losh struct bcache *bc = malloc(sizeof (struct bcache)); 126ca987d46SWarner Losh int disks = bcache_numdev; 127ca987d46SWarner Losh 128ca987d46SWarner Losh if (disks == 0) 129ca987d46SWarner Losh disks = 1; /* safe guard */ 130ca987d46SWarner Losh 131ca987d46SWarner Losh if (bc == NULL) { 132ca987d46SWarner Losh errno = ENOMEM; 133ca987d46SWarner Losh return (bc); 134ca987d46SWarner Losh } 135ca987d46SWarner Losh 136ca987d46SWarner Losh /* 137ca987d46SWarner Losh * the bcache block count must be power of 2 for hash function 138ca987d46SWarner Losh */ 139ca987d46SWarner Losh i = fls(disks) - 1; /* highbit - 1 */ 140ca987d46SWarner Losh if (disks > (1 << i)) /* next power of 2 */ 141ca987d46SWarner Losh i++; 142ca987d46SWarner Losh 143ca987d46SWarner Losh bc->bcache_nblks = bcache_total_nblks >> i; 144ca987d46SWarner Losh bcache_unit_nblks = bc->bcache_nblks; 1456c81fe16SMaxim Sobolev bc->bcache_data = malloc(bc->bcache_nblks * bcache_blksize); 146ca987d46SWarner Losh if (bc->bcache_data == NULL) { 147ca987d46SWarner Losh /* dont error out yet. fall back to 32 blocks and try again */ 148ca987d46SWarner Losh bc->bcache_nblks = 32; 149ca987d46SWarner Losh bc->bcache_data = malloc(bc->bcache_nblks * bcache_blksize + 150ca987d46SWarner Losh sizeof(uint32_t)); 151ca987d46SWarner Losh } 152ca987d46SWarner Losh 153ca987d46SWarner Losh bc->bcache_ctl = malloc(bc->bcache_nblks * sizeof(struct bcachectl)); 154ca987d46SWarner Losh 155ca987d46SWarner Losh if ((bc->bcache_data == NULL) || (bc->bcache_ctl == NULL)) { 156ca987d46SWarner Losh bcache_free_instance(bc); 157ca987d46SWarner Losh errno = ENOMEM; 158ca987d46SWarner Losh return (NULL); 159ca987d46SWarner Losh } 160ca987d46SWarner Losh 161ca987d46SWarner Losh /* Flush the cache */ 162ca987d46SWarner Losh for (i = 0; i < bc->bcache_nblks; i++) { 163ca987d46SWarner Losh bc->bcache_ctl[i].bc_count = -1; 164ca987d46SWarner Losh bc->bcache_ctl[i].bc_blkno = -1; 165ca987d46SWarner Losh } 166ca987d46SWarner Losh bcache_units++; 167ca987d46SWarner Losh bc->ra = BCACHE_READAHEAD; /* optimistic read ahead */ 1680d13f534SMaxim Sobolev bc->bcache_nextblkno = -1; 169ca987d46SWarner Losh return (bc); 170ca987d46SWarner Losh } 171ca987d46SWarner Losh 172ca987d46SWarner Losh void 173ca987d46SWarner Losh bcache_free(void *cache) 174ca987d46SWarner Losh { 175ca987d46SWarner Losh struct bcache *bc = cache; 176ca987d46SWarner Losh 177ca987d46SWarner Losh if (bc == NULL) 178ca987d46SWarner Losh return; 179ca987d46SWarner Losh 180ca987d46SWarner Losh bcache_free_instance(bc); 181ca987d46SWarner Losh bcache_units--; 182ca987d46SWarner Losh } 183ca987d46SWarner Losh 184ca987d46SWarner Losh /* 185ca987d46SWarner Losh * Handle a write request; write directly to the disk, and populate the 186ca987d46SWarner Losh * cache with the new values. 187ca987d46SWarner Losh */ 188ca987d46SWarner Losh static int 189ca987d46SWarner Losh write_strategy(void *devdata, int rw, daddr_t blk, size_t size, 190ca987d46SWarner Losh char *buf, size_t *rsize) 191ca987d46SWarner Losh { 192ca987d46SWarner Losh struct bcache_devdata *dd = (struct bcache_devdata *)devdata; 193ca987d46SWarner Losh struct bcache *bc = dd->dv_cache; 194ca987d46SWarner Losh daddr_t i, nblk; 195ca987d46SWarner Losh 196ca987d46SWarner Losh nblk = size / bcache_blksize; 197ca987d46SWarner Losh 198ca987d46SWarner Losh /* Invalidate the blocks being written */ 199ca987d46SWarner Losh for (i = 0; i < nblk; i++) { 200ca987d46SWarner Losh bcache_invalidate(bc, blk + i); 201ca987d46SWarner Losh } 202ca987d46SWarner Losh 203ca987d46SWarner Losh /* Write the blocks */ 204ca987d46SWarner Losh return (dd->dv_strategy(dd->dv_devdata, rw, blk, size, buf, rsize)); 205ca987d46SWarner Losh } 206ca987d46SWarner Losh 207ca987d46SWarner Losh /* 208ca987d46SWarner Losh * Handle a read request; fill in parts of the request that can 209ca987d46SWarner Losh * be satisfied by the cache, use the supplied strategy routine to do 210ca987d46SWarner Losh * device I/O and then use the I/O results to populate the cache. 211ca987d46SWarner Losh */ 212ca987d46SWarner Losh static int 213ca987d46SWarner Losh read_strategy(void *devdata, int rw, daddr_t blk, size_t size, 214ca987d46SWarner Losh char *buf, size_t *rsize) 215ca987d46SWarner Losh { 216ca987d46SWarner Losh struct bcache_devdata *dd = (struct bcache_devdata *)devdata; 217ca987d46SWarner Losh struct bcache *bc = dd->dv_cache; 218ca987d46SWarner Losh size_t i, nblk, p_size, r_size, complete, ra; 219ca987d46SWarner Losh int result; 220ca987d46SWarner Losh daddr_t p_blk; 221ca987d46SWarner Losh caddr_t p_buf; 222ca987d46SWarner Losh 223ca987d46SWarner Losh if (bc == NULL) { 224ca987d46SWarner Losh errno = ENODEV; 225ca987d46SWarner Losh return (-1); 226ca987d46SWarner Losh } 227ca987d46SWarner Losh 228ca987d46SWarner Losh if (rsize != NULL) 229ca987d46SWarner Losh *rsize = 0; 230ca987d46SWarner Losh 231ca987d46SWarner Losh nblk = size / bcache_blksize; 232ca987d46SWarner Losh if (nblk == 0 && size != 0) 233ca987d46SWarner Losh nblk++; 234ca987d46SWarner Losh result = 0; 235ca987d46SWarner Losh complete = 1; 236ca987d46SWarner Losh 237ca987d46SWarner Losh /* Satisfy any cache hits up front, break on first miss */ 238ca987d46SWarner Losh for (i = 0; i < nblk; i++) { 239ca987d46SWarner Losh if (BCACHE_LOOKUP(bc, (daddr_t)(blk + i))) { 240ca987d46SWarner Losh bcache_misses += (nblk - i); 241ca987d46SWarner Losh complete = 0; 242ca987d46SWarner Losh break; 243ca987d46SWarner Losh } else { 244ca987d46SWarner Losh bcache_hits++; 245ca987d46SWarner Losh } 246ca987d46SWarner Losh } 247ca987d46SWarner Losh 248b841148bSColin Percival /* 249b841148bSColin Percival * Adjust read-ahead size if appropriate. Subject to the requirement 250b841148bSColin Percival * that bc->ra must stay in between MINREADAHEAD and READAHEAD, we 251b841148bSColin Percival * increase it when we notice that readahead was useful and decrease 252b841148bSColin Percival * it when we notice that readahead was not useful. 253b841148bSColin Percival */ 254*04b9b7c5SColin Percival if (complete || (i == bc->ralen && bc->ralen > 0)) { 255ca987d46SWarner Losh if (bc->ra < BCACHE_READAHEAD) 256ca987d46SWarner Losh bc->ra <<= 1; /* increase read ahead */ 257b841148bSColin Percival } else { 258*04b9b7c5SColin Percival if (nblk - i > BCACHE_MINREADAHEAD && bc->ralen > 0 && 259*04b9b7c5SColin Percival bc->ra > BCACHE_MINREADAHEAD) 260b841148bSColin Percival bc->ra >>= 1; /* reduce read ahead */ 261b841148bSColin Percival } 262b841148bSColin Percival 263*04b9b7c5SColin Percival /* Adjust our "unconsumed readahead" value. */ 264*04b9b7c5SColin Percival if (blk == bc->bcache_nextblkno) { 265*04b9b7c5SColin Percival if (nblk > bc->ralen) 266*04b9b7c5SColin Percival bc->ralen = 0; 267*04b9b7c5SColin Percival else 268*04b9b7c5SColin Percival bc->ralen -= nblk; 269*04b9b7c5SColin Percival } 270*04b9b7c5SColin Percival 271b841148bSColin Percival if (complete) { /* whole set was in cache, return it */ 272ca987d46SWarner Losh bcopy(bc->bcache_data + (bcache_blksize * BHASH(bc, blk)), buf, size); 273ca987d46SWarner Losh goto done; 274ca987d46SWarner Losh } 275ca987d46SWarner Losh 276ca987d46SWarner Losh /* 277ca987d46SWarner Losh * Fill in any misses. From check we have i pointing to first missing 278ca987d46SWarner Losh * block, read in all remaining blocks + readahead. 279ca987d46SWarner Losh * We have space at least for nblk - i before bcache wraps. 280ca987d46SWarner Losh */ 281ca987d46SWarner Losh p_blk = blk + i; 282ca987d46SWarner Losh p_buf = bc->bcache_data + (bcache_blksize * BHASH(bc, p_blk)); 283ca987d46SWarner Losh r_size = bc->bcache_nblks - BHASH(bc, p_blk); /* remaining blocks */ 284ca987d46SWarner Losh 285ca987d46SWarner Losh p_size = MIN(r_size, nblk - i); /* read at least those blocks */ 286ca987d46SWarner Losh 287ca987d46SWarner Losh /* 288ca987d46SWarner Losh * The read ahead size setup. 289ca987d46SWarner Losh * While the read ahead can save us IO, it also can complicate things: 290ca987d46SWarner Losh * 1. We do not want to read ahead by wrapping around the 291ca987d46SWarner Losh * bcache end - this would complicate the cache management. 292ca987d46SWarner Losh * 2. We are using bc->ra as dynamic hint for read ahead size, 293ca987d46SWarner Losh * detected cache hits will increase the read-ahead block count, and 294ca987d46SWarner Losh * misses will decrease, see the code above. 295ca987d46SWarner Losh * 3. The bcache is sized by 512B blocks, however, the underlying device 296ca987d46SWarner Losh * may have a larger sector size, and we should perform the IO by 297ca987d46SWarner Losh * taking into account these larger sector sizes. We could solve this by 298ca987d46SWarner Losh * passing the sector size to bcache_allocate(), or by using ioctl(), but 299ca987d46SWarner Losh * in this version we are using the constant, 16 blocks, and are rounding 300ca987d46SWarner Losh * read ahead block count down to multiple of 16. 301ca987d46SWarner Losh * Using the constant has two reasons, we are not entirely sure if the 302ca987d46SWarner Losh * BIOS disk interface is providing the correct value for sector size. 303ca987d46SWarner Losh * And secondly, this way we get the most conservative setup for the ra. 304ca987d46SWarner Losh * 305ca987d46SWarner Losh * The selection of multiple of 16 blocks (8KB) is quite arbitrary, however, 306ca987d46SWarner Losh * we want to cover CDs (2K) and 4K disks. 307ca987d46SWarner Losh * bcache_allocate() will always fall back to a minimum of 32 blocks. 308ca987d46SWarner Losh * Our choice of 16 read ahead blocks will always fit inside the bcache. 309ca987d46SWarner Losh */ 310ca987d46SWarner Losh 311ca987d46SWarner Losh if ((rw & F_NORA) == F_NORA) 312ca987d46SWarner Losh ra = 0; 313ca987d46SWarner Losh else 314ca987d46SWarner Losh ra = bc->bcache_nblks - BHASH(bc, p_blk + p_size); 315ca987d46SWarner Losh 3160d13f534SMaxim Sobolev /* 3170d13f534SMaxim Sobolev * Only trigger read-ahead if we detect two blocks being read 3180d13f534SMaxim Sobolev * sequentially. 3190d13f534SMaxim Sobolev */ 3200d13f534SMaxim Sobolev if ((bc->bcache_nextblkno != blk) && ra != 0) { 3210d13f534SMaxim Sobolev ra = 0; 3220d13f534SMaxim Sobolev } 3230d13f534SMaxim Sobolev 324ca987d46SWarner Losh if (ra != 0 && ra != bc->bcache_nblks) { /* do we have RA space? */ 325ca987d46SWarner Losh ra = MIN(bc->ra, ra - 1); 326ca987d46SWarner Losh ra = rounddown(ra, 16); /* multiple of 16 blocks */ 327*04b9b7c5SColin Percival bc->ralen = ra; 328ca987d46SWarner Losh p_size += ra; 329*04b9b7c5SColin Percival } else { 330*04b9b7c5SColin Percival bc->ralen = 0; 331ca987d46SWarner Losh } 332ca987d46SWarner Losh 333ca987d46SWarner Losh /* invalidate bcache */ 334ca987d46SWarner Losh for (i = 0; i < p_size; i++) { 335ca987d46SWarner Losh bcache_invalidate(bc, p_blk + i); 336ca987d46SWarner Losh } 337ca987d46SWarner Losh 338ca987d46SWarner Losh r_size = 0; 339ca987d46SWarner Losh /* 340ca987d46SWarner Losh * with read-ahead, it may happen we are attempting to read past 341ca987d46SWarner Losh * disk end, as bcache has no information about disk size. 342ca987d46SWarner Losh * in such case we should get partial read if some blocks can be 343ca987d46SWarner Losh * read or error, if no blocks can be read. 344ca987d46SWarner Losh * in either case we should return the data in bcache and only 345ca987d46SWarner Losh * return error if there is no data. 346ca987d46SWarner Losh */ 347ca987d46SWarner Losh rw &= F_MASK; 348ca987d46SWarner Losh result = dd->dv_strategy(dd->dv_devdata, rw, p_blk, 349ca987d46SWarner Losh p_size * bcache_blksize, p_buf, &r_size); 350ca987d46SWarner Losh 351ca987d46SWarner Losh r_size /= bcache_blksize; 352ca987d46SWarner Losh for (i = 0; i < r_size; i++) 353ca987d46SWarner Losh bcache_insert(bc, p_blk + i); 354ca987d46SWarner Losh 355ca987d46SWarner Losh /* update ra statistics */ 356ca987d46SWarner Losh if (r_size != 0) { 357ca987d46SWarner Losh if (r_size < p_size) 358ca987d46SWarner Losh bcache_rablks += (p_size - r_size); 359ca987d46SWarner Losh else 360ca987d46SWarner Losh bcache_rablks += ra; 361ca987d46SWarner Losh } 362ca987d46SWarner Losh 363ca987d46SWarner Losh /* check how much data can we copy */ 364ca987d46SWarner Losh for (i = 0; i < nblk; i++) { 365ca987d46SWarner Losh if (BCACHE_LOOKUP(bc, (daddr_t)(blk + i))) 366ca987d46SWarner Losh break; 367ca987d46SWarner Losh } 368ca987d46SWarner Losh 369ca987d46SWarner Losh if (size > i * bcache_blksize) 370ca987d46SWarner Losh size = i * bcache_blksize; 371ca987d46SWarner Losh 372ca987d46SWarner Losh if (size != 0) { 373ca987d46SWarner Losh bcopy(bc->bcache_data + (bcache_blksize * BHASH(bc, blk)), buf, size); 374ca987d46SWarner Losh result = 0; 375ca987d46SWarner Losh } 376ca987d46SWarner Losh 377ca987d46SWarner Losh done: 3780d13f534SMaxim Sobolev if (result == 0) { 3790d13f534SMaxim Sobolev if (rsize != NULL) 380ca987d46SWarner Losh *rsize = size; 3810d13f534SMaxim Sobolev bc->bcache_nextblkno = blk + (size / DEV_BSIZE); 3820d13f534SMaxim Sobolev } 383ca987d46SWarner Losh return(result); 384ca987d46SWarner Losh } 385ca987d46SWarner Losh 386ca987d46SWarner Losh /* 387ca987d46SWarner Losh * Requests larger than 1/2 cache size will be bypassed and go 388ca987d46SWarner Losh * directly to the disk. XXX tune this. 389ca987d46SWarner Losh */ 390ca987d46SWarner Losh int 391ca987d46SWarner Losh bcache_strategy(void *devdata, int rw, daddr_t blk, size_t size, 392ca987d46SWarner Losh char *buf, size_t *rsize) 393ca987d46SWarner Losh { 394ca987d46SWarner Losh struct bcache_devdata *dd = (struct bcache_devdata *)devdata; 395ca987d46SWarner Losh struct bcache *bc = dd->dv_cache; 396ca987d46SWarner Losh u_int bcache_nblks = 0; 397ca987d46SWarner Losh int nblk, cblk, ret; 398ca987d46SWarner Losh size_t csize, isize, total; 399ca987d46SWarner Losh 400ca987d46SWarner Losh bcache_ops++; 401ca987d46SWarner Losh 402ca987d46SWarner Losh if (bc != NULL) 403ca987d46SWarner Losh bcache_nblks = bc->bcache_nblks; 404ca987d46SWarner Losh 405ca987d46SWarner Losh /* bypass large requests, or when the cache is inactive */ 406ca987d46SWarner Losh if (bc == NULL || 407ca987d46SWarner Losh ((size * 2 / bcache_blksize) > bcache_nblks)) { 4087325df02SKyle Evans DPRINTF("bypass %zu from %qu", size / bcache_blksize, blk); 409ca987d46SWarner Losh bcache_bypasses++; 410ca987d46SWarner Losh rw &= F_MASK; 411ca987d46SWarner Losh return (dd->dv_strategy(dd->dv_devdata, rw, blk, size, buf, rsize)); 412ca987d46SWarner Losh } 413ca987d46SWarner Losh 414ca987d46SWarner Losh switch (rw & F_MASK) { 415ca987d46SWarner Losh case F_READ: 416ca987d46SWarner Losh nblk = size / bcache_blksize; 417ca987d46SWarner Losh if (size != 0 && nblk == 0) 418ca987d46SWarner Losh nblk++; /* read at least one block */ 419ca987d46SWarner Losh 420ca987d46SWarner Losh ret = 0; 421ca987d46SWarner Losh total = 0; 422ca987d46SWarner Losh while(size) { 423ca987d46SWarner Losh cblk = bcache_nblks - BHASH(bc, blk); /* # of blocks left */ 424ca987d46SWarner Losh cblk = MIN(cblk, nblk); 425ca987d46SWarner Losh 426ca987d46SWarner Losh if (size <= bcache_blksize) 427ca987d46SWarner Losh csize = size; 428ca987d46SWarner Losh else 429ca987d46SWarner Losh csize = cblk * bcache_blksize; 430ca987d46SWarner Losh 431ca987d46SWarner Losh ret = read_strategy(devdata, rw, blk, csize, buf+total, &isize); 432ca987d46SWarner Losh 433ca987d46SWarner Losh /* 434ca987d46SWarner Losh * we may have error from read ahead, if we have read some data 435ca987d46SWarner Losh * return partial read. 436ca987d46SWarner Losh */ 437ca987d46SWarner Losh if (ret != 0 || isize == 0) { 438ca987d46SWarner Losh if (total != 0) 439ca987d46SWarner Losh ret = 0; 440ca987d46SWarner Losh break; 441ca987d46SWarner Losh } 442ca987d46SWarner Losh blk += isize / bcache_blksize; 443ca987d46SWarner Losh total += isize; 444ca987d46SWarner Losh size -= isize; 445ca987d46SWarner Losh nblk = size / bcache_blksize; 446ca987d46SWarner Losh } 447ca987d46SWarner Losh 448ca987d46SWarner Losh if (rsize) 449ca987d46SWarner Losh *rsize = total; 450ca987d46SWarner Losh 451ca987d46SWarner Losh return (ret); 452ca987d46SWarner Losh case F_WRITE: 453ca987d46SWarner Losh return write_strategy(devdata, F_WRITE, blk, size, buf, rsize); 454ca987d46SWarner Losh } 455ca987d46SWarner Losh return -1; 456ca987d46SWarner Losh } 457ca987d46SWarner Losh 458ca987d46SWarner Losh /* 459ca987d46SWarner Losh * Free allocated bcache instance 460ca987d46SWarner Losh */ 461ca987d46SWarner Losh static void 462ca987d46SWarner Losh bcache_free_instance(struct bcache *bc) 463ca987d46SWarner Losh { 464ca987d46SWarner Losh if (bc != NULL) { 465ca987d46SWarner Losh free(bc->bcache_ctl); 466ca987d46SWarner Losh free(bc->bcache_data); 467ca987d46SWarner Losh free(bc); 468ca987d46SWarner Losh } 469ca987d46SWarner Losh } 470ca987d46SWarner Losh 471ca987d46SWarner Losh /* 472ca987d46SWarner Losh * Insert a block into the cache. 473ca987d46SWarner Losh */ 474ca987d46SWarner Losh static void 475ca987d46SWarner Losh bcache_insert(struct bcache *bc, daddr_t blkno) 476ca987d46SWarner Losh { 477ca987d46SWarner Losh u_int cand; 478ca987d46SWarner Losh 479ca987d46SWarner Losh cand = BHASH(bc, blkno); 480ca987d46SWarner Losh 4817325df02SKyle Evans DPRINTF("insert blk %llu -> %u # %d", blkno, cand, bcache_bcount); 482ca987d46SWarner Losh bc->bcache_ctl[cand].bc_blkno = blkno; 483ca987d46SWarner Losh bc->bcache_ctl[cand].bc_count = bcache_bcount++; 484ca987d46SWarner Losh } 485ca987d46SWarner Losh 486ca987d46SWarner Losh /* 487ca987d46SWarner Losh * Invalidate a block from the cache. 488ca987d46SWarner Losh */ 489ca987d46SWarner Losh static void 490ca987d46SWarner Losh bcache_invalidate(struct bcache *bc, daddr_t blkno) 491ca987d46SWarner Losh { 492ca987d46SWarner Losh u_int i; 493ca987d46SWarner Losh 494ca987d46SWarner Losh i = BHASH(bc, blkno); 495ca987d46SWarner Losh if (bc->bcache_ctl[i].bc_blkno == blkno) { 496ca987d46SWarner Losh bc->bcache_ctl[i].bc_count = -1; 497ca987d46SWarner Losh bc->bcache_ctl[i].bc_blkno = -1; 4987325df02SKyle Evans DPRINTF("invalidate blk %llu", blkno); 499ca987d46SWarner Losh } 500ca987d46SWarner Losh } 501ca987d46SWarner Losh 502ca987d46SWarner Losh #ifndef BOOT2 503ca987d46SWarner Losh COMMAND_SET(bcachestat, "bcachestat", "get disk block cache stats", command_bcache); 504ca987d46SWarner Losh 505ca987d46SWarner Losh static int 506726f09bcSToomas Soome command_bcache(int argc, char *argv[] __unused) 507ca987d46SWarner Losh { 508ca987d46SWarner Losh if (argc != 1) { 509ca987d46SWarner Losh command_errmsg = "wrong number of arguments"; 510ca987d46SWarner Losh return(CMD_ERROR); 511ca987d46SWarner Losh } 512ca987d46SWarner Losh 51313dcff6bSToomas Soome printf("\ncache blocks: %u\n", bcache_total_nblks); 51413dcff6bSToomas Soome printf("cache blocksz: %u\n", bcache_blksize); 51513dcff6bSToomas Soome printf("cache readahead: %u\n", bcache_rablks); 51613dcff6bSToomas Soome printf("unit cache blocks: %u\n", bcache_unit_nblks); 51713dcff6bSToomas Soome printf("cached units: %u\n", bcache_units); 51813dcff6bSToomas Soome printf("%u ops %d bypasses %u hits %u misses\n", bcache_ops, 519ca987d46SWarner Losh bcache_bypasses, bcache_hits, bcache_misses); 520ca987d46SWarner Losh return(CMD_OK); 521ca987d46SWarner Losh } 522ca987d46SWarner Losh #endif 523