1*7c478bd9Sstevel@tonic-gate /*- 2*7c478bd9Sstevel@tonic-gate * See the file LICENSE for redistribution information. 3*7c478bd9Sstevel@tonic-gate * 4*7c478bd9Sstevel@tonic-gate * Copyright (c) 1996, 1997, 1998 5*7c478bd9Sstevel@tonic-gate * Sleepycat Software. All rights reserved. 6*7c478bd9Sstevel@tonic-gate */ 7*7c478bd9Sstevel@tonic-gate #include "config.h" 8*7c478bd9Sstevel@tonic-gate 9*7c478bd9Sstevel@tonic-gate #ifndef lint 10*7c478bd9Sstevel@tonic-gate static const char sccsid[] = "@(#)mp_fget.c 10.53 (Sleepycat) 11/16/98"; 11*7c478bd9Sstevel@tonic-gate #endif /* not lint */ 12*7c478bd9Sstevel@tonic-gate 13*7c478bd9Sstevel@tonic-gate #ifndef NO_SYSTEM_INCLUDES 14*7c478bd9Sstevel@tonic-gate #include <sys/types.h> 15*7c478bd9Sstevel@tonic-gate 16*7c478bd9Sstevel@tonic-gate #include <errno.h> 17*7c478bd9Sstevel@tonic-gate #include <string.h> 18*7c478bd9Sstevel@tonic-gate #endif 19*7c478bd9Sstevel@tonic-gate 20*7c478bd9Sstevel@tonic-gate #include "db_int.h" 21*7c478bd9Sstevel@tonic-gate #include "shqueue.h" 22*7c478bd9Sstevel@tonic-gate #include "db_shash.h" 23*7c478bd9Sstevel@tonic-gate #include "mp.h" 24*7c478bd9Sstevel@tonic-gate #include "common_ext.h" 25*7c478bd9Sstevel@tonic-gate 26*7c478bd9Sstevel@tonic-gate /* 27*7c478bd9Sstevel@tonic-gate * memp_fget -- 28*7c478bd9Sstevel@tonic-gate * Get a page from the file. 29*7c478bd9Sstevel@tonic-gate */ 30*7c478bd9Sstevel@tonic-gate int 31*7c478bd9Sstevel@tonic-gate memp_fget(dbmfp, pgnoaddr, flags, addrp) 32*7c478bd9Sstevel@tonic-gate DB_MPOOLFILE *dbmfp; 33*7c478bd9Sstevel@tonic-gate db_pgno_t *pgnoaddr; 34*7c478bd9Sstevel@tonic-gate u_int32_t flags; 35*7c478bd9Sstevel@tonic-gate void *addrp; 36*7c478bd9Sstevel@tonic-gate { 37*7c478bd9Sstevel@tonic-gate BH *bhp; 38*7c478bd9Sstevel@tonic-gate DB_MPOOL *dbmp; 39*7c478bd9Sstevel@tonic-gate MPOOL *mp; 40*7c478bd9Sstevel@tonic-gate MPOOLFILE *mfp; 41*7c478bd9Sstevel@tonic-gate size_t bucket, mf_offset; 42*7c478bd9Sstevel@tonic-gate u_int32_t st_hsearch; 43*7c478bd9Sstevel@tonic-gate int b_incr, first, ret; 44*7c478bd9Sstevel@tonic-gate 45*7c478bd9Sstevel@tonic-gate dbmp = dbmfp->dbmp; 46*7c478bd9Sstevel@tonic-gate mp = dbmp->mp; 47*7c478bd9Sstevel@tonic-gate mfp = dbmfp->mfp; 48*7c478bd9Sstevel@tonic-gate 49*7c478bd9Sstevel@tonic-gate MP_PANIC_CHECK(dbmp); 50*7c478bd9Sstevel@tonic-gate 51*7c478bd9Sstevel@tonic-gate /* 52*7c478bd9Sstevel@tonic-gate * Validate arguments. 53*7c478bd9Sstevel@tonic-gate * 54*7c478bd9Sstevel@tonic-gate * !!! 55*7c478bd9Sstevel@tonic-gate * Don't test for DB_MPOOL_CREATE and DB_MPOOL_NEW flags for readonly 56*7c478bd9Sstevel@tonic-gate * files here, and create non-existent pages in readonly files if the 57*7c478bd9Sstevel@tonic-gate * flags are set, later. The reason is that the hash access method 58*7c478bd9Sstevel@tonic-gate * wants to get empty pages that don't really exist in readonly files. 59*7c478bd9Sstevel@tonic-gate * The only alternative is for hash to write the last "bucket" all the 60*7c478bd9Sstevel@tonic-gate * time, which we don't want to do because one of our big goals in life 61*7c478bd9Sstevel@tonic-gate * is to keep database files small. It's sleazy as hell, but we catch 62*7c478bd9Sstevel@tonic-gate * any attempt to actually write the file in memp_fput(). 63*7c478bd9Sstevel@tonic-gate */ 64*7c478bd9Sstevel@tonic-gate #define OKFLAGS (DB_MPOOL_CREATE | DB_MPOOL_LAST | DB_MPOOL_NEW) 65*7c478bd9Sstevel@tonic-gate if (flags != 0) { 66*7c478bd9Sstevel@tonic-gate if ((ret = 67*7c478bd9Sstevel@tonic-gate __db_fchk(dbmp->dbenv, "memp_fget", flags, OKFLAGS)) != 0) 68*7c478bd9Sstevel@tonic-gate return (ret); 69*7c478bd9Sstevel@tonic-gate 70*7c478bd9Sstevel@tonic-gate switch (flags) { 71*7c478bd9Sstevel@tonic-gate case DB_MPOOL_CREATE: 72*7c478bd9Sstevel@tonic-gate case DB_MPOOL_LAST: 73*7c478bd9Sstevel@tonic-gate case DB_MPOOL_NEW: 74*7c478bd9Sstevel@tonic-gate case 0: 75*7c478bd9Sstevel@tonic-gate break; 76*7c478bd9Sstevel@tonic-gate default: 77*7c478bd9Sstevel@tonic-gate return (__db_ferr(dbmp->dbenv, "memp_fget", 1)); 78*7c478bd9Sstevel@tonic-gate } 79*7c478bd9Sstevel@tonic-gate } 80*7c478bd9Sstevel@tonic-gate 81*7c478bd9Sstevel@tonic-gate #ifdef DIAGNOSTIC 82*7c478bd9Sstevel@tonic-gate /* 83*7c478bd9Sstevel@tonic-gate * XXX 84*7c478bd9Sstevel@tonic-gate * We want to switch threads as often as possible. Yield every time 85*7c478bd9Sstevel@tonic-gate * we get a new page to ensure contention. 86*7c478bd9Sstevel@tonic-gate */ 87*7c478bd9Sstevel@tonic-gate if (DB_GLOBAL(db_pageyield)) 88*7c478bd9Sstevel@tonic-gate __os_yield(1); 89*7c478bd9Sstevel@tonic-gate #endif 90*7c478bd9Sstevel@tonic-gate 91*7c478bd9Sstevel@tonic-gate /* Initialize remaining local variables. */ 92*7c478bd9Sstevel@tonic-gate mf_offset = R_OFFSET(dbmp, mfp); 93*7c478bd9Sstevel@tonic-gate bhp = NULL; 94*7c478bd9Sstevel@tonic-gate st_hsearch = 0; 95*7c478bd9Sstevel@tonic-gate b_incr = ret = 0; 96*7c478bd9Sstevel@tonic-gate 97*7c478bd9Sstevel@tonic-gate /* Determine the hash bucket where this page will live. */ 98*7c478bd9Sstevel@tonic-gate bucket = BUCKET(mp, mf_offset, *pgnoaddr); 99*7c478bd9Sstevel@tonic-gate 100*7c478bd9Sstevel@tonic-gate LOCKREGION(dbmp); 101*7c478bd9Sstevel@tonic-gate 102*7c478bd9Sstevel@tonic-gate /* 103*7c478bd9Sstevel@tonic-gate * Check for the last or last + 1 page requests. 104*7c478bd9Sstevel@tonic-gate * 105*7c478bd9Sstevel@tonic-gate * Examine and update the file's last_pgno value. We don't care if 106*7c478bd9Sstevel@tonic-gate * the last_pgno value immediately changes due to another thread -- 107*7c478bd9Sstevel@tonic-gate * at this instant in time, the value is correct. We do increment the 108*7c478bd9Sstevel@tonic-gate * current last_pgno value if the thread is asking for a new page, 109*7c478bd9Sstevel@tonic-gate * however, to ensure that two threads creating pages don't get the 110*7c478bd9Sstevel@tonic-gate * same one. 111*7c478bd9Sstevel@tonic-gate */ 112*7c478bd9Sstevel@tonic-gate if (LF_ISSET(DB_MPOOL_LAST | DB_MPOOL_NEW)) { 113*7c478bd9Sstevel@tonic-gate if (LF_ISSET(DB_MPOOL_NEW)) 114*7c478bd9Sstevel@tonic-gate ++mfp->last_pgno; 115*7c478bd9Sstevel@tonic-gate *pgnoaddr = mfp->last_pgno; 116*7c478bd9Sstevel@tonic-gate bucket = BUCKET(mp, mf_offset, mfp->last_pgno); 117*7c478bd9Sstevel@tonic-gate 118*7c478bd9Sstevel@tonic-gate if (LF_ISSET(DB_MPOOL_NEW)) 119*7c478bd9Sstevel@tonic-gate goto alloc; 120*7c478bd9Sstevel@tonic-gate } 121*7c478bd9Sstevel@tonic-gate 122*7c478bd9Sstevel@tonic-gate /* 123*7c478bd9Sstevel@tonic-gate * If mmap'ing the file and the page is not past the end of the file, 124*7c478bd9Sstevel@tonic-gate * just return a pointer. 125*7c478bd9Sstevel@tonic-gate * 126*7c478bd9Sstevel@tonic-gate * The page may be past the end of the file, so check the page number 127*7c478bd9Sstevel@tonic-gate * argument against the original length of the file. If we previously 128*7c478bd9Sstevel@tonic-gate * returned pages past the original end of the file, last_pgno will 129*7c478bd9Sstevel@tonic-gate * have been updated to match the "new" end of the file, and checking 130*7c478bd9Sstevel@tonic-gate * against it would return pointers past the end of the mmap'd region. 131*7c478bd9Sstevel@tonic-gate * 132*7c478bd9Sstevel@tonic-gate * If another process has opened the file for writing since we mmap'd 133*7c478bd9Sstevel@tonic-gate * it, we will start playing the game by their rules, i.e. everything 134*7c478bd9Sstevel@tonic-gate * goes through the cache. All pages previously returned will be safe, 135*7c478bd9Sstevel@tonic-gate * as long as the correct locking protocol was observed. 136*7c478bd9Sstevel@tonic-gate * 137*7c478bd9Sstevel@tonic-gate * XXX 138*7c478bd9Sstevel@tonic-gate * We don't discard the map because we don't know when all of the 139*7c478bd9Sstevel@tonic-gate * pages will have been discarded from the process' address space. 140*7c478bd9Sstevel@tonic-gate * It would be possible to do so by reference counting the open 141*7c478bd9Sstevel@tonic-gate * pages from the mmap, but it's unclear to me that it's worth it. 142*7c478bd9Sstevel@tonic-gate */ 143*7c478bd9Sstevel@tonic-gate if (dbmfp->addr != NULL && F_ISSET(mfp, MP_CAN_MMAP)) 144*7c478bd9Sstevel@tonic-gate if (*pgnoaddr > mfp->orig_last_pgno) { 145*7c478bd9Sstevel@tonic-gate /* 146*7c478bd9Sstevel@tonic-gate * !!! 147*7c478bd9Sstevel@tonic-gate * See the comment above about non-existent pages and 148*7c478bd9Sstevel@tonic-gate * the hash access method. 149*7c478bd9Sstevel@tonic-gate */ 150*7c478bd9Sstevel@tonic-gate if (!LF_ISSET(DB_MPOOL_CREATE)) { 151*7c478bd9Sstevel@tonic-gate __db_err(dbmp->dbenv, 152*7c478bd9Sstevel@tonic-gate "%s: page %lu doesn't exist", 153*7c478bd9Sstevel@tonic-gate __memp_fn(dbmfp), (u_long)*pgnoaddr); 154*7c478bd9Sstevel@tonic-gate ret = EINVAL; 155*7c478bd9Sstevel@tonic-gate goto err; 156*7c478bd9Sstevel@tonic-gate } 157*7c478bd9Sstevel@tonic-gate } else { 158*7c478bd9Sstevel@tonic-gate *(void **)addrp = 159*7c478bd9Sstevel@tonic-gate R_ADDR(dbmfp, *pgnoaddr * mfp->stat.st_pagesize); 160*7c478bd9Sstevel@tonic-gate ++mp->stat.st_map; 161*7c478bd9Sstevel@tonic-gate ++mfp->stat.st_map; 162*7c478bd9Sstevel@tonic-gate goto done; 163*7c478bd9Sstevel@tonic-gate } 164*7c478bd9Sstevel@tonic-gate 165*7c478bd9Sstevel@tonic-gate /* Search the hash chain for the page. */ 166*7c478bd9Sstevel@tonic-gate for (bhp = SH_TAILQ_FIRST(&dbmp->htab[bucket], __bh); 167*7c478bd9Sstevel@tonic-gate bhp != NULL; bhp = SH_TAILQ_NEXT(bhp, hq, __bh)) { 168*7c478bd9Sstevel@tonic-gate ++st_hsearch; 169*7c478bd9Sstevel@tonic-gate if (bhp->pgno != *pgnoaddr || bhp->mf_offset != mf_offset) 170*7c478bd9Sstevel@tonic-gate continue; 171*7c478bd9Sstevel@tonic-gate 172*7c478bd9Sstevel@tonic-gate /* Increment the reference count. */ 173*7c478bd9Sstevel@tonic-gate if (bhp->ref == UINT16_T_MAX) { 174*7c478bd9Sstevel@tonic-gate __db_err(dbmp->dbenv, 175*7c478bd9Sstevel@tonic-gate "%s: page %lu: reference count overflow", 176*7c478bd9Sstevel@tonic-gate __memp_fn(dbmfp), (u_long)bhp->pgno); 177*7c478bd9Sstevel@tonic-gate ret = EINVAL; 178*7c478bd9Sstevel@tonic-gate goto err; 179*7c478bd9Sstevel@tonic-gate } 180*7c478bd9Sstevel@tonic-gate 181*7c478bd9Sstevel@tonic-gate /* 182*7c478bd9Sstevel@tonic-gate * Increment the reference count. We may discard the region 183*7c478bd9Sstevel@tonic-gate * lock as we evaluate and/or read the buffer, so we need to 184*7c478bd9Sstevel@tonic-gate * ensure that it doesn't move and that its contents remain 185*7c478bd9Sstevel@tonic-gate * unchanged. 186*7c478bd9Sstevel@tonic-gate */ 187*7c478bd9Sstevel@tonic-gate ++bhp->ref; 188*7c478bd9Sstevel@tonic-gate b_incr = 1; 189*7c478bd9Sstevel@tonic-gate 190*7c478bd9Sstevel@tonic-gate /* 191*7c478bd9Sstevel@tonic-gate * Any buffer we find might be trouble. 192*7c478bd9Sstevel@tonic-gate * 193*7c478bd9Sstevel@tonic-gate * BH_LOCKED -- 194*7c478bd9Sstevel@tonic-gate * I/O is in progress. Because we've incremented the buffer 195*7c478bd9Sstevel@tonic-gate * reference count, we know the buffer can't move. Unlock 196*7c478bd9Sstevel@tonic-gate * the region lock, wait for the I/O to complete, and reacquire 197*7c478bd9Sstevel@tonic-gate * the region. 198*7c478bd9Sstevel@tonic-gate */ 199*7c478bd9Sstevel@tonic-gate for (first = 1; F_ISSET(bhp, BH_LOCKED); first = 0) { 200*7c478bd9Sstevel@tonic-gate UNLOCKREGION(dbmp); 201*7c478bd9Sstevel@tonic-gate 202*7c478bd9Sstevel@tonic-gate /* 203*7c478bd9Sstevel@tonic-gate * Explicitly yield the processor if it's not the first 204*7c478bd9Sstevel@tonic-gate * pass through this loop -- if we don't, we might end 205*7c478bd9Sstevel@tonic-gate * up running to the end of our CPU quantum as we will 206*7c478bd9Sstevel@tonic-gate * simply be swapping between the two locks. 207*7c478bd9Sstevel@tonic-gate */ 208*7c478bd9Sstevel@tonic-gate if (!first) 209*7c478bd9Sstevel@tonic-gate __os_yield(1); 210*7c478bd9Sstevel@tonic-gate 211*7c478bd9Sstevel@tonic-gate LOCKBUFFER(dbmp, bhp); 212*7c478bd9Sstevel@tonic-gate /* Wait for I/O to finish... */ 213*7c478bd9Sstevel@tonic-gate UNLOCKBUFFER(dbmp, bhp); 214*7c478bd9Sstevel@tonic-gate LOCKREGION(dbmp); 215*7c478bd9Sstevel@tonic-gate } 216*7c478bd9Sstevel@tonic-gate 217*7c478bd9Sstevel@tonic-gate /* 218*7c478bd9Sstevel@tonic-gate * BH_TRASH -- 219*7c478bd9Sstevel@tonic-gate * The contents of the buffer are garbage. Shouldn't happen, 220*7c478bd9Sstevel@tonic-gate * and this read is likely to fail, but might as well try. 221*7c478bd9Sstevel@tonic-gate */ 222*7c478bd9Sstevel@tonic-gate if (F_ISSET(bhp, BH_TRASH)) 223*7c478bd9Sstevel@tonic-gate goto reread; 224*7c478bd9Sstevel@tonic-gate 225*7c478bd9Sstevel@tonic-gate /* 226*7c478bd9Sstevel@tonic-gate * BH_CALLPGIN -- 227*7c478bd9Sstevel@tonic-gate * The buffer was converted so it could be written, and the 228*7c478bd9Sstevel@tonic-gate * contents need to be converted again. 229*7c478bd9Sstevel@tonic-gate */ 230*7c478bd9Sstevel@tonic-gate if (F_ISSET(bhp, BH_CALLPGIN)) { 231*7c478bd9Sstevel@tonic-gate if ((ret = __memp_pg(dbmfp, bhp, 1)) != 0) 232*7c478bd9Sstevel@tonic-gate goto err; 233*7c478bd9Sstevel@tonic-gate F_CLR(bhp, BH_CALLPGIN); 234*7c478bd9Sstevel@tonic-gate } 235*7c478bd9Sstevel@tonic-gate 236*7c478bd9Sstevel@tonic-gate ++mp->stat.st_cache_hit; 237*7c478bd9Sstevel@tonic-gate ++mfp->stat.st_cache_hit; 238*7c478bd9Sstevel@tonic-gate *(void **)addrp = bhp->buf; 239*7c478bd9Sstevel@tonic-gate goto done; 240*7c478bd9Sstevel@tonic-gate } 241*7c478bd9Sstevel@tonic-gate 242*7c478bd9Sstevel@tonic-gate alloc: /* Allocate new buffer header and data space. */ 243*7c478bd9Sstevel@tonic-gate if ((ret = __memp_alloc(dbmp, sizeof(BH) - 244*7c478bd9Sstevel@tonic-gate sizeof(u_int8_t) + mfp->stat.st_pagesize, NULL, &bhp)) != 0) 245*7c478bd9Sstevel@tonic-gate goto err; 246*7c478bd9Sstevel@tonic-gate 247*7c478bd9Sstevel@tonic-gate #ifdef DIAGNOSTIC 248*7c478bd9Sstevel@tonic-gate if ((ALIGNTYPE)bhp->buf & (sizeof(size_t) - 1)) { 249*7c478bd9Sstevel@tonic-gate __db_err(dbmp->dbenv, 250*7c478bd9Sstevel@tonic-gate "Internal error: BH data NOT size_t aligned."); 251*7c478bd9Sstevel@tonic-gate ret = EINVAL; 252*7c478bd9Sstevel@tonic-gate goto err; 253*7c478bd9Sstevel@tonic-gate } 254*7c478bd9Sstevel@tonic-gate #endif 255*7c478bd9Sstevel@tonic-gate /* Initialize the BH fields. */ 256*7c478bd9Sstevel@tonic-gate memset(bhp, 0, sizeof(BH)); 257*7c478bd9Sstevel@tonic-gate LOCKINIT(dbmp, &bhp->mutex); 258*7c478bd9Sstevel@tonic-gate bhp->ref = 1; 259*7c478bd9Sstevel@tonic-gate bhp->pgno = *pgnoaddr; 260*7c478bd9Sstevel@tonic-gate bhp->mf_offset = mf_offset; 261*7c478bd9Sstevel@tonic-gate 262*7c478bd9Sstevel@tonic-gate /* 263*7c478bd9Sstevel@tonic-gate * Prepend the bucket header to the head of the appropriate MPOOL 264*7c478bd9Sstevel@tonic-gate * bucket hash list. Append the bucket header to the tail of the 265*7c478bd9Sstevel@tonic-gate * MPOOL LRU chain. 266*7c478bd9Sstevel@tonic-gate */ 267*7c478bd9Sstevel@tonic-gate SH_TAILQ_INSERT_HEAD(&dbmp->htab[bucket], bhp, hq, __bh); 268*7c478bd9Sstevel@tonic-gate SH_TAILQ_INSERT_TAIL(&mp->bhq, bhp, q); 269*7c478bd9Sstevel@tonic-gate 270*7c478bd9Sstevel@tonic-gate /* 271*7c478bd9Sstevel@tonic-gate * If we created the page, zero it out and continue. 272*7c478bd9Sstevel@tonic-gate * 273*7c478bd9Sstevel@tonic-gate * !!! 274*7c478bd9Sstevel@tonic-gate * Note: DB_MPOOL_NEW specifically doesn't call the pgin function. 275*7c478bd9Sstevel@tonic-gate * If DB_MPOOL_CREATE is used, then the application's pgin function 276*7c478bd9Sstevel@tonic-gate * has to be able to handle pages of 0's -- if it uses DB_MPOOL_NEW, 277*7c478bd9Sstevel@tonic-gate * it can detect all of its page creates, and not bother. 278*7c478bd9Sstevel@tonic-gate * 279*7c478bd9Sstevel@tonic-gate * Otherwise, read the page into memory, optionally creating it if 280*7c478bd9Sstevel@tonic-gate * DB_MPOOL_CREATE is set. 281*7c478bd9Sstevel@tonic-gate */ 282*7c478bd9Sstevel@tonic-gate if (LF_ISSET(DB_MPOOL_NEW)) { 283*7c478bd9Sstevel@tonic-gate if (mfp->clear_len == 0) 284*7c478bd9Sstevel@tonic-gate memset(bhp->buf, 0, mfp->stat.st_pagesize); 285*7c478bd9Sstevel@tonic-gate else { 286*7c478bd9Sstevel@tonic-gate memset(bhp->buf, 0, mfp->clear_len); 287*7c478bd9Sstevel@tonic-gate #ifdef DIAGNOSTIC 288*7c478bd9Sstevel@tonic-gate memset(bhp->buf + mfp->clear_len, 0xdb, 289*7c478bd9Sstevel@tonic-gate mfp->stat.st_pagesize - mfp->clear_len); 290*7c478bd9Sstevel@tonic-gate #endif 291*7c478bd9Sstevel@tonic-gate } 292*7c478bd9Sstevel@tonic-gate 293*7c478bd9Sstevel@tonic-gate ++mp->stat.st_page_create; 294*7c478bd9Sstevel@tonic-gate ++mfp->stat.st_page_create; 295*7c478bd9Sstevel@tonic-gate } else { 296*7c478bd9Sstevel@tonic-gate /* 297*7c478bd9Sstevel@tonic-gate * It's possible for the read function to fail, which means 298*7c478bd9Sstevel@tonic-gate * that we fail as well. Note, the __memp_pgread() function 299*7c478bd9Sstevel@tonic-gate * discards the region lock, so the buffer must be pinned 300*7c478bd9Sstevel@tonic-gate * down so that it cannot move and its contents are unchanged. 301*7c478bd9Sstevel@tonic-gate */ 302*7c478bd9Sstevel@tonic-gate reread: if ((ret = __memp_pgread(dbmfp, 303*7c478bd9Sstevel@tonic-gate bhp, LF_ISSET(DB_MPOOL_CREATE))) != 0) { 304*7c478bd9Sstevel@tonic-gate /* 305*7c478bd9Sstevel@tonic-gate * !!! 306*7c478bd9Sstevel@tonic-gate * Discard the buffer unless another thread is waiting 307*7c478bd9Sstevel@tonic-gate * on our I/O to complete. Regardless, the header has 308*7c478bd9Sstevel@tonic-gate * the BH_TRASH flag set. 309*7c478bd9Sstevel@tonic-gate */ 310*7c478bd9Sstevel@tonic-gate if (bhp->ref == 1) 311*7c478bd9Sstevel@tonic-gate __memp_bhfree(dbmp, mfp, bhp, 1); 312*7c478bd9Sstevel@tonic-gate goto err; 313*7c478bd9Sstevel@tonic-gate } 314*7c478bd9Sstevel@tonic-gate 315*7c478bd9Sstevel@tonic-gate ++mp->stat.st_cache_miss; 316*7c478bd9Sstevel@tonic-gate ++mfp->stat.st_cache_miss; 317*7c478bd9Sstevel@tonic-gate } 318*7c478bd9Sstevel@tonic-gate 319*7c478bd9Sstevel@tonic-gate /* 320*7c478bd9Sstevel@tonic-gate * If we're returning a page after our current notion of the last-page, 321*7c478bd9Sstevel@tonic-gate * update our information. Note, there's no way to un-instantiate this 322*7c478bd9Sstevel@tonic-gate * page, it's going to exist whether it's returned to us dirty or not. 323*7c478bd9Sstevel@tonic-gate */ 324*7c478bd9Sstevel@tonic-gate if (bhp->pgno > mfp->last_pgno) 325*7c478bd9Sstevel@tonic-gate mfp->last_pgno = bhp->pgno; 326*7c478bd9Sstevel@tonic-gate 327*7c478bd9Sstevel@tonic-gate ++mp->stat.st_page_clean; 328*7c478bd9Sstevel@tonic-gate *(void **)addrp = bhp->buf; 329*7c478bd9Sstevel@tonic-gate 330*7c478bd9Sstevel@tonic-gate done: /* Update the chain search statistics. */ 331*7c478bd9Sstevel@tonic-gate if (st_hsearch) { 332*7c478bd9Sstevel@tonic-gate ++mp->stat.st_hash_searches; 333*7c478bd9Sstevel@tonic-gate if (st_hsearch > mp->stat.st_hash_longest) 334*7c478bd9Sstevel@tonic-gate mp->stat.st_hash_longest = st_hsearch; 335*7c478bd9Sstevel@tonic-gate mp->stat.st_hash_examined += st_hsearch; 336*7c478bd9Sstevel@tonic-gate } 337*7c478bd9Sstevel@tonic-gate 338*7c478bd9Sstevel@tonic-gate ++dbmfp->pinref; 339*7c478bd9Sstevel@tonic-gate 340*7c478bd9Sstevel@tonic-gate UNLOCKREGION(dbmp); 341*7c478bd9Sstevel@tonic-gate 342*7c478bd9Sstevel@tonic-gate return (0); 343*7c478bd9Sstevel@tonic-gate 344*7c478bd9Sstevel@tonic-gate err: /* Discard our reference. */ 345*7c478bd9Sstevel@tonic-gate if (b_incr) 346*7c478bd9Sstevel@tonic-gate --bhp->ref; 347*7c478bd9Sstevel@tonic-gate UNLOCKREGION(dbmp); 348*7c478bd9Sstevel@tonic-gate 349*7c478bd9Sstevel@tonic-gate *(void **)addrp = NULL; 350*7c478bd9Sstevel@tonic-gate return (ret); 351*7c478bd9Sstevel@tonic-gate } 352