17c478bd9Sstevel@tonic-gate /* 27c478bd9Sstevel@tonic-gate * CDDL HEADER START 37c478bd9Sstevel@tonic-gate * 47c478bd9Sstevel@tonic-gate * The contents of this file are subject to the terms of the 5*80d34432Sfrankho * Common Development and Distribution License (the "License"). 6*80d34432Sfrankho * You may not use this file except in compliance with the License. 77c478bd9Sstevel@tonic-gate * 87c478bd9Sstevel@tonic-gate * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 97c478bd9Sstevel@tonic-gate * or http://www.opensolaris.org/os/licensing. 107c478bd9Sstevel@tonic-gate * See the License for the specific language governing permissions 117c478bd9Sstevel@tonic-gate * and limitations under the License. 127c478bd9Sstevel@tonic-gate * 137c478bd9Sstevel@tonic-gate * When distributing Covered Code, include this CDDL HEADER in each 147c478bd9Sstevel@tonic-gate * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 157c478bd9Sstevel@tonic-gate * If applicable, add the following below this CDDL HEADER, with the 167c478bd9Sstevel@tonic-gate * fields enclosed by brackets "[]" replaced with your own identifying 177c478bd9Sstevel@tonic-gate * information: Portions Copyright [yyyy] [name of copyright owner] 187c478bd9Sstevel@tonic-gate * 197c478bd9Sstevel@tonic-gate * CDDL HEADER END 207c478bd9Sstevel@tonic-gate */ 217c478bd9Sstevel@tonic-gate /* 22*80d34432Sfrankho * Copyright 2007 Sun Microsystems, Inc. All rights reserved. 237c478bd9Sstevel@tonic-gate * Use is subject to license terms. 247c478bd9Sstevel@tonic-gate */ 257c478bd9Sstevel@tonic-gate 267c478bd9Sstevel@tonic-gate #pragma ident "%Z%%M% %I% %E% SMI" 277c478bd9Sstevel@tonic-gate 287c478bd9Sstevel@tonic-gate #include <sys/systm.h> 297c478bd9Sstevel@tonic-gate #include <sys/types.h> 307c478bd9Sstevel@tonic-gate #include <sys/vnode.h> 317c478bd9Sstevel@tonic-gate #include <sys/errno.h> 327c478bd9Sstevel@tonic-gate #include <sys/sysmacros.h> 337c478bd9Sstevel@tonic-gate #include <sys/debug.h> 347c478bd9Sstevel@tonic-gate #include <sys/kmem.h> 357c478bd9Sstevel@tonic-gate #include <sys/conf.h> 367c478bd9Sstevel@tonic-gate #include <sys/proc.h> 377c478bd9Sstevel@tonic-gate #include <sys/taskq.h> 387c478bd9Sstevel@tonic-gate #include <sys/cmn_err.h> 397c478bd9Sstevel@tonic-gate #include <sys/fs/ufs_inode.h> 407c478bd9Sstevel@tonic-gate #include <sys/fs/ufs_filio.h> 417c478bd9Sstevel@tonic-gate #include <sys/fs/ufs_log.h> 427c478bd9Sstevel@tonic-gate #include <sys/fs/ufs_bio.h> 437c478bd9Sstevel@tonic-gate 447c478bd9Sstevel@tonic-gate /* 457c478bd9Sstevel@tonic-gate * FILE SYSTEM INTERFACE TO TRANSACTION OPERATIONS (TOP; like VOP) 467c478bd9Sstevel@tonic-gate */ 477c478bd9Sstevel@tonic-gate 487c478bd9Sstevel@tonic-gate uint_t topkey; /* tsd transaction key */ 497c478bd9Sstevel@tonic-gate 507c478bd9Sstevel@tonic-gate /* 517c478bd9Sstevel@tonic-gate * declare a delta 527c478bd9Sstevel@tonic-gate */ 537c478bd9Sstevel@tonic-gate void 547c478bd9Sstevel@tonic-gate top_delta( 557c478bd9Sstevel@tonic-gate ufsvfs_t *ufsvfsp, 567c478bd9Sstevel@tonic-gate offset_t mof, 577c478bd9Sstevel@tonic-gate off_t nb, 587c478bd9Sstevel@tonic-gate delta_t dtyp, 597c478bd9Sstevel@tonic-gate int (*func)(), 607c478bd9Sstevel@tonic-gate ulong_t arg) 617c478bd9Sstevel@tonic-gate { 627c478bd9Sstevel@tonic-gate ml_unit_t *ul = ufsvfsp->vfs_log; 637c478bd9Sstevel@tonic-gate threadtrans_t *tp = tsd_get(topkey); 647c478bd9Sstevel@tonic-gate 657c478bd9Sstevel@tonic-gate ASSERT(ufsvfsp->vfs_dev == ul->un_dev); 667c478bd9Sstevel@tonic-gate ASSERT(nb); 677c478bd9Sstevel@tonic-gate ASSERT(((ul->un_debug & (MT_TRANSACT|MT_MATAMAP)) == 0) || 687c478bd9Sstevel@tonic-gate top_delta_debug(ul, mof, nb, dtyp)); 697c478bd9Sstevel@tonic-gate 707c478bd9Sstevel@tonic-gate deltamap_add(ul->un_deltamap, mof, nb, dtyp, func, arg, tp); 717c478bd9Sstevel@tonic-gate 727c478bd9Sstevel@tonic-gate ul->un_logmap->mtm_ref = 1; /* for roll thread's heuristic */ 737c478bd9Sstevel@tonic-gate if (tp) { 747c478bd9Sstevel@tonic-gate tp->any_deltas = 1; 757c478bd9Sstevel@tonic-gate } 767c478bd9Sstevel@tonic-gate } 777c478bd9Sstevel@tonic-gate 787c478bd9Sstevel@tonic-gate /* 797c478bd9Sstevel@tonic-gate * cancel a delta 807c478bd9Sstevel@tonic-gate */ 817c478bd9Sstevel@tonic-gate void 827c478bd9Sstevel@tonic-gate top_cancel(ufsvfs_t *ufsvfsp, offset_t mof, off_t nb, int flags) 837c478bd9Sstevel@tonic-gate { 847c478bd9Sstevel@tonic-gate ml_unit_t *ul = ufsvfsp->vfs_log; 857c478bd9Sstevel@tonic-gate int metadata = flags & (I_DIR|I_IBLK|I_SHAD|I_QUOTA); 867c478bd9Sstevel@tonic-gate 877c478bd9Sstevel@tonic-gate ASSERT(ufsvfsp->vfs_dev == ul->un_dev); 887c478bd9Sstevel@tonic-gate ASSERT(nb); 897c478bd9Sstevel@tonic-gate ASSERT(((ul->un_debug & (MT_TRANSACT|MT_MATAMAP)) == 0) || 907c478bd9Sstevel@tonic-gate (!(flags & metadata) || 917c478bd9Sstevel@tonic-gate top_delta_debug(ul, mof, nb, DT_CANCEL))); 927c478bd9Sstevel@tonic-gate 937c478bd9Sstevel@tonic-gate if (metadata) 947c478bd9Sstevel@tonic-gate deltamap_del(ul->un_deltamap, mof, nb); 957c478bd9Sstevel@tonic-gate 967c478bd9Sstevel@tonic-gate logmap_cancel(ul, mof, nb, metadata); 977c478bd9Sstevel@tonic-gate 987c478bd9Sstevel@tonic-gate /* 997c478bd9Sstevel@tonic-gate * needed for the roll thread's heuristic 1007c478bd9Sstevel@tonic-gate */ 1017c478bd9Sstevel@tonic-gate ul->un_logmap->mtm_ref = 1; 1027c478bd9Sstevel@tonic-gate } 1037c478bd9Sstevel@tonic-gate 1047c478bd9Sstevel@tonic-gate /* 1057c478bd9Sstevel@tonic-gate * check if this delta has been canceled (metadata -> userdata) 1067c478bd9Sstevel@tonic-gate */ 1077c478bd9Sstevel@tonic-gate int 1087c478bd9Sstevel@tonic-gate top_iscancel(ufsvfs_t *ufsvfsp, offset_t mof, off_t nb) 1097c478bd9Sstevel@tonic-gate { 1107c478bd9Sstevel@tonic-gate ml_unit_t *ul = ufsvfsp->vfs_log; 1117c478bd9Sstevel@tonic-gate 1127c478bd9Sstevel@tonic-gate ASSERT(ufsvfsp->vfs_dev == ul->un_dev); 1137c478bd9Sstevel@tonic-gate ASSERT(nb); 1147c478bd9Sstevel@tonic-gate if (logmap_iscancel(ul->un_logmap, mof, nb)) 1157c478bd9Sstevel@tonic-gate return (1); 1167c478bd9Sstevel@tonic-gate if (ul->un_flags & LDL_ERROR) 1177c478bd9Sstevel@tonic-gate return (1); 1187c478bd9Sstevel@tonic-gate return (0); 1197c478bd9Sstevel@tonic-gate } 1207c478bd9Sstevel@tonic-gate 1217c478bd9Sstevel@tonic-gate /* 1227c478bd9Sstevel@tonic-gate * put device into error state 1237c478bd9Sstevel@tonic-gate */ 1247c478bd9Sstevel@tonic-gate void 1257c478bd9Sstevel@tonic-gate top_seterror(ufsvfs_t *ufsvfsp) 1267c478bd9Sstevel@tonic-gate { 1277c478bd9Sstevel@tonic-gate ml_unit_t *ul = ufsvfsp->vfs_log; 1287c478bd9Sstevel@tonic-gate 1297c478bd9Sstevel@tonic-gate ASSERT(ufsvfsp->vfs_dev == ul->un_dev); 1307c478bd9Sstevel@tonic-gate ldl_seterror(ul, "ufs is forcing a ufs log error"); 1317c478bd9Sstevel@tonic-gate } 1327c478bd9Sstevel@tonic-gate 1337c478bd9Sstevel@tonic-gate /* 1347c478bd9Sstevel@tonic-gate * issue a empty sync op to help empty the delta/log map or the log 1357c478bd9Sstevel@tonic-gate */ 1367c478bd9Sstevel@tonic-gate static void 1377c478bd9Sstevel@tonic-gate top_issue_sync(void *arg) 1387c478bd9Sstevel@tonic-gate { 1397c478bd9Sstevel@tonic-gate ufsvfs_t *ufsvfsp = (ufsvfs_t *)arg; 1407c478bd9Sstevel@tonic-gate ml_unit_t *ul = (ml_unit_t *)ufsvfsp->vfs_log; 1417c478bd9Sstevel@tonic-gate mt_map_t *mtm = ul->un_logmap; 1427c478bd9Sstevel@tonic-gate int error = 0; 1437c478bd9Sstevel@tonic-gate 1447c478bd9Sstevel@tonic-gate if ((curthread->t_flag & T_DONTBLOCK) == 0) 1457c478bd9Sstevel@tonic-gate curthread->t_flag |= T_DONTBLOCK; 1467c478bd9Sstevel@tonic-gate top_begin_sync(ufsvfsp, TOP_COMMIT_ASYNC, 0, &error); 1477c478bd9Sstevel@tonic-gate if (!error) { 1487c478bd9Sstevel@tonic-gate top_end_sync(ufsvfsp, &error, TOP_COMMIT_ASYNC, 0); 1497c478bd9Sstevel@tonic-gate } 1507c478bd9Sstevel@tonic-gate 1517c478bd9Sstevel@tonic-gate /* 1527c478bd9Sstevel@tonic-gate * If we are a taskq thread, decrement mtm_taskq_sync_count and 1537c478bd9Sstevel@tonic-gate * wake up the thread waiting on the mtm_cv if the mtm_taskq_sync_count 1547c478bd9Sstevel@tonic-gate * hits zero. 1557c478bd9Sstevel@tonic-gate */ 1567c478bd9Sstevel@tonic-gate 1577c478bd9Sstevel@tonic-gate if (taskq_member(system_taskq, curthread)) { 1587c478bd9Sstevel@tonic-gate mutex_enter(&mtm->mtm_lock); 1597c478bd9Sstevel@tonic-gate mtm->mtm_taskq_sync_count--; 1607c478bd9Sstevel@tonic-gate if (mtm->mtm_taskq_sync_count == 0) { 1617c478bd9Sstevel@tonic-gate cv_signal(&mtm->mtm_cv); 1627c478bd9Sstevel@tonic-gate } 1637c478bd9Sstevel@tonic-gate mutex_exit(&mtm->mtm_lock); 1647c478bd9Sstevel@tonic-gate } 1657c478bd9Sstevel@tonic-gate } 1667c478bd9Sstevel@tonic-gate 1677c478bd9Sstevel@tonic-gate /* 1687c478bd9Sstevel@tonic-gate * MOBY TRANSACTION ROUTINES 1697c478bd9Sstevel@tonic-gate * begin a moby transaction 1707c478bd9Sstevel@tonic-gate * sync ops enter until first sync op finishes 1717c478bd9Sstevel@tonic-gate * async ops enter until last sync op finishes 1727c478bd9Sstevel@tonic-gate * end a moby transaction 1737c478bd9Sstevel@tonic-gate * outstanding deltas are pushed thru log 1747c478bd9Sstevel@tonic-gate * log buffer is committed (incore only) 1757c478bd9Sstevel@tonic-gate * next trans is open to async ops 1767c478bd9Sstevel@tonic-gate * log buffer is committed on the log 1777c478bd9Sstevel@tonic-gate * next trans is open to sync ops 1787c478bd9Sstevel@tonic-gate */ 1797c478bd9Sstevel@tonic-gate 1807c478bd9Sstevel@tonic-gate /*ARGSUSED*/ 1817c478bd9Sstevel@tonic-gate void 1827c478bd9Sstevel@tonic-gate top_begin_sync(ufsvfs_t *ufsvfsp, top_t topid, ulong_t size, int *error) 1837c478bd9Sstevel@tonic-gate { 1847c478bd9Sstevel@tonic-gate ml_unit_t *ul = ufsvfsp->vfs_log; 1857c478bd9Sstevel@tonic-gate mt_map_t *mtm = ul->un_logmap; 1867c478bd9Sstevel@tonic-gate threadtrans_t *tp; 1877c478bd9Sstevel@tonic-gate ushort_t seq; 1887c478bd9Sstevel@tonic-gate 1897c478bd9Sstevel@tonic-gate ASSERT(ufsvfsp->vfs_dev == ul->un_dev); 1907c478bd9Sstevel@tonic-gate ASSERT(error != NULL); 1917c478bd9Sstevel@tonic-gate ASSERT(*error == 0); 1927c478bd9Sstevel@tonic-gate 1937c478bd9Sstevel@tonic-gate mutex_enter(&mtm->mtm_lock); 1947c478bd9Sstevel@tonic-gate if (topid == TOP_FSYNC) { 1957c478bd9Sstevel@tonic-gate /* 1967c478bd9Sstevel@tonic-gate * Error the fsync immediately if this is an nfs thread 1977c478bd9Sstevel@tonic-gate * and its last transaction has already been committed. 1987c478bd9Sstevel@tonic-gate * The only transactions outstanding are those 1997c478bd9Sstevel@tonic-gate * where no commit has even started 2007c478bd9Sstevel@tonic-gate * (last_async_tid == mtm->mtm_tid) 2017c478bd9Sstevel@tonic-gate * or those where a commit is in progress 2027c478bd9Sstevel@tonic-gate * (last_async_tid == mtm->mtm_committid) 2037c478bd9Sstevel@tonic-gate */ 2047c478bd9Sstevel@tonic-gate if (curthread->t_flag & T_DONTPEND) { 2057c478bd9Sstevel@tonic-gate tp = tsd_get(topkey); 2067c478bd9Sstevel@tonic-gate if (tp && (tp->last_async_tid != mtm->mtm_tid) && 2077c478bd9Sstevel@tonic-gate (tp->last_async_tid != mtm->mtm_committid)) { 2087c478bd9Sstevel@tonic-gate mutex_exit(&mtm->mtm_lock); 2097c478bd9Sstevel@tonic-gate *error = 1; 2107c478bd9Sstevel@tonic-gate return; 2117c478bd9Sstevel@tonic-gate } 2127c478bd9Sstevel@tonic-gate } 2137c478bd9Sstevel@tonic-gate 2147c478bd9Sstevel@tonic-gate /* 2157c478bd9Sstevel@tonic-gate * If there's already other synchronous transactions 2167c478bd9Sstevel@tonic-gate * and we haven't allowed async ones to start yet 2177c478bd9Sstevel@tonic-gate * then just wait for the commit to complete. 2187c478bd9Sstevel@tonic-gate */ 2197c478bd9Sstevel@tonic-gate if (((mtm->mtm_closed & (TOP_SYNC | TOP_ASYNC)) == 2207c478bd9Sstevel@tonic-gate (TOP_SYNC | TOP_ASYNC)) || mtm->mtm_activesync) { 2217c478bd9Sstevel@tonic-gate seq = mtm->mtm_seq; 2227c478bd9Sstevel@tonic-gate do { 2237c478bd9Sstevel@tonic-gate cv_wait(&mtm->mtm_cv_commit, &mtm->mtm_lock); 2247c478bd9Sstevel@tonic-gate } while (seq == mtm->mtm_seq); 2257c478bd9Sstevel@tonic-gate mutex_exit(&mtm->mtm_lock); 2267c478bd9Sstevel@tonic-gate *error = 1; 2277c478bd9Sstevel@tonic-gate return; 2287c478bd9Sstevel@tonic-gate } 2297c478bd9Sstevel@tonic-gate if (mtm->mtm_closed & TOP_SYNC) { 2307c478bd9Sstevel@tonic-gate /* 2317c478bd9Sstevel@tonic-gate * We know we're in the window where a thread is 2327c478bd9Sstevel@tonic-gate * committing a transaction in top_end_sync() and 2337c478bd9Sstevel@tonic-gate * has allowed async threads to start but hasn't 2347c478bd9Sstevel@tonic-gate * got the completion on the commit write to 2357c478bd9Sstevel@tonic-gate * allow sync threads to start. 2367c478bd9Sstevel@tonic-gate * So wait for that commit completion then retest 2377c478bd9Sstevel@tonic-gate * for the quick nfs check and if that fails 2387c478bd9Sstevel@tonic-gate * go on to start a transaction 2397c478bd9Sstevel@tonic-gate */ 2407c478bd9Sstevel@tonic-gate seq = mtm->mtm_seq; 2417c478bd9Sstevel@tonic-gate do { 2427c478bd9Sstevel@tonic-gate cv_wait(&mtm->mtm_cv_commit, &mtm->mtm_lock); 2437c478bd9Sstevel@tonic-gate } while (seq == mtm->mtm_seq); 2447c478bd9Sstevel@tonic-gate 2457c478bd9Sstevel@tonic-gate /* tp is set above if T_DONTPEND */ 2467c478bd9Sstevel@tonic-gate if ((curthread->t_flag & T_DONTPEND) && tp && 2477c478bd9Sstevel@tonic-gate (tp->last_async_tid != mtm->mtm_tid) && 2487c478bd9Sstevel@tonic-gate (tp->last_async_tid != mtm->mtm_committid)) { 2497c478bd9Sstevel@tonic-gate mutex_exit(&mtm->mtm_lock); 2507c478bd9Sstevel@tonic-gate *error = 1; 2517c478bd9Sstevel@tonic-gate return; 2527c478bd9Sstevel@tonic-gate } 2537c478bd9Sstevel@tonic-gate } 2547c478bd9Sstevel@tonic-gate } 2557c478bd9Sstevel@tonic-gate retry: 2567c478bd9Sstevel@tonic-gate mtm->mtm_ref = 1; 2577c478bd9Sstevel@tonic-gate /* 2587c478bd9Sstevel@tonic-gate * current transaction closed to sync ops; try for next transaction 2597c478bd9Sstevel@tonic-gate */ 2607c478bd9Sstevel@tonic-gate if ((mtm->mtm_closed & TOP_SYNC) && !panicstr) { 2617c478bd9Sstevel@tonic-gate ulong_t resv; 2627c478bd9Sstevel@tonic-gate 2637c478bd9Sstevel@tonic-gate /* 2647c478bd9Sstevel@tonic-gate * We know a commit is in progress, if we are trying to 2657c478bd9Sstevel@tonic-gate * commit and we haven't allowed async ones to start yet, 2667c478bd9Sstevel@tonic-gate * then just wait for the commit completion 2677c478bd9Sstevel@tonic-gate */ 2687c478bd9Sstevel@tonic-gate if ((size == TOP_COMMIT_SIZE) && 2697c478bd9Sstevel@tonic-gate (((mtm->mtm_closed & (TOP_SYNC | TOP_ASYNC)) == 2707c478bd9Sstevel@tonic-gate (TOP_SYNC | TOP_ASYNC)) || (mtm->mtm_activesync))) { 2717c478bd9Sstevel@tonic-gate seq = mtm->mtm_seq; 2727c478bd9Sstevel@tonic-gate do { 2737c478bd9Sstevel@tonic-gate cv_wait(&mtm->mtm_cv_commit, &mtm->mtm_lock); 2747c478bd9Sstevel@tonic-gate } while (seq == mtm->mtm_seq); 2757c478bd9Sstevel@tonic-gate mutex_exit(&mtm->mtm_lock); 2767c478bd9Sstevel@tonic-gate *error = 1; 2777c478bd9Sstevel@tonic-gate return; 2787c478bd9Sstevel@tonic-gate } 2797c478bd9Sstevel@tonic-gate 2807c478bd9Sstevel@tonic-gate /* 2817c478bd9Sstevel@tonic-gate * next transaction is full; try for next transaction 2827c478bd9Sstevel@tonic-gate */ 2837c478bd9Sstevel@tonic-gate resv = size + ul->un_resv_wantin + ul->un_resv; 2847c478bd9Sstevel@tonic-gate if (resv > ul->un_maxresv) { 2857c478bd9Sstevel@tonic-gate cv_wait(&mtm->mtm_cv_commit, &mtm->mtm_lock); 2867c478bd9Sstevel@tonic-gate goto retry; 2877c478bd9Sstevel@tonic-gate } 2887c478bd9Sstevel@tonic-gate /* 2897c478bd9Sstevel@tonic-gate * we are in the next transaction; wait for it to start 2907c478bd9Sstevel@tonic-gate */ 2917c478bd9Sstevel@tonic-gate mtm->mtm_wantin++; 2927c478bd9Sstevel@tonic-gate ul->un_resv_wantin += size; 2937c478bd9Sstevel@tonic-gate /* 2947c478bd9Sstevel@tonic-gate * The corresponding cv_broadcast wakes up 2957c478bd9Sstevel@tonic-gate * all threads that have been validated to go into 2967c478bd9Sstevel@tonic-gate * the next transaction. However, because spurious 2977c478bd9Sstevel@tonic-gate * cv_wait wakeups are possible we use a sequence 2987c478bd9Sstevel@tonic-gate * number to check that the commit and cv_broadcast 2997c478bd9Sstevel@tonic-gate * has really occurred. We couldn't use mtm_tid 3007c478bd9Sstevel@tonic-gate * because on error that doesn't get incremented. 3017c478bd9Sstevel@tonic-gate */ 3027c478bd9Sstevel@tonic-gate seq = mtm->mtm_seq; 3037c478bd9Sstevel@tonic-gate do { 3047c478bd9Sstevel@tonic-gate cv_wait(&mtm->mtm_cv_commit, &mtm->mtm_lock); 3057c478bd9Sstevel@tonic-gate } while (seq == mtm->mtm_seq); 3067c478bd9Sstevel@tonic-gate } else { 3077c478bd9Sstevel@tonic-gate /* 3087c478bd9Sstevel@tonic-gate * if the current transaction is full; try the next one 3097c478bd9Sstevel@tonic-gate */ 3107c478bd9Sstevel@tonic-gate if (size && (ul->un_resv && ((size + ul->un_resv) > 3117c478bd9Sstevel@tonic-gate ul->un_maxresv)) && !panicstr) { 3127c478bd9Sstevel@tonic-gate /* 3137c478bd9Sstevel@tonic-gate * log is over reserved and no one will unresv the space 3147c478bd9Sstevel@tonic-gate * so generate empty sync op to unresv the space 3157c478bd9Sstevel@tonic-gate */ 3167c478bd9Sstevel@tonic-gate if (mtm->mtm_activesync == 0) { 3177c478bd9Sstevel@tonic-gate mutex_exit(&mtm->mtm_lock); 3187c478bd9Sstevel@tonic-gate top_issue_sync(ufsvfsp); 3197c478bd9Sstevel@tonic-gate mutex_enter(&mtm->mtm_lock); 3207c478bd9Sstevel@tonic-gate goto retry; 3217c478bd9Sstevel@tonic-gate } 3227c478bd9Sstevel@tonic-gate cv_wait(&mtm->mtm_cv_commit, &mtm->mtm_lock); 3237c478bd9Sstevel@tonic-gate goto retry; 3247c478bd9Sstevel@tonic-gate } 3257c478bd9Sstevel@tonic-gate /* 3267c478bd9Sstevel@tonic-gate * we are in the current transaction 3277c478bd9Sstevel@tonic-gate */ 3287c478bd9Sstevel@tonic-gate mtm->mtm_active++; 3297c478bd9Sstevel@tonic-gate mtm->mtm_activesync++; 3307c478bd9Sstevel@tonic-gate ul->un_resv += size; 3317c478bd9Sstevel@tonic-gate } 3327c478bd9Sstevel@tonic-gate 3337c478bd9Sstevel@tonic-gate ASSERT(mtm->mtm_active > 0); 3347c478bd9Sstevel@tonic-gate ASSERT(mtm->mtm_activesync > 0); 3357c478bd9Sstevel@tonic-gate mutex_exit(&mtm->mtm_lock); 3367c478bd9Sstevel@tonic-gate 3377c478bd9Sstevel@tonic-gate ASSERT(((ul->un_debug & MT_TRANSACT) == 0) || 3387c478bd9Sstevel@tonic-gate top_begin_debug(ul, topid, size)); 3397c478bd9Sstevel@tonic-gate } 3407c478bd9Sstevel@tonic-gate 3417c478bd9Sstevel@tonic-gate int tryfail_cnt; 3427c478bd9Sstevel@tonic-gate 3437c478bd9Sstevel@tonic-gate int 3447c478bd9Sstevel@tonic-gate top_begin_async(ufsvfs_t *ufsvfsp, top_t topid, ulong_t size, int tryasync) 3457c478bd9Sstevel@tonic-gate { 3467c478bd9Sstevel@tonic-gate ml_unit_t *ul = ufsvfsp->vfs_log; 3477c478bd9Sstevel@tonic-gate mt_map_t *mtm = ul->un_logmap; 3487c478bd9Sstevel@tonic-gate threadtrans_t *tp; 3497c478bd9Sstevel@tonic-gate 3507c478bd9Sstevel@tonic-gate ASSERT(ufsvfsp->vfs_dev == ul->un_dev); 3517c478bd9Sstevel@tonic-gate 3527c478bd9Sstevel@tonic-gate tp = tsd_get(topkey); 3537c478bd9Sstevel@tonic-gate if (tp == NULL) { 3547c478bd9Sstevel@tonic-gate tp = kmem_zalloc(sizeof (threadtrans_t), KM_SLEEP); 3557c478bd9Sstevel@tonic-gate (void) tsd_set(topkey, tp); 3567c478bd9Sstevel@tonic-gate } 3577c478bd9Sstevel@tonic-gate tp->deltas_size = 0; 3587c478bd9Sstevel@tonic-gate tp->any_deltas = 0; 3597c478bd9Sstevel@tonic-gate 3607c478bd9Sstevel@tonic-gate mutex_enter(&mtm->mtm_lock); 3617c478bd9Sstevel@tonic-gate retry: 3627c478bd9Sstevel@tonic-gate mtm->mtm_ref = 1; 3637c478bd9Sstevel@tonic-gate /* 3647c478bd9Sstevel@tonic-gate * current transaction closed to async ops; try for next transaction 3657c478bd9Sstevel@tonic-gate */ 3667c478bd9Sstevel@tonic-gate if ((mtm->mtm_closed & TOP_ASYNC) && !panicstr) { 3677c478bd9Sstevel@tonic-gate if (tryasync) { 3687c478bd9Sstevel@tonic-gate mutex_exit(&mtm->mtm_lock); 3697c478bd9Sstevel@tonic-gate tryfail_cnt++; 3707c478bd9Sstevel@tonic-gate return (EWOULDBLOCK); 3717c478bd9Sstevel@tonic-gate } 3727c478bd9Sstevel@tonic-gate cv_wait(&mtm->mtm_cv_next, &mtm->mtm_lock); 3737c478bd9Sstevel@tonic-gate goto retry; 3747c478bd9Sstevel@tonic-gate } 3757c478bd9Sstevel@tonic-gate 3767c478bd9Sstevel@tonic-gate /* 3777c478bd9Sstevel@tonic-gate * if the current transaction is full; try the next one 3787c478bd9Sstevel@tonic-gate */ 3797c478bd9Sstevel@tonic-gate if (((size + ul->un_resv + ul->un_resv_wantin) > ul->un_maxresv) && 3807c478bd9Sstevel@tonic-gate !panicstr) { 3817c478bd9Sstevel@tonic-gate /* 3827c478bd9Sstevel@tonic-gate * log is overreserved and no one will unresv the space 3837c478bd9Sstevel@tonic-gate * so generate empty sync op to unresv the space 3847c478bd9Sstevel@tonic-gate * We need TOP_SYNC_FORCED because we want to know when 3857c478bd9Sstevel@tonic-gate * a top_end_sync is completed. 3867c478bd9Sstevel@tonic-gate * mtm_taskq_sync_count is needed because we want to keep track 3877c478bd9Sstevel@tonic-gate * of the pending top_issue_sync dispatches so that during 3887c478bd9Sstevel@tonic-gate * forced umount we can wait for these to complete. 3897c478bd9Sstevel@tonic-gate * mtm_taskq_sync_count is decremented in top_issue_sync and 3907c478bd9Sstevel@tonic-gate * can remain set even after top_end_sync completes. 3917c478bd9Sstevel@tonic-gate * We have a window between the clearing of TOP_SYNC_FORCED 3927c478bd9Sstevel@tonic-gate * flag and the decrementing of mtm_taskq_sync_count. 3937c478bd9Sstevel@tonic-gate * If in this window new async transactions start consuming 3947c478bd9Sstevel@tonic-gate * log space, the log can get overreserved. 3957c478bd9Sstevel@tonic-gate * Subsequently a new async transaction would fail to generate 3967c478bd9Sstevel@tonic-gate * an empty sync transaction via the taskq, since it finds 3977c478bd9Sstevel@tonic-gate * the mtm_taskq_sync_count set. This can cause a hang. 3987c478bd9Sstevel@tonic-gate * Hence we do not test for mtm_taskq_sync_count being zero. 3997c478bd9Sstevel@tonic-gate * Instead, the TOP_SYNC_FORCED flag is tested here. 4007c478bd9Sstevel@tonic-gate */ 4017c478bd9Sstevel@tonic-gate if ((mtm->mtm_activesync == 0) && 4027c478bd9Sstevel@tonic-gate (!(mtm->mtm_closed & TOP_SYNC_FORCED))) { 4037c478bd9Sstevel@tonic-gate /* 4047c478bd9Sstevel@tonic-gate * Set flag to stop multiple forced empty 4057c478bd9Sstevel@tonic-gate * sync transactions. Increment mtm_taskq_sync_count. 4067c478bd9Sstevel@tonic-gate */ 4077c478bd9Sstevel@tonic-gate mtm->mtm_closed |= TOP_SYNC_FORCED; 4087c478bd9Sstevel@tonic-gate mtm->mtm_taskq_sync_count++; 4097c478bd9Sstevel@tonic-gate mutex_exit(&mtm->mtm_lock); 4107c478bd9Sstevel@tonic-gate (void) taskq_dispatch(system_taskq, 4117c478bd9Sstevel@tonic-gate top_issue_sync, ufsvfsp, TQ_SLEEP); 4127c478bd9Sstevel@tonic-gate if (tryasync) { 4137c478bd9Sstevel@tonic-gate tryfail_cnt++; 4147c478bd9Sstevel@tonic-gate return (EWOULDBLOCK); 4157c478bd9Sstevel@tonic-gate } 4167c478bd9Sstevel@tonic-gate mutex_enter(&mtm->mtm_lock); 4177c478bd9Sstevel@tonic-gate goto retry; 4187c478bd9Sstevel@tonic-gate } 4197c478bd9Sstevel@tonic-gate if (tryasync) { 4207c478bd9Sstevel@tonic-gate mutex_exit(&mtm->mtm_lock); 4217c478bd9Sstevel@tonic-gate tryfail_cnt++; 4227c478bd9Sstevel@tonic-gate return (EWOULDBLOCK); 4237c478bd9Sstevel@tonic-gate } 4247c478bd9Sstevel@tonic-gate cv_wait(&mtm->mtm_cv_next, &mtm->mtm_lock); 4257c478bd9Sstevel@tonic-gate goto retry; 4267c478bd9Sstevel@tonic-gate } 4277c478bd9Sstevel@tonic-gate /* 4287c478bd9Sstevel@tonic-gate * we are in the current transaction 4297c478bd9Sstevel@tonic-gate */ 4307c478bd9Sstevel@tonic-gate mtm->mtm_active++; 4317c478bd9Sstevel@tonic-gate ul->un_resv += size; 4327c478bd9Sstevel@tonic-gate 4337c478bd9Sstevel@tonic-gate ASSERT(mtm->mtm_active > 0); 4347c478bd9Sstevel@tonic-gate mutex_exit(&mtm->mtm_lock); 4357c478bd9Sstevel@tonic-gate 4367c478bd9Sstevel@tonic-gate ASSERT(((ul->un_debug & MT_TRANSACT) == 0) || 4377c478bd9Sstevel@tonic-gate top_begin_debug(ul, topid, size)); 4387c478bd9Sstevel@tonic-gate return (0); 4397c478bd9Sstevel@tonic-gate } 4407c478bd9Sstevel@tonic-gate 4417c478bd9Sstevel@tonic-gate /*ARGSUSED*/ 4427c478bd9Sstevel@tonic-gate void 4437c478bd9Sstevel@tonic-gate top_end_sync(ufsvfs_t *ufsvfsp, int *ep, top_t topid, ulong_t size) 4447c478bd9Sstevel@tonic-gate { 4457c478bd9Sstevel@tonic-gate ml_unit_t *ul = ufsvfsp->vfs_log; 4467c478bd9Sstevel@tonic-gate mt_map_t *mtm = ul->un_logmap; 4477c478bd9Sstevel@tonic-gate mapentry_t *cancellist; 4487c478bd9Sstevel@tonic-gate uint32_t tid; 4497c478bd9Sstevel@tonic-gate 4507c478bd9Sstevel@tonic-gate ASSERT(ufsvfsp->vfs_dev == ul->un_dev); 4517c478bd9Sstevel@tonic-gate ASSERT(((ul->un_debug & MT_TRANSACT) == 0) || 4527c478bd9Sstevel@tonic-gate top_end_debug(ul, mtm, topid, size)); 4537c478bd9Sstevel@tonic-gate 4547c478bd9Sstevel@tonic-gate mutex_enter(&mtm->mtm_lock); 4557c478bd9Sstevel@tonic-gate tid = mtm->mtm_tid; 4567c478bd9Sstevel@tonic-gate 4577c478bd9Sstevel@tonic-gate mtm->mtm_activesync--; 4587c478bd9Sstevel@tonic-gate mtm->mtm_active--; 4597c478bd9Sstevel@tonic-gate 4607c478bd9Sstevel@tonic-gate mtm->mtm_ref = 1; 4617c478bd9Sstevel@tonic-gate 4627c478bd9Sstevel@tonic-gate /* 4637c478bd9Sstevel@tonic-gate * wait for last syncop to complete 4647c478bd9Sstevel@tonic-gate */ 4657c478bd9Sstevel@tonic-gate if (mtm->mtm_activesync || panicstr) { 4667c478bd9Sstevel@tonic-gate ushort_t seq = mtm->mtm_seq; 4677c478bd9Sstevel@tonic-gate 4687c478bd9Sstevel@tonic-gate mtm->mtm_closed = TOP_SYNC; 4697c478bd9Sstevel@tonic-gate 4707c478bd9Sstevel@tonic-gate do { 4717c478bd9Sstevel@tonic-gate cv_wait(&mtm->mtm_cv_commit, &mtm->mtm_lock); 4727c478bd9Sstevel@tonic-gate } while (seq == mtm->mtm_seq); 4737c478bd9Sstevel@tonic-gate mutex_exit(&mtm->mtm_lock); 4747c478bd9Sstevel@tonic-gate goto out; 4757c478bd9Sstevel@tonic-gate } 4767c478bd9Sstevel@tonic-gate /* 4777c478bd9Sstevel@tonic-gate * last syncop; close current transaction to all ops 4787c478bd9Sstevel@tonic-gate */ 4797c478bd9Sstevel@tonic-gate mtm->mtm_closed = TOP_SYNC|TOP_ASYNC; 4807c478bd9Sstevel@tonic-gate 4817c478bd9Sstevel@tonic-gate /* 4827c478bd9Sstevel@tonic-gate * wait for last asyncop to finish 4837c478bd9Sstevel@tonic-gate */ 4847c478bd9Sstevel@tonic-gate while (mtm->mtm_active) { 4857c478bd9Sstevel@tonic-gate cv_wait(&mtm->mtm_cv_eot, &mtm->mtm_lock); 4867c478bd9Sstevel@tonic-gate } 4877c478bd9Sstevel@tonic-gate 4887c478bd9Sstevel@tonic-gate /* 4897c478bd9Sstevel@tonic-gate * push dirty metadata thru the log 4907c478bd9Sstevel@tonic-gate */ 4917c478bd9Sstevel@tonic-gate deltamap_push(ul); 4927c478bd9Sstevel@tonic-gate 4937c478bd9Sstevel@tonic-gate ASSERT(((ul->un_debug & MT_FORCEROLL) == 0) || 4947c478bd9Sstevel@tonic-gate top_roll_debug(ul)); 4957c478bd9Sstevel@tonic-gate 4967c478bd9Sstevel@tonic-gate mtm->mtm_tid = tid + 1; /* can overflow to 0 */ 4977c478bd9Sstevel@tonic-gate 4987c478bd9Sstevel@tonic-gate /* 4997c478bd9Sstevel@tonic-gate * Empty the cancellist, but save it for logmap_free_cancel 5007c478bd9Sstevel@tonic-gate */ 5017c478bd9Sstevel@tonic-gate mutex_enter(&mtm->mtm_mutex); 5027c478bd9Sstevel@tonic-gate cancellist = mtm->mtm_cancel; 5037c478bd9Sstevel@tonic-gate mtm->mtm_cancel = NULL; 5047c478bd9Sstevel@tonic-gate mutex_exit(&mtm->mtm_mutex); 5057c478bd9Sstevel@tonic-gate 5067c478bd9Sstevel@tonic-gate /* 5077c478bd9Sstevel@tonic-gate * allow async ops 5087c478bd9Sstevel@tonic-gate */ 5097c478bd9Sstevel@tonic-gate ASSERT(mtm->mtm_active == 0); 5107c478bd9Sstevel@tonic-gate ul->un_resv = 0; /* unreserve the log space */ 5117c478bd9Sstevel@tonic-gate mtm->mtm_closed = TOP_SYNC; 5127c478bd9Sstevel@tonic-gate /* 5137c478bd9Sstevel@tonic-gate * Hold the un_log_mutex here until we are done writing 5147c478bd9Sstevel@tonic-gate * the commit record to prevent any more deltas to be written 5157c478bd9Sstevel@tonic-gate * to the log after we allow async operations. 5167c478bd9Sstevel@tonic-gate */ 5177c478bd9Sstevel@tonic-gate mutex_enter(&ul->un_log_mutex); 5187c478bd9Sstevel@tonic-gate mutex_exit(&mtm->mtm_lock); 5197c478bd9Sstevel@tonic-gate cv_broadcast(&mtm->mtm_cv_next); 5207c478bd9Sstevel@tonic-gate 5217c478bd9Sstevel@tonic-gate /* 5227c478bd9Sstevel@tonic-gate * asynchronously write the commit record, 5237c478bd9Sstevel@tonic-gate */ 5247c478bd9Sstevel@tonic-gate logmap_commit(ul, tid); 5257c478bd9Sstevel@tonic-gate 5267c478bd9Sstevel@tonic-gate /* 5277c478bd9Sstevel@tonic-gate * wait for outstanding log writes (e.g., commits) to finish 5287c478bd9Sstevel@tonic-gate */ 5297c478bd9Sstevel@tonic-gate ldl_waito(ul); 5307c478bd9Sstevel@tonic-gate 5317c478bd9Sstevel@tonic-gate /* 5327c478bd9Sstevel@tonic-gate * Now that we are sure the commit has been written to the log 5337c478bd9Sstevel@tonic-gate * we can free any canceled deltas. If we free them before 5347c478bd9Sstevel@tonic-gate * guaranteeing that the commit was written, we could panic before 5357c478bd9Sstevel@tonic-gate * the commit, but after an async thread has allocated and written 5367c478bd9Sstevel@tonic-gate * to canceled freed block. 5377c478bd9Sstevel@tonic-gate */ 5387c478bd9Sstevel@tonic-gate 5397c478bd9Sstevel@tonic-gate logmap_free_cancel(mtm, &cancellist); 5407c478bd9Sstevel@tonic-gate mutex_exit(&ul->un_log_mutex); 5417c478bd9Sstevel@tonic-gate 5427c478bd9Sstevel@tonic-gate /* 5437c478bd9Sstevel@tonic-gate * now, allow all ops 5447c478bd9Sstevel@tonic-gate */ 5457c478bd9Sstevel@tonic-gate mutex_enter(&mtm->mtm_lock); 5467c478bd9Sstevel@tonic-gate mtm->mtm_active += mtm->mtm_wantin; 5477c478bd9Sstevel@tonic-gate ul->un_resv += ul->un_resv_wantin; 5487c478bd9Sstevel@tonic-gate mtm->mtm_activesync = mtm->mtm_wantin; 5497c478bd9Sstevel@tonic-gate mtm->mtm_wantin = 0; 5507c478bd9Sstevel@tonic-gate mtm->mtm_closed = 0; 5517c478bd9Sstevel@tonic-gate ul->un_resv_wantin = 0; 5527c478bd9Sstevel@tonic-gate mtm->mtm_committid = mtm->mtm_tid; 5537c478bd9Sstevel@tonic-gate mtm->mtm_seq++; 5547c478bd9Sstevel@tonic-gate mutex_exit(&mtm->mtm_lock); 5557c478bd9Sstevel@tonic-gate 5567c478bd9Sstevel@tonic-gate /* 5577c478bd9Sstevel@tonic-gate * Finish any other synchronous transactions and 5587c478bd9Sstevel@tonic-gate * start any waiting new synchronous transactions 5597c478bd9Sstevel@tonic-gate */ 5607c478bd9Sstevel@tonic-gate cv_broadcast(&mtm->mtm_cv_commit); 5617c478bd9Sstevel@tonic-gate 5627c478bd9Sstevel@tonic-gate /* 5637c478bd9Sstevel@tonic-gate * if the logmap is getting full; roll something 5647c478bd9Sstevel@tonic-gate */ 5657c478bd9Sstevel@tonic-gate if (logmap_need_roll_sync(mtm)) { 5667c478bd9Sstevel@tonic-gate logmap_forceroll_nowait(mtm); 5677c478bd9Sstevel@tonic-gate } 5687c478bd9Sstevel@tonic-gate 5697c478bd9Sstevel@tonic-gate out: 5707c478bd9Sstevel@tonic-gate if (ul->un_flags & LDL_ERROR) 5717c478bd9Sstevel@tonic-gate *ep = EIO; 5727c478bd9Sstevel@tonic-gate } 5737c478bd9Sstevel@tonic-gate 5747c478bd9Sstevel@tonic-gate /*ARGSUSED*/ 5757c478bd9Sstevel@tonic-gate void 5767c478bd9Sstevel@tonic-gate top_end_async(ufsvfs_t *ufsvfsp, top_t topid, ulong_t size) 5777c478bd9Sstevel@tonic-gate { 5787c478bd9Sstevel@tonic-gate ml_unit_t *ul = ufsvfsp->vfs_log; 5797c478bd9Sstevel@tonic-gate mt_map_t *mtm = ul->un_logmap; 5807c478bd9Sstevel@tonic-gate threadtrans_t *tp = tsd_get(topkey); 5817c478bd9Sstevel@tonic-gate int wakeup_needed = 0; 5827c478bd9Sstevel@tonic-gate 5837c478bd9Sstevel@tonic-gate ASSERT(tp); 5847c478bd9Sstevel@tonic-gate ASSERT(ufsvfsp->vfs_dev == ul->un_dev); 5857c478bd9Sstevel@tonic-gate ASSERT(((ul->un_debug & MT_TRANSACT) == 0) || 5867c478bd9Sstevel@tonic-gate top_end_debug(ul, mtm, topid, size)); 5877c478bd9Sstevel@tonic-gate 5887c478bd9Sstevel@tonic-gate mutex_enter(&mtm->mtm_lock); 5897c478bd9Sstevel@tonic-gate 5907c478bd9Sstevel@tonic-gate if (size > tp->deltas_size) { 5917c478bd9Sstevel@tonic-gate ul->un_resv -= (size - tp->deltas_size); 5927c478bd9Sstevel@tonic-gate } 5937c478bd9Sstevel@tonic-gate if (tp->any_deltas) { 5947c478bd9Sstevel@tonic-gate tp->last_async_tid = mtm->mtm_tid; 5957c478bd9Sstevel@tonic-gate } 5967c478bd9Sstevel@tonic-gate mtm->mtm_ref = 1; 5977c478bd9Sstevel@tonic-gate 5987c478bd9Sstevel@tonic-gate mtm->mtm_active--; 5997c478bd9Sstevel@tonic-gate if ((mtm->mtm_active == 0) && 6007c478bd9Sstevel@tonic-gate (mtm->mtm_closed == (TOP_SYNC|TOP_ASYNC))) { 6017c478bd9Sstevel@tonic-gate wakeup_needed = 1; 6027c478bd9Sstevel@tonic-gate } 6037c478bd9Sstevel@tonic-gate mutex_exit(&mtm->mtm_lock); 6047c478bd9Sstevel@tonic-gate if (wakeup_needed) 6057c478bd9Sstevel@tonic-gate cv_signal(&mtm->mtm_cv_eot); 6067c478bd9Sstevel@tonic-gate 6077c478bd9Sstevel@tonic-gate /* 6087c478bd9Sstevel@tonic-gate * Generate a sync op if the log, logmap, or deltamap are heavily used. 6097c478bd9Sstevel@tonic-gate * Unless we are possibly holding any VM locks, since if we are holding 6107c478bd9Sstevel@tonic-gate * any VM locks and we issue a top_end_sync(), we could deadlock. 6117c478bd9Sstevel@tonic-gate */ 6127c478bd9Sstevel@tonic-gate if ((mtm->mtm_activesync == 0) && 6137c478bd9Sstevel@tonic-gate !(mtm->mtm_closed & TOP_SYNC) && 6147c478bd9Sstevel@tonic-gate (deltamap_need_commit(ul->un_deltamap) || 6157c478bd9Sstevel@tonic-gate logmap_need_commit(mtm) || 6167c478bd9Sstevel@tonic-gate ldl_need_commit(ul)) && 6177c478bd9Sstevel@tonic-gate (topid != TOP_GETPAGE)) { 6187c478bd9Sstevel@tonic-gate top_issue_sync(ufsvfsp); 6197c478bd9Sstevel@tonic-gate } 6207c478bd9Sstevel@tonic-gate /* 6217c478bd9Sstevel@tonic-gate * roll something from the log if the logmap is too full 6227c478bd9Sstevel@tonic-gate */ 6237c478bd9Sstevel@tonic-gate if (logmap_need_roll_async(mtm)) 6247c478bd9Sstevel@tonic-gate logmap_forceroll_nowait(mtm); 6257c478bd9Sstevel@tonic-gate } 6267c478bd9Sstevel@tonic-gate 6277c478bd9Sstevel@tonic-gate /* 6287c478bd9Sstevel@tonic-gate * Called from roll thread; 6297c478bd9Sstevel@tonic-gate * buffer set for reading master 6307c478bd9Sstevel@tonic-gate * Returns 6317c478bd9Sstevel@tonic-gate * 0 - success, can continue with next buffer 6327c478bd9Sstevel@tonic-gate * 1 - failure due to logmap deltas being in use 6337c478bd9Sstevel@tonic-gate */ 6347c478bd9Sstevel@tonic-gate int 6357c478bd9Sstevel@tonic-gate top_read_roll(rollbuf_t *rbp, ml_unit_t *ul) 6367c478bd9Sstevel@tonic-gate { 6377c478bd9Sstevel@tonic-gate buf_t *bp = &rbp->rb_bh; 6387c478bd9Sstevel@tonic-gate offset_t mof = ldbtob(bp->b_blkno); 6397c478bd9Sstevel@tonic-gate 6407c478bd9Sstevel@tonic-gate /* 6417c478bd9Sstevel@tonic-gate * get a list of deltas 6427c478bd9Sstevel@tonic-gate */ 6437c478bd9Sstevel@tonic-gate if (logmap_list_get_roll(ul->un_logmap, mof, rbp)) { 6447c478bd9Sstevel@tonic-gate /* logmap deltas are in use */ 6457c478bd9Sstevel@tonic-gate return (1); 6467c478bd9Sstevel@tonic-gate } 6477c478bd9Sstevel@tonic-gate 6487c478bd9Sstevel@tonic-gate /* 6497c478bd9Sstevel@tonic-gate * no deltas were found, nothing to roll 6507c478bd9Sstevel@tonic-gate */ 6517c478bd9Sstevel@tonic-gate if (rbp->rb_age == NULL) { 6527c478bd9Sstevel@tonic-gate bp->b_flags |= B_INVAL; 6537c478bd9Sstevel@tonic-gate return (0); 6547c478bd9Sstevel@tonic-gate } 6557c478bd9Sstevel@tonic-gate 6567c478bd9Sstevel@tonic-gate /* 6577c478bd9Sstevel@tonic-gate * If there is one cached roll buffer that cover all the deltas then 6587c478bd9Sstevel@tonic-gate * we can use that instead of copying to a separate roll buffer. 6597c478bd9Sstevel@tonic-gate */ 6607c478bd9Sstevel@tonic-gate if (rbp->rb_crb) { 6617c478bd9Sstevel@tonic-gate rbp->rb_bh.b_blkno = lbtodb(rbp->rb_crb->c_mof); 6627c478bd9Sstevel@tonic-gate return (0); 6637c478bd9Sstevel@tonic-gate } 6647c478bd9Sstevel@tonic-gate 6657c478bd9Sstevel@tonic-gate /* 6667c478bd9Sstevel@tonic-gate * Set up the read. 6677c478bd9Sstevel@tonic-gate * If no read is needed logmap_setup_read() returns 0. 6687c478bd9Sstevel@tonic-gate */ 6697c478bd9Sstevel@tonic-gate if (logmap_setup_read(rbp->rb_age, rbp)) { 6707c478bd9Sstevel@tonic-gate /* 6717c478bd9Sstevel@tonic-gate * async read the data from master 6727c478bd9Sstevel@tonic-gate */ 6737c478bd9Sstevel@tonic-gate logstats.ls_rreads.value.ui64++; 6747c478bd9Sstevel@tonic-gate bp->b_bcount = MAPBLOCKSIZE; 6757c478bd9Sstevel@tonic-gate (void) bdev_strategy(bp); 6767c478bd9Sstevel@tonic-gate lwp_stat_update(LWP_STAT_INBLK, 1); 6777c478bd9Sstevel@tonic-gate } else { 6787c478bd9Sstevel@tonic-gate sema_v(&bp->b_io); /* mark read as complete */ 6797c478bd9Sstevel@tonic-gate } 6807c478bd9Sstevel@tonic-gate return (0); 6817c478bd9Sstevel@tonic-gate } 6827c478bd9Sstevel@tonic-gate 6837c478bd9Sstevel@tonic-gate int ufs_crb_enable = 1; 6847c478bd9Sstevel@tonic-gate 6857c478bd9Sstevel@tonic-gate /* 6867c478bd9Sstevel@tonic-gate * move deltas from deltamap into the log 6877c478bd9Sstevel@tonic-gate */ 6887c478bd9Sstevel@tonic-gate void 6897c478bd9Sstevel@tonic-gate top_log(ufsvfs_t *ufsvfsp, char *va, offset_t vamof, off_t nb, 6907c478bd9Sstevel@tonic-gate caddr_t buf, uint32_t bufsz) 6917c478bd9Sstevel@tonic-gate { 6927c478bd9Sstevel@tonic-gate ml_unit_t *ul = ufsvfsp->vfs_log; 6937c478bd9Sstevel@tonic-gate mapentry_t *me; 6947c478bd9Sstevel@tonic-gate offset_t hmof; 6957c478bd9Sstevel@tonic-gate uint32_t hnb, nb1; 6967c478bd9Sstevel@tonic-gate 6977c478bd9Sstevel@tonic-gate /* 6987c478bd9Sstevel@tonic-gate * needed for the roll thread's heuristic 6997c478bd9Sstevel@tonic-gate */ 7007c478bd9Sstevel@tonic-gate ul->un_logmap->mtm_ref = 1; 7017c478bd9Sstevel@tonic-gate 7027c478bd9Sstevel@tonic-gate if (buf && ufs_crb_enable) { 7037c478bd9Sstevel@tonic-gate ASSERT((bufsz & DEV_BMASK) == 0); 7047c478bd9Sstevel@tonic-gate /* 7057c478bd9Sstevel@tonic-gate * Move any deltas to the logmap. Split requests that 7067c478bd9Sstevel@tonic-gate * straddle MAPBLOCKSIZE hash boundaries (i.e. summary info). 7077c478bd9Sstevel@tonic-gate */ 7087c478bd9Sstevel@tonic-gate for (hmof = vamof - (va - buf), nb1 = nb; bufsz; 7097c478bd9Sstevel@tonic-gate bufsz -= hnb, hmof += hnb, buf += hnb, nb1 -= hnb) { 7107c478bd9Sstevel@tonic-gate hnb = MAPBLOCKSIZE - (hmof & MAPBLOCKOFF); 7117c478bd9Sstevel@tonic-gate if (hnb > bufsz) 7127c478bd9Sstevel@tonic-gate hnb = bufsz; 7137c478bd9Sstevel@tonic-gate me = deltamap_remove(ul->un_deltamap, 7147c478bd9Sstevel@tonic-gate MAX(hmof, vamof), MIN(hnb, nb1)); 7157c478bd9Sstevel@tonic-gate if (me) { 7167c478bd9Sstevel@tonic-gate logmap_add_buf(ul, va, hmof, me, buf, hnb); 7177c478bd9Sstevel@tonic-gate } 7187c478bd9Sstevel@tonic-gate } 7197c478bd9Sstevel@tonic-gate } else { 7207c478bd9Sstevel@tonic-gate /* 7217c478bd9Sstevel@tonic-gate * if there are deltas 7227c478bd9Sstevel@tonic-gate */ 7237c478bd9Sstevel@tonic-gate me = deltamap_remove(ul->un_deltamap, vamof, nb); 7247c478bd9Sstevel@tonic-gate if (me) { 7257c478bd9Sstevel@tonic-gate /* 7267c478bd9Sstevel@tonic-gate * move to logmap 7277c478bd9Sstevel@tonic-gate */ 7287c478bd9Sstevel@tonic-gate logmap_add(ul, va, vamof, me); 7297c478bd9Sstevel@tonic-gate } 7307c478bd9Sstevel@tonic-gate } 7317c478bd9Sstevel@tonic-gate 7327c478bd9Sstevel@tonic-gate ASSERT((ul->un_matamap == NULL) || 7337c478bd9Sstevel@tonic-gate matamap_within(ul->un_matamap, vamof, nb)); 7347c478bd9Sstevel@tonic-gate } 7357c478bd9Sstevel@tonic-gate 7367c478bd9Sstevel@tonic-gate 7377c478bd9Sstevel@tonic-gate static void 7387c478bd9Sstevel@tonic-gate top_threadtrans_destroy(void *tp) 7397c478bd9Sstevel@tonic-gate { 7407c478bd9Sstevel@tonic-gate kmem_free(tp, sizeof (threadtrans_t)); 7417c478bd9Sstevel@tonic-gate } 7427c478bd9Sstevel@tonic-gate 7437c478bd9Sstevel@tonic-gate void 7447c478bd9Sstevel@tonic-gate _init_top(void) 7457c478bd9Sstevel@tonic-gate { 7467c478bd9Sstevel@tonic-gate ASSERT(top_init_debug()); 7477c478bd9Sstevel@tonic-gate 7487c478bd9Sstevel@tonic-gate /* 7497c478bd9Sstevel@tonic-gate * set up the delta layer 7507c478bd9Sstevel@tonic-gate */ 7517c478bd9Sstevel@tonic-gate _init_map(); 7527c478bd9Sstevel@tonic-gate 7537c478bd9Sstevel@tonic-gate /* 7547c478bd9Sstevel@tonic-gate * Initialise the thread specific data transaction key 7557c478bd9Sstevel@tonic-gate */ 7567c478bd9Sstevel@tonic-gate tsd_create(&topkey, top_threadtrans_destroy); 7577c478bd9Sstevel@tonic-gate } 758