Lines Matching refs:txg
295 zil_parse_lr_func_t *parse_lr_func, void *arg, uint64_t txg) in zil_parse() argument
334 if ((error = parse_blk_func(zilog, &blk, arg, txg)) != 0) in zil_parse()
353 if ((error = parse_lr_func(zilog, lr, arg, txg)) != 0) in zil_parse()
442 zil_alloc_lwb(zilog_t *zilog, blkptr_t *bp, uint64_t txg) in zil_alloc_lwb() argument
450 lwb->lwb_max_txg = txg; in zil_alloc_lwb()
473 zilog_dirty(zilog_t *zilog, uint64_t txg) in zilog_dirty() argument
481 if (txg_list_add(&dp->dp_dirty_zilogs, zilog, txg)) { in zilog_dirty()
507 uint64_t txg = 0; in zil_create() local
531 txg = dmu_tx_get_txg(tx); in zil_create()
534 zio_free_zil(zilog->zl_spa, txg, &blk); in zil_create()
538 error = zio_alloc_zil(zilog->zl_spa, txg, &blk, NULL, in zil_create()
549 lwb = zil_alloc_lwb(zilog, &blk, txg); in zil_create()
558 txg_wait_synced(zilog->zl_dmu_pool, txg); in zil_create()
581 uint64_t txg; in zil_destroy() local
596 txg = dmu_tx_get_txg(tx); in zil_destroy()
600 ASSERT3U(zilog->zl_destroy_txg, <, txg); in zil_destroy()
601 zilog->zl_destroy_txg = txg; in zil_destroy()
611 zio_free_zil(zilog->zl_spa, txg, &lwb->lwb_blk); in zil_destroy()
933 uint64_t txg; in zil_lwb_write_start() local
959 txg = dmu_tx_get_txg(tx); in zil_lwb_write_start()
992 error = zio_alloc_zil(spa, txg, bp, &lwb->lwb_blk, zil_blksz, in zil_lwb_write_start()
995 ASSERT3U(bp->blk_birth, ==, txg); in zil_lwb_write_start()
1002 nlwb = zil_alloc_lwb(zilog, bp, txg); in zil_lwb_write_start()
1042 uint64_t txg = lrc->lrc_txg; in zil_lwb_commit() local
1071 txg_wait_synced(zilog->zl_dmu_pool, txg); in zil_lwb_commit()
1085 if (txg > spa_freeze_txg(zilog->zl_spa)) in zil_lwb_commit()
1086 txg_wait_synced(zilog->zl_dmu_pool, txg); in zil_lwb_commit()
1102 txg_wait_synced(zilog->zl_dmu_pool, txg); in zil_lwb_commit()
1121 lwb->lwb_max_txg = MAX(lwb->lwb_max_txg, txg); in zil_lwb_commit()
1208 uint64_t otxg, txg; in zil_remove_async() local
1223 for (txg = otxg; txg < (otxg + TXG_CONCURRENT_STATES); txg++) { in zil_remove_async()
1224 itxg_t *itxg = &zilog->zl_itxg[txg & TXG_MASK]; in zil_remove_async()
1227 if (itxg->itxg_txg != txg) { in zil_remove_async()
1252 uint64_t txg; in zil_itx_assign() local
1273 txg = ZILTEST_TXG; in zil_itx_assign()
1275 txg = dmu_tx_get_txg(tx); in zil_itx_assign()
1277 itxg = &zilog->zl_itxg[txg & TXG_MASK]; in zil_itx_assign()
1280 if (itxg->itxg_txg != txg) { in zil_itx_assign()
1292 itxg->itxg_txg = txg; in zil_itx_assign()
1323 zilog_dirty(zilog, txg); in zil_itx_assign()
1375 uint64_t otxg, txg; in zil_get_commit_list() local
1384 for (txg = otxg; txg < (otxg + TXG_CONCURRENT_STATES); txg++) { in zil_get_commit_list()
1385 itxg_t *itxg = &zilog->zl_itxg[txg & TXG_MASK]; in zil_get_commit_list()
1388 if (itxg->itxg_txg != txg) { in zil_get_commit_list()
1408 uint64_t otxg, txg; in zil_async_to_sync() local
1418 for (txg = otxg; txg < (otxg + TXG_CONCURRENT_STATES); txg++) { in zil_async_to_sync()
1419 itxg_t *itxg = &zilog->zl_itxg[txg & TXG_MASK]; in zil_async_to_sync()
1422 if (itxg->itxg_txg != txg) { in zil_async_to_sync()
1457 uint64_t txg; in zil_commit_writer() local
1488 txg = itx->itx_lr.lrc_txg; in zil_commit_writer()
1489 ASSERT(txg); in zil_commit_writer()
1491 if (txg > spa_last_synced_txg(spa) || txg > spa_freeze_txg(spa)) in zil_commit_writer()
1593 uint64_t txg = dmu_tx_get_txg(tx); in zil_sync() local
1595 uint64_t *replayed_seq = &zilog->zl_replayed_seq[txg & TXG_MASK]; in zil_sync()
1615 if (zilog->zl_destroy_txg == txg) { in zil_sync()
1639 if (lwb->lwb_buf != NULL || lwb->lwb_max_txg > txg) in zil_sync()
1642 zio_free_zil(spa, txg, &lwb->lwb_blk); in zil_sync()
1790 uint64_t txg = 0; in zil_close() local
1803 txg = lwb->lwb_max_txg; in zil_close()
1805 if (txg) in zil_close()
1806 txg_wait_synced(zilog->zl_dmu_pool, txg); in zil_close()