1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or https://opensource.org/licenses/CDDL-1.0. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 /* 22 * Copyright (c) 2005, 2010, Oracle and/or its affiliates. All rights reserved. 23 * Copyright (c) 2015, 2018 by Delphix. All rights reserved. 24 * Copyright (c) 2022 by Pawel Jakub Dawidek 25 */ 26 27 28 #include <sys/types.h> 29 #include <sys/param.h> 30 #include <sys/sysmacros.h> 31 #include <sys/cmn_err.h> 32 #include <sys/kmem.h> 33 #include <sys/thread.h> 34 #include <sys/file.h> 35 #include <sys/vfs.h> 36 #include <sys/zfs_znode.h> 37 #include <sys/zfs_dir.h> 38 #include <sys/zil.h> 39 #include <sys/zil_impl.h> 40 #include <sys/byteorder.h> 41 #include <sys/policy.h> 42 #include <sys/stat.h> 43 #include <sys/acl.h> 44 #include <sys/dmu.h> 45 #include <sys/dbuf.h> 46 #include <sys/spa.h> 47 #include <sys/zfs_fuid.h> 48 #include <sys/dsl_dataset.h> 49 50 /* 51 * These zfs_log_* functions must be called within a dmu tx, in one 52 * of 2 contexts depending on zilog->z_replay: 53 * 54 * Non replay mode 55 * --------------- 56 * We need to record the transaction so that if it is committed to 57 * the Intent Log then it can be replayed. An intent log transaction 58 * structure (itx_t) is allocated and all the information necessary to 59 * possibly replay the transaction is saved in it. The itx is then assigned 60 * a sequence number and inserted in the in-memory list anchored in the zilog. 61 * 62 * Replay mode 63 * ----------- 64 * We need to mark the intent log record as replayed in the log header. 65 * This is done in the same transaction as the replay so that they 66 * commit atomically. 67 */ 68 69 int 70 zfs_log_create_txtype(zil_create_t type, vsecattr_t *vsecp, vattr_t *vap) 71 { 72 int isxvattr = (vap->va_mask & ATTR_XVATTR); 73 switch (type) { 74 case Z_FILE: 75 if (vsecp == NULL && !isxvattr) 76 return (TX_CREATE); 77 if (vsecp && isxvattr) 78 return (TX_CREATE_ACL_ATTR); 79 if (vsecp) 80 return (TX_CREATE_ACL); 81 else 82 return (TX_CREATE_ATTR); 83 case Z_DIR: 84 if (vsecp == NULL && !isxvattr) 85 return (TX_MKDIR); 86 if (vsecp && isxvattr) 87 return (TX_MKDIR_ACL_ATTR); 88 if (vsecp) 89 return (TX_MKDIR_ACL); 90 else 91 return (TX_MKDIR_ATTR); 92 case Z_XATTRDIR: 93 return (TX_MKXATTR); 94 } 95 ASSERT(0); 96 return (TX_MAX_TYPE); 97 } 98 99 /* 100 * build up the log data necessary for logging xvattr_t 101 * First lr_attr_t is initialized. following the lr_attr_t 102 * is the mapsize and attribute bitmap copied from the xvattr_t. 103 * Following the bitmap and bitmapsize two 64 bit words are reserved 104 * for the create time which may be set. Following the create time 105 * records a single 64 bit integer which has the bits to set on 106 * replay for the xvattr. 107 */ 108 static void 109 zfs_log_xvattr(lr_attr_t *lrattr, xvattr_t *xvap) 110 { 111 xoptattr_t *xoap; 112 113 xoap = xva_getxoptattr(xvap); 114 ASSERT(xoap); 115 116 lrattr->lr_attr_masksize = xvap->xva_mapsize; 117 uint32_t *bitmap = &lrattr->lr_attr_bitmap; 118 for (int i = 0; i != xvap->xva_mapsize; i++, bitmap++) 119 *bitmap = xvap->xva_reqattrmap[i]; 120 121 lr_attr_end_t *end = (lr_attr_end_t *)bitmap; 122 end->lr_attr_attrs = 0; 123 end->lr_attr_crtime[0] = 0; 124 end->lr_attr_crtime[1] = 0; 125 memset(end->lr_attr_scanstamp, 0, AV_SCANSTAMP_SZ); 126 127 if (XVA_ISSET_REQ(xvap, XAT_READONLY)) 128 end->lr_attr_attrs |= (xoap->xoa_readonly == 0) ? 0 : 129 XAT0_READONLY; 130 if (XVA_ISSET_REQ(xvap, XAT_HIDDEN)) 131 end->lr_attr_attrs |= (xoap->xoa_hidden == 0) ? 0 : 132 XAT0_HIDDEN; 133 if (XVA_ISSET_REQ(xvap, XAT_SYSTEM)) 134 end->lr_attr_attrs |= (xoap->xoa_system == 0) ? 0 : 135 XAT0_SYSTEM; 136 if (XVA_ISSET_REQ(xvap, XAT_ARCHIVE)) 137 end->lr_attr_attrs |= (xoap->xoa_archive == 0) ? 0 : 138 XAT0_ARCHIVE; 139 if (XVA_ISSET_REQ(xvap, XAT_IMMUTABLE)) 140 end->lr_attr_attrs |= (xoap->xoa_immutable == 0) ? 0 : 141 XAT0_IMMUTABLE; 142 if (XVA_ISSET_REQ(xvap, XAT_NOUNLINK)) 143 end->lr_attr_attrs |= (xoap->xoa_nounlink == 0) ? 0 : 144 XAT0_NOUNLINK; 145 if (XVA_ISSET_REQ(xvap, XAT_APPENDONLY)) 146 end->lr_attr_attrs |= (xoap->xoa_appendonly == 0) ? 0 : 147 XAT0_APPENDONLY; 148 if (XVA_ISSET_REQ(xvap, XAT_OPAQUE)) 149 end->lr_attr_attrs |= (xoap->xoa_opaque == 0) ? 0 : 150 XAT0_APPENDONLY; 151 if (XVA_ISSET_REQ(xvap, XAT_NODUMP)) 152 end->lr_attr_attrs |= (xoap->xoa_nodump == 0) ? 0 : 153 XAT0_NODUMP; 154 if (XVA_ISSET_REQ(xvap, XAT_AV_QUARANTINED)) 155 end->lr_attr_attrs |= (xoap->xoa_av_quarantined == 0) ? 0 : 156 XAT0_AV_QUARANTINED; 157 if (XVA_ISSET_REQ(xvap, XAT_AV_MODIFIED)) 158 end->lr_attr_attrs |= (xoap->xoa_av_modified == 0) ? 0 : 159 XAT0_AV_MODIFIED; 160 if (XVA_ISSET_REQ(xvap, XAT_CREATETIME)) 161 ZFS_TIME_ENCODE(&xoap->xoa_createtime, end->lr_attr_crtime); 162 if (XVA_ISSET_REQ(xvap, XAT_AV_SCANSTAMP)) { 163 ASSERT(!XVA_ISSET_REQ(xvap, XAT_PROJID)); 164 165 memcpy(end->lr_attr_scanstamp, xoap->xoa_av_scanstamp, 166 AV_SCANSTAMP_SZ); 167 } else if (XVA_ISSET_REQ(xvap, XAT_PROJID)) { 168 /* 169 * XAT_PROJID and XAT_AV_SCANSTAMP will never be valid 170 * at the same time, so we can share the same space. 171 */ 172 memcpy(end->lr_attr_scanstamp, &xoap->xoa_projid, 173 sizeof (uint64_t)); 174 } 175 if (XVA_ISSET_REQ(xvap, XAT_REPARSE)) 176 end->lr_attr_attrs |= (xoap->xoa_reparse == 0) ? 0 : 177 XAT0_REPARSE; 178 if (XVA_ISSET_REQ(xvap, XAT_OFFLINE)) 179 end->lr_attr_attrs |= (xoap->xoa_offline == 0) ? 0 : 180 XAT0_OFFLINE; 181 if (XVA_ISSET_REQ(xvap, XAT_SPARSE)) 182 end->lr_attr_attrs |= (xoap->xoa_sparse == 0) ? 0 : 183 XAT0_SPARSE; 184 if (XVA_ISSET_REQ(xvap, XAT_PROJINHERIT)) 185 end->lr_attr_attrs |= (xoap->xoa_projinherit == 0) ? 0 : 186 XAT0_PROJINHERIT; 187 } 188 189 static void * 190 zfs_log_fuid_ids(zfs_fuid_info_t *fuidp, void *start) 191 { 192 zfs_fuid_t *zfuid; 193 uint64_t *fuidloc = start; 194 195 /* First copy in the ACE FUIDs */ 196 for (zfuid = list_head(&fuidp->z_fuids); zfuid; 197 zfuid = list_next(&fuidp->z_fuids, zfuid)) { 198 *fuidloc++ = zfuid->z_logfuid; 199 } 200 return (fuidloc); 201 } 202 203 204 static void * 205 zfs_log_fuid_domains(zfs_fuid_info_t *fuidp, void *start) 206 { 207 zfs_fuid_domain_t *zdomain; 208 209 /* now copy in the domain info, if any */ 210 if (fuidp->z_domain_str_sz != 0) { 211 for (zdomain = list_head(&fuidp->z_domains); zdomain; 212 zdomain = list_next(&fuidp->z_domains, zdomain)) { 213 memcpy(start, zdomain->z_domain, 214 strlen(zdomain->z_domain) + 1); 215 start = (caddr_t)start + 216 strlen(zdomain->z_domain) + 1; 217 } 218 } 219 return (start); 220 } 221 222 /* 223 * If zp is an xattr node, check whether the xattr owner is unlinked. 224 * We don't want to log anything if the owner is unlinked. 225 */ 226 static int 227 zfs_xattr_owner_unlinked(znode_t *zp) 228 { 229 int unlinked = 0; 230 znode_t *dzp; 231 #ifdef __FreeBSD__ 232 znode_t *tzp = zp; 233 234 /* 235 * zrele drops the vnode lock which violates the VOP locking contract 236 * on FreeBSD. See comment at the top of zfs_replay.c for more detail. 237 */ 238 /* 239 * if zp is XATTR node, keep walking up via z_xattr_parent until we 240 * get the owner 241 */ 242 while (tzp->z_pflags & ZFS_XATTR) { 243 ASSERT3U(zp->z_xattr_parent, !=, 0); 244 if (zfs_zget(ZTOZSB(tzp), tzp->z_xattr_parent, &dzp) != 0) { 245 unlinked = 1; 246 break; 247 } 248 249 if (tzp != zp) 250 zrele(tzp); 251 tzp = dzp; 252 unlinked = tzp->z_unlinked; 253 } 254 if (tzp != zp) 255 zrele(tzp); 256 #else 257 zhold(zp); 258 /* 259 * if zp is XATTR node, keep walking up via z_xattr_parent until we 260 * get the owner 261 */ 262 while (zp->z_pflags & ZFS_XATTR) { 263 ASSERT3U(zp->z_xattr_parent, !=, 0); 264 if (zfs_zget(ZTOZSB(zp), zp->z_xattr_parent, &dzp) != 0) { 265 unlinked = 1; 266 break; 267 } 268 269 zrele(zp); 270 zp = dzp; 271 unlinked = zp->z_unlinked; 272 } 273 zrele(zp); 274 #endif 275 return (unlinked); 276 } 277 278 /* 279 * Handles TX_CREATE, TX_CREATE_ATTR, TX_MKDIR, TX_MKDIR_ATTR and 280 * TK_MKXATTR transactions. 281 * 282 * TX_CREATE and TX_MKDIR are standard creates, but they may have FUID 283 * domain information appended prior to the name. In this case the 284 * uid/gid in the log record will be a log centric FUID. 285 * 286 * TX_CREATE_ACL_ATTR and TX_MKDIR_ACL_ATTR handle special creates that 287 * may contain attributes, ACL and optional fuid information. 288 * 289 * TX_CREATE_ACL and TX_MKDIR_ACL handle special creates that specify 290 * and ACL and normal users/groups in the ACEs. 291 * 292 * There may be an optional xvattr attribute information similar 293 * to zfs_log_setattr. 294 * 295 * Also, after the file name "domain" strings may be appended. 296 */ 297 void 298 zfs_log_create(zilog_t *zilog, dmu_tx_t *tx, uint64_t txtype, 299 znode_t *dzp, znode_t *zp, const char *name, vsecattr_t *vsecp, 300 zfs_fuid_info_t *fuidp, vattr_t *vap) 301 { 302 itx_t *itx; 303 _lr_create_t *lr; 304 lr_acl_create_t *lracl = NULL; 305 uint8_t *lrdata; 306 size_t aclsize = 0; 307 size_t xvatsize = 0; 308 size_t txsize; 309 xvattr_t *xvap = (xvattr_t *)vap; 310 size_t namesize = strlen(name) + 1; 311 size_t fuidsz = 0; 312 313 if (zil_replaying(zilog, tx) || zfs_xattr_owner_unlinked(dzp)) 314 return; 315 316 /* 317 * If we have FUIDs present then add in space for 318 * domains and ACE fuid's if any. 319 */ 320 if (fuidp) { 321 fuidsz += fuidp->z_domain_str_sz; 322 fuidsz += fuidp->z_fuid_cnt * sizeof (uint64_t); 323 } 324 325 if (vap->va_mask & ATTR_XVATTR) 326 xvatsize = ZIL_XVAT_SIZE(xvap->xva_mapsize); 327 328 if ((int)txtype == TX_CREATE_ATTR || (int)txtype == TX_MKDIR_ATTR || 329 (int)txtype == TX_CREATE || (int)txtype == TX_MKDIR || 330 (int)txtype == TX_MKXATTR) { 331 txsize = sizeof (lr_create_t) + namesize + fuidsz + xvatsize; 332 itx = zil_itx_create(txtype, txsize); 333 lr_create_t *lrc = (lr_create_t *)&itx->itx_lr; 334 lrdata = &lrc->lr_data[0]; 335 } else { 336 txsize = 337 sizeof (lr_acl_create_t) + namesize + fuidsz + 338 ZIL_ACE_LENGTH(aclsize) + xvatsize; 339 itx = zil_itx_create(txtype, txsize); 340 lracl = (lr_acl_create_t *)&itx->itx_lr; 341 lrdata = &lracl->lr_data[0]; 342 } 343 344 345 lr = (_lr_create_t *)&itx->itx_lr; 346 lr->lr_doid = dzp->z_id; 347 lr->lr_foid = zp->z_id; 348 /* Store dnode slot count in 8 bits above object id. */ 349 LR_FOID_SET_SLOTS(lr->lr_foid, zp->z_dnodesize >> DNODE_SHIFT); 350 lr->lr_mode = zp->z_mode; 351 if (!IS_EPHEMERAL(KUID_TO_SUID(ZTOUID(zp)))) { 352 lr->lr_uid = (uint64_t)KUID_TO_SUID(ZTOUID(zp)); 353 } else { 354 lr->lr_uid = fuidp->z_fuid_owner; 355 } 356 if (!IS_EPHEMERAL(KGID_TO_SGID(ZTOGID(zp)))) { 357 lr->lr_gid = (uint64_t)KGID_TO_SGID(ZTOGID(zp)); 358 } else { 359 lr->lr_gid = fuidp->z_fuid_group; 360 } 361 (void) sa_lookup(zp->z_sa_hdl, SA_ZPL_GEN(ZTOZSB(zp)), &lr->lr_gen, 362 sizeof (uint64_t)); 363 (void) sa_lookup(zp->z_sa_hdl, SA_ZPL_CRTIME(ZTOZSB(zp)), 364 lr->lr_crtime, sizeof (uint64_t) * 2); 365 366 if (sa_lookup(zp->z_sa_hdl, SA_ZPL_RDEV(ZTOZSB(zp)), &lr->lr_rdev, 367 sizeof (lr->lr_rdev)) != 0) 368 lr->lr_rdev = 0; 369 370 /* 371 * Fill in xvattr info if any 372 */ 373 if (vap->va_mask & ATTR_XVATTR) { 374 zfs_log_xvattr((lr_attr_t *)lrdata, xvap); 375 lrdata = &lrdata[xvatsize]; 376 } 377 378 /* Now fill in any ACL info */ 379 380 if (vsecp) { 381 ASSERT3P(lracl, !=, NULL); 382 lracl->lr_aclcnt = vsecp->vsa_aclcnt; 383 lracl->lr_acl_bytes = aclsize; 384 lracl->lr_domcnt = fuidp ? fuidp->z_domain_cnt : 0; 385 lracl->lr_fuidcnt = fuidp ? fuidp->z_fuid_cnt : 0; 386 if (vsecp->vsa_aclflags & VSA_ACE_ACLFLAGS) 387 lracl->lr_acl_flags = (uint64_t)vsecp->vsa_aclflags; 388 else 389 lracl->lr_acl_flags = 0; 390 391 memcpy(lrdata, vsecp->vsa_aclentp, aclsize); 392 lrdata = &lrdata[ZIL_ACE_LENGTH(aclsize)]; 393 } 394 395 /* drop in FUID info */ 396 if (fuidp) { 397 lrdata = zfs_log_fuid_ids(fuidp, lrdata); 398 lrdata = zfs_log_fuid_domains(fuidp, lrdata); 399 } 400 /* 401 * Now place file name in log record 402 */ 403 memcpy(lrdata, name, namesize); 404 405 zil_itx_assign(zilog, itx, tx); 406 } 407 408 /* 409 * Handles both TX_REMOVE and TX_RMDIR transactions. 410 */ 411 void 412 zfs_log_remove(zilog_t *zilog, dmu_tx_t *tx, uint64_t txtype, 413 znode_t *dzp, const char *name, uint64_t foid, boolean_t unlinked) 414 { 415 itx_t *itx; 416 lr_remove_t *lr; 417 size_t namesize = strlen(name) + 1; 418 419 if (zil_replaying(zilog, tx) || zfs_xattr_owner_unlinked(dzp)) 420 return; 421 422 itx = zil_itx_create(txtype, sizeof (*lr) + namesize); 423 lr = (lr_remove_t *)&itx->itx_lr; 424 lr->lr_doid = dzp->z_id; 425 memcpy(&lr->lr_data[0], name, namesize); 426 427 itx->itx_oid = foid; 428 429 /* 430 * Object ids can be re-instantiated in the next txg so 431 * remove any async transactions to avoid future leaks. 432 * This can happen if a fsync occurs on the re-instantiated 433 * object for a WR_INDIRECT or WR_NEED_COPY write, which gets 434 * the new file data and flushes a write record for the old object. 435 */ 436 if (unlinked) { 437 ASSERT((txtype & ~TX_CI) == TX_REMOVE); 438 zil_remove_async(zilog, foid); 439 } 440 zil_itx_assign(zilog, itx, tx); 441 } 442 443 /* 444 * Handles TX_LINK transactions. 445 */ 446 void 447 zfs_log_link(zilog_t *zilog, dmu_tx_t *tx, uint64_t txtype, 448 znode_t *dzp, znode_t *zp, const char *name) 449 { 450 itx_t *itx; 451 lr_link_t *lr; 452 size_t namesize = strlen(name) + 1; 453 454 if (zil_replaying(zilog, tx)) 455 return; 456 457 itx = zil_itx_create(txtype, sizeof (*lr) + namesize); 458 lr = (lr_link_t *)&itx->itx_lr; 459 lr->lr_doid = dzp->z_id; 460 lr->lr_link_obj = zp->z_id; 461 memcpy(&lr->lr_data[0], name, namesize); 462 463 zil_itx_assign(zilog, itx, tx); 464 } 465 466 /* 467 * Handles TX_SYMLINK transactions. 468 */ 469 void 470 zfs_log_symlink(zilog_t *zilog, dmu_tx_t *tx, uint64_t txtype, 471 znode_t *dzp, znode_t *zp, const char *name, const char *link) 472 { 473 itx_t *itx; 474 _lr_create_t *lr; 475 lr_create_t *lrc; 476 size_t namesize = strlen(name) + 1; 477 size_t linksize = strlen(link) + 1; 478 479 if (zil_replaying(zilog, tx)) 480 return; 481 482 itx = zil_itx_create(txtype, sizeof (*lrc) + namesize + linksize); 483 lrc = (lr_create_t *)&itx->itx_lr; 484 lr = &lrc->lr_create; 485 lr->lr_doid = dzp->z_id; 486 lr->lr_foid = zp->z_id; 487 lr->lr_uid = KUID_TO_SUID(ZTOUID(zp)); 488 lr->lr_gid = KGID_TO_SGID(ZTOGID(zp)); 489 lr->lr_mode = zp->z_mode; 490 (void) sa_lookup(zp->z_sa_hdl, SA_ZPL_GEN(ZTOZSB(zp)), &lr->lr_gen, 491 sizeof (uint64_t)); 492 (void) sa_lookup(zp->z_sa_hdl, SA_ZPL_CRTIME(ZTOZSB(zp)), 493 lr->lr_crtime, sizeof (uint64_t) * 2); 494 memcpy(&lrc->lr_data[0], name, namesize); 495 memcpy(&lrc->lr_data[namesize], link, linksize); 496 497 zil_itx_assign(zilog, itx, tx); 498 } 499 500 static void 501 do_zfs_log_rename(zilog_t *zilog, dmu_tx_t *tx, uint64_t txtype, znode_t *sdzp, 502 const char *sname, znode_t *tdzp, const char *dname, znode_t *szp) 503 { 504 itx_t *itx; 505 _lr_rename_t *lr; 506 lr_rename_t *lrr; 507 size_t snamesize = strlen(sname) + 1; 508 size_t dnamesize = strlen(dname) + 1; 509 510 if (zil_replaying(zilog, tx)) 511 return; 512 513 itx = zil_itx_create(txtype, sizeof (*lr) + snamesize + dnamesize); 514 lrr = (lr_rename_t *)&itx->itx_lr; 515 lr = &lrr->lr_rename; 516 lr->lr_sdoid = sdzp->z_id; 517 lr->lr_tdoid = tdzp->z_id; 518 memcpy(&lrr->lr_data[0], sname, snamesize); 519 memcpy(&lrr->lr_data[snamesize], dname, dnamesize); 520 itx->itx_oid = szp->z_id; 521 522 zil_itx_assign(zilog, itx, tx); 523 } 524 525 /* 526 * Handles TX_RENAME transactions. 527 */ 528 void 529 zfs_log_rename(zilog_t *zilog, dmu_tx_t *tx, uint64_t txtype, znode_t *sdzp, 530 const char *sname, znode_t *tdzp, const char *dname, znode_t *szp) 531 { 532 txtype |= TX_RENAME; 533 do_zfs_log_rename(zilog, tx, txtype, sdzp, sname, tdzp, dname, szp); 534 } 535 536 /* 537 * Handles TX_RENAME_EXCHANGE transactions. 538 */ 539 void 540 zfs_log_rename_exchange(zilog_t *zilog, dmu_tx_t *tx, uint64_t txtype, 541 znode_t *sdzp, const char *sname, znode_t *tdzp, const char *dname, 542 znode_t *szp) 543 { 544 txtype |= TX_RENAME_EXCHANGE; 545 do_zfs_log_rename(zilog, tx, txtype, sdzp, sname, tdzp, dname, szp); 546 } 547 548 /* 549 * Handles TX_RENAME_WHITEOUT transactions. 550 * 551 * Unfortunately we cannot reuse do_zfs_log_rename because we we need to call 552 * zfs_mknode() on replay which requires stashing bits as with TX_CREATE. 553 */ 554 void 555 zfs_log_rename_whiteout(zilog_t *zilog, dmu_tx_t *tx, uint64_t txtype, 556 znode_t *sdzp, const char *sname, znode_t *tdzp, const char *dname, 557 znode_t *szp, znode_t *wzp) 558 { 559 itx_t *itx; 560 lr_rename_whiteout_t *lr; 561 size_t snamesize = strlen(sname) + 1; 562 size_t dnamesize = strlen(dname) + 1; 563 564 if (zil_replaying(zilog, tx)) 565 return; 566 567 txtype |= TX_RENAME_WHITEOUT; 568 itx = zil_itx_create(txtype, sizeof (*lr) + snamesize + dnamesize); 569 lr = (lr_rename_whiteout_t *)&itx->itx_lr; 570 lr->lr_rename.lr_sdoid = sdzp->z_id; 571 lr->lr_rename.lr_tdoid = tdzp->z_id; 572 573 /* 574 * RENAME_WHITEOUT will create an entry at the source znode, so we need 575 * to store the same data that the equivalent call to zfs_log_create() 576 * would. 577 */ 578 lr->lr_wfoid = wzp->z_id; 579 LR_FOID_SET_SLOTS(lr->lr_wfoid, wzp->z_dnodesize >> DNODE_SHIFT); 580 (void) sa_lookup(wzp->z_sa_hdl, SA_ZPL_GEN(ZTOZSB(wzp)), &lr->lr_wgen, 581 sizeof (uint64_t)); 582 (void) sa_lookup(wzp->z_sa_hdl, SA_ZPL_CRTIME(ZTOZSB(wzp)), 583 lr->lr_wcrtime, sizeof (uint64_t) * 2); 584 lr->lr_wmode = wzp->z_mode; 585 lr->lr_wuid = (uint64_t)KUID_TO_SUID(ZTOUID(wzp)); 586 lr->lr_wgid = (uint64_t)KGID_TO_SGID(ZTOGID(wzp)); 587 588 /* 589 * This rdev will always be makdevice(0, 0) but because the ZIL log and 590 * replay code needs to be platform independent (and there is no 591 * platform independent makdev()) we need to copy the one created 592 * during the rename operation. 593 */ 594 (void) sa_lookup(wzp->z_sa_hdl, SA_ZPL_RDEV(ZTOZSB(wzp)), &lr->lr_wrdev, 595 sizeof (lr->lr_wrdev)); 596 597 memcpy(&lr->lr_data[0], sname, snamesize); 598 memcpy(&lr->lr_data[snamesize], dname, dnamesize); 599 itx->itx_oid = szp->z_id; 600 601 zil_itx_assign(zilog, itx, tx); 602 } 603 604 /* 605 * zfs_log_write() handles TX_WRITE transactions. The specified callback is 606 * called as soon as the write is on stable storage (be it via a DMU sync or a 607 * ZIL commit). 608 */ 609 static int64_t zfs_immediate_write_sz = 32768; 610 611 void 612 zfs_log_write(zilog_t *zilog, dmu_tx_t *tx, int txtype, 613 znode_t *zp, offset_t off, ssize_t resid, boolean_t commit, 614 boolean_t o_direct, zil_callback_t callback, void *callback_data) 615 { 616 dmu_buf_impl_t *db = (dmu_buf_impl_t *)sa_get_db(zp->z_sa_hdl); 617 uint32_t blocksize = zp->z_blksz; 618 itx_wr_state_t write_state; 619 uint64_t gen = 0; 620 ssize_t size = resid; 621 622 if (zil_replaying(zilog, tx) || zp->z_unlinked || 623 zfs_xattr_owner_unlinked(zp)) { 624 if (callback != NULL) 625 callback(callback_data); 626 return; 627 } 628 629 if (zilog->zl_logbias == ZFS_LOGBIAS_THROUGHPUT || o_direct) 630 write_state = WR_INDIRECT; 631 else if (!spa_has_slogs(zilog->zl_spa) && 632 resid >= zfs_immediate_write_sz) 633 write_state = WR_INDIRECT; 634 else if (commit) 635 write_state = WR_COPIED; 636 else 637 write_state = WR_NEED_COPY; 638 639 (void) sa_lookup(zp->z_sa_hdl, SA_ZPL_GEN(ZTOZSB(zp)), &gen, 640 sizeof (gen)); 641 642 while (resid) { 643 itx_t *itx; 644 lr_write_t *lr; 645 itx_wr_state_t wr_state = write_state; 646 ssize_t len = resid; 647 648 /* 649 * A WR_COPIED record must fit entirely in one log block. 650 * Large writes can use WR_NEED_COPY, which the ZIL will 651 * split into multiple records across several log blocks 652 * if necessary. 653 */ 654 if (wr_state == WR_COPIED && 655 resid > zil_max_copied_data(zilog)) 656 wr_state = WR_NEED_COPY; 657 else if (wr_state == WR_INDIRECT) 658 len = MIN(blocksize - P2PHASE(off, blocksize), resid); 659 660 itx = zil_itx_create(txtype, sizeof (*lr) + 661 (wr_state == WR_COPIED ? len : 0)); 662 lr = (lr_write_t *)&itx->itx_lr; 663 664 /* 665 * For WR_COPIED records, copy the data into the lr_write_t. 666 */ 667 if (wr_state == WR_COPIED) { 668 int err; 669 DB_DNODE_ENTER(db); 670 err = dmu_read_by_dnode(DB_DNODE(db), off, len, 671 &lr->lr_data[0], DMU_READ_NO_PREFETCH); 672 DB_DNODE_EXIT(db); 673 if (err != 0) { 674 zil_itx_destroy(itx); 675 itx = zil_itx_create(txtype, sizeof (*lr)); 676 lr = (lr_write_t *)&itx->itx_lr; 677 wr_state = WR_NEED_COPY; 678 } 679 } 680 681 itx->itx_wr_state = wr_state; 682 lr->lr_foid = zp->z_id; 683 lr->lr_offset = off; 684 lr->lr_length = len; 685 lr->lr_blkoff = 0; 686 BP_ZERO(&lr->lr_blkptr); 687 688 itx->itx_private = ZTOZSB(zp); 689 itx->itx_sync = (zp->z_sync_cnt != 0); 690 itx->itx_gen = gen; 691 692 itx->itx_callback = callback; 693 itx->itx_callback_data = callback_data; 694 zil_itx_assign(zilog, itx, tx); 695 696 off += len; 697 resid -= len; 698 } 699 700 if (write_state == WR_COPIED || write_state == WR_NEED_COPY) { 701 dsl_pool_wrlog_count(zilog->zl_dmu_pool, size, tx->tx_txg); 702 } 703 } 704 705 /* 706 * Handles TX_TRUNCATE transactions. 707 */ 708 void 709 zfs_log_truncate(zilog_t *zilog, dmu_tx_t *tx, int txtype, 710 znode_t *zp, uint64_t off, uint64_t len) 711 { 712 itx_t *itx; 713 lr_truncate_t *lr; 714 715 if (zil_replaying(zilog, tx) || zp->z_unlinked || 716 zfs_xattr_owner_unlinked(zp)) 717 return; 718 719 itx = zil_itx_create(txtype, sizeof (*lr)); 720 lr = (lr_truncate_t *)&itx->itx_lr; 721 lr->lr_foid = zp->z_id; 722 lr->lr_offset = off; 723 lr->lr_length = len; 724 725 itx->itx_sync = (zp->z_sync_cnt != 0); 726 zil_itx_assign(zilog, itx, tx); 727 } 728 729 /* 730 * Handles TX_SETATTR transactions. 731 */ 732 void 733 zfs_log_setattr(zilog_t *zilog, dmu_tx_t *tx, int txtype, 734 znode_t *zp, vattr_t *vap, uint_t mask_applied, zfs_fuid_info_t *fuidp) 735 { 736 itx_t *itx; 737 lr_setattr_t *lr; 738 xvattr_t *xvap = (xvattr_t *)vap; 739 size_t recsize = sizeof (lr_setattr_t); 740 uint8_t *start; 741 742 if (zil_replaying(zilog, tx) || zp->z_unlinked) 743 return; 744 745 /* 746 * If XVATTR set, then log record size needs to allow 747 * for lr_attr_t + xvattr mask, mapsize and create time 748 * plus actual attribute values 749 */ 750 if (vap->va_mask & ATTR_XVATTR) 751 recsize = sizeof (*lr) + ZIL_XVAT_SIZE(xvap->xva_mapsize); 752 753 if (fuidp) 754 recsize += fuidp->z_domain_str_sz; 755 756 itx = zil_itx_create(txtype, recsize); 757 lr = (lr_setattr_t *)&itx->itx_lr; 758 lr->lr_foid = zp->z_id; 759 lr->lr_mask = (uint64_t)mask_applied; 760 lr->lr_mode = (uint64_t)vap->va_mode; 761 if ((mask_applied & ATTR_UID) && IS_EPHEMERAL(vap->va_uid)) 762 lr->lr_uid = fuidp->z_fuid_owner; 763 else 764 lr->lr_uid = (uint64_t)vap->va_uid; 765 766 if ((mask_applied & ATTR_GID) && IS_EPHEMERAL(vap->va_gid)) 767 lr->lr_gid = fuidp->z_fuid_group; 768 else 769 lr->lr_gid = (uint64_t)vap->va_gid; 770 771 lr->lr_size = (uint64_t)vap->va_size; 772 ZFS_TIME_ENCODE(&vap->va_atime, lr->lr_atime); 773 ZFS_TIME_ENCODE(&vap->va_mtime, lr->lr_mtime); 774 start = &lr->lr_data[0]; 775 if (vap->va_mask & ATTR_XVATTR) { 776 zfs_log_xvattr((lr_attr_t *)start, xvap); 777 start = &lr->lr_data[ZIL_XVAT_SIZE(xvap->xva_mapsize)]; 778 } 779 780 /* 781 * Now stick on domain information if any on end 782 */ 783 784 if (fuidp) 785 (void) zfs_log_fuid_domains(fuidp, start); 786 787 itx->itx_sync = (zp->z_sync_cnt != 0); 788 zil_itx_assign(zilog, itx, tx); 789 } 790 791 /* 792 * Handles TX_SETSAXATTR transactions. 793 */ 794 void 795 zfs_log_setsaxattr(zilog_t *zilog, dmu_tx_t *tx, int txtype, 796 znode_t *zp, const char *name, const void *value, size_t size) 797 { 798 itx_t *itx; 799 lr_setsaxattr_t *lr; 800 size_t recsize = sizeof (lr_setsaxattr_t); 801 int namelen; 802 803 if (zil_replaying(zilog, tx) || zp->z_unlinked) 804 return; 805 806 namelen = strlen(name) + 1; 807 recsize += (namelen + size); 808 itx = zil_itx_create(txtype, recsize); 809 lr = (lr_setsaxattr_t *)&itx->itx_lr; 810 lr->lr_foid = zp->z_id; 811 memcpy(&lr->lr_data[0], name, namelen); 812 if (value != NULL) { 813 memcpy(&lr->lr_data[namelen], value, size); 814 lr->lr_size = size; 815 } else { 816 lr->lr_size = 0; 817 } 818 819 itx->itx_sync = (zp->z_sync_cnt != 0); 820 zil_itx_assign(zilog, itx, tx); 821 } 822 823 /* 824 * Handles TX_ACL transactions. 825 */ 826 void 827 zfs_log_acl(zilog_t *zilog, dmu_tx_t *tx, znode_t *zp, 828 vsecattr_t *vsecp, zfs_fuid_info_t *fuidp) 829 { 830 itx_t *itx; 831 lr_acl_v0_t *lrv0; 832 lr_acl_t *lr; 833 int txtype; 834 int lrsize; 835 size_t txsize; 836 size_t aclbytes = vsecp->vsa_aclentsz; 837 838 if (zil_replaying(zilog, tx) || zp->z_unlinked) 839 return; 840 841 txtype = (ZTOZSB(zp)->z_version < ZPL_VERSION_FUID) ? 842 TX_ACL_V0 : TX_ACL; 843 844 if (txtype == TX_ACL) 845 lrsize = sizeof (*lr); 846 else 847 lrsize = sizeof (*lrv0); 848 849 txsize = lrsize + 850 ((txtype == TX_ACL) ? ZIL_ACE_LENGTH(aclbytes) : aclbytes) + 851 (fuidp ? fuidp->z_domain_str_sz : 0) + 852 sizeof (uint64_t) * (fuidp ? fuidp->z_fuid_cnt : 0); 853 854 itx = zil_itx_create(txtype, txsize); 855 856 lr = (lr_acl_t *)&itx->itx_lr; 857 lr->lr_foid = zp->z_id; 858 if (txtype == TX_ACL) { 859 lr->lr_acl_bytes = aclbytes; 860 lr->lr_domcnt = fuidp ? fuidp->z_domain_cnt : 0; 861 lr->lr_fuidcnt = fuidp ? fuidp->z_fuid_cnt : 0; 862 if (vsecp->vsa_mask & VSA_ACE_ACLFLAGS) 863 lr->lr_acl_flags = (uint64_t)vsecp->vsa_aclflags; 864 else 865 lr->lr_acl_flags = 0; 866 } 867 lr->lr_aclcnt = (uint64_t)vsecp->vsa_aclcnt; 868 869 if (txtype == TX_ACL_V0) { 870 lrv0 = (lr_acl_v0_t *)lr; 871 memcpy(&lrv0->lr_data[0], vsecp->vsa_aclentp, aclbytes); 872 } else { 873 uint8_t *start = &lr->lr_data[0]; 874 875 memcpy(start, vsecp->vsa_aclentp, aclbytes); 876 877 start = &lr->lr_data[ZIL_ACE_LENGTH(aclbytes)]; 878 879 if (fuidp) { 880 start = zfs_log_fuid_ids(fuidp, start); 881 (void) zfs_log_fuid_domains(fuidp, start); 882 } 883 } 884 885 itx->itx_sync = (zp->z_sync_cnt != 0); 886 zil_itx_assign(zilog, itx, tx); 887 } 888 889 /* 890 * Handles TX_CLONE_RANGE transactions. 891 */ 892 void 893 zfs_log_clone_range(zilog_t *zilog, dmu_tx_t *tx, int txtype, znode_t *zp, 894 uint64_t off, uint64_t len, uint64_t blksz, const blkptr_t *bps, 895 size_t nbps) 896 { 897 itx_t *itx; 898 lr_clone_range_t *lr; 899 uint64_t partlen, max_log_data; 900 size_t partnbps; 901 902 if (zil_replaying(zilog, tx) || zp->z_unlinked) 903 return; 904 905 max_log_data = zil_max_log_data(zilog, sizeof (lr_clone_range_t)); 906 907 while (nbps > 0) { 908 partnbps = MIN(nbps, max_log_data / sizeof (bps[0])); 909 partlen = partnbps * blksz; 910 ASSERT3U(partlen, <, len + blksz); 911 partlen = MIN(partlen, len); 912 913 itx = zil_itx_create(txtype, 914 sizeof (*lr) + sizeof (bps[0]) * partnbps); 915 lr = (lr_clone_range_t *)&itx->itx_lr; 916 lr->lr_foid = zp->z_id; 917 lr->lr_offset = off; 918 lr->lr_length = partlen; 919 lr->lr_blksz = blksz; 920 lr->lr_nbps = partnbps; 921 memcpy(lr->lr_bps, bps, sizeof (bps[0]) * partnbps); 922 923 itx->itx_sync = (zp->z_sync_cnt != 0); 924 925 zil_itx_assign(zilog, itx, tx); 926 927 bps += partnbps; 928 ASSERT3U(nbps, >=, partnbps); 929 nbps -= partnbps; 930 off += partlen; 931 ASSERT3U(len, >=, partlen); 932 len -= partlen; 933 } 934 } 935 936 ZFS_MODULE_PARAM(zfs, zfs_, immediate_write_sz, S64, ZMOD_RW, 937 "Largest data block to write to zil"); 938