1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License, Version 1.0 only 6 * (the "License"). You may not use this file except in compliance 7 * with the License. 8 * 9 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 10 * or http://www.opensolaris.org/os/licensing. 11 * See the License for the specific language governing permissions 12 * and limitations under the License. 13 * 14 * When distributing Covered Code, include this CDDL HEADER in each 15 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 16 * If applicable, add the following below this CDDL HEADER, with the 17 * fields enclosed by brackets "[]" replaced with your own identifying 18 * information: Portions Copyright [yyyy] [name of copyright owner] 19 * 20 * CDDL HEADER END 21 */ 22 /* 23 * Copyright 2005 Sun Microsystems, Inc. All rights reserved. 24 * Use is subject to license terms. 25 */ 26 27 #pragma ident "%Z%%M% %I% %E% SMI" 28 29 /* 30 * FMD Log File Subsystem 31 * 32 * Events are written to one of two log files as they are received or created; 33 * the error log tracks all ereport.* events received on the inbound event 34 * transport, and the fault log tracks all list.* events generated by fmd or 35 * its client modules. In addition, we use the same log file format to cache 36 * state and events associated with ASRUs that are named in a diagnosis. 37 * 38 * The log files use the exacct format manipulated by libexacct(3LIB) and 39 * originally defined in PSARC 1999/119. However, the exacct library was 40 * designed primarily for read-only clients and without the synchronous i/o 41 * considerations and seeking required for fmd, so we use libexacct here only 42 * to read and write the file headers and to pack data from memory into a file 43 * bytestream. All of the i/o and file offset manipulations are performed by 44 * the fmd code below. Our exacct file management uses the following grammar: 45 * 46 * file := hdr toc event* 47 * hdr := EXD_FMA_LABEL EXD_FMA_VERSION EXD_FMA_OSREL EXD_FMA_OSVER 48 * EXD_FMA_PLAT EXD_FMA_UUID 49 * toc := EXD_FMA_OFFSET 50 * event := EXD_FMA_TODSEC EXD_FMA_TODNSEC EXD_FMA_NVLIST evref* or legacy evref 51 * evref := EXD_FMA_UUID EXD_FMA_OFFSET 52 * legacy evref := EXD_FMA_MAJOR EXD_FMA_MINOR EXD_FMA_INODE EXD_FMA_OFFSET 53 * 54 * Any event can be uniquely identified by the tuple (file, offset) where file 55 * is encoded as (uuid) when we are cross-linking files. For legacy file 56 * formats we still support encoding the reference as (major, minor, inode). 57 * Note that we break out of the file's dev_t into its two 32-bit components to 58 * permit development of either 32-bit or 64-bit log readers and writers; the 59 * LFS APIs do not yet export a 64-bit dev_t to fstat64(), so there is no way 60 * for a 32-bit application to retrieve and store a 64-bit dev_t. 61 * 62 * In order to replay events in the event of an fmd crash, events are initially 63 * written to the error log using the group catalog tag EXD_GROUP_RFMA by the 64 * fmd_log_append() function. Later, once an event transitions from the 65 * received state to one of its other states (see fmd_event.c for details), 66 * fmd_log_commit() is used to overwrite the tag with EXD_GROUP_FMA, indicating 67 * that the event is fully processed and no longer needs to be replayed. 68 */ 69 70 #include <sys/types.h> 71 #include <sys/mkdev.h> 72 #include <sys/statvfs.h> 73 #include <sys/fm/protocol.h> 74 #include <sys/exacct_impl.h> 75 #include <uuid/uuid.h> 76 77 #include <unistd.h> 78 #include <limits.h> 79 #include <fcntl.h> 80 #include <ctype.h> 81 82 #include <fmd_alloc.h> 83 #include <fmd_error.h> 84 #include <fmd_string.h> 85 #include <fmd_event.h> 86 #include <fmd_conf.h> 87 #include <fmd_subr.h> 88 #include <fmd_case.h> 89 #include <fmd_log.h> 90 91 #include <fmd.h> 92 93 #define CAT_FMA_RGROUP (EXT_GROUP | EXC_DEFAULT | EXD_GROUP_RFMA) 94 #define CAT_FMA_GROUP (EXT_GROUP | EXC_DEFAULT | EXD_GROUP_FMA) 95 96 #define CAT_FMA_LABEL (EXT_STRING | EXC_DEFAULT | EXD_FMA_LABEL) 97 #define CAT_FMA_VERSION (EXT_STRING | EXC_DEFAULT | EXD_FMA_VERSION) 98 #define CAT_FMA_OSREL (EXT_STRING | EXC_DEFAULT | EXD_FMA_OSREL) 99 #define CAT_FMA_OSVER (EXT_STRING | EXC_DEFAULT | EXD_FMA_OSVER) 100 #define CAT_FMA_PLAT (EXT_STRING | EXC_DEFAULT | EXD_FMA_PLAT) 101 #define CAT_FMA_UUID (EXT_STRING | EXC_DEFAULT | EXD_FMA_UUID) 102 #define CAT_FMA_TODSEC (EXT_UINT64 | EXC_DEFAULT | EXD_FMA_TODSEC) 103 #define CAT_FMA_TODNSEC (EXT_UINT64 | EXC_DEFAULT | EXD_FMA_TODNSEC) 104 #define CAT_FMA_NVLIST (EXT_RAW | EXC_DEFAULT | EXD_FMA_NVLIST) 105 #define CAT_FMA_MAJOR (EXT_UINT32 | EXC_DEFAULT | EXD_FMA_MAJOR) 106 #define CAT_FMA_MINOR (EXT_UINT32 | EXC_DEFAULT | EXD_FMA_MINOR) 107 #define CAT_FMA_INODE (EXT_UINT64 | EXC_DEFAULT | EXD_FMA_INODE) 108 #define CAT_FMA_OFFSET (EXT_UINT64 | EXC_DEFAULT | EXD_FMA_OFFSET) 109 110 static ssize_t 111 fmd_log_write(fmd_log_t *lp, const void *buf, size_t n) 112 { 113 ssize_t resid = n; 114 ssize_t len; 115 116 ASSERT(MUTEX_HELD(&lp->log_lock)); 117 118 while (resid != 0) { 119 if ((len = write(lp->log_fd, buf, resid)) <= 0) 120 break; 121 122 resid -= len; 123 buf = (char *)buf + len; 124 } 125 126 if (resid == n && n != 0) 127 return (-1); 128 129 return (n - resid); 130 } 131 132 static int 133 fmd_log_write_hdr(fmd_log_t *lp, const char *tag) 134 { 135 ea_object_t hdr, toc, i0, i1, i2, i3, i4, i5, i6; 136 const char *osrel, *osver, *plat; 137 off64_t off = 0; 138 int err = 0; 139 uuid_t uuid; 140 141 (void) fmd_conf_getprop(fmd.d_conf, "osrelease", &osrel); 142 (void) fmd_conf_getprop(fmd.d_conf, "osversion", &osver); 143 (void) fmd_conf_getprop(fmd.d_conf, "platform", &plat); 144 (void) fmd_conf_getprop(fmd.d_conf, "uuidlen", &lp->log_uuidlen); 145 146 lp->log_uuid = fmd_zalloc(lp->log_uuidlen + 1, FMD_SLEEP); 147 uuid_generate(uuid); 148 uuid_unparse(uuid, lp->log_uuid); 149 150 err |= ea_set_group(&hdr, CAT_FMA_GROUP); 151 err |= ea_set_group(&toc, CAT_FMA_GROUP); 152 153 err |= ea_set_item(&i0, CAT_FMA_LABEL, tag, 0); 154 err |= ea_set_item(&i1, CAT_FMA_VERSION, fmd.d_version, 0); 155 err |= ea_set_item(&i2, CAT_FMA_OSREL, osrel, 0); 156 err |= ea_set_item(&i3, CAT_FMA_OSVER, osver, 0); 157 err |= ea_set_item(&i4, CAT_FMA_PLAT, plat, 0); 158 err |= ea_set_item(&i5, CAT_FMA_UUID, lp->log_uuid, 0); 159 err |= ea_set_item(&i6, CAT_FMA_OFFSET, &off, 0); 160 161 (void) ea_attach_to_group(&hdr, &i0); 162 (void) ea_attach_to_group(&hdr, &i1); 163 (void) ea_attach_to_group(&hdr, &i2); 164 (void) ea_attach_to_group(&hdr, &i3); 165 (void) ea_attach_to_group(&hdr, &i4); 166 (void) ea_attach_to_group(&hdr, &i5); 167 (void) ea_attach_to_group(&toc, &i6); 168 169 if (err == 0) { 170 size_t hdr_size = ea_pack_object(&hdr, NULL, 0); 171 size_t toc_size = ea_pack_object(&toc, NULL, 0); 172 173 size_t size = hdr_size + toc_size; 174 void *buf = fmd_alloc(size, FMD_SLEEP); 175 176 (void) ea_pack_object(&hdr, buf, hdr_size); 177 (void) ea_pack_object(&toc, (char *)buf + hdr_size, toc_size); 178 179 if ((lp->log_off = lseek64(lp->log_fd, 0, SEEK_END)) == -1L) 180 fmd_panic("failed to seek log %s", lp->log_name); 181 182 if (fmd_log_write(lp, buf, size) != size) 183 err = errno; /* save errno for fmd_set_errno() below */ 184 185 fmd_free(buf, size); 186 187 lp->log_toc = lp->log_off + hdr_size; 188 lp->log_beg = lp->log_off + hdr_size + toc_size; 189 lp->log_off = lp->log_off + hdr_size + toc_size; 190 191 if (lp->log_off != lseek64(lp->log_fd, 0, SEEK_END)) 192 fmd_panic("eof off != log_off 0x%llx\n", lp->log_off); 193 } else 194 err = EFMD_LOG_EXACCT; 195 196 (void) ea_free_item(&i0, EUP_ALLOC); 197 (void) ea_free_item(&i1, EUP_ALLOC); 198 (void) ea_free_item(&i2, EUP_ALLOC); 199 (void) ea_free_item(&i3, EUP_ALLOC); 200 (void) ea_free_item(&i4, EUP_ALLOC); 201 (void) ea_free_item(&i5, EUP_ALLOC); 202 (void) ea_free_item(&i6, EUP_ALLOC); 203 204 return (err ? fmd_set_errno(err) : 0); 205 } 206 207 static int 208 fmd_log_check_err(fmd_log_t *lp, int err, const char *msg) 209 { 210 int eaerr = ea_error(); 211 char buf[BUFSIZ]; 212 213 (void) snprintf(buf, sizeof (buf), "%s: %s: %s\n", 214 lp->log_name, msg, eaerr != EXR_OK ? 215 fmd_ea_strerror(eaerr) : "catalog tag mismatch"); 216 217 fmd_error(err, buf); 218 return (fmd_set_errno(err)); 219 } 220 221 static int 222 fmd_log_check_hdr(fmd_log_t *lp, const char *tag) 223 { 224 int got_version = 0, got_label = 0; 225 ea_object_t *grp, *obj; 226 off64_t hdr_off, hdr_size; 227 int dvers, fvers; 228 const char *p; 229 230 ea_clear(&lp->log_ea); /* resync exacct file */ 231 232 if ((hdr_off = lseek64(lp->log_fd, 0, SEEK_CUR)) == -1L) 233 fmd_panic("failed to seek log %s", lp->log_name); 234 235 /* 236 * Read the first group of log meta-data: the write-once read-only 237 * file header. We read all records in this group, ignoring all but 238 * the VERSION and LABEL, which are required and must be verified. 239 */ 240 if ((grp = ea_get_object_tree(&lp->log_ea, 1)) == NULL || 241 grp->eo_catalog != CAT_FMA_GROUP) { 242 ea_free_object(grp, EUP_ALLOC); 243 return (fmd_log_check_err(lp, EFMD_LOG_INVAL, 244 "invalid fma hdr record group")); 245 } 246 247 for (obj = grp->eo_group.eg_objs; obj != NULL; obj = obj->eo_next) { 248 switch (obj->eo_catalog) { 249 case CAT_FMA_VERSION: 250 for (dvers = 0, p = fmd.d_version; 251 *p != '\0'; p++) { 252 if (isdigit(*p)) 253 dvers = dvers * 10 + (*p - '0'); 254 else 255 break; 256 } 257 258 for (fvers = 0, p = obj->eo_item.ei_string; 259 *p != '\0'; p++) { 260 if (isdigit(*p)) 261 fvers = fvers * 10 + (*p - '0'); 262 else 263 break; 264 } 265 266 if (fvers > dvers) { 267 fmd_error(EFMD_LOG_INVAL, "%s: log version " 268 "%s is not supported by this daemon\n", 269 lp->log_name, obj->eo_item.ei_string); 270 ea_free_object(grp, EUP_ALLOC); 271 return (fmd_set_errno(EFMD_LOG_VERSION)); 272 } 273 274 got_version++; 275 break; 276 277 case CAT_FMA_LABEL: 278 if (strcmp(obj->eo_item.ei_string, tag) != 0) { 279 fmd_error(EFMD_LOG_INVAL, "%s: log tag '%s' " 280 "does not matched expected tag '%s'\n", 281 lp->log_name, obj->eo_item.ei_string, tag); 282 ea_free_object(grp, EUP_ALLOC); 283 return (fmd_set_errno(EFMD_LOG_INVAL)); 284 } 285 got_label++; 286 break; 287 case CAT_FMA_UUID: 288 lp->log_uuid = fmd_strdup(obj->eo_item.ei_string, 289 FMD_SLEEP); 290 lp->log_uuidlen = strlen(lp->log_uuid); 291 break; 292 } 293 } 294 295 hdr_size = ea_pack_object(grp, NULL, 0); 296 ea_free_object(grp, EUP_ALLOC); 297 298 if (!got_version || !got_label) { 299 fmd_error(EFMD_LOG_INVAL, "%s: fmd hdr record group did not " 300 "include mandatory version and/or label\n", lp->log_name); 301 return (fmd_set_errno(EFMD_LOG_INVAL)); 302 } 303 304 /* 305 * Read the second group of log meta-data: the table of contents. We 306 * expect this group to contain an OFFSET object indicating the current 307 * value of log_skip. We save this in our fmd_log_t and then return. 308 */ 309 if ((grp = ea_get_object_tree(&lp->log_ea, 1)) == NULL || 310 grp->eo_catalog != CAT_FMA_GROUP || grp->eo_group.eg_nobjs < 1 || 311 grp->eo_group.eg_objs->eo_catalog != CAT_FMA_OFFSET) { 312 ea_free_object(grp, EUP_ALLOC); 313 return (fmd_log_check_err(lp, EFMD_LOG_INVAL, 314 "invalid fma toc record group")); 315 } 316 317 lp->log_toc = hdr_off + hdr_size; 318 lp->log_beg = hdr_off + hdr_size + ea_pack_object(grp, NULL, 0); 319 lp->log_off = lseek64(lp->log_fd, 0, SEEK_END); 320 lp->log_skip = grp->eo_group.eg_objs->eo_item.ei_uint64; 321 322 if (lp->log_skip > lp->log_off) { 323 fmd_error(EFMD_LOG_INVAL, "%s: skip %llx exceeds file size; " 324 "resetting to zero\n", lp->log_name, lp->log_skip); 325 lp->log_skip = 0; 326 } 327 328 ea_free_object(grp, EUP_ALLOC); 329 return (0); 330 } 331 332 static int 333 fmd_log_open_exacct(fmd_log_t *lp, int aflags, int oflags) 334 { 335 int fd = dup(lp->log_fd); 336 const char *creator; 337 338 (void) fmd_conf_getprop(fmd.d_conf, "log.creator", &creator); 339 340 if (ea_fdopen(&lp->log_ea, fd, creator, aflags, oflags) != 0) { 341 fmd_error(EFMD_LOG_EXACCT, "%s: failed to open log file: %s\n", 342 lp->log_name, fmd_ea_strerror(ea_error())); 343 (void) close(fd); 344 return (fmd_set_errno(EFMD_LOG_EXACCT)); 345 } 346 347 lp->log_flags |= FMD_LF_EAOPEN; 348 return (0); 349 } 350 351 static fmd_log_t * 352 fmd_log_xopen(const char *root, const char *name, const char *tag, int oflags) 353 { 354 fmd_log_t *lp = fmd_zalloc(sizeof (fmd_log_t), FMD_SLEEP); 355 356 char buf[PATH_MAX]; 357 size_t len; 358 int err; 359 360 (void) pthread_mutex_init(&lp->log_lock, NULL); 361 (void) pthread_cond_init(&lp->log_cv, NULL); 362 (void) pthread_mutex_lock(&lp->log_lock); 363 364 len = strlen(root) + strlen(name) + 2; /* for "/" and "\0" */ 365 lp->log_name = fmd_alloc(len, FMD_SLEEP); 366 (void) snprintf(lp->log_name, len, "%s/%s", root, name); 367 lp->log_tag = fmd_strdup(tag, FMD_SLEEP); 368 (void) fmd_conf_getprop(fmd.d_conf, "log.minfree", &lp->log_minfree); 369 370 if (strcmp(lp->log_tag, FMD_LOG_ERROR) == 0) 371 lp->log_flags |= FMD_LF_REPLAY; 372 373 top: 374 if ((lp->log_fd = open64(lp->log_name, oflags, 0644)) == -1 || 375 fstat64(lp->log_fd, &lp->log_stat) == -1) { 376 fmd_error(EFMD_LOG_OPEN, "failed to open log %s", lp->log_name); 377 fmd_log_close(lp); 378 return (NULL); 379 } 380 381 /* 382 * If our open() created the log file, use libexacct to write a header 383 * and position the file just after the header (EO_TAIL). If the log 384 * file already existed, use libexacct to validate the header and again 385 * position the file just after the header (EO_HEAD). Note that we lie 386 * to libexacct about 'oflags' in order to achieve the desired result. 387 */ 388 if (lp->log_stat.st_size == 0) { 389 err = fmd_log_open_exacct(lp, EO_VALID_HDR | EO_TAIL, 390 O_CREAT | O_WRONLY) || fmd_log_write_hdr(lp, tag); 391 } else { 392 err = fmd_log_open_exacct(lp, EO_VALID_HDR | EO_HEAD, 393 O_RDONLY) || fmd_log_check_hdr(lp, tag); 394 } 395 396 /* 397 * If ea_fdopen() failed and the log was pre-existing, attempt to move 398 * it aside and start a new one. If we created the log but failed to 399 * initialize it, then we have no choice but to give up (e.g. EROFS). 400 */ 401 if (err) { 402 fmd_error(EFMD_LOG_OPEN, 403 "failed to initialize log %s", lp->log_name); 404 405 if (lp->log_flags & FMD_LF_EAOPEN) { 406 lp->log_flags &= ~FMD_LF_EAOPEN; 407 (void) ea_close(&lp->log_ea); 408 } 409 410 (void) close(lp->log_fd); 411 lp->log_fd = -1; 412 413 if (lp->log_stat.st_size != 0 && snprintf(buf, 414 sizeof (buf), "%s-", lp->log_name) < PATH_MAX && 415 rename(lp->log_name, buf) == 0) { 416 TRACE((FMD_DBG_LOG, "mv %s to %s", lp->log_name, buf)); 417 if (oflags & O_CREAT) 418 goto top; 419 } 420 421 fmd_log_close(lp); 422 return (NULL); 423 } 424 425 lp->log_refs++; 426 (void) pthread_mutex_unlock(&lp->log_lock); 427 428 return (lp); 429 } 430 431 fmd_log_t * 432 fmd_log_tryopen(const char *root, const char *name, const char *tag) 433 { 434 return (fmd_log_xopen(root, name, tag, O_RDWR | O_SYNC)); 435 } 436 437 fmd_log_t * 438 fmd_log_open(const char *root, const char *name, const char *tag) 439 { 440 return (fmd_log_xopen(root, name, tag, O_RDWR | O_CREAT | O_SYNC)); 441 } 442 443 void 444 fmd_log_close(fmd_log_t *lp) 445 { 446 ASSERT(MUTEX_HELD(&lp->log_lock)); 447 ASSERT(lp->log_refs == 0); 448 449 if ((lp->log_flags & FMD_LF_EAOPEN) && ea_close(&lp->log_ea) != 0) { 450 fmd_error(EFMD_LOG_CLOSE, "failed to close log %s: %s\n", 451 lp->log_name, fmd_ea_strerror(ea_error())); 452 } 453 454 if (lp->log_fd >= 0 && close(lp->log_fd) != 0) { 455 fmd_error(EFMD_LOG_CLOSE, 456 "failed to close log %s", lp->log_name); 457 } 458 459 fmd_strfree(lp->log_name); 460 fmd_strfree(lp->log_tag); 461 if (lp->log_uuid != NULL) 462 fmd_free(lp->log_uuid, lp->log_uuidlen + 1); 463 464 fmd_free(lp, sizeof (fmd_log_t)); 465 } 466 467 void 468 fmd_log_hold_pending(fmd_log_t *lp) 469 { 470 (void) pthread_mutex_lock(&lp->log_lock); 471 472 lp->log_refs++; 473 ASSERT(lp->log_refs != 0); 474 475 if (lp->log_flags & FMD_LF_REPLAY) { 476 lp->log_pending++; 477 ASSERT(lp->log_pending != 0); 478 } 479 480 (void) pthread_mutex_unlock(&lp->log_lock); 481 } 482 483 void 484 fmd_log_hold(fmd_log_t *lp) 485 { 486 (void) pthread_mutex_lock(&lp->log_lock); 487 lp->log_refs++; 488 ASSERT(lp->log_refs != 0); 489 (void) pthread_mutex_unlock(&lp->log_lock); 490 } 491 492 void 493 fmd_log_rele(fmd_log_t *lp) 494 { 495 (void) pthread_mutex_lock(&lp->log_lock); 496 ASSERT(lp->log_refs != 0); 497 498 if (--lp->log_refs == 0) 499 fmd_log_close(lp); 500 else 501 (void) pthread_mutex_unlock(&lp->log_lock); 502 } 503 504 void 505 fmd_log_append(fmd_log_t *lp, fmd_event_t *e, fmd_case_t *cp) 506 { 507 fmd_event_impl_t *ep = (fmd_event_impl_t *)e; 508 fmd_case_impl_t *cip = (fmd_case_impl_t *)cp; 509 int err = 0; 510 511 ea_object_t grp0, grp1, i0, i1, i2, *items; 512 ea_object_t **fe = NULL; 513 size_t nvsize, easize, itsize, frsize; 514 char *nvbuf, *eabuf; 515 statvfs64_t stv; 516 517 (void) pthread_mutex_lock(&ep->ev_lock); 518 519 ASSERT(ep->ev_flags & FMD_EVF_VOLATILE); 520 ASSERT(ep->ev_log == NULL); 521 522 (void) nvlist_size(ep->ev_nvl, &nvsize, NV_ENCODE_XDR); 523 nvbuf = fmd_alloc(nvsize, FMD_SLEEP); 524 (void) nvlist_pack(ep->ev_nvl, &nvbuf, &nvsize, NV_ENCODE_XDR, 0); 525 526 if (lp->log_flags & FMD_LF_REPLAY) 527 err |= ea_set_group(&grp0, CAT_FMA_RGROUP); 528 else 529 err |= ea_set_group(&grp0, CAT_FMA_GROUP); 530 531 err |= ea_set_item(&i0, CAT_FMA_TODSEC, &ep->ev_time.ftv_sec, 0); 532 err |= ea_set_item(&i1, CAT_FMA_TODNSEC, &ep->ev_time.ftv_nsec, 0); 533 err |= ea_set_item(&i2, CAT_FMA_NVLIST, nvbuf, nvsize); 534 535 if (err != 0) { 536 (void) pthread_mutex_unlock(&ep->ev_lock); 537 err = EFMD_LOG_EXACCT; 538 goto exerr; 539 } 540 541 (void) ea_attach_to_group(&grp0, &i0); 542 (void) ea_attach_to_group(&grp0, &i1); 543 (void) ea_attach_to_group(&grp0, &i2); 544 545 /* 546 * If this event has a case associated with it (i.e. it is a list), 547 * then allocate a block of ea_object_t's and fill in a group for 548 * each event saved in the case's item list. For each such group, 549 * we attach it to grp1, which in turn will be attached to grp0. 550 */ 551 if (cp != NULL) { 552 ea_object_t *egrp, *ip, **fp; 553 fmd_event_impl_t *eip; 554 fmd_case_item_t *cit; 555 556 (void) ea_set_group(&grp1, CAT_FMA_GROUP); 557 frsize = sizeof (ea_object_t *) * cip->ci_nitems; 558 itsize = sizeof (ea_object_t) * cip->ci_nitems * 5; 559 items = ip = fmd_alloc(itsize, FMD_SLEEP); 560 561 for (cit = cip->ci_items; cit != NULL; cit = cit->cit_next) { 562 major_t maj; 563 minor_t min; 564 565 eip = (fmd_event_impl_t *)cit->cit_event; 566 567 if (eip->ev_log == NULL) 568 continue; /* event was never logged */ 569 570 maj = major(eip->ev_log->log_stat.st_dev); 571 min = minor(eip->ev_log->log_stat.st_dev); 572 573 (void) ea_set_group(ip, CAT_FMA_GROUP); 574 egrp = ip++; /* first obj is group */ 575 576 /* 577 * If the event log file is in legacy format, 578 * then write the xref to the file in the legacy 579 * maj/min/inode method else write it using the 580 * file uuid. 581 */ 582 if (eip->ev_log->log_uuid == NULL) { 583 (void) ea_set_item(ip, CAT_FMA_MAJOR, &maj, 0); 584 (void) ea_attach_to_group(egrp, ip++); 585 (void) ea_set_item(ip, CAT_FMA_MINOR, &min, 0); 586 (void) ea_attach_to_group(egrp, ip++); 587 (void) ea_set_item(ip, CAT_FMA_INODE, 588 &eip->ev_log->log_stat.st_ino, 0); 589 (void) ea_attach_to_group(egrp, ip++); 590 } else { 591 if (ea_set_item(ip, CAT_FMA_UUID, 592 eip->ev_log->log_uuid, 0) == -1) { 593 err = EFMD_LOG_EXACCT; 594 goto exerrcp; 595 } 596 if (fe == NULL) 597 fe = fp = fmd_zalloc(frsize, FMD_SLEEP); 598 *fp++ = ip; 599 (void) ea_attach_to_group(egrp, ip++); 600 } 601 (void) ea_set_item(ip, CAT_FMA_OFFSET, &eip->ev_off, 0); 602 (void) ea_attach_to_group(egrp, ip++); 603 (void) ea_attach_to_group(&grp1, egrp); 604 } 605 (void) ea_attach_to_group(&grp0, &grp1); 606 } 607 608 easize = ea_pack_object(&grp0, NULL, 0); 609 eabuf = fmd_alloc(easize, FMD_SLEEP); 610 (void) ea_pack_object(&grp0, eabuf, easize); 611 612 /* 613 * Before writing the record, check to see if this would cause the free 614 * space in the filesystem to drop below our minfree threshold. If so, 615 * don't bother attempting the write and instead pretend it failed. As 616 * fmd(1M) runs as root, it will be able to access the space "reserved" 617 * for root, and therefore can run the system of out of disk space in a 618 * heavy error load situation, violating the basic design principle of 619 * fmd(1M) that we don't want to make a bad situation even worse. 620 */ 621 (void) pthread_mutex_lock(&lp->log_lock); 622 623 if (lp->log_minfree != 0 && fstatvfs64(lp->log_fd, &stv) == 0 && 624 stv.f_bavail * stv.f_frsize < lp->log_minfree + easize) { 625 626 TRACE((FMD_DBG_LOG, "append %s crosses minfree", lp->log_tag)); 627 err = EFMD_LOG_MINFREE; 628 629 } else if (fmd_log_write(lp, eabuf, easize) == easize) { 630 TRACE((FMD_DBG_LOG, "append %s %p off=0x%llx", 631 lp->log_tag, (void *)ep, (u_longlong_t)lp->log_off)); 632 633 ep->ev_flags &= ~FMD_EVF_VOLATILE; 634 ep->ev_log = lp; 635 ep->ev_off = lp->log_off; 636 ep->ev_len = easize; 637 638 if (lp->log_flags & FMD_LF_REPLAY) { 639 lp->log_pending++; 640 ASSERT(lp->log_pending != 0); 641 } 642 643 lp->log_refs++; 644 ASSERT(lp->log_refs != 0); 645 lp->log_off += easize; 646 } else { 647 err = errno; /* save errno for fmd_error() call below */ 648 649 /* 650 * If we can't write append the record, seek the file back to 651 * the original location and truncate it there in order to make 652 * sure the file is always in a sane state w.r.t. libexacct. 653 */ 654 (void) lseek64(lp->log_fd, lp->log_off, SEEK_SET); 655 (void) ftruncate64(lp->log_fd, lp->log_off); 656 } 657 658 (void) pthread_mutex_unlock(&lp->log_lock); 659 (void) pthread_mutex_unlock(&ep->ev_lock); 660 661 fmd_free(eabuf, easize); 662 663 exerrcp: 664 if (cp != NULL) { 665 if (fe != NULL) { 666 ea_object_t **fp = fe; 667 int i = 0; 668 669 for (; *fp != NULL && i < cip->ci_nitems; i++) 670 (void) ea_free_item(*fp++, EUP_ALLOC); 671 fmd_free(fe, frsize); 672 } 673 674 fmd_free(items, itsize); 675 } 676 677 exerr: 678 fmd_free(nvbuf, nvsize); 679 680 (void) ea_free_item(&i0, EUP_ALLOC); 681 (void) ea_free_item(&i1, EUP_ALLOC); 682 (void) ea_free_item(&i2, EUP_ALLOC); 683 684 /* 685 * Keep track of out-of-space errors using global statistics. As we're 686 * out of disk space, it's unlikely the EFMD_LOG_APPEND will be logged. 687 */ 688 if (err == ENOSPC || err == EFMD_LOG_MINFREE) { 689 fmd_stat_t *sp; 690 691 if (lp == fmd.d_errlog) 692 sp = &fmd.d_stats->ds_err_enospc; 693 else if (lp == fmd.d_fltlog) 694 sp = &fmd.d_stats->ds_flt_enospc; 695 else 696 sp = &fmd.d_stats->ds_oth_enospc; 697 698 (void) pthread_mutex_lock(&fmd.d_stats_lock); 699 sp->fmds_value.ui64++; 700 (void) pthread_mutex_unlock(&fmd.d_stats_lock); 701 } 702 703 if (err != 0) { 704 fmd_error(EFMD_LOG_APPEND, "failed to log_append %s %p: %s\n", 705 lp->log_tag, (void *)ep, fmd_strerror(err)); 706 } 707 } 708 709 /* 710 * Commit an event to the log permanently, indicating that it should not be 711 * replayed on restart. This is done by overwriting the event group's catalog 712 * code with EXD_GROUP_FMA (from EXD_GROUP_RFMA used in fmd_log_append()). We 713 * use pwrite64() to update the existing word directly, using somewhat guilty 714 * knowledge that exacct stores the 32-bit catalog word first for each object. 715 * Since we are overwriting an existing log location using pwrite64() and hold 716 * the event lock, we do not need to hold the log_lock during the i/o. 717 */ 718 void 719 fmd_log_commit(fmd_log_t *lp, fmd_event_t *e) 720 { 721 fmd_event_impl_t *ep = (fmd_event_impl_t *)e; 722 ea_catalog_t c; 723 int err = 0; 724 725 if (!(lp->log_flags & FMD_LF_REPLAY)) 726 return; /* log does not require replay tagging */ 727 728 ASSERT(MUTEX_HELD(&ep->ev_lock)); 729 ASSERT(ep->ev_log == lp && ep->ev_off != 0); 730 731 c = CAT_FMA_GROUP; 732 exacct_order32(&c); 733 734 if (pwrite64(lp->log_fd, &c, sizeof (c), ep->ev_off) == sizeof (c)) { 735 TRACE((FMD_DBG_LOG, "commit %s %p", lp->log_tag, (void *)ep)); 736 ep->ev_flags &= ~FMD_EVF_REPLAY; 737 738 /* 739 * If we have committed the event, check to see if the TOC skip 740 * offset needs to be updated, and decrement the pending count. 741 */ 742 (void) pthread_mutex_lock(&lp->log_lock); 743 744 if (lp->log_skip == ep->ev_off) { 745 lp->log_flags |= FMD_LF_DIRTY; 746 lp->log_skip += ep->ev_len; 747 } 748 749 ASSERT(lp->log_pending != 0); 750 lp->log_pending--; 751 752 (void) pthread_mutex_unlock(&lp->log_lock); 753 (void) pthread_cond_broadcast(&lp->log_cv); 754 755 } else { 756 fmd_error(EFMD_LOG_COMMIT, "failed to log_commit %s %p: %s\n", 757 lp->log_tag, (void *)ep, fmd_strerror(err)); 758 } 759 } 760 761 /* 762 * If we need to destroy an event and it wasn't able to be committed, we permit 763 * the owner to decommit from ever trying again. This operation decrements the 764 * pending count on the log and broadcasts to anyone waiting on log_cv. 765 */ 766 void 767 fmd_log_decommit(fmd_log_t *lp, fmd_event_t *e) 768 { 769 fmd_event_impl_t *ep = (fmd_event_impl_t *)e; 770 771 if (!(lp->log_flags & FMD_LF_REPLAY)) 772 return; /* log does not require replay tagging */ 773 774 ASSERT(MUTEX_HELD(&ep->ev_lock)); 775 ASSERT(ep->ev_log == lp); 776 777 (void) pthread_mutex_lock(&lp->log_lock); 778 779 TRACE((FMD_DBG_LOG, "decommit %s %p", lp->log_tag, (void *)ep)); 780 ep->ev_flags &= ~FMD_EVF_REPLAY; 781 782 ASSERT(lp->log_pending != 0); 783 lp->log_pending--; 784 785 (void) pthread_mutex_unlock(&lp->log_lock); 786 (void) pthread_cond_broadcast(&lp->log_cv); 787 } 788 789 static fmd_event_t * 790 fmd_log_unpack(fmd_log_t *lp, ea_object_t *grp, off64_t off) 791 { 792 fmd_timeval_t ftv = { -1ULL, -1ULL }; 793 nvlist_t *nvl = NULL; 794 795 ea_object_t *obj; 796 char *class; 797 int err; 798 799 for (obj = grp->eo_group.eg_objs; obj != NULL; obj = obj->eo_next) { 800 switch (obj->eo_catalog) { 801 case CAT_FMA_NVLIST: 802 if ((err = nvlist_xunpack(obj->eo_item.ei_raw, 803 obj->eo_item.ei_size, &nvl, &fmd.d_nva)) != 0) { 804 fmd_error(EFMD_LOG_UNPACK, "failed to unpack " 805 "log nvpair: %s\n", fmd_strerror(err)); 806 return (NULL); 807 } 808 break; 809 810 case CAT_FMA_TODSEC: 811 ftv.ftv_sec = obj->eo_item.ei_uint64; 812 break; 813 814 case CAT_FMA_TODNSEC: 815 ftv.ftv_nsec = obj->eo_item.ei_uint64; 816 break; 817 } 818 } 819 820 if (nvl == NULL || ftv.ftv_sec == -1ULL || ftv.ftv_nsec == -1ULL) { 821 fmd_error(EFMD_LOG_UNPACK, "failed to unpack log event: " 822 "required object(s) missing from record group\n"); 823 nvlist_free(nvl); 824 return (NULL); 825 } 826 827 if (nvlist_lookup_string(nvl, FM_CLASS, &class) != 0) { 828 fmd_error(EFMD_LOG_UNPACK, "failed to unpack log event: " 829 "record is missing required '%s' nvpair\n", FM_CLASS); 830 nvlist_free(nvl); 831 return (NULL); 832 } 833 834 return (fmd_event_recreate(FMD_EVT_PROTOCOL, 835 &ftv, nvl, class, lp, off, ea_pack_object(grp, NULL, 0))); 836 } 837 838 /* 839 * Replay event(s) from the specified log by invoking the specified callback 840 * function 'func' for each event. If the log has the FMD_LF_REPLAY flag set, 841 * we replay all events after log_skip that have the FMA_RGROUP group tag. 842 * This mode is used for the error telemetry log. If the log does not have 843 * this flag set (used for ASRU logs), only the most recent event is replayed. 844 */ 845 void 846 fmd_log_replay(fmd_log_t *lp, fmd_log_f *func, void *data) 847 { 848 ea_object_t obj, *grp; 849 ea_object_type_t type; 850 ea_catalog_t c; 851 fmd_event_t *ep; 852 off64_t off, skp; 853 uint_t n = 0; 854 855 (void) pthread_mutex_lock(&lp->log_lock); 856 857 if (lp->log_stat.st_size == 0 && (lp->log_flags & FMD_LF_REPLAY)) { 858 (void) pthread_mutex_unlock(&lp->log_lock); 859 return; /* we just created this log: never replay events */ 860 } 861 862 while (lp->log_flags & FMD_LF_BUSY) 863 (void) pthread_cond_wait(&lp->log_cv, &lp->log_lock); 864 865 if (lp->log_off == lp->log_beg) { 866 (void) pthread_mutex_unlock(&lp->log_lock); 867 return; /* no records appended yet */ 868 } 869 870 lp->log_flags |= FMD_LF_BUSY; 871 skp = lp->log_skip; 872 ea_clear(&lp->log_ea); /* resync exacct file */ 873 874 /* 875 * If FMD_LF_REPLAY is set, begin our replay at either log_skip (if it 876 * is non-zero) or at log_beg. Otherwise replay from the end (log_off) 877 */ 878 if (lp->log_flags & FMD_LF_REPLAY) { 879 off = MAX(lp->log_beg, lp->log_skip); 880 c = CAT_FMA_RGROUP; 881 } else { 882 off = lp->log_off; 883 c = CAT_FMA_GROUP; 884 } 885 886 if (lseek64(lp->log_fd, off, SEEK_SET) != off) { 887 fmd_panic("failed to seek %s to 0x%llx\n", 888 lp->log_name, (u_longlong_t)off); 889 } 890 891 /* 892 * If FMD_LF_REPLAY is not set, back up to the start of the previous 893 * object and make sure this object is an EO_GROUP; otherwise return. 894 */ 895 if (!(lp->log_flags & FMD_LF_REPLAY) && 896 (type = ea_previous_object(&lp->log_ea, &obj)) != EO_GROUP) { 897 fmd_error(EFMD_LOG_REPLAY, "last log object is of unexpected " 898 "type %d (log may be truncated or corrupt)\n", type); 899 goto out; 900 } 901 902 while ((grp = ea_get_object_tree(&lp->log_ea, 1)) != NULL) { 903 if (!(lp->log_flags & FMD_LF_REPLAY)) 904 off -= ea_pack_object(grp, NULL, 0); 905 else if (n == 0 && grp->eo_catalog == CAT_FMA_GROUP) 906 skp = off; /* update skip */ 907 908 /* 909 * We temporarily drop log_lock around the call to unpack the 910 * event, hold it, and perform the callback, because these 911 * operations may try to acquire log_lock to bump log_refs. 912 * We cannot lose control because the FMD_LF_BUSY flag is set. 913 */ 914 (void) pthread_mutex_unlock(&lp->log_lock); 915 916 if (grp->eo_catalog == c && 917 (ep = fmd_log_unpack(lp, grp, off)) != NULL) { 918 919 TRACE((FMD_DBG_LOG, "replay %s %p off %llx", 920 lp->log_tag, (void *)ep, (u_longlong_t)off)); 921 922 fmd_event_hold(ep); 923 func(lp, ep, data); 924 fmd_event_rele(ep); 925 n++; 926 } 927 928 (void) pthread_mutex_lock(&lp->log_lock); 929 off += ea_pack_object(grp, NULL, 0); 930 ea_free_object(grp, EUP_ALLOC); 931 } 932 933 if (ea_error() != EXR_EOF) { 934 fmd_error(EFMD_LOG_REPLAY, "failed to replay %s event at " 935 "offset 0x%llx: %s\n", lp->log_name, (u_longlong_t)off, 936 fmd_ea_strerror(ea_error())); 937 } 938 939 if (n == 0) 940 skp = off; /* if no replays, move skip to where we ended up */ 941 942 out: 943 if (lseek64(lp->log_fd, lp->log_off, SEEK_SET) != lp->log_off) { 944 fmd_panic("failed to seek %s to 0x%llx\n", 945 lp->log_name, (u_longlong_t)lp->log_off); 946 } 947 948 if (skp != lp->log_skip) { 949 lp->log_flags |= FMD_LF_DIRTY; 950 lp->log_skip = skp; 951 } 952 953 lp->log_flags &= ~FMD_LF_BUSY; 954 (void) pthread_mutex_unlock(&lp->log_lock); 955 (void) pthread_cond_broadcast(&lp->log_cv); 956 } 957 958 void 959 fmd_log_update(fmd_log_t *lp) 960 { 961 ea_object_t toc, item; 962 off64_t skip = 0; 963 size_t size; 964 void *buf; 965 966 (void) pthread_mutex_lock(&lp->log_lock); 967 968 if (lp->log_flags & FMD_LF_DIRTY) { 969 lp->log_flags &= ~FMD_LF_DIRTY; 970 skip = lp->log_skip; 971 } 972 973 (void) pthread_mutex_unlock(&lp->log_lock); 974 975 /* 976 * If the skip needs to be updated, construct a TOC record group 977 * containing the skip offset and overwrite the TOC in-place. 978 */ 979 if (skip != 0 && ea_set_group(&toc, CAT_FMA_GROUP) == 0 && 980 ea_set_item(&item, CAT_FMA_OFFSET, &skip, 0) == 0) { 981 982 (void) ea_attach_to_group(&toc, &item); 983 size = ea_pack_object(&toc, NULL, 0); 984 buf = fmd_alloc(size, FMD_SLEEP); 985 986 (void) ea_pack_object(&toc, buf, size); 987 ASSERT(lp->log_toc + size == lp->log_beg); 988 989 if (pwrite64(lp->log_fd, buf, size, lp->log_toc) == size) { 990 TRACE((FMD_DBG_LOG, "updated skip to %llx", skip)); 991 } else { 992 fmd_error(EFMD_LOG_UPDATE, 993 "failed to log_update %s", lp->log_tag); 994 } 995 996 fmd_free(buf, size); 997 (void) ea_free_item(&item, EUP_ALLOC); 998 } 999 } 1000 1001 /* 1002 * Rotate the specified log by renaming its underlying file to a staging file 1003 * that can be handed off to logadm(1M) or an administrator script. If the 1004 * rename succeeds, open a new log file using the old path and return it. 1005 * Note that we are relying our caller to use some higher-level mechanism to 1006 * ensure that fmd_log_rotate() cannot be called while other threads are 1007 * attempting fmd_log_append() using the same log (fmd's d_log_lock is used 1008 * for the global errlog and fltlog). 1009 */ 1010 fmd_log_t * 1011 fmd_log_rotate(fmd_log_t *lp) 1012 { 1013 char npath[PATH_MAX]; 1014 fmd_log_t *nlp; 1015 1016 (void) snprintf(npath, sizeof (npath), "%s.0-", lp->log_name); 1017 (void) pthread_mutex_lock(&lp->log_lock); 1018 1019 /* 1020 * Check for any pending commits to drain before proceeding. We can't 1021 * rotate the log out if commits are pending because if we die after 1022 * the log is moved aside, we won't be able to replay them on restart. 1023 */ 1024 if (lp->log_pending != 0) { 1025 (void) pthread_mutex_unlock(&lp->log_lock); 1026 (void) fmd_set_errno(EFMD_LOG_ROTBUSY); 1027 return (NULL); 1028 } 1029 1030 if (rename(lp->log_name, npath) != 0) { 1031 (void) pthread_mutex_unlock(&lp->log_lock); 1032 fmd_error(EFMD_LOG_ROTATE, "failed to rename %s", lp->log_name); 1033 (void) fmd_set_errno(EFMD_LOG_ROTATE); 1034 return (NULL); 1035 } 1036 1037 if ((nlp = fmd_log_open("", lp->log_name, lp->log_tag)) == NULL) { 1038 (void) rename(npath, lp->log_name); 1039 (void) pthread_mutex_unlock(&lp->log_lock); 1040 fmd_error(EFMD_LOG_ROTATE, "failed to reopen %s", lp->log_name); 1041 (void) fmd_set_errno(EFMD_LOG_ROTATE); 1042 return (NULL); 1043 } 1044 1045 /* 1046 * If we've rotated the log, no pending events exist so we don't have 1047 * any more commits coming, and our caller should have arranged for 1048 * no more calls to append. As such, we can close log_fd for good. 1049 */ 1050 if (lp->log_flags & FMD_LF_EAOPEN) { 1051 (void) ea_close(&lp->log_ea); 1052 lp->log_flags &= ~FMD_LF_EAOPEN; 1053 } 1054 1055 (void) close(lp->log_fd); 1056 lp->log_fd = -1; 1057 1058 (void) pthread_mutex_unlock(&lp->log_lock); 1059 return (nlp); 1060 } 1061