/* * CDDL HEADER START * * The contents of this file are subject to the terms of the * Common Development and Distribution License (the "License"). * You may not use this file except in compliance with the License. * * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE * or http://www.opensolaris.org/os/licensing. * See the License for the specific language governing permissions * and limitations under the License. * * When distributing Covered Code, include this CDDL HEADER in each * file and include the License file at usr/src/OPENSOLARIS.LICENSE. * If applicable, add the following below this CDDL HEADER, with the * fields enclosed by brackets "[]" replaced with your own identifying * information: Portions Copyright [yyyy] [name of copyright owner] * * CDDL HEADER END */ /* * Copyright 2007 Sun Microsystems, Inc. All rights reserved. * Use is subject to license terms. */ #pragma ident "%Z%%M% %I% %E% SMI" /* * FMD Dynamic Reconfiguration (DR) Event Handling * * Fault manager scheme plug-ins must track characteristics of individual * pieces of hardware. As these components can be added or removed by a DR * operation, we need to provide a means by which plug-ins can determine when * they need to re-examine the current configuration. We provide a simple * mechanism whereby this task can be implemented using lazy evaluation: a * simple 64-bit generation counter is maintained and incremented on *any* DR. * Schemes can store the generation number in scheme-specific data structures, * and then revalidate their contents if the current generation number has * changed since the resource information was cached. This method saves time, * avoids the complexity of direct participation in DR, avoids the need for * resource-specific processing of DR events, and is relatively easy to port * to other systems that support dynamic reconfiguration. * * The dr generation is only incremented in response to hardware changes. Since * ASRUs can be in any scheme, including the device scheme, we must also be * aware of software configuration changes which may affect the resource cache. * In addition, we take a snapshot of the topology whenever a reconfiguration * event occurs and notify any modules of the change. */ #include #include #include #include #include #include #include #include #undef MUTEX_HELD #undef RW_READ_HELD #undef RW_WRITE_HELD #include #include #include #include #include #include #include #include static void fmd_dr_event(sysevent_t *sep) { uint64_t gen; fmd_event_t *e; const char *class = sysevent_get_class_name(sep); hrtime_t evtime; fmd_topo_t *ftp, *prev; boolean_t update_topo = B_FALSE; /* * The dr generation is only changed in response to DR events. */ if (strcmp(class, EC_DR) == 0) { update_topo = B_TRUE; (void) pthread_mutex_lock(&fmd.d_stats_lock); gen = fmd.d_stats->ds_dr_gen.fmds_value.ui64++; (void) pthread_mutex_unlock(&fmd.d_stats_lock); TRACE((FMD_DBG_XPRT, "dr event %p, gen=%llu", (void *)sep, gen)); } /* * Take a topo snapshot and notify modules of the change. Picking an * accurate time here is difficult. On one hand, we have the timestamp * of the underlying sysevent, indicating when the reconfiguration event * occurred. On the other hand, we are taking the topo snapshot * asynchronously, and hence the timestamp of the snapshot is the * current time. Pretending this topo snapshot was valid at the time * the sysevent was posted seems wrong, so we instead opt for the * current time as an upper bound on the snapshot validity. * * Along these lines, we keep track of the last time we dispatched a * topo snapshot. If the sysevent occurred before the last topo * snapshot, then don't bother dispatching another topo change event. * We've already indicated (to the best of our ability) the change in * topology. This prevents endless topo snapshots in response to a * flurry of sysevents. */ sysevent_get_time(sep, &evtime); prev = fmd_topo_hold(); if (evtime <= prev->ft_time && fmd.d_clockops == &fmd_timeops_native) { fmd_topo_rele(prev); return; } fmd_topo_rele(prev); if (update_topo) fmd_topo_update(); ftp = fmd_topo_hold(); e = fmd_event_create(FMD_EVT_TOPO, ftp->ft_time, NULL, ftp); fmd_modhash_dispatch(fmd.d_mod_hash, e); } void fmd_dr_init(void) { const char *dr_subclasses[] = { ESC_DR_AP_STATE_CHANGE }; const char *zfs_subclasses[] = { ESC_ZFS_VDEV_CLEAR, ESC_ZFS_VDEV_REMOVE, ESC_ZFS_POOL_DESTROY }; const char *dev_subclasses[] = { EC_SUB_ALL }; if (geteuid() != 0) return; /* legacy sysevent mechanism is still root-only */ if ((fmd.d_dr_hdl = sysevent_bind_handle(fmd_dr_event)) == NULL) fmd_error(EFMD_EXIT, "failed to bind handle for DR sysevent"); if (sysevent_subscribe_event(fmd.d_dr_hdl, EC_DR, dr_subclasses, sizeof (dr_subclasses) / sizeof (char *)) == -1) fmd_error(EFMD_EXIT, "failed to subscribe to DR sysevents"); if (sysevent_subscribe_event(fmd.d_dr_hdl, EC_DEVFS, dev_subclasses, sizeof (dev_subclasses) / sizeof (char *)) == -1) fmd_error(EFMD_EXIT, "failed to subscribe to devfs sysevents"); if (sysevent_subscribe_event(fmd.d_dr_hdl, EC_ZFS, zfs_subclasses, sizeof (zfs_subclasses) / sizeof (char *)) == -1) fmd_error(EFMD_EXIT, "failed to subscribe to ZFS sysevents"); } void fmd_dr_fini(void) { if (fmd.d_dr_hdl != NULL) { sysevent_unsubscribe_event(fmd.d_dr_hdl, EC_DR); sysevent_unbind_handle(fmd.d_dr_hdl); } }