/* * CDDL HEADER START * * The contents of this file are subject to the terms of the * Common Development and Distribution License (the "License"). * You may not use this file except in compliance with the License. * * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE * or http://www.opensolaris.org/os/licensing. * See the License for the specific language governing permissions * and limitations under the License. * * When distributing Covered Code, include this CDDL HEADER in each * file and include the License file at usr/src/OPENSOLARIS.LICENSE. * If applicable, add the following below this CDDL HEADER, with the * fields enclosed by brackets "[]" replaced with your own identifying * information: Portions Copyright [yyyy] [name of copyright owner] * * CDDL HEADER END */ /* * Copyright 2006 Sun Microsystems, Inc. All rights reserved. * Use is subject to license terms. */ #pragma ident "%Z%%M% %I% %E% SMI" /* * PX Interrupt Block implementation */ #include #include #include #include /* panicstr */ #include #include #include /* intr_dist_add */ #include #include #include #include "px_obj.h" /*LINTLIBRARY*/ static void px_ib_intr_redist(void *arg, int32_t weight_max, int32_t weight); static void px_ib_cpu_ticks_to_ih_nsec(px_ib_t *ib_p, px_ih_t *ih_p, uint32_t cpu_id); static uint_t px_ib_intr_reset(void *arg); static void px_fill_in_intr_devs(pcitool_intr_dev_t *dev, char *driver_name, char *path_name, int instance); extern uint64_t xc_tick_jump_limit; int px_ib_attach(px_t *px_p) { dev_info_t *dip = px_p->px_dip; px_ib_t *ib_p; sysino_t sysino; px_fault_t *fault_p = &px_p->px_fault; DBG(DBG_IB, dip, "px_ib_attach\n"); if (px_lib_intr_devino_to_sysino(px_p->px_dip, px_p->px_inos[PX_INTR_PEC], &sysino) != DDI_SUCCESS) return (DDI_FAILURE); /* * Allocate interrupt block state structure and link it to * the px state structure. */ ib_p = kmem_zalloc(sizeof (px_ib_t), KM_SLEEP); px_p->px_ib_p = ib_p; ib_p->ib_px_p = px_p; ib_p->ib_ino_lst = (px_ib_ino_info_t *)NULL; mutex_init(&ib_p->ib_intr_lock, NULL, MUTEX_DRIVER, NULL); mutex_init(&ib_p->ib_ino_lst_mutex, NULL, MUTEX_DRIVER, NULL); bus_func_register(BF_TYPE_RESINTR, px_ib_intr_reset, ib_p); intr_dist_add_weighted(px_ib_intr_redist, ib_p); /* * Initialize PEC fault data structure */ fault_p->px_fh_dip = dip; fault_p->px_fh_sysino = sysino; fault_p->px_err_func = px_err_dmc_pec_intr; fault_p->px_intr_ino = px_p->px_inos[PX_INTR_PEC]; return (DDI_SUCCESS); } void px_ib_detach(px_t *px_p) { px_ib_t *ib_p = px_p->px_ib_p; dev_info_t *dip = px_p->px_dip; DBG(DBG_IB, dip, "px_ib_detach\n"); bus_func_unregister(BF_TYPE_RESINTR, px_ib_intr_reset, ib_p); intr_dist_rem_weighted(px_ib_intr_redist, ib_p); mutex_destroy(&ib_p->ib_ino_lst_mutex); mutex_destroy(&ib_p->ib_intr_lock); px_ib_free_ino_all(ib_p); px_p->px_ib_p = NULL; kmem_free(ib_p, sizeof (px_ib_t)); } void px_ib_intr_enable(px_t *px_p, cpuid_t cpu_id, devino_t ino) { px_ib_t *ib_p = px_p->px_ib_p; sysino_t sysino; /* * Determine the cpu for the interrupt */ mutex_enter(&ib_p->ib_intr_lock); DBG(DBG_IB, px_p->px_dip, "px_ib_intr_enable: ino=%x cpu_id=%x\n", ino, cpu_id); if (px_lib_intr_devino_to_sysino(px_p->px_dip, ino, &sysino) != DDI_SUCCESS) { DBG(DBG_IB, px_p->px_dip, "px_ib_intr_enable: px_intr_devino_to_sysino() failed\n"); mutex_exit(&ib_p->ib_intr_lock); return; } PX_INTR_ENABLE(px_p->px_dip, sysino, cpu_id); px_lib_intr_setstate(px_p->px_dip, sysino, INTR_IDLE_STATE); mutex_exit(&ib_p->ib_intr_lock); } /*ARGSUSED*/ void px_ib_intr_disable(px_ib_t *ib_p, devino_t ino, int wait) { sysino_t sysino; mutex_enter(&ib_p->ib_intr_lock); DBG(DBG_IB, ib_p->ib_px_p->px_dip, "px_ib_intr_disable: ino=%x\n", ino); /* Disable the interrupt */ if (px_lib_intr_devino_to_sysino(ib_p->ib_px_p->px_dip, ino, &sysino) != DDI_SUCCESS) { DBG(DBG_IB, ib_p->ib_px_p->px_dip, "px_ib_intr_disable: px_intr_devino_to_sysino() failed\n"); mutex_exit(&ib_p->ib_intr_lock); return; } PX_INTR_DISABLE(ib_p->ib_px_p->px_dip, sysino); mutex_exit(&ib_p->ib_intr_lock); } void px_ib_intr_dist_en(dev_info_t *dip, cpuid_t cpu_id, devino_t ino, boolean_t wait_flag) { uint32_t old_cpu_id; sysino_t sysino; intr_valid_state_t enabled = 0; hrtime_t start_time, prev, curr, interval, jump; hrtime_t intr_timeout; intr_state_t intr_state; int e = DDI_SUCCESS; DBG(DBG_IB, dip, "px_ib_intr_dist_en: ino=0x%x\n", ino); if (px_lib_intr_devino_to_sysino(dip, ino, &sysino) != DDI_SUCCESS) { DBG(DBG_IB, dip, "px_ib_intr_dist_en: " "px_intr_devino_to_sysino() failed, ino 0x%x\n", ino); return; } /* Skip enabling disabled interrupts */ if (px_lib_intr_getvalid(dip, sysino, &enabled) != DDI_SUCCESS) { DBG(DBG_IB, dip, "px_ib_intr_dist_en: px_intr_getvalid() " "failed, sysino 0x%x\n", sysino); return; } if (!enabled) return; /* Done if redistributed onto the same cpuid */ if (px_lib_intr_gettarget(dip, sysino, &old_cpu_id) != DDI_SUCCESS) { DBG(DBG_IB, dip, "px_ib_intr_dist_en: " "px_intr_gettarget() failed\n"); return; } if (cpu_id == old_cpu_id) return; if (!wait_flag) goto done; /* Busy wait on pending interrupts */ PX_INTR_DISABLE(dip, sysino); intr_timeout = px_intrpend_timeout; jump = TICK_TO_NSEC(xc_tick_jump_limit); for (curr = start_time = gethrtime(); !panicstr && ((e = px_lib_intr_getstate(dip, sysino, &intr_state)) == DDI_SUCCESS) && (intr_state == INTR_DELIVERED_STATE); /* */) { /* * If we have a really large jump in hrtime, it is most * probably because we entered the debugger (or OBP, * in general). So, we adjust the timeout accordingly * to prevent declaring an interrupt timeout. The * master-interrupt mechanism in OBP should deliver * the interrupts properly. */ prev = curr; curr = gethrtime(); interval = curr - prev; if (interval > jump) intr_timeout += interval; if (curr - start_time > intr_timeout) { cmn_err(CE_WARN, "%s%d: px_ib_intr_dist_en: sysino 0x%lx(ino 0x%x) " "from cpu id 0x%x to 0x%x timeout", ddi_driver_name(dip), ddi_get_instance(dip), sysino, ino, old_cpu_id, cpu_id); e = DDI_FAILURE; break; } } if (e != DDI_SUCCESS) DBG(DBG_IB, dip, "px_ib_intr_dist_en: failed, " "ino 0x%x sysino 0x%x\n", ino, sysino); done: PX_INTR_ENABLE(dip, sysino, cpu_id); } static void px_ib_cpu_ticks_to_ih_nsec(px_ib_t *ib_p, px_ih_t *ih_p, uint32_t cpu_id) { extern kmutex_t pxintr_ks_template_lock; hrtime_t ticks; /* * Because we are updating two fields in ih_t we must lock * pxintr_ks_template_lock to prevent someone from reading the * kstats after we set ih_ticks to 0 and before we increment * ih_nsec to compensate. * * We must also protect against the interrupt arriving and incrementing * ih_ticks between the time we read it and when we reset it to 0. * To do this we use atomic_swap. */ ASSERT(MUTEX_HELD(&ib_p->ib_ino_lst_mutex)); mutex_enter(&pxintr_ks_template_lock); ticks = atomic_swap_64(&ih_p->ih_ticks, 0); ih_p->ih_nsec += (uint64_t)tick2ns(ticks, cpu_id); mutex_exit(&pxintr_ks_template_lock); } /* * Redistribute interrupts of the specified weight. The first call has a weight * of weight_max, which can be used to trigger initialization for * redistribution. The inos with weight [weight_max, inf.) should be processed * on the "weight == weight_max" call. This first call is followed by calls * of decreasing weights, inos of that weight should be processed. The final * call specifies a weight of zero, this can be used to trigger processing of * stragglers. */ static void px_ib_intr_redist(void *arg, int32_t weight_max, int32_t weight) { px_ib_t *ib_p = (px_ib_t *)arg; px_t *px_p = ib_p->ib_px_p; dev_info_t *dip = px_p->px_dip; px_ib_ino_info_t *ino_p; px_ih_t *ih_lst; int32_t dweight = 0; int i; /* Redistribute internal interrupts */ if (weight == 0) { mutex_enter(&ib_p->ib_intr_lock); px_ib_intr_dist_en(dip, intr_dist_cpuid(), px_p->px_inos[PX_INTR_PEC], B_FALSE); mutex_exit(&ib_p->ib_intr_lock); px_cb_intr_redist(px_p); } /* Redistribute device interrupts */ mutex_enter(&ib_p->ib_ino_lst_mutex); for (ino_p = ib_p->ib_ino_lst; ino_p; ino_p = ino_p->ino_next) { uint32_t orig_cpuid; /* * Recomputes the sum of interrupt weights of devices that * share the same ino upon first call marked by * (weight == weight_max). */ if (weight == weight_max) { ino_p->ino_intr_weight = 0; for (i = 0, ih_lst = ino_p->ino_ih_head; i < ino_p->ino_ih_size; i++, ih_lst = ih_lst->ih_next) { dweight = i_ddi_get_intr_weight(ih_lst->ih_dip); if (dweight > 0) ino_p->ino_intr_weight += dweight; } } /* * As part of redistributing weighted interrupts over cpus, * nexus redistributes device interrupts and updates * cpu weight. The purpose is for the most light weighted * cpu to take the next interrupt and gain weight, therefore * attention demanding device gains more cpu attention by * making itself heavy. */ if ((weight == ino_p->ino_intr_weight) || ((weight >= weight_max) && (ino_p->ino_intr_weight >= weight_max))) { orig_cpuid = ino_p->ino_cpuid; if (cpu[orig_cpuid] == NULL) orig_cpuid = CPU->cpu_id; /* select cpuid to target and mark ino established */ ino_p->ino_cpuid = intr_dist_cpuid(); /* Add device weight to targeted cpu. */ for (i = 0, ih_lst = ino_p->ino_ih_head; i < ino_p->ino_ih_size; i++, ih_lst = ih_lst->ih_next) { dweight = i_ddi_get_intr_weight(ih_lst->ih_dip); intr_dist_cpuid_add_device_weight( ino_p->ino_cpuid, ih_lst->ih_dip, dweight); /* * Different cpus may have different clock * speeds. to account for this, whenever an * interrupt is moved to a new CPU, we * convert the accumulated ticks into nsec, * based upon the clock rate of the prior * CPU. * * It is possible that the prior CPU no longer * exists. In this case, fall back to using * this CPU's clock rate. * * Note that the value in ih_ticks has already * been corrected for any power savings mode * which might have been in effect. */ px_ib_cpu_ticks_to_ih_nsec(ib_p, ih_lst, orig_cpuid); } /* enable interrupt on new targeted cpu */ px_ib_intr_dist_en(dip, ino_p->ino_cpuid, ino_p->ino_ino, B_TRUE); } } mutex_exit(&ib_p->ib_ino_lst_mutex); } /* * Reset interrupts to IDLE. This function is called during * panic handling after redistributing interrupts; it's needed to * support dumping to network devices after 'sync' from OBP. * * N.B. This routine runs in a context where all other threads * are permanently suspended. */ static uint_t px_ib_intr_reset(void *arg) { px_ib_t *ib_p = (px_ib_t *)arg; DBG(DBG_IB, ib_p->ib_px_p->px_dip, "px_ib_intr_reset\n"); if (px_lib_intr_reset(ib_p->ib_px_p->px_dip) != DDI_SUCCESS) return (BF_FATAL); return (BF_NONE); } /* * Locate ino_info structure on ib_p->ib_ino_lst according to ino# * returns NULL if not found. */ px_ib_ino_info_t * px_ib_locate_ino(px_ib_t *ib_p, devino_t ino_num) { px_ib_ino_info_t *ino_p = ib_p->ib_ino_lst; ASSERT(MUTEX_HELD(&ib_p->ib_ino_lst_mutex)); for (; ino_p && ino_p->ino_ino != ino_num; ino_p = ino_p->ino_next); return (ino_p); } px_ib_ino_info_t * px_ib_new_ino(px_ib_t *ib_p, devino_t ino_num, px_ih_t *ih_p) { px_ib_ino_info_t *ino_p = kmem_alloc(sizeof (px_ib_ino_info_t), KM_SLEEP); sysino_t sysino; ino_p->ino_ino = ino_num; ino_p->ino_ib_p = ib_p; ino_p->ino_unclaimed = 0; if (px_lib_intr_devino_to_sysino(ib_p->ib_px_p->px_dip, ino_p->ino_ino, &sysino) != DDI_SUCCESS) return (NULL); ino_p->ino_sysino = sysino; /* * Cannot disable interrupt since we might share slot */ ih_p->ih_next = ih_p; ino_p->ino_ih_head = ih_p; ino_p->ino_ih_tail = ih_p; ino_p->ino_ih_start = ih_p; ino_p->ino_ih_size = 1; ino_p->ino_next = ib_p->ib_ino_lst; ib_p->ib_ino_lst = ino_p; return (ino_p); } /* * The ino_p is retrieved by previous call to px_ib_locate_ino(). */ void px_ib_delete_ino(px_ib_t *ib_p, px_ib_ino_info_t *ino_p) { px_ib_ino_info_t *list = ib_p->ib_ino_lst; ASSERT(MUTEX_HELD(&ib_p->ib_ino_lst_mutex)); if (list == ino_p) ib_p->ib_ino_lst = list->ino_next; else { for (; list->ino_next != ino_p; list = list->ino_next); list->ino_next = ino_p->ino_next; } } /* * Free all ino when we are detaching. */ void px_ib_free_ino_all(px_ib_t *ib_p) { px_ib_ino_info_t *tmp = ib_p->ib_ino_lst; px_ib_ino_info_t *next = NULL; while (tmp) { next = tmp->ino_next; kmem_free(tmp, sizeof (px_ib_ino_info_t)); tmp = next; } } int px_ib_ino_add_intr(px_t *px_p, px_ib_ino_info_t *ino_p, px_ih_t *ih_p) { px_ib_t *ib_p = ino_p->ino_ib_p; devino_t ino = ino_p->ino_ino; sysino_t sysino = ino_p->ino_sysino; dev_info_t *dip = px_p->px_dip; cpuid_t curr_cpu; hrtime_t start_time; intr_state_t intr_state; int ret = DDI_SUCCESS; ASSERT(MUTEX_HELD(&ib_p->ib_ino_lst_mutex)); ASSERT(ib_p == px_p->px_ib_p); DBG(DBG_IB, dip, "px_ib_ino_add_intr ino=%x\n", ino_p->ino_ino); /* Disable the interrupt */ if ((ret = px_lib_intr_gettarget(dip, sysino, &curr_cpu)) != DDI_SUCCESS) { DBG(DBG_IB, dip, "px_ib_ino_add_intr px_intr_gettarget() failed\n"); return (ret); } PX_INTR_DISABLE(dip, sysino); /* Busy wait on pending interrupt */ for (start_time = gethrtime(); !panicstr && ((ret = px_lib_intr_getstate(dip, sysino, &intr_state)) == DDI_SUCCESS) && (intr_state == INTR_DELIVERED_STATE); /* */) { if (gethrtime() - start_time > px_intrpend_timeout) { cmn_err(CE_WARN, "%s%d: px_ib_ino_add_intr: pending " "sysino 0x%lx(ino 0x%x) timeout", ddi_driver_name(dip), ddi_get_instance(dip), sysino, ino); ret = DDI_FAILURE; break; } } if (ret != DDI_SUCCESS) { DBG(DBG_IB, dip, "px_ib_ino_add_intr: failed, " "ino 0x%x sysino 0x%x\n", ino, sysino); return (ret); } /* Link up px_ispec_t portion of the ppd */ ih_p->ih_next = ino_p->ino_ih_head; ino_p->ino_ih_tail->ih_next = ih_p; ino_p->ino_ih_tail = ih_p; ino_p->ino_ih_start = ino_p->ino_ih_head; ino_p->ino_ih_size++; /* * If the interrupt was previously blocked (left in pending state) * because of jabber we need to clear the pending state in case the * jabber has gone away. */ if (ino_p->ino_unclaimed > px_unclaimed_intr_max) { cmn_err(CE_WARN, "%s%d: px_ib_ino_add_intr: ino 0x%x has been unblocked", ddi_driver_name(dip), ddi_get_instance(dip), ino); ino_p->ino_unclaimed = 0; if ((ret = px_lib_intr_setstate(dip, sysino, INTR_IDLE_STATE)) != DDI_SUCCESS) { DBG(DBG_IB, px_p->px_dip, "px_ib_ino_add_intr px_intr_setstate failed\n"); return (ret); } } /* Re-enable interrupt */ PX_INTR_ENABLE(dip, sysino, curr_cpu); return (ret); } /* * Removes px_ispec_t from the ino's link list. * uses hardware mutex to lock out interrupt threads. * Side effects: interrupt belongs to that ino is turned off on return. * if we are sharing PX slot with other inos, the caller needs * to turn it back on. */ int px_ib_ino_rem_intr(px_t *px_p, px_ib_ino_info_t *ino_p, px_ih_t *ih_p) { devino_t ino = ino_p->ino_ino; sysino_t sysino = ino_p->ino_sysino; dev_info_t *dip = px_p->px_dip; px_ih_t *ih_lst = ino_p->ino_ih_head; hrtime_t start_time; intr_state_t intr_state; int i, ret = DDI_SUCCESS; ASSERT(MUTEX_HELD(&ino_p->ino_ib_p->ib_ino_lst_mutex)); DBG(DBG_IB, px_p->px_dip, "px_ib_ino_rem_intr ino=%x\n", ino_p->ino_ino); /* Disable the interrupt */ PX_INTR_DISABLE(px_p->px_dip, sysino); if (ino_p->ino_ih_size == 1) { if (ih_lst != ih_p) goto not_found; /* No need to set head/tail as ino_p will be freed */ goto reset; } /* Busy wait on pending interrupt */ for (start_time = gethrtime(); !panicstr && ((ret = px_lib_intr_getstate(dip, sysino, &intr_state)) == DDI_SUCCESS) && (intr_state == INTR_DELIVERED_STATE); /* */) { if (gethrtime() - start_time > px_intrpend_timeout) { cmn_err(CE_WARN, "%s%d: px_ib_ino_rem_intr: pending " "sysino 0x%lx(ino 0x%x) timeout", ddi_driver_name(dip), ddi_get_instance(dip), sysino, ino); ret = DDI_FAILURE; break; } } if (ret != DDI_SUCCESS) { DBG(DBG_IB, dip, "px_ib_ino_rem_intr: failed, " "ino 0x%x sysino 0x%x\n", ino, sysino); return (ret); } /* * If the interrupt was previously blocked (left in pending state) * because of jabber we need to clear the pending state in case the * jabber has gone away. */ if (ino_p->ino_unclaimed > px_unclaimed_intr_max) { cmn_err(CE_WARN, "%s%d: px_ib_ino_rem_intr: " "ino 0x%x has been unblocked", ddi_driver_name(dip), ddi_get_instance(dip), ino); ino_p->ino_unclaimed = 0; if ((ret = px_lib_intr_setstate(dip, sysino, INTR_IDLE_STATE)) != DDI_SUCCESS) { DBG(DBG_IB, px_p->px_dip, "px_ib_ino_rem_intr px_intr_setstate failed\n"); return (ret); } } /* Search the link list for ih_p */ for (i = 0; (i < ino_p->ino_ih_size) && (ih_lst->ih_next != ih_p); i++, ih_lst = ih_lst->ih_next); if (ih_lst->ih_next != ih_p) goto not_found; /* Remove ih_p from the link list and maintain the head/tail */ ih_lst->ih_next = ih_p->ih_next; if (ino_p->ino_ih_head == ih_p) ino_p->ino_ih_head = ih_p->ih_next; if (ino_p->ino_ih_tail == ih_p) ino_p->ino_ih_tail = ih_lst; ino_p->ino_ih_start = ino_p->ino_ih_head; reset: if (ih_p->ih_config_handle) pci_config_teardown(&ih_p->ih_config_handle); if (ih_p->ih_ksp != NULL) kstat_delete(ih_p->ih_ksp); kmem_free(ih_p, sizeof (px_ih_t)); ino_p->ino_ih_size--; return (ret); not_found: DBG(DBG_R_INTX, ino_p->ino_ib_p->ib_px_p->px_dip, "ino_p=%x does not have ih_p=%x\n", ino_p, ih_p); return (DDI_FAILURE); } px_ih_t * px_ib_ino_locate_intr(px_ib_ino_info_t *ino_p, dev_info_t *rdip, uint32_t inum, msiq_rec_type_t rec_type, msgcode_t msg_code) { px_ih_t *ih_lst = ino_p->ino_ih_head; int i; for (i = 0; i < ino_p->ino_ih_size; i++, ih_lst = ih_lst->ih_next) { if ((ih_lst->ih_dip == rdip) && (ih_lst->ih_inum == inum) && (ih_lst->ih_rec_type == rec_type) && (ih_lst->ih_msg_code == msg_code)) return (ih_lst); } return ((px_ih_t *)NULL); } px_ih_t * px_ib_alloc_ih(dev_info_t *rdip, uint32_t inum, uint_t (*int_handler)(caddr_t int_handler_arg1, caddr_t int_handler_arg2), caddr_t int_handler_arg1, caddr_t int_handler_arg2, msiq_rec_type_t rec_type, msgcode_t msg_code) { px_ih_t *ih_p; ih_p = kmem_alloc(sizeof (px_ih_t), KM_SLEEP); ih_p->ih_dip = rdip; ih_p->ih_inum = inum; ih_p->ih_intr_state = PX_INTR_STATE_DISABLE; ih_p->ih_handler = int_handler; ih_p->ih_handler_arg1 = int_handler_arg1; ih_p->ih_handler_arg2 = int_handler_arg2; ih_p->ih_config_handle = NULL; ih_p->ih_rec_type = rec_type; ih_p->ih_msg_code = msg_code; ih_p->ih_nsec = 0; ih_p->ih_ticks = 0; ih_p->ih_ksp = NULL; return (ih_p); } int px_ib_update_intr_state(px_t *px_p, dev_info_t *rdip, uint_t inum, devino_t ino, uint_t new_intr_state, msiq_rec_type_t rec_type, msgcode_t msg_code) { px_ib_t *ib_p = px_p->px_ib_p; px_ib_ino_info_t *ino_p; px_ih_t *ih_p; int ret = DDI_FAILURE; DBG(DBG_IB, px_p->px_dip, "ib_update_intr_state: %s%d " "inum %x devino %x state %x\n", ddi_driver_name(rdip), ddi_get_instance(rdip), inum, ino, new_intr_state); mutex_enter(&ib_p->ib_ino_lst_mutex); if (ino_p = px_ib_locate_ino(ib_p, ino)) { if (ih_p = px_ib_ino_locate_intr(ino_p, rdip, inum, rec_type, msg_code)) { ih_p->ih_intr_state = new_intr_state; ret = DDI_SUCCESS; } } mutex_exit(&ib_p->ib_ino_lst_mutex); return (ret); } static void px_fill_in_intr_devs(pcitool_intr_dev_t *dev, char *driver_name, char *path_name, int instance) { (void) strncpy(dev->driver_name, driver_name, MAXMODCONFNAME-1); dev->driver_name[MAXMODCONFNAME] = '\0'; (void) strncpy(dev->path, path_name, MAXPATHLEN-1); dev->dev_inst = instance; } /* * Return the dips or number of dips associated with a given interrupt block. * Size of dips array arg is passed in as dips_ret arg. * Number of dips returned is returned in dips_ret arg. * Array of dips gets returned in the dips argument. * Function returns number of dips existing for the given interrupt block. * * Note: this function assumes an enabled/valid INO, which is why it returns * the px node and (Internal) when it finds no other devices (and *devs_ret > 0) */ uint8_t pxtool_ib_get_ino_devs( px_t *px_p, uint32_t ino, uint8_t *devs_ret, pcitool_intr_dev_t *devs) { px_ib_t *ib_p = px_p->px_ib_p; px_ib_ino_info_t *ino_p; px_ih_t *ih_p; uint32_t num_devs = 0; char pathname[MAXPATHLEN]; int i; mutex_enter(&ib_p->ib_ino_lst_mutex); ino_p = px_ib_locate_ino(ib_p, ino); if (ino_p != NULL) { num_devs = ino_p->ino_ih_size; for (i = 0, ih_p = ino_p->ino_ih_head; ((i < ino_p->ino_ih_size) && (i < *devs_ret)); i++, ih_p = ih_p->ih_next) { (void) ddi_pathname(ih_p->ih_dip, pathname); px_fill_in_intr_devs(&devs[i], (char *)ddi_driver_name(ih_p->ih_dip), pathname, ddi_get_instance(ih_p->ih_dip)); } *devs_ret = i; } else if (*devs_ret > 0) { (void) ddi_pathname(px_p->px_dip, pathname); strcat(pathname, " (Internal)"); px_fill_in_intr_devs(&devs[0], (char *)ddi_driver_name(px_p->px_dip), pathname, ddi_get_instance(px_p->px_dip)); num_devs = *devs_ret = 1; } mutex_exit(&ib_p->ib_ino_lst_mutex); return (num_devs); } void px_ib_log_new_cpu(px_ib_t *ib_p, uint32_t old_cpu_id, uint32_t new_cpu_id, uint32_t ino) { px_ib_ino_info_t *ino_p; mutex_enter(&ib_p->ib_ino_lst_mutex); /* Log in OS data structures the new CPU. */ ino_p = px_ib_locate_ino(ib_p, ino); if (ino_p != NULL) { /* Log in OS data structures the new CPU. */ ino_p->ino_cpuid = new_cpu_id; /* Account for any residual time to be logged for old cpu. */ px_ib_cpu_ticks_to_ih_nsec(ib_p, ino_p->ino_ih_head, old_cpu_id); } mutex_exit(&ib_p->ib_ino_lst_mutex); }