/* * CDDL HEADER START * * The contents of this file are subject to the terms of the * Common Development and Distribution License (the "License"). * You may not use this file except in compliance with the License. * * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE * or http://www.opensolaris.org/os/licensing. * See the License for the specific language governing permissions * and limitations under the License. * * When distributing Covered Code, include this CDDL HEADER in each * file and include the License file at usr/src/OPENSOLARIS.LICENSE. * If applicable, add the following below this CDDL HEADER, with the * fields enclosed by brackets "[]" replaced with your own identifying * information: Portions Copyright [yyyy] [name of copyright owner] * * CDDL HEADER END */ /* * Copyright (c) 1993, 2010, Oracle and/or its affiliates. All rights reserved. * Copyright 2019 Peter Tribble. */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include /* * fpRAS implementation structures. */ struct fpras_chkfn *fpras_chkfnaddrs[FPRAS_NCOPYOPS]; struct fpras_chkfngrp *fpras_chkfngrps; struct fpras_chkfngrp *fpras_chkfngrps_base; int fpras_frequency = -1; int64_t fpras_interval = -1; /* * Increase unix symbol table size as a work around for 6828121 */ int alloc_mem_bermuda_triangle; /* * Halt idling cpus optimization * * This optimation is only enabled in platforms that have * the CPU halt support. The cpu_halt_cpu() support is provided * in the cpu module and it is referenced here with a pragma weak. * The presence of this routine automatically enable the halt idling * cpus functionality if the global switch enable_halt_idle_cpus * is set (default is set). * */ #pragma weak cpu_halt_cpu extern void cpu_halt_cpu(); /* * Defines for the idle_state_transition DTrace probe * * The probe fires when the CPU undergoes an idle state change (e.g. halting) * The agument passed is the state to which the CPU is transitioning. * * The states are defined here. */ #define IDLE_STATE_NORMAL 0 #define IDLE_STATE_HALTED 1 int enable_halt_idle_cpus = 1; /* global switch */ uint_t cp_haltset_fanout = 3; void setup_trap_table(void) { intr_init(CPU); /* init interrupt request free list */ setwstate(WSTATE_KERN); prom_set_traptable(&trap_table); } void mach_fpras() { if (fpras_implemented && !fpras_disable) { int i; struct fpras_chkfngrp *fcgp; size_t chkfngrpsallocsz; /* * Note that we size off of NCPU and setup for * all those possibilities regardless of whether * the cpu id is present or not. We do this so that * we don't have any construction or destruction * activity to perform at DR time, and it's not * costly in memory. We require block alignment. */ chkfngrpsallocsz = NCPU * sizeof (struct fpras_chkfngrp); fpras_chkfngrps_base = kmem_alloc(chkfngrpsallocsz, KM_SLEEP); if (IS_P2ALIGNED((uintptr_t)fpras_chkfngrps_base, 64)) { fpras_chkfngrps = fpras_chkfngrps_base; } else { kmem_free(fpras_chkfngrps_base, chkfngrpsallocsz); chkfngrpsallocsz += 64; fpras_chkfngrps_base = kmem_alloc(chkfngrpsallocsz, KM_SLEEP); fpras_chkfngrps = (struct fpras_chkfngrp *) P2ROUNDUP((uintptr_t)fpras_chkfngrps_base, 64); } /* * Copy our check function into place for each copy operation * and each cpu id. */ fcgp = &fpras_chkfngrps[0]; for (i = 0; i < FPRAS_NCOPYOPS; ++i) bcopy((void *)fpras_chkfn_type1, &fcgp->fpras_fn[i], sizeof (struct fpras_chkfn)); for (i = 1; i < NCPU; ++i) *(&fpras_chkfngrps[i]) = *fcgp; /* * At definition fpras_frequency is set to -1, and it will * still have that value unless changed in /etc/system (not * strictly supported, but not preventable). The following * both sets the default and sanity checks anything from * /etc/system. */ if (fpras_frequency < 0) fpras_frequency = FPRAS_DEFAULT_FREQUENCY; /* * Now calculate fpras_interval. When fpras_interval * becomes non-negative fpras checks will commence * (copies before this point in boot will bypass fpras). * Our stores of instructions must be visible; no need * to flush as they're never been executed before. */ membar_producer(); fpras_interval = (fpras_frequency == 0) ? 0 : sys_tick_freq / fpras_frequency; } } void mach_hw_copy_limit(void) { if (!fpu_exists) { use_hw_bcopy = 0; hw_copy_limit_1 = 0; hw_copy_limit_2 = 0; hw_copy_limit_4 = 0; hw_copy_limit_8 = 0; use_hw_bzero = 0; } } void load_tod_module() { /* * Load tod driver module for the tod part found on this system. * Recompute the cpu frequency/delays based on tod as tod part * tends to keep time more accurately. */ if (tod_module_name == NULL || modload("tod", tod_module_name) == -1) halt("Can't load tod module"); } void mach_memscrub(void) { /* * Startup memory scrubber, if not running fpu emulation code. */ #ifndef _HW_MEMSCRUB_SUPPORT if (fpu_exists) { if (memscrub_init()) { cmn_err(CE_WARN, "Memory scrubber failed to initialize"); } } #endif /* _HW_MEMSCRUB_SUPPORT */ } /* * Halt the present CPU until awoken via an interrupt. * This routine should only be invoked if cpu_halt_cpu() * exists and is supported, see mach_cpu_halt_idle() */ void cpu_halt(void) { cpu_t *cpup = CPU; processorid_t cpu_sid = cpup->cpu_seqid; cpupart_t *cp = cpup->cpu_part; int hset_update = 1; volatile int *p = &cpup->cpu_disp->disp_nrunnable; uint_t s; /* * If this CPU is online then we should notate our halting * by adding ourselves to the partition's halted CPU * bitset. This allows other CPUs to find/awaken us when * work becomes available. */ if (CPU->cpu_flags & CPU_OFFLINE) hset_update = 0; /* * Add ourselves to the partition's halted CPUs bitset * and set our HALTED flag, if necessary. * * When a thread becomes runnable, it is placed on the queue * and then the halted cpu bitset is checked to determine who * (if anyone) should be awoken. We therefore need to first * add ourselves to the halted bitset, and then check if there * is any work available. The order is important to prevent a race * that can lead to work languishing on a run queue somewhere while * this CPU remains halted. * * Either the producing CPU will see we're halted and will awaken us, * or this CPU will see the work available in disp_anywork() */ if (hset_update) { cpup->cpu_disp_flags |= CPU_DISP_HALTED; membar_producer(); bitset_atomic_add(&cp->cp_haltset, cpu_sid); } /* * Check to make sure there's really nothing to do. * Work destined for this CPU may become available after * this check. We'll be notified through the clearing of our * bit in the halted CPU bitset, and a poke. */ if (disp_anywork()) { if (hset_update) { cpup->cpu_disp_flags &= ~CPU_DISP_HALTED; bitset_atomic_del(&cp->cp_haltset, cpu_sid); } return; } /* * We're on our way to being halted. Wait until something becomes * runnable locally or we are awaken (i.e. removed from the halt set). * Note that the call to hv_cpu_yield() can return even if we have * nothing to do. * * Disable interrupts now, so that we'll awaken immediately * after halting if someone tries to poke us between now and * the time we actually halt. * * We check for the presence of our bit after disabling interrupts. * If it's cleared, we'll return. If the bit is cleared after * we check then the poke will pop us out of the halted state. * Also, if the offlined CPU has been brought back on-line, then * we return as well. * * The ordering of the poke and the clearing of the bit by cpu_wakeup * is important. * cpu_wakeup() must clear, then poke. * cpu_halt() must disable interrupts, then check for the bit. * * The check for anything locally runnable is here for performance * and isn't needed for correctness. disp_nrunnable ought to be * in our cache still, so it's inexpensive to check, and if there * is anything runnable we won't have to wait for the poke. * * Any interrupt will awaken the cpu from halt. Looping here * will filter spurious interrupts that wake us up, but don't * represent a need for us to head back out to idle(). This * will enable the idle loop to be more efficient and sleep in * the processor pipeline for a larger percent of the time, * which returns useful cycles to the peer hardware strand * that shares the pipeline. */ s = disable_vec_intr(); while (*p == 0 && ((hset_update && bitset_in_set(&cp->cp_haltset, cpu_sid)) || (!hset_update && (CPU->cpu_flags & CPU_OFFLINE)))) { DTRACE_PROBE1(idle__state__transition, uint_t, IDLE_STATE_HALTED); (void) cpu_halt_cpu(); DTRACE_PROBE1(idle__state__transition, uint_t, IDLE_STATE_NORMAL); enable_vec_intr(s); s = disable_vec_intr(); } /* * We're no longer halted */ enable_vec_intr(s); if (hset_update) { cpup->cpu_disp_flags &= ~CPU_DISP_HALTED; bitset_atomic_del(&cp->cp_haltset, cpu_sid); } } /* * If "cpu" is halted, then wake it up clearing its halted bit in advance. * Otherwise, see if other CPUs in the cpu partition are halted and need to * be woken up so that they can steal the thread we placed on this CPU. * This function is only used on MP systems. * This function should only be invoked if cpu_halt_cpu() * exists and is supported, see mach_cpu_halt_idle() */ static void cpu_wakeup(cpu_t *cpu, int bound) { uint_t cpu_found; processorid_t cpu_sid; cpupart_t *cp; cp = cpu->cpu_part; cpu_sid = cpu->cpu_seqid; if (bitset_in_set(&cp->cp_haltset, cpu_sid)) { /* * Clear the halted bit for that CPU since it will be * poked in a moment. */ bitset_atomic_del(&cp->cp_haltset, cpu_sid); /* * We may find the current CPU present in the halted cpu bitset * if we're in the context of an interrupt that occurred * before we had a chance to clear our bit in cpu_halt(). * Poking ourself is obviously unnecessary, since if * we're here, we're not halted. */ if (cpu != CPU) poke_cpu(cpu->cpu_id); return; } else { /* * This cpu isn't halted, but it's idle or undergoing a * context switch. No need to awaken anyone else. */ if (cpu->cpu_thread == cpu->cpu_idle_thread || cpu->cpu_disp_flags & CPU_DISP_DONTSTEAL) return; } /* * No need to wake up other CPUs if this is for a bound thread. */ if (bound) return; /* * The CPU specified for wakeup isn't currently halted, so check * to see if there are any other halted CPUs in the partition, * and if there are then awaken one. * * If possible, try to select a CPU close to the target, since this * will likely trigger a migration. */ do { cpu_found = bitset_find(&cp->cp_haltset); if (cpu_found == (uint_t)-1) return; } while (bitset_atomic_test_and_del(&cp->cp_haltset, cpu_found) < 0); if (cpu_found != CPU->cpu_seqid) poke_cpu(cpu_seq[cpu_found]->cpu_id); } void mach_cpu_halt_idle(void) { if (enable_halt_idle_cpus) { if (&cpu_halt_cpu) { idle_cpu = cpu_halt; disp_enq_thread = cpu_wakeup; } } } /*ARGSUSED*/ int cpu_intrq_setup(struct cpu *cp) { /* Interrupt mondo queues not applicable to sun4u */ return (0); } /*ARGSUSED*/ void cpu_intrq_cleanup(struct cpu *cp) { /* Interrupt mondo queues not applicable to sun4u */ } /*ARGSUSED*/ void cpu_intrq_register(struct cpu *cp) { /* Interrupt/error queues not applicable to sun4u */ } /*ARGSUSED*/ void mach_htraptrace_setup(int cpuid) { /* Setup hypervisor traptrace buffer, not applicable to sun4u */ } /*ARGSUSED*/ void mach_htraptrace_configure(int cpuid) { /* enable/ disable hypervisor traptracing, not applicable to sun4u */ } /*ARGSUSED*/ void mach_htraptrace_cleanup(int cpuid) { /* cleanup hypervisor traptrace buffer, not applicable to sun4u */ } void mach_descrip_startup_init(void) { /* * Only for sun4v. * Initialize Machine description framework during startup. */ } void mach_descrip_startup_fini(void) { /* * Only for sun4v. * Clean up Machine Description framework during startup. */ } void mach_descrip_init(void) { /* * Only for sun4v. * Initialize Machine description framework. */ } void hsvc_setup(void) { /* Setup hypervisor services, not applicable to sun4u */ } void load_mach_drivers(void) { /* Currently no machine class (sun4u) specific drivers to load */ } /* * Find a physically contiguous area of twice the largest ecache size * to be used while doing displacement flush of ecaches. */ uint64_t ecache_flush_address(void) { struct memlist *pmem; uint64_t flush_size; uint64_t ret_val; flush_size = ecache_size * 2; for (pmem = phys_install; pmem; pmem = pmem->ml_next) { ret_val = P2ROUNDUP(pmem->ml_address, ecache_size); if (ret_val + flush_size <= pmem->ml_address + pmem->ml_size) return (ret_val); } return ((uint64_t)-1); } /* * Called with the memlist lock held to say that phys_install has * changed. */ void phys_install_has_changed(void) { /* * Get the new address into a temporary just in case panicking * involves use of ecache_flushaddr. */ uint64_t new_addr; new_addr = ecache_flush_address(); if (new_addr == (uint64_t)-1) { cmn_err(CE_PANIC, "ecache_flush_address(): failed, ecache_size=%x", ecache_size); /*NOTREACHED*/ } ecache_flushaddr = new_addr; membar_producer(); }