/* * CDDL HEADER START * * The contents of this file are subject to the terms of the * Common Development and Distribution License (the "License"). * You may not use this file except in compliance with the License. * * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE * or http://www.opensolaris.org/os/licensing. * See the License for the specific language governing permissions * and limitations under the License. * * When distributing Covered Code, include this CDDL HEADER in each * file and include the License file at usr/src/OPENSOLARIS.LICENSE. * If applicable, add the following below this CDDL HEADER, with the * fields enclosed by brackets "[]" replaced with your own identifying * information: Portions Copyright [yyyy] [name of copyright owner] * * CDDL HEADER END */ /* * Copyright (c) 2007, 2010, Oracle and/or its affiliates. All rights reserved. */ #include #include #include #include #include #include #include #include #include #include static int pwrnow_init(cpu_t *); static void pwrnow_fini(cpu_t *); static void pwrnow_power(cpuset_t, uint32_t); static void pwrnow_stop(cpu_t *); static boolean_t pwrnow_cpb_supported(void); /* * Interfaces for modules implementing AMD's PowerNow!. */ cpupm_state_ops_t pwrnow_ops = { "PowerNow! Technology", pwrnow_init, pwrnow_fini, pwrnow_power, pwrnow_stop }; /* * Error returns */ #define PWRNOW_RET_SUCCESS 0x00 #define PWRNOW_RET_NO_PM 0x01 #define PWRNOW_RET_UNSUP_STATE 0x02 #define PWRNOW_RET_TRANS_INCOMPLETE 0x03 #define PWRNOW_LATENCY_WAIT 10 /* * MSR registers for changing and reading processor power state. */ #define PWRNOW_PERF_CTL_MSR 0xC0010062 #define PWRNOW_PERF_STATUS_MSR 0xC0010063 #define AMD_CPUID_PSTATE_HARDWARE (1<<7) #define AMD_CPUID_TSC_CONSTANT (1<<8) #define AMD_CPUID_CPB (1<<9) /* * Debugging support */ #ifdef DEBUG volatile int pwrnow_debug = 0; #define PWRNOW_DEBUG(arglist) if (pwrnow_debug) printf arglist; #else #define PWRNOW_DEBUG(arglist) #endif /* * Write the ctrl register. */ static void write_ctrl(cpu_acpi_handle_t handle, uint32_t ctrl) { cpu_acpi_pct_t *pct_ctrl; uint64_t reg; pct_ctrl = CPU_ACPI_PCT_CTRL(handle); switch (pct_ctrl->cr_addrspace_id) { case ACPI_ADR_SPACE_FIXED_HARDWARE: reg = ctrl; wrmsr(PWRNOW_PERF_CTL_MSR, reg); break; default: DTRACE_PROBE1(pwrnow_ctrl_unsupported_type, uint8_t, pct_ctrl->cr_addrspace_id); return; } DTRACE_PROBE1(pwrnow_ctrl_write, uint32_t, ctrl); } /* * Transition the current processor to the requested state. */ static int pwrnow_pstate_transition(xc_arg_t arg1, xc_arg_t arg2 __unused, xc_arg_t arg3 __unused) { uint32_t req_state = (uint32_t)arg1; cpupm_mach_state_t *mach_state = (cpupm_mach_state_t *)CPU->cpu_m.mcpu_pm_mach_state; cpu_acpi_handle_t handle = mach_state->ms_acpi_handle; cpu_acpi_pstate_t *req_pstate; uint32_t ctrl; req_pstate = (cpu_acpi_pstate_t *)CPU_ACPI_PSTATES(handle); req_pstate += req_state; DTRACE_PROBE1(pwrnow_transition_freq, uint32_t, CPU_ACPI_FREQ(req_pstate)); /* * Initiate the processor p-state change. */ ctrl = CPU_ACPI_PSTATE_CTRL(req_pstate); write_ctrl(handle, ctrl); if (mach_state->ms_turbo != NULL) cpupm_record_turbo_info(mach_state->ms_turbo, mach_state->ms_pstate.cma_state.pstate, req_state); mach_state->ms_pstate.cma_state.pstate = req_state; cpu_set_curr_clock((uint64_t)CPU_ACPI_FREQ(req_pstate) * 1000000); return (0); } static void pwrnow_power(cpuset_t set, uint32_t req_state) { /* * If thread is already running on target CPU then just * make the transition request. Otherwise, we'll need to * make a cross-call. */ kpreempt_disable(); if (CPU_IN_SET(set, CPU->cpu_id)) { (void) pwrnow_pstate_transition(req_state, 0, 0); CPUSET_DEL(set, CPU->cpu_id); } if (!CPUSET_ISNULL(set)) { xc_call((xc_arg_t)req_state, 0, 0, CPUSET2BV(set), pwrnow_pstate_transition); } kpreempt_enable(); } /* * Validate that this processor supports PowerNow! and if so, * get the P-state data from ACPI and cache it. */ static int pwrnow_init(cpu_t *cp) { cpupm_mach_state_t *mach_state = (cpupm_mach_state_t *)cp->cpu_m.mcpu_pm_mach_state; cpu_acpi_handle_t handle = mach_state->ms_acpi_handle; cpu_acpi_pct_t *pct_stat; static int logged = 0; PWRNOW_DEBUG(("pwrnow_init: processor %d\n", cp->cpu_id)); /* * Cache the P-state specific ACPI data. */ if (cpu_acpi_cache_pstate_data(handle) != 0) { if (!logged) { cmn_err(CE_NOTE, "!PowerNow! support is being " "disabled due to errors parsing ACPI P-state " "objects exported by BIOS."); logged = 1; } pwrnow_fini(cp); return (PWRNOW_RET_NO_PM); } pct_stat = CPU_ACPI_PCT_STATUS(handle); switch (pct_stat->cr_addrspace_id) { case ACPI_ADR_SPACE_FIXED_HARDWARE: PWRNOW_DEBUG(("Transitions will use fixed hardware\n")); break; default: cmn_err(CE_WARN, "!_PCT configured for unsupported " "addrspace = %d.", pct_stat->cr_addrspace_id); cmn_err(CE_NOTE, "!CPU power management will not function."); pwrnow_fini(cp); return (PWRNOW_RET_NO_PM); } cpupm_alloc_domains(cp, CPUPM_P_STATES); /* * Check for Core Performance Boost support */ if (pwrnow_cpb_supported()) mach_state->ms_turbo = cpupm_turbo_init(cp); PWRNOW_DEBUG(("Processor %d succeeded.\n", cp->cpu_id)) return (PWRNOW_RET_SUCCESS); } /* * Free resources allocated by pwrnow_init(). */ static void pwrnow_fini(cpu_t *cp) { cpupm_mach_state_t *mach_state = (cpupm_mach_state_t *)(cp->cpu_m.mcpu_pm_mach_state); cpu_acpi_handle_t handle = mach_state->ms_acpi_handle; cpupm_free_domains(&cpupm_pstate_domains); cpu_acpi_free_pstate_data(handle); if (mach_state->ms_turbo != NULL) cpupm_turbo_fini(mach_state->ms_turbo); mach_state->ms_turbo = NULL; } boolean_t pwrnow_supported() { struct cpuid_regs cpu_regs; /* Required features */ if (!is_x86_feature(x86_featureset, X86FSET_CPUID) || !is_x86_feature(x86_featureset, X86FSET_MSR)) { PWRNOW_DEBUG(("No CPUID or MSR support.")); return (B_FALSE); } /* * Get the Advanced Power Management Information. */ cpu_regs.cp_eax = 0x80000007; (void) __cpuid_insn(&cpu_regs); /* * We currently only support CPU power management of * processors that are P-state TSC invariant */ if (!(cpu_regs.cp_edx & AMD_CPUID_TSC_CONSTANT)) { PWRNOW_DEBUG(("No support for CPUs that are not P-state " "TSC invariant.\n")); return (B_FALSE); } /* * We only support the "Fire and Forget" style of PowerNow! (i.e., * single MSR write to change speed). */ if (!(cpu_regs.cp_edx & AMD_CPUID_PSTATE_HARDWARE)) { PWRNOW_DEBUG(("Hardware P-State control is not supported.\n")); return (B_FALSE); } return (B_TRUE); } static boolean_t pwrnow_cpb_supported(void) { struct cpuid_regs cpu_regs; /* Required features */ if (!is_x86_feature(x86_featureset, X86FSET_CPUID) || !is_x86_feature(x86_featureset, X86FSET_MSR)) { PWRNOW_DEBUG(("No CPUID or MSR support.")); return (B_FALSE); } /* * Get the Advanced Power Management Information. */ cpu_regs.cp_eax = 0x80000007; (void) __cpuid_insn(&cpu_regs); if (!(cpu_regs.cp_edx & AMD_CPUID_CPB)) return (B_FALSE); return (B_TRUE); } static void pwrnow_stop(cpu_t *cp) { cpupm_mach_state_t *mach_state = (cpupm_mach_state_t *)(cp->cpu_m.mcpu_pm_mach_state); cpu_acpi_handle_t handle = mach_state->ms_acpi_handle; cpupm_remove_domains(cp, CPUPM_P_STATES, &cpupm_pstate_domains); cpu_acpi_free_pstate_data(handle); if (mach_state->ms_turbo != NULL) cpupm_turbo_fini(mach_state->ms_turbo); mach_state->ms_turbo = NULL; }