1 /*
2 * CDDL HEADER START
3 *
4 * The contents of this file are subject to the terms of the
5 * Common Development and Distribution License (the "License").
6 * You may not use this file except in compliance with the License.
7 *
8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9 * or http://www.opensolaris.org/os/licensing.
10 * See the License for the specific language governing permissions
11 * and limitations under the License.
12 *
13 * When distributing Covered Code, include this CDDL HEADER in each
14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15 * If applicable, add the following below this CDDL HEADER, with the
16 * fields enclosed by brackets "[]" replaced with your own identifying
17 * information: Portions Copyright [yyyy] [name of copyright owner]
18 *
19 * CDDL HEADER END
20 */
21 /*
22 * Copyright (c) 2007, 2010, Oracle and/or its affiliates. All rights reserved.
23 */
24
25 #include <sys/x86_archext.h>
26 #include <sys/machsystm.h>
27 #include <sys/x_call.h>
28 #include <sys/acpi/acpi.h>
29 #include <sys/acpica.h>
30 #include <sys/pwrnow.h>
31 #include <sys/cpu_acpi.h>
32 #include <sys/cpupm.h>
33 #include <sys/dtrace.h>
34 #include <sys/sdt.h>
35
36 static int pwrnow_init(cpu_t *);
37 static void pwrnow_fini(cpu_t *);
38 static void pwrnow_power(cpuset_t, uint32_t);
39 static void pwrnow_stop(cpu_t *);
40
41 static boolean_t pwrnow_cpb_supported(void);
42
43 /*
44 * Interfaces for modules implementing AMD's PowerNow!.
45 */
46 cpupm_state_ops_t pwrnow_ops = {
47 "PowerNow! Technology",
48 pwrnow_init,
49 pwrnow_fini,
50 pwrnow_power,
51 pwrnow_stop
52 };
53
54 /*
55 * Error returns
56 */
57 #define PWRNOW_RET_SUCCESS 0x00
58 #define PWRNOW_RET_NO_PM 0x01
59 #define PWRNOW_RET_UNSUP_STATE 0x02
60 #define PWRNOW_RET_TRANS_INCOMPLETE 0x03
61
62 #define PWRNOW_LATENCY_WAIT 10
63
64 /*
65 * MSR registers for changing and reading processor power state.
66 */
67 #define PWRNOW_PERF_CTL_MSR 0xC0010062
68 #define PWRNOW_PERF_STATUS_MSR 0xC0010063
69
70 #define AMD_CPUID_PSTATE_HARDWARE (1<<7)
71 #define AMD_CPUID_TSC_CONSTANT (1<<8)
72 #define AMD_CPUID_CPB (1<<9)
73
74 /*
75 * Debugging support
76 */
77 #ifdef DEBUG
78 volatile int pwrnow_debug = 0;
79 #define PWRNOW_DEBUG(arglist) if (pwrnow_debug) printf arglist;
80 #else
81 #define PWRNOW_DEBUG(arglist)
82 #endif
83
84 /*
85 * Write the ctrl register.
86 */
87 static void
write_ctrl(cpu_acpi_handle_t handle,uint32_t ctrl)88 write_ctrl(cpu_acpi_handle_t handle, uint32_t ctrl)
89 {
90 cpu_acpi_pct_t *pct_ctrl;
91 uint64_t reg;
92
93 pct_ctrl = CPU_ACPI_PCT_CTRL(handle);
94
95 switch (pct_ctrl->cr_addrspace_id) {
96 case ACPI_ADR_SPACE_FIXED_HARDWARE:
97 reg = ctrl;
98 wrmsr(PWRNOW_PERF_CTL_MSR, reg);
99 break;
100
101 default:
102 DTRACE_PROBE1(pwrnow_ctrl_unsupported_type, uint8_t,
103 pct_ctrl->cr_addrspace_id);
104 return;
105 }
106
107 DTRACE_PROBE1(pwrnow_ctrl_write, uint32_t, ctrl);
108 }
109
110 /*
111 * Transition the current processor to the requested state.
112 */
113 static void
pwrnow_pstate_transition(uint32_t req_state)114 pwrnow_pstate_transition(uint32_t req_state)
115 {
116 cpupm_mach_state_t *mach_state =
117 (cpupm_mach_state_t *)CPU->cpu_m.mcpu_pm_mach_state;
118 cpu_acpi_handle_t handle = mach_state->ms_acpi_handle;
119 cpu_acpi_pstate_t *req_pstate;
120 uint32_t ctrl;
121
122 req_pstate = (cpu_acpi_pstate_t *)CPU_ACPI_PSTATES(handle);
123 req_pstate += req_state;
124
125 DTRACE_PROBE1(pwrnow_transition_freq, uint32_t,
126 CPU_ACPI_FREQ(req_pstate));
127
128 /*
129 * Initiate the processor p-state change.
130 */
131 ctrl = CPU_ACPI_PSTATE_CTRL(req_pstate);
132 write_ctrl(handle, ctrl);
133
134 if (mach_state->ms_turbo != NULL)
135 cpupm_record_turbo_info(mach_state->ms_turbo,
136 mach_state->ms_pstate.cma_state.pstate, req_state);
137
138 mach_state->ms_pstate.cma_state.pstate = req_state;
139 cpu_set_curr_clock((uint64_t)CPU_ACPI_FREQ(req_pstate) * 1000000);
140 }
141
142 static void
pwrnow_power(cpuset_t set,uint32_t req_state)143 pwrnow_power(cpuset_t set, uint32_t req_state)
144 {
145 /*
146 * If thread is already running on target CPU then just
147 * make the transition request. Otherwise, we'll need to
148 * make a cross-call.
149 */
150 kpreempt_disable();
151 if (CPU_IN_SET(set, CPU->cpu_id)) {
152 pwrnow_pstate_transition(req_state);
153 CPUSET_DEL(set, CPU->cpu_id);
154 }
155 if (!CPUSET_ISNULL(set)) {
156 xc_call((xc_arg_t)req_state, NULL, NULL,
157 CPUSET2BV(set), (xc_func_t)pwrnow_pstate_transition);
158 }
159 kpreempt_enable();
160 }
161
162 /*
163 * Validate that this processor supports PowerNow! and if so,
164 * get the P-state data from ACPI and cache it.
165 */
166 static int
pwrnow_init(cpu_t * cp)167 pwrnow_init(cpu_t *cp)
168 {
169 cpupm_mach_state_t *mach_state =
170 (cpupm_mach_state_t *)cp->cpu_m.mcpu_pm_mach_state;
171 cpu_acpi_handle_t handle = mach_state->ms_acpi_handle;
172 cpu_acpi_pct_t *pct_stat;
173 static int logged = 0;
174
175 PWRNOW_DEBUG(("pwrnow_init: processor %d\n", cp->cpu_id));
176
177 /*
178 * Cache the P-state specific ACPI data.
179 */
180 if (cpu_acpi_cache_pstate_data(handle) != 0) {
181 if (!logged) {
182 cmn_err(CE_NOTE, "!PowerNow! support is being "
183 "disabled due to errors parsing ACPI P-state "
184 "objects exported by BIOS.");
185 logged = 1;
186 }
187 pwrnow_fini(cp);
188 return (PWRNOW_RET_NO_PM);
189 }
190
191 pct_stat = CPU_ACPI_PCT_STATUS(handle);
192 switch (pct_stat->cr_addrspace_id) {
193 case ACPI_ADR_SPACE_FIXED_HARDWARE:
194 PWRNOW_DEBUG(("Transitions will use fixed hardware\n"));
195 break;
196 default:
197 cmn_err(CE_WARN, "!_PCT configured for unsupported "
198 "addrspace = %d.", pct_stat->cr_addrspace_id);
199 cmn_err(CE_NOTE, "!CPU power management will not function.");
200 pwrnow_fini(cp);
201 return (PWRNOW_RET_NO_PM);
202 }
203
204 cpupm_alloc_domains(cp, CPUPM_P_STATES);
205
206 /*
207 * Check for Core Performance Boost support
208 */
209 if (pwrnow_cpb_supported())
210 mach_state->ms_turbo = cpupm_turbo_init(cp);
211
212 PWRNOW_DEBUG(("Processor %d succeeded.\n", cp->cpu_id))
213 return (PWRNOW_RET_SUCCESS);
214 }
215
216 /*
217 * Free resources allocated by pwrnow_init().
218 */
219 static void
pwrnow_fini(cpu_t * cp)220 pwrnow_fini(cpu_t *cp)
221 {
222 cpupm_mach_state_t *mach_state =
223 (cpupm_mach_state_t *)(cp->cpu_m.mcpu_pm_mach_state);
224 cpu_acpi_handle_t handle = mach_state->ms_acpi_handle;
225
226 cpupm_free_domains(&cpupm_pstate_domains);
227 cpu_acpi_free_pstate_data(handle);
228
229 if (mach_state->ms_turbo != NULL)
230 cpupm_turbo_fini(mach_state->ms_turbo);
231 mach_state->ms_turbo = NULL;
232 }
233
234 boolean_t
pwrnow_supported()235 pwrnow_supported()
236 {
237 struct cpuid_regs cpu_regs;
238
239 /* Required features */
240 if (!is_x86_feature(x86_featureset, X86FSET_CPUID) ||
241 !is_x86_feature(x86_featureset, X86FSET_MSR)) {
242 PWRNOW_DEBUG(("No CPUID or MSR support."));
243 return (B_FALSE);
244 }
245
246 /*
247 * Get the Advanced Power Management Information.
248 */
249 cpu_regs.cp_eax = 0x80000007;
250 (void) __cpuid_insn(&cpu_regs);
251
252 /*
253 * We currently only support CPU power management of
254 * processors that are P-state TSC invariant
255 */
256 if (!(cpu_regs.cp_edx & AMD_CPUID_TSC_CONSTANT)) {
257 PWRNOW_DEBUG(("No support for CPUs that are not P-state "
258 "TSC invariant.\n"));
259 return (B_FALSE);
260 }
261
262 /*
263 * We only support the "Fire and Forget" style of PowerNow! (i.e.,
264 * single MSR write to change speed).
265 */
266 if (!(cpu_regs.cp_edx & AMD_CPUID_PSTATE_HARDWARE)) {
267 PWRNOW_DEBUG(("Hardware P-State control is not supported.\n"));
268 return (B_FALSE);
269 }
270 return (B_TRUE);
271 }
272
273 static boolean_t
pwrnow_cpb_supported(void)274 pwrnow_cpb_supported(void)
275 {
276 struct cpuid_regs cpu_regs;
277
278 /* Required features */
279 if (!is_x86_feature(x86_featureset, X86FSET_CPUID) ||
280 !is_x86_feature(x86_featureset, X86FSET_MSR)) {
281 PWRNOW_DEBUG(("No CPUID or MSR support."));
282 return (B_FALSE);
283 }
284
285 /*
286 * Get the Advanced Power Management Information.
287 */
288 cpu_regs.cp_eax = 0x80000007;
289 (void) __cpuid_insn(&cpu_regs);
290
291 if (!(cpu_regs.cp_edx & AMD_CPUID_CPB))
292 return (B_FALSE);
293
294 return (B_TRUE);
295 }
296
297 static void
pwrnow_stop(cpu_t * cp)298 pwrnow_stop(cpu_t *cp)
299 {
300 cpupm_mach_state_t *mach_state =
301 (cpupm_mach_state_t *)(cp->cpu_m.mcpu_pm_mach_state);
302 cpu_acpi_handle_t handle = mach_state->ms_acpi_handle;
303
304 cpupm_remove_domains(cp, CPUPM_P_STATES, &cpupm_pstate_domains);
305 cpu_acpi_free_pstate_data(handle);
306
307 if (mach_state->ms_turbo != NULL)
308 cpupm_turbo_fini(mach_state->ms_turbo);
309 mach_state->ms_turbo = NULL;
310 }
311