xref: /titanic_41/usr/src/uts/intel/ia32/os/cpc_subr.c (revision 4a6822d07d6d3f9ffe6907ef5f10d11dcadd75c6)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 /*
22  * Copyright 2007 Sun Microsystems, Inc.  All rights reserved.
23  * Use is subject to license terms.
24  */
25 
26 #pragma ident	"%Z%%M%	%I%	%E% SMI"
27 
28 /*
29  * x86-specific routines used by the CPU Performance counter driver.
30  */
31 
32 #include <sys/types.h>
33 #include <sys/time.h>
34 #include <sys/atomic.h>
35 #include <sys/regset.h>
36 #include <sys/privregs.h>
37 #include <sys/x86_archext.h>
38 #include <sys/cpuvar.h>
39 #include <sys/machcpuvar.h>
40 #include <sys/archsystm.h>
41 #include <sys/cpc_pcbe.h>
42 #include <sys/cpc_impl.h>
43 #include <sys/x_call.h>
44 #include <sys/cmn_err.h>
45 #include <sys/cmt.h>
46 #include <sys/spl.h>
47 #include <sys/apic.h>
48 
49 static const uint64_t allstopped = 0;
50 static kcpc_ctx_t *(*overflow_intr_handler)(caddr_t);
51 
52 int kcpc_hw_overflow_intr_installed;		/* set by APIC code */
53 extern kcpc_ctx_t *kcpc_overflow_intr(caddr_t arg, uint64_t bitmap);
54 
55 extern int kcpc_counts_include_idle; /* Project Private /etc/system variable */
56 
57 void (*kcpc_hw_enable_cpc_intr)(void);		/* set by APIC code */
58 
59 int
60 kcpc_hw_add_ovf_intr(kcpc_ctx_t *(*handler)(caddr_t))
61 {
62 	if (x86_type != X86_TYPE_P6)
63 		return (0);
64 	overflow_intr_handler = handler;
65 	return (ipltospl(APIC_PCINT_IPL));
66 }
67 
68 void
69 kcpc_hw_rem_ovf_intr(void)
70 {
71 	overflow_intr_handler = NULL;
72 }
73 
74 /*
75  * Hook used on P4 systems to catch online/offline events.
76  */
77 /*ARGSUSED*/
78 static int
79 kcpc_cpu_setup(cpu_setup_t what, int cpuid, void *arg)
80 {
81 	pg_cmt_t	*chip_pg;
82 	int		active_cpus_cnt;
83 
84 	if (what != CPU_ON)
85 		return (0);
86 
87 	/*
88 	 * If any CPU-bound contexts exist, we don't need to invalidate
89 	 * anything, as no per-LWP contexts can coexist.
90 	 */
91 	if (kcpc_cpuctx)
92 		return (0);
93 
94 	/*
95 	 * If this chip now has more than 1 active cpu, we must invalidate all
96 	 * contexts in the system.
97 	 */
98 	chip_pg = (pg_cmt_t *)pghw_find_pg(cpu[cpuid], PGHW_CHIP);
99 	if (chip_pg != NULL) {
100 		active_cpus_cnt = GROUP_SIZE(&chip_pg->cmt_cpus_actv);
101 		if (active_cpus_cnt > 1)
102 			kcpc_invalidate_all();
103 	}
104 
105 	return (0);
106 }
107 
108 static kmutex_t cpu_setup_lock;	/* protects setup_registered */
109 static int setup_registered;
110 
111 void
112 kcpc_hw_init(cpu_t *cp)
113 {
114 	kthread_t *t = cp->cpu_idle_thread;
115 
116 	if (x86_feature & X86_HTT) {
117 		mutex_enter(&cpu_setup_lock);
118 		if (setup_registered == 0) {
119 			mutex_enter(&cpu_lock);
120 			register_cpu_setup_func(kcpc_cpu_setup, NULL);
121 			mutex_exit(&cpu_lock);
122 			setup_registered = 1;
123 		}
124 		mutex_exit(&cpu_setup_lock);
125 	}
126 
127 	mutex_init(&cp->cpu_cpc_ctxlock, "cpu_cpc_ctxlock", MUTEX_DEFAULT, 0);
128 
129 	if (kcpc_counts_include_idle)
130 		return;
131 
132 	installctx(t, cp, kcpc_idle_save, kcpc_idle_restore,
133 	    NULL, NULL, NULL, NULL);
134 }
135 
136 void
137 kcpc_hw_fini(cpu_t *cp)
138 {
139 	ASSERT(cp->cpu_idle_thread == NULL);
140 
141 	mutex_destroy(&cp->cpu_cpc_ctxlock);
142 }
143 
144 #define	BITS(v, u, l)	\
145 	(((v) >> (l)) & ((1 << (1 + (u) - (l))) - 1))
146 
147 #define	PCBE_NAMELEN 30	/* Enough Room for pcbe.manuf.model.family.stepping */
148 
149 /*
150  * Examine the processor and load an appropriate PCBE.
151  */
152 int
153 kcpc_hw_load_pcbe(void)
154 {
155 	return (kcpc_pcbe_tryload(cpuid_getvendorstr(CPU), cpuid_getfamily(CPU),
156 	    cpuid_getmodel(CPU), cpuid_getstep(CPU)));
157 }
158 
159 static int
160 kcpc_remotestop_func(void)
161 {
162 	ASSERT(CPU->cpu_cpc_ctx != NULL);
163 	pcbe_ops->pcbe_allstop();
164 	atomic_or_uint(&CPU->cpu_cpc_ctx->kc_flags, KCPC_CTX_INVALID_STOPPED);
165 
166 	return (0);
167 }
168 
169 /*
170  * Ensure the counters are stopped on the given processor.
171  *
172  * Callers must ensure kernel preemption is disabled.
173  */
174 void
175 kcpc_remote_stop(cpu_t *cp)
176 {
177 	cpuset_t set;
178 
179 	CPUSET_ZERO(set);
180 
181 	CPUSET_ADD(set, cp->cpu_id);
182 
183 	xc_sync(0, 0, 0, X_CALL_HIPRI, set, (xc_func_t)kcpc_remotestop_func);
184 }
185 
186 /*
187  * Called by the generic framework to check if it's OK to bind a set to a CPU.
188  */
189 int
190 kcpc_hw_cpu_hook(processorid_t cpuid, ulong_t *kcpc_cpumap)
191 {
192 	cpu_t		*cpu, *p;
193 	pg_t		*chip_pg;
194 	pg_cpu_itr_t	itr;
195 
196 	if ((x86_feature & X86_HTT) == 0)
197 		return (0);
198 
199 	/*
200 	 * Only one logical CPU on each Pentium 4 HT CPU may be bound to at
201 	 * once.
202 	 *
203 	 * This loop is protected by holding cpu_lock, in order to properly
204 	 * access the cpu_t of the desired cpu.
205 	 */
206 	mutex_enter(&cpu_lock);
207 	if ((cpu = cpu_get(cpuid)) == NULL) {
208 		mutex_exit(&cpu_lock);
209 		return (-1);
210 	}
211 
212 	chip_pg = (pg_t *)pghw_find_pg(cpu, PGHW_CHIP);
213 
214 	PG_CPU_ITR_INIT(chip_pg, itr);
215 	while ((p = pg_cpu_next(&itr)) != NULL) {
216 		if (p == cpu)
217 			continue;
218 		if (BT_TEST(kcpc_cpumap, p->cpu_id)) {
219 			mutex_exit(&cpu_lock);
220 			return (-1);
221 		}
222 	}
223 
224 	mutex_exit(&cpu_lock);
225 	return (0);
226 }
227 
228 /*
229  * Called by the generic framework to check if it's OK to bind a set to an LWP.
230  */
231 int
232 kcpc_hw_lwp_hook(void)
233 {
234 	pg_cmt_t	*chip;
235 	group_t		*chips;
236 	group_iter_t	i;
237 
238 	if ((x86_feature & X86_HTT) == 0)
239 		return (0);
240 
241 	/*
242 	 * Only one CPU per chip may be online.
243 	 */
244 	mutex_enter(&cpu_lock);
245 
246 	chips = pghw_set_lookup(PGHW_CHIP);
247 	if (chips == NULL) {
248 		mutex_exit(&cpu_lock);
249 		return (0);
250 	}
251 
252 	group_iter_init(&i);
253 	while ((chip = group_iterate(chips, &i)) != NULL) {
254 		if (GROUP_SIZE(&chip->cmt_cpus_actv) > 1) {
255 			mutex_exit(&cpu_lock);
256 			return (-1);
257 		}
258 	}
259 
260 	mutex_exit(&cpu_lock);
261 	return (0);
262 }
263