xref: /titanic_50/usr/src/uts/sun4u/os/mach_startup.c (revision ff3124eff995e6cd8ebd8c6543648e0670920034)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 /*
22  * Copyright 2008 Sun Microsystems, Inc.  All rights reserved.
23  * Use is subject to license terms.
24  */
25 
26 #pragma ident	"%Z%%M%	%I%	%E% SMI"
27 
28 #include <sys/machsystm.h>
29 #include <sys/archsystm.h>
30 #include <sys/vm.h>
31 #include <sys/cpu.h>
32 #include <sys/cpupart.h>
33 #include <sys/atomic.h>
34 #include <sys/reboot.h>
35 #include <sys/kdi.h>
36 #include <sys/bootconf.h>
37 #include <sys/memlist_plat.h>
38 #include <sys/memlist_impl.h>
39 #include <sys/prom_plat.h>
40 #include <sys/prom_isa.h>
41 #include <sys/autoconf.h>
42 #include <sys/intreg.h>
43 #include <sys/ivintr.h>
44 #include <sys/fpu/fpusystm.h>
45 #include <sys/iommutsb.h>
46 #include <vm/vm_dep.h>
47 #include <vm/seg_kmem.h>
48 #include <vm/seg_kpm.h>
49 #include <vm/seg_map.h>
50 #include <vm/seg_kp.h>
51 #include <sys/sysconf.h>
52 #include <vm/hat_sfmmu.h>
53 #include <sys/kobj.h>
54 #include <sys/sun4asi.h>
55 #include <sys/clconf.h>
56 #include <sys/platform_module.h>
57 #include <sys/panic.h>
58 #include <sys/cpu_sgnblk_defs.h>
59 #include <sys/clock.h>
60 #include <sys/fpras_impl.h>
61 #include <sys/prom_debug.h>
62 #include <sys/traptrace.h>
63 #include <sys/memnode.h>
64 #include <sys/mem_cage.h>
65 
66 /*
67  * fpRAS implementation structures.
68  */
69 struct fpras_chkfn *fpras_chkfnaddrs[FPRAS_NCOPYOPS];
70 struct fpras_chkfngrp *fpras_chkfngrps;
71 struct fpras_chkfngrp *fpras_chkfngrps_base;
72 int fpras_frequency = -1;
73 int64_t fpras_interval = -1;
74 
75 /*
76  * Halt idling cpus optimization
77  *
78  * This optimation is only enabled in platforms that have
79  * the CPU halt support. The cpu_halt_cpu() support is provided
80  * in the cpu module and it is referenced here with a pragma weak.
81  * The presence of this routine automatically enable the halt idling
82  * cpus functionality if the global switch enable_halt_idle_cpus
83  * is set (default is set).
84  *
85  */
86 #pragma weak	cpu_halt_cpu
87 extern void	cpu_halt_cpu();
88 
89 /*
90  * Defines for the idle_state_transition DTrace probe
91  *
92  * The probe fires when the CPU undergoes an idle state change (e.g. halting)
93  * The agument passed is the state to which the CPU is transitioning.
94  *
95  * The states are defined here.
96  */
97 #define	IDLE_STATE_NORMAL 0
98 #define	IDLE_STATE_HALTED 1
99 
100 int		enable_halt_idle_cpus = 1; /* global switch */
101 
102 void
103 setup_trap_table(void)
104 {
105 	intr_init(CPU);			/* init interrupt request free list */
106 	setwstate(WSTATE_KERN);
107 	prom_set_traptable(&trap_table);
108 }
109 
110 void
111 mach_fpras()
112 {
113 	if (fpras_implemented && !fpras_disable) {
114 		int i;
115 		struct fpras_chkfngrp *fcgp;
116 		size_t chkfngrpsallocsz;
117 
118 		/*
119 		 * Note that we size off of NCPU and setup for
120 		 * all those possibilities regardless of whether
121 		 * the cpu id is present or not.  We do this so that
122 		 * we don't have any construction or destruction
123 		 * activity to perform at DR time, and it's not
124 		 * costly in memory.  We require block alignment.
125 		 */
126 		chkfngrpsallocsz = NCPU * sizeof (struct fpras_chkfngrp);
127 		fpras_chkfngrps_base = kmem_alloc(chkfngrpsallocsz, KM_SLEEP);
128 		if (IS_P2ALIGNED((uintptr_t)fpras_chkfngrps_base, 64)) {
129 			fpras_chkfngrps = fpras_chkfngrps_base;
130 		} else {
131 			kmem_free(fpras_chkfngrps_base, chkfngrpsallocsz);
132 			chkfngrpsallocsz += 64;
133 			fpras_chkfngrps_base = kmem_alloc(chkfngrpsallocsz,
134 			    KM_SLEEP);
135 			fpras_chkfngrps = (struct fpras_chkfngrp *)
136 			    P2ROUNDUP((uintptr_t)fpras_chkfngrps_base, 64);
137 		}
138 
139 		/*
140 		 * Copy our check function into place for each copy operation
141 		 * and each cpu id.
142 		 */
143 		fcgp = &fpras_chkfngrps[0];
144 		for (i = 0; i < FPRAS_NCOPYOPS; ++i)
145 			bcopy((void *)fpras_chkfn_type1, &fcgp->fpras_fn[i],
146 			    sizeof (struct fpras_chkfn));
147 		for (i = 1; i < NCPU; ++i)
148 			*(&fpras_chkfngrps[i]) = *fcgp;
149 
150 		/*
151 		 * At definition fpras_frequency is set to -1, and it will
152 		 * still have that value unless changed in /etc/system (not
153 		 * strictly supported, but not preventable).  The following
154 		 * both sets the default and sanity checks anything from
155 		 * /etc/system.
156 		 */
157 		if (fpras_frequency < 0)
158 			fpras_frequency = FPRAS_DEFAULT_FREQUENCY;
159 
160 		/*
161 		 * Now calculate fpras_interval.  When fpras_interval
162 		 * becomes non-negative fpras checks will commence
163 		 * (copies before this point in boot will bypass fpras).
164 		 * Our stores of instructions must be visible; no need
165 		 * to flush as they're never been executed before.
166 		 */
167 		membar_producer();
168 		fpras_interval = (fpras_frequency == 0) ?
169 		    0 : sys_tick_freq / fpras_frequency;
170 	}
171 }
172 
173 void
174 mach_hw_copy_limit(void)
175 {
176 	if (!fpu_exists) {
177 		use_hw_bcopy = 0;
178 		hw_copy_limit_1 = 0;
179 		hw_copy_limit_2 = 0;
180 		hw_copy_limit_4 = 0;
181 		hw_copy_limit_8 = 0;
182 		use_hw_bzero = 0;
183 	}
184 }
185 
186 void
187 load_tod_module()
188 {
189 	/*
190 	 * Load tod driver module for the tod part found on this system.
191 	 * Recompute the cpu frequency/delays based on tod as tod part
192 	 * tends to keep time more accurately.
193 	 */
194 	if (tod_module_name == NULL || modload("tod", tod_module_name) == -1)
195 		halt("Can't load tod module");
196 }
197 
198 void
199 mach_memscrub(void)
200 {
201 	/*
202 	 * Startup memory scrubber, if not running fpu emulation code.
203 	 */
204 
205 #ifndef _HW_MEMSCRUB_SUPPORT
206 	if (fpu_exists) {
207 		if (memscrub_init()) {
208 			cmn_err(CE_WARN,
209 			    "Memory scrubber failed to initialize");
210 		}
211 	}
212 #endif /* _HW_MEMSCRUB_SUPPORT */
213 }
214 
215 /*
216  * Halt the calling CPU until awoken via an interrupt
217  * This routine should only be invoked if cpu_halt_cpu()
218  * exists and is supported, see mach_cpu_halt_idle()
219  */
220 static void
221 cpu_halt(void)
222 {
223 	cpu_t		*cpup = CPU;
224 	processorid_t	cpun = cpup->cpu_id;
225 	cpupart_t	*cp = cpup->cpu_part;
226 	int		hset_update = 1;
227 	uint_t		pstate;
228 	extern uint_t	getpstate(void);
229 	extern void	setpstate(uint_t);
230 
231 	/*
232 	 * If this CPU is online, and there's multiple CPUs
233 	 * in the system, then we should notate our halting
234 	 * by adding ourselves to the partition's halted CPU
235 	 * bitmap. This allows other CPUs to find/awaken us when
236 	 * work becomes available.
237 	 */
238 	if (CPU->cpu_flags & CPU_OFFLINE || ncpus == 1)
239 		hset_update = 0;
240 
241 	/*
242 	 * Add ourselves to the partition's halted CPUs bitmask
243 	 * and set our HALTED flag, if necessary.
244 	 *
245 	 * When a thread becomes runnable, it is placed on the queue
246 	 * and then the halted cpuset is checked to determine who
247 	 * (if anyone) should be awoken. We therefore need to first
248 	 * add ourselves to the halted cpuset, and then check if there
249 	 * is any work available.
250 	 */
251 	if (hset_update) {
252 		cpup->cpu_disp_flags |= CPU_DISP_HALTED;
253 		membar_producer();
254 		CPUSET_ATOMIC_ADD(cp->cp_mach->mc_haltset, cpun);
255 	}
256 
257 	/*
258 	 * Check to make sure there's really nothing to do.
259 	 * Work destined for this CPU may become available after
260 	 * this check. We'll be notified through the clearing of our
261 	 * bit in the halted CPU bitmask, and a poke.
262 	 */
263 	if (disp_anywork()) {
264 		if (hset_update) {
265 			cpup->cpu_disp_flags &= ~CPU_DISP_HALTED;
266 			CPUSET_ATOMIC_DEL(cp->cp_mach->mc_haltset, cpun);
267 		}
268 		return;
269 	}
270 
271 	/*
272 	 * We're on our way to being halted.
273 	 *
274 	 * Disable interrupts now, so that we'll awaken immediately
275 	 * after halting if someone tries to poke us between now and
276 	 * the time we actually halt.
277 	 *
278 	 * We check for the presence of our bit after disabling interrupts.
279 	 * If it's cleared, we'll return. If the bit is cleared after
280 	 * we check then the poke will pop us out of the halted state.
281 	 *
282 	 * The ordering of the poke and the clearing of the bit by cpu_wakeup
283 	 * is important.
284 	 * cpu_wakeup() must clear, then poke.
285 	 * cpu_halt() must disable interrupts, then check for the bit.
286 	 */
287 	pstate = getpstate();
288 	setpstate(pstate & ~PSTATE_IE);
289 
290 	if (hset_update && !CPU_IN_SET(cp->cp_mach->mc_haltset, cpun)) {
291 		cpup->cpu_disp_flags &= ~CPU_DISP_HALTED;
292 		setpstate(pstate);
293 		return;
294 	}
295 
296 	/*
297 	 * The check for anything locally runnable is here for performance
298 	 * and isn't needed for correctness. disp_nrunnable ought to be
299 	 * in our cache still, so it's inexpensive to check, and if there
300 	 * is anything runnable we won't have to wait for the poke.
301 	 */
302 	if (cpup->cpu_disp->disp_nrunnable != 0) {
303 		if (hset_update) {
304 			cpup->cpu_disp_flags &= ~CPU_DISP_HALTED;
305 			CPUSET_ATOMIC_DEL(cp->cp_mach->mc_haltset, cpun);
306 		}
307 		setpstate(pstate);
308 		return;
309 	}
310 
311 	/*
312 	 * Halt the strand.
313 	 */
314 	if (&cpu_halt_cpu) {
315 		DTRACE_PROBE1(idle__state__transition,
316 		    uint_t, IDLE_STATE_HALTED);
317 
318 		cpu_halt_cpu();
319 
320 		DTRACE_PROBE1(idle__state__transition,
321 		    uint_t, IDLE_STATE_NORMAL);
322 	}
323 
324 	/*
325 	 * We're no longer halted
326 	 */
327 	setpstate(pstate);
328 	if (hset_update) {
329 		cpup->cpu_disp_flags &= ~CPU_DISP_HALTED;
330 		CPUSET_ATOMIC_DEL(cp->cp_mach->mc_haltset, cpun);
331 	}
332 }
333 
334 /*
335  * If "cpu" is halted, then wake it up clearing its halted bit in advance.
336  * Otherwise, see if other CPUs in the cpu partition are halted and need to
337  * be woken up so that they can steal the thread we placed on this CPU.
338  * This function is only used on MP systems.
339  * This function should only be invoked if cpu_halt_cpu()
340  * exists and is supported, see mach_cpu_halt_idle()
341  */
342 static void
343 cpu_wakeup(cpu_t *cpu, int bound)
344 {
345 	uint_t		cpu_found;
346 	int		result;
347 	cpupart_t	*cp;
348 
349 	cp = cpu->cpu_part;
350 	if (CPU_IN_SET(cp->cp_mach->mc_haltset, cpu->cpu_id)) {
351 		/*
352 		 * Clear the halted bit for that CPU since it will be
353 		 * poked in a moment.
354 		 */
355 		CPUSET_ATOMIC_DEL(cp->cp_mach->mc_haltset, cpu->cpu_id);
356 		/*
357 		 * We may find the current CPU present in the halted cpuset
358 		 * if we're in the context of an interrupt that occurred
359 		 * before we had a chance to clear our bit in cpu_halt().
360 		 * Poking ourself is obviously unnecessary, since if
361 		 * we're here, we're not halted.
362 		 */
363 		if (cpu != CPU)
364 			poke_cpu(cpu->cpu_id);
365 		return;
366 	} else {
367 		/*
368 		 * This cpu isn't halted, but it's idle or undergoing a
369 		 * context switch. No need to awaken anyone else.
370 		 */
371 		if (cpu->cpu_thread == cpu->cpu_idle_thread ||
372 		    cpu->cpu_disp_flags & CPU_DISP_DONTSTEAL)
373 			return;
374 	}
375 
376 	/*
377 	 * No need to wake up other CPUs if the thread we just enqueued
378 	 * is bound.
379 	 */
380 	if (bound)
381 		return;
382 
383 	/*
384 	 * See if there's any other halted CPUs. If there are, then
385 	 * select one, and awaken it.
386 	 * It's possible that after we find a CPU, somebody else
387 	 * will awaken it before we get the chance.
388 	 * In that case, look again.
389 	 */
390 	do {
391 		CPUSET_FIND(cp->cp_mach->mc_haltset, cpu_found);
392 		if (cpu_found == CPUSET_NOTINSET)
393 			return;
394 
395 		ASSERT(cpu_found >= 0 && cpu_found < NCPU);
396 		CPUSET_ATOMIC_XDEL(cp->cp_mach->mc_haltset, cpu_found, result);
397 	} while (result < 0);
398 
399 	if (cpu_found != CPU->cpu_id)
400 		poke_cpu(cpu_found);
401 }
402 
403 void
404 mach_cpu_halt_idle()
405 {
406 	if (enable_halt_idle_cpus) {
407 		if (&cpu_halt_cpu) {
408 			idle_cpu = cpu_halt;
409 			disp_enq_thread = cpu_wakeup;
410 		}
411 	}
412 }
413 
414 /*ARGSUSED*/
415 int
416 cpu_intrq_setup(struct cpu *cp)
417 {
418 	/* Interrupt mondo queues not applicable to sun4u */
419 	return (0);
420 }
421 
422 /*ARGSUSED*/
423 void
424 cpu_intrq_cleanup(struct cpu *cp)
425 {
426 	/* Interrupt mondo queues not applicable to sun4u */
427 }
428 
429 /*ARGSUSED*/
430 void
431 cpu_intrq_register(struct cpu *cp)
432 {
433 	/* Interrupt/error queues not applicable to sun4u */
434 }
435 
436 /*ARGSUSED*/
437 void
438 mach_htraptrace_setup(int cpuid)
439 {
440 	/* Setup hypervisor traptrace buffer, not applicable to sun4u */
441 }
442 
443 /*ARGSUSED*/
444 void
445 mach_htraptrace_configure(int cpuid)
446 {
447 	/* enable/ disable hypervisor traptracing, not applicable to sun4u */
448 }
449 
450 /*ARGSUSED*/
451 void
452 mach_htraptrace_cleanup(int cpuid)
453 {
454 	/* cleanup hypervisor traptrace buffer, not applicable to sun4u */
455 }
456 
457 void
458 mach_descrip_startup_init(void)
459 {
460 	/*
461 	 * Only for sun4v.
462 	 * Initialize Machine description framework during startup.
463 	 */
464 }
465 void
466 mach_descrip_startup_fini(void)
467 {
468 	/*
469 	 * Only for sun4v.
470 	 * Clean up Machine Description framework during startup.
471 	 */
472 }
473 
474 void
475 mach_descrip_init(void)
476 {
477 	/*
478 	 * Only for sun4v.
479 	 * Initialize Machine description framework.
480 	 */
481 }
482 
483 void
484 hsvc_setup(void)
485 {
486 	/* Setup hypervisor services, not applicable to sun4u */
487 }
488 
489 void
490 load_mach_drivers(void)
491 {
492 	/* Currently no machine class (sun4u) specific drivers to load */
493 }
494 
495 /*
496  * Return true if the machine we're running on is a Positron.
497  * (Positron is an unsupported developers platform.)
498  */
499 int
500 iam_positron(void)
501 {
502 	char model[32];
503 	const char proto_model[] = "SUNW,501-2732";
504 	pnode_t root = prom_rootnode();
505 
506 	if (prom_getproplen(root, "model") != sizeof (proto_model))
507 		return (0);
508 
509 	(void) prom_getprop(root, "model", model);
510 	if (strcmp(model, proto_model) == 0)
511 		return (1);
512 	return (0);
513 }
514 
515 /*
516  * Find a physically contiguous area of twice the largest ecache size
517  * to be used while doing displacement flush of ecaches.
518  */
519 uint64_t
520 ecache_flush_address(void)
521 {
522 	struct memlist *pmem;
523 	uint64_t flush_size;
524 	uint64_t ret_val;
525 
526 	flush_size = ecache_size * 2;
527 	for (pmem = phys_install; pmem; pmem = pmem->next) {
528 		ret_val = P2ROUNDUP(pmem->address, ecache_size);
529 		if (ret_val + flush_size <= pmem->address + pmem->size)
530 			return (ret_val);
531 	}
532 	return ((uint64_t)-1);
533 }
534 
535 /*
536  * Called with the memlist lock held to say that phys_install has
537  * changed.
538  */
539 void
540 phys_install_has_changed(void)
541 {
542 	/*
543 	 * Get the new address into a temporary just in case panicking
544 	 * involves use of ecache_flushaddr.
545 	 */
546 	uint64_t new_addr;
547 
548 	new_addr = ecache_flush_address();
549 	if (new_addr == (uint64_t)-1) {
550 		cmn_err(CE_PANIC,
551 		    "ecache_flush_address(): failed, ecache_size=%x",
552 		    ecache_size);
553 		/*NOTREACHED*/
554 	}
555 	ecache_flushaddr = new_addr;
556 	membar_producer();
557 }
558