xref: /titanic_50/usr/src/uts/sun4u/os/mach_startup.c (revision 7e749adb01c386454877ad532459e92a9f19a702)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 /*
22  * Copyright 2010 Sun Microsystems, Inc.  All rights reserved.
23  * Use is subject to license terms.
24  */
25 
26 #include <sys/machsystm.h>
27 #include <sys/archsystm.h>
28 #include <sys/vm.h>
29 #include <sys/cpu.h>
30 #include <sys/cpupart.h>
31 #include <sys/cmt.h>
32 #include <sys/bitset.h>
33 #include <sys/reboot.h>
34 #include <sys/kdi.h>
35 #include <sys/bootconf.h>
36 #include <sys/memlist_plat.h>
37 #include <sys/memlist_impl.h>
38 #include <sys/prom_plat.h>
39 #include <sys/prom_isa.h>
40 #include <sys/autoconf.h>
41 #include <sys/intreg.h>
42 #include <sys/ivintr.h>
43 #include <sys/fpu/fpusystm.h>
44 #include <sys/iommutsb.h>
45 #include <vm/vm_dep.h>
46 #include <vm/seg_kmem.h>
47 #include <vm/seg_kpm.h>
48 #include <vm/seg_map.h>
49 #include <vm/seg_kp.h>
50 #include <sys/sysconf.h>
51 #include <vm/hat_sfmmu.h>
52 #include <sys/kobj.h>
53 #include <sys/sun4asi.h>
54 #include <sys/clconf.h>
55 #include <sys/platform_module.h>
56 #include <sys/panic.h>
57 #include <sys/cpu_sgnblk_defs.h>
58 #include <sys/clock.h>
59 #include <sys/fpras_impl.h>
60 #include <sys/prom_debug.h>
61 #include <sys/traptrace.h>
62 #include <sys/memnode.h>
63 #include <sys/mem_cage.h>
64 
65 /*
66  * fpRAS implementation structures.
67  */
68 struct fpras_chkfn *fpras_chkfnaddrs[FPRAS_NCOPYOPS];
69 struct fpras_chkfngrp *fpras_chkfngrps;
70 struct fpras_chkfngrp *fpras_chkfngrps_base;
71 int fpras_frequency = -1;
72 int64_t fpras_interval = -1;
73 
74 /*
75  * Increase unix symbol table size as a work around for 6828121
76  */
77 int alloc_mem_bermuda_triangle;
78 
79 /*
80  * Halt idling cpus optimization
81  *
82  * This optimation is only enabled in platforms that have
83  * the CPU halt support. The cpu_halt_cpu() support is provided
84  * in the cpu module and it is referenced here with a pragma weak.
85  * The presence of this routine automatically enable the halt idling
86  * cpus functionality if the global switch enable_halt_idle_cpus
87  * is set (default is set).
88  *
89  */
90 #pragma weak	cpu_halt_cpu
91 extern void	cpu_halt_cpu();
92 
93 /*
94  * Defines for the idle_state_transition DTrace probe
95  *
96  * The probe fires when the CPU undergoes an idle state change (e.g. halting)
97  * The agument passed is the state to which the CPU is transitioning.
98  *
99  * The states are defined here.
100  */
101 #define	IDLE_STATE_NORMAL 0
102 #define	IDLE_STATE_HALTED 1
103 
104 int		enable_halt_idle_cpus = 1; /* global switch */
105 
106 void
107 setup_trap_table(void)
108 {
109 	intr_init(CPU);			/* init interrupt request free list */
110 	setwstate(WSTATE_KERN);
111 	prom_set_traptable(&trap_table);
112 }
113 
114 void
115 mach_fpras()
116 {
117 	if (fpras_implemented && !fpras_disable) {
118 		int i;
119 		struct fpras_chkfngrp *fcgp;
120 		size_t chkfngrpsallocsz;
121 
122 		/*
123 		 * Note that we size off of NCPU and setup for
124 		 * all those possibilities regardless of whether
125 		 * the cpu id is present or not.  We do this so that
126 		 * we don't have any construction or destruction
127 		 * activity to perform at DR time, and it's not
128 		 * costly in memory.  We require block alignment.
129 		 */
130 		chkfngrpsallocsz = NCPU * sizeof (struct fpras_chkfngrp);
131 		fpras_chkfngrps_base = kmem_alloc(chkfngrpsallocsz, KM_SLEEP);
132 		if (IS_P2ALIGNED((uintptr_t)fpras_chkfngrps_base, 64)) {
133 			fpras_chkfngrps = fpras_chkfngrps_base;
134 		} else {
135 			kmem_free(fpras_chkfngrps_base, chkfngrpsallocsz);
136 			chkfngrpsallocsz += 64;
137 			fpras_chkfngrps_base = kmem_alloc(chkfngrpsallocsz,
138 			    KM_SLEEP);
139 			fpras_chkfngrps = (struct fpras_chkfngrp *)
140 			    P2ROUNDUP((uintptr_t)fpras_chkfngrps_base, 64);
141 		}
142 
143 		/*
144 		 * Copy our check function into place for each copy operation
145 		 * and each cpu id.
146 		 */
147 		fcgp = &fpras_chkfngrps[0];
148 		for (i = 0; i < FPRAS_NCOPYOPS; ++i)
149 			bcopy((void *)fpras_chkfn_type1, &fcgp->fpras_fn[i],
150 			    sizeof (struct fpras_chkfn));
151 		for (i = 1; i < NCPU; ++i)
152 			*(&fpras_chkfngrps[i]) = *fcgp;
153 
154 		/*
155 		 * At definition fpras_frequency is set to -1, and it will
156 		 * still have that value unless changed in /etc/system (not
157 		 * strictly supported, but not preventable).  The following
158 		 * both sets the default and sanity checks anything from
159 		 * /etc/system.
160 		 */
161 		if (fpras_frequency < 0)
162 			fpras_frequency = FPRAS_DEFAULT_FREQUENCY;
163 
164 		/*
165 		 * Now calculate fpras_interval.  When fpras_interval
166 		 * becomes non-negative fpras checks will commence
167 		 * (copies before this point in boot will bypass fpras).
168 		 * Our stores of instructions must be visible; no need
169 		 * to flush as they're never been executed before.
170 		 */
171 		membar_producer();
172 		fpras_interval = (fpras_frequency == 0) ?
173 		    0 : sys_tick_freq / fpras_frequency;
174 	}
175 }
176 
177 void
178 mach_hw_copy_limit(void)
179 {
180 	if (!fpu_exists) {
181 		use_hw_bcopy = 0;
182 		hw_copy_limit_1 = 0;
183 		hw_copy_limit_2 = 0;
184 		hw_copy_limit_4 = 0;
185 		hw_copy_limit_8 = 0;
186 		use_hw_bzero = 0;
187 	}
188 }
189 
190 void
191 load_tod_module()
192 {
193 	/*
194 	 * Load tod driver module for the tod part found on this system.
195 	 * Recompute the cpu frequency/delays based on tod as tod part
196 	 * tends to keep time more accurately.
197 	 */
198 	if (tod_module_name == NULL || modload("tod", tod_module_name) == -1)
199 		halt("Can't load tod module");
200 }
201 
202 void
203 mach_memscrub(void)
204 {
205 	/*
206 	 * Startup memory scrubber, if not running fpu emulation code.
207 	 */
208 
209 #ifndef _HW_MEMSCRUB_SUPPORT
210 	if (fpu_exists) {
211 		if (memscrub_init()) {
212 			cmn_err(CE_WARN,
213 			    "Memory scrubber failed to initialize");
214 		}
215 	}
216 #endif /* _HW_MEMSCRUB_SUPPORT */
217 }
218 
219 /*
220  * Halt the present CPU until awoken via an interrupt.
221  * This routine should only be invoked if cpu_halt_cpu()
222  * exists and is supported, see mach_cpu_halt_idle()
223  */
224 void
225 cpu_halt(void)
226 {
227 	cpu_t *cpup = CPU;
228 	processorid_t cpu_sid = cpup->cpu_seqid;
229 	cpupart_t *cp = cpup->cpu_part;
230 	int hset_update = 1;
231 	volatile int *p = &cpup->cpu_disp->disp_nrunnable;
232 	uint_t s;
233 
234 	/*
235 	 * If this CPU is online then we should notate our halting
236 	 * by adding ourselves to the partition's halted CPU
237 	 * bitset. This allows other CPUs to find/awaken us when
238 	 * work becomes available.
239 	 */
240 	if (CPU->cpu_flags & CPU_OFFLINE)
241 		hset_update = 0;
242 
243 	/*
244 	 * Add ourselves to the partition's halted CPUs bitset
245 	 * and set our HALTED flag, if necessary.
246 	 *
247 	 * When a thread becomes runnable, it is placed on the queue
248 	 * and then the halted cpu bitset is checked to determine who
249 	 * (if anyone) should be awoken. We therefore need to first
250 	 * add ourselves to the halted bitset, and then check if there
251 	 * is any work available.  The order is important to prevent a race
252 	 * that can lead to work languishing on a run queue somewhere while
253 	 * this CPU remains halted.
254 	 *
255 	 * Either the producing CPU will see we're halted and will awaken us,
256 	 * or this CPU will see the work available in disp_anywork()
257 	 */
258 	if (hset_update) {
259 		cpup->cpu_disp_flags |= CPU_DISP_HALTED;
260 		membar_producer();
261 		bitset_atomic_add(&cp->cp_haltset, cpu_sid);
262 	}
263 
264 	/*
265 	 * Check to make sure there's really nothing to do.
266 	 * Work destined for this CPU may become available after
267 	 * this check. We'll be notified through the clearing of our
268 	 * bit in the halted CPU bitset, and a poke.
269 	 */
270 	if (disp_anywork()) {
271 		if (hset_update) {
272 			cpup->cpu_disp_flags &= ~CPU_DISP_HALTED;
273 			bitset_atomic_del(&cp->cp_haltset, cpu_sid);
274 		}
275 		return;
276 	}
277 
278 	/*
279 	 * We're on our way to being halted.  Wait until something becomes
280 	 * runnable locally or we are awaken (i.e. removed from the halt set).
281 	 * Note that the call to hv_cpu_yield() can return even if we have
282 	 * nothing to do.
283 	 *
284 	 * Disable interrupts now, so that we'll awaken immediately
285 	 * after halting if someone tries to poke us between now and
286 	 * the time we actually halt.
287 	 *
288 	 * We check for the presence of our bit after disabling interrupts.
289 	 * If it's cleared, we'll return. If the bit is cleared after
290 	 * we check then the poke will pop us out of the halted state.
291 	 * Also, if the offlined CPU has been brought back on-line, then
292 	 * we return as well.
293 	 *
294 	 * The ordering of the poke and the clearing of the bit by cpu_wakeup
295 	 * is important.
296 	 * cpu_wakeup() must clear, then poke.
297 	 * cpu_halt() must disable interrupts, then check for the bit.
298 	 *
299 	 * The check for anything locally runnable is here for performance
300 	 * and isn't needed for correctness. disp_nrunnable ought to be
301 	 * in our cache still, so it's inexpensive to check, and if there
302 	 * is anything runnable we won't have to wait for the poke.
303 	 *
304 	 * Any interrupt will awaken the cpu from halt. Looping here
305 	 * will filter spurious interrupts that wake us up, but don't
306 	 * represent a need for us to head back out to idle().  This
307 	 * will enable the idle loop to be more efficient and sleep in
308 	 * the processor pipeline for a larger percent of the time,
309 	 * which returns useful cycles to the peer hardware strand
310 	 * that shares the pipeline.
311 	 */
312 	s = disable_vec_intr();
313 	while (*p == 0 &&
314 	    ((hset_update && bitset_in_set(&cp->cp_haltset, cpu_sid)) ||
315 	    (!hset_update && (CPU->cpu_flags & CPU_OFFLINE)))) {
316 
317 		DTRACE_PROBE1(idle__state__transition,
318 		    uint_t, IDLE_STATE_HALTED);
319 		(void) cpu_halt_cpu();
320 		DTRACE_PROBE1(idle__state__transition,
321 		    uint_t, IDLE_STATE_NORMAL);
322 
323 		enable_vec_intr(s);
324 		s = disable_vec_intr();
325 	}
326 
327 	/*
328 	 * We're no longer halted
329 	 */
330 	enable_vec_intr(s);
331 	if (hset_update) {
332 		cpup->cpu_disp_flags &= ~CPU_DISP_HALTED;
333 		bitset_atomic_del(&cp->cp_haltset, cpu_sid);
334 	}
335 }
336 
337 /*
338  * If "cpu" is halted, then wake it up clearing its halted bit in advance.
339  * Otherwise, see if other CPUs in the cpu partition are halted and need to
340  * be woken up so that they can steal the thread we placed on this CPU.
341  * This function is only used on MP systems.
342  * This function should only be invoked if cpu_halt_cpu()
343  * exists and is supported, see mach_cpu_halt_idle()
344  */
345 static void
346 cpu_wakeup(cpu_t *cpu, int bound)
347 {
348 	uint_t		cpu_found;
349 	processorid_t	cpu_sid;
350 	cpupart_t	*cp;
351 
352 	cp = cpu->cpu_part;
353 	cpu_sid = cpu->cpu_seqid;
354 	if (bitset_in_set(&cp->cp_haltset, cpu_sid)) {
355 		/*
356 		 * Clear the halted bit for that CPU since it will be
357 		 * poked in a moment.
358 		 */
359 		bitset_atomic_del(&cp->cp_haltset, cpu_sid);
360 		/*
361 		 * We may find the current CPU present in the halted cpu bitset
362 		 * if we're in the context of an interrupt that occurred
363 		 * before we had a chance to clear our bit in cpu_halt().
364 		 * Poking ourself is obviously unnecessary, since if
365 		 * we're here, we're not halted.
366 		 */
367 		if (cpu != CPU)
368 			poke_cpu(cpu->cpu_id);
369 		return;
370 	} else {
371 		/*
372 		 * This cpu isn't halted, but it's idle or undergoing a
373 		 * context switch. No need to awaken anyone else.
374 		 */
375 		if (cpu->cpu_thread == cpu->cpu_idle_thread ||
376 		    cpu->cpu_disp_flags & CPU_DISP_DONTSTEAL)
377 			return;
378 	}
379 
380 	/*
381 	 * No need to wake up other CPUs if this is for a bound thread.
382 	 */
383 	if (bound)
384 		return;
385 
386 	/*
387 	 * The CPU specified for wakeup isn't currently halted, so check
388 	 * to see if there are any other halted CPUs in the partition,
389 	 * and if there are then awaken one.
390 	 *
391 	 * If possible, try to select a CPU close to the target, since this
392 	 * will likely trigger a migration.
393 	 */
394 	do {
395 		cpu_found = bitset_find(&cp->cp_haltset);
396 		if (cpu_found == (uint_t)-1)
397 			return;
398 	} while (bitset_atomic_test_and_del(&cp->cp_haltset, cpu_found) < 0);
399 
400 	if (cpu_found != CPU->cpu_seqid)
401 		poke_cpu(cpu_seq[cpu_found]->cpu_id);
402 }
403 
404 void
405 mach_cpu_halt_idle(void)
406 {
407 	if (enable_halt_idle_cpus) {
408 		if (&cpu_halt_cpu) {
409 			idle_cpu = cpu_halt;
410 			disp_enq_thread = cpu_wakeup;
411 		}
412 	}
413 }
414 
415 /*ARGSUSED*/
416 int
417 cpu_intrq_setup(struct cpu *cp)
418 {
419 	/* Interrupt mondo queues not applicable to sun4u */
420 	return (0);
421 }
422 
423 /*ARGSUSED*/
424 void
425 cpu_intrq_cleanup(struct cpu *cp)
426 {
427 	/* Interrupt mondo queues not applicable to sun4u */
428 }
429 
430 /*ARGSUSED*/
431 void
432 cpu_intrq_register(struct cpu *cp)
433 {
434 	/* Interrupt/error queues not applicable to sun4u */
435 }
436 
437 /*ARGSUSED*/
438 void
439 mach_htraptrace_setup(int cpuid)
440 {
441 	/* Setup hypervisor traptrace buffer, not applicable to sun4u */
442 }
443 
444 /*ARGSUSED*/
445 void
446 mach_htraptrace_configure(int cpuid)
447 {
448 	/* enable/ disable hypervisor traptracing, not applicable to sun4u */
449 }
450 
451 /*ARGSUSED*/
452 void
453 mach_htraptrace_cleanup(int cpuid)
454 {
455 	/* cleanup hypervisor traptrace buffer, not applicable to sun4u */
456 }
457 
458 void
459 mach_descrip_startup_init(void)
460 {
461 	/*
462 	 * Only for sun4v.
463 	 * Initialize Machine description framework during startup.
464 	 */
465 }
466 void
467 mach_descrip_startup_fini(void)
468 {
469 	/*
470 	 * Only for sun4v.
471 	 * Clean up Machine Description framework during startup.
472 	 */
473 }
474 
475 void
476 mach_descrip_init(void)
477 {
478 	/*
479 	 * Only for sun4v.
480 	 * Initialize Machine description framework.
481 	 */
482 }
483 
484 void
485 hsvc_setup(void)
486 {
487 	/* Setup hypervisor services, not applicable to sun4u */
488 }
489 
490 void
491 load_mach_drivers(void)
492 {
493 	/* Currently no machine class (sun4u) specific drivers to load */
494 }
495 
496 /*
497  * Return true if the machine we're running on is a Positron.
498  * (Positron is an unsupported developers platform.)
499  */
500 int
501 iam_positron(void)
502 {
503 	char model[32];
504 	const char proto_model[] = "SUNW,501-2732";
505 	pnode_t root = prom_rootnode();
506 
507 	if (prom_getproplen(root, "model") != sizeof (proto_model))
508 		return (0);
509 
510 	(void) prom_getprop(root, "model", model);
511 	if (strcmp(model, proto_model) == 0)
512 		return (1);
513 	return (0);
514 }
515 
516 /*
517  * Find a physically contiguous area of twice the largest ecache size
518  * to be used while doing displacement flush of ecaches.
519  */
520 uint64_t
521 ecache_flush_address(void)
522 {
523 	struct memlist *pmem;
524 	uint64_t flush_size;
525 	uint64_t ret_val;
526 
527 	flush_size = ecache_size * 2;
528 	for (pmem = phys_install; pmem; pmem = pmem->ml_next) {
529 		ret_val = P2ROUNDUP(pmem->ml_address, ecache_size);
530 		if (ret_val + flush_size <= pmem->ml_address + pmem->ml_size)
531 			return (ret_val);
532 	}
533 	return ((uint64_t)-1);
534 }
535 
536 /*
537  * Called with the memlist lock held to say that phys_install has
538  * changed.
539  */
540 void
541 phys_install_has_changed(void)
542 {
543 	/*
544 	 * Get the new address into a temporary just in case panicking
545 	 * involves use of ecache_flushaddr.
546 	 */
547 	uint64_t new_addr;
548 
549 	new_addr = ecache_flush_address();
550 	if (new_addr == (uint64_t)-1) {
551 		cmn_err(CE_PANIC,
552 		    "ecache_flush_address(): failed, ecache_size=%x",
553 		    ecache_size);
554 		/*NOTREACHED*/
555 	}
556 	ecache_flushaddr = new_addr;
557 	membar_producer();
558 }
559