xref: /titanic_41/usr/src/uts/common/os/lgrp.c (revision d6e098b6e116b68f7b561b91cc6b9dcce32ec0ec)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 /*
22  * Copyright 2009 Sun Microsystems, Inc.  All rights reserved.
23  * Use is subject to license terms.
24  */
25 
26 /*
27  * Basic NUMA support in terms of locality groups
28  *
29  * Solaris needs to know which CPUs, memory, etc. are near each other to
30  * provide good performance on NUMA machines by optimizing for locality.
31  * In order to do this, a new abstraction called a "locality group (lgroup)"
32  * has been introduced to keep track of which CPU-like and memory-like hardware
33  * resources are close to each other.  Currently, latency is the only measure
34  * used to determine how to group hardware resources into lgroups, but this
35  * does not limit the groupings to be based solely on latency.  Other factors
36  * may be used to determine the groupings in the future.
37  *
38  * Lgroups are organized into a hieararchy or topology that represents the
39  * latency topology of the machine.  There is always at least a root lgroup in
40  * the system.  It represents all the hardware resources in the machine at a
41  * latency big enough that any hardware resource can at least access any other
42  * hardware resource within that latency.  A Uniform Memory Access (UMA)
43  * machine is represented with one lgroup (the root).  In contrast, a NUMA
44  * machine is represented at least by the root lgroup and some number of leaf
45  * lgroups where the leaf lgroups contain the hardware resources within the
46  * least latency of each other and the root lgroup still contains all the
47  * resources in the machine.  Some number of intermediate lgroups may exist
48  * which represent more levels of locality than just the local latency of the
49  * leaf lgroups and the system latency of the root lgroup.  Non-leaf lgroups
50  * (eg. root and intermediate lgroups) contain the next nearest resources to
51  * its children lgroups.  Thus, the lgroup hierarchy from a given leaf lgroup
52  * to the root lgroup shows the hardware resources from closest to farthest
53  * from the leaf lgroup such that each successive ancestor lgroup contains
54  * the next nearest resources at the next level of locality from the previous.
55  *
56  * The kernel uses the lgroup abstraction to know how to allocate resources
57  * near a given process/thread.  At fork() and lwp/thread_create() time, a
58  * "home" lgroup is chosen for a thread.  This is done by picking the lgroup
59  * with the lowest load average.  Binding to a processor or processor set will
60  * change the home lgroup for a thread.  The scheduler has been modified to try
61  * to dispatch a thread on a CPU in its home lgroup.  Physical memory
62  * allocation is lgroup aware too, so memory will be allocated from the current
63  * thread's home lgroup if possible.  If the desired resources are not
64  * available, the kernel traverses the lgroup hierarchy going to the parent
65  * lgroup to find resources at the next level of locality until it reaches the
66  * root lgroup.
67  */
68 
69 #include <sys/lgrp.h>
70 #include <sys/lgrp_user.h>
71 #include <sys/types.h>
72 #include <sys/mman.h>
73 #include <sys/param.h>
74 #include <sys/var.h>
75 #include <sys/thread.h>
76 #include <sys/cpuvar.h>
77 #include <sys/cpupart.h>
78 #include <sys/kmem.h>
79 #include <vm/seg.h>
80 #include <vm/seg_kmem.h>
81 #include <vm/seg_spt.h>
82 #include <vm/seg_vn.h>
83 #include <vm/as.h>
84 #include <sys/atomic.h>
85 #include <sys/systm.h>
86 #include <sys/errno.h>
87 #include <sys/cmn_err.h>
88 #include <sys/kstat.h>
89 #include <sys/sysmacros.h>
90 #include <sys/pg.h>
91 #include <sys/promif.h>
92 #include <sys/sdt.h>
93 
94 lgrp_gen_t	lgrp_gen = 0;		/* generation of lgroup hierarchy */
95 lgrp_t *lgrp_table[NLGRPS_MAX]; /* table of all initialized lgrp_t structs */
96 				/* indexed by lgrp_id */
97 int	nlgrps;			/* number of lgroups in machine */
98 int	lgrp_alloc_hint = -1;	/* hint for where to try to allocate next */
99 int	lgrp_alloc_max = 0;	/* max lgroup ID allocated so far */
100 
101 /*
102  * Kstat data for lgroups.
103  *
104  * Actual kstat data is collected in lgrp_stats array.
105  * The lgrp_kstat_data array of named kstats is used to extract data from
106  * lgrp_stats and present it to kstat framework. It is protected from partallel
107  * modifications by lgrp_kstat_mutex. This may cause some contention when
108  * several kstat commands run in parallel but this is not the
109  * performance-critical path.
110  */
111 extern struct lgrp_stats lgrp_stats[];	/* table of per-lgrp stats */
112 
113 /*
114  * Declare kstat names statically for enums as defined in the header file.
115  */
116 LGRP_KSTAT_NAMES;
117 
118 static void	lgrp_kstat_init(void);
119 static int	lgrp_kstat_extract(kstat_t *, int);
120 static void	lgrp_kstat_reset(lgrp_id_t);
121 
122 static struct kstat_named lgrp_kstat_data[LGRP_NUM_STATS];
123 static kmutex_t lgrp_kstat_mutex;
124 
125 
126 /*
127  * max number of lgroups supported by the platform
128  */
129 int	nlgrpsmax = 0;
130 
131 /*
132  * The root lgroup. Represents the set of resources at the system wide
133  * level of locality.
134  */
135 lgrp_t		*lgrp_root = NULL;
136 
137 /*
138  * During system bootstrap cp_default does not contain the list of lgrp load
139  * averages (cp_lgrploads). The list is allocated after the first CPU is brought
140  * on-line when cp_default is initialized by cpupart_initialize_default().
141  * Configuring CPU0 may create a two-level topology with root and one leaf node
142  * containing CPU0. This topology is initially constructed in a special
143  * statically allocated 2-element lpl list lpl_bootstrap_list and later cloned
144  * to cp_default when cp_default is initialized. The lpl_bootstrap_list is used
145  * for all lpl operations until cp_default is fully constructed.
146  *
147  * The lpl_bootstrap_list is maintained by the code in lgrp.c. Every other
148  * consumer who needs default lpl should use lpl_bootstrap which is a pointer to
149  * the first element of lpl_bootstrap_list.
150  *
151  * CPUs that are added to the system, but have not yet been assigned to an
152  * lgrp will use lpl_bootstrap as a default lpl. This is necessary because
153  * on some architectures (x86) it's possible for the slave CPU startup thread
154  * to enter the dispatcher or allocate memory before calling lgrp_cpu_init().
155  */
156 #define	LPL_BOOTSTRAP_SIZE 2
157 static lpl_t	lpl_bootstrap_list[LPL_BOOTSTRAP_SIZE];
158 lpl_t		*lpl_bootstrap;
159 static lpl_t	*lpl_bootstrap_rset[LPL_BOOTSTRAP_SIZE];
160 static int	lpl_bootstrap_id2rset[LPL_BOOTSTRAP_SIZE];
161 
162 /*
163  * If cp still references the bootstrap lpl, it has not yet been added to
164  * an lgrp. lgrp_mem_choose() uses this macro to detect the case where
165  * a thread is trying to allocate memory close to a CPU that has no lgrp.
166  */
167 #define	LGRP_CPU_HAS_NO_LGRP(cp)	((cp)->cpu_lpl == lpl_bootstrap)
168 
169 static lgrp_t	lroot;
170 
171 /*
172  * Size, in bytes, beyond which random memory allocation policy is applied
173  * to non-shared memory.  Default is the maximum size, so random memory
174  * allocation won't be used for non-shared memory by default.
175  */
176 size_t	lgrp_privm_random_thresh = (size_t)(-1);
177 
178 /* the maximum effect that a single thread can have on it's lgroup's load */
179 #define	LGRP_LOADAVG_MAX_EFFECT(ncpu) \
180 	((lgrp_loadavg_max_effect) / (ncpu))
181 uint32_t	lgrp_loadavg_max_effect = LGRP_LOADAVG_THREAD_MAX;
182 
183 
184 /*
185  * Size, in bytes, beyond which random memory allocation policy is applied to
186  * shared memory.  Default is 8MB (2 ISM pages).
187  */
188 size_t	lgrp_shm_random_thresh = 8*1024*1024;
189 
190 /*
191  * Whether to do processor set aware memory allocation by default
192  */
193 int	lgrp_mem_pset_aware = 0;
194 
195 /*
196  * Set the default memory allocation policy for root lgroup
197  */
198 lgrp_mem_policy_t	lgrp_mem_policy_root = LGRP_MEM_POLICY_RANDOM;
199 
200 /*
201  * Set the default memory allocation policy.  For most platforms,
202  * next touch is sufficient, but some platforms may wish to override
203  * this.
204  */
205 lgrp_mem_policy_t	lgrp_mem_default_policy = LGRP_MEM_POLICY_NEXT;
206 
207 
208 /*
209  * lgroup CPU event handlers
210  */
211 static void	lgrp_cpu_init(struct cpu *);
212 static void	lgrp_cpu_fini(struct cpu *, lgrp_id_t);
213 static lgrp_t	*lgrp_cpu_to_lgrp(struct cpu *);
214 
215 /*
216  * lgroup memory event handlers
217  */
218 static void	lgrp_mem_init(int, lgrp_handle_t, boolean_t);
219 static void	lgrp_mem_fini(int, lgrp_handle_t, boolean_t);
220 static void	lgrp_mem_rename(int, lgrp_handle_t, lgrp_handle_t);
221 
222 /*
223  * lgroup CPU partition event handlers
224  */
225 static void	lgrp_part_add_cpu(struct cpu *, lgrp_id_t);
226 static void	lgrp_part_del_cpu(struct cpu *);
227 
228 /*
229  * lgroup framework initialization
230  */
231 static void	lgrp_main_init(void);
232 static void	lgrp_main_mp_init(void);
233 static void	lgrp_root_init(void);
234 static void	lgrp_setup(void);
235 
236 /*
237  * lpl topology
238  */
239 static void	lpl_init(lpl_t *, lpl_t *, lgrp_t *);
240 static void	lpl_clear(lpl_t *);
241 static void	lpl_leaf_insert(lpl_t *, struct cpupart *);
242 static void	lpl_leaf_remove(lpl_t *, struct cpupart *);
243 static void	lpl_rset_add(lpl_t *, lpl_t *);
244 static void	lpl_rset_del(lpl_t *, lpl_t *);
245 static int	lpl_rset_contains(lpl_t *, lpl_t *);
246 static void	lpl_cpu_adjcnt(lpl_act_t, struct cpu *);
247 static void	lpl_child_update(lpl_t *, struct cpupart *);
248 static int	lpl_pick(lpl_t *, lpl_t *);
249 static void	lpl_verify_wrapper(struct cpupart *);
250 
251 /*
252  * defines for lpl topology verifier return codes
253  */
254 
255 #define	LPL_TOPO_CORRECT			0
256 #define	LPL_TOPO_PART_HAS_NO_LPL		-1
257 #define	LPL_TOPO_CPUS_NOT_EMPTY			-2
258 #define	LPL_TOPO_LGRP_MISMATCH			-3
259 #define	LPL_TOPO_MISSING_PARENT			-4
260 #define	LPL_TOPO_PARENT_MISMATCH		-5
261 #define	LPL_TOPO_BAD_CPUCNT			-6
262 #define	LPL_TOPO_RSET_MISMATCH			-7
263 #define	LPL_TOPO_LPL_ORPHANED			-8
264 #define	LPL_TOPO_LPL_BAD_NCPU			-9
265 #define	LPL_TOPO_RSET_MSSNG_LF			-10
266 #define	LPL_TOPO_CPU_HAS_BAD_LPL		-11
267 #define	LPL_TOPO_NONLEAF_HAS_CPUS		-12
268 #define	LPL_TOPO_LGRP_NOT_LEAF			-13
269 #define	LPL_TOPO_BAD_RSETCNT			-14
270 
271 /*
272  * Return whether lgroup optimizations should be enabled on this system
273  */
274 int
275 lgrp_optimizations(void)
276 {
277 	/*
278 	 * System must have more than 2 lgroups to enable lgroup optimizations
279 	 *
280 	 * XXX This assumes that a 2 lgroup system has an empty root lgroup
281 	 * with one child lgroup containing all the resources. A 2 lgroup
282 	 * system with a root lgroup directly containing CPUs or memory might
283 	 * need lgroup optimizations with its child lgroup, but there
284 	 * isn't such a machine for now....
285 	 */
286 	if (nlgrps > 2)
287 		return (1);
288 
289 	return (0);
290 }
291 
292 /*
293  * Setup root lgroup
294  */
295 static void
296 lgrp_root_init(void)
297 {
298 	lgrp_handle_t	hand;
299 	int		i;
300 	lgrp_id_t	id;
301 
302 	/*
303 	 * Create the "root" lgroup
304 	 */
305 	ASSERT(nlgrps == 0);
306 	id = nlgrps++;
307 
308 	lgrp_root = &lroot;
309 
310 	lgrp_root->lgrp_cpu = NULL;
311 	lgrp_root->lgrp_mnodes = 0;
312 	lgrp_root->lgrp_nmnodes = 0;
313 	hand = lgrp_plat_root_hand();
314 	lgrp_root->lgrp_plathand = hand;
315 
316 	lgrp_root->lgrp_id = id;
317 	lgrp_root->lgrp_cpucnt = 0;
318 	lgrp_root->lgrp_childcnt = 0;
319 	klgrpset_clear(lgrp_root->lgrp_children);
320 	klgrpset_clear(lgrp_root->lgrp_leaves);
321 	lgrp_root->lgrp_parent = NULL;
322 	lgrp_root->lgrp_latency = lgrp_plat_latency(hand, hand);
323 
324 	for (i = 0; i < LGRP_RSRC_COUNT; i++)
325 		klgrpset_clear(lgrp_root->lgrp_set[i]);
326 
327 	lgrp_root->lgrp_kstat = NULL;
328 
329 	lgrp_table[id] = lgrp_root;
330 
331 	/*
332 	 * Setup initial lpl list for CPU0 and initial t0 home.
333 	 * The only lpl space we have so far is lpl_bootstrap. It is used for
334 	 * all topology operations until cp_default is initialized at which
335 	 * point t0.t_lpl will be updated.
336 	 */
337 	lpl_bootstrap = lpl_bootstrap_list;
338 	t0.t_lpl = lpl_bootstrap;
339 	cp_default.cp_nlgrploads = LPL_BOOTSTRAP_SIZE;
340 	lpl_bootstrap_list[1].lpl_lgrpid = 1;
341 
342 	/*
343 	 * Set up the bootstrap rset
344 	 * Since the bootstrap toplogy has just the root, and a leaf,
345 	 * the rset contains just the leaf, and both lpls can use the same rset
346 	 */
347 	lpl_bootstrap_rset[0] = &lpl_bootstrap_list[1];
348 	lpl_bootstrap_list[0].lpl_rset_sz = 1;
349 	lpl_bootstrap_list[0].lpl_rset = lpl_bootstrap_rset;
350 	lpl_bootstrap_list[0].lpl_id2rset = lpl_bootstrap_id2rset;
351 
352 	lpl_bootstrap_list[1].lpl_rset_sz = 1;
353 	lpl_bootstrap_list[1].lpl_rset = lpl_bootstrap_rset;
354 	lpl_bootstrap_list[1].lpl_id2rset = lpl_bootstrap_id2rset;
355 
356 	cp_default.cp_lgrploads = lpl_bootstrap;
357 }
358 
359 /*
360  * Initialize the lgroup framework and allow the platform to do the same
361  *
362  * This happens in stages during boot and is all funnelled through this routine
363  * (see definition of lgrp_init_stages_t to see what happens at each stage and
364  * when)
365  */
366 void
367 lgrp_init(lgrp_init_stages_t stage)
368 {
369 	/*
370 	 * Initialize the platform
371 	 */
372 	lgrp_plat_init(stage);
373 
374 	switch (stage) {
375 	case LGRP_INIT_STAGE1:
376 		/*
377 		 * Set max number of lgroups supported on this platform which
378 		 * must be less than the max number of lgroups supported by the
379 		 * common lgroup framework (eg. NLGRPS_MAX is max elements in
380 		 * lgrp_table[], etc.)
381 		 */
382 		nlgrpsmax = lgrp_plat_max_lgrps();
383 		ASSERT(nlgrpsmax <= NLGRPS_MAX);
384 		break;
385 
386 	case LGRP_INIT_STAGE2:
387 		lgrp_setup();
388 		break;
389 
390 	case LGRP_INIT_STAGE4:
391 		lgrp_main_init();
392 		break;
393 
394 	case LGRP_INIT_STAGE5:
395 		lgrp_main_mp_init();
396 		break;
397 
398 	default:
399 		break;
400 	}
401 }
402 
403 /*
404  * Create the root and cpu0's lgroup, and set t0's home.
405  */
406 static void
407 lgrp_setup(void)
408 {
409 	/*
410 	 * Setup the root lgroup
411 	 */
412 	lgrp_root_init();
413 
414 	/*
415 	 * Add cpu0 to an lgroup
416 	 */
417 	lgrp_config(LGRP_CONFIG_CPU_ADD, (uintptr_t)CPU, 0);
418 	lgrp_config(LGRP_CONFIG_CPU_ONLINE, (uintptr_t)CPU, 0);
419 }
420 
421 /*
422  * true when lgrp initialization has been completed.
423  */
424 int	lgrp_initialized = 0;
425 
426 /*
427  * True when lgrp topology is constructed.
428  */
429 int	lgrp_topo_initialized = 0;
430 
431 /*
432  * Init routine called after startup(), /etc/system has been processed,
433  * and cpu0 has been added to an lgroup.
434  */
435 static void
436 lgrp_main_init(void)
437 {
438 	cpu_t		*cp = CPU;
439 	lgrp_id_t	lgrpid;
440 	int		i;
441 	extern void	pg_cpu0_reinit();
442 
443 	/*
444 	 * Enforce a valid lgrp_mem_default_policy
445 	 */
446 	if ((lgrp_mem_default_policy <= LGRP_MEM_POLICY_DEFAULT) ||
447 	    (lgrp_mem_default_policy >= LGRP_NUM_MEM_POLICIES) ||
448 	    (lgrp_mem_default_policy == LGRP_MEM_POLICY_NEXT_SEG))
449 		lgrp_mem_default_policy = LGRP_MEM_POLICY_NEXT;
450 
451 	/*
452 	 * See if mpo should be disabled.
453 	 * This may happen in the case of null proc LPA on Starcat.
454 	 * The platform won't be able to detect null proc LPA until after
455 	 * cpu0 and memory have already been added to lgroups.
456 	 * When and if it is detected, the Starcat platform will return
457 	 * a different platform handle for cpu0 which is what we check for
458 	 * here. If mpo should be disabled move cpu0 to it's rightful place
459 	 * (the root), and destroy the remaining lgroups. This effectively
460 	 * provides an UMA lgroup topology.
461 	 */
462 	lgrpid = cp->cpu_lpl->lpl_lgrpid;
463 	if (lgrp_table[lgrpid]->lgrp_plathand !=
464 	    lgrp_plat_cpu_to_hand(cp->cpu_id)) {
465 		lgrp_part_del_cpu(cp);
466 		lgrp_cpu_fini(cp, lgrpid);
467 
468 		lgrp_cpu_init(cp);
469 		lgrp_part_add_cpu(cp, cp->cpu_lpl->lpl_lgrpid);
470 
471 		ASSERT(cp->cpu_lpl->lpl_lgrpid == LGRP_ROOTID);
472 
473 		/*
474 		 * Notify the PG subsystem that the CPU's lgrp
475 		 * association has changed
476 		 */
477 		pg_cpu0_reinit();
478 
479 		/*
480 		 * Destroy all lgroups except for root
481 		 */
482 		for (i = 0; i <= lgrp_alloc_max; i++) {
483 			if (LGRP_EXISTS(lgrp_table[i]) &&
484 			    lgrp_table[i] != lgrp_root)
485 				lgrp_destroy(lgrp_table[i]);
486 		}
487 
488 		/*
489 		 * Fix up root to point at itself for leaves and resources
490 		 * and not have any children
491 		 */
492 		lgrp_root->lgrp_childcnt = 0;
493 		klgrpset_clear(lgrp_root->lgrp_children);
494 		klgrpset_clear(lgrp_root->lgrp_leaves);
495 		klgrpset_add(lgrp_root->lgrp_leaves, LGRP_ROOTID);
496 		klgrpset_clear(lgrp_root->lgrp_set[LGRP_RSRC_MEM]);
497 		klgrpset_add(lgrp_root->lgrp_set[LGRP_RSRC_MEM], LGRP_ROOTID);
498 	}
499 
500 	/*
501 	 * Initialize kstats framework.
502 	 */
503 	lgrp_kstat_init();
504 	/*
505 	 * cpu0 is finally where it should be, so create it's lgroup's kstats
506 	 */
507 	mutex_enter(&cpu_lock);
508 	lgrp_kstat_create(cp);
509 	mutex_exit(&cpu_lock);
510 
511 	lgrp_initialized = 1;
512 }
513 
514 /*
515  * Finish lgrp initialization after all CPUS are brought on-line.
516  * This routine is called after start_other_cpus().
517  */
518 static void
519 lgrp_main_mp_init(void)
520 {
521 	klgrpset_t changed;
522 
523 	/*
524 	 * Update lgroup topology (if necessary)
525 	 */
526 	klgrpset_clear(changed);
527 	(void) lgrp_topo_update(lgrp_table, lgrp_alloc_max + 1, &changed);
528 	lgrp_topo_initialized = 1;
529 }
530 
531 /*
532  * Change latency of lgroup with specified lgroup platform handle (if one is
533  * given) or change all lgroups with old latency to new latency
534  */
535 void
536 lgrp_latency_change(lgrp_handle_t hand, u_longlong_t oldtime,
537     u_longlong_t newtime)
538 {
539 	lgrp_t		*lgrp;
540 	int		i;
541 
542 	for (i = 0; i <= lgrp_alloc_max; i++) {
543 		lgrp = lgrp_table[i];
544 
545 		if (!LGRP_EXISTS(lgrp))
546 			continue;
547 
548 		if ((hand == LGRP_NULL_HANDLE &&
549 		    lgrp->lgrp_latency == oldtime) ||
550 		    (hand != LGRP_NULL_HANDLE && lgrp->lgrp_plathand == hand))
551 			lgrp->lgrp_latency = (int)newtime;
552 	}
553 }
554 
555 /*
556  * Handle lgroup (re)configuration events (eg. addition of CPU, etc.)
557  */
558 void
559 lgrp_config(lgrp_config_flag_t event, uintptr_t resource, uintptr_t where)
560 {
561 	klgrpset_t	changed;
562 	cpu_t		*cp;
563 	lgrp_id_t	id;
564 	int		rc;
565 
566 	switch (event) {
567 	/*
568 	 * The following (re)configuration events are common code
569 	 * initiated. lgrp_plat_config() is called here to inform the
570 	 * platform of the reconfiguration event.
571 	 */
572 	case LGRP_CONFIG_CPU_ADD:
573 		cp = (cpu_t *)resource;
574 
575 		/*
576 		 * Initialize the new CPU's lgrp related next/prev
577 		 * links, and give it a bootstrap lpl so that it can
578 		 * survive should it need to enter the dispatcher.
579 		 */
580 		cp->cpu_next_lpl = cp;
581 		cp->cpu_prev_lpl = cp;
582 		cp->cpu_next_lgrp = cp;
583 		cp->cpu_prev_lgrp = cp;
584 		cp->cpu_lpl = lpl_bootstrap;
585 
586 		lgrp_plat_config(event, resource);
587 		atomic_inc_32(&lgrp_gen);
588 
589 		break;
590 	case LGRP_CONFIG_CPU_DEL:
591 		lgrp_plat_config(event, resource);
592 		atomic_inc_32(&lgrp_gen);
593 
594 		break;
595 	case LGRP_CONFIG_CPU_ONLINE:
596 		cp = (cpu_t *)resource;
597 		lgrp_cpu_init(cp);
598 		lgrp_part_add_cpu(cp, cp->cpu_lpl->lpl_lgrpid);
599 		rc = lpl_topo_verify(cp->cpu_part);
600 		if (rc != LPL_TOPO_CORRECT) {
601 			panic("lpl_topo_verify failed: %d", rc);
602 		}
603 		lgrp_plat_config(event, resource);
604 		atomic_inc_32(&lgrp_gen);
605 
606 		break;
607 	case LGRP_CONFIG_CPU_OFFLINE:
608 		cp = (cpu_t *)resource;
609 		id = cp->cpu_lpl->lpl_lgrpid;
610 		lgrp_part_del_cpu(cp);
611 		lgrp_cpu_fini(cp, id);
612 		rc = lpl_topo_verify(cp->cpu_part);
613 		if (rc != LPL_TOPO_CORRECT) {
614 			panic("lpl_topo_verify failed: %d", rc);
615 		}
616 		lgrp_plat_config(event, resource);
617 		atomic_inc_32(&lgrp_gen);
618 
619 		break;
620 	case LGRP_CONFIG_CPUPART_ADD:
621 		cp = (cpu_t *)resource;
622 		lgrp_part_add_cpu((cpu_t *)resource, (lgrp_id_t)where);
623 		rc = lpl_topo_verify(cp->cpu_part);
624 		if (rc != LPL_TOPO_CORRECT) {
625 			panic("lpl_topo_verify failed: %d", rc);
626 		}
627 		lgrp_plat_config(event, resource);
628 
629 		break;
630 	case LGRP_CONFIG_CPUPART_DEL:
631 		cp = (cpu_t *)resource;
632 		lgrp_part_del_cpu((cpu_t *)resource);
633 		rc = lpl_topo_verify(cp->cpu_part);
634 		if (rc != LPL_TOPO_CORRECT) {
635 			panic("lpl_topo_verify failed: %d", rc);
636 		}
637 		lgrp_plat_config(event, resource);
638 
639 		break;
640 	/*
641 	 * The following events are initiated by the memnode
642 	 * subsystem.
643 	 */
644 	case LGRP_CONFIG_MEM_ADD:
645 		lgrp_mem_init((int)resource, where, B_FALSE);
646 		atomic_inc_32(&lgrp_gen);
647 
648 		break;
649 	case LGRP_CONFIG_MEM_DEL:
650 		lgrp_mem_fini((int)resource, where, B_FALSE);
651 		atomic_inc_32(&lgrp_gen);
652 
653 		break;
654 	case LGRP_CONFIG_MEM_RENAME: {
655 		lgrp_config_mem_rename_t *ren_arg =
656 		    (lgrp_config_mem_rename_t *)where;
657 
658 		lgrp_mem_rename((int)resource,
659 		    ren_arg->lmem_rename_from,
660 		    ren_arg->lmem_rename_to);
661 		atomic_inc_32(&lgrp_gen);
662 
663 		break;
664 	}
665 	case LGRP_CONFIG_GEN_UPDATE:
666 		atomic_inc_32(&lgrp_gen);
667 
668 		break;
669 	case LGRP_CONFIG_FLATTEN:
670 		if (where == 0)
671 			lgrp_topo_levels = (int)resource;
672 		else
673 			(void) lgrp_topo_flatten(resource,
674 			    lgrp_table, lgrp_alloc_max, &changed);
675 
676 		break;
677 	/*
678 	 * Update any lgroups with old latency to new latency
679 	 */
680 	case LGRP_CONFIG_LAT_CHANGE_ALL:
681 		lgrp_latency_change(LGRP_NULL_HANDLE, (u_longlong_t)resource,
682 		    (u_longlong_t)where);
683 
684 		break;
685 	/*
686 	 * Update lgroup with specified lgroup platform handle to have
687 	 * new latency
688 	 */
689 	case LGRP_CONFIG_LAT_CHANGE:
690 		lgrp_latency_change((lgrp_handle_t)resource, 0,
691 		    (u_longlong_t)where);
692 
693 		break;
694 	case LGRP_CONFIG_NOP:
695 
696 		break;
697 	default:
698 		break;
699 	}
700 
701 }
702 
703 /*
704  * Called to add lgrp info into cpu structure from cpu_add_unit;
705  * do not assume cpu is in cpu[] yet!
706  *
707  * CPUs are brought online with all other CPUs paused so we can't
708  * allocate memory or we could deadlock the system, so we rely on
709  * the platform to statically allocate as much space as we need
710  * for the lgrp structs and stats.
711  */
712 static void
713 lgrp_cpu_init(struct cpu *cp)
714 {
715 	klgrpset_t	changed;
716 	int		count;
717 	lgrp_handle_t	hand;
718 	int		first_cpu;
719 	lgrp_t		*my_lgrp;
720 	lgrp_id_t	lgrpid;
721 	struct cpu	*cptr;
722 
723 	/*
724 	 * This is the first time through if the resource set
725 	 * for the root lgroup is empty. After cpu0 has been
726 	 * initially added to an lgroup, the root's CPU resource
727 	 * set can never be empty, since the system's last CPU
728 	 * cannot be offlined.
729 	 */
730 	if (klgrpset_isempty(lgrp_root->lgrp_set[LGRP_RSRC_CPU])) {
731 		/*
732 		 * First time through.
733 		 */
734 		first_cpu = 1;
735 	} else {
736 		/*
737 		 * If cpu0 needs to move lgroups, we may come
738 		 * through here again, at which time cpu_lock won't
739 		 * be held, and lgrp_initialized will be false.
740 		 */
741 		ASSERT(MUTEX_HELD(&cpu_lock) || !lgrp_initialized);
742 		ASSERT(cp->cpu_part != NULL);
743 		first_cpu = 0;
744 	}
745 
746 	hand = lgrp_plat_cpu_to_hand(cp->cpu_id);
747 	my_lgrp = lgrp_hand_to_lgrp(hand);
748 
749 	if (my_lgrp == NULL) {
750 		/*
751 		 * Create new lgrp and add it to lgroup topology
752 		 */
753 		my_lgrp = lgrp_create();
754 		my_lgrp->lgrp_plathand = hand;
755 		my_lgrp->lgrp_latency = lgrp_plat_latency(hand, hand);
756 		lgrpid = my_lgrp->lgrp_id;
757 		klgrpset_add(my_lgrp->lgrp_leaves, lgrpid);
758 		klgrpset_add(my_lgrp->lgrp_set[LGRP_RSRC_CPU], lgrpid);
759 
760 		count = 0;
761 		klgrpset_clear(changed);
762 		count += lgrp_leaf_add(my_lgrp, lgrp_table, lgrp_alloc_max + 1,
763 		    &changed);
764 		/*
765 		 * May have added new intermediate lgroups, so need to add
766 		 * resources other than CPUs which are added below
767 		 */
768 		(void) lgrp_mnode_update(changed, NULL);
769 	} else if (my_lgrp->lgrp_latency == 0 && lgrp_plat_latency(hand, hand)
770 	    > 0) {
771 		/*
772 		 * Leaf lgroup was created, but latency wasn't available
773 		 * then.  So, set latency for it and fill in rest of lgroup
774 		 * topology  now that we know how far it is from other leaf
775 		 * lgroups.
776 		 */
777 		lgrpid = my_lgrp->lgrp_id;
778 		klgrpset_clear(changed);
779 		if (!klgrpset_ismember(my_lgrp->lgrp_set[LGRP_RSRC_CPU],
780 		    lgrpid))
781 			klgrpset_add(my_lgrp->lgrp_set[LGRP_RSRC_CPU], lgrpid);
782 		count = lgrp_leaf_add(my_lgrp, lgrp_table, lgrp_alloc_max + 1,
783 		    &changed);
784 
785 		/*
786 		 * May have added new intermediate lgroups, so need to add
787 		 * resources other than CPUs which are added below
788 		 */
789 		(void) lgrp_mnode_update(changed, NULL);
790 	} else if (!klgrpset_ismember(my_lgrp->lgrp_set[LGRP_RSRC_CPU],
791 	    my_lgrp->lgrp_id)) {
792 		int	i;
793 
794 		/*
795 		 * Update existing lgroup and lgroups containing it with CPU
796 		 * resource
797 		 */
798 		lgrpid = my_lgrp->lgrp_id;
799 		klgrpset_add(my_lgrp->lgrp_set[LGRP_RSRC_CPU], lgrpid);
800 		for (i = 0; i <= lgrp_alloc_max; i++) {
801 			lgrp_t		*lgrp;
802 
803 			lgrp = lgrp_table[i];
804 			if (!LGRP_EXISTS(lgrp) ||
805 			    !lgrp_rsets_member(lgrp->lgrp_set, lgrpid))
806 				continue;
807 
808 			klgrpset_add(lgrp->lgrp_set[LGRP_RSRC_CPU], lgrpid);
809 		}
810 	}
811 
812 	lgrpid = my_lgrp->lgrp_id;
813 	cp->cpu_lpl = &cp->cpu_part->cp_lgrploads[lgrpid];
814 
815 	/*
816 	 * For multi-lgroup systems, need to setup lpl for CPU0 or CPU0 will
817 	 * end up in lpl for lgroup 0 whether it is supposed to be in there or
818 	 * not since none of lgroup IDs in the lpl's have been set yet.
819 	 */
820 	if (first_cpu && nlgrpsmax > 1 && lgrpid != cp->cpu_lpl->lpl_lgrpid)
821 		cp->cpu_lpl->lpl_lgrpid = lgrpid;
822 
823 	/*
824 	 * link the CPU into the lgrp's CPU list
825 	 */
826 	if (my_lgrp->lgrp_cpucnt == 0) {
827 		my_lgrp->lgrp_cpu = cp;
828 		cp->cpu_next_lgrp = cp->cpu_prev_lgrp = cp;
829 	} else {
830 		cptr = my_lgrp->lgrp_cpu;
831 		cp->cpu_next_lgrp = cptr;
832 		cp->cpu_prev_lgrp = cptr->cpu_prev_lgrp;
833 		cptr->cpu_prev_lgrp->cpu_next_lgrp = cp;
834 		cptr->cpu_prev_lgrp = cp;
835 	}
836 	my_lgrp->lgrp_cpucnt++;
837 }
838 
839 lgrp_t *
840 lgrp_create(void)
841 {
842 	lgrp_t		*my_lgrp;
843 	lgrp_id_t	lgrpid;
844 	int		i;
845 
846 	ASSERT(!lgrp_initialized || MUTEX_HELD(&cpu_lock));
847 
848 	/*
849 	 * Find an open slot in the lgroup table and recycle unused lgroup
850 	 * left there if any
851 	 */
852 	my_lgrp = NULL;
853 	if (lgrp_alloc_hint == -1)
854 		/*
855 		 * Allocate from end when hint not set yet because no lgroups
856 		 * have been deleted yet
857 		 */
858 		lgrpid = nlgrps++;
859 	else {
860 		/*
861 		 * Start looking for next open slot from hint and leave hint
862 		 * at slot allocated
863 		 */
864 		for (i = lgrp_alloc_hint; i < nlgrpsmax; i++) {
865 			my_lgrp = lgrp_table[i];
866 			if (!LGRP_EXISTS(my_lgrp)) {
867 				lgrpid = i;
868 				nlgrps++;
869 				break;
870 			}
871 		}
872 		lgrp_alloc_hint = lgrpid;
873 	}
874 
875 	/*
876 	 * Keep track of max lgroup ID allocated so far to cut down on searches
877 	 */
878 	if (lgrpid > lgrp_alloc_max)
879 		lgrp_alloc_max = lgrpid;
880 
881 	/*
882 	 * Need to allocate new lgroup if next open slot didn't have one
883 	 * for recycling
884 	 */
885 	if (my_lgrp == NULL)
886 		my_lgrp = lgrp_plat_alloc(lgrpid);
887 
888 	if (nlgrps > nlgrpsmax || my_lgrp == NULL)
889 		panic("Too many lgrps for platform (%d)", nlgrps);
890 
891 	my_lgrp->lgrp_id = lgrpid;
892 	my_lgrp->lgrp_latency = 0;
893 	my_lgrp->lgrp_plathand = LGRP_NULL_HANDLE;
894 	my_lgrp->lgrp_parent = NULL;
895 	my_lgrp->lgrp_childcnt = 0;
896 	my_lgrp->lgrp_mnodes = (mnodeset_t)0;
897 	my_lgrp->lgrp_nmnodes = 0;
898 	klgrpset_clear(my_lgrp->lgrp_children);
899 	klgrpset_clear(my_lgrp->lgrp_leaves);
900 	for (i = 0; i < LGRP_RSRC_COUNT; i++)
901 		klgrpset_clear(my_lgrp->lgrp_set[i]);
902 
903 	my_lgrp->lgrp_cpu = NULL;
904 	my_lgrp->lgrp_cpucnt = 0;
905 
906 	if (my_lgrp->lgrp_kstat != NULL)
907 		lgrp_kstat_reset(lgrpid);
908 
909 	lgrp_table[my_lgrp->lgrp_id] = my_lgrp;
910 
911 	return (my_lgrp);
912 }
913 
914 void
915 lgrp_destroy(lgrp_t *lgrp)
916 {
917 	int		i;
918 
919 	/*
920 	 * Unless this lgroup is being destroyed on behalf of
921 	 * the boot CPU, cpu_lock must be held
922 	 */
923 	ASSERT(!lgrp_initialized || MUTEX_HELD(&cpu_lock));
924 
925 	if (nlgrps == 1)
926 		cmn_err(CE_PANIC, "Can't destroy only lgroup!");
927 
928 	if (!LGRP_EXISTS(lgrp))
929 		return;
930 
931 	/*
932 	 * Set hint to lgroup being deleted and try to keep lower numbered
933 	 * hints to facilitate finding empty slots
934 	 */
935 	if (lgrp_alloc_hint == -1 || lgrp->lgrp_id < lgrp_alloc_hint)
936 		lgrp_alloc_hint = lgrp->lgrp_id;
937 
938 	/*
939 	 * Mark this lgroup to be recycled by setting its lgroup ID to
940 	 * LGRP_NONE and clear relevant fields
941 	 */
942 	lgrp->lgrp_id = LGRP_NONE;
943 	lgrp->lgrp_latency = 0;
944 	lgrp->lgrp_plathand = LGRP_NULL_HANDLE;
945 	lgrp->lgrp_parent = NULL;
946 	lgrp->lgrp_childcnt = 0;
947 
948 	klgrpset_clear(lgrp->lgrp_children);
949 	klgrpset_clear(lgrp->lgrp_leaves);
950 	for (i = 0; i < LGRP_RSRC_COUNT; i++)
951 		klgrpset_clear(lgrp->lgrp_set[i]);
952 
953 	lgrp->lgrp_mnodes = (mnodeset_t)0;
954 	lgrp->lgrp_nmnodes = 0;
955 
956 	lgrp->lgrp_cpu = NULL;
957 	lgrp->lgrp_cpucnt = 0;
958 
959 	nlgrps--;
960 }
961 
962 /*
963  * Initialize kstat data. Called from lgrp intialization code.
964  */
965 static void
966 lgrp_kstat_init(void)
967 {
968 	lgrp_stat_t	stat;
969 
970 	mutex_init(&lgrp_kstat_mutex, NULL, MUTEX_DEFAULT, NULL);
971 
972 	for (stat = 0; stat < LGRP_NUM_STATS; stat++)
973 		kstat_named_init(&lgrp_kstat_data[stat],
974 		    lgrp_kstat_names[stat], KSTAT_DATA_INT64);
975 }
976 
977 /*
978  * initialize an lgrp's kstats if needed
979  * called with cpu_lock held but not with cpus paused.
980  * we don't tear these down now because we don't know about
981  * memory leaving the lgrp yet...
982  */
983 
984 void
985 lgrp_kstat_create(cpu_t *cp)
986 {
987 	kstat_t		*lgrp_kstat;
988 	lgrp_id_t	lgrpid;
989 	lgrp_t		*my_lgrp;
990 
991 	ASSERT(MUTEX_HELD(&cpu_lock));
992 
993 	lgrpid = cp->cpu_lpl->lpl_lgrpid;
994 	my_lgrp = lgrp_table[lgrpid];
995 
996 	if (my_lgrp->lgrp_kstat != NULL)
997 		return; /* already initialized */
998 
999 	lgrp_kstat = kstat_create("lgrp", lgrpid, NULL, "misc",
1000 	    KSTAT_TYPE_NAMED, LGRP_NUM_STATS,
1001 	    KSTAT_FLAG_VIRTUAL | KSTAT_FLAG_WRITABLE);
1002 
1003 	if (lgrp_kstat != NULL) {
1004 		lgrp_kstat->ks_lock = &lgrp_kstat_mutex;
1005 		lgrp_kstat->ks_private = my_lgrp;
1006 		lgrp_kstat->ks_data = &lgrp_kstat_data;
1007 		lgrp_kstat->ks_update = lgrp_kstat_extract;
1008 		my_lgrp->lgrp_kstat = lgrp_kstat;
1009 		kstat_install(lgrp_kstat);
1010 	}
1011 }
1012 
1013 /*
1014  * this will do something when we manage to remove now unused lgrps
1015  */
1016 
1017 /* ARGSUSED */
1018 void
1019 lgrp_kstat_destroy(cpu_t *cp)
1020 {
1021 	ASSERT(MUTEX_HELD(&cpu_lock));
1022 }
1023 
1024 /*
1025  * Called when a CPU is off-lined.
1026  */
1027 static void
1028 lgrp_cpu_fini(struct cpu *cp, lgrp_id_t lgrpid)
1029 {
1030 	lgrp_t *my_lgrp;
1031 	struct cpu *prev;
1032 	struct cpu *next;
1033 
1034 	ASSERT(MUTEX_HELD(&cpu_lock) || !lgrp_initialized);
1035 
1036 	prev = cp->cpu_prev_lgrp;
1037 	next = cp->cpu_next_lgrp;
1038 
1039 	prev->cpu_next_lgrp = next;
1040 	next->cpu_prev_lgrp = prev;
1041 
1042 	/*
1043 	 * just because I'm paranoid doesn't mean...
1044 	 */
1045 
1046 	cp->cpu_next_lgrp = cp->cpu_prev_lgrp = NULL;
1047 
1048 	my_lgrp = lgrp_table[lgrpid];
1049 	my_lgrp->lgrp_cpucnt--;
1050 
1051 	/*
1052 	 * Removing last CPU in lgroup, so update lgroup topology
1053 	 */
1054 	if (my_lgrp->lgrp_cpucnt == 0) {
1055 		klgrpset_t	changed;
1056 		int		count;
1057 		int		i;
1058 
1059 		my_lgrp->lgrp_cpu = NULL;
1060 
1061 		/*
1062 		 * Remove this lgroup from its lgroup CPU resources and remove
1063 		 * lgroup from lgroup topology if it doesn't have any more
1064 		 * resources in it now
1065 		 */
1066 		klgrpset_del(my_lgrp->lgrp_set[LGRP_RSRC_CPU], lgrpid);
1067 		if (lgrp_rsets_empty(my_lgrp->lgrp_set)) {
1068 			count = 0;
1069 			klgrpset_clear(changed);
1070 			count += lgrp_leaf_delete(my_lgrp, lgrp_table,
1071 			    lgrp_alloc_max + 1, &changed);
1072 			return;
1073 		}
1074 
1075 		/*
1076 		 * This lgroup isn't empty, so just remove it from CPU
1077 		 * resources of any lgroups that contain it as such
1078 		 */
1079 		for (i = 0; i <= lgrp_alloc_max; i++) {
1080 			lgrp_t		*lgrp;
1081 
1082 			lgrp = lgrp_table[i];
1083 			if (!LGRP_EXISTS(lgrp) ||
1084 			    !klgrpset_ismember(lgrp->lgrp_set[LGRP_RSRC_CPU],
1085 			    lgrpid))
1086 				continue;
1087 
1088 			klgrpset_del(lgrp->lgrp_set[LGRP_RSRC_CPU], lgrpid);
1089 		}
1090 		return;
1091 	}
1092 
1093 	if (my_lgrp->lgrp_cpu == cp)
1094 		my_lgrp->lgrp_cpu = next;
1095 
1096 }
1097 
1098 /*
1099  * Update memory nodes in target lgroups and return ones that get changed
1100  */
1101 int
1102 lgrp_mnode_update(klgrpset_t target, klgrpset_t *changed)
1103 {
1104 	int	count;
1105 	int	i;
1106 	int	j;
1107 	lgrp_t	*lgrp;
1108 	lgrp_t	*lgrp_rsrc;
1109 
1110 	count = 0;
1111 	if (changed)
1112 		klgrpset_clear(*changed);
1113 
1114 	if (klgrpset_isempty(target))
1115 		return (0);
1116 
1117 	/*
1118 	 * Find each lgroup in target lgroups
1119 	 */
1120 	for (i = 0; i <= lgrp_alloc_max; i++) {
1121 		/*
1122 		 * Skip any lgroups that don't exist or aren't in target group
1123 		 */
1124 		lgrp = lgrp_table[i];
1125 		if (!klgrpset_ismember(target, i) || !LGRP_EXISTS(lgrp)) {
1126 			continue;
1127 		}
1128 
1129 		/*
1130 		 * Initialize memnodes for intermediate lgroups to 0
1131 		 * and update them from scratch since they may have completely
1132 		 * changed
1133 		 */
1134 		if (lgrp->lgrp_childcnt && lgrp != lgrp_root) {
1135 			lgrp->lgrp_mnodes = (mnodeset_t)0;
1136 			lgrp->lgrp_nmnodes = 0;
1137 		}
1138 
1139 		/*
1140 		 * Update memory nodes of of target lgroup with memory nodes
1141 		 * from each lgroup in its lgroup memory resource set
1142 		 */
1143 		for (j = 0; j <= lgrp_alloc_max; j++) {
1144 			int	k;
1145 
1146 			/*
1147 			 * Skip any lgroups that don't exist or aren't in
1148 			 * memory resources of target lgroup
1149 			 */
1150 			lgrp_rsrc = lgrp_table[j];
1151 			if (!LGRP_EXISTS(lgrp_rsrc) ||
1152 			    !klgrpset_ismember(lgrp->lgrp_set[LGRP_RSRC_MEM],
1153 			    j))
1154 				continue;
1155 
1156 			/*
1157 			 * Update target lgroup's memnodes to include memnodes
1158 			 * of this lgroup
1159 			 */
1160 			for (k = 0; k < sizeof (mnodeset_t) * NBBY; k++) {
1161 				mnodeset_t	mnode_mask;
1162 
1163 				mnode_mask = (mnodeset_t)1 << k;
1164 				if ((lgrp_rsrc->lgrp_mnodes & mnode_mask) &&
1165 				    !(lgrp->lgrp_mnodes & mnode_mask)) {
1166 					lgrp->lgrp_mnodes |= mnode_mask;
1167 					lgrp->lgrp_nmnodes++;
1168 				}
1169 			}
1170 			count++;
1171 			if (changed)
1172 				klgrpset_add(*changed, lgrp->lgrp_id);
1173 		}
1174 	}
1175 
1176 	return (count);
1177 }
1178 
1179 /*
1180  * Memory copy-rename. Called when the "mnode" containing the kernel cage memory
1181  * is moved from one board to another. The "from" and "to" arguments specify the
1182  * source and the destination of the move.
1183  *
1184  * See plat_lgrp_config() for a detailed description of the copy-rename
1185  * semantics.
1186  *
1187  * The lgrp_mem_rename() is called by the platform copy-rename code to update
1188  * the lgroup topology which is changing as memory moves from one lgroup to
1189  * another. It removes the mnode from the source lgroup and re-inserts it in the
1190  * target lgroup.
1191  *
1192  * The lgrp_mem_rename() function passes a flag to lgrp_mem_init() and
1193  * lgrp_mem_fini() telling that the insertion and deleteion are part of a DR
1194  * copy-rename operation.
1195  *
1196  * There is one case which requires special handling. If the system contains
1197  * only two boards (mnodes), the lgrp_mem_fini() removes the only mnode from the
1198  * lgroup hierarchy. This mnode is soon re-inserted back in the hierarchy by
1199  * lgrp_mem_init), but there is a window when the system has no memory in the
1200  * lgroup hierarchy. If another thread tries to allocate memory during this
1201  * window, the allocation will fail, although the system has physical memory.
1202  * This may cause a system panic or a deadlock (some sleeping memory allocations
1203  * happen with cpu_lock held which prevents lgrp_mem_init() from re-inserting
1204  * the mnode back).
1205  *
1206  * The lgrp_memnode_choose() function walks the lgroup hierarchy looking for the
1207  * lgrp with non-empty lgrp_mnodes. To deal with the special case above,
1208  * lgrp_mem_fini() does not remove the last mnode from the lroot->lgrp_mnodes,
1209  * but it updates the rest of the lgroup topology as if the mnode was actually
1210  * removed. The lgrp_mem_init() function recognizes that the mnode being
1211  * inserted represents such a special case and updates the topology
1212  * appropriately.
1213  */
1214 void
1215 lgrp_mem_rename(int mnode, lgrp_handle_t from, lgrp_handle_t to)
1216 {
1217 	/*
1218 	 * Remove the memory from the source node and add it to the destination
1219 	 * node.
1220 	 */
1221 	lgrp_mem_fini(mnode, from, B_TRUE);
1222 	lgrp_mem_init(mnode, to, B_TRUE);
1223 }
1224 
1225 /*
1226  * Called to indicate that the lgrp with platform handle "hand" now
1227  * contains the memory identified by "mnode".
1228  *
1229  * LOCKING for this routine is a bit tricky. Usually it is called without
1230  * cpu_lock and it must must grab cpu_lock here to prevent racing with other
1231  * callers. During DR of the board containing the caged memory it may be called
1232  * with cpu_lock already held and CPUs paused.
1233  *
1234  * If the insertion is part of the DR copy-rename and the inserted mnode (and
1235  * only this mnode) is already present in the lgrp_root->lgrp_mnodes set, we are
1236  * dealing with the special case of DR copy-rename described in
1237  * lgrp_mem_rename().
1238  */
1239 void
1240 lgrp_mem_init(int mnode, lgrp_handle_t hand, boolean_t is_copy_rename)
1241 {
1242 	klgrpset_t	changed;
1243 	int		count;
1244 	int		i;
1245 	lgrp_t		*my_lgrp;
1246 	lgrp_id_t	lgrpid;
1247 	mnodeset_t	mnodes_mask = ((mnodeset_t)1 << mnode);
1248 	boolean_t	drop_lock = B_FALSE;
1249 	boolean_t	need_synch = B_FALSE;
1250 
1251 	/*
1252 	 * Grab CPU lock (if we haven't already)
1253 	 */
1254 	if (!MUTEX_HELD(&cpu_lock)) {
1255 		mutex_enter(&cpu_lock);
1256 		drop_lock = B_TRUE;
1257 	}
1258 
1259 	/*
1260 	 * This routine may be called from a context where we already
1261 	 * hold cpu_lock, and have already paused cpus.
1262 	 */
1263 	if (!cpus_paused())
1264 		need_synch = B_TRUE;
1265 
1266 	/*
1267 	 * Check if this mnode is already configured and return immediately if
1268 	 * it is.
1269 	 *
1270 	 * NOTE: in special case of copy-rename of the only remaining mnode,
1271 	 * lgrp_mem_fini() refuses to remove the last mnode from the root, so we
1272 	 * recognize this case and continue as usual, but skip the update to
1273 	 * the lgrp_mnodes and the lgrp_nmnodes. This restores the inconsistency
1274 	 * in topology, temporarily introduced by lgrp_mem_fini().
1275 	 */
1276 	if (! (is_copy_rename && (lgrp_root->lgrp_mnodes == mnodes_mask)) &&
1277 	    lgrp_root->lgrp_mnodes & mnodes_mask) {
1278 		if (drop_lock)
1279 			mutex_exit(&cpu_lock);
1280 		return;
1281 	}
1282 
1283 	/*
1284 	 * Update lgroup topology with new memory resources, keeping track of
1285 	 * which lgroups change
1286 	 */
1287 	count = 0;
1288 	klgrpset_clear(changed);
1289 	my_lgrp = lgrp_hand_to_lgrp(hand);
1290 	if (my_lgrp == NULL) {
1291 		/* new lgrp */
1292 		my_lgrp = lgrp_create();
1293 		lgrpid = my_lgrp->lgrp_id;
1294 		my_lgrp->lgrp_plathand = hand;
1295 		my_lgrp->lgrp_latency = lgrp_plat_latency(hand, hand);
1296 		klgrpset_add(my_lgrp->lgrp_leaves, lgrpid);
1297 		klgrpset_add(my_lgrp->lgrp_set[LGRP_RSRC_MEM], lgrpid);
1298 
1299 		if (need_synch)
1300 			pause_cpus(NULL, NULL);
1301 		count = lgrp_leaf_add(my_lgrp, lgrp_table, lgrp_alloc_max + 1,
1302 		    &changed);
1303 		if (need_synch)
1304 			start_cpus();
1305 	} else if (my_lgrp->lgrp_latency == 0 && lgrp_plat_latency(hand, hand)
1306 	    > 0) {
1307 		/*
1308 		 * Leaf lgroup was created, but latency wasn't available
1309 		 * then.  So, set latency for it and fill in rest of lgroup
1310 		 * topology  now that we know how far it is from other leaf
1311 		 * lgroups.
1312 		 */
1313 		klgrpset_clear(changed);
1314 		lgrpid = my_lgrp->lgrp_id;
1315 		if (!klgrpset_ismember(my_lgrp->lgrp_set[LGRP_RSRC_MEM],
1316 		    lgrpid))
1317 			klgrpset_add(my_lgrp->lgrp_set[LGRP_RSRC_MEM], lgrpid);
1318 		if (need_synch)
1319 			pause_cpus(NULL, NULL);
1320 		count = lgrp_leaf_add(my_lgrp, lgrp_table, lgrp_alloc_max + 1,
1321 		    &changed);
1322 		if (need_synch)
1323 			start_cpus();
1324 	} else if (!klgrpset_ismember(my_lgrp->lgrp_set[LGRP_RSRC_MEM],
1325 	    my_lgrp->lgrp_id)) {
1326 		/*
1327 		 * Add new lgroup memory resource to existing lgroup
1328 		 */
1329 		lgrpid = my_lgrp->lgrp_id;
1330 		klgrpset_add(my_lgrp->lgrp_set[LGRP_RSRC_MEM], lgrpid);
1331 		klgrpset_add(changed, lgrpid);
1332 		count++;
1333 		for (i = 0; i <= lgrp_alloc_max; i++) {
1334 			lgrp_t		*lgrp;
1335 
1336 			lgrp = lgrp_table[i];
1337 			if (!LGRP_EXISTS(lgrp) ||
1338 			    !lgrp_rsets_member(lgrp->lgrp_set, lgrpid))
1339 				continue;
1340 
1341 			klgrpset_add(lgrp->lgrp_set[LGRP_RSRC_MEM], lgrpid);
1342 			klgrpset_add(changed, lgrp->lgrp_id);
1343 			count++;
1344 		}
1345 	}
1346 
1347 	/*
1348 	 * Add memory node to lgroup and remove lgroup from ones that need
1349 	 * to be updated
1350 	 */
1351 	if (!(my_lgrp->lgrp_mnodes & mnodes_mask)) {
1352 		my_lgrp->lgrp_mnodes |= mnodes_mask;
1353 		my_lgrp->lgrp_nmnodes++;
1354 	}
1355 	klgrpset_del(changed, lgrpid);
1356 
1357 	/*
1358 	 * Update memory node information for all lgroups that changed and
1359 	 * contain new memory node as a resource
1360 	 */
1361 	if (count)
1362 		(void) lgrp_mnode_update(changed, NULL);
1363 
1364 	if (drop_lock)
1365 		mutex_exit(&cpu_lock);
1366 }
1367 
1368 /*
1369  * Called to indicate that the lgroup associated with the platform
1370  * handle "hand" no longer contains given memory node
1371  *
1372  * LOCKING for this routine is a bit tricky. Usually it is called without
1373  * cpu_lock and it must must grab cpu_lock here to prevent racing with other
1374  * callers. During DR of the board containing the caged memory it may be called
1375  * with cpu_lock already held and CPUs paused.
1376  *
1377  * If the deletion is part of the DR copy-rename and the deleted mnode is the
1378  * only one present in the lgrp_root->lgrp_mnodes, all the topology is updated,
1379  * but lgrp_root->lgrp_mnodes is left intact. Later, lgrp_mem_init() will insert
1380  * the same mnode back into the topology. See lgrp_mem_rename() and
1381  * lgrp_mem_init() for additional details.
1382  */
1383 void
1384 lgrp_mem_fini(int mnode, lgrp_handle_t hand, boolean_t is_copy_rename)
1385 {
1386 	klgrpset_t	changed;
1387 	int		count;
1388 	int		i;
1389 	lgrp_t		*my_lgrp;
1390 	lgrp_id_t	lgrpid;
1391 	mnodeset_t	mnodes_mask;
1392 	boolean_t	drop_lock = B_FALSE;
1393 	boolean_t	need_synch = B_FALSE;
1394 
1395 	/*
1396 	 * Grab CPU lock (if we haven't already)
1397 	 */
1398 	if (!MUTEX_HELD(&cpu_lock)) {
1399 		mutex_enter(&cpu_lock);
1400 		drop_lock = B_TRUE;
1401 	}
1402 
1403 	/*
1404 	 * This routine may be called from a context where we already
1405 	 * hold cpu_lock and have already paused cpus.
1406 	 */
1407 	if (!cpus_paused())
1408 		need_synch = B_TRUE;
1409 
1410 	my_lgrp = lgrp_hand_to_lgrp(hand);
1411 
1412 	/*
1413 	 * The lgrp *must* be pre-existing
1414 	 */
1415 	ASSERT(my_lgrp != NULL);
1416 
1417 	/*
1418 	 * Delete memory node from lgroups which contain it
1419 	 */
1420 	mnodes_mask = ((mnodeset_t)1 << mnode);
1421 	for (i = 0; i <= lgrp_alloc_max; i++) {
1422 		lgrp_t *lgrp = lgrp_table[i];
1423 		/*
1424 		 * Skip any non-existent lgroups and any lgroups that don't
1425 		 * contain leaf lgroup of memory as a memory resource
1426 		 */
1427 		if (!LGRP_EXISTS(lgrp) ||
1428 		    !(lgrp->lgrp_mnodes & mnodes_mask))
1429 			continue;
1430 
1431 		/*
1432 		 * Avoid removing the last mnode from the root in the DR
1433 		 * copy-rename case. See lgrp_mem_rename() for details.
1434 		 */
1435 		if (is_copy_rename &&
1436 		    (lgrp == lgrp_root) && (lgrp->lgrp_mnodes == mnodes_mask))
1437 			continue;
1438 
1439 		/*
1440 		 * Remove memory node from lgroup.
1441 		 */
1442 		lgrp->lgrp_mnodes &= ~mnodes_mask;
1443 		lgrp->lgrp_nmnodes--;
1444 		ASSERT(lgrp->lgrp_nmnodes >= 0);
1445 	}
1446 	ASSERT(lgrp_root->lgrp_nmnodes > 0);
1447 
1448 	/*
1449 	 * Don't need to update lgroup topology if this lgroup still has memory.
1450 	 *
1451 	 * In the special case of DR copy-rename with the only mnode being
1452 	 * removed, the lgrp_mnodes for the root is always non-zero, but we
1453 	 * still need to update the lgroup topology.
1454 	 */
1455 	if ((my_lgrp->lgrp_nmnodes > 0) &&
1456 	    !(is_copy_rename && (my_lgrp == lgrp_root) &&
1457 	    (my_lgrp->lgrp_mnodes == mnodes_mask))) {
1458 		if (drop_lock)
1459 			mutex_exit(&cpu_lock);
1460 		return;
1461 	}
1462 
1463 	/*
1464 	 * This lgroup does not contain any memory now
1465 	 */
1466 	klgrpset_clear(my_lgrp->lgrp_set[LGRP_RSRC_MEM]);
1467 
1468 	/*
1469 	 * Remove this lgroup from lgroup topology if it does not contain any
1470 	 * resources now
1471 	 */
1472 	lgrpid = my_lgrp->lgrp_id;
1473 	count = 0;
1474 	klgrpset_clear(changed);
1475 	if (lgrp_rsets_empty(my_lgrp->lgrp_set)) {
1476 		/*
1477 		 * Delete lgroup when no more resources
1478 		 */
1479 		if (need_synch)
1480 			pause_cpus(NULL, NULL);
1481 		count = lgrp_leaf_delete(my_lgrp, lgrp_table,
1482 		    lgrp_alloc_max + 1, &changed);
1483 		ASSERT(count > 0);
1484 		if (need_synch)
1485 			start_cpus();
1486 	} else {
1487 		/*
1488 		 * Remove lgroup from memory resources of any lgroups that
1489 		 * contain it as such
1490 		 */
1491 		for (i = 0; i <= lgrp_alloc_max; i++) {
1492 			lgrp_t		*lgrp;
1493 
1494 			lgrp = lgrp_table[i];
1495 			if (!LGRP_EXISTS(lgrp) ||
1496 			    !klgrpset_ismember(lgrp->lgrp_set[LGRP_RSRC_MEM],
1497 			    lgrpid))
1498 				continue;
1499 
1500 			klgrpset_del(lgrp->lgrp_set[LGRP_RSRC_MEM], lgrpid);
1501 		}
1502 	}
1503 	if (drop_lock)
1504 		mutex_exit(&cpu_lock);
1505 }
1506 
1507 /*
1508  * Return lgroup with given platform handle
1509  */
1510 lgrp_t *
1511 lgrp_hand_to_lgrp(lgrp_handle_t hand)
1512 {
1513 	int	i;
1514 	lgrp_t	*lgrp;
1515 
1516 	if (hand == LGRP_NULL_HANDLE)
1517 		return (NULL);
1518 
1519 	for (i = 0; i <= lgrp_alloc_max; i++) {
1520 		lgrp = lgrp_table[i];
1521 		if (LGRP_EXISTS(lgrp) && lgrp->lgrp_plathand == hand)
1522 			return (lgrp);
1523 	}
1524 	return (NULL);
1525 }
1526 
1527 /*
1528  * Return the home lgroup of the current thread.
1529  * We must do this with kernel preemption disabled, since we don't want our
1530  * thread to be re-homed while we're poking around with its lpl, and the lpl
1531  * should never be NULL.
1532  *
1533  * NOTE: Can't guarantee that lgroup will be valid once kernel preemption
1534  * is enabled because of DR.  Callers can use disable kernel preemption
1535  * around this call to guarantee that the lgroup will be valid beyond this
1536  * routine, since kernel preemption can be recursive.
1537  */
1538 lgrp_t *
1539 lgrp_home_lgrp(void)
1540 {
1541 	lgrp_t	*lgrp;
1542 	lpl_t	*lpl;
1543 
1544 	kpreempt_disable();
1545 
1546 	lpl = curthread->t_lpl;
1547 	ASSERT(lpl != NULL);
1548 	ASSERT(lpl->lpl_lgrpid >= 0 && lpl->lpl_lgrpid <= lgrp_alloc_max);
1549 	ASSERT(LGRP_EXISTS(lgrp_table[lpl->lpl_lgrpid]));
1550 	lgrp = lgrp_table[lpl->lpl_lgrpid];
1551 
1552 	kpreempt_enable();
1553 
1554 	return (lgrp);
1555 }
1556 
1557 /*
1558  * Return ID of home lgroup for given thread
1559  * (See comments for lgrp_home_lgrp() for special care and handling
1560  * instructions)
1561  */
1562 lgrp_id_t
1563 lgrp_home_id(kthread_t *t)
1564 {
1565 	lgrp_id_t	lgrp;
1566 	lpl_t		*lpl;
1567 
1568 	ASSERT(t != NULL);
1569 	/*
1570 	 * We'd like to ASSERT(MUTEX_HELD(&ttoproc(t)->p_lock)), but we
1571 	 * cannot since the HAT layer can call into this routine to
1572 	 * determine the locality for its data structures in the context
1573 	 * of a page fault.
1574 	 */
1575 
1576 	kpreempt_disable();
1577 
1578 	lpl = t->t_lpl;
1579 	ASSERT(lpl != NULL);
1580 	ASSERT(lpl->lpl_lgrpid >= 0 && lpl->lpl_lgrpid <= lgrp_alloc_max);
1581 	lgrp = lpl->lpl_lgrpid;
1582 
1583 	kpreempt_enable();
1584 
1585 	return (lgrp);
1586 }
1587 
1588 /*
1589  * Return lgroup containing the physical memory for the given page frame number
1590  */
1591 lgrp_t *
1592 lgrp_pfn_to_lgrp(pfn_t pfn)
1593 {
1594 	lgrp_handle_t	hand;
1595 	int		i;
1596 	lgrp_t		*lgrp;
1597 
1598 	hand = lgrp_plat_pfn_to_hand(pfn);
1599 	if (hand != LGRP_NULL_HANDLE)
1600 		for (i = 0; i <= lgrp_alloc_max; i++) {
1601 			lgrp = lgrp_table[i];
1602 			if (LGRP_EXISTS(lgrp) && lgrp->lgrp_plathand == hand)
1603 				return (lgrp);
1604 		}
1605 	return (NULL);
1606 }
1607 
1608 /*
1609  * Return lgroup containing the physical memory for the given page frame number
1610  */
1611 lgrp_t *
1612 lgrp_phys_to_lgrp(u_longlong_t physaddr)
1613 {
1614 	lgrp_handle_t	hand;
1615 	int		i;
1616 	lgrp_t		*lgrp;
1617 	pfn_t		pfn;
1618 
1619 	pfn = btop(physaddr);
1620 	hand = lgrp_plat_pfn_to_hand(pfn);
1621 	if (hand != LGRP_NULL_HANDLE)
1622 		for (i = 0; i <= lgrp_alloc_max; i++) {
1623 			lgrp = lgrp_table[i];
1624 			if (LGRP_EXISTS(lgrp) && lgrp->lgrp_plathand == hand)
1625 				return (lgrp);
1626 		}
1627 	return (NULL);
1628 }
1629 
1630 /*
1631  * Return the leaf lgroup containing the given CPU
1632  *
1633  * The caller needs to take precautions necessary to prevent
1634  * "cpu", and it's lpl from going away across a call to this function.
1635  * hint: kpreempt_disable()/kpreempt_enable()
1636  */
1637 static lgrp_t *
1638 lgrp_cpu_to_lgrp(cpu_t *cpu)
1639 {
1640 	return (cpu->cpu_lpl->lpl_lgrp);
1641 }
1642 
1643 /*
1644  * Return the sum of the partition loads in an lgrp divided by
1645  * the number of CPUs in the lgrp.  This is our best approximation
1646  * of an 'lgroup load average' for a useful per-lgroup kstat.
1647  */
1648 static uint64_t
1649 lgrp_sum_loadavgs(lgrp_t *lgrp)
1650 {
1651 	cpu_t *cpu;
1652 	int ncpu;
1653 	uint64_t loads = 0;
1654 
1655 	mutex_enter(&cpu_lock);
1656 
1657 	cpu = lgrp->lgrp_cpu;
1658 	ncpu = lgrp->lgrp_cpucnt;
1659 
1660 	if (cpu == NULL || ncpu == 0) {
1661 		mutex_exit(&cpu_lock);
1662 		return (0ull);
1663 	}
1664 
1665 	do {
1666 		loads += cpu->cpu_lpl->lpl_loadavg;
1667 		cpu = cpu->cpu_next_lgrp;
1668 	} while (cpu != lgrp->lgrp_cpu);
1669 
1670 	mutex_exit(&cpu_lock);
1671 
1672 	return (loads / ncpu);
1673 }
1674 
1675 void
1676 lgrp_stat_add(lgrp_id_t lgrpid, lgrp_stat_t stat, int64_t val)
1677 {
1678 	struct lgrp_stats *pstats;
1679 
1680 	/*
1681 	 * Verify that the caller isn't trying to add to
1682 	 * a statistic for an lgroup that has gone away
1683 	 */
1684 	if (lgrpid < 0 || lgrpid > lgrp_alloc_max)
1685 		return;
1686 
1687 	pstats = &lgrp_stats[lgrpid];
1688 	atomic_add_64((uint64_t *)LGRP_STAT_WRITE_PTR(pstats, stat), val);
1689 }
1690 
1691 int64_t
1692 lgrp_stat_read(lgrp_id_t lgrpid, lgrp_stat_t stat)
1693 {
1694 	uint64_t val;
1695 	struct lgrp_stats *pstats;
1696 
1697 	if (lgrpid < 0 || lgrpid > lgrp_alloc_max)
1698 		return ((int64_t)0);
1699 
1700 	pstats = &lgrp_stats[lgrpid];
1701 	LGRP_STAT_READ(pstats, stat, val);
1702 	return (val);
1703 }
1704 
1705 /*
1706  * Reset all kstats for lgrp specified by its lgrpid.
1707  */
1708 static void
1709 lgrp_kstat_reset(lgrp_id_t lgrpid)
1710 {
1711 	lgrp_stat_t stat;
1712 
1713 	if (lgrpid < 0 || lgrpid > lgrp_alloc_max)
1714 		return;
1715 
1716 	for (stat = 0; stat < LGRP_NUM_COUNTER_STATS; stat++) {
1717 		LGRP_STAT_RESET(&lgrp_stats[lgrpid], stat);
1718 	}
1719 }
1720 
1721 /*
1722  * Collect all per-lgrp statistics for the lgrp associated with this
1723  * kstat, and store them in the ks_data array.
1724  *
1725  * The superuser can reset all the running counter statistics for an
1726  * lgrp by writing to any of the lgrp's stats.
1727  */
1728 static int
1729 lgrp_kstat_extract(kstat_t *ksp, int rw)
1730 {
1731 	lgrp_stat_t		stat;
1732 	struct kstat_named	*ksd;
1733 	lgrp_t			*lgrp;
1734 	lgrp_id_t		lgrpid;
1735 
1736 	lgrp = (lgrp_t *)ksp->ks_private;
1737 
1738 	ksd = (struct kstat_named *)ksp->ks_data;
1739 	ASSERT(ksd == (struct kstat_named *)&lgrp_kstat_data);
1740 
1741 	lgrpid = lgrp->lgrp_id;
1742 
1743 	if (lgrpid == LGRP_NONE) {
1744 		/*
1745 		 * Return all zeroes as stats for freed lgrp.
1746 		 */
1747 		for (stat = 0; stat < LGRP_NUM_COUNTER_STATS; stat++) {
1748 			ksd[stat].value.i64 = 0;
1749 		}
1750 		ksd[stat + LGRP_NUM_CPUS].value.i64 = 0;
1751 		ksd[stat + LGRP_NUM_PG_INSTALL].value.i64 = 0;
1752 		ksd[stat + LGRP_NUM_PG_AVAIL].value.i64 = 0;
1753 		ksd[stat + LGRP_NUM_PG_FREE].value.i64 = 0;
1754 		ksd[stat + LGRP_LOADAVG].value.i64 = 0;
1755 	} else if (rw != KSTAT_WRITE) {
1756 		/*
1757 		 * Handle counter stats
1758 		 */
1759 		for (stat = 0; stat < LGRP_NUM_COUNTER_STATS; stat++) {
1760 			ksd[stat].value.i64 = lgrp_stat_read(lgrpid, stat);
1761 		}
1762 
1763 		/*
1764 		 * Handle kernel data snapshot stats
1765 		 */
1766 		ksd[stat + LGRP_NUM_CPUS].value.i64 = lgrp->lgrp_cpucnt;
1767 		ksd[stat + LGRP_NUM_PG_INSTALL].value.i64 =
1768 		    lgrp_mem_size(lgrpid, LGRP_MEM_SIZE_INSTALL);
1769 		ksd[stat + LGRP_NUM_PG_AVAIL].value.i64 =
1770 		    lgrp_mem_size(lgrpid, LGRP_MEM_SIZE_AVAIL);
1771 		ksd[stat + LGRP_NUM_PG_FREE].value.i64 =
1772 		    lgrp_mem_size(lgrpid, LGRP_MEM_SIZE_FREE);
1773 		ksd[stat + LGRP_LOADAVG].value.i64 = lgrp_sum_loadavgs(lgrp);
1774 		ksd[stat + LGRP_LOADAVG_SCALE].value.i64 =
1775 		    lgrp_loadavg_max_effect;
1776 	} else {
1777 		lgrp_kstat_reset(lgrpid);
1778 	}
1779 
1780 	return (0);
1781 }
1782 
1783 int
1784 lgrp_query_cpu(processorid_t id, lgrp_id_t *lp)
1785 {
1786 	cpu_t	*cp;
1787 
1788 	mutex_enter(&cpu_lock);
1789 
1790 	if ((cp = cpu_get(id)) == NULL) {
1791 		mutex_exit(&cpu_lock);
1792 		return (EINVAL);
1793 	}
1794 
1795 	if (cpu_is_offline(cp) || cpu_is_poweredoff(cp)) {
1796 		mutex_exit(&cpu_lock);
1797 		return (EINVAL);
1798 	}
1799 
1800 	ASSERT(cp->cpu_lpl != NULL);
1801 
1802 	*lp = cp->cpu_lpl->lpl_lgrpid;
1803 
1804 	mutex_exit(&cpu_lock);
1805 
1806 	return (0);
1807 }
1808 
1809 int
1810 lgrp_query_load(processorid_t id, lgrp_load_t *lp)
1811 {
1812 	cpu_t *cp;
1813 
1814 	mutex_enter(&cpu_lock);
1815 
1816 	if ((cp = cpu_get(id)) == NULL) {
1817 		mutex_exit(&cpu_lock);
1818 		return (EINVAL);
1819 	}
1820 
1821 	ASSERT(cp->cpu_lpl != NULL);
1822 
1823 	*lp = cp->cpu_lpl->lpl_loadavg;
1824 
1825 	mutex_exit(&cpu_lock);
1826 
1827 	return (0);
1828 }
1829 
1830 /*
1831  * Add a resource named by lpl_leaf to rset of lpl_target
1832  *
1833  * This routine also adjusts ncpu and nrset if the call succeeds in adding a
1834  * resource. It is adjusted here, as this is presently the only place that we
1835  * can be certain a resource addition has succeeded.
1836  *
1837  * We keep the list of rsets sorted so that the dispatcher can quickly walk the
1838  * list in order until it reaches a NULL.  (This list is required to be NULL
1839  * terminated, too).  This is done so that we can mark start pos + 1, so that
1840  * each lpl is traversed sequentially, but in a different order.  We hope this
1841  * will improve performance a bit.  (Hopefully, less read-to-own traffic...)
1842  */
1843 
1844 void
1845 lpl_rset_add(lpl_t *lpl_target, lpl_t *lpl_leaf)
1846 {
1847 	int		i;
1848 	int		entry_slot = 0;
1849 
1850 	/* return if leaf is already present */
1851 	for (i = 0; i < lpl_target->lpl_nrset; i++) {
1852 		if (lpl_target->lpl_rset[i] == lpl_leaf) {
1853 			return;
1854 		}
1855 
1856 		if (lpl_target->lpl_rset[i]->lpl_lgrpid >
1857 		    lpl_leaf->lpl_lgrpid) {
1858 			break;
1859 		}
1860 	}
1861 
1862 	/* insert leaf, update counts */
1863 	entry_slot = i;
1864 	i = lpl_target->lpl_nrset++;
1865 
1866 	/*
1867 	 * Start at the end of the rset array and work backwards towards the
1868 	 * slot into which the new lpl will be inserted. This effectively
1869 	 * preserves the current ordering by scooting everybody over one entry,
1870 	 * and placing the new entry into the space created.
1871 	 */
1872 	while (i-- > entry_slot) {
1873 		lpl_target->lpl_rset[i + 1] = lpl_target->lpl_rset[i];
1874 		lpl_target->lpl_id2rset[lpl_target->lpl_rset[i]->lpl_lgrpid] =
1875 		    i + 1;
1876 	}
1877 
1878 	lpl_target->lpl_rset[entry_slot] = lpl_leaf;
1879 	lpl_target->lpl_id2rset[lpl_leaf->lpl_lgrpid] = entry_slot;
1880 
1881 	lpl_target->lpl_ncpu += lpl_leaf->lpl_ncpu;
1882 }
1883 
1884 /*
1885  * Update each of lpl_parent's children with a reference to their parent.
1886  * The lgrp topology is used as the reference since it is fully
1887  * consistent and correct at this point.
1888  * This should be called after any potential change in lpl_parent's
1889  * rset.
1890  */
1891 static void
1892 lpl_child_update(lpl_t *lpl_parent, struct cpupart *cp)
1893 {
1894 	klgrpset_t	children;
1895 	int		i;
1896 
1897 	children = lgrp_table[lpl_parent->lpl_lgrpid]->lgrp_children;
1898 	if (klgrpset_isempty(children))
1899 		return; /* nothing to do */
1900 
1901 	for (i = 0; i <= lgrp_alloc_max; i++) {
1902 		if (klgrpset_ismember(children, i)) {
1903 			/*
1904 			 * (Re)set the parent. It may be incorrect if
1905 			 * lpl_parent is new in the topology.
1906 			 */
1907 			cp->cp_lgrploads[i].lpl_parent = lpl_parent;
1908 		}
1909 	}
1910 }
1911 
1912 /*
1913  * Delete resource lpl_leaf from rset of lpl_target, assuming it's there.
1914  *
1915  * This routine also adjusts ncpu and nrset if the call succeeds in deleting a
1916  * resource. The values are adjusted here, as this is the only place that we can
1917  * be certain a resource was successfully deleted.
1918  */
1919 void
1920 lpl_rset_del(lpl_t *lpl_target, lpl_t *lpl_leaf)
1921 {
1922 	int i;
1923 	lpl_t *leaf;
1924 
1925 	if (lpl_target->lpl_nrset == 0)
1926 		return;
1927 
1928 	/* find leaf in intermediate node */
1929 	for (i = 0; i < lpl_target->lpl_nrset; i++) {
1930 		if (lpl_target->lpl_rset[i] == lpl_leaf)
1931 			break;
1932 	}
1933 
1934 	/* return if leaf not found */
1935 	if (lpl_target->lpl_rset[i] != lpl_leaf)
1936 		return;
1937 
1938 	/* prune leaf, compress array */
1939 	lpl_target->lpl_rset[lpl_target->lpl_nrset--] = NULL;
1940 	lpl_target->lpl_id2rset[lpl_leaf->lpl_lgrpid] = -1;
1941 	lpl_target->lpl_ncpu--;
1942 	do {
1943 		lpl_target->lpl_rset[i] = lpl_target->lpl_rset[i + 1];
1944 		/*
1945 		 * Update the lgrp id <=> rset mapping
1946 		 */
1947 		if ((leaf = lpl_target->lpl_rset[i]) != NULL) {
1948 			lpl_target->lpl_id2rset[leaf->lpl_lgrpid] = i;
1949 		}
1950 	} while (i++ < lpl_target->lpl_nrset);
1951 }
1952 
1953 /*
1954  * Check to see if the resource set of the target lpl contains the
1955  * supplied leaf lpl.  This returns 1 if the lpl is found, 0 if it is not.
1956  */
1957 
1958 int
1959 lpl_rset_contains(lpl_t *lpl_target, lpl_t *lpl_leaf)
1960 {
1961 	int i;
1962 
1963 	for (i = 0; i < lpl_target->lpl_nrset; i++) {
1964 		if (lpl_target->lpl_rset[i] == lpl_leaf)
1965 			return (1);
1966 	}
1967 
1968 	return (0);
1969 }
1970 
1971 /*
1972  * Called when we change cpu lpl membership.  This increments or decrements the
1973  * per-cpu counter in every lpl in which our leaf appears.
1974  */
1975 void
1976 lpl_cpu_adjcnt(lpl_act_t act, cpu_t *cp)
1977 {
1978 	cpupart_t	*cpupart;
1979 	lgrp_t		*lgrp_leaf;
1980 	lgrp_t		*lgrp_cur;
1981 	lpl_t		*lpl_leaf;
1982 	lpl_t		*lpl_cur;
1983 	int		i;
1984 
1985 	ASSERT(act == LPL_DECREMENT || act == LPL_INCREMENT);
1986 
1987 	cpupart = cp->cpu_part;
1988 	lpl_leaf = cp->cpu_lpl;
1989 	lgrp_leaf = lgrp_table[lpl_leaf->lpl_lgrpid];
1990 
1991 	for (i = 0; i <= lgrp_alloc_max; i++) {
1992 		lgrp_cur = lgrp_table[i];
1993 
1994 		/*
1995 		 * Don't adjust if the lgrp isn't there, if we're the leaf lpl
1996 		 * for the cpu in question, or if the current lgrp and leaf
1997 		 * don't share the same resources.
1998 		 */
1999 
2000 		if (!LGRP_EXISTS(lgrp_cur) || (lgrp_cur == lgrp_leaf) ||
2001 		    !klgrpset_intersects(lgrp_leaf->lgrp_set[LGRP_RSRC_CPU],
2002 		    lgrp_cur->lgrp_set[LGRP_RSRC_CPU]))
2003 			continue;
2004 
2005 
2006 		lpl_cur = &cpupart->cp_lgrploads[lgrp_cur->lgrp_id];
2007 
2008 		if (lpl_cur->lpl_nrset > 0) {
2009 			if (act == LPL_INCREMENT) {
2010 				lpl_cur->lpl_ncpu++;
2011 			} else if (act == LPL_DECREMENT) {
2012 				lpl_cur->lpl_ncpu--;
2013 			}
2014 		}
2015 	}
2016 }
2017 
2018 /*
2019  * Initialize lpl with given resources and specified lgrp
2020  */
2021 void
2022 lpl_init(lpl_t *lpl, lpl_t *lpl_leaf, lgrp_t *lgrp)
2023 {
2024 	lpl->lpl_lgrpid = lgrp->lgrp_id;
2025 	lpl->lpl_loadavg = 0;
2026 	if (lpl == lpl_leaf)
2027 		lpl->lpl_ncpu = 1;
2028 	else
2029 		lpl->lpl_ncpu = lpl_leaf->lpl_ncpu;
2030 	lpl->lpl_nrset = 1;
2031 	lpl->lpl_rset[0] = lpl_leaf;
2032 	lpl->lpl_id2rset[lpl_leaf->lpl_lgrpid] = 0;
2033 	lpl->lpl_lgrp = lgrp;
2034 	lpl->lpl_parent = NULL; /* set by lpl_leaf_insert() */
2035 	lpl->lpl_cpus = NULL; /* set by lgrp_part_add_cpu() */
2036 }
2037 
2038 /*
2039  * Clear an unused lpl
2040  */
2041 void
2042 lpl_clear(lpl_t *lpl)
2043 {
2044 	/*
2045 	 * Clear out all fields in the lpl except:
2046 	 *    lpl_lgrpid - to facilitate debugging
2047 	 *    lpl_rset, lpl_rset_sz, lpl_id2rset - rset array references / size
2048 	 *
2049 	 * Note that the lpl's rset and id2rset mapping are cleared as well.
2050 	 */
2051 	lpl->lpl_loadavg = 0;
2052 	lpl->lpl_ncpu = 0;
2053 	lpl->lpl_lgrp = NULL;
2054 	lpl->lpl_parent = NULL;
2055 	lpl->lpl_cpus = NULL;
2056 	lpl->lpl_nrset = 0;
2057 	lpl->lpl_homed_time = 0;
2058 	bzero(lpl->lpl_rset, sizeof (lpl->lpl_rset[0]) * lpl->lpl_rset_sz);
2059 	bzero(lpl->lpl_id2rset,
2060 	    sizeof (lpl->lpl_id2rset[0]) * lpl->lpl_rset_sz);
2061 }
2062 
2063 /*
2064  * Given a CPU-partition, verify that the lpl topology in the CPU-partition
2065  * is in sync with the lgroup toplogy in the system.  The lpl topology may not
2066  * make full use of all of the lgroup topology, but this checks to make sure
2067  * that for the parts that it does use, it has correctly understood the
2068  * relationships that exist. This function returns
2069  * 0 if the topology is correct, and a non-zero error code, for non-debug
2070  * kernels if incorrect.  Asserts are spread throughout the code to aid in
2071  * debugging on a DEBUG kernel.
2072  */
2073 int
2074 lpl_topo_verify(cpupart_t *cpupart)
2075 {
2076 	lgrp_t		*lgrp;
2077 	lpl_t		*lpl;
2078 	klgrpset_t	rset;
2079 	klgrpset_t	cset;
2080 	cpu_t		*cpu;
2081 	cpu_t		*cp_start;
2082 	int		i;
2083 	int		j;
2084 	int		sum;
2085 
2086 	/* topology can't be incorrect if it doesn't exist */
2087 	if (!lgrp_topo_initialized || !lgrp_initialized)
2088 		return (LPL_TOPO_CORRECT);
2089 
2090 	ASSERT(cpupart != NULL);
2091 
2092 	for (i = 0; i <= lgrp_alloc_max; i++) {
2093 		lgrp = lgrp_table[i];
2094 		lpl = NULL;
2095 		/* make sure lpls are allocated */
2096 		ASSERT(cpupart->cp_lgrploads);
2097 		if (!cpupart->cp_lgrploads)
2098 			return (LPL_TOPO_PART_HAS_NO_LPL);
2099 
2100 		lpl = &cpupart->cp_lgrploads[i];
2101 		/* make sure our index is good */
2102 		ASSERT(i < cpupart->cp_nlgrploads);
2103 
2104 		/* if lgroup doesn't exist, make sure lpl is empty */
2105 		if (!LGRP_EXISTS(lgrp)) {
2106 			ASSERT(lpl->lpl_ncpu == 0);
2107 			if (lpl->lpl_ncpu > 0) {
2108 				return (LPL_TOPO_CPUS_NOT_EMPTY);
2109 			} else {
2110 				continue;
2111 			}
2112 		}
2113 
2114 		/* verify that lgroup and lpl are identically numbered */
2115 		ASSERT(lgrp->lgrp_id == lpl->lpl_lgrpid);
2116 
2117 		/* if lgroup isn't in our partition, make sure lpl is empty */
2118 		if (!klgrpset_intersects(lgrp->lgrp_leaves,
2119 		    cpupart->cp_lgrpset)) {
2120 			ASSERT(lpl->lpl_ncpu == 0);
2121 			if (lpl->lpl_ncpu > 0) {
2122 				return (LPL_TOPO_CPUS_NOT_EMPTY);
2123 			}
2124 			/*
2125 			 * lpl is empty, and lgroup isn't in partition.  verify
2126 			 * that lpl doesn't show up in anyone else's rsets (in
2127 			 * this partition, anyway)
2128 			 */
2129 			for (j = 0; j < cpupart->cp_nlgrploads; j++) {
2130 				lpl_t *i_lpl; /* lpl we're iterating over */
2131 
2132 				i_lpl = &cpupart->cp_lgrploads[j];
2133 
2134 				ASSERT(!lpl_rset_contains(i_lpl, lpl));
2135 				if (lpl_rset_contains(i_lpl, lpl)) {
2136 					return (LPL_TOPO_LPL_ORPHANED);
2137 				}
2138 			}
2139 			/* lgroup is empty, and everything is ok. continue */
2140 			continue;
2141 		}
2142 
2143 
2144 		/* lgroup is in this partition, now check it against lpl */
2145 
2146 		/* do both have matching lgrps? */
2147 		ASSERT(lgrp == lpl->lpl_lgrp);
2148 		if (lgrp != lpl->lpl_lgrp) {
2149 			return (LPL_TOPO_LGRP_MISMATCH);
2150 		}
2151 
2152 		/* do the parent lgroups exist and do they match? */
2153 		if (lgrp->lgrp_parent) {
2154 			ASSERT(lpl->lpl_parent);
2155 			ASSERT(lgrp->lgrp_parent->lgrp_id ==
2156 			    lpl->lpl_parent->lpl_lgrpid);
2157 
2158 			if (!lpl->lpl_parent) {
2159 				return (LPL_TOPO_MISSING_PARENT);
2160 			} else if (lgrp->lgrp_parent->lgrp_id !=
2161 			    lpl->lpl_parent->lpl_lgrpid) {
2162 				return (LPL_TOPO_PARENT_MISMATCH);
2163 			}
2164 		}
2165 
2166 		/* only leaf lgroups keep a cpucnt, only check leaves */
2167 		if ((lpl->lpl_nrset == 1) && (lpl == lpl->lpl_rset[0])) {
2168 
2169 			/* verify that lgrp is also a leaf */
2170 			ASSERT((lgrp->lgrp_childcnt == 0) &&
2171 			    (klgrpset_ismember(lgrp->lgrp_leaves,
2172 			    lpl->lpl_lgrpid)));
2173 
2174 			if ((lgrp->lgrp_childcnt > 0) ||
2175 			    (!klgrpset_ismember(lgrp->lgrp_leaves,
2176 			    lpl->lpl_lgrpid))) {
2177 				return (LPL_TOPO_LGRP_NOT_LEAF);
2178 			}
2179 
2180 			ASSERT((lgrp->lgrp_cpucnt >= lpl->lpl_ncpu) &&
2181 			    (lpl->lpl_ncpu > 0));
2182 			if ((lgrp->lgrp_cpucnt < lpl->lpl_ncpu) ||
2183 			    (lpl->lpl_ncpu <= 0)) {
2184 				return (LPL_TOPO_BAD_CPUCNT);
2185 			}
2186 
2187 			/*
2188 			 * Check that lpl_ncpu also matches the number of
2189 			 * cpus in the lpl's linked list.  This only exists in
2190 			 * leaves, but they should always match.
2191 			 */
2192 			j = 0;
2193 			cpu = cp_start = lpl->lpl_cpus;
2194 			while (cpu != NULL) {
2195 				j++;
2196 
2197 				/* check to make sure cpu's lpl is leaf lpl */
2198 				ASSERT(cpu->cpu_lpl == lpl);
2199 				if (cpu->cpu_lpl != lpl) {
2200 					return (LPL_TOPO_CPU_HAS_BAD_LPL);
2201 				}
2202 
2203 				/* check next cpu */
2204 				if ((cpu = cpu->cpu_next_lpl) != cp_start) {
2205 					continue;
2206 				} else {
2207 					cpu = NULL;
2208 				}
2209 			}
2210 
2211 			ASSERT(j == lpl->lpl_ncpu);
2212 			if (j != lpl->lpl_ncpu) {
2213 				return (LPL_TOPO_LPL_BAD_NCPU);
2214 			}
2215 
2216 			/*
2217 			 * Also, check that leaf lpl is contained in all
2218 			 * intermediate lpls that name the leaf as a descendant
2219 			 */
2220 			for (j = 0; j <= lgrp_alloc_max; j++) {
2221 				klgrpset_t intersect;
2222 				lgrp_t *lgrp_cand;
2223 				lpl_t *lpl_cand;
2224 
2225 				lgrp_cand = lgrp_table[j];
2226 				intersect = klgrpset_intersects(
2227 				    lgrp_cand->lgrp_set[LGRP_RSRC_CPU],
2228 				    cpupart->cp_lgrpset);
2229 
2230 				if (!LGRP_EXISTS(lgrp_cand) ||
2231 				    !klgrpset_intersects(lgrp_cand->lgrp_leaves,
2232 				    cpupart->cp_lgrpset) ||
2233 				    (intersect == 0))
2234 					continue;
2235 
2236 				lpl_cand =
2237 				    &cpupart->cp_lgrploads[lgrp_cand->lgrp_id];
2238 
2239 				if (klgrpset_ismember(intersect,
2240 				    lgrp->lgrp_id)) {
2241 					ASSERT(lpl_rset_contains(lpl_cand,
2242 					    lpl));
2243 
2244 					if (!lpl_rset_contains(lpl_cand, lpl)) {
2245 						return (LPL_TOPO_RSET_MSSNG_LF);
2246 					}
2247 				}
2248 			}
2249 
2250 		} else { /* non-leaf specific checks */
2251 
2252 			/*
2253 			 * Non-leaf lpls should have lpl_cpus == NULL
2254 			 * verify that this is so
2255 			 */
2256 			ASSERT(lpl->lpl_cpus == NULL);
2257 			if (lpl->lpl_cpus != NULL) {
2258 				return (LPL_TOPO_NONLEAF_HAS_CPUS);
2259 			}
2260 
2261 			/*
2262 			 * verify that the sum of the cpus in the leaf resources
2263 			 * is equal to the total ncpu in the intermediate
2264 			 */
2265 			for (j = sum = 0; j < lpl->lpl_nrset; j++) {
2266 				sum += lpl->lpl_rset[j]->lpl_ncpu;
2267 			}
2268 
2269 			ASSERT(sum == lpl->lpl_ncpu);
2270 			if (sum != lpl->lpl_ncpu) {
2271 				return (LPL_TOPO_LPL_BAD_NCPU);
2272 			}
2273 		}
2274 
2275 		/*
2276 		 * Check the rset of the lpl in question.  Make sure that each
2277 		 * rset contains a subset of the resources in
2278 		 * lgrp_set[LGRP_RSRC_CPU] and in cp_lgrpset.  This also makes
2279 		 * sure that each rset doesn't include resources that are
2280 		 * outside of that set.  (Which would be resources somehow not
2281 		 * accounted for).
2282 		 */
2283 		klgrpset_clear(rset);
2284 		for (j = 0; j < lpl->lpl_nrset; j++) {
2285 			klgrpset_add(rset, lpl->lpl_rset[j]->lpl_lgrpid);
2286 		}
2287 		klgrpset_copy(cset, rset);
2288 		/* make sure lpl rset matches lgrp rset */
2289 		klgrpset_diff(rset, lgrp->lgrp_set[LGRP_RSRC_CPU]);
2290 		/* make sure rset is contained with in partition, too */
2291 		klgrpset_diff(cset, cpupart->cp_lgrpset);
2292 
2293 		ASSERT(klgrpset_isempty(rset) && klgrpset_isempty(cset));
2294 		if (!klgrpset_isempty(rset) || !klgrpset_isempty(cset)) {
2295 			return (LPL_TOPO_RSET_MISMATCH);
2296 		}
2297 
2298 		/*
2299 		 * check to make sure lpl_nrset matches the number of rsets
2300 		 * contained in the lpl
2301 		 */
2302 		for (j = 0; j < lpl->lpl_nrset; j++) {
2303 			if (lpl->lpl_rset[j] == NULL)
2304 				break;
2305 		}
2306 
2307 		ASSERT(j == lpl->lpl_nrset);
2308 		if (j != lpl->lpl_nrset) {
2309 			return (LPL_TOPO_BAD_RSETCNT);
2310 		}
2311 
2312 	}
2313 	return (LPL_TOPO_CORRECT);
2314 }
2315 
2316 /*
2317  * Flatten lpl topology to given number of levels.  This is presently only
2318  * implemented for a flatten to 2 levels, which will prune out the intermediates
2319  * and home the leaf lpls to the root lpl.
2320  */
2321 int
2322 lpl_topo_flatten(int levels)
2323 {
2324 	int		i;
2325 	uint_t		sum;
2326 	lgrp_t		*lgrp_cur;
2327 	lpl_t		*lpl_cur;
2328 	lpl_t		*lpl_root;
2329 	cpupart_t	*cp;
2330 
2331 	if (levels != 2)
2332 		return (0);
2333 
2334 	/* called w/ cpus paused - grab no locks! */
2335 	ASSERT(MUTEX_HELD(&cpu_lock) || curthread->t_preempt > 0 ||
2336 	    !lgrp_initialized);
2337 
2338 	cp = cp_list_head;
2339 	do {
2340 		lpl_root = &cp->cp_lgrploads[lgrp_root->lgrp_id];
2341 		ASSERT(LGRP_EXISTS(lgrp_root) && (lpl_root->lpl_ncpu > 0));
2342 
2343 		for (i = 0; i <= lgrp_alloc_max; i++) {
2344 			lgrp_cur = lgrp_table[i];
2345 			lpl_cur = &cp->cp_lgrploads[i];
2346 
2347 			if ((lgrp_cur == lgrp_root) ||
2348 			    (!LGRP_EXISTS(lgrp_cur) &&
2349 			    (lpl_cur->lpl_ncpu == 0)))
2350 				continue;
2351 
2352 			if (!LGRP_EXISTS(lgrp_cur) && (lpl_cur->lpl_ncpu > 0)) {
2353 				/*
2354 				 * this should be a deleted intermediate, so
2355 				 * clear it
2356 				 */
2357 				lpl_clear(lpl_cur);
2358 			} else if ((lpl_cur->lpl_nrset == 1) &&
2359 			    (lpl_cur->lpl_rset[0] == lpl_cur) &&
2360 			    ((lpl_cur->lpl_parent->lpl_ncpu == 0) ||
2361 			    (!LGRP_EXISTS(lpl_cur->lpl_parent->lpl_lgrp)))) {
2362 				/*
2363 				 * this is a leaf whose parent was deleted, or
2364 				 * whose parent had their lgrp deleted.  (And
2365 				 * whose parent will soon be deleted).  Point
2366 				 * this guy back to the root lpl.
2367 				 */
2368 				lpl_cur->lpl_parent = lpl_root;
2369 				lpl_rset_add(lpl_root, lpl_cur);
2370 			}
2371 
2372 		}
2373 
2374 		/*
2375 		 * Now that we're done, make sure the count on the root lpl is
2376 		 * correct, and update the hints of the children for the sake of
2377 		 * thoroughness
2378 		 */
2379 		for (i = sum = 0; i < lpl_root->lpl_nrset; i++) {
2380 			sum += lpl_root->lpl_rset[i]->lpl_ncpu;
2381 		}
2382 		lpl_root->lpl_ncpu = sum;
2383 		lpl_child_update(lpl_root, cp);
2384 
2385 		cp = cp->cp_next;
2386 	} while (cp != cp_list_head);
2387 
2388 	return (levels);
2389 }
2390 
2391 /*
2392  * Insert a lpl into the resource hierarchy and create any additional lpls that
2393  * are necessary to represent the varying states of locality for the cpu
2394  * resoruces newly added to the partition.
2395  *
2396  * This routine is clever enough that it can correctly add resources from the
2397  * new leaf into both direct and indirect resource sets in the hierarchy.  (Ie,
2398  * those for which the lpl is a leaf as opposed to simply a named equally local
2399  * resource).  The one special case that needs additional processing is when a
2400  * new intermediate lpl is introduced.  Since the main loop only traverses
2401  * looking to add the leaf resource where it does not yet exist, additional work
2402  * is necessary to add other leaf resources that may need to exist in the newly
2403  * created intermediate.  This is performed by the second inner loop, and is
2404  * only done when the check for more than one overlapping resource succeeds.
2405  */
2406 
2407 void
2408 lpl_leaf_insert(lpl_t *lpl_leaf, cpupart_t *cpupart)
2409 {
2410 	int		i;
2411 	int		j;
2412 	int		rset_num_intersect;
2413 	lgrp_t		*lgrp_cur;
2414 	lpl_t		*lpl_cur;
2415 	lpl_t		*lpl_parent;
2416 	lgrp_id_t	parent_id;
2417 	klgrpset_t	rset_intersect; /* resources in cpupart and lgrp */
2418 
2419 	for (i = 0; i <= lgrp_alloc_max; i++) {
2420 		lgrp_cur = lgrp_table[i];
2421 
2422 		/*
2423 		 * Don't insert if the lgrp isn't there, if the leaf isn't
2424 		 * contained within the current lgrp, or if the current lgrp has
2425 		 * no leaves in this partition
2426 		 */
2427 
2428 		if (!LGRP_EXISTS(lgrp_cur) ||
2429 		    !klgrpset_ismember(lgrp_cur->lgrp_set[LGRP_RSRC_CPU],
2430 		    lpl_leaf->lpl_lgrpid) ||
2431 		    !klgrpset_intersects(lgrp_cur->lgrp_leaves,
2432 		    cpupart->cp_lgrpset))
2433 			continue;
2434 
2435 		lpl_cur = &cpupart->cp_lgrploads[lgrp_cur->lgrp_id];
2436 		if (lgrp_cur->lgrp_parent != NULL) {
2437 			/* if lgrp has a parent, assign it properly */
2438 			parent_id = lgrp_cur->lgrp_parent->lgrp_id;
2439 			lpl_parent = &cpupart->cp_lgrploads[parent_id];
2440 		} else {
2441 			/* if not, make sure parent ptr gets set to null */
2442 			lpl_parent = NULL;
2443 		}
2444 
2445 		if (lpl_cur == lpl_leaf) {
2446 			/*
2447 			 * Almost all leaf state was initialized elsewhere.  The
2448 			 * only thing left to do is to set the parent.
2449 			 */
2450 			lpl_cur->lpl_parent = lpl_parent;
2451 			continue;
2452 		}
2453 
2454 		lpl_clear(lpl_cur);
2455 		lpl_init(lpl_cur, lpl_leaf, lgrp_cur);
2456 
2457 		lpl_cur->lpl_parent = lpl_parent;
2458 
2459 		/* does new lpl need to be populated with other resources? */
2460 		rset_intersect =
2461 		    klgrpset_intersects(lgrp_cur->lgrp_set[LGRP_RSRC_CPU],
2462 		    cpupart->cp_lgrpset);
2463 		klgrpset_nlgrps(rset_intersect, rset_num_intersect);
2464 
2465 		if (rset_num_intersect > 1) {
2466 			/*
2467 			 * If so, figure out what lpls have resources that
2468 			 * intersect this one, and add them.
2469 			 */
2470 			for (j = 0; j <= lgrp_alloc_max; j++) {
2471 				lgrp_t	*lgrp_cand;	/* candidate lgrp */
2472 				lpl_t	*lpl_cand;	/* candidate lpl */
2473 
2474 				lgrp_cand = lgrp_table[j];
2475 				if (!LGRP_EXISTS(lgrp_cand) ||
2476 				    !klgrpset_ismember(rset_intersect,
2477 				    lgrp_cand->lgrp_id))
2478 					continue;
2479 				lpl_cand =
2480 				    &cpupart->cp_lgrploads[lgrp_cand->lgrp_id];
2481 				lpl_rset_add(lpl_cur, lpl_cand);
2482 			}
2483 		}
2484 		/*
2485 		 * This lpl's rset has changed. Update the hint in it's
2486 		 * children.
2487 		 */
2488 		lpl_child_update(lpl_cur, cpupart);
2489 	}
2490 }
2491 
2492 /*
2493  * remove a lpl from the hierarchy of resources, clearing its state when
2494  * finished.  If the lpls at the intermediate levels of the hierarchy have no
2495  * remaining resources, or no longer name a leaf resource in the cpu-partition,
2496  * delete them as well.
2497  */
2498 
2499 void
2500 lpl_leaf_remove(lpl_t *lpl_leaf, cpupart_t *cpupart)
2501 {
2502 	int		i;
2503 	lgrp_t		*lgrp_cur;
2504 	lpl_t		*lpl_cur;
2505 	klgrpset_t	leaf_intersect;	/* intersection of leaves */
2506 
2507 	for (i = 0; i <= lgrp_alloc_max; i++) {
2508 		lgrp_cur = lgrp_table[i];
2509 
2510 		/*
2511 		 * Don't attempt to remove from lgrps that aren't there, that
2512 		 * don't contain our leaf, or from the leaf itself. (We do that
2513 		 * later)
2514 		 */
2515 
2516 		if (!LGRP_EXISTS(lgrp_cur))
2517 			continue;
2518 
2519 		lpl_cur = &cpupart->cp_lgrploads[lgrp_cur->lgrp_id];
2520 
2521 		if (!klgrpset_ismember(lgrp_cur->lgrp_set[LGRP_RSRC_CPU],
2522 		    lpl_leaf->lpl_lgrpid) ||
2523 		    (lpl_cur == lpl_leaf)) {
2524 			continue;
2525 		}
2526 
2527 		/*
2528 		 * This is a slightly sleazy simplification in that we have
2529 		 * already marked the cp_lgrpset as no longer containing the
2530 		 * leaf we've deleted.  Any lpls that pass the above checks
2531 		 * based upon lgrp membership but not necessarily cpu-part
2532 		 * membership also get cleared by the checks below.  Currently
2533 		 * this is harmless, as the lpls should be empty anyway.
2534 		 *
2535 		 * In particular, we want to preserve lpls that have additional
2536 		 * leaf resources, even though we don't yet have a processor
2537 		 * architecture that represents resources this way.
2538 		 */
2539 
2540 		leaf_intersect = klgrpset_intersects(lgrp_cur->lgrp_leaves,
2541 		    cpupart->cp_lgrpset);
2542 
2543 		lpl_rset_del(lpl_cur, lpl_leaf);
2544 		if ((lpl_cur->lpl_nrset == 0) || (!leaf_intersect)) {
2545 			lpl_clear(lpl_cur);
2546 		} else {
2547 			/*
2548 			 * Update this lpl's children
2549 			 */
2550 			lpl_child_update(lpl_cur, cpupart);
2551 		}
2552 	}
2553 	lpl_clear(lpl_leaf);
2554 }
2555 
2556 /*
2557  * add a cpu to a partition in terms of lgrp load avg bookeeping
2558  *
2559  * The lpl (cpu partition load average information) is now arranged in a
2560  * hierarchical fashion whereby resources that are closest, ie. most local, to
2561  * the cpu in question are considered to be leaves in a tree of resources.
2562  * There are two general cases for cpu additon:
2563  *
2564  * 1. A lpl structure that contains resources already in the hierarchy tree.
2565  * In this case, all of the associated lpl relationships have been defined, and
2566  * all that is necessary is that we link the new cpu into the per-lpl list of
2567  * cpus, and increment the ncpu count of all places where this cpu resource will
2568  * be accounted for.  lpl_cpu_adjcnt updates the cpu count, and the cpu pointer
2569  * pushing is accomplished by this routine.
2570  *
2571  * 2. The lpl to contain the resources in this cpu-partition for this lgrp does
2572  * not exist yet.  In this case, it is necessary to build the leaf lpl, and
2573  * construct the hierarchy of state necessary to name it's more distant
2574  * resources, if they should exist.  The leaf structure is initialized by this
2575  * routine, as is the cpu-partition state for the lgrp membership.  This routine
2576  * also calls lpl_leaf_insert() which inserts the named lpl into the hierarchy
2577  * and builds all of the "ancestoral" state necessary to identify resources at
2578  * differing levels of locality.
2579  */
2580 void
2581 lgrp_part_add_cpu(cpu_t *cp, lgrp_id_t lgrpid)
2582 {
2583 	cpupart_t	*cpupart;
2584 	lgrp_t		*lgrp_leaf;
2585 	lpl_t		*lpl_leaf;
2586 
2587 	/* called sometimes w/ cpus paused - grab no locks */
2588 	ASSERT(MUTEX_HELD(&cpu_lock) || !lgrp_initialized);
2589 
2590 	cpupart = cp->cpu_part;
2591 	lgrp_leaf = lgrp_table[lgrpid];
2592 
2593 	/* don't add non-existent lgrp */
2594 	ASSERT(LGRP_EXISTS(lgrp_leaf));
2595 	lpl_leaf = &cpupart->cp_lgrploads[lgrpid];
2596 	cp->cpu_lpl = lpl_leaf;
2597 
2598 	/* only leaf lpls contain cpus */
2599 
2600 	if (lpl_leaf->lpl_ncpu++ == 0) {
2601 		lpl_init(lpl_leaf, lpl_leaf, lgrp_leaf);
2602 		klgrpset_add(cpupart->cp_lgrpset, lgrpid);
2603 		lpl_leaf_insert(lpl_leaf, cpupart);
2604 	} else {
2605 		/*
2606 		 * the lpl should already exist in the parent, so just update
2607 		 * the count of available CPUs
2608 		 */
2609 		lpl_cpu_adjcnt(LPL_INCREMENT, cp);
2610 	}
2611 
2612 	/* link cpu into list of cpus in lpl */
2613 
2614 	if (lpl_leaf->lpl_cpus) {
2615 		cp->cpu_next_lpl = lpl_leaf->lpl_cpus;
2616 		cp->cpu_prev_lpl = lpl_leaf->lpl_cpus->cpu_prev_lpl;
2617 		lpl_leaf->lpl_cpus->cpu_prev_lpl->cpu_next_lpl = cp;
2618 		lpl_leaf->lpl_cpus->cpu_prev_lpl = cp;
2619 	} else {
2620 		/*
2621 		 * We increment ncpu immediately after we create a new leaf
2622 		 * lpl, so assert that ncpu == 1 for the case where we don't
2623 		 * have any cpu pointers yet.
2624 		 */
2625 		ASSERT(lpl_leaf->lpl_ncpu == 1);
2626 		lpl_leaf->lpl_cpus = cp->cpu_next_lpl = cp->cpu_prev_lpl = cp;
2627 	}
2628 
2629 }
2630 
2631 
2632 /*
2633  * remove a cpu from a partition in terms of lgrp load avg bookeeping
2634  *
2635  * The lpl (cpu partition load average information) is now arranged in a
2636  * hierarchical fashion whereby resources that are closest, ie. most local, to
2637  * the cpu in question are considered to be leaves in a tree of resources.
2638  * There are two removal cases in question:
2639  *
2640  * 1. Removal of the resource in the leaf leaves other resources remaining in
2641  * that leaf.  (Another cpu still exists at this level of locality).  In this
2642  * case, the count of available cpus is decremented in all assocated lpls by
2643  * calling lpl_adj_cpucnt(), and the pointer to the removed cpu is pruned
2644  * from the per-cpu lpl list.
2645  *
2646  * 2. Removal of the resource results in the lpl containing no resources.  (It's
2647  * empty)  In this case, all of what has occurred for the first step must take
2648  * place; however, additionally we must remove the lpl structure itself, prune
2649  * out any stranded lpls that do not directly name a leaf resource, and mark the
2650  * cpu partition in question as no longer containing resources from the lgrp of
2651  * the lpl that has been delted.  Cpu-partition changes are handled by this
2652  * method, but the lpl_leaf_remove function deals with the details of pruning
2653  * out the empty lpl and any of its orphaned direct ancestors.
2654  */
2655 void
2656 lgrp_part_del_cpu(cpu_t *cp)
2657 {
2658 	lpl_t		*lpl;
2659 	lpl_t		*leaf_lpl;
2660 	lgrp_t		*lgrp_leaf;
2661 
2662 	/* called sometimes w/ cpus paused - grab no locks */
2663 
2664 	ASSERT(MUTEX_HELD(&cpu_lock) || !lgrp_initialized);
2665 
2666 	lpl = leaf_lpl = cp->cpu_lpl;
2667 	lgrp_leaf = leaf_lpl->lpl_lgrp;
2668 
2669 	/* don't delete a leaf that isn't there */
2670 	ASSERT(LGRP_EXISTS(lgrp_leaf));
2671 
2672 	/* no double-deletes */
2673 	ASSERT(lpl->lpl_ncpu);
2674 	if (--lpl->lpl_ncpu == 0) {
2675 		/*
2676 		 * This was the last cpu in this lgroup for this partition,
2677 		 * clear its bit in the partition's lgroup bitmask
2678 		 */
2679 		klgrpset_del(cp->cpu_part->cp_lgrpset, lpl->lpl_lgrpid);
2680 
2681 		/* eliminate remaning lpl link pointers in cpu, lpl */
2682 		lpl->lpl_cpus = cp->cpu_next_lpl = cp->cpu_prev_lpl = NULL;
2683 
2684 		lpl_leaf_remove(leaf_lpl, cp->cpu_part);
2685 	} else {
2686 
2687 		/* unlink cpu from lists of cpus in lpl */
2688 		cp->cpu_prev_lpl->cpu_next_lpl = cp->cpu_next_lpl;
2689 		cp->cpu_next_lpl->cpu_prev_lpl = cp->cpu_prev_lpl;
2690 		if (lpl->lpl_cpus == cp) {
2691 			lpl->lpl_cpus = cp->cpu_next_lpl;
2692 		}
2693 
2694 		/*
2695 		 * Update the cpu count in the lpls associated with parent
2696 		 * lgroups.
2697 		 */
2698 		lpl_cpu_adjcnt(LPL_DECREMENT, cp);
2699 
2700 	}
2701 	/* clear cpu's lpl ptr when we're all done */
2702 	cp->cpu_lpl = NULL;
2703 }
2704 
2705 /*
2706  * Recompute load average for the specified partition/lgrp fragment.
2707  *
2708  * We rely on the fact that this routine is called from the clock thread
2709  * at a point before the clock thread can block (i.e. before its first
2710  * lock request).  Since the clock thread can not be preempted (since it
2711  * runs at highest priority), we know that cpu partitions can not change
2712  * (since doing so would require either the repartition requester or the
2713  * cpu_pause thread to run on this cpu), so we can update the cpu's load
2714  * without grabbing cpu_lock.
2715  */
2716 void
2717 lgrp_loadavg(lpl_t *lpl, uint_t nrcpus, int ageflag)
2718 {
2719 	uint_t		ncpu;
2720 	int64_t		old, new, f;
2721 
2722 	/*
2723 	 * 1 - exp(-1/(20 * ncpu)) << 13 = 400 for 1 cpu...
2724 	 */
2725 	static short expval[] = {
2726 	    0, 3196, 1618, 1083,
2727 	    814, 652, 543, 466,
2728 	    408, 363, 326, 297,
2729 	    272, 251, 233, 218,
2730 	    204, 192, 181, 172,
2731 	    163, 155, 148, 142,
2732 	    136, 130, 125, 121,
2733 	    116, 112, 109, 105
2734 	};
2735 
2736 	/* ASSERT (called from clock level) */
2737 
2738 	if ((lpl == NULL) ||	/* we're booting - this is easiest for now */
2739 	    ((ncpu = lpl->lpl_ncpu) == 0)) {
2740 		return;
2741 	}
2742 
2743 	for (;;) {
2744 
2745 		if (ncpu >= sizeof (expval) / sizeof (expval[0]))
2746 			f = expval[1]/ncpu; /* good approx. for large ncpu */
2747 		else
2748 			f = expval[ncpu];
2749 
2750 		/*
2751 		 * Modify the load average atomically to avoid losing
2752 		 * anticipatory load updates (see lgrp_move_thread()).
2753 		 */
2754 		if (ageflag) {
2755 			/*
2756 			 * We're supposed to both update and age the load.
2757 			 * This happens 10 times/sec. per cpu.  We do a
2758 			 * little hoop-jumping to avoid integer overflow.
2759 			 */
2760 			int64_t		q, r;
2761 
2762 			do {
2763 				old = new = lpl->lpl_loadavg;
2764 				q = (old  >> 16) << 7;
2765 				r = (old  & 0xffff) << 7;
2766 				new += ((long long)(nrcpus - q) * f -
2767 				    ((r * f) >> 16)) >> 7;
2768 
2769 				/*
2770 				 * Check for overflow
2771 				 */
2772 				if (new > LGRP_LOADAVG_MAX)
2773 					new = LGRP_LOADAVG_MAX;
2774 				else if (new < 0)
2775 					new = 0;
2776 			} while (atomic_cas_32((lgrp_load_t *)&lpl->lpl_loadavg,
2777 			    old, new) != old);
2778 		} else {
2779 			/*
2780 			 * We're supposed to update the load, but not age it.
2781 			 * This option is used to update the load (which either
2782 			 * has already been aged in this 1/10 sec. interval or
2783 			 * soon will be) to account for a remotely executing
2784 			 * thread.
2785 			 */
2786 			do {
2787 				old = new = lpl->lpl_loadavg;
2788 				new += f;
2789 				/*
2790 				 * Check for overflow
2791 				 * Underflow not possible here
2792 				 */
2793 				if (new < old)
2794 					new = LGRP_LOADAVG_MAX;
2795 			} while (atomic_cas_32((lgrp_load_t *)&lpl->lpl_loadavg,
2796 			    old, new) != old);
2797 		}
2798 
2799 		/*
2800 		 * Do the same for this lpl's parent
2801 		 */
2802 		if ((lpl = lpl->lpl_parent) == NULL)
2803 			break;
2804 		ncpu = lpl->lpl_ncpu;
2805 	}
2806 }
2807 
2808 /*
2809  * Initialize lpl topology in the target based on topology currently present in
2810  * lpl_bootstrap.
2811  *
2812  * lpl_topo_bootstrap is only called once from cpupart_initialize_default() to
2813  * initialize cp_default list of lpls. Up to this point all topology operations
2814  * were performed using lpl_bootstrap. Now cp_default has its own list of lpls
2815  * and all subsequent lpl operations should use it instead of lpl_bootstrap. The
2816  * `target' points to the list of lpls in cp_default and `size' is the size of
2817  * this list.
2818  *
2819  * This function walks the lpl topology in lpl_bootstrap and does for things:
2820  *
2821  * 1) Copies all fields from lpl_bootstrap to the target.
2822  *
2823  * 2) Sets CPU0 lpl pointer to the correct element of the target list.
2824  *
2825  * 3) Updates lpl_parent pointers to point to the lpls in the target list
2826  *    instead of lpl_bootstrap.
2827  *
2828  * 4) Updates pointers in the resource list of the target to point to the lpls
2829  *    in the target list instead of lpl_bootstrap.
2830  *
2831  * After lpl_topo_bootstrap() completes, target contains the same information
2832  * that would be present there if it were used during boot instead of
2833  * lpl_bootstrap. There is no need in information in lpl_bootstrap after this
2834  * and it is bzeroed.
2835  */
2836 void
2837 lpl_topo_bootstrap(lpl_t *target, int size)
2838 {
2839 	lpl_t	*lpl = lpl_bootstrap;
2840 	lpl_t	*target_lpl = target;
2841 	lpl_t	**rset;
2842 	int	*id2rset;
2843 	int	sz;
2844 	int	howmany;
2845 	int	id;
2846 	int	i;
2847 
2848 	/*
2849 	 * The only target that should be passed here is cp_default lpl list.
2850 	 */
2851 	ASSERT(target == cp_default.cp_lgrploads);
2852 	ASSERT(size == cp_default.cp_nlgrploads);
2853 	ASSERT(!lgrp_topo_initialized);
2854 	ASSERT(ncpus == 1);
2855 
2856 	howmany = MIN(LPL_BOOTSTRAP_SIZE, size);
2857 	for (i = 0; i < howmany; i++, lpl++, target_lpl++) {
2858 		/*
2859 		 * Copy all fields from lpl, except for the rset,
2860 		 * lgrp id <=> rset mapping storage,
2861 		 * and amount of storage
2862 		 */
2863 		rset = target_lpl->lpl_rset;
2864 		id2rset = target_lpl->lpl_id2rset;
2865 		sz = target_lpl->lpl_rset_sz;
2866 
2867 		*target_lpl = *lpl;
2868 
2869 		target_lpl->lpl_rset_sz = sz;
2870 		target_lpl->lpl_rset = rset;
2871 		target_lpl->lpl_id2rset = id2rset;
2872 
2873 		/*
2874 		 * Substitute CPU0 lpl pointer with one relative to target.
2875 		 */
2876 		if (lpl->lpl_cpus == CPU) {
2877 			ASSERT(CPU->cpu_lpl == lpl);
2878 			CPU->cpu_lpl = target_lpl;
2879 		}
2880 
2881 		/*
2882 		 * Substitute parent information with parent relative to target.
2883 		 */
2884 		if (lpl->lpl_parent != NULL)
2885 			target_lpl->lpl_parent = (lpl_t *)
2886 			    (((uintptr_t)lpl->lpl_parent -
2887 			    (uintptr_t)lpl_bootstrap) +
2888 			    (uintptr_t)target);
2889 
2890 		/*
2891 		 * Walk over resource set substituting pointers relative to
2892 		 * lpl_bootstrap's rset to pointers relative to target's
2893 		 */
2894 		ASSERT(lpl->lpl_nrset <= 1);
2895 
2896 		for (id = 0; id < lpl->lpl_nrset; id++) {
2897 			if (lpl->lpl_rset[id] != NULL) {
2898 				target_lpl->lpl_rset[id] = (lpl_t *)
2899 				    (((uintptr_t)lpl->lpl_rset[id] -
2900 				    (uintptr_t)lpl_bootstrap) +
2901 				    (uintptr_t)target);
2902 			}
2903 			target_lpl->lpl_id2rset[id] =
2904 			    lpl->lpl_id2rset[id];
2905 		}
2906 	}
2907 
2908 	/*
2909 	 * Clean up the bootstrap lpls since we have switched over to the
2910 	 * actual lpl array in the default cpu partition.
2911 	 *
2912 	 * We still need to keep one empty lpl around for newly starting
2913 	 * slave CPUs to reference should they need to make it through the
2914 	 * dispatcher prior to their lgrp/lpl initialization.
2915 	 *
2916 	 * The lpl related dispatcher code has been designed to work properly
2917 	 * (and without extra checks) for this special case of a zero'ed
2918 	 * bootstrap lpl. Such an lpl appears to the dispatcher as an lpl
2919 	 * with lgrpid 0 and an empty resource set. Iteration over the rset
2920 	 * array by the dispatcher is also NULL terminated for this reason.
2921 	 *
2922 	 * This provides the desired behaviour for an uninitialized CPU.
2923 	 * It shouldn't see any other CPU to either dispatch to or steal
2924 	 * from until it is properly initialized.
2925 	 */
2926 	bzero(lpl_bootstrap_list, sizeof (lpl_bootstrap_list));
2927 	bzero(lpl_bootstrap_id2rset, sizeof (lpl_bootstrap_id2rset));
2928 	bzero(lpl_bootstrap_rset, sizeof (lpl_bootstrap_rset));
2929 
2930 	lpl_bootstrap_list[0].lpl_rset = lpl_bootstrap_rset;
2931 	lpl_bootstrap_list[0].lpl_id2rset = lpl_bootstrap_id2rset;
2932 }
2933 
2934 /*
2935  * If the lowest load among the lgroups a process' threads are currently
2936  * spread across is greater than lgrp_expand_proc_thresh, we'll consider
2937  * expanding the process to a new lgroup.
2938  */
2939 #define	LGRP_EXPAND_PROC_THRESH_DEFAULT 62250
2940 lgrp_load_t	lgrp_expand_proc_thresh = LGRP_EXPAND_PROC_THRESH_DEFAULT;
2941 
2942 #define	LGRP_EXPAND_PROC_THRESH(ncpu) \
2943 	((lgrp_expand_proc_thresh) / (ncpu))
2944 
2945 /*
2946  * A process will be expanded to a new lgroup only if the difference between
2947  * the lowest load on the lgroups the process' thread's are currently spread
2948  * across and the lowest load on the other lgroups in the process' partition
2949  * is greater than lgrp_expand_proc_diff.
2950  */
2951 #define	LGRP_EXPAND_PROC_DIFF_DEFAULT 60000
2952 lgrp_load_t	lgrp_expand_proc_diff = LGRP_EXPAND_PROC_DIFF_DEFAULT;
2953 
2954 #define	LGRP_EXPAND_PROC_DIFF(ncpu) \
2955 	((lgrp_expand_proc_diff) / (ncpu))
2956 
2957 /*
2958  * The loadavg tolerance accounts for "noise" inherent in the load, which may
2959  * be present due to impreciseness of the load average decay algorithm.
2960  *
2961  * The default tolerance is lgrp_loadavg_max_effect. Note that the tunable
2962  * tolerance is scaled by the number of cpus in the lgroup just like
2963  * lgrp_loadavg_max_effect. For example, if lgrp_loadavg_tolerance = 0x10000,
2964  * and ncpu = 4, then lgrp_choose will consider differences in lgroup loads
2965  * of: 0x10000 / 4 => 0x4000 or greater to be significant.
2966  */
2967 uint32_t	lgrp_loadavg_tolerance = LGRP_LOADAVG_THREAD_MAX;
2968 #define	LGRP_LOADAVG_TOLERANCE(ncpu)	\
2969 	((lgrp_loadavg_tolerance) / ncpu)
2970 
2971 /*
2972  * lgrp_choose() will choose root lgroup as home when lowest lgroup load
2973  * average is above this threshold
2974  */
2975 uint32_t	lgrp_load_thresh = UINT32_MAX;
2976 
2977 /*
2978  * lgrp_choose() will try to skip any lgroups with less memory
2979  * than this free when choosing a home lgroup
2980  */
2981 pgcnt_t	lgrp_mem_free_thresh = 0;
2982 
2983 /*
2984  * When choosing between similarly loaded lgroups, lgrp_choose() will pick
2985  * one based on one of the following policies:
2986  * - Random selection
2987  * - Pseudo round robin placement
2988  * - Longest time since a thread was last placed
2989  */
2990 #define	LGRP_CHOOSE_RANDOM	1
2991 #define	LGRP_CHOOSE_RR		2
2992 #define	LGRP_CHOOSE_TIME	3
2993 
2994 int	lgrp_choose_policy = LGRP_CHOOSE_TIME;
2995 
2996 /*
2997  * Choose a suitable leaf lgroup for a kthread.  The kthread is assumed not to
2998  * be bound to a CPU or processor set.
2999  *
3000  * Arguments:
3001  *	t		The thread
3002  *	cpupart		The partition the thread belongs to.
3003  *
3004  * NOTE: Should at least be called with the cpu_lock held, kernel preemption
3005  *	 disabled, or thread_lock held (at splhigh) to protect against the CPU
3006  *	 partitions changing out from under us and assumes that given thread is
3007  *	 protected.  Also, called sometimes w/ cpus paused or kernel preemption
3008  *	 disabled, so don't grab any locks because we should never block under
3009  *	 those conditions.
3010  */
3011 lpl_t *
3012 lgrp_choose(kthread_t *t, cpupart_t *cpupart)
3013 {
3014 	lgrp_load_t	bestload, bestrload;
3015 	int		lgrpid_offset, lgrp_count;
3016 	lgrp_id_t	lgrpid, lgrpid_start;
3017 	lpl_t		*lpl, *bestlpl, *bestrlpl;
3018 	klgrpset_t	lgrpset;
3019 	proc_t		*p;
3020 
3021 	ASSERT(t != NULL);
3022 	ASSERT(MUTEX_HELD(&cpu_lock) || curthread->t_preempt > 0 ||
3023 	    THREAD_LOCK_HELD(t));
3024 	ASSERT(cpupart != NULL);
3025 
3026 	p = t->t_procp;
3027 
3028 	/* A process should always be in an active partition */
3029 	ASSERT(!klgrpset_isempty(cpupart->cp_lgrpset));
3030 
3031 	bestlpl = bestrlpl = NULL;
3032 	bestload = bestrload = LGRP_LOADAVG_MAX;
3033 	lgrpset = cpupart->cp_lgrpset;
3034 
3035 	switch (lgrp_choose_policy) {
3036 	case LGRP_CHOOSE_RR:
3037 		lgrpid = cpupart->cp_lgrp_hint;
3038 		do {
3039 			if (++lgrpid > lgrp_alloc_max)
3040 				lgrpid = 0;
3041 		} while (!klgrpset_ismember(lgrpset, lgrpid));
3042 
3043 		break;
3044 	default:
3045 	case LGRP_CHOOSE_TIME:
3046 	case LGRP_CHOOSE_RANDOM:
3047 		klgrpset_nlgrps(lgrpset, lgrp_count);
3048 		lgrpid_offset =
3049 		    (((ushort_t)(gethrtime() >> 4)) % lgrp_count) + 1;
3050 		for (lgrpid = 0; ; lgrpid++) {
3051 			if (klgrpset_ismember(lgrpset, lgrpid)) {
3052 				if (--lgrpid_offset == 0)
3053 					break;
3054 			}
3055 		}
3056 		break;
3057 	}
3058 
3059 	lgrpid_start = lgrpid;
3060 
3061 	DTRACE_PROBE2(lgrp_choose_start, lgrp_id_t, lgrpid_start,
3062 	    lgrp_id_t, cpupart->cp_lgrp_hint);
3063 
3064 	/*
3065 	 * Use lgroup affinities (if any) to choose best lgroup
3066 	 *
3067 	 * NOTE: Assumes that thread is protected from going away and its
3068 	 *	 lgroup affinities won't change (ie. p_lock, or
3069 	 *	 thread_lock() being held and/or CPUs paused)
3070 	 */
3071 	if (t->t_lgrp_affinity) {
3072 		lpl = lgrp_affinity_best(t, cpupart, lgrpid_start, B_FALSE);
3073 		if (lpl != NULL)
3074 			return (lpl);
3075 	}
3076 
3077 	ASSERT(klgrpset_ismember(lgrpset, lgrpid_start));
3078 
3079 	do {
3080 		pgcnt_t	npgs;
3081 
3082 		/*
3083 		 * Skip any lgroups outside of thread's pset
3084 		 */
3085 		if (!klgrpset_ismember(lgrpset, lgrpid)) {
3086 			if (++lgrpid > lgrp_alloc_max)
3087 				lgrpid = 0;	/* wrap the search */
3088 			continue;
3089 		}
3090 
3091 		/*
3092 		 * Skip any non-leaf lgroups
3093 		 */
3094 		if (lgrp_table[lgrpid]->lgrp_childcnt != 0)
3095 			continue;
3096 
3097 		/*
3098 		 * Skip any lgroups without enough free memory
3099 		 * (when threshold set to nonzero positive value)
3100 		 */
3101 		if (lgrp_mem_free_thresh > 0) {
3102 			npgs = lgrp_mem_size(lgrpid, LGRP_MEM_SIZE_FREE);
3103 			if (npgs < lgrp_mem_free_thresh) {
3104 				if (++lgrpid > lgrp_alloc_max)
3105 					lgrpid = 0;	/* wrap the search */
3106 				continue;
3107 			}
3108 		}
3109 
3110 		lpl = &cpupart->cp_lgrploads[lgrpid];
3111 		if (klgrpset_isempty(p->p_lgrpset) ||
3112 		    klgrpset_ismember(p->p_lgrpset, lgrpid)) {
3113 			/*
3114 			 * Either this is a new process or the process already
3115 			 * has threads on this lgrp, so this is a preferred
3116 			 * lgroup for the thread.
3117 			 */
3118 			if (bestlpl == NULL ||
3119 			    lpl_pick(lpl, bestlpl)) {
3120 				bestload = lpl->lpl_loadavg;
3121 				bestlpl = lpl;
3122 			}
3123 		} else {
3124 			/*
3125 			 * The process doesn't have any threads on this lgrp,
3126 			 * but we're willing to consider this lgrp if the load
3127 			 * difference is big enough to justify splitting up
3128 			 * the process' threads.
3129 			 */
3130 			if (bestrlpl == NULL ||
3131 			    lpl_pick(lpl, bestrlpl)) {
3132 				bestrload = lpl->lpl_loadavg;
3133 				bestrlpl = lpl;
3134 			}
3135 		}
3136 		if (++lgrpid > lgrp_alloc_max)
3137 			lgrpid = 0;	/* wrap the search */
3138 	} while (lgrpid != lgrpid_start);
3139 
3140 	/*
3141 	 * Return root lgroup if threshold isn't set to maximum value and
3142 	 * lowest lgroup load average more than a certain threshold
3143 	 */
3144 	if (lgrp_load_thresh != UINT32_MAX &&
3145 	    bestload >= lgrp_load_thresh && bestrload >= lgrp_load_thresh)
3146 		return (&cpupart->cp_lgrploads[lgrp_root->lgrp_id]);
3147 
3148 	/*
3149 	 * If all the lgroups over which the thread's process is spread are
3150 	 * heavily loaded, or otherwise undesirable, we'll consider placing
3151 	 * the thread on one of the other leaf lgroups in the thread's
3152 	 * partition.
3153 	 */
3154 	if ((bestlpl == NULL) ||
3155 	    ((bestload > LGRP_EXPAND_PROC_THRESH(bestlpl->lpl_ncpu)) &&
3156 	    (bestrload < bestload) &&	/* paranoid about wraparound */
3157 	    (bestrload + LGRP_EXPAND_PROC_DIFF(bestrlpl->lpl_ncpu) <
3158 	    bestload))) {
3159 		bestlpl = bestrlpl;
3160 	}
3161 
3162 	if (bestlpl == NULL) {
3163 		/*
3164 		 * No lgroup looked particularly good, but we still
3165 		 * have to pick something. Go with the randomly selected
3166 		 * legal lgroup we started with above.
3167 		 */
3168 		bestlpl = &cpupart->cp_lgrploads[lgrpid_start];
3169 	}
3170 
3171 	cpupart->cp_lgrp_hint = bestlpl->lpl_lgrpid;
3172 	bestlpl->lpl_homed_time = gethrtime_unscaled();
3173 
3174 	ASSERT(bestlpl->lpl_ncpu > 0);
3175 	return (bestlpl);
3176 }
3177 
3178 /*
3179  * Decide if lpl1 is a better candidate than lpl2 for lgrp homing.
3180  * Returns non-zero if lpl1 is a better candidate, and 0 otherwise.
3181  */
3182 static int
3183 lpl_pick(lpl_t *lpl1, lpl_t *lpl2)
3184 {
3185 	lgrp_load_t	l1, l2;
3186 	lgrp_load_t	tolerance = LGRP_LOADAVG_TOLERANCE(lpl1->lpl_ncpu);
3187 
3188 	l1 = lpl1->lpl_loadavg;
3189 	l2 = lpl2->lpl_loadavg;
3190 
3191 	if ((l1 + tolerance < l2) && (l1 < l2)) {
3192 		/* lpl1 is significantly less loaded than lpl2 */
3193 		return (1);
3194 	}
3195 
3196 	if (lgrp_choose_policy == LGRP_CHOOSE_TIME &&
3197 	    l1 + tolerance >= l2 && l1 < l2 &&
3198 	    lpl1->lpl_homed_time < lpl2->lpl_homed_time) {
3199 		/*
3200 		 * lpl1's load is within the tolerance of lpl2. We're
3201 		 * willing to consider it be to better however if
3202 		 * it has been longer since we last homed a thread there
3203 		 */
3204 		return (1);
3205 	}
3206 
3207 	return (0);
3208 }
3209 
3210 /*
3211  * lgrp_trthr_moves counts the number of times main thread (t_tid = 1) of a
3212  * process that uses text replication changed home lgrp. This info is used by
3213  * segvn asyncronous thread to detect if it needs to recheck what lgrps
3214  * should be used for text replication.
3215  */
3216 static uint64_t lgrp_trthr_moves = 0;
3217 
3218 uint64_t
3219 lgrp_get_trthr_migrations(void)
3220 {
3221 	return (lgrp_trthr_moves);
3222 }
3223 
3224 void
3225 lgrp_update_trthr_migrations(uint64_t incr)
3226 {
3227 	atomic_add_64(&lgrp_trthr_moves, incr);
3228 }
3229 
3230 /*
3231  * An LWP is expected to be assigned to an lgroup for at least this long
3232  * for its anticipatory load to be justified.  NOTE that this value should
3233  * not be set extremely huge (say, larger than 100 years), to avoid problems
3234  * with overflow in the calculation that uses it.
3235  */
3236 #define	LGRP_MIN_NSEC	(NANOSEC / 10)		/* 1/10 of a second */
3237 hrtime_t lgrp_min_nsec = LGRP_MIN_NSEC;
3238 
3239 /*
3240  * Routine to change a thread's lgroup affiliation.  This routine updates
3241  * the thread's kthread_t struct and its process' proc_t struct to note the
3242  * thread's new lgroup affiliation, and its lgroup affinities.
3243  *
3244  * Note that this is the only routine that modifies a thread's t_lpl field,
3245  * and that adds in or removes anticipatory load.
3246  *
3247  * If the thread is exiting, newlpl is NULL.
3248  *
3249  * Locking:
3250  * The following lock must be held on entry:
3251  *	cpu_lock, kpreempt_disable(), or thread_lock -- to assure t's new lgrp
3252  *		doesn't get removed from t's partition
3253  *
3254  * This routine is not allowed to grab any locks, since it may be called
3255  * with cpus paused (such as from cpu_offline).
3256  */
3257 void
3258 lgrp_move_thread(kthread_t *t, lpl_t *newlpl, int do_lgrpset_delete)
3259 {
3260 	proc_t		*p;
3261 	lpl_t		*lpl, *oldlpl;
3262 	lgrp_id_t	oldid;
3263 	kthread_t	*tp;
3264 	uint_t		ncpu;
3265 	lgrp_load_t	old, new;
3266 
3267 	ASSERT(t);
3268 	ASSERT(MUTEX_HELD(&cpu_lock) || curthread->t_preempt > 0 ||
3269 	    THREAD_LOCK_HELD(t));
3270 
3271 	/*
3272 	 * If not changing lpls, just return
3273 	 */
3274 	if ((oldlpl = t->t_lpl) == newlpl)
3275 		return;
3276 
3277 	/*
3278 	 * Make sure the thread's lwp hasn't exited (if so, this thread is now
3279 	 * associated with process 0 rather than with its original process).
3280 	 */
3281 	if (t->t_proc_flag & TP_LWPEXIT) {
3282 		if (newlpl != NULL) {
3283 			t->t_lpl = newlpl;
3284 		}
3285 		return;
3286 	}
3287 
3288 	p = ttoproc(t);
3289 
3290 	/*
3291 	 * If the thread had a previous lgroup, update its process' p_lgrpset
3292 	 * to account for it being moved from its old lgroup.
3293 	 */
3294 	if ((oldlpl != NULL) &&	/* thread had a previous lgroup */
3295 	    (p->p_tlist != NULL)) {
3296 		oldid = oldlpl->lpl_lgrpid;
3297 
3298 		if (newlpl != NULL)
3299 			lgrp_stat_add(oldid, LGRP_NUM_MIGR, 1);
3300 
3301 		if ((do_lgrpset_delete) &&
3302 		    (klgrpset_ismember(p->p_lgrpset, oldid))) {
3303 			for (tp = p->p_tlist->t_forw; ; tp = tp->t_forw) {
3304 				/*
3305 				 * Check if a thread other than the thread
3306 				 * that's moving is assigned to the same
3307 				 * lgroup as the thread that's moving.  Note
3308 				 * that we have to compare lgroup IDs, rather
3309 				 * than simply comparing t_lpl's, since the
3310 				 * threads may belong to different partitions
3311 				 * but be assigned to the same lgroup.
3312 				 */
3313 				ASSERT(tp->t_lpl != NULL);
3314 
3315 				if ((tp != t) &&
3316 				    (tp->t_lpl->lpl_lgrpid == oldid)) {
3317 					/*
3318 					 * Another thread is assigned to the
3319 					 * same lgroup as the thread that's
3320 					 * moving, p_lgrpset doesn't change.
3321 					 */
3322 					break;
3323 				} else if (tp == p->p_tlist) {
3324 					/*
3325 					 * No other thread is assigned to the
3326 					 * same lgroup as the exiting thread,
3327 					 * clear the lgroup's bit in p_lgrpset.
3328 					 */
3329 					klgrpset_del(p->p_lgrpset, oldid);
3330 					break;
3331 				}
3332 			}
3333 		}
3334 
3335 		/*
3336 		 * If this thread was assigned to its old lgroup for such a
3337 		 * short amount of time that the anticipatory load that was
3338 		 * added on its behalf has aged very little, remove that
3339 		 * anticipatory load.
3340 		 */
3341 		if ((t->t_anttime + lgrp_min_nsec > gethrtime()) &&
3342 		    ((ncpu = oldlpl->lpl_ncpu) > 0)) {
3343 			lpl = oldlpl;
3344 			for (;;) {
3345 				do {
3346 					old = new = lpl->lpl_loadavg;
3347 					new -= LGRP_LOADAVG_MAX_EFFECT(ncpu);
3348 					if (new > old) {
3349 						/*
3350 						 * this can happen if the load
3351 						 * average was aged since we
3352 						 * added in the anticipatory
3353 						 * load
3354 						 */
3355 						new = 0;
3356 					}
3357 				} while (atomic_cas_32(
3358 				    (lgrp_load_t *)&lpl->lpl_loadavg, old,
3359 				    new) != old);
3360 
3361 				lpl = lpl->lpl_parent;
3362 				if (lpl == NULL)
3363 					break;
3364 
3365 				ncpu = lpl->lpl_ncpu;
3366 				ASSERT(ncpu > 0);
3367 			}
3368 		}
3369 	}
3370 	/*
3371 	 * If the thread has a new lgroup (i.e. it's not exiting), update its
3372 	 * t_lpl and its process' p_lgrpset, and apply an anticipatory load
3373 	 * to its new lgroup to account for its move to its new lgroup.
3374 	 */
3375 	if (newlpl != NULL) {
3376 		/*
3377 		 * This thread is moving to a new lgroup
3378 		 */
3379 		t->t_lpl = newlpl;
3380 		if (t->t_tid == 1 && p->p_t1_lgrpid != newlpl->lpl_lgrpid) {
3381 			p->p_t1_lgrpid = newlpl->lpl_lgrpid;
3382 			membar_producer();
3383 			if (p->p_tr_lgrpid != LGRP_NONE &&
3384 			    p->p_tr_lgrpid != p->p_t1_lgrpid) {
3385 				lgrp_update_trthr_migrations(1);
3386 			}
3387 		}
3388 
3389 		/*
3390 		 * Reflect move in load average of new lgroup
3391 		 * unless it is root lgroup
3392 		 */
3393 		if (lgrp_table[newlpl->lpl_lgrpid] == lgrp_root)
3394 			return;
3395 
3396 		if (!klgrpset_ismember(p->p_lgrpset, newlpl->lpl_lgrpid)) {
3397 			klgrpset_add(p->p_lgrpset, newlpl->lpl_lgrpid);
3398 		}
3399 
3400 		/*
3401 		 * It'll take some time for the load on the new lgroup
3402 		 * to reflect this thread's placement on it.  We'd
3403 		 * like not, however, to have all threads between now
3404 		 * and then also piling on to this lgroup.  To avoid
3405 		 * this pileup, we anticipate the load this thread
3406 		 * will generate on its new lgroup.  The goal is to
3407 		 * make the lgroup's load appear as though the thread
3408 		 * had been there all along.  We're very conservative
3409 		 * in calculating this anticipatory load, we assume
3410 		 * the worst case case (100% CPU-bound thread).  This
3411 		 * may be modified in the future to be more accurate.
3412 		 */
3413 		lpl = newlpl;
3414 		for (;;) {
3415 			ncpu = lpl->lpl_ncpu;
3416 			ASSERT(ncpu > 0);
3417 			do {
3418 				old = new = lpl->lpl_loadavg;
3419 				new += LGRP_LOADAVG_MAX_EFFECT(ncpu);
3420 				/*
3421 				 * Check for overflow
3422 				 * Underflow not possible here
3423 				 */
3424 				if (new < old)
3425 					new = UINT32_MAX;
3426 			} while (atomic_cas_32((lgrp_load_t *)&lpl->lpl_loadavg,
3427 			    old, new) != old);
3428 
3429 			lpl = lpl->lpl_parent;
3430 			if (lpl == NULL)
3431 				break;
3432 		}
3433 		t->t_anttime = gethrtime();
3434 	}
3435 }
3436 
3437 /*
3438  * Return lgroup memory allocation policy given advice from madvise(3C)
3439  */
3440 lgrp_mem_policy_t
3441 lgrp_madv_to_policy(uchar_t advice, size_t size, int type)
3442 {
3443 	switch (advice) {
3444 	case MADV_ACCESS_LWP:
3445 		return (LGRP_MEM_POLICY_NEXT);
3446 	case MADV_ACCESS_MANY:
3447 		return (LGRP_MEM_POLICY_RANDOM);
3448 	default:
3449 		return (lgrp_mem_policy_default(size, type));
3450 	}
3451 }
3452 
3453 /*
3454  * Figure out default policy
3455  */
3456 lgrp_mem_policy_t
3457 lgrp_mem_policy_default(size_t size, int type)
3458 {
3459 	cpupart_t		*cp;
3460 	lgrp_mem_policy_t	policy;
3461 	size_t			pset_mem_size;
3462 
3463 	/*
3464 	 * Randomly allocate memory across lgroups for shared memory
3465 	 * beyond a certain threshold
3466 	 */
3467 	if ((type != MAP_SHARED && size > lgrp_privm_random_thresh) ||
3468 	    (type == MAP_SHARED && size > lgrp_shm_random_thresh)) {
3469 		/*
3470 		 * Get total memory size of current thread's pset
3471 		 */
3472 		kpreempt_disable();
3473 		cp = curthread->t_cpupart;
3474 		klgrpset_totalsize(cp->cp_lgrpset, pset_mem_size);
3475 		kpreempt_enable();
3476 
3477 		/*
3478 		 * Choose policy to randomly allocate memory across
3479 		 * lgroups in pset if it will fit and is not default
3480 		 * partition.  Otherwise, allocate memory randomly
3481 		 * across machine.
3482 		 */
3483 		if (lgrp_mem_pset_aware && size < pset_mem_size)
3484 			policy = LGRP_MEM_POLICY_RANDOM_PSET;
3485 		else
3486 			policy = LGRP_MEM_POLICY_RANDOM;
3487 	} else
3488 		/*
3489 		 * Apply default policy for private memory and
3490 		 * shared memory under the respective random
3491 		 * threshold.
3492 		 */
3493 		policy = lgrp_mem_default_policy;
3494 
3495 	return (policy);
3496 }
3497 
3498 /*
3499  * Get memory allocation policy for this segment
3500  */
3501 lgrp_mem_policy_info_t *
3502 lgrp_mem_policy_get(struct seg *seg, caddr_t vaddr)
3503 {
3504 	lgrp_mem_policy_info_t	*policy_info;
3505 	extern struct seg_ops	segspt_ops;
3506 	extern struct seg_ops	segspt_shmops;
3507 
3508 	/*
3509 	 * This is for binary compatibility to protect against third party
3510 	 * segment drivers which haven't recompiled to allow for
3511 	 * SEGOP_GETPOLICY()
3512 	 */
3513 	if (seg->s_ops != &segvn_ops && seg->s_ops != &segspt_ops &&
3514 	    seg->s_ops != &segspt_shmops)
3515 		return (NULL);
3516 
3517 	policy_info = NULL;
3518 	if (seg->s_ops->getpolicy != NULL)
3519 		policy_info = SEGOP_GETPOLICY(seg, vaddr);
3520 
3521 	return (policy_info);
3522 }
3523 
3524 /*
3525  * Set policy for allocating private memory given desired policy, policy info,
3526  * size in bytes of memory that policy is being applied.
3527  * Return 0 if policy wasn't set already and 1 if policy was set already
3528  */
3529 int
3530 lgrp_privm_policy_set(lgrp_mem_policy_t policy,
3531     lgrp_mem_policy_info_t *policy_info, size_t size)
3532 {
3533 
3534 	ASSERT(policy_info != NULL);
3535 
3536 	if (policy == LGRP_MEM_POLICY_DEFAULT)
3537 		policy = lgrp_mem_policy_default(size, MAP_PRIVATE);
3538 
3539 	/*
3540 	 * Policy set already?
3541 	 */
3542 	if (policy == policy_info->mem_policy)
3543 		return (1);
3544 
3545 	/*
3546 	 * Set policy
3547 	 */
3548 	policy_info->mem_policy = policy;
3549 	policy_info->mem_lgrpid = LGRP_NONE;
3550 
3551 	return (0);
3552 }
3553 
3554 
3555 /*
3556  * Get shared memory allocation policy with given tree and offset
3557  */
3558 lgrp_mem_policy_info_t *
3559 lgrp_shm_policy_get(struct anon_map *amp, ulong_t anon_index, vnode_t *vp,
3560     u_offset_t vn_off)
3561 {
3562 	u_offset_t		off;
3563 	lgrp_mem_policy_info_t	*policy_info;
3564 	lgrp_shm_policy_seg_t	*policy_seg;
3565 	lgrp_shm_locality_t	*shm_locality;
3566 	avl_tree_t		*tree;
3567 	avl_index_t		where;
3568 
3569 	/*
3570 	 * Get policy segment tree from anon_map or vnode and use specified
3571 	 * anon index or vnode offset as offset
3572 	 *
3573 	 * Assume that no lock needs to be held on anon_map or vnode, since
3574 	 * they should be protected by their reference count which must be
3575 	 * nonzero for an existing segment
3576 	 */
3577 	if (amp) {
3578 		ASSERT(amp->refcnt != 0);
3579 		shm_locality = amp->locality;
3580 		if (shm_locality == NULL)
3581 			return (NULL);
3582 		tree = shm_locality->loc_tree;
3583 		off = ptob(anon_index);
3584 	} else if (vp) {
3585 		shm_locality = vp->v_locality;
3586 		if (shm_locality == NULL)
3587 			return (NULL);
3588 		ASSERT(shm_locality->loc_count != 0);
3589 		tree = shm_locality->loc_tree;
3590 		off = vn_off;
3591 	}
3592 
3593 	if (tree == NULL)
3594 		return (NULL);
3595 
3596 	/*
3597 	 * Lookup policy segment for offset into shared object and return
3598 	 * policy info
3599 	 */
3600 	rw_enter(&shm_locality->loc_lock, RW_READER);
3601 	policy_info = NULL;
3602 	policy_seg = avl_find(tree, &off, &where);
3603 	if (policy_seg)
3604 		policy_info = &policy_seg->shm_policy;
3605 	rw_exit(&shm_locality->loc_lock);
3606 
3607 	return (policy_info);
3608 }
3609 
3610 /*
3611  * Default memory allocation policy for kernel segmap pages
3612  */
3613 lgrp_mem_policy_t	lgrp_segmap_default_policy = LGRP_MEM_POLICY_RANDOM;
3614 
3615 /*
3616  * Return lgroup to use for allocating memory
3617  * given the segment and address
3618  *
3619  * There isn't any mutual exclusion that exists between calls
3620  * to this routine and DR, so this routine and whomever calls it
3621  * should be mindful of the possibility that the lgrp returned
3622  * may be deleted. If this happens, dereferences of the lgrp
3623  * pointer will still be safe, but the resources in the lgrp will
3624  * be gone, and LGRP_EXISTS() will no longer be true.
3625  */
3626 lgrp_t *
3627 lgrp_mem_choose(struct seg *seg, caddr_t vaddr, size_t pgsz)
3628 {
3629 	int			i;
3630 	lgrp_t			*lgrp;
3631 	klgrpset_t		lgrpset;
3632 	int			lgrps_spanned;
3633 	unsigned long		off;
3634 	lgrp_mem_policy_t	policy;
3635 	lgrp_mem_policy_info_t	*policy_info;
3636 	ushort_t		random;
3637 	int			stat = 0;
3638 	extern struct seg	*segkmap;
3639 
3640 	/*
3641 	 * Just return null if the lgrp framework hasn't finished
3642 	 * initializing or if this is a UMA machine.
3643 	 */
3644 	if (nlgrps == 1 || !lgrp_initialized)
3645 		return (lgrp_root);
3646 
3647 	/*
3648 	 * Get memory allocation policy for this segment
3649 	 */
3650 	policy = lgrp_mem_default_policy;
3651 	if (seg != NULL) {
3652 		if (seg->s_as == &kas) {
3653 			if (seg == segkmap)
3654 				policy = lgrp_segmap_default_policy;
3655 			if (policy == LGRP_MEM_POLICY_RANDOM_PROC ||
3656 			    policy == LGRP_MEM_POLICY_RANDOM_PSET)
3657 				policy = LGRP_MEM_POLICY_RANDOM;
3658 		} else {
3659 			policy_info = lgrp_mem_policy_get(seg, vaddr);
3660 			if (policy_info != NULL) {
3661 				policy = policy_info->mem_policy;
3662 				if (policy == LGRP_MEM_POLICY_NEXT_SEG) {
3663 					lgrp_id_t id = policy_info->mem_lgrpid;
3664 					ASSERT(id != LGRP_NONE);
3665 					ASSERT(id < NLGRPS_MAX);
3666 					lgrp = lgrp_table[id];
3667 					if (!LGRP_EXISTS(lgrp)) {
3668 						policy = LGRP_MEM_POLICY_NEXT;
3669 					} else {
3670 						lgrp_stat_add(id,
3671 						    LGRP_NUM_NEXT_SEG, 1);
3672 						return (lgrp);
3673 					}
3674 				}
3675 			}
3676 		}
3677 	}
3678 	lgrpset = 0;
3679 
3680 	/*
3681 	 * Initialize lgroup to home by default
3682 	 */
3683 	lgrp = lgrp_home_lgrp();
3684 
3685 	/*
3686 	 * When homing threads on root lgrp, override default memory
3687 	 * allocation policies with root lgroup memory allocation policy
3688 	 */
3689 	if (lgrp == lgrp_root)
3690 		policy = lgrp_mem_policy_root;
3691 
3692 	/*
3693 	 * Implement policy
3694 	 */
3695 	switch (policy) {
3696 	case LGRP_MEM_POLICY_NEXT_CPU:
3697 
3698 		/*
3699 		 * Return lgroup of current CPU which faulted on memory
3700 		 * If the CPU isn't currently in an lgrp, then opt to
3701 		 * allocate from the root.
3702 		 *
3703 		 * Kernel preemption needs to be disabled here to prevent
3704 		 * the current CPU from going away before lgrp is found.
3705 		 */
3706 		if (LGRP_CPU_HAS_NO_LGRP(CPU)) {
3707 			lgrp = lgrp_root;
3708 		} else {
3709 			kpreempt_disable();
3710 			lgrp = lgrp_cpu_to_lgrp(CPU);
3711 			kpreempt_enable();
3712 		}
3713 		break;
3714 
3715 	case LGRP_MEM_POLICY_NEXT:
3716 	case LGRP_MEM_POLICY_DEFAULT:
3717 	default:
3718 
3719 		/*
3720 		 * Just return current thread's home lgroup
3721 		 * for default policy (next touch)
3722 		 * If the thread is homed to the root,
3723 		 * then the default policy is random across lgroups.
3724 		 * Fallthrough to the random case.
3725 		 */
3726 		if (lgrp != lgrp_root) {
3727 			if (policy == LGRP_MEM_POLICY_NEXT)
3728 				lgrp_stat_add(lgrp->lgrp_id, LGRP_NUM_NEXT, 1);
3729 			else
3730 				lgrp_stat_add(lgrp->lgrp_id,
3731 				    LGRP_NUM_DEFAULT, 1);
3732 			break;
3733 		}
3734 		/* LINTED fallthrough on case statement */
3735 	case LGRP_MEM_POLICY_RANDOM:
3736 
3737 		/*
3738 		 * Return a random leaf lgroup with memory
3739 		 */
3740 		lgrpset = lgrp_root->lgrp_set[LGRP_RSRC_MEM];
3741 		/*
3742 		 * Count how many lgroups are spanned
3743 		 */
3744 		klgrpset_nlgrps(lgrpset, lgrps_spanned);
3745 
3746 		/*
3747 		 * There may be no memnodes in the root lgroup during DR copy
3748 		 * rename on a system with only two boards (memnodes)
3749 		 * configured. In this case just return the root lgrp.
3750 		 */
3751 		if (lgrps_spanned == 0) {
3752 			lgrp = lgrp_root;
3753 			break;
3754 		}
3755 
3756 		/*
3757 		 * Pick a random offset within lgroups spanned
3758 		 * and return lgroup at that offset
3759 		 */
3760 		random = (ushort_t)gethrtime() >> 4;
3761 		off = random % lgrps_spanned;
3762 		ASSERT(off <= lgrp_alloc_max);
3763 
3764 		for (i = 0; i <= lgrp_alloc_max; i++) {
3765 			if (!klgrpset_ismember(lgrpset, i))
3766 				continue;
3767 			if (off)
3768 				off--;
3769 			else {
3770 				lgrp = lgrp_table[i];
3771 				lgrp_stat_add(lgrp->lgrp_id, LGRP_NUM_RANDOM,
3772 				    1);
3773 				break;
3774 			}
3775 		}
3776 		break;
3777 
3778 	case LGRP_MEM_POLICY_RANDOM_PROC:
3779 
3780 		/*
3781 		 * Grab copy of bitmask of lgroups spanned by
3782 		 * this process
3783 		 */
3784 		klgrpset_copy(lgrpset, curproc->p_lgrpset);
3785 		stat = LGRP_NUM_RANDOM_PROC;
3786 
3787 		/* LINTED fallthrough on case statement */
3788 	case LGRP_MEM_POLICY_RANDOM_PSET:
3789 
3790 		if (!stat)
3791 			stat = LGRP_NUM_RANDOM_PSET;
3792 
3793 		if (klgrpset_isempty(lgrpset)) {
3794 			/*
3795 			 * Grab copy of bitmask of lgroups spanned by
3796 			 * this processor set
3797 			 */
3798 			kpreempt_disable();
3799 			klgrpset_copy(lgrpset,
3800 			    curthread->t_cpupart->cp_lgrpset);
3801 			kpreempt_enable();
3802 		}
3803 
3804 		/*
3805 		 * Count how many lgroups are spanned
3806 		 */
3807 		klgrpset_nlgrps(lgrpset, lgrps_spanned);
3808 		ASSERT(lgrps_spanned <= nlgrps);
3809 
3810 		/*
3811 		 * Probably lgrps_spanned should be always non-zero, but to be
3812 		 * on the safe side we return lgrp_root if it is empty.
3813 		 */
3814 		if (lgrps_spanned == 0) {
3815 			lgrp = lgrp_root;
3816 			break;
3817 		}
3818 
3819 		/*
3820 		 * Pick a random offset within lgroups spanned
3821 		 * and return lgroup at that offset
3822 		 */
3823 		random = (ushort_t)gethrtime() >> 4;
3824 		off = random % lgrps_spanned;
3825 		ASSERT(off <= lgrp_alloc_max);
3826 
3827 		for (i = 0; i <= lgrp_alloc_max; i++) {
3828 			if (!klgrpset_ismember(lgrpset, i))
3829 				continue;
3830 			if (off)
3831 				off--;
3832 			else {
3833 				lgrp = lgrp_table[i];
3834 				lgrp_stat_add(lgrp->lgrp_id, LGRP_NUM_RANDOM,
3835 				    1);
3836 				break;
3837 			}
3838 		}
3839 		break;
3840 
3841 	case LGRP_MEM_POLICY_ROUNDROBIN:
3842 
3843 		/*
3844 		 * Use offset within segment to determine
3845 		 * offset from home lgroup to choose for
3846 		 * next lgroup to allocate memory from
3847 		 */
3848 		off = ((unsigned long)(vaddr - seg->s_base) / pgsz) %
3849 		    (lgrp_alloc_max + 1);
3850 
3851 		kpreempt_disable();
3852 		lgrpset = lgrp_root->lgrp_set[LGRP_RSRC_MEM];
3853 		i = lgrp->lgrp_id;
3854 		kpreempt_enable();
3855 
3856 		while (off > 0) {
3857 			i = (i + 1) % (lgrp_alloc_max + 1);
3858 			lgrp = lgrp_table[i];
3859 			if (klgrpset_ismember(lgrpset, i))
3860 				off--;
3861 		}
3862 		lgrp_stat_add(lgrp->lgrp_id, LGRP_NUM_ROUNDROBIN, 1);
3863 
3864 		break;
3865 	}
3866 
3867 	ASSERT(lgrp != NULL);
3868 	return (lgrp);
3869 }
3870 
3871 /*
3872  * Return the number of pages in an lgroup
3873  *
3874  * NOTE: NUMA test (numat) driver uses this, so changing arguments or semantics
3875  *	 could cause tests that rely on the numat driver to fail....
3876  */
3877 pgcnt_t
3878 lgrp_mem_size(lgrp_id_t lgrpid, lgrp_mem_query_t query)
3879 {
3880 	lgrp_t *lgrp;
3881 
3882 	lgrp = lgrp_table[lgrpid];
3883 	if (!LGRP_EXISTS(lgrp) ||
3884 	    klgrpset_isempty(lgrp->lgrp_set[LGRP_RSRC_MEM]) ||
3885 	    !klgrpset_ismember(lgrp->lgrp_set[LGRP_RSRC_MEM], lgrpid))
3886 		return (0);
3887 
3888 	return (lgrp_plat_mem_size(lgrp->lgrp_plathand, query));
3889 }
3890 
3891 /*
3892  * Initialize lgroup shared memory allocation policy support
3893  */
3894 void
3895 lgrp_shm_policy_init(struct anon_map *amp, vnode_t *vp)
3896 {
3897 	lgrp_shm_locality_t	*shm_locality;
3898 
3899 	/*
3900 	 * Initialize locality field in anon_map
3901 	 * Don't need any locks because this is called when anon_map is
3902 	 * allocated, but not used anywhere yet.
3903 	 */
3904 	if (amp) {
3905 		ANON_LOCK_ENTER(&amp->a_rwlock, RW_WRITER);
3906 		if (amp->locality == NULL) {
3907 			/*
3908 			 * Allocate and initialize shared memory locality info
3909 			 * and set anon_map locality pointer to it
3910 			 * Drop lock across kmem_alloc(KM_SLEEP)
3911 			 */
3912 			ANON_LOCK_EXIT(&amp->a_rwlock);
3913 			shm_locality = kmem_alloc(sizeof (*shm_locality),
3914 			    KM_SLEEP);
3915 			rw_init(&shm_locality->loc_lock, NULL, RW_DEFAULT,
3916 			    NULL);
3917 			shm_locality->loc_count = 1;	/* not used for amp */
3918 			shm_locality->loc_tree = NULL;
3919 
3920 			/*
3921 			 * Reacquire lock and check to see whether anyone beat
3922 			 * us to initializing the locality info
3923 			 */
3924 			ANON_LOCK_ENTER(&amp->a_rwlock, RW_WRITER);
3925 			if (amp->locality != NULL) {
3926 				rw_destroy(&shm_locality->loc_lock);
3927 				kmem_free(shm_locality,
3928 				    sizeof (*shm_locality));
3929 			} else
3930 				amp->locality = shm_locality;
3931 		}
3932 		ANON_LOCK_EXIT(&amp->a_rwlock);
3933 		return;
3934 	}
3935 
3936 	/*
3937 	 * Allocate shared vnode policy info if vnode is not locality aware yet
3938 	 */
3939 	mutex_enter(&vp->v_lock);
3940 	if ((vp->v_flag & V_LOCALITY) == 0) {
3941 		/*
3942 		 * Allocate and initialize shared memory locality info
3943 		 */
3944 		mutex_exit(&vp->v_lock);
3945 		shm_locality = kmem_alloc(sizeof (*shm_locality), KM_SLEEP);
3946 		rw_init(&shm_locality->loc_lock, NULL, RW_DEFAULT, NULL);
3947 		shm_locality->loc_count = 1;
3948 		shm_locality->loc_tree = NULL;
3949 
3950 		/*
3951 		 * Point vnode locality field at shared vnode policy info
3952 		 * and set locality aware flag in vnode
3953 		 */
3954 		mutex_enter(&vp->v_lock);
3955 		if ((vp->v_flag & V_LOCALITY) == 0) {
3956 			vp->v_locality = shm_locality;
3957 			vp->v_flag |= V_LOCALITY;
3958 		} else {
3959 			/*
3960 			 * Lost race so free locality info and increment count.
3961 			 */
3962 			rw_destroy(&shm_locality->loc_lock);
3963 			kmem_free(shm_locality, sizeof (*shm_locality));
3964 			shm_locality = vp->v_locality;
3965 			shm_locality->loc_count++;
3966 		}
3967 		mutex_exit(&vp->v_lock);
3968 
3969 		return;
3970 	}
3971 
3972 	/*
3973 	 * Increment reference count of number of segments mapping this vnode
3974 	 * shared
3975 	 */
3976 	shm_locality = vp->v_locality;
3977 	shm_locality->loc_count++;
3978 	mutex_exit(&vp->v_lock);
3979 }
3980 
3981 /*
3982  * Destroy the given shared memory policy segment tree
3983  */
3984 void
3985 lgrp_shm_policy_tree_destroy(avl_tree_t *tree)
3986 {
3987 	lgrp_shm_policy_seg_t	*cur;
3988 	lgrp_shm_policy_seg_t	*next;
3989 
3990 	if (tree == NULL)
3991 		return;
3992 
3993 	cur = (lgrp_shm_policy_seg_t *)avl_first(tree);
3994 	while (cur != NULL) {
3995 		next = AVL_NEXT(tree, cur);
3996 		avl_remove(tree, cur);
3997 		kmem_free(cur, sizeof (*cur));
3998 		cur = next;
3999 	}
4000 	kmem_free(tree, sizeof (avl_tree_t));
4001 }
4002 
4003 /*
4004  * Uninitialize lgroup shared memory allocation policy support
4005  */
4006 void
4007 lgrp_shm_policy_fini(struct anon_map *amp, vnode_t *vp)
4008 {
4009 	lgrp_shm_locality_t	*shm_locality;
4010 
4011 	/*
4012 	 * For anon_map, deallocate shared memory policy tree and
4013 	 * zero locality field
4014 	 * Don't need any locks because anon_map is being freed
4015 	 */
4016 	if (amp) {
4017 		if (amp->locality == NULL)
4018 			return;
4019 		shm_locality = amp->locality;
4020 		shm_locality->loc_count = 0;	/* not really used for amp */
4021 		rw_destroy(&shm_locality->loc_lock);
4022 		lgrp_shm_policy_tree_destroy(shm_locality->loc_tree);
4023 		kmem_free(shm_locality, sizeof (*shm_locality));
4024 		amp->locality = 0;
4025 		return;
4026 	}
4027 
4028 	/*
4029 	 * For vnode, decrement reference count of segments mapping this vnode
4030 	 * shared and delete locality info if reference count drops to 0
4031 	 */
4032 	mutex_enter(&vp->v_lock);
4033 	shm_locality = vp->v_locality;
4034 	shm_locality->loc_count--;
4035 
4036 	if (shm_locality->loc_count == 0) {
4037 		rw_destroy(&shm_locality->loc_lock);
4038 		lgrp_shm_policy_tree_destroy(shm_locality->loc_tree);
4039 		kmem_free(shm_locality, sizeof (*shm_locality));
4040 		vp->v_locality = 0;
4041 		vp->v_flag &= ~V_LOCALITY;
4042 	}
4043 	mutex_exit(&vp->v_lock);
4044 }
4045 
4046 /*
4047  * Compare two shared memory policy segments
4048  * Used by AVL tree code for searching
4049  */
4050 int
4051 lgrp_shm_policy_compar(const void *x, const void *y)
4052 {
4053 	lgrp_shm_policy_seg_t *a = (lgrp_shm_policy_seg_t *)x;
4054 	lgrp_shm_policy_seg_t *b = (lgrp_shm_policy_seg_t *)y;
4055 
4056 	if (a->shm_off < b->shm_off)
4057 		return (-1);
4058 	if (a->shm_off >= b->shm_off + b->shm_size)
4059 		return (1);
4060 	return (0);
4061 }
4062 
4063 /*
4064  * Concatenate seg1 with seg2 and remove seg2
4065  */
4066 static int
4067 lgrp_shm_policy_concat(avl_tree_t *tree, lgrp_shm_policy_seg_t *seg1,
4068     lgrp_shm_policy_seg_t *seg2)
4069 {
4070 	if (!seg1 || !seg2 ||
4071 	    seg1->shm_off + seg1->shm_size != seg2->shm_off ||
4072 	    seg1->shm_policy.mem_policy != seg2->shm_policy.mem_policy)
4073 		return (-1);
4074 
4075 	seg1->shm_size += seg2->shm_size;
4076 	avl_remove(tree, seg2);
4077 	kmem_free(seg2, sizeof (*seg2));
4078 	return (0);
4079 }
4080 
4081 /*
4082  * Split segment at given offset and return rightmost (uppermost) segment
4083  * Assumes that there are no overlapping segments
4084  */
4085 static lgrp_shm_policy_seg_t *
4086 lgrp_shm_policy_split(avl_tree_t *tree, lgrp_shm_policy_seg_t *seg,
4087     u_offset_t off)
4088 {
4089 	lgrp_shm_policy_seg_t	*newseg;
4090 	avl_index_t		where;
4091 
4092 	ASSERT(seg != NULL);
4093 	ASSERT(off >= seg->shm_off && off <= seg->shm_off + seg->shm_size);
4094 
4095 	if (!seg || off < seg->shm_off || off > seg->shm_off +
4096 	    seg->shm_size)
4097 		return (NULL);
4098 
4099 	if (off == seg->shm_off || off == seg->shm_off + seg->shm_size)
4100 		return (seg);
4101 
4102 	/*
4103 	 * Adjust size of left segment and allocate new (right) segment
4104 	 */
4105 	newseg = kmem_alloc(sizeof (lgrp_shm_policy_seg_t), KM_SLEEP);
4106 	newseg->shm_policy = seg->shm_policy;
4107 	newseg->shm_off = off;
4108 	newseg->shm_size = seg->shm_size - (off - seg->shm_off);
4109 	seg->shm_size = off - seg->shm_off;
4110 
4111 	/*
4112 	 * Find where to insert new segment in AVL tree and insert it
4113 	 */
4114 	(void) avl_find(tree, &off, &where);
4115 	avl_insert(tree, newseg, where);
4116 
4117 	return (newseg);
4118 }
4119 
4120 /*
4121  * Set shared memory allocation policy on specified shared object at given
4122  * offset and length
4123  *
4124  * Return 0 if policy wasn't set already, 1 if policy was set already, and
4125  * -1 if can't set policy.
4126  */
4127 int
4128 lgrp_shm_policy_set(lgrp_mem_policy_t policy, struct anon_map *amp,
4129     ulong_t anon_index, vnode_t *vp, u_offset_t vn_off, size_t len)
4130 {
4131 	u_offset_t		eoff;
4132 	lgrp_shm_policy_seg_t	*next;
4133 	lgrp_shm_policy_seg_t	*newseg;
4134 	u_offset_t		off;
4135 	u_offset_t		oldeoff;
4136 	lgrp_shm_policy_seg_t	*prev;
4137 	int			retval;
4138 	lgrp_shm_policy_seg_t	*seg;
4139 	lgrp_shm_locality_t	*shm_locality;
4140 	avl_tree_t		*tree;
4141 	avl_index_t		where;
4142 
4143 	ASSERT(amp || vp);
4144 	ASSERT((len & PAGEOFFSET) == 0);
4145 
4146 	if (len == 0)
4147 		return (-1);
4148 
4149 	retval = 0;
4150 
4151 	/*
4152 	 * Get locality info and starting offset into shared object
4153 	 * Try anon map first and then vnode
4154 	 * Assume that no locks need to be held on anon_map or vnode, since
4155 	 * it should be protected by its reference count which must be nonzero
4156 	 * for an existing segment.
4157 	 */
4158 	if (amp) {
4159 		/*
4160 		 * Get policy info from anon_map
4161 		 *
4162 		 */
4163 		ASSERT(amp->refcnt != 0);
4164 		if (amp->locality == NULL)
4165 			lgrp_shm_policy_init(amp, NULL);
4166 		shm_locality = amp->locality;
4167 		off = ptob(anon_index);
4168 	} else if (vp) {
4169 		/*
4170 		 * Get policy info from vnode
4171 		 */
4172 		if ((vp->v_flag & V_LOCALITY) == 0 || vp->v_locality == NULL)
4173 			lgrp_shm_policy_init(NULL, vp);
4174 		shm_locality = vp->v_locality;
4175 		ASSERT(shm_locality->loc_count != 0);
4176 		off = vn_off;
4177 	} else
4178 		return (-1);
4179 
4180 	ASSERT((off & PAGEOFFSET) == 0);
4181 
4182 	/*
4183 	 * Figure out default policy
4184 	 */
4185 	if (policy == LGRP_MEM_POLICY_DEFAULT)
4186 		policy = lgrp_mem_policy_default(len, MAP_SHARED);
4187 
4188 	/*
4189 	 * Create AVL tree if there isn't one yet
4190 	 * and set locality field to point at it
4191 	 */
4192 	rw_enter(&shm_locality->loc_lock, RW_WRITER);
4193 	tree = shm_locality->loc_tree;
4194 	if (!tree) {
4195 		rw_exit(&shm_locality->loc_lock);
4196 
4197 		tree = kmem_alloc(sizeof (avl_tree_t), KM_SLEEP);
4198 
4199 		rw_enter(&shm_locality->loc_lock, RW_WRITER);
4200 		if (shm_locality->loc_tree == NULL) {
4201 			avl_create(tree, lgrp_shm_policy_compar,
4202 			    sizeof (lgrp_shm_policy_seg_t),
4203 			    offsetof(lgrp_shm_policy_seg_t, shm_tree));
4204 			shm_locality->loc_tree = tree;
4205 		} else {
4206 			/*
4207 			 * Another thread managed to set up the tree
4208 			 * before we could. Free the tree we allocated
4209 			 * and use the one that's already there.
4210 			 */
4211 			kmem_free(tree, sizeof (*tree));
4212 			tree = shm_locality->loc_tree;
4213 		}
4214 	}
4215 
4216 	/*
4217 	 * Set policy
4218 	 *
4219 	 * Need to maintain hold on writer's lock to keep tree from
4220 	 * changing out from under us
4221 	 */
4222 	while (len != 0) {
4223 		/*
4224 		 * Find policy segment for specified offset into shared object
4225 		 */
4226 		seg = avl_find(tree, &off, &where);
4227 
4228 		/*
4229 		 * Didn't find any existing segment that contains specified
4230 		 * offset, so allocate new segment, insert it, and concatenate
4231 		 * with adjacent segments if possible
4232 		 */
4233 		if (seg == NULL) {
4234 			newseg = kmem_alloc(sizeof (lgrp_shm_policy_seg_t),
4235 			    KM_SLEEP);
4236 			newseg->shm_policy.mem_policy = policy;
4237 			newseg->shm_policy.mem_lgrpid = LGRP_NONE;
4238 			newseg->shm_off = off;
4239 			avl_insert(tree, newseg, where);
4240 
4241 			/*
4242 			 * Check to see whether new segment overlaps with next
4243 			 * one, set length of new segment accordingly, and
4244 			 * calculate remaining length and next offset
4245 			 */
4246 			seg = AVL_NEXT(tree, newseg);
4247 			if (seg == NULL || off + len <= seg->shm_off) {
4248 				newseg->shm_size = len;
4249 				len = 0;
4250 			} else {
4251 				newseg->shm_size = seg->shm_off - off;
4252 				off = seg->shm_off;
4253 				len -= newseg->shm_size;
4254 			}
4255 
4256 			/*
4257 			 * Try to concatenate new segment with next and
4258 			 * previous ones, since they might have the same policy
4259 			 * now.  Grab previous and next segments first because
4260 			 * they will change on concatenation.
4261 			 */
4262 			prev =  AVL_PREV(tree, newseg);
4263 			next = AVL_NEXT(tree, newseg);
4264 			(void) lgrp_shm_policy_concat(tree, newseg, next);
4265 			(void) lgrp_shm_policy_concat(tree, prev, newseg);
4266 
4267 			continue;
4268 		}
4269 
4270 		eoff = off + len;
4271 		oldeoff = seg->shm_off + seg->shm_size;
4272 
4273 		/*
4274 		 * Policy set already?
4275 		 */
4276 		if (policy == seg->shm_policy.mem_policy) {
4277 			/*
4278 			 * Nothing left to do if offset and length
4279 			 * fall within this segment
4280 			 */
4281 			if (eoff <= oldeoff) {
4282 				retval = 1;
4283 				break;
4284 			} else {
4285 				len = eoff - oldeoff;
4286 				off = oldeoff;
4287 				continue;
4288 			}
4289 		}
4290 
4291 		/*
4292 		 * Specified offset and length match existing segment exactly
4293 		 */
4294 		if (off == seg->shm_off && len == seg->shm_size) {
4295 			/*
4296 			 * Set policy and update current length
4297 			 */
4298 			seg->shm_policy.mem_policy = policy;
4299 			seg->shm_policy.mem_lgrpid = LGRP_NONE;
4300 			len = 0;
4301 
4302 			/*
4303 			 * Try concatenating new segment with previous and next
4304 			 * segments, since they might have the same policy now.
4305 			 * Grab previous and next segments first because they
4306 			 * will change on concatenation.
4307 			 */
4308 			prev =  AVL_PREV(tree, seg);
4309 			next = AVL_NEXT(tree, seg);
4310 			(void) lgrp_shm_policy_concat(tree, seg, next);
4311 			(void) lgrp_shm_policy_concat(tree, prev, seg);
4312 		} else {
4313 			/*
4314 			 * Specified offset and length only apply to part of
4315 			 * existing segment
4316 			 */
4317 
4318 			/*
4319 			 * New segment starts in middle of old one, so split
4320 			 * new one off near beginning of old one
4321 			 */
4322 			newseg = NULL;
4323 			if (off > seg->shm_off) {
4324 				newseg = lgrp_shm_policy_split(tree, seg, off);
4325 
4326 				/*
4327 				 * New segment ends where old one did, so try
4328 				 * to concatenate with next segment
4329 				 */
4330 				if (eoff == oldeoff) {
4331 					newseg->shm_policy.mem_policy = policy;
4332 					newseg->shm_policy.mem_lgrpid =
4333 					    LGRP_NONE;
4334 					(void) lgrp_shm_policy_concat(tree,
4335 					    newseg, AVL_NEXT(tree, newseg));
4336 					break;
4337 				}
4338 			}
4339 
4340 			/*
4341 			 * New segment ends before old one, so split off end of
4342 			 * old one
4343 			 */
4344 			if (eoff < oldeoff) {
4345 				if (newseg) {
4346 					(void) lgrp_shm_policy_split(tree,
4347 					    newseg, eoff);
4348 					newseg->shm_policy.mem_policy = policy;
4349 					newseg->shm_policy.mem_lgrpid =
4350 					    LGRP_NONE;
4351 				} else {
4352 					(void) lgrp_shm_policy_split(tree, seg,
4353 					    eoff);
4354 					seg->shm_policy.mem_policy = policy;
4355 					seg->shm_policy.mem_lgrpid = LGRP_NONE;
4356 				}
4357 
4358 				if (off == seg->shm_off)
4359 					(void) lgrp_shm_policy_concat(tree,
4360 					    AVL_PREV(tree, seg), seg);
4361 				break;
4362 			}
4363 
4364 			/*
4365 			 * Calculate remaining length and next offset
4366 			 */
4367 			len = eoff - oldeoff;
4368 			off = oldeoff;
4369 		}
4370 	}
4371 
4372 	rw_exit(&shm_locality->loc_lock);
4373 	return (retval);
4374 }
4375 
4376 /*
4377  * Return the best memnode from which to allocate memory given
4378  * an lgroup.
4379  *
4380  * "c" is for cookie, which is good enough for me.
4381  * It references a cookie struct that should be zero'ed to initialize.
4382  * The cookie should live on the caller's stack.
4383  *
4384  * The routine returns -1 when:
4385  *	- traverse is 0, and all the memnodes in "lgrp" have been returned.
4386  *	- traverse is 1, and all the memnodes in the system have been
4387  *	  returned.
4388  */
4389 int
4390 lgrp_memnode_choose(lgrp_mnode_cookie_t *c)
4391 {
4392 	lgrp_t		*lp = c->lmc_lgrp;
4393 	mnodeset_t	nodes = c->lmc_nodes;
4394 	int		cnt = c->lmc_cnt;
4395 	int		offset, mnode;
4396 
4397 	extern int	max_mem_nodes;
4398 
4399 	/*
4400 	 * If the set is empty, and the caller is willing, traverse
4401 	 * up the hierarchy until we find a non-empty set.
4402 	 */
4403 	while (nodes == (mnodeset_t)0 || cnt <= 0) {
4404 		if (c->lmc_scope == LGRP_SRCH_LOCAL ||
4405 		    ((lp = lp->lgrp_parent) == NULL))
4406 			return (-1);
4407 
4408 		nodes = lp->lgrp_mnodes & ~(c->lmc_tried);
4409 		cnt = lp->lgrp_nmnodes - c->lmc_ntried;
4410 	}
4411 
4412 	/*
4413 	 * Select a memnode by picking one at a "random" offset.
4414 	 * Because of DR, memnodes can come and go at any time.
4415 	 * This code must be able to cope with the possibility
4416 	 * that the nodes count "cnt" is inconsistent with respect
4417 	 * to the number of elements actually in "nodes", and
4418 	 * therefore that the offset chosen could be greater than
4419 	 * the number of elements in the set (some memnodes may
4420 	 * have dissapeared just before cnt was read).
4421 	 * If this happens, the search simply wraps back to the
4422 	 * beginning of the set.
4423 	 */
4424 	ASSERT(nodes != (mnodeset_t)0 && cnt > 0);
4425 	offset = c->lmc_rand % cnt;
4426 	do {
4427 		for (mnode = 0; mnode < max_mem_nodes; mnode++)
4428 			if (nodes & ((mnodeset_t)1 << mnode))
4429 				if (!offset--)
4430 					break;
4431 	} while (mnode >= max_mem_nodes);
4432 
4433 	/* Found a node. Store state before returning. */
4434 	c->lmc_lgrp = lp;
4435 	c->lmc_nodes = (nodes & ~((mnodeset_t)1 << mnode));
4436 	c->lmc_cnt = cnt - 1;
4437 	c->lmc_tried = (c->lmc_tried | ((mnodeset_t)1 << mnode));
4438 	c->lmc_ntried++;
4439 
4440 	return (mnode);
4441 }
4442