17c478bd9Sstevel@tonic-gate /* 27c478bd9Sstevel@tonic-gate * CDDL HEADER START 37c478bd9Sstevel@tonic-gate * 47c478bd9Sstevel@tonic-gate * The contents of this file are subject to the terms of the 5e824d57fSjohnlev * Common Development and Distribution License (the "License"). 6e824d57fSjohnlev * You may not use this file except in compliance with the License. 77c478bd9Sstevel@tonic-gate * 87c478bd9Sstevel@tonic-gate * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 97c478bd9Sstevel@tonic-gate * or http://www.opensolaris.org/os/licensing. 107c478bd9Sstevel@tonic-gate * See the License for the specific language governing permissions 117c478bd9Sstevel@tonic-gate * and limitations under the License. 127c478bd9Sstevel@tonic-gate * 137c478bd9Sstevel@tonic-gate * When distributing Covered Code, include this CDDL HEADER in each 147c478bd9Sstevel@tonic-gate * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 157c478bd9Sstevel@tonic-gate * If applicable, add the following below this CDDL HEADER, with the 167c478bd9Sstevel@tonic-gate * fields enclosed by brackets "[]" replaced with your own identifying 177c478bd9Sstevel@tonic-gate * information: Portions Copyright [yyyy] [name of copyright owner] 187c478bd9Sstevel@tonic-gate * 197c478bd9Sstevel@tonic-gate * CDDL HEADER END 207c478bd9Sstevel@tonic-gate */ 217c478bd9Sstevel@tonic-gate /* 220542eecfSRafael Vanoni * Copyright (c) 1996, 2010, Oracle and/or its affiliates. All rights reserved. 237c478bd9Sstevel@tonic-gate */ 247c478bd9Sstevel@tonic-gate 257c478bd9Sstevel@tonic-gate #include <sys/types.h> 267c478bd9Sstevel@tonic-gate #include <sys/systm.h> 277c478bd9Sstevel@tonic-gate #include <sys/cmn_err.h> 287c478bd9Sstevel@tonic-gate #include <sys/cpuvar.h> 297c478bd9Sstevel@tonic-gate #include <sys/thread.h> 307c478bd9Sstevel@tonic-gate #include <sys/disp.h> 317c478bd9Sstevel@tonic-gate #include <sys/kmem.h> 327c478bd9Sstevel@tonic-gate #include <sys/debug.h> 337c478bd9Sstevel@tonic-gate #include <sys/cpupart.h> 347c478bd9Sstevel@tonic-gate #include <sys/pset.h> 357c478bd9Sstevel@tonic-gate #include <sys/var.h> 367c478bd9Sstevel@tonic-gate #include <sys/cyclic.h> 377c478bd9Sstevel@tonic-gate #include <sys/lgrp.h> 38fb2f18f8Sesaxe #include <sys/pghw.h> 397c478bd9Sstevel@tonic-gate #include <sys/loadavg.h> 407c478bd9Sstevel@tonic-gate #include <sys/class.h> 417c478bd9Sstevel@tonic-gate #include <sys/fss.h> 427c478bd9Sstevel@tonic-gate #include <sys/pool.h> 437c478bd9Sstevel@tonic-gate #include <sys/pool_pset.h> 447c478bd9Sstevel@tonic-gate #include <sys/policy.h> 457c478bd9Sstevel@tonic-gate 467c478bd9Sstevel@tonic-gate /* 477c478bd9Sstevel@tonic-gate * Calling pool_lock() protects the pools configuration, which includes 487c478bd9Sstevel@tonic-gate * CPU partitions. cpu_lock protects the CPU partition list, and prevents 497c478bd9Sstevel@tonic-gate * partitions from being created or destroyed while the lock is held. 507c478bd9Sstevel@tonic-gate * The lock ordering with respect to related locks is: 517c478bd9Sstevel@tonic-gate * 527c478bd9Sstevel@tonic-gate * pool_lock() ---> cpu_lock ---> pidlock --> p_lock 537c478bd9Sstevel@tonic-gate * 547c478bd9Sstevel@tonic-gate * Blocking memory allocations may be made while holding "pool_lock" 557c478bd9Sstevel@tonic-gate * or cpu_lock. 567c478bd9Sstevel@tonic-gate */ 577c478bd9Sstevel@tonic-gate 587c478bd9Sstevel@tonic-gate /* 597c478bd9Sstevel@tonic-gate * The cp_default partition is allocated statically, but its lgroup load average 607c478bd9Sstevel@tonic-gate * (lpl) list is allocated dynamically after kmem subsystem is initialized. This 617c478bd9Sstevel@tonic-gate * saves some memory since the space allocated reflects the actual number of 627c478bd9Sstevel@tonic-gate * lgroups supported by the platform. The lgrp facility provides a temporary 637c478bd9Sstevel@tonic-gate * space to hold lpl information during system bootstrap. 647c478bd9Sstevel@tonic-gate */ 657c478bd9Sstevel@tonic-gate 667c478bd9Sstevel@tonic-gate cpupart_t *cp_list_head; 677c478bd9Sstevel@tonic-gate cpupart_t cp_default; 687c478bd9Sstevel@tonic-gate static cpupartid_t cp_id_next; 697c478bd9Sstevel@tonic-gate uint_t cp_numparts; 707c478bd9Sstevel@tonic-gate uint_t cp_numparts_nonempty; 717c478bd9Sstevel@tonic-gate 727c478bd9Sstevel@tonic-gate /* 737c478bd9Sstevel@tonic-gate * Need to limit total number of partitions to avoid slowing down the 747c478bd9Sstevel@tonic-gate * clock code too much. The clock code traverses the list of 757c478bd9Sstevel@tonic-gate * partitions and needs to be able to execute in a reasonable amount 767c478bd9Sstevel@tonic-gate * of time (less than 1/hz seconds). The maximum is sized based on 777c478bd9Sstevel@tonic-gate * max_ncpus so it shouldn't be a problem unless there are large 787c478bd9Sstevel@tonic-gate * numbers of empty partitions. 797c478bd9Sstevel@tonic-gate */ 807c478bd9Sstevel@tonic-gate static uint_t cp_max_numparts; 817c478bd9Sstevel@tonic-gate 827c478bd9Sstevel@tonic-gate /* 837c478bd9Sstevel@tonic-gate * Processor sets and CPU partitions are different but related concepts. 847c478bd9Sstevel@tonic-gate * A processor set is a user-level abstraction allowing users to create 857c478bd9Sstevel@tonic-gate * sets of CPUs and bind threads exclusively to those sets. A CPU 867c478bd9Sstevel@tonic-gate * partition is a kernel dispatcher object consisting of a set of CPUs 877c478bd9Sstevel@tonic-gate * and a global dispatch queue. The processor set abstraction is 887c478bd9Sstevel@tonic-gate * implemented via a CPU partition, and currently there is a 1-1 897c478bd9Sstevel@tonic-gate * mapping between processor sets and partitions (excluding the default 907c478bd9Sstevel@tonic-gate * partition, which is not visible as a processor set). Hence, the 917c478bd9Sstevel@tonic-gate * numbering for processor sets and CPU partitions is identical. This 927c478bd9Sstevel@tonic-gate * may not always be true in the future, and these macros could become 937c478bd9Sstevel@tonic-gate * less trivial if we support e.g. a processor set containing multiple 947c478bd9Sstevel@tonic-gate * CPU partitions. 957c478bd9Sstevel@tonic-gate */ 967c478bd9Sstevel@tonic-gate #define PSTOCP(psid) ((cpupartid_t)((psid) == PS_NONE ? CP_DEFAULT : (psid))) 977c478bd9Sstevel@tonic-gate #define CPTOPS(cpid) ((psetid_t)((cpid) == CP_DEFAULT ? PS_NONE : (cpid))) 987c478bd9Sstevel@tonic-gate 990b70c467Sakolb static int cpupart_unbind_threads(cpupart_t *, boolean_t); 1000b70c467Sakolb 1017c478bd9Sstevel@tonic-gate /* 1027c478bd9Sstevel@tonic-gate * Find a CPU partition given a processor set ID. 1037c478bd9Sstevel@tonic-gate */ 1047c478bd9Sstevel@tonic-gate static cpupart_t * 1057c478bd9Sstevel@tonic-gate cpupart_find_all(psetid_t psid) 1067c478bd9Sstevel@tonic-gate { 1077c478bd9Sstevel@tonic-gate cpupart_t *cp; 1087c478bd9Sstevel@tonic-gate cpupartid_t cpid = PSTOCP(psid); 1097c478bd9Sstevel@tonic-gate 1107c478bd9Sstevel@tonic-gate ASSERT(MUTEX_HELD(&cpu_lock)); 1117c478bd9Sstevel@tonic-gate 1127c478bd9Sstevel@tonic-gate /* default partition not visible as a processor set */ 1137c478bd9Sstevel@tonic-gate if (psid == CP_DEFAULT) 1147c478bd9Sstevel@tonic-gate return (NULL); 1157c478bd9Sstevel@tonic-gate 1167c478bd9Sstevel@tonic-gate if (psid == PS_MYID) 1177c478bd9Sstevel@tonic-gate return (curthread->t_cpupart); 1187c478bd9Sstevel@tonic-gate 1197c478bd9Sstevel@tonic-gate cp = cp_list_head; 1207c478bd9Sstevel@tonic-gate do { 1217c478bd9Sstevel@tonic-gate if (cp->cp_id == cpid) 1227c478bd9Sstevel@tonic-gate return (cp); 1237c478bd9Sstevel@tonic-gate cp = cp->cp_next; 1247c478bd9Sstevel@tonic-gate } while (cp != cp_list_head); 1257c478bd9Sstevel@tonic-gate return (NULL); 1267c478bd9Sstevel@tonic-gate } 1277c478bd9Sstevel@tonic-gate 1287c478bd9Sstevel@tonic-gate /* 1297c478bd9Sstevel@tonic-gate * Find a CPU partition given a processor set ID if the processor set 1307c478bd9Sstevel@tonic-gate * should be visible from the calling zone. 1317c478bd9Sstevel@tonic-gate */ 1327c478bd9Sstevel@tonic-gate cpupart_t * 1337c478bd9Sstevel@tonic-gate cpupart_find(psetid_t psid) 1347c478bd9Sstevel@tonic-gate { 1357c478bd9Sstevel@tonic-gate cpupart_t *cp; 1367c478bd9Sstevel@tonic-gate 1377c478bd9Sstevel@tonic-gate ASSERT(MUTEX_HELD(&cpu_lock)); 1387c478bd9Sstevel@tonic-gate cp = cpupart_find_all(psid); 1397c478bd9Sstevel@tonic-gate if (cp != NULL && !INGLOBALZONE(curproc) && pool_pset_enabled() && 1407c478bd9Sstevel@tonic-gate zone_pset_get(curproc->p_zone) != CPTOPS(cp->cp_id)) 1417c478bd9Sstevel@tonic-gate return (NULL); 1427c478bd9Sstevel@tonic-gate return (cp); 1437c478bd9Sstevel@tonic-gate } 1447c478bd9Sstevel@tonic-gate 1457c478bd9Sstevel@tonic-gate static int 1467c478bd9Sstevel@tonic-gate cpupart_kstat_update(kstat_t *ksp, int rw) 1477c478bd9Sstevel@tonic-gate { 1487c478bd9Sstevel@tonic-gate cpupart_t *cp = (cpupart_t *)ksp->ks_private; 1497c478bd9Sstevel@tonic-gate cpupart_kstat_t *cpksp = ksp->ks_data; 1507c478bd9Sstevel@tonic-gate 1517c478bd9Sstevel@tonic-gate if (rw == KSTAT_WRITE) 1527c478bd9Sstevel@tonic-gate return (EACCES); 1537c478bd9Sstevel@tonic-gate 1547c478bd9Sstevel@tonic-gate cpksp->cpk_updates.value.ui64 = cp->cp_updates; 1557c478bd9Sstevel@tonic-gate cpksp->cpk_runnable.value.ui64 = cp->cp_nrunnable_cum; 1567c478bd9Sstevel@tonic-gate cpksp->cpk_waiting.value.ui64 = cp->cp_nwaiting_cum; 1577c478bd9Sstevel@tonic-gate cpksp->cpk_ncpus.value.ui32 = cp->cp_ncpus; 1587c478bd9Sstevel@tonic-gate cpksp->cpk_avenrun_1min.value.ui32 = cp->cp_hp_avenrun[0] >> 1597c478bd9Sstevel@tonic-gate (16 - FSHIFT); 1607c478bd9Sstevel@tonic-gate cpksp->cpk_avenrun_5min.value.ui32 = cp->cp_hp_avenrun[1] >> 1617c478bd9Sstevel@tonic-gate (16 - FSHIFT); 1627c478bd9Sstevel@tonic-gate cpksp->cpk_avenrun_15min.value.ui32 = cp->cp_hp_avenrun[2] >> 1637c478bd9Sstevel@tonic-gate (16 - FSHIFT); 1647c478bd9Sstevel@tonic-gate return (0); 1657c478bd9Sstevel@tonic-gate } 1667c478bd9Sstevel@tonic-gate 1677c478bd9Sstevel@tonic-gate static void 1687c478bd9Sstevel@tonic-gate cpupart_kstat_create(cpupart_t *cp) 1697c478bd9Sstevel@tonic-gate { 1707c478bd9Sstevel@tonic-gate kstat_t *ksp; 1717c478bd9Sstevel@tonic-gate zoneid_t zoneid; 1727c478bd9Sstevel@tonic-gate 1737c478bd9Sstevel@tonic-gate ASSERT(MUTEX_HELD(&cpu_lock)); 1747c478bd9Sstevel@tonic-gate 1757c478bd9Sstevel@tonic-gate /* 1767c478bd9Sstevel@tonic-gate * We have a bit of a chicken-egg problem since this code will 1777c478bd9Sstevel@tonic-gate * get called to create the kstats for CP_DEFAULT before the 1787c478bd9Sstevel@tonic-gate * pools framework gets initialized. We circumvent the problem 1797c478bd9Sstevel@tonic-gate * by special-casing cp_default. 1807c478bd9Sstevel@tonic-gate */ 1817c478bd9Sstevel@tonic-gate if (cp != &cp_default && pool_pset_enabled()) 1827c478bd9Sstevel@tonic-gate zoneid = GLOBAL_ZONEID; 1837c478bd9Sstevel@tonic-gate else 1847c478bd9Sstevel@tonic-gate zoneid = ALL_ZONES; 1857c478bd9Sstevel@tonic-gate ksp = kstat_create_zone("unix", cp->cp_id, "pset", "misc", 1867c478bd9Sstevel@tonic-gate KSTAT_TYPE_NAMED, 1877c478bd9Sstevel@tonic-gate sizeof (cpupart_kstat_t) / sizeof (kstat_named_t), 0, zoneid); 1887c478bd9Sstevel@tonic-gate if (ksp != NULL) { 1897c478bd9Sstevel@tonic-gate cpupart_kstat_t *cpksp = ksp->ks_data; 1907c478bd9Sstevel@tonic-gate 1917c478bd9Sstevel@tonic-gate kstat_named_init(&cpksp->cpk_updates, "updates", 1927c478bd9Sstevel@tonic-gate KSTAT_DATA_UINT64); 1937c478bd9Sstevel@tonic-gate kstat_named_init(&cpksp->cpk_runnable, "runnable", 1947c478bd9Sstevel@tonic-gate KSTAT_DATA_UINT64); 1957c478bd9Sstevel@tonic-gate kstat_named_init(&cpksp->cpk_waiting, "waiting", 1967c478bd9Sstevel@tonic-gate KSTAT_DATA_UINT64); 1977c478bd9Sstevel@tonic-gate kstat_named_init(&cpksp->cpk_ncpus, "ncpus", 1987c478bd9Sstevel@tonic-gate KSTAT_DATA_UINT32); 1997c478bd9Sstevel@tonic-gate kstat_named_init(&cpksp->cpk_avenrun_1min, "avenrun_1min", 2007c478bd9Sstevel@tonic-gate KSTAT_DATA_UINT32); 2017c478bd9Sstevel@tonic-gate kstat_named_init(&cpksp->cpk_avenrun_5min, "avenrun_5min", 2027c478bd9Sstevel@tonic-gate KSTAT_DATA_UINT32); 2037c478bd9Sstevel@tonic-gate kstat_named_init(&cpksp->cpk_avenrun_15min, "avenrun_15min", 2047c478bd9Sstevel@tonic-gate KSTAT_DATA_UINT32); 2057c478bd9Sstevel@tonic-gate 2067c478bd9Sstevel@tonic-gate ksp->ks_update = cpupart_kstat_update; 2077c478bd9Sstevel@tonic-gate ksp->ks_private = cp; 2087c478bd9Sstevel@tonic-gate 2097c478bd9Sstevel@tonic-gate kstat_install(ksp); 2107c478bd9Sstevel@tonic-gate } 2117c478bd9Sstevel@tonic-gate cp->cp_kstat = ksp; 2127c478bd9Sstevel@tonic-gate } 2137c478bd9Sstevel@tonic-gate 2147c478bd9Sstevel@tonic-gate /* 2156890d023SEric Saxe * Initialize the cpupart's lgrp partions (lpls) 2166890d023SEric Saxe */ 2176890d023SEric Saxe static void 2186890d023SEric Saxe cpupart_lpl_initialize(cpupart_t *cp) 2196890d023SEric Saxe { 2206890d023SEric Saxe int i, sz; 2216890d023SEric Saxe 2226890d023SEric Saxe sz = cp->cp_nlgrploads = lgrp_plat_max_lgrps(); 2236890d023SEric Saxe cp->cp_lgrploads = kmem_zalloc(sizeof (lpl_t) * sz, KM_SLEEP); 2246890d023SEric Saxe 2256890d023SEric Saxe for (i = 0; i < sz; i++) { 2266890d023SEric Saxe /* 2276890d023SEric Saxe * The last entry of the lpl's resource set is always NULL 2286890d023SEric Saxe * by design (to facilitate iteration)...hence the "oversizing" 2296890d023SEric Saxe * by 1. 2306890d023SEric Saxe */ 2316890d023SEric Saxe cp->cp_lgrploads[i].lpl_rset_sz = sz + 1; 2326890d023SEric Saxe cp->cp_lgrploads[i].lpl_rset = 2336890d023SEric Saxe kmem_zalloc(sizeof (struct lgrp_ld *) * (sz + 1), KM_SLEEP); 2346890d023SEric Saxe cp->cp_lgrploads[i].lpl_id2rset = 2356890d023SEric Saxe kmem_zalloc(sizeof (int) * (sz + 1), KM_SLEEP); 2366890d023SEric Saxe cp->cp_lgrploads[i].lpl_lgrpid = i; 2376890d023SEric Saxe } 2386890d023SEric Saxe } 2396890d023SEric Saxe 2406890d023SEric Saxe /* 2416890d023SEric Saxe * Teardown the cpupart's lgrp partitions 2426890d023SEric Saxe */ 2436890d023SEric Saxe static void 2446890d023SEric Saxe cpupart_lpl_teardown(cpupart_t *cp) 2456890d023SEric Saxe { 2466890d023SEric Saxe int i, sz; 2476890d023SEric Saxe lpl_t *lpl; 2486890d023SEric Saxe 2496890d023SEric Saxe for (i = 0; i < cp->cp_nlgrploads; i++) { 2506890d023SEric Saxe lpl = &cp->cp_lgrploads[i]; 2516890d023SEric Saxe 2526890d023SEric Saxe sz = lpl->lpl_rset_sz; 2536890d023SEric Saxe kmem_free(lpl->lpl_rset, sizeof (struct lgrp_ld *) * sz); 2546890d023SEric Saxe kmem_free(lpl->lpl_id2rset, sizeof (int) * sz); 2556890d023SEric Saxe lpl->lpl_rset = NULL; 2566890d023SEric Saxe lpl->lpl_id2rset = NULL; 2576890d023SEric Saxe } 2586890d023SEric Saxe kmem_free(cp->cp_lgrploads, sizeof (lpl_t) * cp->cp_nlgrploads); 2596890d023SEric Saxe cp->cp_lgrploads = NULL; 2606890d023SEric Saxe } 2616890d023SEric Saxe 2626890d023SEric Saxe /* 2637c478bd9Sstevel@tonic-gate * Initialize the default partition and kpreempt disp queue. 2647c478bd9Sstevel@tonic-gate */ 2657c478bd9Sstevel@tonic-gate void 2667c478bd9Sstevel@tonic-gate cpupart_initialize_default(void) 2677c478bd9Sstevel@tonic-gate { 2687c478bd9Sstevel@tonic-gate lgrp_id_t i; 2697c478bd9Sstevel@tonic-gate 2707c478bd9Sstevel@tonic-gate cp_list_head = &cp_default; 2717c478bd9Sstevel@tonic-gate cp_default.cp_next = &cp_default; 2727c478bd9Sstevel@tonic-gate cp_default.cp_prev = &cp_default; 2737c478bd9Sstevel@tonic-gate cp_default.cp_id = CP_DEFAULT; 2747c478bd9Sstevel@tonic-gate cp_default.cp_kp_queue.disp_maxrunpri = -1; 2757c478bd9Sstevel@tonic-gate cp_default.cp_kp_queue.disp_max_unbound_pri = -1; 2767c478bd9Sstevel@tonic-gate cp_default.cp_kp_queue.disp_cpu = NULL; 2777c478bd9Sstevel@tonic-gate cp_default.cp_gen = 0; 2787c478bd9Sstevel@tonic-gate cp_default.cp_loadavg.lg_cur = 0; 2797c478bd9Sstevel@tonic-gate cp_default.cp_loadavg.lg_len = 0; 2807c478bd9Sstevel@tonic-gate cp_default.cp_loadavg.lg_total = 0; 2817c478bd9Sstevel@tonic-gate for (i = 0; i < S_LOADAVG_SZ; i++) { 2827c478bd9Sstevel@tonic-gate cp_default.cp_loadavg.lg_loads[i] = 0; 2837c478bd9Sstevel@tonic-gate } 2847c478bd9Sstevel@tonic-gate DISP_LOCK_INIT(&cp_default.cp_kp_queue.disp_lock); 2857c478bd9Sstevel@tonic-gate cp_id_next = CP_DEFAULT + 1; 2867c478bd9Sstevel@tonic-gate cpupart_kstat_create(&cp_default); 2877c478bd9Sstevel@tonic-gate cp_numparts = 1; 2887c478bd9Sstevel@tonic-gate if (cp_max_numparts == 0) /* allow for /etc/system tuning */ 2897c478bd9Sstevel@tonic-gate cp_max_numparts = max_ncpus * 2 + 1; 2907c478bd9Sstevel@tonic-gate /* 2917c478bd9Sstevel@tonic-gate * Allocate space for cp_default list of lgrploads 2927c478bd9Sstevel@tonic-gate */ 2936890d023SEric Saxe cpupart_lpl_initialize(&cp_default); 2947c478bd9Sstevel@tonic-gate 2957c478bd9Sstevel@tonic-gate /* 2967c478bd9Sstevel@tonic-gate * The initial lpl topology is created in a special lpl list 2977c478bd9Sstevel@tonic-gate * lpl_bootstrap. It should be copied to cp_default. 2987c478bd9Sstevel@tonic-gate * NOTE: lpl_topo_bootstrap() also updates CPU0 cpu_lpl pointer to point 2997c478bd9Sstevel@tonic-gate * to the correct lpl in the cp_default.cp_lgrploads list. 3007c478bd9Sstevel@tonic-gate */ 3017c478bd9Sstevel@tonic-gate lpl_topo_bootstrap(cp_default.cp_lgrploads, 3027c478bd9Sstevel@tonic-gate cp_default.cp_nlgrploads); 3037c478bd9Sstevel@tonic-gate 3046890d023SEric Saxe 3057c478bd9Sstevel@tonic-gate cp_default.cp_attr = PSET_NOESCAPE; 3067c478bd9Sstevel@tonic-gate cp_numparts_nonempty = 1; 3077c478bd9Sstevel@tonic-gate /* 3087c478bd9Sstevel@tonic-gate * Set t0's home 3097c478bd9Sstevel@tonic-gate */ 3107c478bd9Sstevel@tonic-gate t0.t_lpl = &cp_default.cp_lgrploads[LGRP_ROOTID]; 311fb2f18f8Sesaxe 312fb2f18f8Sesaxe bitset_init(&cp_default.cp_cmt_pgs); 3130542eecfSRafael Vanoni bitset_init_fanout(&cp_default.cp_haltset, cp_haltset_fanout); 3140542eecfSRafael Vanoni 3156890d023SEric Saxe bitset_resize(&cp_default.cp_haltset, max_ncpus); 3167c478bd9Sstevel@tonic-gate } 3177c478bd9Sstevel@tonic-gate 3187c478bd9Sstevel@tonic-gate 3197c478bd9Sstevel@tonic-gate static int 3207c478bd9Sstevel@tonic-gate cpupart_move_cpu(cpu_t *cp, cpupart_t *newpp, int forced) 3217c478bd9Sstevel@tonic-gate { 3227c478bd9Sstevel@tonic-gate cpupart_t *oldpp; 3237c478bd9Sstevel@tonic-gate cpu_t *ncp, *newlist; 3247c478bd9Sstevel@tonic-gate kthread_t *t; 3257c478bd9Sstevel@tonic-gate int move_threads = 1; 3267c478bd9Sstevel@tonic-gate lgrp_id_t lgrpid; 3277c478bd9Sstevel@tonic-gate proc_t *p; 3287c478bd9Sstevel@tonic-gate int lgrp_diff_lpl; 3297c478bd9Sstevel@tonic-gate lpl_t *cpu_lpl; 3307c478bd9Sstevel@tonic-gate int ret; 3310b70c467Sakolb boolean_t unbind_all_threads = (forced != 0); 3327c478bd9Sstevel@tonic-gate 3337c478bd9Sstevel@tonic-gate ASSERT(MUTEX_HELD(&cpu_lock)); 3347c478bd9Sstevel@tonic-gate ASSERT(newpp != NULL); 3357c478bd9Sstevel@tonic-gate 3367c478bd9Sstevel@tonic-gate oldpp = cp->cpu_part; 3377c478bd9Sstevel@tonic-gate ASSERT(oldpp != NULL); 3387c478bd9Sstevel@tonic-gate ASSERT(oldpp->cp_ncpus > 0); 3397c478bd9Sstevel@tonic-gate 3407c478bd9Sstevel@tonic-gate if (newpp == oldpp) { 3417c478bd9Sstevel@tonic-gate /* 3427c478bd9Sstevel@tonic-gate * Don't need to do anything. 3437c478bd9Sstevel@tonic-gate */ 3447c478bd9Sstevel@tonic-gate return (0); 3457c478bd9Sstevel@tonic-gate } 3467c478bd9Sstevel@tonic-gate 3477c478bd9Sstevel@tonic-gate cpu_state_change_notify(cp->cpu_id, CPU_CPUPART_OUT); 3487c478bd9Sstevel@tonic-gate 3497c478bd9Sstevel@tonic-gate if (!disp_bound_partition(cp, 0)) { 3507c478bd9Sstevel@tonic-gate /* 3517c478bd9Sstevel@tonic-gate * Don't need to move threads if there are no threads in 3527c478bd9Sstevel@tonic-gate * the partition. Note that threads can't enter the 3537c478bd9Sstevel@tonic-gate * partition while we're holding cpu_lock. 3547c478bd9Sstevel@tonic-gate */ 3557c478bd9Sstevel@tonic-gate move_threads = 0; 3567c478bd9Sstevel@tonic-gate } else if (oldpp->cp_ncpus == 1) { 3570b70c467Sakolb /* 3580b70c467Sakolb * The last CPU is removed from a partition which has threads 3590b70c467Sakolb * running in it. Some of these threads may be bound to this 3600b70c467Sakolb * CPU. 3610b70c467Sakolb * 3620b70c467Sakolb * Attempt to unbind threads from the CPU and from the processor 3630b70c467Sakolb * set. Note that no threads should be bound to this CPU since 3640b70c467Sakolb * cpupart_move_threads will refuse to move bound threads to 3650b70c467Sakolb * other CPUs. 3660b70c467Sakolb */ 3670b70c467Sakolb (void) cpu_unbind(oldpp->cp_cpulist->cpu_id, B_FALSE); 3680b70c467Sakolb (void) cpupart_unbind_threads(oldpp, B_FALSE); 3690b70c467Sakolb 3700b70c467Sakolb if (!disp_bound_partition(cp, 0)) { 3710b70c467Sakolb /* 3720b70c467Sakolb * No bound threads in this partition any more 3730b70c467Sakolb */ 3740b70c467Sakolb move_threads = 0; 3750b70c467Sakolb } else { 3760b70c467Sakolb /* 3770b70c467Sakolb * There are still threads bound to the partition 3780b70c467Sakolb */ 3797c478bd9Sstevel@tonic-gate cpu_state_change_notify(cp->cpu_id, CPU_CPUPART_IN); 3807c478bd9Sstevel@tonic-gate return (EBUSY); 3817c478bd9Sstevel@tonic-gate } 3820b70c467Sakolb } 3837c478bd9Sstevel@tonic-gate 3840b70c467Sakolb /* 3850b70c467Sakolb * If forced flag is set unbind any threads from this CPU. 3860b70c467Sakolb * Otherwise unbind soft-bound threads only. 3870b70c467Sakolb */ 3880b70c467Sakolb if ((ret = cpu_unbind(cp->cpu_id, unbind_all_threads)) != 0) { 3897c478bd9Sstevel@tonic-gate cpu_state_change_notify(cp->cpu_id, CPU_CPUPART_IN); 3907c478bd9Sstevel@tonic-gate return (ret); 3917c478bd9Sstevel@tonic-gate } 3927c478bd9Sstevel@tonic-gate 3937c478bd9Sstevel@tonic-gate /* 3947c478bd9Sstevel@tonic-gate * Stop further threads weak binding to this cpu. 3957c478bd9Sstevel@tonic-gate */ 3967c478bd9Sstevel@tonic-gate cpu_inmotion = cp; 3977c478bd9Sstevel@tonic-gate membar_enter(); 3987c478bd9Sstevel@tonic-gate 399fb2f18f8Sesaxe /* 400fb2f18f8Sesaxe * Notify the Processor Groups subsystem that the CPU 401fb2f18f8Sesaxe * will be moving cpu partitions. This is done before 402fb2f18f8Sesaxe * CPUs are paused to provide an opportunity for any 403fb2f18f8Sesaxe * needed memory allocations. 404fb2f18f8Sesaxe */ 405fb2f18f8Sesaxe pg_cpupart_out(cp, oldpp); 406fb2f18f8Sesaxe pg_cpupart_in(cp, newpp); 407fb2f18f8Sesaxe 4087c478bd9Sstevel@tonic-gate again: 4097c478bd9Sstevel@tonic-gate if (move_threads) { 4107c478bd9Sstevel@tonic-gate int loop_count; 4117c478bd9Sstevel@tonic-gate /* 4127c478bd9Sstevel@tonic-gate * Check for threads strong or weak bound to this CPU. 4137c478bd9Sstevel@tonic-gate */ 4147c478bd9Sstevel@tonic-gate for (loop_count = 0; disp_bound_threads(cp, 0); loop_count++) { 4157c478bd9Sstevel@tonic-gate if (loop_count >= 5) { 4167c478bd9Sstevel@tonic-gate cpu_state_change_notify(cp->cpu_id, 4177c478bd9Sstevel@tonic-gate CPU_CPUPART_IN); 418fb2f18f8Sesaxe pg_cpupart_out(cp, newpp); 419fb2f18f8Sesaxe pg_cpupart_in(cp, oldpp); 4207c478bd9Sstevel@tonic-gate cpu_inmotion = NULL; 4217c478bd9Sstevel@tonic-gate return (EBUSY); /* some threads still bound */ 4227c478bd9Sstevel@tonic-gate } 4237c478bd9Sstevel@tonic-gate delay(1); 4247c478bd9Sstevel@tonic-gate } 4257c478bd9Sstevel@tonic-gate } 4267c478bd9Sstevel@tonic-gate 4277c478bd9Sstevel@tonic-gate /* 4287c478bd9Sstevel@tonic-gate * Before we actually start changing data structures, notify 4297c478bd9Sstevel@tonic-gate * the cyclic subsystem that we want to move this CPU out of its 4307c478bd9Sstevel@tonic-gate * partition. 4317c478bd9Sstevel@tonic-gate */ 4327c478bd9Sstevel@tonic-gate if (!cyclic_move_out(cp)) { 4337c478bd9Sstevel@tonic-gate /* 4347c478bd9Sstevel@tonic-gate * This CPU must be the last CPU in a processor set with 4357c478bd9Sstevel@tonic-gate * a bound cyclic. 4367c478bd9Sstevel@tonic-gate */ 4377c478bd9Sstevel@tonic-gate cpu_state_change_notify(cp->cpu_id, CPU_CPUPART_IN); 438fb2f18f8Sesaxe pg_cpupart_out(cp, newpp); 439fb2f18f8Sesaxe pg_cpupart_in(cp, oldpp); 4407c478bd9Sstevel@tonic-gate cpu_inmotion = NULL; 4417c478bd9Sstevel@tonic-gate return (EBUSY); 4427c478bd9Sstevel@tonic-gate } 4437c478bd9Sstevel@tonic-gate 444*0ed5c46eSJosef 'Jeff' Sipek pause_cpus(cp, NULL); 4457c478bd9Sstevel@tonic-gate 4467c478bd9Sstevel@tonic-gate if (move_threads) { 4477c478bd9Sstevel@tonic-gate /* 4487c478bd9Sstevel@tonic-gate * The thread on cpu before the pause thread may have read 4497c478bd9Sstevel@tonic-gate * cpu_inmotion before we raised the barrier above. Check 4507c478bd9Sstevel@tonic-gate * again. 4517c478bd9Sstevel@tonic-gate */ 4527c478bd9Sstevel@tonic-gate if (disp_bound_threads(cp, 1)) { 4537c478bd9Sstevel@tonic-gate start_cpus(); 4547c478bd9Sstevel@tonic-gate goto again; 4557c478bd9Sstevel@tonic-gate } 4567c478bd9Sstevel@tonic-gate 4577c478bd9Sstevel@tonic-gate } 4587c478bd9Sstevel@tonic-gate 4597c478bd9Sstevel@tonic-gate /* 460fb2f18f8Sesaxe * Now that CPUs are paused, let the PG subsystem perform 461fb2f18f8Sesaxe * any necessary data structure updates. 4627c478bd9Sstevel@tonic-gate */ 463fb2f18f8Sesaxe pg_cpupart_move(cp, oldpp, newpp); 4647c478bd9Sstevel@tonic-gate 4657c478bd9Sstevel@tonic-gate /* save this cpu's lgroup -- it'll be the same in the new partition */ 4667c478bd9Sstevel@tonic-gate lgrpid = cp->cpu_lpl->lpl_lgrpid; 4677c478bd9Sstevel@tonic-gate 4687c478bd9Sstevel@tonic-gate cpu_lpl = cp->cpu_lpl; 4697c478bd9Sstevel@tonic-gate /* 4707c478bd9Sstevel@tonic-gate * let the lgroup framework know cp has left the partition 4717c478bd9Sstevel@tonic-gate */ 4727c478bd9Sstevel@tonic-gate lgrp_config(LGRP_CONFIG_CPUPART_DEL, (uintptr_t)cp, lgrpid); 4737c478bd9Sstevel@tonic-gate 4747c478bd9Sstevel@tonic-gate /* move out of old partition */ 4757c478bd9Sstevel@tonic-gate oldpp->cp_ncpus--; 4767c478bd9Sstevel@tonic-gate if (oldpp->cp_ncpus > 0) { 4777c478bd9Sstevel@tonic-gate 4787c478bd9Sstevel@tonic-gate ncp = cp->cpu_prev_part->cpu_next_part = cp->cpu_next_part; 4797c478bd9Sstevel@tonic-gate cp->cpu_next_part->cpu_prev_part = cp->cpu_prev_part; 4807c478bd9Sstevel@tonic-gate if (oldpp->cp_cpulist == cp) { 4817c478bd9Sstevel@tonic-gate oldpp->cp_cpulist = ncp; 4827c478bd9Sstevel@tonic-gate } 4837c478bd9Sstevel@tonic-gate } else { 4847c478bd9Sstevel@tonic-gate ncp = oldpp->cp_cpulist = NULL; 4857c478bd9Sstevel@tonic-gate cp_numparts_nonempty--; 4867c478bd9Sstevel@tonic-gate ASSERT(cp_numparts_nonempty != 0); 4877c478bd9Sstevel@tonic-gate } 4887c478bd9Sstevel@tonic-gate oldpp->cp_gen++; 4897c478bd9Sstevel@tonic-gate 4907c478bd9Sstevel@tonic-gate /* move into new partition */ 4917c478bd9Sstevel@tonic-gate newlist = newpp->cp_cpulist; 4927c478bd9Sstevel@tonic-gate if (newlist == NULL) { 4937c478bd9Sstevel@tonic-gate newpp->cp_cpulist = cp->cpu_next_part = cp->cpu_prev_part = cp; 4947c478bd9Sstevel@tonic-gate cp_numparts_nonempty++; 4957c478bd9Sstevel@tonic-gate ASSERT(cp_numparts_nonempty != 0); 4967c478bd9Sstevel@tonic-gate } else { 4977c478bd9Sstevel@tonic-gate cp->cpu_next_part = newlist; 4987c478bd9Sstevel@tonic-gate cp->cpu_prev_part = newlist->cpu_prev_part; 4997c478bd9Sstevel@tonic-gate newlist->cpu_prev_part->cpu_next_part = cp; 5007c478bd9Sstevel@tonic-gate newlist->cpu_prev_part = cp; 5017c478bd9Sstevel@tonic-gate } 5027c478bd9Sstevel@tonic-gate cp->cpu_part = newpp; 5037c478bd9Sstevel@tonic-gate newpp->cp_ncpus++; 5047c478bd9Sstevel@tonic-gate newpp->cp_gen++; 5057c478bd9Sstevel@tonic-gate 5066890d023SEric Saxe ASSERT(bitset_is_null(&newpp->cp_haltset)); 5076890d023SEric Saxe ASSERT(bitset_is_null(&oldpp->cp_haltset)); 5087c478bd9Sstevel@tonic-gate 5097c478bd9Sstevel@tonic-gate /* 5107c478bd9Sstevel@tonic-gate * let the lgroup framework know cp has entered the partition 5117c478bd9Sstevel@tonic-gate */ 5127c478bd9Sstevel@tonic-gate lgrp_config(LGRP_CONFIG_CPUPART_ADD, (uintptr_t)cp, lgrpid); 5137c478bd9Sstevel@tonic-gate 5147c478bd9Sstevel@tonic-gate /* 5157c478bd9Sstevel@tonic-gate * If necessary, move threads off processor. 5167c478bd9Sstevel@tonic-gate */ 5177c478bd9Sstevel@tonic-gate if (move_threads) { 5187c478bd9Sstevel@tonic-gate ASSERT(ncp != NULL); 5197c478bd9Sstevel@tonic-gate 5207c478bd9Sstevel@tonic-gate /* 5217c478bd9Sstevel@tonic-gate * Walk thru the active process list to look for 5227c478bd9Sstevel@tonic-gate * threads that need to have a new home lgroup, 5237c478bd9Sstevel@tonic-gate * or the last CPU they run on is the same CPU 5247c478bd9Sstevel@tonic-gate * being moved out of the partition. 5257c478bd9Sstevel@tonic-gate */ 5267c478bd9Sstevel@tonic-gate 5277c478bd9Sstevel@tonic-gate for (p = practive; p != NULL; p = p->p_next) { 5287c478bd9Sstevel@tonic-gate 5297c478bd9Sstevel@tonic-gate t = p->p_tlist; 5307c478bd9Sstevel@tonic-gate 5317c478bd9Sstevel@tonic-gate if (t == NULL) 5327c478bd9Sstevel@tonic-gate continue; 5337c478bd9Sstevel@tonic-gate 5347c478bd9Sstevel@tonic-gate lgrp_diff_lpl = 0; 5357c478bd9Sstevel@tonic-gate 5367c478bd9Sstevel@tonic-gate do { 5377c478bd9Sstevel@tonic-gate 5387c478bd9Sstevel@tonic-gate ASSERT(t->t_lpl != NULL); 5397c478bd9Sstevel@tonic-gate 5407c478bd9Sstevel@tonic-gate /* 5417c478bd9Sstevel@tonic-gate * Update the count of how many threads are 5427c478bd9Sstevel@tonic-gate * in this CPU's lgroup but have a different lpl 5437c478bd9Sstevel@tonic-gate */ 5447c478bd9Sstevel@tonic-gate 5457c478bd9Sstevel@tonic-gate if (t->t_lpl != cpu_lpl && 5467c478bd9Sstevel@tonic-gate t->t_lpl->lpl_lgrpid == lgrpid) 5477c478bd9Sstevel@tonic-gate lgrp_diff_lpl++; 5487c478bd9Sstevel@tonic-gate /* 5497c478bd9Sstevel@tonic-gate * If the lgroup that t is assigned to no 5507c478bd9Sstevel@tonic-gate * longer has any CPUs in t's partition, 5517c478bd9Sstevel@tonic-gate * we'll have to choose a new lgroup for t. 5527c478bd9Sstevel@tonic-gate */ 5537c478bd9Sstevel@tonic-gate 5547c478bd9Sstevel@tonic-gate if (!LGRP_CPUS_IN_PART(t->t_lpl->lpl_lgrpid, 5557c478bd9Sstevel@tonic-gate t->t_cpupart)) { 5567c478bd9Sstevel@tonic-gate lgrp_move_thread(t, 5577c478bd9Sstevel@tonic-gate lgrp_choose(t, t->t_cpupart), 0); 5587c478bd9Sstevel@tonic-gate } 5597c478bd9Sstevel@tonic-gate 5607c478bd9Sstevel@tonic-gate /* 5617c478bd9Sstevel@tonic-gate * make sure lpl points to our own partition 5627c478bd9Sstevel@tonic-gate */ 5637c478bd9Sstevel@tonic-gate ASSERT(t->t_lpl >= t->t_cpupart->cp_lgrploads && 5647c478bd9Sstevel@tonic-gate (t->t_lpl < t->t_cpupart->cp_lgrploads + 5657c478bd9Sstevel@tonic-gate t->t_cpupart->cp_nlgrploads)); 5667c478bd9Sstevel@tonic-gate 5677c478bd9Sstevel@tonic-gate ASSERT(t->t_lpl->lpl_ncpu > 0); 5687c478bd9Sstevel@tonic-gate 5697c478bd9Sstevel@tonic-gate /* Update CPU last ran on if it was this CPU */ 5707c478bd9Sstevel@tonic-gate if (t->t_cpu == cp && t->t_cpupart == oldpp && 5717c478bd9Sstevel@tonic-gate t->t_bound_cpu != cp) { 5727c478bd9Sstevel@tonic-gate t->t_cpu = disp_lowpri_cpu(ncp, 5737c478bd9Sstevel@tonic-gate t->t_lpl, t->t_pri, NULL); 5747c478bd9Sstevel@tonic-gate } 5757c478bd9Sstevel@tonic-gate t = t->t_forw; 5767c478bd9Sstevel@tonic-gate } while (t != p->p_tlist); 5777c478bd9Sstevel@tonic-gate 5787c478bd9Sstevel@tonic-gate /* 5797c478bd9Sstevel@tonic-gate * Didn't find any threads in the same lgroup as this 5807c478bd9Sstevel@tonic-gate * CPU with a different lpl, so remove the lgroup from 5817c478bd9Sstevel@tonic-gate * the process lgroup bitmask. 5827c478bd9Sstevel@tonic-gate */ 5837c478bd9Sstevel@tonic-gate 5847c478bd9Sstevel@tonic-gate if (lgrp_diff_lpl) 5857c478bd9Sstevel@tonic-gate klgrpset_del(p->p_lgrpset, lgrpid); 5867c478bd9Sstevel@tonic-gate } 5877c478bd9Sstevel@tonic-gate 5887c478bd9Sstevel@tonic-gate /* 5897c478bd9Sstevel@tonic-gate * Walk thread list looking for threads that need to be 5907c478bd9Sstevel@tonic-gate * rehomed, since there are some threads that are not in 5917c478bd9Sstevel@tonic-gate * their process's p_tlist. 5927c478bd9Sstevel@tonic-gate */ 5937c478bd9Sstevel@tonic-gate 5947c478bd9Sstevel@tonic-gate t = curthread; 5957c478bd9Sstevel@tonic-gate 5967c478bd9Sstevel@tonic-gate do { 5977c478bd9Sstevel@tonic-gate ASSERT(t != NULL && t->t_lpl != NULL); 5987c478bd9Sstevel@tonic-gate 5997c478bd9Sstevel@tonic-gate /* 6007c478bd9Sstevel@tonic-gate * If the lgroup that t is assigned to no 6017c478bd9Sstevel@tonic-gate * longer has any CPUs in t's partition, 6027c478bd9Sstevel@tonic-gate * we'll have to choose a new lgroup for t. 6037c478bd9Sstevel@tonic-gate * Also, choose best lgroup for home when 6047c478bd9Sstevel@tonic-gate * thread has specified lgroup affinities, 6057c478bd9Sstevel@tonic-gate * since there may be an lgroup with more 6067c478bd9Sstevel@tonic-gate * affinity available after moving CPUs 6077c478bd9Sstevel@tonic-gate * around. 6087c478bd9Sstevel@tonic-gate */ 6097c478bd9Sstevel@tonic-gate if (!LGRP_CPUS_IN_PART(t->t_lpl->lpl_lgrpid, 6107c478bd9Sstevel@tonic-gate t->t_cpupart) || t->t_lgrp_affinity) { 6117c478bd9Sstevel@tonic-gate lgrp_move_thread(t, 6127c478bd9Sstevel@tonic-gate lgrp_choose(t, t->t_cpupart), 1); 6137c478bd9Sstevel@tonic-gate } 6147c478bd9Sstevel@tonic-gate 6157c478bd9Sstevel@tonic-gate /* make sure lpl points to our own partition */ 6167c478bd9Sstevel@tonic-gate ASSERT((t->t_lpl >= t->t_cpupart->cp_lgrploads) && 6177c478bd9Sstevel@tonic-gate (t->t_lpl < t->t_cpupart->cp_lgrploads + 6187c478bd9Sstevel@tonic-gate t->t_cpupart->cp_nlgrploads)); 6197c478bd9Sstevel@tonic-gate 6207c478bd9Sstevel@tonic-gate ASSERT(t->t_lpl->lpl_ncpu > 0); 6217c478bd9Sstevel@tonic-gate 6227c478bd9Sstevel@tonic-gate /* Update CPU last ran on if it was this CPU */ 6237c478bd9Sstevel@tonic-gate if (t->t_cpu == cp && t->t_cpupart == oldpp && 6247c478bd9Sstevel@tonic-gate t->t_bound_cpu != cp) { 6257c478bd9Sstevel@tonic-gate t->t_cpu = disp_lowpri_cpu(ncp, t->t_lpl, 6267c478bd9Sstevel@tonic-gate t->t_pri, NULL); 6277c478bd9Sstevel@tonic-gate } 6287c478bd9Sstevel@tonic-gate 6297c478bd9Sstevel@tonic-gate t = t->t_next; 6307c478bd9Sstevel@tonic-gate } while (t != curthread); 6317c478bd9Sstevel@tonic-gate 6327c478bd9Sstevel@tonic-gate /* 6337c478bd9Sstevel@tonic-gate * Clear off the CPU's run queue, and the kp queue if the 6347c478bd9Sstevel@tonic-gate * partition is now empty. 6357c478bd9Sstevel@tonic-gate */ 6367c478bd9Sstevel@tonic-gate disp_cpu_inactive(cp); 6377c478bd9Sstevel@tonic-gate 6387c478bd9Sstevel@tonic-gate /* 6397c478bd9Sstevel@tonic-gate * Make cp switch to a thread from the new partition. 6407c478bd9Sstevel@tonic-gate */ 6417c478bd9Sstevel@tonic-gate cp->cpu_runrun = 1; 6427c478bd9Sstevel@tonic-gate cp->cpu_kprunrun = 1; 6437c478bd9Sstevel@tonic-gate } 6447c478bd9Sstevel@tonic-gate 6457c478bd9Sstevel@tonic-gate cpu_inmotion = NULL; 6467c478bd9Sstevel@tonic-gate start_cpus(); 6477c478bd9Sstevel@tonic-gate 6487c478bd9Sstevel@tonic-gate /* 6497c478bd9Sstevel@tonic-gate * Let anyone interested know that cpu has been added to the set. 6507c478bd9Sstevel@tonic-gate */ 6517c478bd9Sstevel@tonic-gate cpu_state_change_notify(cp->cpu_id, CPU_CPUPART_IN); 6527c478bd9Sstevel@tonic-gate 6537c478bd9Sstevel@tonic-gate /* 6547c478bd9Sstevel@tonic-gate * Now let the cyclic subsystem know that it can reshuffle cyclics 6557c478bd9Sstevel@tonic-gate * bound to the new processor set. 6567c478bd9Sstevel@tonic-gate */ 6577c478bd9Sstevel@tonic-gate cyclic_move_in(cp); 6587c478bd9Sstevel@tonic-gate 6597c478bd9Sstevel@tonic-gate return (0); 6607c478bd9Sstevel@tonic-gate } 6617c478bd9Sstevel@tonic-gate 6627c478bd9Sstevel@tonic-gate /* 6637c478bd9Sstevel@tonic-gate * Check if thread can be moved to a new cpu partition. Called by 6647c478bd9Sstevel@tonic-gate * cpupart_move_thread() and pset_bind_start(). 6657c478bd9Sstevel@tonic-gate */ 6667c478bd9Sstevel@tonic-gate int 6677c478bd9Sstevel@tonic-gate cpupart_movable_thread(kthread_id_t tp, cpupart_t *cp, int ignore) 6687c478bd9Sstevel@tonic-gate { 6697c478bd9Sstevel@tonic-gate ASSERT(MUTEX_HELD(&cpu_lock)); 6707c478bd9Sstevel@tonic-gate ASSERT(MUTEX_HELD(&ttoproc(tp)->p_lock)); 6717c478bd9Sstevel@tonic-gate ASSERT(cp != NULL); 6727c478bd9Sstevel@tonic-gate ASSERT(THREAD_LOCK_HELD(tp)); 6737c478bd9Sstevel@tonic-gate 6747c478bd9Sstevel@tonic-gate /* 6757c478bd9Sstevel@tonic-gate * CPU-bound threads can't be moved. 6767c478bd9Sstevel@tonic-gate */ 6777c478bd9Sstevel@tonic-gate if (!ignore) { 6787c478bd9Sstevel@tonic-gate cpu_t *boundcpu = tp->t_bound_cpu ? tp->t_bound_cpu : 6797c478bd9Sstevel@tonic-gate tp->t_weakbound_cpu; 6807c478bd9Sstevel@tonic-gate if (boundcpu != NULL && boundcpu->cpu_part != cp) 6817c478bd9Sstevel@tonic-gate return (EBUSY); 6827c478bd9Sstevel@tonic-gate } 68335a5a358SJonathan Adams 68435a5a358SJonathan Adams if (tp->t_cid == sysdccid) { 68535a5a358SJonathan Adams return (EINVAL); /* For now, sysdc threads can't move */ 68635a5a358SJonathan Adams } 68735a5a358SJonathan Adams 6887c478bd9Sstevel@tonic-gate return (0); 6897c478bd9Sstevel@tonic-gate } 6907c478bd9Sstevel@tonic-gate 6917c478bd9Sstevel@tonic-gate /* 6927c478bd9Sstevel@tonic-gate * Move thread to new partition. If ignore is non-zero, then CPU 6937c478bd9Sstevel@tonic-gate * bindings should be ignored (this is used when destroying a 6947c478bd9Sstevel@tonic-gate * partition). 6957c478bd9Sstevel@tonic-gate */ 6967c478bd9Sstevel@tonic-gate static int 6977c478bd9Sstevel@tonic-gate cpupart_move_thread(kthread_id_t tp, cpupart_t *newpp, int ignore, 6987c478bd9Sstevel@tonic-gate void *projbuf, void *zonebuf) 6997c478bd9Sstevel@tonic-gate { 7007c478bd9Sstevel@tonic-gate cpupart_t *oldpp = tp->t_cpupart; 7017c478bd9Sstevel@tonic-gate int ret; 7027c478bd9Sstevel@tonic-gate 7037c478bd9Sstevel@tonic-gate ASSERT(MUTEX_HELD(&cpu_lock)); 7047c478bd9Sstevel@tonic-gate ASSERT(MUTEX_HELD(&pidlock)); 7057c478bd9Sstevel@tonic-gate ASSERT(MUTEX_HELD(&ttoproc(tp)->p_lock)); 7067c478bd9Sstevel@tonic-gate ASSERT(newpp != NULL); 7077c478bd9Sstevel@tonic-gate 7087c478bd9Sstevel@tonic-gate if (newpp->cp_cpulist == NULL) 7097c478bd9Sstevel@tonic-gate return (EINVAL); 7107c478bd9Sstevel@tonic-gate 7117c478bd9Sstevel@tonic-gate /* 7127c478bd9Sstevel@tonic-gate * Check for errors first. 7137c478bd9Sstevel@tonic-gate */ 7147c478bd9Sstevel@tonic-gate thread_lock(tp); 7157c478bd9Sstevel@tonic-gate if ((ret = cpupart_movable_thread(tp, newpp, ignore)) != 0) { 7167c478bd9Sstevel@tonic-gate thread_unlock(tp); 7177c478bd9Sstevel@tonic-gate return (ret); 7187c478bd9Sstevel@tonic-gate } 7197c478bd9Sstevel@tonic-gate 7207c478bd9Sstevel@tonic-gate /* move the thread */ 7217c478bd9Sstevel@tonic-gate if (oldpp != newpp) { 7227c478bd9Sstevel@tonic-gate /* 7237c478bd9Sstevel@tonic-gate * Make the thread switch to the new partition. 7247c478bd9Sstevel@tonic-gate */ 7257c478bd9Sstevel@tonic-gate tp->t_cpupart = newpp; 7267c478bd9Sstevel@tonic-gate ASSERT(tp->t_lpl != NULL); 7277c478bd9Sstevel@tonic-gate /* 7287c478bd9Sstevel@tonic-gate * Leave the thread on the same lgroup if possible; otherwise 7297c478bd9Sstevel@tonic-gate * choose a new lgroup for it. In either case, update its 7307c478bd9Sstevel@tonic-gate * t_lpl. 7317c478bd9Sstevel@tonic-gate */ 7327c478bd9Sstevel@tonic-gate if (LGRP_CPUS_IN_PART(tp->t_lpl->lpl_lgrpid, newpp) && 7337c478bd9Sstevel@tonic-gate tp->t_lgrp_affinity == NULL) { 7347c478bd9Sstevel@tonic-gate /* 7357c478bd9Sstevel@tonic-gate * The thread's lgroup has CPUs in the thread's new 7367c478bd9Sstevel@tonic-gate * partition, so the thread can stay assigned to the 7377c478bd9Sstevel@tonic-gate * same lgroup. Update its t_lpl to point to the 7387c478bd9Sstevel@tonic-gate * lpl_t for its lgroup in its new partition. 7397c478bd9Sstevel@tonic-gate */ 7407c478bd9Sstevel@tonic-gate lgrp_move_thread(tp, &tp->t_cpupart->\ 7417c478bd9Sstevel@tonic-gate cp_lgrploads[tp->t_lpl->lpl_lgrpid], 1); 7427c478bd9Sstevel@tonic-gate } else { 7437c478bd9Sstevel@tonic-gate /* 7447c478bd9Sstevel@tonic-gate * The thread's lgroup has no cpus in its new 7457c478bd9Sstevel@tonic-gate * partition or it has specified lgroup affinities, 7467c478bd9Sstevel@tonic-gate * so choose the best lgroup for the thread and 7477c478bd9Sstevel@tonic-gate * assign it to that lgroup. 7487c478bd9Sstevel@tonic-gate */ 7497c478bd9Sstevel@tonic-gate lgrp_move_thread(tp, lgrp_choose(tp, tp->t_cpupart), 7507c478bd9Sstevel@tonic-gate 1); 7517c478bd9Sstevel@tonic-gate } 7527c478bd9Sstevel@tonic-gate /* 7537c478bd9Sstevel@tonic-gate * make sure lpl points to our own partition 7547c478bd9Sstevel@tonic-gate */ 7557c478bd9Sstevel@tonic-gate ASSERT((tp->t_lpl >= tp->t_cpupart->cp_lgrploads) && 7567c478bd9Sstevel@tonic-gate (tp->t_lpl < tp->t_cpupart->cp_lgrploads + 7577c478bd9Sstevel@tonic-gate tp->t_cpupart->cp_nlgrploads)); 7587c478bd9Sstevel@tonic-gate 7597c478bd9Sstevel@tonic-gate ASSERT(tp->t_lpl->lpl_ncpu > 0); 7607c478bd9Sstevel@tonic-gate 7617c478bd9Sstevel@tonic-gate if (tp->t_state == TS_ONPROC) { 7627c478bd9Sstevel@tonic-gate cpu_surrender(tp); 7637c478bd9Sstevel@tonic-gate } else if (tp->t_state == TS_RUN) { 7647c478bd9Sstevel@tonic-gate (void) dispdeq(tp); 7657c478bd9Sstevel@tonic-gate setbackdq(tp); 7667c478bd9Sstevel@tonic-gate } 7677c478bd9Sstevel@tonic-gate } 7687c478bd9Sstevel@tonic-gate 7697c478bd9Sstevel@tonic-gate /* 7707c478bd9Sstevel@tonic-gate * Our binding has changed; set TP_CHANGEBIND. 7717c478bd9Sstevel@tonic-gate */ 7727c478bd9Sstevel@tonic-gate tp->t_proc_flag |= TP_CHANGEBIND; 7737c478bd9Sstevel@tonic-gate aston(tp); 7747c478bd9Sstevel@tonic-gate 7757c478bd9Sstevel@tonic-gate thread_unlock(tp); 7767c478bd9Sstevel@tonic-gate fss_changepset(tp, newpp, projbuf, zonebuf); 7777c478bd9Sstevel@tonic-gate 7787c478bd9Sstevel@tonic-gate return (0); /* success */ 7797c478bd9Sstevel@tonic-gate } 7807c478bd9Sstevel@tonic-gate 7817c478bd9Sstevel@tonic-gate 7827c478bd9Sstevel@tonic-gate /* 7837c478bd9Sstevel@tonic-gate * This function binds a thread to a partition. Must be called with the 7847c478bd9Sstevel@tonic-gate * p_lock of the containing process held (to keep the thread from going 7857c478bd9Sstevel@tonic-gate * away), and thus also with cpu_lock held (since cpu_lock must be 7867c478bd9Sstevel@tonic-gate * acquired before p_lock). If ignore is non-zero, then CPU bindings 7877c478bd9Sstevel@tonic-gate * should be ignored (this is used when destroying a partition). 7887c478bd9Sstevel@tonic-gate */ 7897c478bd9Sstevel@tonic-gate int 7907c478bd9Sstevel@tonic-gate cpupart_bind_thread(kthread_id_t tp, psetid_t psid, int ignore, void *projbuf, 7917c478bd9Sstevel@tonic-gate void *zonebuf) 7927c478bd9Sstevel@tonic-gate { 7937c478bd9Sstevel@tonic-gate cpupart_t *newpp; 7947c478bd9Sstevel@tonic-gate 7957c478bd9Sstevel@tonic-gate ASSERT(pool_lock_held()); 7967c478bd9Sstevel@tonic-gate ASSERT(MUTEX_HELD(&cpu_lock)); 7977c478bd9Sstevel@tonic-gate ASSERT(MUTEX_HELD(&pidlock)); 7987c478bd9Sstevel@tonic-gate ASSERT(MUTEX_HELD(&ttoproc(tp)->p_lock)); 7997c478bd9Sstevel@tonic-gate 8007c478bd9Sstevel@tonic-gate if (psid == PS_NONE) 8017c478bd9Sstevel@tonic-gate newpp = &cp_default; 8027c478bd9Sstevel@tonic-gate else { 8037c478bd9Sstevel@tonic-gate newpp = cpupart_find(psid); 8047c478bd9Sstevel@tonic-gate if (newpp == NULL) { 8057c478bd9Sstevel@tonic-gate return (EINVAL); 8067c478bd9Sstevel@tonic-gate } 8077c478bd9Sstevel@tonic-gate } 8087c478bd9Sstevel@tonic-gate return (cpupart_move_thread(tp, newpp, ignore, projbuf, zonebuf)); 8097c478bd9Sstevel@tonic-gate } 8107c478bd9Sstevel@tonic-gate 8117c478bd9Sstevel@tonic-gate 8127c478bd9Sstevel@tonic-gate /* 8137c478bd9Sstevel@tonic-gate * Create a new partition. On MP systems, this also allocates a 8147c478bd9Sstevel@tonic-gate * kpreempt disp queue for that partition. 8157c478bd9Sstevel@tonic-gate */ 8167c478bd9Sstevel@tonic-gate int 8177c478bd9Sstevel@tonic-gate cpupart_create(psetid_t *psid) 8187c478bd9Sstevel@tonic-gate { 8197c478bd9Sstevel@tonic-gate cpupart_t *pp; 8207c478bd9Sstevel@tonic-gate 8217c478bd9Sstevel@tonic-gate ASSERT(pool_lock_held()); 8227c478bd9Sstevel@tonic-gate 8237c478bd9Sstevel@tonic-gate pp = kmem_zalloc(sizeof (cpupart_t), KM_SLEEP); 8247c478bd9Sstevel@tonic-gate pp->cp_nlgrploads = lgrp_plat_max_lgrps(); 8257c478bd9Sstevel@tonic-gate pp->cp_lgrploads = kmem_zalloc(sizeof (lpl_t) * pp->cp_nlgrploads, 8267c478bd9Sstevel@tonic-gate KM_SLEEP); 8277c478bd9Sstevel@tonic-gate 8287c478bd9Sstevel@tonic-gate mutex_enter(&cpu_lock); 8297c478bd9Sstevel@tonic-gate if (cp_numparts == cp_max_numparts) { 8307c478bd9Sstevel@tonic-gate mutex_exit(&cpu_lock); 8317c478bd9Sstevel@tonic-gate kmem_free(pp->cp_lgrploads, sizeof (lpl_t) * pp->cp_nlgrploads); 8327c478bd9Sstevel@tonic-gate pp->cp_lgrploads = NULL; 8337c478bd9Sstevel@tonic-gate kmem_free(pp, sizeof (cpupart_t)); 8347c478bd9Sstevel@tonic-gate return (ENOMEM); 8357c478bd9Sstevel@tonic-gate } 8367c478bd9Sstevel@tonic-gate cp_numparts++; 8377c478bd9Sstevel@tonic-gate /* find the next free partition ID */ 8387c478bd9Sstevel@tonic-gate while (cpupart_find(CPTOPS(cp_id_next)) != NULL) 8397c478bd9Sstevel@tonic-gate cp_id_next++; 8407c478bd9Sstevel@tonic-gate pp->cp_id = cp_id_next++; 8417c478bd9Sstevel@tonic-gate pp->cp_ncpus = 0; 8427c478bd9Sstevel@tonic-gate pp->cp_cpulist = NULL; 8437c478bd9Sstevel@tonic-gate pp->cp_attr = 0; 8447c478bd9Sstevel@tonic-gate klgrpset_clear(pp->cp_lgrpset); 8457c478bd9Sstevel@tonic-gate pp->cp_kp_queue.disp_maxrunpri = -1; 8467c478bd9Sstevel@tonic-gate pp->cp_kp_queue.disp_max_unbound_pri = -1; 8477c478bd9Sstevel@tonic-gate pp->cp_kp_queue.disp_cpu = NULL; 8487c478bd9Sstevel@tonic-gate pp->cp_gen = 0; 8497c478bd9Sstevel@tonic-gate DISP_LOCK_INIT(&pp->cp_kp_queue.disp_lock); 8507c478bd9Sstevel@tonic-gate *psid = CPTOPS(pp->cp_id); 8517c478bd9Sstevel@tonic-gate disp_kp_alloc(&pp->cp_kp_queue, v.v_nglobpris); 8527c478bd9Sstevel@tonic-gate cpupart_kstat_create(pp); 8536890d023SEric Saxe cpupart_lpl_initialize(pp); 8546890d023SEric Saxe 855fb2f18f8Sesaxe bitset_init(&pp->cp_cmt_pgs); 8567c478bd9Sstevel@tonic-gate 8577c478bd9Sstevel@tonic-gate /* 8580542eecfSRafael Vanoni * Initialize and size the partition's bitset of halted CPUs. 8596890d023SEric Saxe */ 8600542eecfSRafael Vanoni bitset_init_fanout(&pp->cp_haltset, cp_haltset_fanout); 8616890d023SEric Saxe bitset_resize(&pp->cp_haltset, max_ncpus); 8626890d023SEric Saxe 8636890d023SEric Saxe /* 8647c478bd9Sstevel@tonic-gate * Pause all CPUs while changing the partition list, to make sure 8657c478bd9Sstevel@tonic-gate * the clock thread (which traverses the list without holding 8667c478bd9Sstevel@tonic-gate * cpu_lock) isn't running. 8677c478bd9Sstevel@tonic-gate */ 868*0ed5c46eSJosef 'Jeff' Sipek pause_cpus(NULL, NULL); 8697c478bd9Sstevel@tonic-gate pp->cp_next = cp_list_head; 8707c478bd9Sstevel@tonic-gate pp->cp_prev = cp_list_head->cp_prev; 8717c478bd9Sstevel@tonic-gate cp_list_head->cp_prev->cp_next = pp; 8727c478bd9Sstevel@tonic-gate cp_list_head->cp_prev = pp; 8737c478bd9Sstevel@tonic-gate start_cpus(); 8747c478bd9Sstevel@tonic-gate mutex_exit(&cpu_lock); 8757c478bd9Sstevel@tonic-gate 8767c478bd9Sstevel@tonic-gate return (0); 8777c478bd9Sstevel@tonic-gate } 8787c478bd9Sstevel@tonic-gate 8797c478bd9Sstevel@tonic-gate /* 8800b70c467Sakolb * Move threads from specified partition to cp_default. If `force' is specified, 8810b70c467Sakolb * move all threads, otherwise move only soft-bound threads. 8827c478bd9Sstevel@tonic-gate */ 8830b70c467Sakolb static int 8840b70c467Sakolb cpupart_unbind_threads(cpupart_t *pp, boolean_t unbind_all) 8857c478bd9Sstevel@tonic-gate { 8867c478bd9Sstevel@tonic-gate void *projbuf, *zonebuf; 8877c478bd9Sstevel@tonic-gate kthread_t *t; 8887c478bd9Sstevel@tonic-gate proc_t *p; 8890b70c467Sakolb int err = 0; 8900b70c467Sakolb psetid_t psid = pp->cp_id; 8917c478bd9Sstevel@tonic-gate 8927c478bd9Sstevel@tonic-gate ASSERT(pool_lock_held()); 8930b70c467Sakolb ASSERT(MUTEX_HELD(&cpu_lock)); 8947c478bd9Sstevel@tonic-gate 8957c478bd9Sstevel@tonic-gate if (pp == NULL || pp == &cp_default) { 8967c478bd9Sstevel@tonic-gate return (EINVAL); 8977c478bd9Sstevel@tonic-gate } 8987c478bd9Sstevel@tonic-gate 8997c478bd9Sstevel@tonic-gate /* 9007c478bd9Sstevel@tonic-gate * Pre-allocate enough buffers for FSS for all active projects and 9017c478bd9Sstevel@tonic-gate * for all active zones on the system. Unused buffers will be 9027c478bd9Sstevel@tonic-gate * freed later by fss_freebuf(). 9037c478bd9Sstevel@tonic-gate */ 9047c478bd9Sstevel@tonic-gate projbuf = fss_allocbuf(FSS_NPROJ_BUF, FSS_ALLOC_PROJ); 9057c478bd9Sstevel@tonic-gate zonebuf = fss_allocbuf(FSS_NPROJ_BUF, FSS_ALLOC_ZONE); 9067c478bd9Sstevel@tonic-gate 9077c478bd9Sstevel@tonic-gate mutex_enter(&pidlock); 9087c478bd9Sstevel@tonic-gate t = curthread; 9097c478bd9Sstevel@tonic-gate do { 9107c478bd9Sstevel@tonic-gate if (t->t_bind_pset == psid) { 9117c478bd9Sstevel@tonic-gate again: p = ttoproc(t); 9127c478bd9Sstevel@tonic-gate mutex_enter(&p->p_lock); 9137c478bd9Sstevel@tonic-gate if (ttoproc(t) != p) { 9147c478bd9Sstevel@tonic-gate /* 9157c478bd9Sstevel@tonic-gate * lwp_exit has changed this thread's process 9167c478bd9Sstevel@tonic-gate * pointer before we grabbed its p_lock. 9177c478bd9Sstevel@tonic-gate */ 9187c478bd9Sstevel@tonic-gate mutex_exit(&p->p_lock); 9197c478bd9Sstevel@tonic-gate goto again; 9207c478bd9Sstevel@tonic-gate } 9210b70c467Sakolb 9220b70c467Sakolb /* 9230b70c467Sakolb * Can only unbind threads which have revocable binding 9240b70c467Sakolb * unless force unbinding requested. 9250b70c467Sakolb */ 9260b70c467Sakolb if (unbind_all || TB_PSET_IS_SOFT(t)) { 9277c478bd9Sstevel@tonic-gate err = cpupart_bind_thread(t, PS_NONE, 1, 9287c478bd9Sstevel@tonic-gate projbuf, zonebuf); 9297c478bd9Sstevel@tonic-gate if (err) { 9307c478bd9Sstevel@tonic-gate mutex_exit(&p->p_lock); 9317c478bd9Sstevel@tonic-gate mutex_exit(&pidlock); 9327c478bd9Sstevel@tonic-gate fss_freebuf(projbuf, FSS_ALLOC_PROJ); 9337c478bd9Sstevel@tonic-gate fss_freebuf(zonebuf, FSS_ALLOC_ZONE); 9347c478bd9Sstevel@tonic-gate return (err); 9357c478bd9Sstevel@tonic-gate } 9367c478bd9Sstevel@tonic-gate t->t_bind_pset = PS_NONE; 9370b70c467Sakolb } 9387c478bd9Sstevel@tonic-gate mutex_exit(&p->p_lock); 9397c478bd9Sstevel@tonic-gate } 9407c478bd9Sstevel@tonic-gate t = t->t_next; 9417c478bd9Sstevel@tonic-gate } while (t != curthread); 9427c478bd9Sstevel@tonic-gate 9437c478bd9Sstevel@tonic-gate mutex_exit(&pidlock); 9447c478bd9Sstevel@tonic-gate fss_freebuf(projbuf, FSS_ALLOC_PROJ); 9457c478bd9Sstevel@tonic-gate fss_freebuf(zonebuf, FSS_ALLOC_ZONE); 9460b70c467Sakolb return (err); 9470b70c467Sakolb } 9480b70c467Sakolb 9490b70c467Sakolb /* 9500b70c467Sakolb * Destroy a partition. 9510b70c467Sakolb */ 9520b70c467Sakolb int 9530b70c467Sakolb cpupart_destroy(psetid_t psid) 9540b70c467Sakolb { 9550b70c467Sakolb cpu_t *cp, *first_cp; 9560b70c467Sakolb cpupart_t *pp, *newpp; 9570b70c467Sakolb int err = 0; 9580b70c467Sakolb 9590b70c467Sakolb ASSERT(pool_lock_held()); 9600b70c467Sakolb mutex_enter(&cpu_lock); 9610b70c467Sakolb 9620b70c467Sakolb pp = cpupart_find(psid); 9630b70c467Sakolb if (pp == NULL || pp == &cp_default) { 9640b70c467Sakolb mutex_exit(&cpu_lock); 9650b70c467Sakolb return (EINVAL); 9660b70c467Sakolb } 9670b70c467Sakolb 9680b70c467Sakolb /* 9690b70c467Sakolb * Unbind all the threads currently bound to the partition. 9700b70c467Sakolb */ 9710b70c467Sakolb err = cpupart_unbind_threads(pp, B_TRUE); 9720b70c467Sakolb if (err) { 9730b70c467Sakolb mutex_exit(&cpu_lock); 9740b70c467Sakolb return (err); 9750b70c467Sakolb } 9767c478bd9Sstevel@tonic-gate 9777c478bd9Sstevel@tonic-gate newpp = &cp_default; 9787c478bd9Sstevel@tonic-gate while ((cp = pp->cp_cpulist) != NULL) { 9797c478bd9Sstevel@tonic-gate if (err = cpupart_move_cpu(cp, newpp, 0)) { 9807c478bd9Sstevel@tonic-gate mutex_exit(&cpu_lock); 9817c478bd9Sstevel@tonic-gate return (err); 9827c478bd9Sstevel@tonic-gate } 9837c478bd9Sstevel@tonic-gate } 9847c478bd9Sstevel@tonic-gate 985fb2f18f8Sesaxe ASSERT(bitset_is_null(&pp->cp_cmt_pgs)); 9866890d023SEric Saxe ASSERT(bitset_is_null(&pp->cp_haltset)); 9877c478bd9Sstevel@tonic-gate 9887c478bd9Sstevel@tonic-gate /* 9896890d023SEric Saxe * Teardown the partition's group of active CMT PGs and halted 9906890d023SEric Saxe * CPUs now that they have all left. 991fb2f18f8Sesaxe */ 992fb2f18f8Sesaxe bitset_fini(&pp->cp_cmt_pgs); 9936890d023SEric Saxe bitset_fini(&pp->cp_haltset); 994fb2f18f8Sesaxe 995fb2f18f8Sesaxe /* 9967c478bd9Sstevel@tonic-gate * Reset the pointers in any offline processors so they won't 9977c478bd9Sstevel@tonic-gate * try to rejoin the destroyed partition when they're turned 9987c478bd9Sstevel@tonic-gate * online. 9997c478bd9Sstevel@tonic-gate */ 10007c478bd9Sstevel@tonic-gate first_cp = cp = CPU; 10017c478bd9Sstevel@tonic-gate do { 10027c478bd9Sstevel@tonic-gate if (cp->cpu_part == pp) { 10037c478bd9Sstevel@tonic-gate ASSERT(cp->cpu_flags & CPU_OFFLINE); 10047c478bd9Sstevel@tonic-gate cp->cpu_part = newpp; 10057c478bd9Sstevel@tonic-gate } 10067c478bd9Sstevel@tonic-gate cp = cp->cpu_next; 10077c478bd9Sstevel@tonic-gate } while (cp != first_cp); 10087c478bd9Sstevel@tonic-gate 10097c478bd9Sstevel@tonic-gate /* 10107c478bd9Sstevel@tonic-gate * Pause all CPUs while changing the partition list, to make sure 10117c478bd9Sstevel@tonic-gate * the clock thread (which traverses the list without holding 10127c478bd9Sstevel@tonic-gate * cpu_lock) isn't running. 10137c478bd9Sstevel@tonic-gate */ 1014*0ed5c46eSJosef 'Jeff' Sipek pause_cpus(NULL, NULL); 10157c478bd9Sstevel@tonic-gate pp->cp_prev->cp_next = pp->cp_next; 10167c478bd9Sstevel@tonic-gate pp->cp_next->cp_prev = pp->cp_prev; 10177c478bd9Sstevel@tonic-gate if (cp_list_head == pp) 10187c478bd9Sstevel@tonic-gate cp_list_head = pp->cp_next; 10197c478bd9Sstevel@tonic-gate start_cpus(); 10207c478bd9Sstevel@tonic-gate 10217c478bd9Sstevel@tonic-gate if (cp_id_next > pp->cp_id) 10227c478bd9Sstevel@tonic-gate cp_id_next = pp->cp_id; 10237c478bd9Sstevel@tonic-gate 10247c478bd9Sstevel@tonic-gate if (pp->cp_kstat) 10257c478bd9Sstevel@tonic-gate kstat_delete(pp->cp_kstat); 10267c478bd9Sstevel@tonic-gate 10277c478bd9Sstevel@tonic-gate cp_numparts--; 10287c478bd9Sstevel@tonic-gate 10297c478bd9Sstevel@tonic-gate disp_kp_free(&pp->cp_kp_queue); 10306890d023SEric Saxe 10316890d023SEric Saxe cpupart_lpl_teardown(pp); 10326890d023SEric Saxe 10337c478bd9Sstevel@tonic-gate kmem_free(pp, sizeof (cpupart_t)); 10347c478bd9Sstevel@tonic-gate mutex_exit(&cpu_lock); 10357c478bd9Sstevel@tonic-gate 10367c478bd9Sstevel@tonic-gate return (err); 10377c478bd9Sstevel@tonic-gate } 10387c478bd9Sstevel@tonic-gate 10397c478bd9Sstevel@tonic-gate 10407c478bd9Sstevel@tonic-gate /* 10417c478bd9Sstevel@tonic-gate * Return the ID of the partition to which the specified processor belongs. 10427c478bd9Sstevel@tonic-gate */ 10437c478bd9Sstevel@tonic-gate psetid_t 10447c478bd9Sstevel@tonic-gate cpupart_query_cpu(cpu_t *cp) 10457c478bd9Sstevel@tonic-gate { 10467c478bd9Sstevel@tonic-gate ASSERT(MUTEX_HELD(&cpu_lock)); 10477c478bd9Sstevel@tonic-gate 10487c478bd9Sstevel@tonic-gate return (CPTOPS(cp->cpu_part->cp_id)); 10497c478bd9Sstevel@tonic-gate } 10507c478bd9Sstevel@tonic-gate 10517c478bd9Sstevel@tonic-gate 10527c478bd9Sstevel@tonic-gate /* 10537c478bd9Sstevel@tonic-gate * Attach a processor to an existing partition. 10547c478bd9Sstevel@tonic-gate */ 10557c478bd9Sstevel@tonic-gate int 10567c478bd9Sstevel@tonic-gate cpupart_attach_cpu(psetid_t psid, cpu_t *cp, int forced) 10577c478bd9Sstevel@tonic-gate { 10587c478bd9Sstevel@tonic-gate cpupart_t *pp; 10597c478bd9Sstevel@tonic-gate int err; 10607c478bd9Sstevel@tonic-gate 10617c478bd9Sstevel@tonic-gate ASSERT(pool_lock_held()); 10627c478bd9Sstevel@tonic-gate ASSERT(MUTEX_HELD(&cpu_lock)); 10637c478bd9Sstevel@tonic-gate 10647c478bd9Sstevel@tonic-gate pp = cpupart_find(psid); 10657c478bd9Sstevel@tonic-gate if (pp == NULL) 10667c478bd9Sstevel@tonic-gate return (EINVAL); 10677c478bd9Sstevel@tonic-gate if (cp->cpu_flags & CPU_OFFLINE) 10687c478bd9Sstevel@tonic-gate return (EINVAL); 10697c478bd9Sstevel@tonic-gate 10707c478bd9Sstevel@tonic-gate err = cpupart_move_cpu(cp, pp, forced); 10717c478bd9Sstevel@tonic-gate return (err); 10727c478bd9Sstevel@tonic-gate } 10737c478bd9Sstevel@tonic-gate 10747c478bd9Sstevel@tonic-gate /* 10757c478bd9Sstevel@tonic-gate * Get a list of cpus belonging to the partition. If numcpus is NULL, 10767c478bd9Sstevel@tonic-gate * this just checks for a valid partition. If numcpus is non-NULL but 10777c478bd9Sstevel@tonic-gate * cpulist is NULL, the current number of cpus is stored in *numcpus. 10787c478bd9Sstevel@tonic-gate * If both are non-NULL, the current number of cpus is stored in *numcpus, 10797c478bd9Sstevel@tonic-gate * and a list of those cpus up to the size originally in *numcpus is 10807c478bd9Sstevel@tonic-gate * stored in cpulist[]. Also, store the processor set id in *psid. 10817c478bd9Sstevel@tonic-gate * This is useful in case the processor set id passed in was PS_MYID. 10827c478bd9Sstevel@tonic-gate */ 10837c478bd9Sstevel@tonic-gate int 10847c478bd9Sstevel@tonic-gate cpupart_get_cpus(psetid_t *psid, processorid_t *cpulist, uint_t *numcpus) 10857c478bd9Sstevel@tonic-gate { 10867c478bd9Sstevel@tonic-gate cpupart_t *pp; 10877c478bd9Sstevel@tonic-gate uint_t ncpus; 10887c478bd9Sstevel@tonic-gate cpu_t *c; 10897c478bd9Sstevel@tonic-gate int i; 10907c478bd9Sstevel@tonic-gate 10917c478bd9Sstevel@tonic-gate mutex_enter(&cpu_lock); 10927c478bd9Sstevel@tonic-gate pp = cpupart_find(*psid); 10937c478bd9Sstevel@tonic-gate if (pp == NULL) { 10947c478bd9Sstevel@tonic-gate mutex_exit(&cpu_lock); 10957c478bd9Sstevel@tonic-gate return (EINVAL); 10967c478bd9Sstevel@tonic-gate } 10977c478bd9Sstevel@tonic-gate *psid = CPTOPS(pp->cp_id); 10987c478bd9Sstevel@tonic-gate ncpus = pp->cp_ncpus; 10997c478bd9Sstevel@tonic-gate if (numcpus) { 11007c478bd9Sstevel@tonic-gate if (ncpus > *numcpus) { 11017c478bd9Sstevel@tonic-gate /* 11027c478bd9Sstevel@tonic-gate * Only copy as many cpus as were passed in, but 11037c478bd9Sstevel@tonic-gate * pass back the real number. 11047c478bd9Sstevel@tonic-gate */ 11057c478bd9Sstevel@tonic-gate uint_t t = ncpus; 11067c478bd9Sstevel@tonic-gate ncpus = *numcpus; 11077c478bd9Sstevel@tonic-gate *numcpus = t; 11087c478bd9Sstevel@tonic-gate } else 11097c478bd9Sstevel@tonic-gate *numcpus = ncpus; 11107c478bd9Sstevel@tonic-gate 11117c478bd9Sstevel@tonic-gate if (cpulist) { 11127c478bd9Sstevel@tonic-gate c = pp->cp_cpulist; 11137c478bd9Sstevel@tonic-gate for (i = 0; i < ncpus; i++) { 11147c478bd9Sstevel@tonic-gate ASSERT(c != NULL); 11157c478bd9Sstevel@tonic-gate cpulist[i] = c->cpu_id; 11167c478bd9Sstevel@tonic-gate c = c->cpu_next_part; 11177c478bd9Sstevel@tonic-gate } 11187c478bd9Sstevel@tonic-gate } 11197c478bd9Sstevel@tonic-gate } 11207c478bd9Sstevel@tonic-gate mutex_exit(&cpu_lock); 11217c478bd9Sstevel@tonic-gate return (0); 11227c478bd9Sstevel@tonic-gate } 11237c478bd9Sstevel@tonic-gate 11247c478bd9Sstevel@tonic-gate /* 11257c478bd9Sstevel@tonic-gate * Reallocate kpreempt queues for each CPU partition. Called from 11267c478bd9Sstevel@tonic-gate * disp_setup when a new scheduling class is loaded that increases the 11277c478bd9Sstevel@tonic-gate * number of priorities in the system. 11287c478bd9Sstevel@tonic-gate */ 11297c478bd9Sstevel@tonic-gate void 11307c478bd9Sstevel@tonic-gate cpupart_kpqalloc(pri_t npri) 11317c478bd9Sstevel@tonic-gate { 11327c478bd9Sstevel@tonic-gate cpupart_t *cpp; 11337c478bd9Sstevel@tonic-gate 11347c478bd9Sstevel@tonic-gate ASSERT(MUTEX_HELD(&cpu_lock)); 11357c478bd9Sstevel@tonic-gate cpp = cp_list_head; 11367c478bd9Sstevel@tonic-gate do { 11377c478bd9Sstevel@tonic-gate disp_kp_alloc(&cpp->cp_kp_queue, npri); 11387c478bd9Sstevel@tonic-gate cpp = cpp->cp_next; 11397c478bd9Sstevel@tonic-gate } while (cpp != cp_list_head); 11407c478bd9Sstevel@tonic-gate } 11417c478bd9Sstevel@tonic-gate 11427c478bd9Sstevel@tonic-gate int 11437c478bd9Sstevel@tonic-gate cpupart_get_loadavg(psetid_t psid, int *buf, int nelem) 11447c478bd9Sstevel@tonic-gate { 11457c478bd9Sstevel@tonic-gate cpupart_t *cp; 11467c478bd9Sstevel@tonic-gate int i; 11477c478bd9Sstevel@tonic-gate 11487c478bd9Sstevel@tonic-gate ASSERT(nelem >= 0); 11497c478bd9Sstevel@tonic-gate ASSERT(nelem <= LOADAVG_NSTATS); 11507c478bd9Sstevel@tonic-gate ASSERT(MUTEX_HELD(&cpu_lock)); 11517c478bd9Sstevel@tonic-gate 11527c478bd9Sstevel@tonic-gate cp = cpupart_find(psid); 11537c478bd9Sstevel@tonic-gate if (cp == NULL) 11547c478bd9Sstevel@tonic-gate return (EINVAL); 11557c478bd9Sstevel@tonic-gate for (i = 0; i < nelem; i++) 11567c478bd9Sstevel@tonic-gate buf[i] = cp->cp_hp_avenrun[i] >> (16 - FSHIFT); 11577c478bd9Sstevel@tonic-gate 11587c478bd9Sstevel@tonic-gate return (0); 11597c478bd9Sstevel@tonic-gate } 11607c478bd9Sstevel@tonic-gate 11617c478bd9Sstevel@tonic-gate 11627c478bd9Sstevel@tonic-gate uint_t 11637c478bd9Sstevel@tonic-gate cpupart_list(psetid_t *list, uint_t nelem, int flag) 11647c478bd9Sstevel@tonic-gate { 11657c478bd9Sstevel@tonic-gate uint_t numpart = 0; 11667c478bd9Sstevel@tonic-gate cpupart_t *cp; 11677c478bd9Sstevel@tonic-gate 11687c478bd9Sstevel@tonic-gate ASSERT(MUTEX_HELD(&cpu_lock)); 11697c478bd9Sstevel@tonic-gate ASSERT(flag == CP_ALL || flag == CP_NONEMPTY); 11707c478bd9Sstevel@tonic-gate 11717c478bd9Sstevel@tonic-gate if (list != NULL) { 11727c478bd9Sstevel@tonic-gate cp = cp_list_head; 11737c478bd9Sstevel@tonic-gate do { 11747c478bd9Sstevel@tonic-gate if (((flag == CP_ALL) && (cp != &cp_default)) || 11757c478bd9Sstevel@tonic-gate ((flag == CP_NONEMPTY) && (cp->cp_ncpus != 0))) { 11767c478bd9Sstevel@tonic-gate if (numpart == nelem) 11777c478bd9Sstevel@tonic-gate break; 11787c478bd9Sstevel@tonic-gate list[numpart++] = CPTOPS(cp->cp_id); 11797c478bd9Sstevel@tonic-gate } 11807c478bd9Sstevel@tonic-gate cp = cp->cp_next; 11817c478bd9Sstevel@tonic-gate } while (cp != cp_list_head); 11827c478bd9Sstevel@tonic-gate } 11837c478bd9Sstevel@tonic-gate 11847c478bd9Sstevel@tonic-gate ASSERT(numpart < cp_numparts); 11857c478bd9Sstevel@tonic-gate 11867c478bd9Sstevel@tonic-gate if (flag == CP_ALL) 11877c478bd9Sstevel@tonic-gate numpart = cp_numparts - 1; /* leave out default partition */ 11887c478bd9Sstevel@tonic-gate else if (flag == CP_NONEMPTY) 11897c478bd9Sstevel@tonic-gate numpart = cp_numparts_nonempty; 11907c478bd9Sstevel@tonic-gate 11917c478bd9Sstevel@tonic-gate return (numpart); 11927c478bd9Sstevel@tonic-gate } 11937c478bd9Sstevel@tonic-gate 11947c478bd9Sstevel@tonic-gate int 11957c478bd9Sstevel@tonic-gate cpupart_setattr(psetid_t psid, uint_t attr) 11967c478bd9Sstevel@tonic-gate { 11977c478bd9Sstevel@tonic-gate cpupart_t *cp; 11987c478bd9Sstevel@tonic-gate 11997c478bd9Sstevel@tonic-gate ASSERT(pool_lock_held()); 12007c478bd9Sstevel@tonic-gate 12017c478bd9Sstevel@tonic-gate mutex_enter(&cpu_lock); 12027c478bd9Sstevel@tonic-gate if ((cp = cpupart_find(psid)) == NULL) { 12037c478bd9Sstevel@tonic-gate mutex_exit(&cpu_lock); 12047c478bd9Sstevel@tonic-gate return (EINVAL); 12057c478bd9Sstevel@tonic-gate } 12067c478bd9Sstevel@tonic-gate /* 12077c478bd9Sstevel@tonic-gate * PSET_NOESCAPE attribute for default cpu partition is always set 12087c478bd9Sstevel@tonic-gate */ 12097c478bd9Sstevel@tonic-gate if (cp == &cp_default && !(attr & PSET_NOESCAPE)) { 12107c478bd9Sstevel@tonic-gate mutex_exit(&cpu_lock); 12117c478bd9Sstevel@tonic-gate return (EINVAL); 12127c478bd9Sstevel@tonic-gate } 12137c478bd9Sstevel@tonic-gate cp->cp_attr = attr; 12147c478bd9Sstevel@tonic-gate mutex_exit(&cpu_lock); 12157c478bd9Sstevel@tonic-gate return (0); 12167c478bd9Sstevel@tonic-gate } 12177c478bd9Sstevel@tonic-gate 12187c478bd9Sstevel@tonic-gate int 12197c478bd9Sstevel@tonic-gate cpupart_getattr(psetid_t psid, uint_t *attrp) 12207c478bd9Sstevel@tonic-gate { 12217c478bd9Sstevel@tonic-gate cpupart_t *cp; 12227c478bd9Sstevel@tonic-gate 12237c478bd9Sstevel@tonic-gate mutex_enter(&cpu_lock); 12247c478bd9Sstevel@tonic-gate if ((cp = cpupart_find(psid)) == NULL) { 12257c478bd9Sstevel@tonic-gate mutex_exit(&cpu_lock); 12267c478bd9Sstevel@tonic-gate return (EINVAL); 12277c478bd9Sstevel@tonic-gate } 12287c478bd9Sstevel@tonic-gate *attrp = cp->cp_attr; 12297c478bd9Sstevel@tonic-gate mutex_exit(&cpu_lock); 12307c478bd9Sstevel@tonic-gate return (0); 12317c478bd9Sstevel@tonic-gate } 1232