1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 22 /* 23 * Copyright (c) 2007, 2010, Oracle and/or its affiliates. All rights reserved. 24 * Copyright (c) 2019, Joyent, Inc. 25 */ 26 /* 27 * Copyright (c) 2010, Intel Corporation. 28 * All rights reserved. 29 */ 30 31 /* 32 * CPU Module Interface - hardware abstraction. 33 */ 34 35 #ifdef __xpv 36 #include <sys/xpv_user.h> 37 #endif 38 39 #include <sys/types.h> 40 #include <sys/cpu_module.h> 41 #include <sys/kmem.h> 42 #include <sys/x86_archext.h> 43 #include <sys/cpuvar.h> 44 #include <sys/ksynch.h> 45 #include <sys/x_call.h> 46 #include <sys/pghw.h> 47 #include <sys/pci_cfgacc.h> 48 #include <sys/pci_cfgspace.h> 49 #include <sys/archsystm.h> 50 #include <sys/ontrap.h> 51 #include <sys/controlregs.h> 52 #include <sys/sunddi.h> 53 #include <sys/trap.h> 54 #include <sys/mca_x86.h> 55 #include <sys/processor.h> 56 #include <sys/cmn_err.h> 57 #include <sys/nvpair.h> 58 #include <sys/fm/util.h> 59 #include <sys/fm/protocol.h> 60 #include <sys/fm/smb/fmsmb.h> 61 #include <sys/cpu_module_impl.h> 62 63 /* 64 * Variable which determines if the SMBIOS supports x86 generic topology; or 65 * if legacy topolgy enumeration will occur. 66 */ 67 extern int x86gentopo_legacy; 68 69 /* 70 * Outside of this file consumers use the opaque cmi_hdl_t. This 71 * definition is duplicated in the generic_cpu mdb module, so keep 72 * them in-sync when making changes. 73 */ 74 typedef struct cmi_hdl_impl { 75 enum cmi_hdl_class cmih_class; /* Handle nature */ 76 const struct cmi_hdl_ops *cmih_ops; /* Operations vector */ 77 uint_t cmih_chipid; /* Chipid of cpu resource */ 78 uint_t cmih_procnodeid; /* Nodeid of cpu resource */ 79 uint_t cmih_coreid; /* Core within die */ 80 uint_t cmih_strandid; /* Thread within core */ 81 uint_t cmih_procnodes_per_pkg; /* Nodes in a processor */ 82 boolean_t cmih_mstrand; /* cores are multithreaded */ 83 volatile uint32_t *cmih_refcntp; /* Reference count pointer */ 84 uint64_t cmih_msrsrc; /* MSR data source flags */ 85 void *cmih_hdlpriv; /* cmi_hw.c private data */ 86 void *cmih_spec; /* cmi_hdl_{set,get}_specific */ 87 void *cmih_cmi; /* cpu mod control structure */ 88 void *cmih_cmidata; /* cpu mod private data */ 89 const struct cmi_mc_ops *cmih_mcops; /* Memory-controller ops */ 90 void *cmih_mcdata; /* Memory-controller data */ 91 uint64_t cmih_flags; /* See CMIH_F_* below */ 92 uint16_t cmih_smbiosid; /* SMBIOS Type 4 struct ID */ 93 uint_t cmih_smb_chipid; /* SMBIOS factored chipid */ 94 nvlist_t *cmih_smb_bboard; /* SMBIOS bboard nvlist */ 95 } cmi_hdl_impl_t; 96 97 #define IMPLHDL(ophdl) ((cmi_hdl_impl_t *)ophdl) 98 #define HDLOPS(hdl) ((hdl)->cmih_ops) 99 100 #define CMIH_F_INJACTV 0x1ULL 101 #define CMIH_F_DEAD 0x2ULL 102 103 /* 104 * Ops structure for handle operations. 105 */ 106 struct cmi_hdl_ops { 107 /* 108 * These ops are required in an implementation. 109 */ 110 uint_t (*cmio_vendor)(cmi_hdl_impl_t *); 111 const char *(*cmio_vendorstr)(cmi_hdl_impl_t *); 112 uint_t (*cmio_family)(cmi_hdl_impl_t *); 113 uint_t (*cmio_model)(cmi_hdl_impl_t *); 114 uint_t (*cmio_stepping)(cmi_hdl_impl_t *); 115 uint_t (*cmio_chipid)(cmi_hdl_impl_t *); 116 uint_t (*cmio_procnodeid)(cmi_hdl_impl_t *); 117 uint_t (*cmio_coreid)(cmi_hdl_impl_t *); 118 uint_t (*cmio_strandid)(cmi_hdl_impl_t *); 119 uint_t (*cmio_procnodes_per_pkg)(cmi_hdl_impl_t *); 120 uint_t (*cmio_strand_apicid)(cmi_hdl_impl_t *); 121 x86_chiprev_t (*cmio_chiprev)(cmi_hdl_impl_t *); 122 const char *(*cmio_chiprevstr)(cmi_hdl_impl_t *); 123 uint32_t (*cmio_getsockettype)(cmi_hdl_impl_t *); 124 const char *(*cmio_getsocketstr)(cmi_hdl_impl_t *); 125 uint_t (*cmio_chipsig)(cmi_hdl_impl_t *); 126 127 id_t (*cmio_logical_id)(cmi_hdl_impl_t *); 128 /* 129 * These ops are optional in an implementation. 130 */ 131 ulong_t (*cmio_getcr4)(cmi_hdl_impl_t *); 132 void (*cmio_setcr4)(cmi_hdl_impl_t *, ulong_t); 133 cmi_errno_t (*cmio_rdmsr)(cmi_hdl_impl_t *, uint_t, uint64_t *); 134 cmi_errno_t (*cmio_wrmsr)(cmi_hdl_impl_t *, uint_t, uint64_t); 135 cmi_errno_t (*cmio_msrinterpose)(cmi_hdl_impl_t *, uint_t, uint64_t); 136 void (*cmio_int)(cmi_hdl_impl_t *, int); 137 int (*cmio_online)(cmi_hdl_impl_t *, int, int *); 138 uint16_t (*cmio_smbiosid) (cmi_hdl_impl_t *); 139 uint_t (*cmio_smb_chipid)(cmi_hdl_impl_t *); 140 nvlist_t *(*cmio_smb_bboard)(cmi_hdl_impl_t *); 141 }; 142 143 static const struct cmi_hdl_ops cmi_hdl_ops; 144 145 /* 146 * Handles are looked up from contexts such as polling, injection etc 147 * where the context is reasonably well defined (although a poller could 148 * interrupt any old thread holding any old lock). They are also looked 149 * up by machine check handlers, which may strike at inconvenient times 150 * such as during handle initialization or destruction or during handle 151 * lookup (which the #MC handler itself will also have to perform). 152 * 153 * So keeping handles in a linked list makes locking difficult when we 154 * consider #MC handlers. Our solution is to have a look-up table indexed 155 * by that which uniquely identifies a handle - chip/core/strand id - 156 * with each entry a structure including a pointer to a handle 157 * structure for the resource, and a reference count for the handle. 158 * Reference counts are modified atomically. The public cmi_hdl_hold 159 * always succeeds because this can only be used after handle creation 160 * and before the call to destruct, so the hold count is already at least one. 161 * In other functions that lookup a handle (cmi_hdl_lookup, cmi_hdl_any) 162 * we must be certain that the count has not already decrmented to zero 163 * before applying our hold. 164 * 165 * The table is an array of maximum number of chips defined in 166 * CMI_CHIPID_ARR_SZ indexed by the chip id. If the chip is not present, the 167 * entry is NULL. Each entry is a pointer to another array which contains a 168 * list of all strands of the chip. This first level table is allocated when 169 * first we want to populate an entry. The size of the latter (per chip) table 170 * is CMI_MAX_STRANDS_PER_CHIP and it is populated when one of its cpus starts. 171 * 172 * Ideally we should only allocate to the actual number of chips, cores per 173 * chip and strand per core. The number of chips is not available until all 174 * of them are passed. The number of cores and strands are partially available. 175 * For now we stick with the above approach. 176 */ 177 #define CMI_MAX_CHIPID_NBITS 6 /* max chipid of 63 */ 178 #define CMI_MAX_CORES_PER_CHIP_NBITS 4 /* 16 cores per chip max */ 179 #define CMI_MAX_STRANDS_PER_CORE_NBITS 3 /* 8 strands per core max */ 180 181 #define CMI_MAX_CHIPID ((1 << (CMI_MAX_CHIPID_NBITS)) - 1) 182 #define CMI_MAX_CORES_PER_CHIP(cbits) (1 << (cbits)) 183 #define CMI_MAX_COREID(cbits) ((1 << (cbits)) - 1) 184 #define CMI_MAX_STRANDS_PER_CORE(sbits) (1 << (sbits)) 185 #define CMI_MAX_STRANDID(sbits) ((1 << (sbits)) - 1) 186 #define CMI_MAX_STRANDS_PER_CHIP(cbits, sbits) \ 187 (CMI_MAX_CORES_PER_CHIP(cbits) * CMI_MAX_STRANDS_PER_CORE(sbits)) 188 189 #define CMI_CHIPID_ARR_SZ (1 << CMI_MAX_CHIPID_NBITS) 190 191 typedef struct cmi_hdl_ent { 192 volatile uint32_t cmae_refcnt; 193 cmi_hdl_impl_t *cmae_hdlp; 194 } cmi_hdl_ent_t; 195 196 static cmi_hdl_ent_t *cmi_chip_tab[CMI_CHIPID_ARR_SZ]; 197 198 /* 199 * Default values for the number of core and strand bits. 200 */ 201 uint_t cmi_core_nbits = CMI_MAX_CORES_PER_CHIP_NBITS; 202 uint_t cmi_strand_nbits = CMI_MAX_STRANDS_PER_CORE_NBITS; 203 static int cmi_ext_topo_check = 0; 204 205 /* 206 * Controls where we will source PCI config space data. 207 */ 208 #define CMI_PCICFG_FLAG_RD_HWOK 0x0001 209 #define CMI_PCICFG_FLAG_RD_INTERPOSEOK 0X0002 210 #define CMI_PCICFG_FLAG_WR_HWOK 0x0004 211 #define CMI_PCICFG_FLAG_WR_INTERPOSEOK 0X0008 212 213 static uint64_t cmi_pcicfg_flags = 214 CMI_PCICFG_FLAG_RD_HWOK | CMI_PCICFG_FLAG_RD_INTERPOSEOK | 215 CMI_PCICFG_FLAG_WR_HWOK | CMI_PCICFG_FLAG_WR_INTERPOSEOK; 216 217 /* 218 * The flags for individual cpus are kept in their per-cpu handle cmih_msrsrc 219 */ 220 #define CMI_MSR_FLAG_RD_HWOK 0x0001 221 #define CMI_MSR_FLAG_RD_INTERPOSEOK 0x0002 222 #define CMI_MSR_FLAG_WR_HWOK 0x0004 223 #define CMI_MSR_FLAG_WR_INTERPOSEOK 0x0008 224 225 int cmi_call_func_ntv_tries = 3; 226 227 static cmi_errno_t 228 call_func_ntv(int cpuid, xc_func_t func, xc_arg_t arg1, xc_arg_t arg2) 229 { 230 cmi_errno_t rc = -1; 231 int i; 232 233 kpreempt_disable(); 234 235 if (CPU->cpu_id == cpuid) { 236 (*func)(arg1, arg2, (xc_arg_t)&rc); 237 } else { 238 /* 239 * This should not happen for a #MC trap or a poll, so 240 * this is likely an error injection or similar. 241 * We will try to cross call with xc_trycall - we 242 * can't guarantee success with xc_call because 243 * the interrupt code in the case of a #MC may 244 * already hold the xc mutex. 245 */ 246 for (i = 0; i < cmi_call_func_ntv_tries; i++) { 247 cpuset_t cpus; 248 249 CPUSET_ONLY(cpus, cpuid); 250 xc_priority(arg1, arg2, (xc_arg_t)&rc, 251 CPUSET2BV(cpus), func); 252 if (rc != -1) 253 break; 254 255 DELAY(1); 256 } 257 } 258 259 kpreempt_enable(); 260 261 return (rc != -1 ? rc : CMIERR_DEADLOCK); 262 } 263 264 static uint64_t injcnt; 265 266 void 267 cmi_hdl_inj_begin(cmi_hdl_t ophdl) 268 { 269 cmi_hdl_impl_t *hdl = IMPLHDL(ophdl); 270 271 if (hdl != NULL) 272 hdl->cmih_flags |= CMIH_F_INJACTV; 273 if (injcnt++ == 0) { 274 cmn_err(CE_NOTE, "Hardware error injection/simulation " 275 "activity noted"); 276 } 277 } 278 279 void 280 cmi_hdl_inj_end(cmi_hdl_t ophdl) 281 { 282 cmi_hdl_impl_t *hdl = IMPLHDL(ophdl); 283 284 ASSERT(hdl == NULL || hdl->cmih_flags & CMIH_F_INJACTV); 285 if (hdl != NULL) 286 hdl->cmih_flags &= ~CMIH_F_INJACTV; 287 } 288 289 boolean_t 290 cmi_inj_tainted(void) 291 { 292 return (injcnt != 0 ? B_TRUE : B_FALSE); 293 } 294 295 /* 296 * ======================================================= 297 * | MSR Interposition | 298 * | ----------------- | 299 * | | 300 * ------------------------------------------------------- 301 */ 302 303 #define CMI_MSRI_HASHSZ 16 304 #define CMI_MSRI_HASHIDX(hdl, msr) \ 305 ((((uintptr_t)(hdl) >> 3) + (msr)) % (CMI_MSRI_HASHSZ - 1)) 306 307 struct cmi_msri_bkt { 308 kmutex_t msrib_lock; 309 struct cmi_msri_hashent *msrib_head; 310 }; 311 312 struct cmi_msri_hashent { 313 struct cmi_msri_hashent *msrie_next; 314 struct cmi_msri_hashent *msrie_prev; 315 cmi_hdl_impl_t *msrie_hdl; 316 uint_t msrie_msrnum; 317 uint64_t msrie_msrval; 318 }; 319 320 #define CMI_MSRI_MATCH(ent, hdl, req_msr) \ 321 ((ent)->msrie_hdl == (hdl) && (ent)->msrie_msrnum == (req_msr)) 322 323 static struct cmi_msri_bkt msrihash[CMI_MSRI_HASHSZ]; 324 325 static void 326 msri_addent(cmi_hdl_impl_t *hdl, uint_t msr, uint64_t val) 327 { 328 int idx = CMI_MSRI_HASHIDX(hdl, msr); 329 struct cmi_msri_bkt *hbp = &msrihash[idx]; 330 struct cmi_msri_hashent *hep; 331 332 mutex_enter(&hbp->msrib_lock); 333 334 for (hep = hbp->msrib_head; hep != NULL; hep = hep->msrie_next) { 335 if (CMI_MSRI_MATCH(hep, hdl, msr)) 336 break; 337 } 338 339 if (hep != NULL) { 340 hep->msrie_msrval = val; 341 } else { 342 hep = kmem_alloc(sizeof (*hep), KM_SLEEP); 343 hep->msrie_hdl = hdl; 344 hep->msrie_msrnum = msr; 345 hep->msrie_msrval = val; 346 347 if (hbp->msrib_head != NULL) 348 hbp->msrib_head->msrie_prev = hep; 349 hep->msrie_next = hbp->msrib_head; 350 hep->msrie_prev = NULL; 351 hbp->msrib_head = hep; 352 } 353 354 mutex_exit(&hbp->msrib_lock); 355 } 356 357 /* 358 * Look for a match for the given hanlde and msr. Return 1 with valp 359 * filled if a match is found, otherwise return 0 with valp untouched. 360 */ 361 static int 362 msri_lookup(cmi_hdl_impl_t *hdl, uint_t msr, uint64_t *valp) 363 { 364 int idx = CMI_MSRI_HASHIDX(hdl, msr); 365 struct cmi_msri_bkt *hbp = &msrihash[idx]; 366 struct cmi_msri_hashent *hep; 367 368 /* 369 * This function is called during #MC trap handling, so we should 370 * consider the possibility that the hash mutex is held by the 371 * interrupted thread. This should not happen because interposition 372 * is an artificial injection mechanism and the #MC is requested 373 * after adding entries, but just in case of a real #MC at an 374 * unlucky moment we'll use mutex_tryenter here. 375 */ 376 if (!mutex_tryenter(&hbp->msrib_lock)) 377 return (0); 378 379 for (hep = hbp->msrib_head; hep != NULL; hep = hep->msrie_next) { 380 if (CMI_MSRI_MATCH(hep, hdl, msr)) { 381 *valp = hep->msrie_msrval; 382 break; 383 } 384 } 385 386 mutex_exit(&hbp->msrib_lock); 387 388 return (hep != NULL); 389 } 390 391 /* 392 * Remove any interposed value that matches. 393 */ 394 static void 395 msri_rment(cmi_hdl_impl_t *hdl, uint_t msr) 396 { 397 398 int idx = CMI_MSRI_HASHIDX(hdl, msr); 399 struct cmi_msri_bkt *hbp = &msrihash[idx]; 400 struct cmi_msri_hashent *hep; 401 402 if (!mutex_tryenter(&hbp->msrib_lock)) 403 return; 404 405 for (hep = hbp->msrib_head; hep != NULL; hep = hep->msrie_next) { 406 if (CMI_MSRI_MATCH(hep, hdl, msr)) { 407 if (hep->msrie_prev != NULL) 408 hep->msrie_prev->msrie_next = hep->msrie_next; 409 410 if (hep->msrie_next != NULL) 411 hep->msrie_next->msrie_prev = hep->msrie_prev; 412 413 if (hbp->msrib_head == hep) 414 hbp->msrib_head = hep->msrie_next; 415 416 kmem_free(hep, sizeof (*hep)); 417 break; 418 } 419 } 420 421 mutex_exit(&hbp->msrib_lock); 422 } 423 424 /* 425 * ======================================================= 426 * | PCI Config Space Interposition | 427 * | ------------------------------ | 428 * | | 429 * ------------------------------------------------------- 430 */ 431 432 /* 433 * Hash for interposed PCI config space values. We lookup on bus/dev/fun/offset 434 * and then record whether the value stashed was made with a byte, word or 435 * doubleword access; we will only return a hit for an access of the 436 * same size. If you access say a 32-bit register using byte accesses 437 * and then attempt to read the full 32-bit value back you will not obtain 438 * any sort of merged result - you get a lookup miss. 439 */ 440 441 #define CMI_PCII_HASHSZ 16 442 #define CMI_PCII_HASHIDX(b, d, f, o) \ 443 (((b) + (d) + (f) + (o)) % (CMI_PCII_HASHSZ - 1)) 444 445 struct cmi_pcii_bkt { 446 kmutex_t pciib_lock; 447 struct cmi_pcii_hashent *pciib_head; 448 }; 449 450 struct cmi_pcii_hashent { 451 struct cmi_pcii_hashent *pcii_next; 452 struct cmi_pcii_hashent *pcii_prev; 453 int pcii_bus; 454 int pcii_dev; 455 int pcii_func; 456 int pcii_reg; 457 int pcii_asize; 458 uint32_t pcii_val; 459 }; 460 461 #define CMI_PCII_MATCH(ent, b, d, f, r, asz) \ 462 ((ent)->pcii_bus == (b) && (ent)->pcii_dev == (d) && \ 463 (ent)->pcii_func == (f) && (ent)->pcii_reg == (r) && \ 464 (ent)->pcii_asize == (asz)) 465 466 static struct cmi_pcii_bkt pciihash[CMI_PCII_HASHSZ]; 467 468 469 /* 470 * Add a new entry to the PCI interpose hash, overwriting any existing 471 * entry that is found. 472 */ 473 static void 474 pcii_addent(int bus, int dev, int func, int reg, uint32_t val, int asz) 475 { 476 int idx = CMI_PCII_HASHIDX(bus, dev, func, reg); 477 struct cmi_pcii_bkt *hbp = &pciihash[idx]; 478 struct cmi_pcii_hashent *hep; 479 480 cmi_hdl_inj_begin(NULL); 481 482 mutex_enter(&hbp->pciib_lock); 483 484 for (hep = hbp->pciib_head; hep != NULL; hep = hep->pcii_next) { 485 if (CMI_PCII_MATCH(hep, bus, dev, func, reg, asz)) 486 break; 487 } 488 489 if (hep != NULL) { 490 hep->pcii_val = val; 491 } else { 492 hep = kmem_alloc(sizeof (*hep), KM_SLEEP); 493 hep->pcii_bus = bus; 494 hep->pcii_dev = dev; 495 hep->pcii_func = func; 496 hep->pcii_reg = reg; 497 hep->pcii_asize = asz; 498 hep->pcii_val = val; 499 500 if (hbp->pciib_head != NULL) 501 hbp->pciib_head->pcii_prev = hep; 502 hep->pcii_next = hbp->pciib_head; 503 hep->pcii_prev = NULL; 504 hbp->pciib_head = hep; 505 } 506 507 mutex_exit(&hbp->pciib_lock); 508 509 cmi_hdl_inj_end(NULL); 510 } 511 512 /* 513 * Look for a match for the given bus/dev/func/reg; return 1 with valp 514 * filled if a match is found, otherwise return 0 with valp untouched. 515 */ 516 static int 517 pcii_lookup(int bus, int dev, int func, int reg, int asz, uint32_t *valp) 518 { 519 int idx = CMI_PCII_HASHIDX(bus, dev, func, reg); 520 struct cmi_pcii_bkt *hbp = &pciihash[idx]; 521 struct cmi_pcii_hashent *hep; 522 523 if (!mutex_tryenter(&hbp->pciib_lock)) 524 return (0); 525 526 for (hep = hbp->pciib_head; hep != NULL; hep = hep->pcii_next) { 527 if (CMI_PCII_MATCH(hep, bus, dev, func, reg, asz)) { 528 *valp = hep->pcii_val; 529 break; 530 } 531 } 532 533 mutex_exit(&hbp->pciib_lock); 534 535 return (hep != NULL); 536 } 537 538 static void 539 pcii_rment(int bus, int dev, int func, int reg, int asz) 540 { 541 int idx = CMI_PCII_HASHIDX(bus, dev, func, reg); 542 struct cmi_pcii_bkt *hbp = &pciihash[idx]; 543 struct cmi_pcii_hashent *hep; 544 545 mutex_enter(&hbp->pciib_lock); 546 547 for (hep = hbp->pciib_head; hep != NULL; hep = hep->pcii_next) { 548 if (CMI_PCII_MATCH(hep, bus, dev, func, reg, asz)) { 549 if (hep->pcii_prev != NULL) 550 hep->pcii_prev->pcii_next = hep->pcii_next; 551 552 if (hep->pcii_next != NULL) 553 hep->pcii_next->pcii_prev = hep->pcii_prev; 554 555 if (hbp->pciib_head == hep) 556 hbp->pciib_head = hep->pcii_next; 557 558 kmem_free(hep, sizeof (*hep)); 559 break; 560 } 561 } 562 563 mutex_exit(&hbp->pciib_lock); 564 } 565 566 #ifndef __xpv 567 568 /* 569 * ======================================================= 570 * | Native methods | 571 * | -------------- | 572 * | | 573 * | These are used when we are running native on bare- | 574 * | metal, or simply don't know any better. | 575 * --------------------------------------------------------- 576 */ 577 578 #define HDLPRIV(hdl) ((cpu_t *)(hdl)->cmih_hdlpriv) 579 580 static uint_t 581 ntv_vendor(cmi_hdl_impl_t *hdl) 582 { 583 return (cpuid_getvendor(HDLPRIV(hdl))); 584 } 585 586 static const char * 587 ntv_vendorstr(cmi_hdl_impl_t *hdl) 588 { 589 return (cpuid_getvendorstr(HDLPRIV(hdl))); 590 } 591 592 static uint_t 593 ntv_family(cmi_hdl_impl_t *hdl) 594 { 595 return (cpuid_getfamily(HDLPRIV(hdl))); 596 } 597 598 static uint_t 599 ntv_model(cmi_hdl_impl_t *hdl) 600 { 601 return (cpuid_getmodel(HDLPRIV(hdl))); 602 } 603 604 static uint_t 605 ntv_stepping(cmi_hdl_impl_t *hdl) 606 { 607 return (cpuid_getstep(HDLPRIV(hdl))); 608 } 609 610 static uint_t 611 ntv_chipid(cmi_hdl_impl_t *hdl) 612 { 613 return (hdl->cmih_chipid); 614 615 } 616 617 static uint_t 618 ntv_procnodeid(cmi_hdl_impl_t *hdl) 619 { 620 return (hdl->cmih_procnodeid); 621 } 622 623 static uint_t 624 ntv_procnodes_per_pkg(cmi_hdl_impl_t *hdl) 625 { 626 return (hdl->cmih_procnodes_per_pkg); 627 } 628 629 static uint_t 630 ntv_coreid(cmi_hdl_impl_t *hdl) 631 { 632 return (hdl->cmih_coreid); 633 } 634 635 static uint_t 636 ntv_strandid(cmi_hdl_impl_t *hdl) 637 { 638 return (hdl->cmih_strandid); 639 } 640 641 static uint_t 642 ntv_strand_apicid(cmi_hdl_impl_t *hdl) 643 { 644 return (cpuid_get_apicid(HDLPRIV(hdl))); 645 } 646 647 static uint16_t 648 ntv_smbiosid(cmi_hdl_impl_t *hdl) 649 { 650 return (hdl->cmih_smbiosid); 651 } 652 653 static uint_t 654 ntv_smb_chipid(cmi_hdl_impl_t *hdl) 655 { 656 return (hdl->cmih_smb_chipid); 657 } 658 659 static nvlist_t * 660 ntv_smb_bboard(cmi_hdl_impl_t *hdl) 661 { 662 return (hdl->cmih_smb_bboard); 663 } 664 665 static x86_chiprev_t 666 ntv_chiprev(cmi_hdl_impl_t *hdl) 667 { 668 return (cpuid_getchiprev(HDLPRIV(hdl))); 669 } 670 671 static const char * 672 ntv_chiprevstr(cmi_hdl_impl_t *hdl) 673 { 674 return (cpuid_getchiprevstr(HDLPRIV(hdl))); 675 } 676 677 static uint32_t 678 ntv_getsockettype(cmi_hdl_impl_t *hdl) 679 { 680 return (cpuid_getsockettype(HDLPRIV(hdl))); 681 } 682 683 static const char * 684 ntv_getsocketstr(cmi_hdl_impl_t *hdl) 685 { 686 return (cpuid_getsocketstr(HDLPRIV(hdl))); 687 } 688 689 static uint_t 690 ntv_chipsig(cmi_hdl_impl_t *hdl) 691 { 692 return (cpuid_getsig(HDLPRIV(hdl))); 693 } 694 695 static id_t 696 ntv_logical_id(cmi_hdl_impl_t *hdl) 697 { 698 return (HDLPRIV(hdl)->cpu_id); 699 } 700 701 /*ARGSUSED*/ 702 static int 703 ntv_getcr4_xc(xc_arg_t arg1, xc_arg_t arg2, xc_arg_t arg3) 704 { 705 ulong_t *dest = (ulong_t *)arg1; 706 cmi_errno_t *rcp = (cmi_errno_t *)arg3; 707 708 *dest = getcr4(); 709 *rcp = CMI_SUCCESS; 710 711 return (0); 712 } 713 714 static ulong_t 715 ntv_getcr4(cmi_hdl_impl_t *hdl) 716 { 717 cpu_t *cp = HDLPRIV(hdl); 718 ulong_t val; 719 720 (void) call_func_ntv(cp->cpu_id, ntv_getcr4_xc, (xc_arg_t)&val, 0); 721 722 return (val); 723 } 724 725 /*ARGSUSED*/ 726 static int 727 ntv_setcr4_xc(xc_arg_t arg1, xc_arg_t arg2, xc_arg_t arg3) 728 { 729 ulong_t val = (ulong_t)arg1; 730 cmi_errno_t *rcp = (cmi_errno_t *)arg3; 731 732 setcr4(val); 733 *rcp = CMI_SUCCESS; 734 735 return (0); 736 } 737 738 static void 739 ntv_setcr4(cmi_hdl_impl_t *hdl, ulong_t val) 740 { 741 cpu_t *cp = HDLPRIV(hdl); 742 743 (void) call_func_ntv(cp->cpu_id, ntv_setcr4_xc, (xc_arg_t)val, 0); 744 } 745 746 volatile uint32_t cmi_trapped_rdmsr; 747 748 /*ARGSUSED*/ 749 static int 750 ntv_rdmsr_xc(xc_arg_t arg1, xc_arg_t arg2, xc_arg_t arg3) 751 { 752 uint_t msr = (uint_t)arg1; 753 uint64_t *valp = (uint64_t *)arg2; 754 cmi_errno_t *rcp = (cmi_errno_t *)arg3; 755 756 on_trap_data_t otd; 757 758 if (on_trap(&otd, OT_DATA_ACCESS) == 0) { 759 if (checked_rdmsr(msr, valp) == 0) 760 *rcp = CMI_SUCCESS; 761 else 762 *rcp = CMIERR_NOTSUP; 763 } else { 764 *rcp = CMIERR_MSRGPF; 765 atomic_inc_32(&cmi_trapped_rdmsr); 766 } 767 no_trap(); 768 769 return (0); 770 } 771 772 static cmi_errno_t 773 ntv_rdmsr(cmi_hdl_impl_t *hdl, uint_t msr, uint64_t *valp) 774 { 775 cpu_t *cp = HDLPRIV(hdl); 776 777 if (!(hdl->cmih_msrsrc & CMI_MSR_FLAG_RD_HWOK)) 778 return (CMIERR_INTERPOSE); 779 780 return (call_func_ntv(cp->cpu_id, ntv_rdmsr_xc, 781 (xc_arg_t)msr, (xc_arg_t)valp)); 782 } 783 784 volatile uint32_t cmi_trapped_wrmsr; 785 786 /*ARGSUSED*/ 787 static int 788 ntv_wrmsr_xc(xc_arg_t arg1, xc_arg_t arg2, xc_arg_t arg3) 789 { 790 uint_t msr = (uint_t)arg1; 791 uint64_t val = *((uint64_t *)arg2); 792 cmi_errno_t *rcp = (cmi_errno_t *)arg3; 793 on_trap_data_t otd; 794 795 if (on_trap(&otd, OT_DATA_ACCESS) == 0) { 796 if (checked_wrmsr(msr, val) == 0) 797 *rcp = CMI_SUCCESS; 798 else 799 *rcp = CMIERR_NOTSUP; 800 } else { 801 *rcp = CMIERR_MSRGPF; 802 atomic_inc_32(&cmi_trapped_wrmsr); 803 } 804 no_trap(); 805 806 return (0); 807 808 } 809 810 static cmi_errno_t 811 ntv_wrmsr(cmi_hdl_impl_t *hdl, uint_t msr, uint64_t val) 812 { 813 cpu_t *cp = HDLPRIV(hdl); 814 815 if (!(hdl->cmih_msrsrc & CMI_MSR_FLAG_WR_HWOK)) 816 return (CMI_SUCCESS); 817 818 return (call_func_ntv(cp->cpu_id, ntv_wrmsr_xc, 819 (xc_arg_t)msr, (xc_arg_t)&val)); 820 } 821 822 static cmi_errno_t 823 ntv_msrinterpose(cmi_hdl_impl_t *hdl, uint_t msr, uint64_t val) 824 { 825 msri_addent(hdl, msr, val); 826 return (CMI_SUCCESS); 827 } 828 829 /*ARGSUSED*/ 830 static int 831 ntv_int_xc(xc_arg_t arg1, xc_arg_t arg2, xc_arg_t arg3) 832 { 833 cmi_errno_t *rcp = (cmi_errno_t *)arg3; 834 int int_no = (int)arg1; 835 836 if (int_no == T_MCE) 837 int18(); 838 else 839 int_cmci(); 840 *rcp = CMI_SUCCESS; 841 842 return (0); 843 } 844 845 static void 846 ntv_int(cmi_hdl_impl_t *hdl, int int_no) 847 { 848 cpu_t *cp = HDLPRIV(hdl); 849 850 (void) call_func_ntv(cp->cpu_id, ntv_int_xc, (xc_arg_t)int_no, 0); 851 } 852 853 static int 854 ntv_online(cmi_hdl_impl_t *hdl, int new_status, int *old_status) 855 { 856 int rc; 857 processorid_t cpuid = HDLPRIV(hdl)->cpu_id; 858 859 while (mutex_tryenter(&cpu_lock) == 0) { 860 if (hdl->cmih_flags & CMIH_F_DEAD) 861 return (EBUSY); 862 delay(1); 863 } 864 rc = p_online_internal_locked(cpuid, new_status, old_status); 865 mutex_exit(&cpu_lock); 866 867 return (rc); 868 } 869 870 #else /* __xpv */ 871 872 /* 873 * ======================================================= 874 * | xVM dom0 methods | 875 * | ---------------- | 876 * | | 877 * | These are used when we are running as dom0 in | 878 * | a Solaris xVM context. | 879 * --------------------------------------------------------- 880 */ 881 882 #define HDLPRIV(hdl) ((xen_mc_lcpu_cookie_t)(hdl)->cmih_hdlpriv) 883 884 extern uint_t _cpuid_vendorstr_to_vendorcode(char *); 885 886 887 static uint_t 888 xpv_vendor(cmi_hdl_impl_t *hdl) 889 { 890 return (_cpuid_vendorstr_to_vendorcode((char *)xen_physcpu_vendorstr( 891 HDLPRIV(hdl)))); 892 } 893 894 static const char * 895 xpv_vendorstr(cmi_hdl_impl_t *hdl) 896 { 897 return (xen_physcpu_vendorstr(HDLPRIV(hdl))); 898 } 899 900 static uint_t 901 xpv_family(cmi_hdl_impl_t *hdl) 902 { 903 return (xen_physcpu_family(HDLPRIV(hdl))); 904 } 905 906 static uint_t 907 xpv_model(cmi_hdl_impl_t *hdl) 908 { 909 return (xen_physcpu_model(HDLPRIV(hdl))); 910 } 911 912 static uint_t 913 xpv_stepping(cmi_hdl_impl_t *hdl) 914 { 915 return (xen_physcpu_stepping(HDLPRIV(hdl))); 916 } 917 918 static uint_t 919 xpv_chipid(cmi_hdl_impl_t *hdl) 920 { 921 return (hdl->cmih_chipid); 922 } 923 924 static uint_t 925 xpv_procnodeid(cmi_hdl_impl_t *hdl) 926 { 927 return (hdl->cmih_procnodeid); 928 } 929 930 static uint_t 931 xpv_procnodes_per_pkg(cmi_hdl_impl_t *hdl) 932 { 933 return (hdl->cmih_procnodes_per_pkg); 934 } 935 936 static uint_t 937 xpv_coreid(cmi_hdl_impl_t *hdl) 938 { 939 return (hdl->cmih_coreid); 940 } 941 942 static uint_t 943 xpv_strandid(cmi_hdl_impl_t *hdl) 944 { 945 return (hdl->cmih_strandid); 946 } 947 948 static uint_t 949 xpv_strand_apicid(cmi_hdl_impl_t *hdl) 950 { 951 return (xen_physcpu_initial_apicid(HDLPRIV(hdl))); 952 } 953 954 static uint16_t 955 xpv_smbiosid(cmi_hdl_impl_t *hdl) 956 { 957 return (hdl->cmih_smbiosid); 958 } 959 960 static uint_t 961 xpv_smb_chipid(cmi_hdl_impl_t *hdl) 962 { 963 return (hdl->cmih_smb_chipid); 964 } 965 966 static nvlist_t * 967 xpv_smb_bboard(cmi_hdl_impl_t *hdl) 968 { 969 return (hdl->cmih_smb_bboard); 970 } 971 972 extern x86_chiprev_t _cpuid_chiprev(uint_t, uint_t, uint_t, uint_t); 973 974 static x86_chiprev_t 975 xpv_chiprev(cmi_hdl_impl_t *hdl) 976 { 977 return (_cpuid_chiprev(xpv_vendor(hdl), xpv_family(hdl), 978 xpv_model(hdl), xpv_stepping(hdl))); 979 } 980 981 extern const char *_cpuid_chiprevstr(uint_t, uint_t, uint_t, uint_t); 982 983 static const char * 984 xpv_chiprevstr(cmi_hdl_impl_t *hdl) 985 { 986 return (_cpuid_chiprevstr(xpv_vendor(hdl), xpv_family(hdl), 987 xpv_model(hdl), xpv_stepping(hdl))); 988 } 989 990 extern uint32_t _cpuid_skt(uint_t, uint_t, uint_t, uint_t); 991 992 static uint32_t 993 xpv_getsockettype(cmi_hdl_impl_t *hdl) 994 { 995 return (_cpuid_skt(xpv_vendor(hdl), xpv_family(hdl), 996 xpv_model(hdl), xpv_stepping(hdl))); 997 } 998 999 extern const char *_cpuid_sktstr(uint_t, uint_t, uint_t, uint_t); 1000 1001 static const char * 1002 xpv_getsocketstr(cmi_hdl_impl_t *hdl) 1003 { 1004 return (_cpuid_sktstr(xpv_vendor(hdl), xpv_family(hdl), 1005 xpv_model(hdl), xpv_stepping(hdl))); 1006 } 1007 1008 /* ARGSUSED */ 1009 static uint_t 1010 xpv_chipsig(cmi_hdl_impl_t *hdl) 1011 { 1012 return (0); 1013 } 1014 1015 static id_t 1016 xpv_logical_id(cmi_hdl_impl_t *hdl) 1017 { 1018 return (xen_physcpu_logical_id(HDLPRIV(hdl))); 1019 } 1020 1021 static cmi_errno_t 1022 xpv_rdmsr(cmi_hdl_impl_t *hdl, uint_t msr, uint64_t *valp) 1023 { 1024 switch (msr) { 1025 case IA32_MSR_MCG_CAP: 1026 *valp = xen_physcpu_mcg_cap(HDLPRIV(hdl)); 1027 break; 1028 1029 default: 1030 return (CMIERR_NOTSUP); 1031 } 1032 1033 return (CMI_SUCCESS); 1034 } 1035 1036 /* 1037 * Request the hypervisor to write an MSR for us. The hypervisor 1038 * will only accept MCA-related MSRs, as this is for MCA error 1039 * simulation purposes alone. We will pre-screen MSRs for injection 1040 * so we don't bother the HV with bogus requests. We will permit 1041 * injection to any MCA bank register, and to MCG_STATUS. 1042 */ 1043 1044 #define IS_MCA_INJ_MSR(msr) \ 1045 (((msr) >= IA32_MSR_MC(0, CTL) && (msr) <= IA32_MSR_MC(10, MISC)) || \ 1046 (msr) == IA32_MSR_MCG_STATUS) 1047 1048 static cmi_errno_t 1049 xpv_wrmsr_cmn(cmi_hdl_impl_t *hdl, uint_t msr, uint64_t val, boolean_t intpose) 1050 { 1051 xen_mc_t xmc; 1052 struct xen_mc_msrinject *mci = &xmc.u.mc_msrinject; 1053 1054 if (!(hdl->cmih_flags & CMIH_F_INJACTV)) 1055 return (CMIERR_NOTSUP); /* for injection use only! */ 1056 1057 if (!IS_MCA_INJ_MSR(msr)) 1058 return (CMIERR_API); 1059 1060 if (panicstr) 1061 return (CMIERR_DEADLOCK); 1062 1063 mci->mcinj_cpunr = xen_physcpu_logical_id(HDLPRIV(hdl)); 1064 mci->mcinj_flags = intpose ? MC_MSRINJ_F_INTERPOSE : 0; 1065 mci->mcinj_count = 1; /* learn to batch sometime */ 1066 mci->mcinj_msr[0].reg = msr; 1067 mci->mcinj_msr[0].value = val; 1068 1069 return (HYPERVISOR_mca(XEN_MC_msrinject, &xmc) == 1070 0 ? CMI_SUCCESS : CMIERR_NOTSUP); 1071 } 1072 1073 static cmi_errno_t 1074 xpv_wrmsr(cmi_hdl_impl_t *hdl, uint_t msr, uint64_t val) 1075 { 1076 return (xpv_wrmsr_cmn(hdl, msr, val, B_FALSE)); 1077 } 1078 1079 1080 static cmi_errno_t 1081 xpv_msrinterpose(cmi_hdl_impl_t *hdl, uint_t msr, uint64_t val) 1082 { 1083 return (xpv_wrmsr_cmn(hdl, msr, val, B_TRUE)); 1084 } 1085 1086 static void 1087 xpv_int(cmi_hdl_impl_t *hdl, int int_no) 1088 { 1089 xen_mc_t xmc; 1090 struct xen_mc_mceinject *mce = &xmc.u.mc_mceinject; 1091 1092 if (!(hdl->cmih_flags & CMIH_F_INJACTV)) 1093 return; 1094 1095 if (int_no != T_MCE) { 1096 cmn_err(CE_WARN, "xpv_int: int_no %d unimplemented\n", 1097 int_no); 1098 } 1099 1100 mce->mceinj_cpunr = xen_physcpu_logical_id(HDLPRIV(hdl)); 1101 1102 (void) HYPERVISOR_mca(XEN_MC_mceinject, &xmc); 1103 } 1104 1105 static int 1106 xpv_online(cmi_hdl_impl_t *hdl, int new_status, int *old_status) 1107 { 1108 xen_sysctl_t xs; 1109 int op, rc, status; 1110 1111 new_status &= ~P_FORCED; 1112 1113 switch (new_status) { 1114 case P_STATUS: 1115 op = XEN_SYSCTL_CPU_HOTPLUG_STATUS; 1116 break; 1117 case P_FAULTED: 1118 case P_OFFLINE: 1119 op = XEN_SYSCTL_CPU_HOTPLUG_OFFLINE; 1120 break; 1121 case P_ONLINE: 1122 op = XEN_SYSCTL_CPU_HOTPLUG_ONLINE; 1123 break; 1124 default: 1125 return (-1); 1126 } 1127 1128 xs.cmd = XEN_SYSCTL_cpu_hotplug; 1129 xs.interface_version = XEN_SYSCTL_INTERFACE_VERSION; 1130 xs.u.cpu_hotplug.cpu = xen_physcpu_logical_id(HDLPRIV(hdl)); 1131 xs.u.cpu_hotplug.op = op; 1132 1133 if ((rc = HYPERVISOR_sysctl(&xs)) >= 0) { 1134 status = rc; 1135 rc = 0; 1136 switch (status) { 1137 case XEN_CPU_HOTPLUG_STATUS_NEW: 1138 *old_status = P_OFFLINE; 1139 break; 1140 case XEN_CPU_HOTPLUG_STATUS_OFFLINE: 1141 *old_status = P_FAULTED; 1142 break; 1143 case XEN_CPU_HOTPLUG_STATUS_ONLINE: 1144 *old_status = P_ONLINE; 1145 break; 1146 default: 1147 return (-1); 1148 } 1149 } 1150 1151 return (-rc); 1152 } 1153 1154 #endif 1155 1156 /*ARGSUSED*/ 1157 static void * 1158 cpu_search(enum cmi_hdl_class class, uint_t chipid, uint_t coreid, 1159 uint_t strandid) 1160 { 1161 #ifdef __xpv 1162 xen_mc_lcpu_cookie_t cpi; 1163 1164 for (cpi = xen_physcpu_next(NULL); cpi != NULL; 1165 cpi = xen_physcpu_next(cpi)) { 1166 if (xen_physcpu_chipid(cpi) == chipid && 1167 xen_physcpu_coreid(cpi) == coreid && 1168 xen_physcpu_strandid(cpi) == strandid) 1169 return ((void *)cpi); 1170 } 1171 return (NULL); 1172 1173 #else /* __xpv */ 1174 1175 cpu_t *cp, *startcp; 1176 1177 kpreempt_disable(); 1178 cp = startcp = CPU; 1179 do { 1180 if (cmi_ntv_hwchipid(cp) == chipid && 1181 cmi_ntv_hwcoreid(cp) == coreid && 1182 cmi_ntv_hwstrandid(cp) == strandid) { 1183 kpreempt_enable(); 1184 return ((void *)cp); 1185 } 1186 1187 cp = cp->cpu_next; 1188 } while (cp != startcp); 1189 kpreempt_enable(); 1190 return (NULL); 1191 #endif /* __ xpv */ 1192 } 1193 1194 static boolean_t 1195 cpu_is_cmt(void *priv) 1196 { 1197 #ifdef __xpv 1198 return (xen_physcpu_is_cmt((xen_mc_lcpu_cookie_t)priv)); 1199 #else /* __xpv */ 1200 cpu_t *cp = (cpu_t *)priv; 1201 1202 int strands_per_core = cpuid_get_ncpu_per_chip(cp) / 1203 cpuid_get_ncore_per_chip(cp); 1204 1205 return (strands_per_core > 1); 1206 #endif /* __xpv */ 1207 } 1208 1209 /* 1210 * Find the handle entry of a given cpu identified by a <chip,core,strand> 1211 * tuple. 1212 */ 1213 static cmi_hdl_ent_t * 1214 cmi_hdl_ent_lookup(uint_t chipid, uint_t coreid, uint_t strandid) 1215 { 1216 int max_strands = CMI_MAX_STRANDS_PER_CHIP(cmi_core_nbits, 1217 cmi_strand_nbits); 1218 1219 /* 1220 * Allocate per-chip table which contains a list of handle of 1221 * all strands of the chip. 1222 */ 1223 if (cmi_chip_tab[chipid] == NULL) { 1224 size_t sz; 1225 cmi_hdl_ent_t *pg; 1226 1227 sz = max_strands * sizeof (cmi_hdl_ent_t); 1228 pg = kmem_zalloc(sz, KM_SLEEP); 1229 1230 /* test and set the per-chip table if it is not allocated */ 1231 if (atomic_cas_ptr(&cmi_chip_tab[chipid], NULL, pg) != NULL) 1232 kmem_free(pg, sz); /* someone beats us */ 1233 } 1234 1235 return (cmi_chip_tab[chipid] + 1236 ((((coreid) & CMI_MAX_COREID(cmi_core_nbits)) << cmi_strand_nbits) | 1237 ((strandid) & CMI_MAX_STRANDID(cmi_strand_nbits)))); 1238 } 1239 1240 extern void cpuid_get_ext_topo(cpu_t *, uint_t *, uint_t *); 1241 1242 cmi_hdl_t 1243 cmi_hdl_create(enum cmi_hdl_class class, uint_t chipid, uint_t coreid, 1244 uint_t strandid) 1245 { 1246 cmi_hdl_impl_t *hdl; 1247 void *priv; 1248 cmi_hdl_ent_t *ent; 1249 uint_t vendor; 1250 1251 #ifdef __xpv 1252 ASSERT(class == CMI_HDL_SOLARIS_xVM_MCA); 1253 #else 1254 ASSERT(class == CMI_HDL_NATIVE); 1255 #endif 1256 1257 if ((priv = cpu_search(class, chipid, coreid, strandid)) == NULL) 1258 return (NULL); 1259 1260 /* 1261 * Assume all chips in the system are the same type. 1262 * For Intel, attempt to check if extended topology is available 1263 * CPUID.EAX=0xB. If so, get the number of core and strand bits. 1264 */ 1265 #ifdef __xpv 1266 vendor = _cpuid_vendorstr_to_vendorcode( 1267 (char *)xen_physcpu_vendorstr((xen_mc_lcpu_cookie_t)priv)); 1268 #else 1269 vendor = cpuid_getvendor((cpu_t *)priv); 1270 #endif 1271 1272 switch (vendor) { 1273 case X86_VENDOR_Intel: 1274 case X86_VENDOR_AMD: 1275 case X86_VENDOR_HYGON: 1276 if (cmi_ext_topo_check == 0) { 1277 cpuid_get_ext_topo((cpu_t *)priv, &cmi_core_nbits, 1278 &cmi_strand_nbits); 1279 cmi_ext_topo_check = 1; 1280 } 1281 default: 1282 break; 1283 } 1284 1285 if (chipid > CMI_MAX_CHIPID || 1286 coreid > CMI_MAX_COREID(cmi_core_nbits) || 1287 strandid > CMI_MAX_STRANDID(cmi_strand_nbits)) 1288 return (NULL); 1289 1290 hdl = kmem_zalloc(sizeof (*hdl), KM_SLEEP); 1291 1292 hdl->cmih_class = class; 1293 HDLOPS(hdl) = &cmi_hdl_ops; 1294 hdl->cmih_chipid = chipid; 1295 hdl->cmih_coreid = coreid; 1296 hdl->cmih_strandid = strandid; 1297 hdl->cmih_mstrand = cpu_is_cmt(priv); 1298 hdl->cmih_hdlpriv = priv; 1299 #ifdef __xpv 1300 hdl->cmih_msrsrc = CMI_MSR_FLAG_RD_INTERPOSEOK | 1301 CMI_MSR_FLAG_WR_INTERPOSEOK; 1302 1303 /* 1304 * XXX: need hypervisor support for procnodeid, for now assume 1305 * single-node processors (procnodeid = chipid) 1306 */ 1307 hdl->cmih_procnodeid = xen_physcpu_chipid((xen_mc_lcpu_cookie_t)priv); 1308 hdl->cmih_procnodes_per_pkg = 1; 1309 #else /* __xpv */ 1310 hdl->cmih_msrsrc = CMI_MSR_FLAG_RD_HWOK | CMI_MSR_FLAG_RD_INTERPOSEOK | 1311 CMI_MSR_FLAG_WR_HWOK | CMI_MSR_FLAG_WR_INTERPOSEOK; 1312 hdl->cmih_procnodeid = cpuid_get_procnodeid((cpu_t *)priv); 1313 hdl->cmih_procnodes_per_pkg = 1314 cpuid_get_procnodes_per_pkg((cpu_t *)priv); 1315 #endif /* __xpv */ 1316 1317 ent = cmi_hdl_ent_lookup(chipid, coreid, strandid); 1318 if (ent->cmae_refcnt != 0 || ent->cmae_hdlp != NULL) { 1319 /* 1320 * Somehow this (chipid, coreid, strandid) id tuple has 1321 * already been assigned! This indicates that the 1322 * callers logic in determining these values is busted, 1323 * or perhaps undermined by bad BIOS setup. Complain, 1324 * and refuse to initialize this tuple again as bad things 1325 * will happen. 1326 */ 1327 cmn_err(CE_NOTE, "cmi_hdl_create: chipid %d coreid %d " 1328 "strandid %d handle already allocated!", 1329 chipid, coreid, strandid); 1330 kmem_free(hdl, sizeof (*hdl)); 1331 return (NULL); 1332 } 1333 1334 /* 1335 * Once we store a nonzero reference count others can find this 1336 * handle via cmi_hdl_lookup etc. This initial hold on the handle 1337 * is to be dropped only if some other part of cmi initialization 1338 * fails or, if it succeeds, at later cpu deconfigure. Note the 1339 * the module private data we hold in cmih_cmi and cmih_cmidata 1340 * is still NULL at this point (the caller will fill it with 1341 * cmi_hdl_setcmi if it initializes) so consumers of handles 1342 * should always be ready for that possibility. 1343 */ 1344 ent->cmae_hdlp = hdl; 1345 hdl->cmih_refcntp = &ent->cmae_refcnt; 1346 ent->cmae_refcnt = 1; 1347 1348 return ((cmi_hdl_t)hdl); 1349 } 1350 1351 void 1352 cmi_read_smbios(cmi_hdl_t ophdl) 1353 { 1354 1355 uint_t strand_apicid = UINT_MAX; 1356 uint_t chip_inst = UINT_MAX; 1357 uint16_t smb_id = USHRT_MAX; 1358 int rc = 0; 1359 1360 cmi_hdl_impl_t *hdl = IMPLHDL(ophdl); 1361 1362 /* set x86gentopo compatibility */ 1363 fm_smb_fmacompat(); 1364 1365 #ifndef __xpv 1366 strand_apicid = ntv_strand_apicid(hdl); 1367 #else 1368 strand_apicid = xpv_strand_apicid(hdl); 1369 #endif 1370 1371 if (!x86gentopo_legacy) { 1372 /* 1373 * If fm_smb_chipinst() or fm_smb_bboard() fails, 1374 * topo reverts to legacy mode 1375 */ 1376 rc = fm_smb_chipinst(strand_apicid, &chip_inst, &smb_id); 1377 if (rc == 0) { 1378 hdl->cmih_smb_chipid = chip_inst; 1379 hdl->cmih_smbiosid = smb_id; 1380 } else { 1381 #ifdef DEBUG 1382 cmn_err(CE_NOTE, "!cmi reads smbios chip info failed"); 1383 #endif /* DEBUG */ 1384 return; 1385 } 1386 1387 hdl->cmih_smb_bboard = fm_smb_bboard(strand_apicid); 1388 #ifdef DEBUG 1389 if (hdl->cmih_smb_bboard == NULL) 1390 cmn_err(CE_NOTE, 1391 "!cmi reads smbios base boards info failed"); 1392 #endif /* DEBUG */ 1393 } 1394 } 1395 1396 void 1397 cmi_hdl_hold(cmi_hdl_t ophdl) 1398 { 1399 cmi_hdl_impl_t *hdl = IMPLHDL(ophdl); 1400 1401 ASSERT(*hdl->cmih_refcntp != 0); /* must not be the initial hold */ 1402 1403 atomic_inc_32(hdl->cmih_refcntp); 1404 } 1405 1406 static int 1407 cmi_hdl_canref(cmi_hdl_ent_t *ent) 1408 { 1409 volatile uint32_t *refcntp; 1410 uint32_t refcnt; 1411 1412 refcntp = &ent->cmae_refcnt; 1413 refcnt = *refcntp; 1414 1415 if (refcnt == 0) { 1416 /* 1417 * Associated object never existed, is being destroyed, 1418 * or has been destroyed. 1419 */ 1420 return (0); 1421 } 1422 1423 /* 1424 * We cannot use atomic increment here because once the reference 1425 * count reaches zero it must never be bumped up again. 1426 */ 1427 while (refcnt != 0) { 1428 if (atomic_cas_32(refcntp, refcnt, refcnt + 1) == refcnt) 1429 return (1); 1430 refcnt = *refcntp; 1431 } 1432 1433 /* 1434 * Somebody dropped the reference count to 0 after our initial 1435 * check. 1436 */ 1437 return (0); 1438 } 1439 1440 1441 void 1442 cmi_hdl_rele(cmi_hdl_t ophdl) 1443 { 1444 cmi_hdl_impl_t *hdl = IMPLHDL(ophdl); 1445 1446 ASSERT(*hdl->cmih_refcntp > 0); 1447 atomic_dec_32(hdl->cmih_refcntp); 1448 } 1449 1450 void 1451 cmi_hdl_destroy(cmi_hdl_t ophdl) 1452 { 1453 cmi_hdl_impl_t *hdl = IMPLHDL(ophdl); 1454 cmi_hdl_ent_t *ent; 1455 1456 /* Release the reference count held by cmi_hdl_create(). */ 1457 ASSERT(*hdl->cmih_refcntp > 0); 1458 atomic_dec_32(hdl->cmih_refcntp); 1459 hdl->cmih_flags |= CMIH_F_DEAD; 1460 1461 ent = cmi_hdl_ent_lookup(hdl->cmih_chipid, hdl->cmih_coreid, 1462 hdl->cmih_strandid); 1463 /* 1464 * Use busy polling instead of condition variable here because 1465 * cmi_hdl_rele() may be called from #MC handler. 1466 */ 1467 while (cmi_hdl_canref(ent)) { 1468 cmi_hdl_rele(ophdl); 1469 delay(1); 1470 } 1471 ent->cmae_hdlp = NULL; 1472 1473 kmem_free(hdl, sizeof (*hdl)); 1474 } 1475 1476 void 1477 cmi_hdl_setspecific(cmi_hdl_t ophdl, void *arg) 1478 { 1479 IMPLHDL(ophdl)->cmih_spec = arg; 1480 } 1481 1482 void * 1483 cmi_hdl_getspecific(cmi_hdl_t ophdl) 1484 { 1485 return (IMPLHDL(ophdl)->cmih_spec); 1486 } 1487 1488 void 1489 cmi_hdl_setmc(cmi_hdl_t ophdl, const struct cmi_mc_ops *mcops, void *mcdata) 1490 { 1491 cmi_hdl_impl_t *hdl = IMPLHDL(ophdl); 1492 1493 ASSERT(hdl->cmih_mcops == NULL && hdl->cmih_mcdata == NULL); 1494 hdl->cmih_mcops = mcops; 1495 hdl->cmih_mcdata = mcdata; 1496 } 1497 1498 const struct cmi_mc_ops * 1499 cmi_hdl_getmcops(cmi_hdl_t ophdl) 1500 { 1501 return (IMPLHDL(ophdl)->cmih_mcops); 1502 } 1503 1504 void * 1505 cmi_hdl_getmcdata(cmi_hdl_t ophdl) 1506 { 1507 return (IMPLHDL(ophdl)->cmih_mcdata); 1508 } 1509 1510 cmi_hdl_t 1511 cmi_hdl_lookup(enum cmi_hdl_class class, uint_t chipid, uint_t coreid, 1512 uint_t strandid) 1513 { 1514 cmi_hdl_ent_t *ent; 1515 1516 if (chipid > CMI_MAX_CHIPID || 1517 coreid > CMI_MAX_COREID(cmi_core_nbits) || 1518 strandid > CMI_MAX_STRANDID(cmi_strand_nbits)) 1519 return (NULL); 1520 1521 ent = cmi_hdl_ent_lookup(chipid, coreid, strandid); 1522 1523 if (class == CMI_HDL_NEUTRAL) 1524 #ifdef __xpv 1525 class = CMI_HDL_SOLARIS_xVM_MCA; 1526 #else 1527 class = CMI_HDL_NATIVE; 1528 #endif 1529 1530 if (!cmi_hdl_canref(ent)) 1531 return (NULL); 1532 1533 if (ent->cmae_hdlp->cmih_class != class) { 1534 cmi_hdl_rele((cmi_hdl_t)ent->cmae_hdlp); 1535 return (NULL); 1536 } 1537 1538 return ((cmi_hdl_t)ent->cmae_hdlp); 1539 } 1540 1541 cmi_hdl_t 1542 cmi_hdl_any(void) 1543 { 1544 int i, j; 1545 cmi_hdl_ent_t *ent; 1546 int max_strands = CMI_MAX_STRANDS_PER_CHIP(cmi_core_nbits, 1547 cmi_strand_nbits); 1548 1549 for (i = 0; i < CMI_CHIPID_ARR_SZ; i++) { 1550 if (cmi_chip_tab[i] == NULL) 1551 continue; 1552 for (j = 0, ent = cmi_chip_tab[i]; j < max_strands; 1553 j++, ent++) { 1554 if (cmi_hdl_canref(ent)) 1555 return ((cmi_hdl_t)ent->cmae_hdlp); 1556 } 1557 } 1558 1559 return (NULL); 1560 } 1561 1562 void 1563 cmi_hdl_walk(int (*cbfunc)(cmi_hdl_t, void *, void *, void *), 1564 void *arg1, void *arg2, void *arg3) 1565 { 1566 int i, j; 1567 cmi_hdl_ent_t *ent; 1568 int max_strands = CMI_MAX_STRANDS_PER_CHIP(cmi_core_nbits, 1569 cmi_strand_nbits); 1570 1571 for (i = 0; i < CMI_CHIPID_ARR_SZ; i++) { 1572 if (cmi_chip_tab[i] == NULL) 1573 continue; 1574 for (j = 0, ent = cmi_chip_tab[i]; j < max_strands; 1575 j++, ent++) { 1576 if (cmi_hdl_canref(ent)) { 1577 cmi_hdl_impl_t *hdl = ent->cmae_hdlp; 1578 if ((*cbfunc)((cmi_hdl_t)hdl, arg1, arg2, arg3) 1579 == CMI_HDL_WALK_DONE) { 1580 cmi_hdl_rele((cmi_hdl_t)hdl); 1581 return; 1582 } 1583 cmi_hdl_rele((cmi_hdl_t)hdl); 1584 } 1585 } 1586 } 1587 } 1588 1589 void 1590 cmi_hdl_setcmi(cmi_hdl_t ophdl, void *cmi, void *cmidata) 1591 { 1592 IMPLHDL(ophdl)->cmih_cmidata = cmidata; 1593 IMPLHDL(ophdl)->cmih_cmi = cmi; 1594 } 1595 1596 void * 1597 cmi_hdl_getcmi(cmi_hdl_t ophdl) 1598 { 1599 return (IMPLHDL(ophdl)->cmih_cmi); 1600 } 1601 1602 void * 1603 cmi_hdl_getcmidata(cmi_hdl_t ophdl) 1604 { 1605 return (IMPLHDL(ophdl)->cmih_cmidata); 1606 } 1607 1608 enum cmi_hdl_class 1609 cmi_hdl_class(cmi_hdl_t ophdl) 1610 { 1611 return (IMPLHDL(ophdl)->cmih_class); 1612 } 1613 1614 #define CMI_HDL_OPFUNC(what, type) \ 1615 type \ 1616 cmi_hdl_##what(cmi_hdl_t ophdl) \ 1617 { \ 1618 return (HDLOPS(IMPLHDL(ophdl))-> \ 1619 cmio_##what(IMPLHDL(ophdl))); \ 1620 } 1621 1622 /* BEGIN CSTYLED */ 1623 CMI_HDL_OPFUNC(vendor, uint_t) 1624 CMI_HDL_OPFUNC(vendorstr, const char *) 1625 CMI_HDL_OPFUNC(family, uint_t) 1626 CMI_HDL_OPFUNC(model, uint_t) 1627 CMI_HDL_OPFUNC(stepping, uint_t) 1628 CMI_HDL_OPFUNC(chipid, uint_t) 1629 CMI_HDL_OPFUNC(procnodeid, uint_t) 1630 CMI_HDL_OPFUNC(coreid, uint_t) 1631 CMI_HDL_OPFUNC(strandid, uint_t) 1632 CMI_HDL_OPFUNC(procnodes_per_pkg, uint_t) 1633 CMI_HDL_OPFUNC(strand_apicid, uint_t) 1634 CMI_HDL_OPFUNC(chiprev, x86_chiprev_t) 1635 CMI_HDL_OPFUNC(chiprevstr, const char *) 1636 CMI_HDL_OPFUNC(getsockettype, uint32_t) 1637 CMI_HDL_OPFUNC(getsocketstr, const char *) 1638 CMI_HDL_OPFUNC(logical_id, id_t) 1639 CMI_HDL_OPFUNC(smbiosid, uint16_t) 1640 CMI_HDL_OPFUNC(smb_chipid, uint_t) 1641 CMI_HDL_OPFUNC(smb_bboard, nvlist_t *) 1642 CMI_HDL_OPFUNC(chipsig, uint_t) 1643 /* END CSTYLED */ 1644 1645 boolean_t 1646 cmi_hdl_is_cmt(cmi_hdl_t ophdl) 1647 { 1648 return (IMPLHDL(ophdl)->cmih_mstrand); 1649 } 1650 1651 void 1652 cmi_hdl_int(cmi_hdl_t ophdl, int num) 1653 { 1654 if (HDLOPS(IMPLHDL(ophdl))->cmio_int == NULL) 1655 return; 1656 1657 cmi_hdl_inj_begin(ophdl); 1658 HDLOPS(IMPLHDL(ophdl))->cmio_int(IMPLHDL(ophdl), num); 1659 cmi_hdl_inj_end(NULL); 1660 } 1661 1662 int 1663 cmi_hdl_online(cmi_hdl_t ophdl, int new_status, int *old_status) 1664 { 1665 return (HDLOPS(IMPLHDL(ophdl))->cmio_online(IMPLHDL(ophdl), 1666 new_status, old_status)); 1667 } 1668 1669 #ifndef __xpv 1670 /* 1671 * Return hardware chip instance; cpuid_get_chipid provides this directly. 1672 */ 1673 uint_t 1674 cmi_ntv_hwchipid(cpu_t *cp) 1675 { 1676 return (cpuid_get_chipid(cp)); 1677 } 1678 1679 /* 1680 * Return hardware node instance; cpuid_get_procnodeid provides this directly. 1681 */ 1682 uint_t 1683 cmi_ntv_hwprocnodeid(cpu_t *cp) 1684 { 1685 return (cpuid_get_procnodeid(cp)); 1686 } 1687 1688 /* 1689 * Return core instance within a single chip. 1690 */ 1691 uint_t 1692 cmi_ntv_hwcoreid(cpu_t *cp) 1693 { 1694 return (cpuid_get_pkgcoreid(cp)); 1695 } 1696 1697 /* 1698 * Return strand number within a single core. cpuid_get_clogid numbers 1699 * all execution units (strands, or cores in unstranded models) sequentially 1700 * within a single chip. 1701 */ 1702 uint_t 1703 cmi_ntv_hwstrandid(cpu_t *cp) 1704 { 1705 int strands_per_core = cpuid_get_ncpu_per_chip(cp) / 1706 cpuid_get_ncore_per_chip(cp); 1707 1708 return (cpuid_get_clogid(cp) % strands_per_core); 1709 } 1710 1711 static void 1712 cmi_ntv_hwdisable_mce_xc(void) 1713 { 1714 ulong_t cr4; 1715 1716 cr4 = getcr4(); 1717 cr4 = cr4 & (~CR4_MCE); 1718 setcr4(cr4); 1719 } 1720 1721 void 1722 cmi_ntv_hwdisable_mce(cmi_hdl_t hdl) 1723 { 1724 cpuset_t set; 1725 cmi_hdl_impl_t *thdl = IMPLHDL(hdl); 1726 cpu_t *cp = HDLPRIV(thdl); 1727 1728 if (CPU->cpu_id == cp->cpu_id) { 1729 cmi_ntv_hwdisable_mce_xc(); 1730 } else { 1731 CPUSET_ONLY(set, cp->cpu_id); 1732 xc_call(0, 0, 0, CPUSET2BV(set), 1733 (xc_func_t)cmi_ntv_hwdisable_mce_xc); 1734 } 1735 } 1736 1737 #endif /* __xpv */ 1738 1739 void 1740 cmi_hdlconf_rdmsr_nohw(cmi_hdl_t ophdl) 1741 { 1742 cmi_hdl_impl_t *hdl = IMPLHDL(ophdl); 1743 1744 hdl->cmih_msrsrc &= ~CMI_MSR_FLAG_RD_HWOK; 1745 } 1746 1747 void 1748 cmi_hdlconf_wrmsr_nohw(cmi_hdl_t ophdl) 1749 { 1750 cmi_hdl_impl_t *hdl = IMPLHDL(ophdl); 1751 1752 hdl->cmih_msrsrc &= ~CMI_MSR_FLAG_WR_HWOK; 1753 } 1754 1755 cmi_errno_t 1756 cmi_hdl_rdmsr(cmi_hdl_t ophdl, uint_t msr, uint64_t *valp) 1757 { 1758 cmi_hdl_impl_t *hdl = IMPLHDL(ophdl); 1759 1760 /* 1761 * Regardless of the handle class, we first check for am 1762 * interposed value. In the xVM case you probably want to 1763 * place interposed values within the hypervisor itself, but 1764 * we still allow interposing them in dom0 for test and bringup 1765 * purposes. 1766 */ 1767 if ((hdl->cmih_msrsrc & CMI_MSR_FLAG_RD_INTERPOSEOK) && 1768 msri_lookup(hdl, msr, valp)) 1769 return (CMI_SUCCESS); 1770 1771 if (HDLOPS(hdl)->cmio_rdmsr == NULL) 1772 return (CMIERR_NOTSUP); 1773 1774 return (HDLOPS(hdl)->cmio_rdmsr(hdl, msr, valp)); 1775 } 1776 1777 cmi_errno_t 1778 cmi_hdl_wrmsr(cmi_hdl_t ophdl, uint_t msr, uint64_t val) 1779 { 1780 cmi_hdl_impl_t *hdl = IMPLHDL(ophdl); 1781 1782 /* Invalidate any interposed value */ 1783 msri_rment(hdl, msr); 1784 1785 if (HDLOPS(hdl)->cmio_wrmsr == NULL) 1786 return (CMI_SUCCESS); /* pretend all is ok */ 1787 1788 return (HDLOPS(hdl)->cmio_wrmsr(hdl, msr, val)); 1789 } 1790 1791 void 1792 cmi_hdl_enable_mce(cmi_hdl_t ophdl) 1793 { 1794 cmi_hdl_impl_t *hdl = IMPLHDL(ophdl); 1795 ulong_t cr4; 1796 1797 if (HDLOPS(hdl)->cmio_getcr4 == NULL || 1798 HDLOPS(hdl)->cmio_setcr4 == NULL) 1799 return; 1800 1801 cr4 = HDLOPS(hdl)->cmio_getcr4(hdl); 1802 1803 HDLOPS(hdl)->cmio_setcr4(hdl, cr4 | CR4_MCE); 1804 } 1805 1806 void 1807 cmi_hdl_msrinterpose(cmi_hdl_t ophdl, cmi_mca_regs_t *regs, uint_t nregs) 1808 { 1809 cmi_hdl_impl_t *hdl = IMPLHDL(ophdl); 1810 int i; 1811 1812 if (HDLOPS(hdl)->cmio_msrinterpose == NULL) 1813 return; 1814 1815 cmi_hdl_inj_begin(ophdl); 1816 1817 for (i = 0; i < nregs; i++, regs++) 1818 HDLOPS(hdl)->cmio_msrinterpose(hdl, regs->cmr_msrnum, 1819 regs->cmr_msrval); 1820 1821 cmi_hdl_inj_end(ophdl); 1822 } 1823 1824 /*ARGSUSED*/ 1825 void 1826 cmi_hdl_msrforward(cmi_hdl_t ophdl, cmi_mca_regs_t *regs, uint_t nregs) 1827 { 1828 #ifdef __xpv 1829 cmi_hdl_impl_t *hdl = IMPLHDL(ophdl); 1830 int i; 1831 1832 for (i = 0; i < nregs; i++, regs++) 1833 msri_addent(hdl, regs->cmr_msrnum, regs->cmr_msrval); 1834 #endif 1835 } 1836 1837 1838 void 1839 cmi_pcird_nohw(void) 1840 { 1841 cmi_pcicfg_flags &= ~CMI_PCICFG_FLAG_RD_HWOK; 1842 } 1843 1844 void 1845 cmi_pciwr_nohw(void) 1846 { 1847 cmi_pcicfg_flags &= ~CMI_PCICFG_FLAG_WR_HWOK; 1848 } 1849 1850 static uint32_t 1851 cmi_pci_get_cmn(int bus, int dev, int func, int reg, int asz, 1852 int *interpose, ddi_acc_handle_t hdl) 1853 { 1854 uint32_t val; 1855 1856 if (cmi_pcicfg_flags & CMI_PCICFG_FLAG_RD_INTERPOSEOK && 1857 pcii_lookup(bus, dev, func, reg, asz, &val)) { 1858 if (interpose) 1859 *interpose = 1; 1860 return (val); 1861 } 1862 if (interpose) 1863 *interpose = 0; 1864 1865 if (!(cmi_pcicfg_flags & CMI_PCICFG_FLAG_RD_HWOK)) 1866 return (0); 1867 1868 switch (asz) { 1869 case 1: 1870 if (hdl) 1871 val = pci_config_get8(hdl, (off_t)reg); 1872 else 1873 val = pci_cfgacc_get8(NULL, PCI_GETBDF(bus, dev, func), 1874 reg); 1875 break; 1876 case 2: 1877 if (hdl) 1878 val = pci_config_get16(hdl, (off_t)reg); 1879 else 1880 val = pci_cfgacc_get16(NULL, PCI_GETBDF(bus, dev, func), 1881 reg); 1882 break; 1883 case 4: 1884 if (hdl) 1885 val = pci_config_get32(hdl, (off_t)reg); 1886 else 1887 val = pci_cfgacc_get32(NULL, PCI_GETBDF(bus, dev, func), 1888 reg); 1889 break; 1890 default: 1891 val = 0; 1892 } 1893 return (val); 1894 } 1895 1896 uint8_t 1897 cmi_pci_getb(int bus, int dev, int func, int reg, int *interpose, 1898 ddi_acc_handle_t hdl) 1899 { 1900 return ((uint8_t)cmi_pci_get_cmn(bus, dev, func, reg, 1, interpose, 1901 hdl)); 1902 } 1903 1904 uint16_t 1905 cmi_pci_getw(int bus, int dev, int func, int reg, int *interpose, 1906 ddi_acc_handle_t hdl) 1907 { 1908 return ((uint16_t)cmi_pci_get_cmn(bus, dev, func, reg, 2, interpose, 1909 hdl)); 1910 } 1911 1912 uint32_t 1913 cmi_pci_getl(int bus, int dev, int func, int reg, int *interpose, 1914 ddi_acc_handle_t hdl) 1915 { 1916 return (cmi_pci_get_cmn(bus, dev, func, reg, 4, interpose, hdl)); 1917 } 1918 1919 void 1920 cmi_pci_interposeb(int bus, int dev, int func, int reg, uint8_t val) 1921 { 1922 pcii_addent(bus, dev, func, reg, val, 1); 1923 } 1924 1925 void 1926 cmi_pci_interposew(int bus, int dev, int func, int reg, uint16_t val) 1927 { 1928 pcii_addent(bus, dev, func, reg, val, 2); 1929 } 1930 1931 void 1932 cmi_pci_interposel(int bus, int dev, int func, int reg, uint32_t val) 1933 { 1934 pcii_addent(bus, dev, func, reg, val, 4); 1935 } 1936 1937 static void 1938 cmi_pci_put_cmn(int bus, int dev, int func, int reg, int asz, 1939 ddi_acc_handle_t hdl, uint32_t val) 1940 { 1941 /* 1942 * If there is an interposed value for this register invalidate it. 1943 */ 1944 pcii_rment(bus, dev, func, reg, asz); 1945 1946 if (!(cmi_pcicfg_flags & CMI_PCICFG_FLAG_WR_HWOK)) 1947 return; 1948 1949 switch (asz) { 1950 case 1: 1951 if (hdl) 1952 pci_config_put8(hdl, (off_t)reg, (uint8_t)val); 1953 else 1954 pci_cfgacc_put8(NULL, PCI_GETBDF(bus, dev, func), reg, 1955 (uint8_t)val); 1956 break; 1957 1958 case 2: 1959 if (hdl) 1960 pci_config_put16(hdl, (off_t)reg, (uint16_t)val); 1961 else 1962 pci_cfgacc_put16(NULL, PCI_GETBDF(bus, dev, func), reg, 1963 (uint16_t)val); 1964 break; 1965 1966 case 4: 1967 if (hdl) 1968 pci_config_put32(hdl, (off_t)reg, val); 1969 else 1970 pci_cfgacc_put32(NULL, PCI_GETBDF(bus, dev, func), reg, 1971 val); 1972 break; 1973 1974 default: 1975 break; 1976 } 1977 } 1978 1979 void 1980 cmi_pci_putb(int bus, int dev, int func, int reg, ddi_acc_handle_t hdl, 1981 uint8_t val) 1982 { 1983 cmi_pci_put_cmn(bus, dev, func, reg, 1, hdl, val); 1984 } 1985 1986 void 1987 cmi_pci_putw(int bus, int dev, int func, int reg, ddi_acc_handle_t hdl, 1988 uint16_t val) 1989 { 1990 cmi_pci_put_cmn(bus, dev, func, reg, 2, hdl, val); 1991 } 1992 1993 void 1994 cmi_pci_putl(int bus, int dev, int func, int reg, ddi_acc_handle_t hdl, 1995 uint32_t val) 1996 { 1997 cmi_pci_put_cmn(bus, dev, func, reg, 4, hdl, val); 1998 } 1999 2000 static const struct cmi_hdl_ops cmi_hdl_ops = { 2001 #ifdef __xpv 2002 /* 2003 * CMI_HDL_SOLARIS_xVM_MCA - ops when we are an xVM dom0 2004 */ 2005 xpv_vendor, /* cmio_vendor */ 2006 xpv_vendorstr, /* cmio_vendorstr */ 2007 xpv_family, /* cmio_family */ 2008 xpv_model, /* cmio_model */ 2009 xpv_stepping, /* cmio_stepping */ 2010 xpv_chipid, /* cmio_chipid */ 2011 xpv_procnodeid, /* cmio_procnodeid */ 2012 xpv_coreid, /* cmio_coreid */ 2013 xpv_strandid, /* cmio_strandid */ 2014 xpv_procnodes_per_pkg, /* cmio_procnodes_per_pkg */ 2015 xpv_strand_apicid, /* cmio_strand_apicid */ 2016 xpv_chiprev, /* cmio_chiprev */ 2017 xpv_chiprevstr, /* cmio_chiprevstr */ 2018 xpv_getsockettype, /* cmio_getsockettype */ 2019 xpv_getsocketstr, /* cmio_getsocketstr */ 2020 xpv_chipsig, /* cmio_chipsig */ 2021 xpv_logical_id, /* cmio_logical_id */ 2022 NULL, /* cmio_getcr4 */ 2023 NULL, /* cmio_setcr4 */ 2024 xpv_rdmsr, /* cmio_rdmsr */ 2025 xpv_wrmsr, /* cmio_wrmsr */ 2026 xpv_msrinterpose, /* cmio_msrinterpose */ 2027 xpv_int, /* cmio_int */ 2028 xpv_online, /* cmio_online */ 2029 xpv_smbiosid, /* cmio_smbiosid */ 2030 xpv_smb_chipid, /* cmio_smb_chipid */ 2031 xpv_smb_bboard /* cmio_smb_bboard */ 2032 2033 #else /* __xpv */ 2034 2035 /* 2036 * CMI_HDL_NATIVE - ops when apparently running on bare-metal 2037 */ 2038 ntv_vendor, /* cmio_vendor */ 2039 ntv_vendorstr, /* cmio_vendorstr */ 2040 ntv_family, /* cmio_family */ 2041 ntv_model, /* cmio_model */ 2042 ntv_stepping, /* cmio_stepping */ 2043 ntv_chipid, /* cmio_chipid */ 2044 ntv_procnodeid, /* cmio_procnodeid */ 2045 ntv_coreid, /* cmio_coreid */ 2046 ntv_strandid, /* cmio_strandid */ 2047 ntv_procnodes_per_pkg, /* cmio_procnodes_per_pkg */ 2048 ntv_strand_apicid, /* cmio_strand_apicid */ 2049 ntv_chiprev, /* cmio_chiprev */ 2050 ntv_chiprevstr, /* cmio_chiprevstr */ 2051 ntv_getsockettype, /* cmio_getsockettype */ 2052 ntv_getsocketstr, /* cmio_getsocketstr */ 2053 ntv_chipsig, /* cmio_chipsig */ 2054 ntv_logical_id, /* cmio_logical_id */ 2055 ntv_getcr4, /* cmio_getcr4 */ 2056 ntv_setcr4, /* cmio_setcr4 */ 2057 ntv_rdmsr, /* cmio_rdmsr */ 2058 ntv_wrmsr, /* cmio_wrmsr */ 2059 ntv_msrinterpose, /* cmio_msrinterpose */ 2060 ntv_int, /* cmio_int */ 2061 ntv_online, /* cmio_online */ 2062 ntv_smbiosid, /* cmio_smbiosid */ 2063 ntv_smb_chipid, /* cmio_smb_chipid */ 2064 ntv_smb_bboard /* cmio_smb_bboard */ 2065 #endif 2066 }; 2067